Role: Technical Program Manager
Function: Technical Program Management
Location: Bengaluru, India
Industry: AI infrastructure, Cloud Computing
About Company
This role is with a rapidly growing AI infrastructure startup founded in 2025 in Bengaluru by a leadership team with deep product, cloud, and systems experience from global-scale tech companies. The company has built a GenAI-powered private cloud platform that automates and manages complex AI workloads across hybrid, on-prem, edge, and sovereign cloud environments — designed for enterprise sectors where performance, data security, and compliance are critical. Backed by leading global VCs and prominent operators (approx. $10M seed raised), the company is recognized for strong engineering rigor and product clarity. Its platform focuses on AI-native orchestration, deep observability, and cost/performance optimization to help large enterprises deploy and scale AI with confidence. This is an opportunity to join early and shape the future of AI-first cloud infrastructure.
Position Overview
We are hiring TPMs across Storage, Platform/SRE, GPU Compute & AI Inference
We’re looking for Technical Program Managers to lead some of the most critical programs powering our AI-native cloud. Whether your background is in storage, platform engineering/SRE, GPU compute, or AI/ML infrastructure, we have a high-impact ownership area for you.
If you are strong in any one of these verticals, you are encouraged to apply.
Role & Responsibilities
1. Storage, Data Foundation & CDN Programs
- Lead programs for object & file storage, CDN delivery, and capacity expansion.
- Drive data-layer services: Kafka, Spark equivalents, vector stores, data lakes.
- Build scalable, reliable data foundations for AI workloads.
2. Platform Engineering, SRE & Security
- Run programs for observability: logging, metrics, traces, dashboards, alerts.
- Lead security & identity initiatives: RBAC, SSO, key management, audit trails.
- Standardize incident management: SLOs, runbooks, PIRs, rollout strategies.
- Partner with SRE on infrastructure-as-code modules, shared environments, drift detection.
3. GPU Compute, Virtualization & High-Performance Workloads
- Lead programs for vGPU, GPU partitioning, and cluster bootstrapping.
- Drive features for VM templates, accelerated compute, and add-on codification.
- Support high-performance AI workloads through strong infra governance.
4. AI Inference Platform, Model Registry & Developer Tooling
- Drive programs for the inference platform, model registry, and developer tools
- Translate AI platform goals into clear roadmaps, milestones, and success metrics.
- Track adoption, performance, and reliability of AI-serving capabilities.
Must have Criteria
- 4–5 years in Technical Program Management or adjacent roles (infra, SRE, backend, ML, platform).
- Strong understanding of cloud infrastructure, Kubernetes, containers, and CI/CD.
- Hands-on experience with Infrastructure-as-Code (Terraform or similar).
- Familiarity with one or more areas:
- Storage / data platforms
- Platform/SRE / security
- GPU compute / virtualization
- AI/ML workflows / inference
- Bachelor’s degree in CS, Engineering, or equivalent practical experience.
Nice to Have
- Experience with MLOps, model serving, or inference systems.
- Exposure to GPUaaS, HPC, or distributed compute platforms.
- Familiarity with data lakes, ETL, analytics, or vector databases.
- Background in high-growth, fast-scaling, or distributed teams.
- Experience with service meshes, API gateways, or developer platforms.
Apply Now
Share your details below to apply for this job.
Job Description
Role: Technical Program Manager
Function: Technical Program Management
Location: Bengaluru, India
Industry: AI infrastructure, Cloud Computing
About Company
This role is with a rapidly growing AI infrastructure startup founded in 2025 in Bengaluru by a leadership team with deep product, cloud, and systems experience from global-scale tech companies. The company has built a GenAI-powered private cloud platform that automates and manages complex AI workloads across hybrid, on-prem, edge, and sovereign cloud environments — designed for enterprise sectors where performance, data security, and compliance are critical. Backed by leading global VCs and prominent operators (approx. $10M seed raised), the company is recognized for strong engineering rigor and product clarity. Its platform focuses on AI-native orchestration, deep observability, and cost/performance optimization to help large enterprises deploy and scale AI with confidence. This is an opportunity to join early and shape the future of AI-first cloud infrastructure.
Position Overview
We are hiring TPMs across Storage, Platform/SRE, GPU Compute & AI Inference
We’re looking for Technical Program Managers to lead some of the most critical programs powering our AI-native cloud. Whether your background is in storage, platform engineering/SRE, GPU compute, or AI/ML infrastructure, we have a high-impact ownership area for you.
If you are strong in any one of these verticals, you are encouraged to apply.
Role & Responsibilities
1. Storage, Data Foundation & CDN Programs
- Lead programs for object & file storage, CDN delivery, and capacity expansion.
- Drive data-layer services: Kafka, Spark equivalents, vector stores, data lakes.
- Build scalable, reliable data foundations for AI workloads.
2. Platform Engineering, SRE & Security
- Run programs for observability: logging, metrics, traces, dashboards, alerts.
- Lead security & identity initiatives: RBAC, SSO, key management, audit trails.
- Standardize incident management: SLOs, runbooks, PIRs, rollout strategies.
- Partner with SRE on infrastructure-as-code modules, shared environments, drift detection.
3. GPU Compute, Virtualization & High-Performance Workloads
- Lead programs for vGPU, GPU partitioning, and cluster bootstrapping.
- Drive features for VM templates, accelerated compute, and add-on codification.
- Support high-performance AI workloads through strong infra governance.
4. AI Inference Platform, Model Registry & Developer Tooling
- Drive programs for the inference platform, model registry, and developer tools
- Translate AI platform goals into clear roadmaps, milestones, and success metrics.
- Track adoption, performance, and reliability of AI-serving capabilities.
Must have Criteria
- 4–5 years in Technical Program Management or adjacent roles (infra, SRE, backend, ML, platform).
- Strong understanding of cloud infrastructure, Kubernetes, containers, and CI/CD.
- Hands-on experience with Infrastructure-as-Code (Terraform or similar).
- Familiarity with one or more areas:
- Storage / data platforms
- Platform/SRE / security
- GPU compute / virtualization
- AI/ML workflows / inference
- Bachelor’s degree in CS, Engineering, or equivalent practical experience.
Nice to Have
- Experience with MLOps, model serving, or inference systems.
- Exposure to GPUaaS, HPC, or distributed compute platforms.
- Familiarity with data lakes, ETL, analytics, or vector databases.
- Background in high-growth, fast-scaling, or distributed teams.
- Experience with service meshes, API gateways, or developer platforms.
Apply Now
Share your details below to apply for this job.
Application Submitted Successfully!
Thank you for applying to Technical Program Manager. We have received your application and will review it shortly.
You will be redirected shortly...