Nebius logo

Nebius

Senior Technical Product Manager - Serverless AI at Nebius

Amsterdam, Netherlands; Remote - Europe; Remote - United StatesFull-timeRemoteTechnology ProductPosted 7 days ago

About the Role

<div class="content-intro"><p><strong>About Nebius:</strong></p> <p>Nebius is leading a new era in cloud infrastructure for the global AI economy. We are building a full-stack AI cloud platform that supports developers and enterprises from data and model training through to production deployment, without the cost and complexity of building large in-house AI/ML infrastructure.</p> <p>Built by engineers, for engineers. From large-scale GPU orchestration to inference optimization, we own the hard problems across compute, storage, networking and applied AI.</p> <p>Listed on Nasdaq (NBIS) and headquartered in Amsterdam, we have a global footprint with R&amp;D hubs across Europe, the UK, North America and Israel. Our team of 1,500+ includes hundreds of engineers with deep expertise across hardware, software and AI R&amp;D.</p></div><div><strong>The role</strong></div> <div>&nbsp;</div> <div>Nebius Serverless AI is our consumption-based compute platform for running AI workloads — training jobs, inference endpoints, and interactive development environments — without managing infrastructure. Users submit containerized workloads via CLI or UI, access GPU compute with pay-per-second billing, and the platform handles provisioning, lifecycle, and cleanup. We launched GA in Q1 2026 and are now scaling toward 1,000+ users while building the next generation of capabilities: autoscaling, multi-node distributed workloads, and developer-first tooling.</div> <div><br>We are looking for a Senior Technical Product Manager to join the Serverless AI product team. Together you will divide ownership across the product surface — but individually, you will own your areas with full autonomy. This is not a role where you write requirements and hand them off. You will be the person who understands container runtimes, GPU scheduling, cold start optimization, and inference serving deeply enough to make correct technical trade-offs — and also the person who talks to customers, shapes the CLI experience, defines pricing, and drives adoption.</div> <div><br>We are building the next generation of AI cloud — infrastructure designed from the ground up for GPU-intensive workloads, not retrofitted from legacy cloud. This is a lean, high-impact team where every person shapes the product directly. You need to be the kind of PM who amplifies engineering output by making the right calls on what to build and what to skip.</div> <div><br><strong>What success looks like in 12 months:</strong></div> <ul> <li>Serverless AI has clear product-market fit with measurable activation and retention metrics improving quarter over quarter.</li> <li>Multi-node jobs and autoscaling endpoints are shipped and adopted by customers running production workloads.</li> <li>Cold start time is reduced from 1-3 minutes to under 60 seconds for common workloads through a combination of product and infrastructure improvements you drove.</li> <li>Developer experience (CLI, docs, error messages, onboarding flow) sets the standard that developers expect from a next-generation AI cloud.</li> <li>At least 3 product decisions you made are directly attributable to customer conversations or data analysis you conducted.</li> </ul> <div><strong>Your responsibilities will include:</strong></div> <div><br><strong>1. Product Ownership</strong></div> <ul> <li>Co-own the Serverless AI product roadmap — Jobs, Endpoints, and DevPods — taking primary ownership of specific product areas while collaborating closely with the other PM on shared priorities and cross-cutting decisions.</li> <li>Write detailed, technically precise PRDs that engineering teams can execute against. Our PRDs specify CLI syntax, API contracts, state machines, and billing models — not abstract feature descriptions.</li> <li>Make build/buy/defer decisions on capabilities like autoscaling, multi-node orchestration, HTTPS termination, secret injection, and health checking based on customer signal and strategic priorities.</li> </ul> <div><strong>2. Technical Depth:</strong></div> <ul> <li>Understand the full workload lifecycle: container image pull → VM provisioning → GPU attachment → workload execution → cleanup — well enough to identify bottlenecks and propose solutions.</li> <li>Evaluate technical trade-offs in areas like container cold start optimization (image caching, snapshot restore, warm pools), GPU scheduling and bin-packing, and storage mount performance.</li> <li>Work directly with engineers on architecture decisions for distributed training support, endpoint autoscaling policies, and fault tolerance mechanisms.</li> <li>Stay current on the fast-moving serverless GPU infrastructure space — new inference frameworks (vLLM, TensorRT-LLM, SGLang), container runtimes, orchestration approaches — and translate trends into product direction.</li> </ul> <div><strong>3. Customer &amp; Market:</strong></div> <ul> <li>Run customer discovery and feedback sessions with ML engineers and platform teams at AI startups and enterprises. Turn qualitative insight into specific product actions.</li> <li>Analyze usage data, activation funnels, and churn patterns to identify where users get stuck and what features drive retention.</li> <li>Track market dynamics, emerging technologies, and industry trends to inform product strategy and ensure Nebius stays ahead of where the market is heading.</li> <li>Define and iterate on pricing, packaging, and tier strategy for Serverless AI.</li> </ul> <div><strong>4. Go-to-Market:</strong></div> <ul> <li>Own the technical content strategy: quickstart guides, tutorials, reference architectures, and example workloads that reduce time-to-first-job.</li> <li>Partner with marketing on developer-focused campaigns, webinars, and conference presence.</li> <li>Work with Solution Architects and Sales to qualify serverless-fit opportunities and support technical evaluations.</li> </ul> <div><strong>Requirements</strong></div> <div><br><strong>Non-negotiables — you must have hands-on experience with:</strong></div> <ul> <li>You have built, shipped, and iterated on infrastructure or platform products used by developers or ML engineers. Not consumer apps. Not dashboards. Infrastructure.</li> <li>You understand containers at a practical level — Docker, image registries, container runtimes, resource limits, networking. You've debugged why a container won't start, why GPU isn't visible inside it, or why a mount isn't working.</li> <li>You have working knowledge of GPU computing for AI/ML: what GPU types exist and when to use them, how training and inference workloads differ in resource requirements, what vLLM / TensorRT-LLM / Triton are and why they matter.</li> <li>You can read a CLI reference and know if it's well-designed. You've shaped developer-facing APIs, CLIs, or SDKs.</li> <li>You have run real customer discovery — not surveys, but technical conversations with engineers where you learned something that changed your product direction.</li> <li>You have 3+ years of product management experience in cloud infrastructure, AI/ML platforms, or developer tools.</li> </ul> <div><strong>Technical skills we will test in the interview:</strong></div> <ul> <li>Ability to whiteboard a workload lifecycle (submit → schedule → provision → execute → cleanup) and identify failure modes at each step.</li> <li>Understanding of autoscaling trade-offs: scale-to-zero vs. warm pools, scaling metrics (queue depth, latency, utilization), cold start implications.</li> <li>Familiarity with inference serving concepts: batching, model loading, quantization, KV-cache management, multi-model serving.</li> <li>Understanding of distributed training concepts: data parallelism, model parallelism, communication overhead, checkpointing.</li> <li>Ability to reason about pricing models: per-second vs. per-request vs. per-token, and how pricing interacts with product architecture.</li> </ul> <div><strong>It will be an added bonus if you have:</strong></div> <ul> <li>Experience at a serverless or GPU cloud company.</li> <li>Hands-on ML engineering background — you've trained models, deployed inference endpoints, or built ML pipelines yourself.</li> <li>Experience with Kubernetes for ML workloads (Kubeflow, KServe, Ray Serve) and understanding of why many ML teams want to avoid it.</li> <li>Prior experience building a product from early stage to scale in a fast-growing market.</li> <li>Background in systems engineering, distributed systems, or site reliability engineering.</li> </ul> <div><strong>Who thrives in this role</strong></div> <ul> <li>You are more comfortable in a terminal than in a slide deck.</li> <li>You form strong opinions based on data and direct customer signal, and you update them when evidence changes.</li> <li>You are energized by building at pace — small team, fast-evolving product, big opportunity.</li> <li>You care about developer experience at the level of error messages, CLI flag naming, and documentation quality.</li> <li>You'd rather ship a smaller thing that works perfectly than a bigger thing that's mediocre.</li> </ul> <p><strong>About Nebius</strong></p> <p><span data-contrast="auto">Nebius AI is an AI cloud platform with one of the largest GPU capacities in Europe. Launched in November 2023, the Nebius AI platform provides high-end, training-optimized infrastructure for AI practitioners. As an NVIDIA preferred cloud service provider, Nebius AI offers a variety of NVIDIA GPUs for training and inference, as well as a set of tools for efficient multi-node training.</span><span data-ccp-props="{}">&nbsp;</span></p> <p><span data-contrast="auto">Nebius AI owns a data center in Finland, built from the ground up by the company’s R&amp;D team and showcasing our commitment to sustainability. The data center is home to ISEG, the most powerful commercially available supercomputer in Europe and the 16th most powerful globally (Top 500 list, November 2023).&nbsp;</span><span data-ccp-props="{}">&nbsp;</span></p> <p><span data-contrast="auto">Nebius’s headquarters are in Amsterdam, Netherlands, with teams working out of R&amp;D hubs across Europe and the Middle East.</span><span data-ccp-props="{}">&nbsp;</span></p> <p><span data-contrast="auto">Nebius AI is built with the talent of more than 500 highly skilled engineers with a proven track record in developing sophisticated cloud and ML solutions and designing cutting-edge hardware. This allows all the layers of the Nebius AI cloud – from hardware to UI – to be built in-house, distictly differentiating Nebius AI from the majority of specialized clouds: Nebius customers get a true hyperscaler-cloud experience tailored for AI practitioners. We’re growing and expanding our products every day.&nbsp;</span></p> <p><span data-ccp-props="{}">If you’re up to the challenge and are excited about AI and ML as much as we are, join us!</span></p><div class="content-conclusion"><p><strong>Benefits &amp; Perks:</strong></p> <ul> <li>Competitive compensation</li> <li>Career growth and learning opportunities</li> <li>Flexibility and work-life balance</li> <li>Collaborative and innovative culture</li> <li>Opportunity to work on impactful AI projects</li> <li>International environment and talented teams</li> </ul> <p><strong>What's it like to work at Nebius:</strong></p> <p>Fast moving&nbsp;- Bold thinking&nbsp;- Constant growth&nbsp;- Meaningful impact&nbsp;- Trust and real ownership&nbsp;- Opportunity to shape the future of AI&nbsp;</p> <p><strong>Equal Opportunity Statement:</strong></p> <p>Nebius is an equal opportunity employer. We are committed to fostering an inclusive and diverse workplace and to providing equal employment opportunities in all aspects of employment. We do not discriminate on the basis of race, color, religion, sex (including pregnancy), national origin, ancestry, age, disability, genetic information, marital status, veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by applicable law.</p> <p>Applicants must be authorized to work in the country in which they apply and will be required to provide proof of employment eligibility as a condition of hire.&nbsp;</p> <p>If you need accommodations during the application process, please let us know.</p></div>