Nuro logo

Nuro

Program Manager, Data Quality at Nuro

Mountain View, California (HQ)Full-timeSoftware OperationsPosted about 4 hours ago

About the Role

<h2><strong>Who We Are</strong></h2> <p>Nuro is a self-driving technology company on a mission to make autonomy accessible to all. Founded in 2016, Nuro is building the world’s most scalable driver, combining cutting-edge AI with automotive-grade hardware. Nuro licenses its core technology, the Nuro Driver™, to support a wide range of applications, from robotaxis and commercial fleets to personally owned vehicles. With technology proven over years of self-driving deployments, Nuro gives the automakers and mobility platforms a clear path to AVs at commercial scale, empowering a safer, richer, and more connected future.</p> <h2><strong>About the Team</strong></h2> <p><em>Every vehicle that drives itself on a public road starts with a decision someone made about a single annotation. At Nuro, we take that seriously — and we're looking for someone who does too.</em></p> <p>ML Operations is where data becomes decisions. We sit at the intersection of Autonomy Engineering and the globally distributed annotation teams who label the world our vehicles need to understand. We are builders, diagnosticians, and systems thinkers who care deeply about the end product - because we know that what we fix here shows up on the road.&nbsp;</p> <p>This team is collaborative and technically rigorous. We work closely with engineering and we hold each other to a high standard, but we also invest in getting things right together. Here is what the work actually looks like day-to-day:&nbsp;</p> <ul> <li>Investigate annotation faults by querying databases directly to identify root causes, moving beyond reliance on ticket queues for initial troubleshooting.</li> <li>Sitting with Autonomy Engineering leadership, translating a model accuracy regression into a specific labeling workflow problem and walking in with the fix already scoped.</li> <li>Auditing a live pipeline, spotting a systematic edge-case misclassification that has been silently propagating, and building the evidence-based case to restructure how that object type gets labeled.</li> <li>Manage the processes for quality management/processes for the offshore annotation team</li> <li>Being the bridge: the person ML Operations leans on for quality diagnostics, and the person Engineering trusts to understand how a labeling decision upstream shapes what the model learns downstream.</li> </ul> <p>If that sounds like the kind of problem-solving that lights you up, we would love to talk.</p> <h2><strong>About the Role</strong></h2> <p>This is a rare opportunity to own something that genuinely matters. You will co-own a portfolio of active labeling pipelines alongside senior leadership with real authority to set the quality standard, measure against it, and close the gaps. Not to report on pipelines. To make them better.</p> <p>The work connects directly to Nuro's mission. A single systematic flaw in annotation can propagate silently through training and surface as a safety regression on the road. You are the person who finds it before that happens and who builds the systems that prevent it from happening again. That kind of upstream impact is hard to find in most roles. Here, it's the whole job.</p> <p>You'll be well-supported: partnering closely with senior engineering and operations leadership, with the access and visibility to do your best work. What we ask in return is curiosity, rigor, and genuine care about getting it right.</p> <h2><strong>About the Work</strong></h2> <p>You will own the quality diagnostic layer across our labeling pipelines: defining the standard, building the instrumentation, and closing the gaps that matter most to model safety and performance.</p> <p><strong>Own the Quality Standard</strong></p> <ul> <li>Define what 'good' looks like for each data type across active labeling pipelines and instrument pipelines to measure against it continuously, not just at delivery milestones.</li> <li>Build inter-annotator agreement frameworks, taxonomy governance, and sampling methodologies that hold up at offshore production scale.</li> <li>Design scalable processes to reduce systematic errors and support evolving ML training requirements.</li> </ul> <p><strong>Diagnose and Close Gaps</strong></p> <ul> <li>Audit live workflows, query production databases, and trace accuracy failures to their structural root cause then return with an evidence-based plan that fixes the mechanism, not just the symptom.</li> <li>Apply statistical process control thinking to distinguish a labeling error from a labeling system error, and drive the changes that address it.</li> </ul> <p><strong>Bridge Engineering and Machine Learning Operations</strong></p> <ul> <li>Connect ML labeling quality metrics directly to model performance and safety outcomes in close partnership with Autonomy Engineering leadership.</li> <li>Build executive-ready reporting that frames quality gaps as safety and business signals — not operational updates.</li> <li>Drive alignment across engineering, product, and global ops with clear analysis and well-reasoned recommendations.</li> </ul> <h2><strong>About You</strong></h2> <p>We are not looking for someone who has done everything. We are looking for someone who cares about the right things: data quality as a craft, systems thinking as a default, and mission as a motivator. If you are energized by finding what is broken and building what fixes it and you want that work to mean something this role was written for you.</p> <p><strong>What You Bring</strong></p> <ul> <li>5+ years embedded with ML, data operations, or software engineering teams close to the work, not managing from a distance.</li> <li>SQL fluency: you can investigate a labeling anomaly yourself, form a hypothesis, and test it without waiting on a data engineer.</li> <li>Deep experience with ML data pipelines and labeling ecosystems annotation workflows, quality sampling, taxonomy design, and inter-annotator agreement.</li> <li>A systems-level mindset: you identify where quality breaks down structurally and design the mechanism that fixes it, not just the process that patches it.</li> <li>Clear, confident communication: you can translate a nuanced data quality finding into a precise safety or business risk that senior leadership can act on.</li> <li>Experience managing large-scale offshore or globally distributed annotation teams.</li> </ul> <p><strong>Bonus Points</strong></p> <ul> <li>Background in autonomous vehicles, robotics, computer vision, or ML model training.</li> <li>Prior experience in ML engineering, data engineering, or technical consulting.</li> <li>A demonstrated track record of improving training data quality at scale, with metrics to show for it.</li> <li>Bachelor's degree in a technical or business discipline, or equivalent practical experience.</li> </ul> <p>At Nuro, your base pay is one part of your total compensation package. For this position, the reasonably expected base pay range is between $120,270 and $180,410 for the level at which this job has been scoped. Your base pay will depend on several factors, including your experience, qualifications, education, location, and skills. In the event that you are considered for a different level, a higher or lower pay range would apply. This position is also eligible for an annual performance bonus, equity, and a competitive benefits package.</p> <p><em><span style="font-weight: 400;">At Nuro, we celebrate differences and are committed to a diverse workplace that fosters inclusion and psychological safety for all employees. Nuro is proud to be an equal opportunity employer and expressly prohibits any form of workplace discrimination based on race, color, religion, gender, sexual orientation, gender identity or expression, national origin, age, genetic information, disability, veteran status, or any other legally protected characteristics. #LI-DNP</span></em></p>