Bonsai Robotics' mission is to create the next leap forward in agriculture equipment efficiency by creating a new ecosystem of semi-autonomous robotic machinery. Orchards are dusty, hazard-filled, and GPS-denied. The GPS-based autosteer features that have driven row crop efficiencies cannot function in orchards. Our vision, AI, and machine control systems offer human-level environment understanding and local navigation capabilities and will be the platform for a new wave of innovation in agricultural production and management systems.
We simultaneously solve twin crises impacting nut growers and most of specialty agriculture : there is not enough human labor when you need it, and operational expenses are growing dramatically. Our state-of-the-art technology empowers orchard managers to optimize their operations, dramatically reduce operational expenses, and increase profitability. We are pursuing a Bonsai Inside strategy, and partnering with the largest orchard Original Equipment Manufacturers (OEMs) in the retrofitting of existing machines and design of form factors.
About the role
We’re looking for an MLOps engineer who thrives in real-world robotics environments and can own the entire machine learning lifecycle—from data ingestion and labeling to training, evaluation, and performance monitoring. You’ll support a perception stack spanning 2D / 3D object detection, semantic and instance segmentation, depth estimation, and multi-sensor fusion across camera and lidar.
This role is deeply cross-functional : you’ll work with perception engineers, autonomy engineers, field operations, and external labeling teams. The work is fast, tangible, and impacts every vehicle that goes into the field.
What you'll do
- Build and maintain scalable data pipelines for 2D / 3D detection, segmentation, instance segmentation, and depth estimation
- Develop data workflows across multi-camera systems and lidar stored in MCAP format
- Own dataset versioning, metadata tracking, and reproducibility systems.
- Improve training throughput using distributed systems (Ray, PyTorch Lightning, custom launchers).
- Optimize data formats and loaders for large‑scale vision and lidar datasets.
Data Curation & Quality Management
Build automated tools for dataset selection, active learning, hard‑sample mining, and outlier detection.Maintain dashboards and automated checks for dataset health, label quality, class balance, and environment coverage.Partner with field teams to prioritize data collection runs and close the loop between field issues and dataset refreshes.Labeling Operations
Manage internal labelers and external labeling vendors.Define annotation standards for camera and lidar tasks.Build QA workflows, reviewer interfaces, and automated label‑consistency checks.Identify systematic labeling errors and drive corrective processes.Deployment & Model Lifecycle
Build pipelines for continuous evaluation using telemetry from vehicles in the field.Monitor model drift, identify edge cases, and manage regression tests across “golden” datasets.Track on‑vehicle performance signals to flag data needs, degradations, or unexpected behavior.Cross‑Functional Collaboration
Work closely with perception engineers on calibration, sensor models, data schemas, and on‑vehicle inference constraints.Coordinate with autonomy and perception teams to align ML outputs with navigation needs.Work with platform team to integrate ML pipelines into core platform infrastructurePartner with fleet operations to review real‑world performance and prioritize new data collection.Qualifications
4–7+ years industry experience in MLOps, ML infrastructure, data engineering or applied ML engineeringExperience building robust data pipelines for large‑scale vision or lidar datasets.Experience managing and operating cloud infrastructure (e.g., AWS EC2, S3, IAM, autoscaling, spot fleets).Familiarity with ML lifecycle (MLflow, Weights & Biases, Metaflow, Airflow, Ray, etc.).Experience managing labeling workflows or working directly with annotation vendors.Strong debugging instincts across the full stack—from data issues to training failures to evaluation anomalies.Bonus Points For
Experience with PyTorch, CUDA, and common CV / 3D libraries.Experience with multi‑sensor fusion, BEV architectures, or 3D perception.Familiarity with MCAP, ROS2, Foxglove, and real‑time robotics systems.Experience with autonomous vehicle pipelines or industrial / agricultural robotics.Background in active learning or automated label‑quality scoring.Experience building synthetic data augmentations or simulator‑driven dataset expansion.Experience building auto‑labeling pipelinesBonsai Robotics is an Equal Employment Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, creed, religion, sex, sexual orientation, national origin or nationality, ancestry, age, disability, gender identity or expression, marital status or any other category protected by law.
The pay range for this role is :
120,000 - 180,000 USD per year (San Jose, CA)
#J-18808-Ljbffr