
At Baseten we provide all the infrastructure you need to deploy and serve ML models performantly, scalably, and cost-efficiently. Get started in minutes, and avoid getting tangled in complex deployment processes. You can deploy best-in-class open-source models and take advantage of optimized serving for your own models. We also utilize horizontally scalable services that take you from prototype to production, with light-speed inference on infra that autoscales with your traffic. Best in class doesn't mean breaking the bank. Run your models on the best infrastructure without running up costs by taking advantage of our scaled-to-zero feature.

At Baseten we provide all the infrastructure you need to deploy and serve ML models performantly, scalably, and cost-efficiently. Get started in minutes, and avoid getting tangled in complex deployment processes. You can deploy best-in-class open-source models and take advantage of optimized serving for your own models. We also utilize horizontally scalable services that take you from prototype to production, with light-speed inference on infra that autoscales with your traffic. Best in class doesn't mean breaking the bank. Run your models on the best infrastructure without running up costs by taking advantage of our scaled-to-zero feature.
Sector: AI inference / ML infrastructure
Founded: 2019
Headquarters: San Francisco, California
Employee count (reported): 204
Notable recent funding: $300M Series E (company disclosure)
Production inference and serving for machine-learning models (including LLMs) with emphasis on scalability, performance, and cost control.
2019
Software Development
$40M
$75M
$150M
$300M
Company disclosure reporting $300M Series E at $5B valuation
“Baseten has raised late-stage rounds with participation from investors including Bond, IVP, CapitalG, Spark Capital, NVIDIA, Greylock, Conviction, 01 Advisors, BoxGroup, and others.”
| Company |
|---|
About Baseten
Baseten powers mission-critical inference for the world's most dynamic AI companies, like Cursor, Notion, OpenEvidence, Abridge, Clay, Gamma and Writer. By uniting applied AI research, flexible infrastructure, and seamless developer tooling, we enable companies operating at the frontier of AI to bring cutting-edge models into production. We're growing quickly and recently raised our $150M Series D, backed by investors including BOND, IVP, Spark Capital, Greylock, and Conviction. Join us and help build the platform engineers turn to to ship AI products.
THE ROLE
As an at Baseten, you’ll own the roadmap for our core inference and compute infrastructure, ensuring our platform delivers world-class reliability, scalability, and performance. You’ll work closely with engineering teams to define how we handle large-scale distributed systems, optimize GPU utilization, and provide enterprise-grade security and observability. This is a deeply technical role that bridges engineering excellence and customer impact, ensuring Baseten’s infrastructure is a foundation our users can depend on.
EXAMPLE INITIATIVES
You'll get to work on these types of projects as part of our Infrastructure team:
Responsibilities
Requirements
NICE TO HAVE
Benefits
Apply now to embark on a rewarding journey in shaping the future of AI! If you are a motivated individual with a passion for machine learning and a desire to be part of a collaborative and forward-thinking team, we would love to hear from you.
At Baseten, we are committed to fostering a diverse and inclusive workplace. We provide equal employment opportunities to all employees and applicants without regard to race, color, religion, gender, sexual orientation, gender identity or expression, national origin, age, genetic information, disability, or veteran status.