
Grafton Sciences is focused on building physical superintelligence to accelerate scientific discovery. Their mission is to create systems capable of autonomous scientific discovery, utilizing tools that expand our capabilities. They are developing an at-home early detection platform for cancer, leveraging advances in synthetic biology and nanotechnology. With a $42.5 million contract from ARPA-H, Grafton aims to detect over 48 tumors at Stage I, representing a significant portion of the global cancer burden. Their approach integrates computational modeling, automated experimentation, and materials engineering, positioning them as leaders in the field of disease detection and scientific innovation.

Grafton Sciences is focused on building physical superintelligence to accelerate scientific discovery. Their mission is to create systems capable of autonomous scientific discovery, utilizing tools that expand our capabilities. They are developing an at-home early detection platform for cancer, leveraging advances in synthetic biology and nanotechnology. With a $42.5 million contract from ARPA-H, Grafton aims to detect over 48 tumors at Stage I, representing a significant portion of the global cancer burden. Their approach integrates computational modeling, automated experimentation, and materials engineering, positioning them as leaders in the field of disease detection and scientific innovation.
About Grafton Sciences
We’re building AI systems with general physical ability — the capacity to experiment, engineer, or manufacture anything. We believe achieving this is a key step towards building superintelligence. With deep technical roots and real-world progress at scale (e.g., a $42M NIH project), we’re pushing the frontier of physical AI. Joining us means inventing from first principles, owning real systems end-to-end, and helping build a capability the world has never had before.
About the Role
We’re seeking a ML Infrastructure / MLOps Engineer to build and operate the infrastructure that powers large-scale training, fine-tuning, RLHF/DPO pipelines, dataset governance, experiment tracking, and model deployment. You’ll design distributed training systems, containerized model runners, data versioning workflows, and reproducible evaluation pipelines that enable rapid iteration across LLMs, RL agents, and surrogate models. This role sits at the heart of the ML stack, ensuring stability, reliability, and performance across all model development.
Responsibilities
• Build and maintain scalable infrastructure for training, fine-tuning, RLHF/DPO workflows, and distributed experiments.
• Develop data pipelines, dataset versioning systems, experiment tracking tools, and reproducibility frameworks.
• Operate containerized inference and training environments, CI/CD for models, and evaluation automation.
• Collaborate with LLM researchers, RL scientists, data engineering, and systems teams to support rapid iteration and robust model deployment.
Qualifications
• Strong experience with ML infrastructure, distributed training, experiment management, or production ML systems.
• Familiarity with containerization, orchestration, model runners, dataset governance, and evaluation pipelines.
• Ability to design reliable training and deployment workflows that support high-throughput experimentation.
• Comfortable working across ML, infrastructure, data systems, and engineering teams in a fast-paced research environment.
Above all, we look for candidates who can demonstrate world-class excellence.
Compensation
We offer competitive salary, meaningful equity, and benefits