
Human Archive is pioneering the first multimodal data ecosystem for embodied intelligence, focusing on tactile intelligence and unified vision data. Their innovative approach allows robots to sense, grip, and interact like humans by utilizing high-fidelity datasets from real human motion. Additionally, they empower individuals to earn by performing real-world tasks with wearable kits, generating rich data that enhances robotic intelligence. This unique business model positions Human Archive at the forefront of the robotics industry, catering to the growing demand for advanced robotic capabilities.

Human Archive is pioneering the first multimodal data ecosystem for embodied intelligence, focusing on tactile intelligence and unified vision data. Their innovative approach allows robots to sense, grip, and interact like humans by utilizing high-fidelity datasets from real human motion. Additionally, they empower individuals to earn by performing real-world tasks with wearable kits, generating rich data that enhances robotic intelligence. This unique business model positions Human Archive at the forefront of the robotics industry, catering to the growing demand for advanced robotic capabilities.
About Human Archive
Human Archive is a robotics data lab founded by Stanford and UC Berkeley dropouts. We work alongside frontier robotics labs and foundation model research groups to collect large-scale, quality, diverse, multimodal datasets of humans performing everyday tasks across household and industrial environments.
We are lean, technical, and operate at extreme speed, taking on unglamorous and conventionally impossible problems that directly unlock step-function gains in model capability.
The deployment of capable humanoids at scale will permanently redefine human labor. Undesirable physical work will disappear, and human effort will shift toward a new era of abundant creativity. This shift is inevitable, and we are building the infrastructure to accelerate it.
We are assembling the best team to solve the hardest problems in embodied intelligence. You will own meaningful systems from day one and see your work directly impact VLA model capabilities. This is a once-in-a-generation inflection point. If you want to leave your dent on humanity and reshape physical labor markets forever, join us!
About the Role
This role is ideal for an early-career engineer or exceptional dropout who wants real ownership and is interested in possibly joining Human Archive as a founding engineer.
As a Machine Learning Intern, you’ll build the data and ML infrastructure that powers the next generation of safe, intelligent, dexterous, and generalizable robots. You’ll work on in-house Vision-Language-Action (VLA) model benchmarking, annotation software, analytics and data platform products, and multimodal alignment tooling. This is a hands-on role at the intersection of applied machine learning, data infrastructure, and robotics, where your work directly shapes how data is collected, validated, annotated, and converted into high-quality learning signal.
Your mandate is to close the loop between research and data collection by building systems that let us rapidly evaluate new environments, tasks, and sensor modalities through fine-tuning and benchmarking our own VLA models, as well as by building annotation pipelines designed for direct model ingestion. You’ll be expected to ship production-grade systems, make architectural decisions, and own projects end to end.
What We’re Looking For