
Cosine is an autonomous, on-premise coding agent post-trained on human reasoning data to deliver unmatched software-engineering accuracy, security, and speed for regulated enterprises.

Cosine is an autonomous, on-premise coding agent post-trained on human reasoning data to deliver unmatched software-engineering accuracy, security, and speed for regulated enterprises.
What they do: Autonomous, on-premise AI coding agent that reads codebases, plans and executes engineering tasks, runs tests, and drafts PRs
Target customers: Regulated enterprises with large, legacy, or high-security codebases (finance, defence, SaaS, manufacturing)
Deployment: Air-gapped on-prem, customer VPC, or cloud with emphasis on zero data egress
Founding year: 2022
| Company |
|---|
Automating software engineering tasks in large, legacy, or high-security enterprise codebases.
2022
Developer tools / AI for software engineering
Crunchbase lists multiple seed rounds; public profiles redact exact amounts and dates
Crunchbase lists multiple seed rounds; public profiles redact exact amounts and dates
Crunchbase indicates three funding rounds in total but details are obfuscated in public profile
“Has multiple institutional and angel investors (examples named in public profiles include Warrick Shanly and Lakestar)”
We’re looking for an ML engineer to own large-scale training of our Lumen Enterprise models – our open‑source–based software engineering LLMs.
You’ll work on supervised fine-tuning (SFT), and reinforcement learning (RL) and continued pre-training on top of open-source base models to push state-of-the-art performance on real software engineering tasks: reading and modifying large codebases, using tools, and reasoning about complex systems.
If you enjoy working close to the metal with PyTorch and distributed training, and you like making big models actually work in practice, this role is for you.
About the role
In this role you will:
You’ll collaborate closely with infra, product, and research to decide what to train next, how to train it, and how to measure whether it’s actually better for engineers.
What you'll do:
Participate in end-to-end training of Lumen Enterprise SWE models:
Design, implement, and iterate on RL training pipelines
Build and maintain large-scale PyTorch training code:
Operate large multi-node training jobs:
Work on long-context and code-focused training:
Improve evaluation for SWE models:
Collaborate:
What we're looking for (must-haves)
Strong experience training deep learning models in production:
Deep proficiency with PyTorch and its primitives:
Experience training large sequence models or LLMs:
Experience with SFT and RL on top of LLMs:
Strong software engineering background:
Distributed systems / training ops experience:
Data engineering instincts:
Clear communication and ownership:
Nice to have (bonus)
You don’t need all of these, but the more you have, the more you’ll hit the ground running:
Continued pre-training and long-context experience:
Code-focused RL and evaluation:
Experience with modern LLM training stacks:
Serving and online training:
Safety, robustness, and reward shaping:
Open-source contributions or research:
Why this role is interesting
If this sounds like a fit, this is a role where you can meaningfully push the frontier of open-source–based software engineering models.