Join Assert Labs
We are a team of engineers and researchers with an ambitious mission: to move the world toward error-free software. We're doing this by building tools to autonomously review and test code. We are backed by Accel, Guillermo Rauch (Vercel), Thomas Wolf (Hugging Face), David Cramer (Sentry), Charlie Marsh (Astral), and a number of other open source developers, machine learning researchers, and entrepreneurs. If you wish to learn more, read here about our mission and values.
The Culture We Are Building
- -We are assembling a lean group of generalists. While we welcome experience, we're primarily looking for creative, curious problem solvers with a keen interest in open-ended, technically challenging problems.
- -We are looking to hire in-person in San Francisco. We believe that having a community is critical for learning, growth, and camaraderie. We will pay generously for relocation assistance.
- -We want to invest heavily in our team's growth. We will happily sponsor conference attendance (NeurIPS, ICML, RustConf, PyCon, etc.) and encourage projects and research. If there's something that will help you grow as an engineer, researcher, or designer, we actively want to support it.
- -We believe in compensating exceptional talent exceptionally well. We benchmark our compensation against the top 90th percentile for each role, and we're committed to working with candidates we're excited about to ensure they are excited too.
- -We have an academic as well as competitive spirit. We enjoy thought exercises, math puzzles, intellectual inquiry, and vigorous debates.
Role: Machine Learning Engineer
Design, train, and ship models that guide reviewers to the most important changes, resolve merge conflicts more safely, and make diffs understandable through semantic grouping and summarization. You'll combine retrieval, code‑aware modeling, and small‑model techniques to deliver low‑latency, cost‑effective ML in production.
Responsibilities
- -Train and fine‑tune models that predict the next location a reviewer should focus on: ranking files, hunks, and comments
- -Build ML‑assisted merge conflict resolution: align and reconcile hunks using code context/AST, propose safe resolutions, and explain changes
- -Develop semantic change grouping and diff visualization: detect refactors/moves, cluster related edits across files, and generate concise summaries
- -Evaluate retrieval and RAG for pull requests: retrieve relevant context, discussions, and history to improve grouping, prioritization, and summarization
- -Design offline/online evaluation: labeled datasets, ranking metrics (e.g., MAP/NDCG), A/B experiments, and human‑in‑the‑loop feedback
- -Operate efficient inference: apply small‑model techniques (LoRA, supervised fine‑tuning, distillation, quantization)
Qualifications
- -Strong background in deep learning research and implementation, with experience in modern architectures and concepts (transformers, MoE, reasoning, etc.)
- -Track record publishing results at top-tier conferences (NeurIPS, ICML, ICLR)
- -Expertise in Python and ML frameworks (PyTorch, JAX)
- -Experience with MLOps and model optimization (ONNX, TensorRT, serving infrastructure)
- -Familiarity with cost‑effective techniques: LoRA, supervised fine‑tuning, distillation, and quantization for small/efficient models
- -Familiarity with systems programming across the stack: Rust/C/C++, CUDA/GPU optimization, and kernel-level performance tuning
- -Experience with reinforcement learning, particularly in systems optimization and decision-making tasks
Interview Process
Our interview process is designed to be thorough yet efficient, respecting our time as well as yours. We occasionally accelerate the process or skip steps for candidates.
Application
Send your resume, a personal introduction, and evidence of exceptional ability to hiring@assertlabs.dev.
Technical Discussion
A conversation with our engineers about your background and a deep dive into one of your past projects.
Take - Home
A 4-hour assessment of your ability to handle a challenge and deliver work without supervision.
Coding
A 1-hour live coding interview.
System Design
A 1-hour system design interview.
Quantitative Reasoning
A live session testing your ability to think abstractly and reason about complex problems.
Final Onsite
Collaborate with our team on a real problem in our codebase or design process.