We’ve organized every stage and persona in the AI supply chain, informed by real recruiting at frontier companies. Click any row to see matching profiles from our talent graph.







Summary
Known as: Member of Technical Staff, Research Engineer, Research Software Engineer
Trains models from scratch (pre-training), continues training on specialized datasets (mid-training), or distills frontier model capabilities into smaller, deployable form factors. Owns training runs end-to-end: data mixing, hyperparameter tuning, convergence monitoring, and checkpoint evaluation. A large and growing share of training compute goes to distillation — compressing what the frontier model knows into models cheap and fast enough to ship.
Specializations
Where the Work Lives
Runs distributed training across GPU clusters, consuming massive compute for weeks-long runs.
Owns the training recipe — data mixing, hyperparameters, convergence — that turns compute into learned capability.
Candidate Archetypes
Decides what the model learns when — mixture ratios, phase schedules, and ablation-driven recipe changes.
Reads loss and gradient health, tunes schedules and precision, and calls whether a run is sick or salvageable.
Owns stop/branch decisions and checkpoint triage that turn compute into usable model artifacts.
Company Scale
Frontier labs for pretraining. Growth-stage for domain-specific training with the right data and compute.
Featured Roles
If you’re hiring at the AI frontier, let’s talk.