We’ve organized every stage and persona in the AI supply chain, informed by real recruiting at frontier companies. Click any row to see matching profiles from our talent graph.







Summary
Known as: Research Scientist (Safety), Alignment Researcher, Interpretability Researcher, AI Red Team Engineer, Safety Engineer, Adversarial ML Researcher
Research and engineering for alignment, interpretability, and safe model behavior. Makes models do what people intend even when instructions are ambiguous, users try to break them (jailbreaks, prompt injection), or systems take actions over many steps.
Specializations
In many orgs the same people red-team and then fix what they find via RLHF. The split between safety research and safety tuning is an org-design choice, not a hard technical boundary.
Where the Work Lives
Researches alignment techniques, interpretability, and reward modeling to make models do what we intend.
Safety evaluation, red teaming, and adversarial testing that harden models before and during deployment.
Candidate Archetypes
Studies internal representations and circuits to predict and constrain model behavior.
Discovers jailbreak, prompt-injection, and tool-misuse failure modes and turns them into repeatable test assets.
Builds supervision schemes and reward models that hold up as model capability grows.
Company Scale
Frontier labs and safety-focused orgs. Most enterprises do governance, not alignment research.
Featured Roles
If you’re hiring at the AI frontier, let’s talk.