Alignment research, scalable oversight
One of the clearest public anchors for scalable oversight and alignment research in the frontier-model era.
Topic
People exploring planning, tool use, and reasoning-heavy model behavior for longer-horizon tasks.
Start with Jan Leike, Danny Hernandez, Amelia Glaese if you want the clearest first pass through agents & reasoning as it shows up in practice.
This area overlaps heavily with Google DeepMind, Anthropic, AI21. Common institution signals include Anthropic, Google DeepMind, AI21 Labs. Recurring starting points include A Generalist Agent, Constitutional AI: Harmlessness from AI Feedback.
Snapshot
Researchers
110
Related labs
7
Starting points
8
Developed dossiers
14
Useful entry points pulled from the strongest linked researcher dossiers.
Frequent institutions showing up across profiles in this area.
Papers, project pages, and repositories that recur across this part of the field.
A Generalist Agent
18Linked by 18 profiles in this topic
Constitutional AI: Harmlessness from AI Feedback
15Linked by 15 profiles in this topic
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
15Linked by 15 profiles in this topic
Question Decomposition Improves the Faithfulness of Model-Generated Reasoning
9Linked by 9 profiles in this topic
Measuring Faithfulness in Chain-of-Thought Reasoning
8Linked by 8 profiles in this topic
Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
7Linked by 7 profiles in this topic
Jamba: A Hybrid Transformer-Mamba Language Model
7Linked by 7 profiles in this topic
ReAct: Synergizing Reasoning and Acting in Language Models
7Linked by 7 profiles in this topic
Source clusters that repeatedly anchor researchers in this area.
A Generalist Agent
18Used across 18 researcher pages in this topic
Constitutional AI: Harmlessness from AI Feedback
15Used across 15 researcher pages in this topic
Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
15Used across 15 researcher pages in this topic
Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
7Used across 7 researcher pages in this topic
Jamba: A Hybrid Transformer-Mamba Language Model
7Used across 7 researcher pages in this topic
ReAct: Synergizing Reasoning and Acting in Language Models
7Used across 7 researcher pages in this topic
A stronger first pass through agents & reasoning, ranked by profile depth, evidence, and editorial importance.
Alignment research, scalable oversight
One of the clearest public anchors for scalable oversight and alignment research in the frontier-model era.
Alignment via AI feedback (Constitutional AI)
A strong person to follow for how Anthropic moved from assistant training into more explicit evaluation work around model behavior, red-teaming, and chain-of-thought faithfulness.
Gemini (multimodal foundation models)
A useful researcher to follow if you care about the bridge between safety evaluation, human data, and how frontier models are turned into practical tools and benchmarks.
Mixture-of-experts LLMs
A useful person to follow if you care about the bridge between embodied-agent research and modern open-weight language-model systems, rather than treating those worlds as separate.
Mixture-of-experts LLMs
A strong person to know for the Mistral line of open-weight models, especially if you care about the arc from compact performant base models into mixture-of-experts, multimodal systems, and reasoning models.
Reasoning, verification, math
A useful person to study if you care about alignment proposals that try to make superhuman systems legible enough for humans to supervise in practice.
Sequence models, large-scale ML
A high-signal researcher for understanding how DeepMind approaches generality, especially in areas where reinforcement learning, multimodality, and large-scale systems meet.
Deep learning, research leadership
A long-running builder of ML intuition whose influence spans Bayesian methods, reinforcement learning, and recent work on generalist and generative environments.
Faster LLM inference via speculative decoding
An important systems page because he is one of the named authors on speculative decoding, a technique that became part of the mainstream conversation about making large-model inference materially faster without changing outputs.
110 linked profiles.