Back to topics

Topic

Post-Training & Alignment

Researchers shaping model behavior after pretraining, from instruction tuning and preference learning to scalable oversight.

Start with Chris Olah, Dario Amodei, Amanda Askell if you want the clearest first pass through post-training & alignment as it shows up in practice.

This area overlaps heavily with Anthropic, OpenAI, AI21. Common institution signals include Anthropic, OpenAI, Stanford University. Recurring starting points include Constitutional AI: Harmlessness from AI Feedback, Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback.

Snapshot

Researchers

148

Related labs

5

Starting points

8

Developed dossiers

35

Institution Signals

Frequent institutions showing up across profiles in this area.

Anthropic (48)OpenAI (9)Stanford University (6)AI21 Labs (5)Google (5)Google DeepMind (3)Alignment Research Center (2)Conjecture (2)

Canonical Starting Points

Papers, project pages, and repositories that recur across this part of the field.

Frequently Linked Sources

Source clusters that repeatedly anchor researchers in this area.

Researchers To Start With

A stronger first pass through post-training & alignment, ranked by profile depth, evidence, and editorial importance.

All Researchers In This Topic

148 linked profiles.