Deep RL, scientific AI, leadership
Important both as a researcher and as an institution builder whose long-running agenda tied deep RL, multimodal systems, and scientific AI into one coherent lab strategy.
Topic
Researchers who make large-scale training and inference practical through architecture, kernels, sharding, and serving work.
Start with Demis Hassabis, Ashish Vaswani, Stella Biderman if you want the clearest first pass through systems & infrastructure as it shows up in practice.
This area overlaps heavily with AI21, Mistral, Anthropic. Common institution signals include AI21 Labs, Meta, Anthropic. Recurring starting points include Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone, Jamba-1.5: Hybrid Transformer-Mamba Models at Scale.
Snapshot
Researchers
417
Related labs
8
Starting points
8
Developed dossiers
57
Useful entry points pulled from the strongest linked researcher dossiers.
Frequent institutions showing up across profiles in this area.
Papers, project pages, and repositories that recur across this part of the field.
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
122Linked by 122 profiles in this topic
Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
61Linked by 61 profiles in this topic
Jamba: A Hybrid Transformer-Mamba Language Model
61Linked by 61 profiles in this topic
AI21 Jamba Large 1.5 model card
42Linked by 42 profiles in this topic
RWKV: Reinventing RNNs for the Transformer Era
34Linked by 34 profiles in this topic
RWKV (project)
32Linked by 32 profiles in this topic
Mixtral of Experts
23Linked by 23 profiles in this topic
Constitutional AI: Harmlessness from AI Feedback
18Linked by 18 profiles in this topic
Source clusters that repeatedly anchor researchers in this area.
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
122Used across 122 researcher pages in this topic
Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
61Used across 61 researcher pages in this topic
Jamba: A Hybrid Transformer-Mamba Language Model
61Used across 61 researcher pages in this topic
RWKV: Reinventing RNNs for the Transformer Era
34Used across 34 researcher pages in this topic
RWKV (project)
32Used across 32 researcher pages in this topic
Mixtral of Experts
20Used across 20 researcher pages in this topic
A stronger first pass through systems & infrastructure, ranked by profile depth, evidence, and editorial importance.
Deep RL, scientific AI, leadership
Important both as a researcher and as an institution builder whose long-running agenda tied deep RL, multimodal systems, and scientific AI into one coherent lab strategy.
Transformers
A foundational figure in modern sequence modeling whose work on the Transformer changed the technical direction of language and multimodal systems.
Open-source LLMs, datasets
A key open-model ecosystem builder whose work matters because it combines research, public infrastructure, and field-level coordination rather than isolated paper output alone.
ML systems, large-scale infrastructure
Foundational less for any single public paper than for shaping the infrastructure, engineering culture, and systems thinking that make frontier-model research possible.
Computer vision, representation learning
A foundational computer-vision researcher whose work on representations and architectures still shapes modern pretraining and perception systems.
Transformers, Mixture-of-Experts, scaling
One of the most important architecture-level thinkers in modern AI, with influence spanning Transformers, efficient scaling, and mixture-of-experts systems.
Alignment via AI feedback (Constitutional AI)
High-signal for the seam between machine learning and hardware systems, especially where learned optimization methods begin affecting the actual compute infrastructure underneath frontier models.
Alignment via AI feedback (Constitutional AI)
A strong person to follow for the point where machine learning research starts shaping the compute stack itself, especially in chip placement and systems-aware optimization.
Gemini (multimodal foundation models)
A good researcher to follow for the infrastructure side of frontier language models, especially mixture-of-experts scaling, instruction tuning, and the data systems that make very large models usable.
417 linked profiles.