Back to page 1

Researchers — page 26

Showing 3001-3120 of 3,615 researchers

3019
Thibaut Lavril

Open-weight foundation models (LLaMA)

A strong page to keep because he sits on both sides of a major shift in open models: he appears on Meta's LLaMA 2 paper and then on Mistral 7B and Mixtral, which makes him part of the early handoff from the first LLaMA wave into Mistral's open-weight model line.

3080
Tom Ben Gal

Hybrid Transformer–Mamba language models (Jamba)

Useful because it puts a name and a clear role on one of the engineers working at the boundary between research and implementation for AI21’s hybrid-model stack.

3103
Travis Hoppe

Open-source LLMs (EleutherAI)

Worth knowing as one of the early open-data contributors around the EleutherAI orbit, with a profile that mixes work on The Pile with a long tail of small, public NLP and machine-learning experiments.

3109
Tri Dao

Efficient sequence models + attention kernels

One of the clearest researchers to follow for efficient sequence-model systems, especially the line of work that made frontier training and inference materially faster rather than merely cleaner on paper.