Back to page 1

Researchers — page 25

Showing 2881-3000 of 3,615 researchers

2901
Song Han

Streaming + long-context stability (attention sinks)

One of the clearest researchers to follow for efficient AI systems, especially the line of work that makes large models smaller, faster, and easier to deploy without giving up too much quality.

2902
Songlin Yang

Linear transformers via the delta rule

A high-signal researcher for the post-attention design space, especially if you care about the line of work trying to make linear-attention and Delta-rule models actually competitive in real language-model systems.

2978
Tal Ness

Hybrid Transformer–Mamba language models (Jamba)

A worthwhile long-tail researcher page because it makes the data-and-evaluation layer of modern language-model work visible instead of treating frontier systems as if they were only architecture or scaling stories.