Open-source LLMs, datasets
A key open-model ecosystem builder whose work matters because it combines research, public infrastructure, and field-level coordination rather than isolated paper output alone.
Lab & Ecosystem
Open-source contributors who helped bootstrap independent LLM training, datasets, and tooling.
Within 500AI, EleutherAI is most legible through researchers like Stella Biderman, Eric Hallahan, Anish Thite.
This cluster is especially tied to Open Models, Systems & Infrastructure, Evaluation & Benchmarks. Frequent institution signals include EleutherAI, Conjecture, Georgia Institute of Technology. Recurring entry points include EleutherAI (GitHub), GPT-NeoX (GitHub).
Snapshot
Researchers
23
Related topics
6
Starting points
8
Developed dossiers
5
Useful lenses pulled from the strongest researcher profiles in this cluster.
Frequent institutions showing up across linked profiles in this ecosystem.
Repeatedly linked papers, projects, and repositories across this lab cluster.
EleutherAI (GitHub)
22Linked by 22 profiles in this cluster
GPT-NeoX (GitHub)
22Linked by 22 profiles in this cluster
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
10Linked by 10 profiles in this cluster
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
6Linked by 6 profiles in this cluster
GPT-NeoX-20B: An Open-Source Autoregressive Language Model
4Linked by 4 profiles in this cluster
A framework for few-shot language model evaluation
2Linked by 2 profiles in this cluster
Cross-Cultural Transfer of Commonsense Reasoning in LLMs
2Linked by 2 profiles in this cluster
Interpreting Neural Networks through the Polytope Lens
2Linked by 2 profiles in this cluster
Source clusters that repeatedly anchor researcher pages in this ecosystem.
EleutherAI (GitHub)
22Used across 22 researcher pages in this lab cluster
GPT-NeoX (GitHub)
21Used across 21 researcher pages in this lab cluster
RWKV: Reinventing RNNs for the Transformer Era
2Used across 2 researcher pages in this lab cluster
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
1Used across 1 researcher pages in this lab cluster
A stronger first pass through EleutherAI, ranked by profile depth, evidence, and editorial importance.
Open-source LLMs, datasets
A key open-model ecosystem builder whose work matters because it combines research, public infrastructure, and field-level coordination rather than isolated paper output alone.
Open-source LLMs (EleutherAI)
Useful because his footprint runs through the early EleutherAI training stack, GPT-NeoX, and Pythia, which makes the page a better map of open-model infrastructure than a generic one-paper profile.
Open-source LLMs (EleutherAI)
Useful to follow if you care about the practical evaluation layer of open models, especially where benchmark tooling and reproducible comparisons actually shape what the ecosystem measures.
Open models, governance, communication
An important bridge figure between open-weight language-model communities and the modern alignment debate, especially when you want to understand how frontier capability, openness, and control arguments collide in practice.
Open-source LLMs (EleutherAI)
An important open-model researcher for understanding how early public LLM efforts, scaling heuristics, and open data work fed into the broader modern model ecosystem.
Open-source LLMs (EleutherAI)
Important for the bridge between early open-model scaling work and later frontier closed-model systems, especially around architecture and training-stack choices that ended up mattering at both ends of the field.
Open-source LLMs (EleutherAI)
Worth tracking for the open-model side of the field, especially where dataset construction, practical training work, and alignment-flavored thinking meet.
Open-source LLMs, training
A useful anchor for the open-model ecosystem because his path runs from EleutherAI’s training efforts into a more explicit alignment and interpretability agenda at Conjecture.
Open-source LLMs (EleutherAI)
A strong person to follow for the systems side of open models, especially where distributed training, hybrid architectures, and practical efficiency work feed directly into model capability.
23 linked profiles.