Back to researchers

Tri Dao

Efficient sequence models + attention kernels

Co-authored Mamba and FlashAttention/FlashAttention-2: high-leverage work for fast training and inference.

Highlights

MambaFlashAttentionSystemsEfficient sequence models
Focus: Efficient sequence models + attention kernels
Why it matters: Co-authored Mamba and FlashAttention/FlashAttention-2: high-leverage work for fast training and inference.

Research Areas

MambaFlashAttentionSystemsEfficient sequence models
Tri Dao - AI Researcher Profile | 500AI