Back to researchers
Tri Dao
Efficient sequence models + attention kernels
Co-authored Mamba and FlashAttention/FlashAttention-2: high-leverage work for fast training and inference.
Highlights
MambaFlashAttentionSystemsEfficient sequence models
Focus: Efficient sequence models + attention kernels
Why it matters: Co-authored Mamba and FlashAttention/FlashAttention-2: high-leverage work for fast training and inference.
Research Areas
MambaFlashAttentionSystemsEfficient sequence models