α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Tri Dao
Tri Dao
16
papers
9,884
total citations
papers (16)
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
NEURIPS 2022
arXiv
3,551
citations
FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
ICLR 2024
arXiv
2,224
citations
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
ICML 2024
arXiv
1,146
citations
Combining Recurrent, Convolutional, and Continuous-time Models with Linear State Space Layers
NEURIPS 2021
arXiv
977
citations
HiPPO: Recurrent Memory with Optimal Polynomial Projections
NEURIPS 2020
arXiv
838
citations
Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads
ICML 2024
arXiv
549
citations
Caduceus: Bi-Directional Equivariant Long-Range DNA Sequence Modeling
ICML 2024
arXiv
170
citations
Scatterbrain: Unifying Sparse and Low-rank Attention
NEURIPS 2021
arXiv
154
citations
Decentralized Training of Foundation Models in Heterogeneous Environments
NEURIPS 2022
arXiv
126
citations
S4ND: Modeling Images and Videos as Multidimensional Signals with State Spaces
NEURIPS 2022
arXiv
55
citations
Long-Context State-Space Video World Models
ICCV 2025
arXiv
31
citations
Rethinking Neural Operations for Diverse Tasks
NEURIPS 2021
arXiv
25
citations
Transform Once: Efficient Operator Learning in Frequency Domain
NEURIPS 2022
arXiv
24
citations
Hardware-Efficient Attention for Fast Decoding
COLM 2025
arXiv
8
citations
Ladder-Residual: Parallelism-Aware Architecture for Accelerating Large Model Inference with Communication Overlapping
ICML 2025
arXiv
6
citations
Fine-tuning Language Models over Slow Networks using Activation Quantization with Guarantees
NEURIPS 2022
0
citations