α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Beidi Chen
Beidi Chen
17
papers
1,408
total citations
papers (17)
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
ICML 2024
arXiv
371
citations
KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache
ICML 2024
arXiv
368
citations
Scatterbrain: Unifying Sparse and Low-rank Attention
NEURIPS 2021
arXiv
154
citations
Decentralized Training of Foundation Models in Heterogeneous Environments
NEURIPS 2022
arXiv
126
citations
Scan and Snap: Understanding Training Dynamics and Token Composition in 1-layer Transformer
NEURIPS 2023
arXiv
105
citations
Get More with LESS: Synthesizing Recurrence with KV Cache Compression for Efficient LLM Inference
ICML 2024
arXiv
79
citations
ShadowKV: KV Cache in Shadows for High-Throughput Long-Context LLM Inference
ICML 2025
arXiv
65
citations
JoMA: Demystifying Multilayer Transformers via Joint Dynamics of MLP and Attention
ICLR 2024
arXiv
48
citations
HexGen: Generative Inference of Large Language Model over Heterogeneous Environment
ICML 2024
arXiv
34
citations
Laughing Hyena Distillery: Extracting Compact Recurrences From Convolutions
NEURIPS 2023
arXiv
29
citations
LoCoCo: Dropping In Convolutions for Long Context Compression
ICML 2024
arXiv
16
citations
Speculative Prefill: Turbocharging TTFT with Lightweight and Training-Free Token Importance Estimation
ICML 2025
arXiv
8
citations
Zeroth-Order Fine-Tuning of LLMs with Transferable Static Sparsity
ICLR 2025
5
citations
Fine-tuning Language Models over Slow Networks using Activation Quantization with Guarantees
NEURIPS 2022
0
citations
Soft Prompt Recovers Compressed LLMs, Transferably
ICML 2024
0
citations
H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models
NEURIPS 2023
0
citations
Locality Sensitive Teaching
NEURIPS 2021
0
citations