α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Dan Alistarh
Dan Alistarh
23
papers
1,015
total citations
papers (23)
Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning
NEURIPS 2022
arXiv
340
citations
Extreme Compression of Large Language Models via Additive Quantization
ICML 2024
arXiv
160
citations
Adaptive Gradient Quantization for Data-Parallel SGD
NEURIPS 2020
arXiv
100
citations
AC/DC: Alternating Compressed/DeCompressed Training of Deep Neural Networks
NEURIPS 2021
arXiv
78
citations
M-FAC: Efficient Matrix-Free Approximations of Second-Order Information
NEURIPS 2021
arXiv
64
citations
Asynchronous Decentralized SGD with Quantized and Local Updates
NEURIPS 2021
arXiv
59
citations
How Well Do Sparse ImageNet Models Transfer?
CVPR 2022
arXiv
49
citations
RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
ICML 2024
arXiv
48
citations
WoodFisher: Efficient Second-Order Approximation for Neural Network Compression
NEURIPS 2020
arXiv
29
citations
CAP: Correlation-Aware Pruning for Highly-Accurate Sparse Vision Models
NEURIPS 2023
arXiv
21
citations
Distributed Principal Component Analysis with Limited Communication
NEURIPS 2021
arXiv
15
citations
Bias in Pruned Vision Models: In-Depth Analysis and Countermeasures
CVPR 2023
arXiv
12
citations
Knowledge Distillation Performs Partial Variance Reduction
NEURIPS 2023
arXiv
10
citations
Towards Tight Communication Lower Bounds for Distributed Optimisation
NEURIPS 2021
arXiv
10
citations
Error Feedback Can Accurately Compress Preconditioners
ICML 2024
arXiv
6
citations
Wasserstein Distances, Neuronal Entanglement, and Sparsity
ICLR 2025
arXiv
5
citations
Cache Me If You Must: Adaptive Key-Value Quantization for Large Language Models
ICML 2025
arXiv
4
citations
SPADE: Sparsity-Guided Debugging for Deep Neural Networks
ICML 2024
arXiv
2
citations
Hybrid Decentralized Optimization: Leveraging Both First- and Zeroth-Order Optimizers for Faster Convergence
AAAI 2025
arXiv
1
citations
Layer-wise Quantization for Quantized Optimistic Dual Averaging
ICML 2025
arXiv
1
citations
The Journey Matters: Average Parameter Count over Pre-training Unifies Sparse and Dense Scaling Laws
ICLR 2025
arXiv
1
citations
ZipLM: Inference-Aware Structured Pruning of Language Models
NEURIPS 2023
0
citations
Scalable Belief Propagation via Relaxed Scheduling
NEURIPS 2020
0
citations