α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Minjia Zhang
Minjia Zhang
10
papers
1,262
total citations
papers (10)
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
NEURIPS 2022
arXiv
636
citations
Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs
ICLR 2024
arXiv
390
citations
Accelerating Training of Transformer-Based Language Models with Progressive Layer Dropping
NEURIPS 2020
arXiv
120
citations
The Stability-Efficiency Dilemma: Investigating Sequence Length Warmup for Training GPT Models
NEURIPS 2022
arXiv
51
citations
DeepSpeed Data Efficiency: Improving Deep Learning Model Quality and Training Efficiency via Efficient Data Sampling and Routing
AAAI 2024
arXiv
40
citations
NxMTransformer: Semi-Structured Sparsification for Natural Language Understanding via ADMM
NEURIPS 2021
arXiv
24
citations
InstantEdit: Text-Guided Few-Step Image Editing with Piecewise Rectified Flow
ICCV 2025
arXiv
1
citations
HM-ANN: Efficient Billion-Point Nearest Neighbor Search on Heterogeneous Memory
NEURIPS 2020
0
citations
XTC: Extreme Compression for Pre-trained Transformers Made Simple and Efficient
NEURIPS 2022
0
citations
AdaTune: Adaptive Tensor Program Compilation Made Efficient
NEURIPS 2020
0
citations