α
Research
Alpha Leak
Conferences
Topics
Top Authors
Rankings
Browse All
EN
中
Home
/
Authors
/
Zhewei Yao
Zhewei Yao
7
papers
1,620
total citations
papers (7)
ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers
NEURIPS 2022
arXiv
636
citations
ZeroQ: A Novel Zero Shot Quantization Framework
CVPR 2020
arXiv
464
citations
HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
NEURIPS 2020
arXiv
339
citations
Exploring Post-training Quantization in LLMs from Comprehensive Study to Low Rank Compensation
AAAI 2024
arXiv
71
citations
A Statistical Framework for Low-bitwidth Training of Deep Neural Networks
NEURIPS 2020
arXiv
70
citations
DeepSpeed Data Efficiency: Improving Deep Learning Model Quality and Training Efficiency via Efficient Data Sampling and Routing
AAAI 2024
arXiv
40
citations
XTC: Extreme Compression for Pre-trained Transformers Made Simple and Efficient
NEURIPS 2022
0
citations