Cheng Li
Cheng Li
Home
Experience
Publications
Talks
Languages
Contact
Xiaoxia Wu
Latest
ZeroQuant-V2: Exploring Post-training Quantization in LLMs from Comprehensive Study to Low Rank Compensation
A Comprehensive Study on Post-Training Quantization for Large Language Models
Understanding INT4 Quantization for Transformer Models: Latency Speedup, Composability, and Failure Cases
Random-LTD: Random and Layerwise Token Dropping Brings Efficient Training for Large-scale Transformers
Cite
×