Robuta

https://arxiv.org/abs/2303.08302 [2303.08302] ZeroQuant-V2: Exploring Post-training Quantization in LLMs from Comprehensive Study to... Abstract page for arXiv paper 2303.08302: ZeroQuant-V2: Exploring Post-training Quantization in LLMs from Comprehensive Study to Low Rank Compensation post training quantization https://arxiv.org/abs/2310.17723 [2310.17723] ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization... Abstract page for arXiv paper 2310.17723: ZeroQuant-HERO: Hardware-Enhanced Robust Optimized Post-Training Quantization Framework for W8A8 Transformers post trainingzeroquanthero