Quick Review: Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs
作者:XD / 发表: 2023年12月6日 23:51 / 科研学习/ 阅读量:1057
Optimize Weight Rounding via Signed Gradient Descent for the Quantization of Large Language Models
Paper: https://arxiv.org/abs/2309.05516
Code: https://github.com/intel/neural-compressor
Organization: Intel