Compensation Method of Quantized Deep Learning Models for Edge Devices.

ICCE-Taiwan(2023)

引用 0|浏览1
暂无评分
摘要
Quantization is one of the optimization methods for developing deep learning models for edge devices. Through converting the floating-point into 8bit integer or even lower bitwidth, the model’s storage size can be reduced. As the rounding error exists during the quantization process, the model performance decreases. As a result, a method that can recover model performance is needed. In this research, a compensation method for improving the performance of quantized deep learning models is proposed, which make the quantized model can achieve equal or even better performance compared to the original floating-point model.
更多
查看译文
关键词
Quantization,compensation method,edge device application
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要