跳到主要导航 跳到搜索 跳到主要内容

TRQ: Ternary Neural Networks With Residual Quantization

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Ternary neural networks (TNNs) are potential for network acceleration by reducing the full-precision weights in network to ternary ones, e.g., {−1, 0, 1}. However, existing TNNs are mostly calculated based on rule-of-thumb quantization methods by simply thresholding operations, which causes a significant accuracy loss. In this paper, we introduce a stem-residual framework which provides new insight into ternary quantization, termed Ternary Residual Quantization (TRQ), to achieve more powerful TNNs. Rather than directly thresholding operations, TRQ recursively performs quantization on full-precision weights for a refined reconstruction by combining the binarized stem and residual parts.With such a unique quantization process, TRQ endows the quantizer with high flexibility and precision. Furthermore, our TRQ is generic, which can be easily extended to multiple bits through recursively encoded residual for a better recognition accuracy. Extensive experimental results demonstrate that the proposed method yields great recognition accuracy while being accelerated.

源语言英语
主期刊名35th AAAI Conference on Artificial Intelligence, AAAI 2021
出版商Association for the Advancement of Artificial Intelligence
8538-8546
页数9
ISBN(电子版)9781713835974
DOI
出版状态已出版 - 2021
活动35th AAAI Conference on Artificial Intelligence, AAAI 2021 - Virtual, Online
期限: 2 2月 20219 2月 2021

出版系列

姓名35th AAAI Conference on Artificial Intelligence, AAAI 2021
10A

会议

会议35th AAAI Conference on Artificial Intelligence, AAAI 2021
Virtual, Online
时期2/02/219/02/21

指纹

探究 'TRQ: Ternary Neural Networks With Residual Quantization' 的科研主题。它们共同构成独一无二的指纹。

引用此