Page 483 - 《软件学报》2025年第10期
P. 483

软件学报 ISSN 1000-9825, CODEN RUXUEW                                        E-mail: jos@iscas.ac.cn
                 2025,36(10):4880−4892 [doi: 10.13328/j.cnki.jos.007363] [CSTR: 32375.14.jos.007363]  http://www.jos.org.cn
                 ©中国科学院软件研究所版权所有.                                                          Tel: +86-10-62562563



                                                              *
                 用于二值神经网络的加宽和收缩机制

                 韩    凯  1,2 ,    刘传建  3 ,    吴恩华  1,4


                 1
                  (计算机科学国家重点实验室 (中国科学院 软件研究所), 北京 100190)
                 2
                  (中国科学院大学, 北京 100049)
                 3
                  (华为诺亚方舟实验室, 北京 100085)
                 4
                  (澳门大学 科技学院, 澳门 999078)
                 通信作者: 吴恩华, E-mail: ehwu@um.edu.mo
                 摘 要: 二值神经网络      (binary neural network, BNN) 因其较少的计算和存储开销而对业界非常有吸引力, 但其准确
                 率仍然比全精度参数的网络差. 大多数现有方法旨在通过利用更有效的训练技术来提高二值神经网络的性能. 然
                 而, 通过实验发现量化后特征的表示能力远弱于全精度的特征. 因此, 提出一种加宽和收缩机制来构建高精度而紧
                 凑的二值神经网络. 首先, 通过将原始全精度网络中的特征投影到高维量化特征来解决量化特征表示能力弱的问
                 题. 同时, 冗余的量化特征将被消除, 以避免某些特征维度的过度增长. 进而建立一个紧凑但具有足够表示能力的
                 量化神经网络. 基准数据集上的实验结果表明, 该方法能够以更少的参数量和计算量建立高精度二值神经网络, 其
                 准确率与全精度基线模型几乎相同, 例如, 二值量化的                ResNet-18 在  ImageNet 数据集上达到了  70%  的准确率.
                 关键词: 神经网络; 模型量化; 图像分类; 目标检测
                 中图法分类号: TP301

                 中文引用格式: 韩凯, 刘传建, 吴恩华. 用于二值神经网络的加宽和收缩机制. 软件学报, 2025, 36(10): 4880–4892. http://www.jos.
                 org.cn/1000-9825/7363.htm
                 英文引用格式: Han K, Liu CJ, Wu EH. Widening and Squeezing Mechanism for Binary Neural Network. Ruan Jian Xue Bao/Journal
                 of Software, 2025, 36(10): 4880–4892 (in Chinese). http://www.jos.org.cn/1000-9825/7363.htm

                 Widening and Squeezing Mechanism for Binary Neural Network
                                      3
                        1,2
                 HAN Kai , LIU Chuan-Jian , WU En-Hua 1,4
                 1
                 (State Key Laboratory of Computer Science (Institute of Software, Chinese Academy of Sciences), Beijing 100190, China)
                 2
                 (University of Chinese Academy of Sciences, Beijing 100049, China)
                 3
                 (Huawei Noah’s Ark Lab, Beijing 100085, China)
                 4
                 (Faculty of Science and Technology, University of Macau, Macao 999078, China)
                 Abstract:  Binary  neural  networks  (BNNs)  are  highly  appealing  to  the  industry  due  to  their  significantly  reduced  computation  and  storage
                 requirements.  However,  their  accuracy  still  lags  behind  that  of  networks  with  full-precision  parameters.  Most  existing  methods  focus  on
                 improving  the  performance  of  BNNs  through  advanced  training  techniques.  Empirical  findings  reveal  that  the  representation  capability  of
                 quantized  features  is  considerably  weaker  than  that  of  full-precision  features.  To  address  this  limitation,  a  widening  and  squeezing
                 mechanism  is  proposed  to  construct  high-accuracy  yet  compact  BNNs.  Specifically,  features  from  the  original  full-precision  networks  are
                 projected  into  high-dimensional  quantized  features  to  mitigate  the  representation  gap.  Meanwhile,  redundant  quantized  features  are  pruned
                 to  prevent  the  over  growth  of  feature  dimensions.  As  a  result,  a  compact  yet  sufficiently  expressive  quantized  neural  network  is
                 constructed.  Experimental  results  on  benchmark  datasets  demonstrate  that  the  proposed  method  achieves  high-accuracy  BNNs  with
                 significantly  fewer  parameters  and  computations  while  delivering  performance  comparable  to  full-precision  baseline  models.  For  instance,


                 *    基金项目: 国家自然科学基金  (62332015)
                  收稿时间: 2023-11-09; 修改时间: 2024-06-17, 2024-09-18; 采用时间: 2024-11-26; jos 在线出版时间: 2025-07-17
                  CNKI 网络首发时间: 2025-07-18
   478   479   480   481   482   483   484   485   486   487   488