Page 494 - 《软件学报》2025年第10期
P. 494

韩凯 等: 用于二值神经网络的加宽和收缩机制                                                          4891


                  [9]   Zhou SC, Wu YX, Ni ZK, Zhou XY, Wen H, Zou YH. DoReFa-Net: Training low bitwidth convolutional neural networks with low
                     bitwidth gradients. arXiv:1606.06160, 2018.
                 [10]   Hubara  I,  Courbariaux  M,  Soudry  D,  El-Yaniv  R,  Bengio  Y.  Binarized  neural  networks.  In:  Proc.  of  the  30th  Int’l  Conf.  on  Neural
                     Information Processing Systems. Barcelona: Curran Associates Inc., 2016. 4114–4122.
                 [11]   Gong C, Lu Y, Dai SR, Liu FX, Chen XW, Li T. Ultra-low loss quantization method for deep neural network compression. Ruan Jian
                     Xue Bao/Journal of Software, 2021, 32(8): 2391–2407 (in Chinese with English abstract). http://www.jos.org.cn/1000-9825/6189.htm
                     [doi: 10.13328/j.cnki.jos.006189]
                 [12]   Hinton G, Vinyals O, Dean J. Distilling the knowledge in a neural network. arXiv:1503.02531, 2015.
                 [13]   Sandler M, Howard A, Zhu ML, Zhmoginov A, Chen LC. MobileNetV2: Inverted residuals and linear bottlenecks. In: Proc. of the 2018
                     IEEE/CVF  Conf.  on  Computer  Vision  and  Pattern  Recognition.  Salt  Lake  City:  IEEE,  2018.  4510–4520.  [doi: 10.1109/CVPR.2018.
                     00474]
                 [14]   Han K, Wang YH, Tian Q, Guo JY, Xu CJ, Xu C. GhostNet: More features from cheap operations. In: Proc. of the 2020 IEEE/CVF Conf.
                     on Computer Vision and Pattern Recognition. Seattle: IEEE, 2020. 1577–1586. [doi: 10.1109/CVPR42600.2020.00165]
                 [15]   Hu  J.  Architecture  optimization  and  quantization  acceleration  of  convolutional  neural  networks  [Ph.D.  Thesis].  Beijing:  Institute  of
                     Software, Chinese Academy of Sciences, 2023 (in Chinese with English abstract).
                 [16]   Zhang DQ, Yang JL, Ye DQ, Hua G. LQ-Nets: Learned quantization for highly accurate and compact deep neural networks. In: Proc. of
                     the 15th European Conf. on Computer Vision (ECCV). Munich: Springer, 2018. 373–390. [doi: 10.1007/978-3-030-01237-3_23]
                 [17]   Lin MB, Ji RR, Xu ZH, Zhang BC, Wang Y, Wu YJ, Huang FY, Lin CW. Rotated binary neural network. In: Proc. of the 34th Int’l Conf.
                     on Neural Information Processing Systems. Vancouver: Curran Associates Inc., 2020. 7474–7485.
                 [18]   Han  S,  Mao  HZ,  Dally  WJ.  Deep  compression:  Compressing  deep  neural  networks  with  pruning,  trained  quantization  and  Huffman
                     coding. arXiv:1510.00149, 2016.
                 [19]   Courbariaux M, Bengio Y, David JP. BinaryConnect: Training deep neural networks with binary weights during propagations. In: Proc.
                     of the 29th Int’l Conf. on Neural Information Processing Systems. Montreal: MIT Press, 2015. 3123–3131.
                 [20]   Rastegari M, Ordonez V, Redmon J, Farhadi A. XNOR-Net: ImageNet classification using binary convolutional neural networks. In:
                     Proc. of the 14th European Conf. on Computer Vision. Amsterdam: Springer, 2016. 525–542. [doi: 10.1007/978-3-319-46493-0_32]
                 [21]   Liu  ZC,  Wu  BY,  Luo  WH,  Yang  X,  Liu  W,  Cheng  KT.  Bi-Real  net:  Enhancing  the  performance  of  1-bit  CNNs  with  improved
                     representational capability and advanced training algorithm. In: Proc. of the 15th European Conf. on Computer Vision (ECCV). Munich:
                     Springer, 2018. 747–763. [doi: 10.1007/978-3-030-01267-0_44]
                 [22]   Lin XF, Zhao C, Pan W. Towards accurate binary convolutional neural network. In: Proc. of the 31st Int’l Conf. on Neural Information
                     Processing Systems. Long Beach: Curran Associates Inc., 2017. 344–352.
                 [23]   Zhu  SL,  Dong  X,  Su  H.  Binary  ensemble  neural  network:  More  bits  per  network  or  more  networks  per  bit?  In:  Proc.  of  the  2019
                     IEEE/CVF Conf. on Computer Vision and Pattern Recognition. Long Beach: IEEE, 2019. 4918–4927. [doi: 10.1109/CVPR.2019.00506]
                 [24]   He  YH,  Zhang  XY,  Sun  J.  Channel  pruning  for  accelerating  very  deep  neural  networks.  In:  Proc.  of  the  2017  IEEE  Int’l  Conf.  on
                     Computer Vision. Venice: IEEE, 2017. 1398–1406. [doi: 10.1109/ICCV.2017.155]
                 [25]   He Y, Kang GL, Dong XY, Fu YW, Yang Y. Soft filter pruning for accelerating deep convolutional neural networks. In: Proc. of the 27th
                     Int’l Joint Conf. on Artificial Intelligence. Stockholm: AAAI Press, 2018. 2234–2240. [doi: 10.24963/ijcai.2018/309]
                 [26]   Liu Z, Li JG, Shen ZQ, Huang G, Yan SM, Zhang CS. Learning efficient convolutional networks through network slimming. In: Proc. of
                     the 2017 IEEE Int’l Conf. on Computer Vision. Venice: IEEE, 2017. 2755–2763. [doi: 10.1109/ICCV.2017.298]
                 [27]   He Y, Liu P, Wang ZW, Hu ZL, Yang Y. Filter pruning via geometric median for deep convolutional neural networks acceleration. In:
                     Proc. of the 2019 IEEE/CVF Conf. on Computer Vision and Pattern Recognition. Long Beach: IEEE, 2019. 4335–4344. [doi: 10.1109/
                     CVPR.2019.00447]
                 [28]   Tan MH, Gao WF, Li H, Xie J, Gong MG. Universal binary neural networks design by improved differentiable neural architecture search.
                     IEEE Trans. on Circuits and Systems for Video Technology, 2024, 34(10): 9153–9165. [doi: 10.1109/TCSVT.2024.3398691]
                 [29]   Bulat A, Martinez B, Tzimiropoulos G. BATS: Binary architecture search. In: Proc. of the 16th European Conf. on Computer Vision.
                     Glasgow: Springer, 2020. 309–325. [doi: 10.1007/978-3-030-58592-1_19]
                 [30]   Romero A, Ballas N, Kahou SE, Chassang A, Gatta C, Bengio Y. FitNets: Hints for thin deep nets. arXiv:1412.6550, 2015.
                 [31]   Zagoruyko  S,  Komodakis  N.  Paying  more  attention  to  attention:  Improving  the  performance  of  convolutional  neural  networks  via
                     attention transfer. arXiv:1612.03928, 2017.
                 [32]   Bengio Y, Léonard N, Courville A. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv:
                     1308.3432, 2013.
   489   490   491   492   493   494   495   496