Citation: | Ji Lai, Lixin Yang, Dejian Li, Chongfei Shen, Xi Feng, Jizeng Wei, Yu Liu. Design and Optimization of Winograd Convolution on Array Accelerator[J].JOURNAL OF BEIJING INSTITUTE OF TECHNOLOGY, 2023, 32(1): 69-81.doi:10.15918/j.jbit1004-0579.2022.094 |
[1] |
G. Hinton, L. Deng, D. Yu, G. E. Dahl, A. R. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. N. Sainath, and B. Kingsbury,“Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups,”
IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 82-97, 2012.
doi:10.1109/MSP.2012.2205597
|
[2] |
D. Ciregan, U. Meier, and J. Schmidhuber, “Multi-column deep neural networks for image classification,” in
2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3642-3649, 2012.
|
[3] |
J. Morajda, “Neural networks and their economic applications,” in
Artificial Intelligence and Security in Computing Systems. Springer, 2003, pp. 53-62.
|
[4] |
J. L. Patel and R. K. Goyal,“Applications of artificial neural networks in medical science,”
Current Clinical Pharmacology, vol. 2, no. 3, pp. 217-226, 2007.
doi:10.2174/157488407781668811
|
[5] |
H. Malmgren, M. Borga, and L. Niklasson, “Artificial neural networks in medicine and biology,”
in
Proceedings of the ANNIMAB-1 Conference,Göteborg, Sweden, 13-16 May 2000. Springer Science & Business Media, 2012.
|
[6] |
J. Albericio, P. Judd, T. Hetherington, T. Aamodt, N. E. Jerger, and A. Moshovos,“Cnvlutin: Ineffectual-neuron-free deep neural network computing,”
ACM SIGARCH Computer Architecture News, vol. 44, no. 3, pp. 1-13, 2016.
doi:10.1145/3007787.3001138
|
[7] |
A. Krizhevsky, I. Sutskever, and G. E. Hinton,“Imagenet classification with deep convolutional neural networks,”
Communications of the ACM, vol. 60, no. 6, pp. 84-90, 2017.
doi:10.1145/3065386
|
[8] |
K. Fukushima,“A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position,”
Biol. Cybern., vol. 36, pp. 193-202, 1980.
doi:10.1007/BF00344251
|
[9] |
K. Simonyan and A. Zisserman,“Two-stream convolutional networks for action recognition in videos,”
Advances in Neural Information Processing Systems, vol. 27, pp. 568-576, 2014.
|
[10] |
A. Parashar, M. Rhu, A. Mukkara, A. Puglielli, R. Venkatesan, B. Khailany, J. Emer, S. W. Keckler, and W. J. Dally,“SCNN: An accelerator for compressed-sparse convolutional neural networks,”
ACM SIGARCH Computer Architecture News, vol. 45, no. 2, pp. 27-40, 2017.
doi:10.1145/3140659.3080254
|
[11] |
Y. Cai, K. Zhou, X. Xue, M. Wang, and X. Zeng, “Nonvolatile binary CNN accelerator with extremely low standby power using rram for iot applications,” in
2019 IEEE 13th International Conference on ASIC (ASICON), pp. 1-4, 2019.
|
[12] |
A. Lavin and S. Gray, “Fast algorithms for convolutional neural networks,” in
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4013-4021, 2016.
|
[13] |
Y. Cao, X. Wei, T. Qiao, and H. Chen, “FPGA-based accelerator for convolution operations,” in
2019 IEEE International Conference on Signal, Information and Data Processing (ICSIDP), pp. 1-5, 2019.
|
[14] |
Y. Chen, T. Luo, S. Liu, S. Zhang, L. He, J. Wang, L. Li, T. Chen, Z. Xu, N. Sun, and O. Temam, “Dadiannao: A machine-learning supercomputer,” in
2014 47th Annual IEEE/ACM International Symposium on Microarchitecture, pp. 609-622, 2014.
|
[15] |
B. Barabasz and D. Gregg, “Winograd convolution for DNNs: Beyond linear polynomials,” in
International Conference of the Italian Association for Artificial Intelligence, pp. 307-320, 2019.
|
[16] |
K. Vincent, K. Stephano, M. A. Frumkin, B. Ginsburg, and J. Demouth, “On improving the numerical stability of winograd convolutions,” in
5th International Conference on Learning Representations, ICLR 2017, pp. 1-4, 2017.
|
[17] |
L. Meng and J. Brothers, “Efficient winograd convolution via integer arithmetic,”
arXiv preprint,arXiv: 1901.01965, 2019.
|
[18] |
L. Lu, Y. Liang, Q. Xiao, and S. Yan, “Evaluating fast algorithms for convolutional neural networks on FPGAs,” in
2017 IEEE 25th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM), pp. 101-108, 2017.
|
[19] |
D. Huang, X. Zhang, R. Zhang, T. Zhi, D. He, J. Guo, C. Liu, Q. Guo, Z. Du, S. Liu
et al., “DWM: A decomposable winograd method for convolution acceleration,” in
Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 4, pp. 4174-4181, 2020.
|
[20] |
Y. Umuroglu, N. J. Fraser, G. Gambardella, M. Blott, P. Leong, M. Jahre, and K. Vissers, “Finn: A framework for fast, scalable binarized neural network inference,” in
Proceedings of the 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, pp. 65-74, 2017.
|
[21] |
K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770-778, 2016.
|
[22] |
O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei,“Imagenet large scale visual recognition challenge,”
International Journal of Computer Vision, vol. 115, no. 3, pp. 211-252, 2015.
doi:10.1007/s11263-015-0816-y
|
[23] |
Y. Guo, A. Yao, and Y. Chen,“Dynamic network surgery for efficient DNNs,”
Advances in Neural Information Processing Systems, vol. 29, pp. 1387-1395, 2016.
|