參考文獻 |
C. Szegedy, A. Toshev, and D. Erhan, “Deep Neural Networks for object detection,” 2013.
[2] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. Alemi, “Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning.”
[3] M. Tan and Q. V. Le, “MixConv: Mixed Depthwise Convolutional Kernels,” 2019, [Online]. Available: http://arxiv.org/abs/1907.09595
[4] S. Han et al., “ESE: Efficient Speech Recognition Engine with Sparse LSTM on FPGA,” 2016, [Online]. Available: http://arxiv.org/abs/1612.00694
[5] Y.-H. Chen, T. Krishna, J. Emer, and V. Sze, “Eyeriss: An energy-efficient reconfigurable accelerator for deep convolutional neural networks,” IEEE Int. J. Solid State Circuits, vol. 59, no. 1, pp. 262–263, 2016, doi: 10.1109/ISSCC.2016.7418007.
[6] Y. Chen et al., “DaDianNao: A Machine-Learning Supercomputer,” in Proceedings of the Annual International Symposium on Microarchitecture, MICRO, Jan. 2015, vol. 2015-Janua, no. January, pp. 609–622. doi: 10.1109/MICRO.2014.58.
[7] M. McCloskey and N. J. Cohen, “Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem,” Psychol. Learn. Motiv. - Adv. Res. Theory, vol. 24, no. C, pp. 109–165, Jan. 1989, doi: 10.1016/S0079-7421(08)60536-8.
[8] S. A. Rebuffi, A. Kolesnikov, G. Sperl, and C. H. Lampert, “iCaRL: Incremental classifier and representation learning,” in Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, 2017, vol. 2017-Janua, pp. 5533–5542. doi: 10.1109/CVPR.2017.587.
[9] Z. Li and D. Hoiem, “Learning without Forgetting,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, no. 12, pp. 2935–2947, 2018, doi: 10.1109/TPAMI.2017.2773081.
[10] A. Rosenfeld and J. K. Tsotsos, “Incremental Learning through Deep Adaptation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 42, no. 3, pp. 651–663, 2020, doi: 10.1109/TPAMI.2018.2884462.
[11] P. Dhar, R. V. Singh, K. C. Peng, Z. Wu, and R. Chellappa, “Learning without memorizing,” in IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2019, vol. 2019-June, pp. 5133–5141. doi: 10.1109/CVPR.2019.00528.
[12] A. A. Rusu et al., “Progressive Neural Networks,” Jun. 2016, Accessed: Sep. 02, 2020. [Online]. Available: http://arxiv.org/abs/1606.04671
[13] S. S. Sarwar, A. Ankit, and K. Roy, “Incremental Learning in Deep Convolutional Neural Networks Using Partial Network Sharing,” IEEE Access, vol. 8, pp. 4615–4628, 2020, doi: 10.1109/ACCESS.2019.2963056.
[14] L. Pellegrini, G. Graffieti, V. Lomonaco, and D. Maltoni, “Latent Replay for Real-Time Continual Learning,” 2019, [Online]. Available: http://arxiv.org/abs/1912.01100
[15] M. A. Hussain, S.-A. Huang, and T.-H. Tsai, “Learning With Sharing: An Edge-Optimized Incremental Learning Method for Deep Neural Networks,” IEEE Trans. Emerg. Top. Comput., 2022.
[16] J. Howard and S. Gugger, “Fastai: A Layered API for Deep Learning,” Information, vol. 11, no. 2, p. 108, Feb. 2020, doi: 10.3390/info11020108.
[17] M. A. Hussain, C.-L. Lee, and T.-H. Tsai, “An Efficient Incremental Learning Algorithm for Sound Classification,” IEEE Multimed., pp. 1–8, 2022, doi: 10.1109/MMUL.2022.3208923.
[18] M. A. Hussain and T. H. Tsai, “Memory Access Optimization for On-Chip Transfer Learning,” IEEE Trans. Circuits Syst. I Regul. Pap., vol. 68, no. 4, pp. 1507–1519, Feb. 2021, doi: 10.1109/TCSI.2021.3055281.
[19] M. A. Hussain and T.-H. Tsai, “An Efficient and Fast Softmax Hardware Architecture (EFSHA) for Deep Neural Networks,” in 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS), Jun. 2021, pp. 1–4. doi: 10.1109/AICAS51828.2021.9458541.
[20] J. Knoblauch, H. Husain, and T. Diethe, “Optimal Continual Learning has Perfect Memory and is NP-hard,” no. Cl, 2020, [Online]. Available: http://arxiv.org/abs/2006.05188
[21] L. Binyan, W. Yanbo, C. Zhihong, L. Jiayu, and L. Junqin, “Object detection and robotic sorting system in complex industrial environment,” in Proceedings - 2017 Chinese Automation Congress, CAC 2017, Dec. 2017, vol. 2017-January, pp. 7277–7281. doi: 10.1109/CAC.2017.8244092.
[22] C. Carpineti, V. Lomonaco, L. Bedogni, M. Di Felice, and L. Bononi, “Custom Dual Transportation Mode Detection by Smartphone Devices Exploiting Sensor Diversity,” in 2018 IEEE International Conference on Pervasive Computing and Communications Workshops, PerCom Workshops 2018, Oct. 2018, pp. 367–372. doi: 10.1109/PERCOMW.2018.8480119.
[23] P. Casale, O. Pujol, and P. Radeva, “Human activity recognition from accelerometer data using a wearable device,” in International Conference on Pattern Recognition and Image Analysis, 2011, vol. 6669 LNCS, pp. 289–296. doi: 10.1007/978-3-642-21257-4_36/COVER.
[24] D. Roy, P. Panda, and K. Roy, “Tree-CNN: A hierarchical Deep Convolutional Neural Network for incremental learning,” Neural Networks, vol. 121, pp. 148–160, 2020, doi: 10.1016/j.neunet.2019.09.010.
[25] T. Xiao, J. Zhang, K. Yang, Y. Peng, and Z. Zhang, “Error-driven incremental learning in deep convolutional neural network for large-scale image classification,” in MM 2014 - Proceedings of the 2014 ACM Conference on Multimedia, Nov. 2014, pp. 177–186. doi: 10.1145/2647868.2654926.
[26] F. M. Castro, M. J. Marín-Jiménez, N. Guil, C. Schmid, and K. Alahari, “End-to-end incremental learning,” in European Conference on Computer Vision, 2018, vol. 11216 LNCS, pp. 241–257. doi: 10.1007/978-3-030-01258-8_15.
[27] A. G. Howard et al., “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,” Apr. 2017, Accessed: May 23, 2020. [Online]. Available: http://arxiv.org/abs/1704.04861
[28] V. Lomonaco and D. Maltoni, “CORe50: a New Dataset and Benchmark for Continuous Object Recognition,” 2017, vol. 78, pp. 17–26. [Online]. Available: http://proceedings.mlr.press/v78/lomonaco17a.html
[29] J. Zhang et al., “Class-incremental learning via deep model consolidation,” in IEEE Winter Conference on Applications of Computer Vision, WACV 2020, 2020, pp. 1120–1129. doi: 10.1109/WACV45572.2020.9093365.
[30] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2016, vol. 2016-December, pp. 770–778. doi: 10.1109/CVPR.2016.90.
[31] S. A. Rebuffi, A. Kolesnikov, G. Sperl, and C. H. Lampert, “iCaRL: Incremental classifier and representation learning,” in Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Nov. 2017, vol. 2017-January, pp. 5533–5542. doi: 10.1109/CVPR.2017.587.
[32] Y. Wu et al., “Large Scale Incremental Learning,” pp. 374–382, 2019, [Online]. Available: http://arxiv.org/abs/1905.13260
[33] D. Lopez-Paz and M. Ranzato, “Gradient Episodic Memory for Continual Learning,” Adv. Neural Inf. Process. Syst., vol. 2017-December, pp. 6468–6477, Jun. 2017, Accessed: Jan. 24, 2021. [Online]. Available: http://arxiv.org/abs/1706.08840
[34] A. Douillard, M. Cord, C. Ollion, T. Robert, and E. Valle, “PODNet: Pooled Outputs Distillation for Small-Tasks Incremental Learning,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Apr. 2020, vol. 12365 LNCS, pp. 86–102. doi: 10.1007/978-3-030-58565-5_6.
[35] A. Mallya and S. Lazebnik, “PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Dec. 2018, pp. 7765–7773. doi: 10.1109/CVPR.2018.00810.
[36] C. Käding, E. Rodner, A. Freytag, and J. Denzler, “Fine-tuning deep neural networks in continuous learning scenarios,” in Asian Conference on Computer Vision (ACCV), Nov. 2017, vol. 10118 LNCS, pp. 588–605. doi: 10.1007/978-3-319-54526-4_43.
[37] A. Howard et al., “Searching for mobileNetV3,” in Proceedings of the IEEE International Conference on Computer Vision, Oct. 2019, vol. 2019-October, pp. 1314–1324. doi: 10.1109/ICCV.2019.00140.
[38] B. Wu et al., “FBNET: Hardware-aware efficient convnet design via differentiable neural architecture search,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 2019, vol. 2019-June, pp. 10726–10734. doi: 10.1109/CVPR.2019.01099.
[39] M. Tan et al., “MnasNet: Platform-Aware Neural Architecture Search for Mobile,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jul. 2018, vol. 2019-June, pp. 2815–2823. Accessed: May 23, 2020. [Online]. Available: http://arxiv.org/abs/1807.11626
[40] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, “MobileNetV2: Inverted Residuals and Linear Bottlenecks,” in Proceedings - IEEE International Conference on Computer Vision and Pattern Recognition, Jan. 2018, pp. 4510–4520. Accessed: Jun. 06, 2020. [Online]. Available: http://arxiv.org/abs/1801.04381
[41] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” Dec. 2014.
[42] A. Krizhevsky, “Learning Multiple Layers of Features from Tiny Images,” 2009.
[43] L. Fei-Fei, R. Fergus, and P. Perona, “Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories,” in IEEE International Conference on Computer Vision and Pattern Recognition Workshops, 2004, vol. 2004-Janua, no. January. doi: 10.1109/CVPR.2004.383.
[44] C. Wah, S. Banson, P. Welinder, P. Perona, and S. Belongie, “The Caltech-UCSD Birds-200-2011 Dataset,” in IFAC Proceedings Volumes, 2009, vol. 42, no. 15, pp. 50–57. doi: 10.3182/20090902-3-US-2007.0059.
[45] J. Salamon, C. Jacoby, and J. P. Bello, “A Dataset and Taxonomy for Urban Sound Research,” in Proceedings of the 22nd ACM international conference on Multimedia, 2014, pp. 1041–1044. doi: 10.1145/2647868.
[46] K. J. Piczak, “ESC: Dataset for environmental sound classification,” in Proceedings of the 2015 ACM Multimedia Conference, Oct. 2015, pp. 1015–1018. doi: 10.1145/2733373.2806390.
[47] A. Mesaros, T. Heittola, and T. Virtanen, “TUT database for acoustic scene classification and sound event detection,” in European Signal Processing Conference, Nov. 2016, vol. 2016-November, pp. 1128–1132. doi: 10.1109/EUSIPCO.2016.7760424.
[48] J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, and G. Desjardins, “Overcoming catastrophic forgetting in neural networks,” Pnas, pp. 1–6, 2016, [Online]. Available: http://www.pnas.org/content/suppl/2017/03/14/1611835114.DCSupplemental/pnas.201611835SI.pdf
[49] R. Aljundi, F. Babiloni, M. Elhoseiny, M. Rohrbach, and T. Tuytelaars, “Memory Aware Synapses: Learning What (not) to Forget,” in Proceedings - European Conference on Computer Vision, 2018, pp. 144–161.
[50] D. Lopez-Paz and M. Ranzato, “Gradient Episodic Memory for Continual Learning,” in Advances in Neural Information Processing Systems, Jun. 2017, vol. 2017-Decem, pp. 6468–6477. Accessed: Jan. 24, 2021. [Online]. Available: http://arxiv.org/abs/1706.08840
[51] L. Jin, H. Liang, and C. Yang, “Class-Incremental Learning of Convolutional Neural Networks Based on Double Consolidation Mechanism,” IEEE Access, vol. 8, pp. 172553–172562, Sep. 2020, doi: 10.1109/access.2020.3025558.
[52] J. O. Zhang, A. Sax, A. Zamir, L. Guibas, and J. Malik, “Side-Tuning: A Baseline for Network Adaptation via Additive Side Networks,” in European Conference on Computer Vision (ECCV) , Dec. 2019, vol. 12348 LNCS, pp. 698–714. Accessed: Jan. 24, 2021. [Online]. Available: http://arxiv.org/abs/1912.13503
[53] “jetson_stats.” https://github.com/rbonghi/jetson_stats (accessed Nov. 17, 2020).
[54] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization,” in Proceedings of the IEEE International Conference on Computer Vision, Dec. 2017, vol. 2017-October, pp. 618–626. doi: 10.1109/ICCV.2017.74.
[55] A. Aimar et al., “NullHop: A Flexible Convolutional Neural Network Accelerator Based on Sparse Representations of Feature Maps,” IEEE Trans. Neural Networks Learn. Syst., vol. 30, no. 3, pp. 644–656, 2019, doi: 10.1109/TNNLS.2018.2852335.
[56] W. Lu, G. Yan, J. Li, S. Gong, Y. Han, and X. Li, “FlexFlow: A Flexible Dataflow Accelerator Architecture for Convolutional Neural Networks,” in Proceedings - International Symposium on High-Performance Computer Architecture, 2017, pp. 553–564. doi: 10.1109/HPCA.2017.29.
[57] Y. Ma, Y. Cao, S. Vrudhula, and J. S. Seo, “Automatic Compilation of Diverse CNNs Onto High-Performance FPGA Accelerators,” IEEE Trans. Comput. Des. Integr. Circuits Syst., vol. 39, no. 2, pp. 424–437, Feb. 2020, doi: 10.1109/TCAD.2018.2884972.
[58] H. Zhu et al., “A Communication-Aware DNN Accelerator on ImageNet Using In-Memory Entry-Counting Based Algorithm-Circuit-Architecture Co-Design in 65-nm CMOS,” IEEE J. Emerg. Sel. Top. Circuits Syst., vol. 10, no. 3, pp. 283–294, Sep. 2020, doi: 10.1109/JETCAS.2020.3014920.
[59] T. Yuan, W. Liu, J. Han, and F. Lombardi, “High Performance CNN Accelerators Based on Hardware and Algorithm Co-Optimization,” IEEE Trans. Circuits Syst. I Regul. Pap., 2020, doi: 10.1109/TCSI.2020.3030663.
[60] S. Kim, J. Lee, S. Kang, J. Lee, and H. J. Yoo, “A Power-Efficient CNN Accelerator with Similar Feature Skipping for Face Recognition in Mobile Devices,” IEEE Trans. Circuits Syst. I Regul. Pap., vol. 67, no. 4, pp. 1181–1193, Apr. 2020, doi: 10.1109/TCSI.2020.2966243.
[61] X. Zhou, L. Zhang, C. Guo, X. Yin, and C. Zhuo, “A Convolutional Neural Network Accelerator Architecture with Fine-Granular Mixed Precision Configurability,” Sep. 2020, pp. 1–5. doi: 10.1109/iscas45731.2020.9180844.
[62] Z. Xu, Z. Yang, J. Xiong, J. Yang, and X. Chen, “ELFISH: Resource-Aware Federated Learning on Heterogeneous Edge Devices,” Dec. 2019.
[63] S. Caldas, J. Konečny, H. B. McMahan, and A. Talwalkar, “Expanding the Reach of Federated Learning by Reducing Client Resource Requirements,” Dec. 2018, Accessed: Sep. 24, 2020. [Online]. Available: http://arxiv.org/abs/1812.07210
[64] S. Wang et al., “When Edge Meets Learning: Adaptive Control for Resource-Constrained Distributed Machine Learning,” in Proceedings - IEEE INFOCOM, Oct. 2018, vol. 2018-April, pp. 63–71. doi: 10.1109/INFOCOM.2018.8486403.
[65] H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-Efficient Learning of Deep Networks from Decentralized Data,” Proc. 20th Int. Conf. Artif. Intell. Stat. AISTATS 2017, Feb. 2016, Accessed: Sep. 24, 2020. [Online]. Available: http://arxiv.org/abs/1602.05629
[66] P. N. Whatmough, C. Zhou, P. Hansen, S. K. Venkataramanaiah, J. Seo, and M. Mattina, “FixyNN: Efficient Hardware for Mobile Computer Vision via Transfer Learning,” Feb. 2019, Accessed: May 24, 2020. [Online]. Available: http://arxiv.org/abs/1902.11128
[67] D. Han, J. Lee, J. Lee, and H. J. Yoo, “A Low-Power Deep Neural Network Online Learning Processor for Real-Time Object Tracking Application,” IEEE Trans. Circuits Syst. I Regul. Pap., vol. 66, no. 5, pp. 1794–1804, May 2019, doi: 10.1109/TCSI.2018.2880363.
[68] M. Nazemi, G. Pasandi, and M. Pedram, “NullaNet: Training Deep Neural Networks for Reduced-Memory-Access Inference,” Jul. 2018, Accessed: Aug. 04, 2020. [Online]. Available: http://arxiv.org/abs/1807.08716
[69] B. Jacob et al., “Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Dec. 2018, pp. 2704–2713. doi: 10.1109/CVPR.2018.00286.
[70] T. Zhang et al., “A systematic DNN weight pruning framework using alternating direction method of multipliers,” in European Conference on Computer Vision, Sep. 2018, vol. 11212 LNCS, pp. 191–207. doi: 10.1007/978-3-030-01237-3_12.
[71] G. Hinton, O. Vinyals, and J. Dean, “Distilling the Knowledge in a Neural Network,” Mar. 2015, Accessed: Aug. 05, 2020. [Online]. Available: http://arxiv.org/abs/1503.02531
[72] S. Rabanser, O. Shchur, and S. Günnemann, “Introduction to Tensor Decompositions and their Applications in Machine Learning,” Nov. 2017, Accessed: Aug. 05, 2020. [Online]. Available: http://arxiv.org/abs/1711.10781
[73] C. F. Jhu, P. Liu, and J. J. Wu, “Data Pinning and Back Propagation Memory Optimization for Deep Learning on GPU,” in Proceedings - 2018 6th International Symposium on Computing and Networking, CANDAR 2018, Dec. 2018, pp. 19–28. doi: 10.1109/CANDAR.2018.00011.
[74] H. Cui, H. Zhang, G. R. Ganger, P. B. Gibbons, and E. P. Xing, “GeePS: Scalable deep learning on distributed GPUs with a GPU-specialized parameter server,” Apr. 2016. doi: 10.1145/2901318.2901323.
[75] I. Gelado, J. E. Stone, J. Cabezas, S. Patel, N. Navarro, and W. W. Hwu, “An asymmetric distributed shared memory model for heterogeneous parallel systems,” ACM SIGARCH Comput. Archit. News, vol. 38, no. 1, pp. 347–358, Mar. 2010, doi: 10.1145/1735970.1736059.
[76] T. Chen, B. Xu, C. Zhang, and C. Guestrin, “Training Deep Nets with Sublinear Memory Cost,” Apr. 2016, Accessed: Aug. 07, 2020. [Online]. Available: http://arxiv.org/abs/1604.06174
[77] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Commun. ACM, vol. 60, no. 6, pp. 84–90, Jun. 2012, doi: 10.1145/3065386.
[78] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” Sep. 2015.
[79] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2323, 1998, doi: 10.1109/5.726791.
[80] M. D. Zeiler and R. Fergus, “Visualizing and understanding convolutional networks,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2014, vol. 8689 LNCS, no. PART 1, pp. 818–833. doi: 10.1007/978-3-319-10590-1_53.
[81] K. H. Koo, W. H. Ryu, S. M. Lee, B. K. Choi, and C. Jo, “Versatile IO circuit schemes for LPDDR4 with 1.8mW/Gbps/pin power efficiency,” 2014. [Online]. Available: http://www.oldfriend.url.tw/article/IEEE_paper/DDR4_LPDDR4/7_WE3Slides_VersatileIOCircuitSchemesforLPDDR4.pdf
[82] T. J. Yang, Y. H. Chen, and V. Sze, “Designing energy-efficient convolutional neural networks using energy-aware pruning,” in Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Nov. 2017, vol. 2017-Janua, pp. 6071–6079. doi: 10.1109/CVPR.2017.643.
[83] P. Nayak, D. Zhang, and S. Chai, “Bit Efficient Quantization for Deep Neural Networks,” Oct. 2019, Accessed: Sep. 30, 2020. [Online]. Available: http://arxiv.org/abs/1910.04877
[84] “pyRAPL.” https://github.com/powerapi-ng/pyRAPL (accessed Nov. 17, 2020).
[85] S. Chakradhar, M. Sankaradas, V. Jakkula, and S. Cadambi, “A dynamically configurable coprocessor for convolutional neural networks,” ACM SIGARCH Comput. Archit. News, vol. 38, no. 3, p. 247, 2012, doi: 10.1145/1816038.1815993.
[86] Y. Umuroglu et al., “FINN: A framework for fast, scalable binarized neural network inference,” FPGA 2017 - Proc. 2017 ACM/SIGDA Int. Symp. Field-Programmable Gate Arrays, no. February, pp. 65–74, 2017, doi: 10.1145/3020078.3021744.
[87] S. Choi, J. Sim, M. Kang, Y. Choi, H. Kim, and L. S. Kim, “An Energy-Efficient Deep Convolutional Neural Network Training Accelerator for in Situ Personalization on Smart Devices,” IEEE J. Solid-State Circuits, vol. 55, no. 10, pp. 2691–2702, Oct. 2020, doi: 10.1109/JSSC.2020.3005786.
[88] B. Fleischer et al., “A Scalable Multi-TeraOPS Deep Learning Processor Core for AI Training and Inference,” in IEEE Symposium on VLSI Circuits, Digest of Technical Papers, Oct. 2018, vol. 2018-June, pp. 35–36. doi: 10.1109/VLSIC.2018.8502276.
[89] Z. Yuan et al., “Sticker: A 0.41-62.1 TOPS/W 8Bit Neural Network Processor with Multi-Sparsity Compatible Convolution Arrays and Online Tuning Acceleration for Fully Connected Layers,” in IEEE Symposium on VLSI Circuits, Oct. 2018, vol. 2018-June, pp. 33–34. doi: 10.1109/VLSIC.2018.8502404.
[90] J. Lee, J. Lee, D. Han, J. Lee, G. Park, and H. J. Yoo, “LNPU: A 25.3TFLOPS/W Sparse Deep-Neural-Network Learning Processor with Fine-Grained Mixed Precision of FP8-FP16,” in Digest of Technical Papers - IEEE International Solid-State Circuits Conference, Mar. 2019, vol. 2019-Febru, pp. 142–144. doi: 10.1109/ISSCC.2019.8662302.
[91] X. Chen, C. Gao, T. Delbruck, and S.-C. Liu, “EILE: Efficient Incremental Learning on the Edge,” in 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS), Jun. 2021, pp. 1–4. doi: 10.1109/AICAS51828.2021.9458554.
[92] C. Chen, H. Ding, H. Peng, H. Zhu, Y. Wang, and C. J. R. Shi, “OCEAN: An On-Chip Incremental-Learning Enhanced Artificial Neural Network Processor with Multiple Gated-Recurrent-Unit Accelerators,” IEEE J. Emerg. Sel. Top. Circuits Syst., vol. 8, no. 3, pp. 519–530, 2018, doi: 10.1109/JETCAS.2018.2852780.
[93] J. Shin, S. Choi, Y. Choi, and L. S. Kim, “A pragmatic approach to on-device incremental learning system with selective weight updates,” Proc. - Des. Autom. Conf., vol. 2020-July, Jul. 2020, doi: 10.1109/DAC18072.2020.9218507.
[94] C. S. Turner, “A fast binary logarithm algorithm,” IEEE Signal Process. Mag., vol. 27, no. 5, 2010, doi: 10.1109/MSP.2010.937503.
[95] D. Kim, J. Kung, and S. Mukhopadhyay, “A power-aware digital multilayer perceptron accelerator with on-chip training based on approximate computing,” IEEE Trans. Emerg. Top. Comput., vol. 5, no. 2, pp. 164–178, Apr. 2017, doi: 10.1109/TETC.2017.2673548.
[96] C.-H. Lu, Y.-C. Wu, and C.-H. Yang, “A 2.25 TOPS/W Fully-Integrated Deep CNN Learning Processor with On-Chip Training,” in IEEE Asian Solid-State Circuits Conference (A-SSCC), Apr. 2019, pp. 65–68. doi: 10.1109/a-sscc47793.2019.9056967.
|