參考文獻 |
參考文獻
[1] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.
[2] S. Mekruksavanich, A. Jitpattanakul, and N. Hnoohom, “Negative emotion recogni- tion using deep learning for thai language,” in 2020 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT & NCON), pp. 71–74, IEEE, 2020.
[3] D. Issa, M. F. Demirci, and A. Yazici, “Speech emotion recognition with deep con- volutional neural networks,” Biomedical Signal Processing and Control, vol. 59, p. 101894, 2020.
[4] P. Jackson and S. Haq, “Surrey audio-visual expressed emotion (savee) database,” University of Surrey: Guildford, UK, 2014.
[5] S. R. Livingstone and F. A. Russo, “The ryerson audio-visual database of emotional speech and song (ravdess): A dynamic, multimodal set of facial and vocal expressions in north american english,” PloS one, vol. 13, no. 5, p. e0196391, 2018.
[6] C. M. Lee, S. Narayanan, and R. Pieraccini, “Recognition of negative emotions from the speech signal,” in IEEE Workshop on Automatic Speech Recognition and Understanding, 2001. ASRU’01., pp. 240–243, IEEE, 2001.
[7] C.VaudableandL.Devillers,“Negativeemotionsdetectionasanindicatorofdialogs quality in call centers,” in 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5109–5112, IEEE, 2012.
[8] N. Hnoohom, A. Jitpattanakul, P. Inluergsri, P. Wongbudsri, and W. Ployput, “Multi-sensor-based fall detection and activity daily living classification by using en- semble learning,” in 2018 International ECTI Northern Section Conference on Elec- trical, Electronics, Computer and Telecommunications Engineering (ECTI-NCON), pp. 111–115, IEEE, 2018.
[9] N.Hnoohom,S.Mekruksavanich,andA.Jitpattanakul,“Humanactivityrecognition using triaxial acceleration data from smartphone and ensemble learning,” in 2017
41
13th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), pp. 408–412, IEEE, 2017.
[10] H. Ali, M. Hariharan, S. Yaacob, and A. H. Adom, “Facial emotion recognition using empirical mode decomposition,” Expert Systems with Applications, vol. 42, no. 3, pp. 1261–1277, 2015.
[11] A. Schirmer and R. Adolphs, “Emotion perception from face, voice, and touch: com- parisons and convergence,” Trends in cognitive sciences, vol. 21, no. 3, pp. 216–228, 2017.
[12] D.H.HubelandT.N.Wiesel,“Receptivefieldsofsingleneuronesinthecat’sstriate cortex,” The Journal of physiology, vol. 148, no. 3, pp. 574–591, 1959.
[13] K.FukushimaandS.Miyake,“Neocognitron:Aself-organizingneuralnetworkmodel for a mechanism of visual pattern recognition,” in Competition and cooperation in neural nets, pp. 267–285, Springer, 1982.
[14] A. Waibel, T. Hanazawa, G. Hinton, K. Shikano, and K. J. Lang, “Phoneme recog- nition using time-delay neural networks,” IEEE transactions on acoustics, speech, and signal processing, vol. 37, no. 3, pp. 328–339, 1989.
[15] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, pp. 1097–1105, 2012.
[16] J. Dai, Y. Li, K. He, and J. Sun, “R-fcn: Object detection via region-based fully con- volutional networks,” in Advances in neural information processing systems, pp. 379– 387, 2016.
[17] P. Booth, An introduction to human-computer interaction (psychology revivals). Psy- chology Press, 2014.
[18] E. R. Harper, T. Rodden, Y. Rogers, A. Sellen, B. Human, et al., “Human-computer interaction in the year 2020,” 2008.
[19] E. Cambria, A. Hussain, C. Havasi, and C. Eckl, “Sentic computing: Exploitation of common sense for the development of emotion-sensitive systems,” in Development of Multimodal Interfaces: Active Listening and Synchrony, pp. 148–156, Springer, 2010.
[20] K. Patil, P. Zope, and S. Suralkar, “Emotion detection from speech using mfcc & gmm,” Int. J. Eng. Res. Technol.(IJERT), vol. 1, no. 9, 2012.
[21] A. Hassan and R. I. Damper, “Multi-class and hierarchical svms for emotion recog- nition,” 2010.
42
[22] Y.-L.LinandG.Wei,“Speechemotionrecognitionbasedonhmmandsvm,”in2005 international conference on machine learning and cybernetics, vol. 8, pp. 4898–4901, IEEE, 2005.
[23] L. R. Rabiner, “A tutorial on hidden markov models and selected applications in speech recognition,” Proceedings of the IEEE, vol. 77, no. 2, pp. 257–286, 1989.
[24] T. L. Nwe, S. W. Foo, and L. C. De Silva, “Speech emotion recognition using hidden markov models,” Speech communication, vol. 41, no. 4, pp. 603–623, 2003.
[25] A. Nogueiras, A. Moreno, A. Bonafonte, and J. B. Mariño, “Speech emotion recog- nition using hidden markov models,” in Seventh European conference on speech com- munication and technology, 2001.
[26] C.-W. Hsu, C.-C. Chang, C.-J. Lin, et al., “A practical guide to support vector classification,” 2003.
[27] T. Seehapoch and S. Wongthanavasu, “Speech emotion recognition using support vector machines,” in 2013 5th international conference on Knowledge and smart technology (KST), pp. 86–91, IEEE, 2013.
[28] J. Weng, N. Ahuja, and T. S. Huang, “Cresceptron: a self-organizing neural network which grows adaptively,” in [Proceedings 1992] IJCNN International Joint Confer- ence on Neural Networks, vol. 1, pp. 576–581, IEEE, 1992.
[29] D. Bertero and P. Fung, “A first look into a convolutional neural network for speech emotion detection,” in 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp. 5115–5119, IEEE, 2017.
[30] H. Cao, D. G. Cooper, M. K. Keutmann, R. C. Gur, A. Nenkova, and R. Verma, “Crema-d: Crowd-sourced emotional multimodal actors dataset,” IEEE transactions on affective computing, vol. 5, no. 4, pp. 377–390, 2014.
[31] K. Dupuis and M. K. Pichora-Fuller, “Recognition of emotional speech for younger and older talkers: Behavioural findings from the toronto emotional speech set,” Canadian Acoustics, vol. 39, no. 3, pp. 182–183, 2011.
[32] C. Busso, M. Bulut, C.-C. Lee, A. Kazemzadeh, E. Mower, S. Kim, J. N. Chang, S. Lee, and S. S. Narayanan, “Iemocap: Interactive emotional dyadic motion capture database,” Language resources and evaluation, vol. 42, no. 4, pp. 335–359, 2008.
[33] R. Caruana, “Multitask learning,” Machine learning, vol. 28, no. 1, pp. 41–75, 1997. |