參考文獻 |
參考文獻
[1] T.S. Chi, P. Ru and S. Shamma, “Multiresolution spectrotemporal analysis of complex sounds,” , Journal of the Acoustical Society of America, vol. 118, no. 2, pp 887-906, 2005.
[2] Available on: http://neural.cs.washington.edu/
[3] Z.M. Dan and F.S. Monica, “A study about MFCC relevance in emotion classification for SRoL database,” Electrical and Electronics Engineering (ISEEE ,pp.1,4, 11-13 Oct. 2013.
[4] M. Lagunas and A.I. Perez-Neira,”Spectral subtraction and spectral estimation,” IEEE Workshop on Statistical Signal and Array Processing Proceedings of the Tenth, pp.383-386, 2000.
[5] 結合雜訊抑制與帶聲語音重建之語音增強系統,廖育志。
[6] PESQ Introduction http://www.pal-acoustics.com/index.php?a=services&id=143&lang=cn
[7] L. Rabiner, “A tutorial on hidden Markov models and selected applications in speech recognition,” Proceedings of the IEEE , vol.77, no.2, pp.257,286, Feb 1989.
[8] 語音訊號與系統,王小川,第二版,全華圖書,民國98年
[9] C.C. Hsu, T.E Lin, J.H. Chen and T.S. Chi, “Spectro-temporal subband Wiener filter for speech enhancement,” IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.4001,4004, 25-30 March 2012
[10] J. Woojay and B. H. Juang, “Speech analysis in a model of the central auditory system,” IEEE Transactions on Audio, Speech, and Language Processing, vol. 15, no. 6, pp.1802-1817, Aug. 2008.
[11] X. Yang, K. Wang, and S. A. Shamma, “Auditory representations of acoustic signals,” IEEE Trans. Inf. Theory, vol. 38, no. 2, pp. 824–839, Mar. 1992.
[12] Y. Ephraim, “Gain-adapted hidden markov models for recognition of clean and noisy speech,” IEEE Trans. Signal Process., vol. 40, no. 6,pp. 1303–1316, Jun. 1992.
[13] C. Corredor-Ardoy, L. Lamel, M. Adda-Decker and J-L Gauvain, “Multilingual phone recognition of spontaneous telephone speech,” IEEE International Conference on Acoustics, Speech and Signal Processing, vol.1, no., pp.413,416 vol.1, 12-15 May 1998
[14] Chunghwa Digits Database. [Online] Available: http://www.aclclp.org.tw/use_mat_c.php
[15] X. Q. Zhao and J. Wang, “A new noisy speech recognition method,” in Proc. IEEE International Symposium on Communications and Information Technology, Oct. 2005, pp.292-296.
[16] S. H. Choi, H. K. Kim, and H. S. Lee, “Speech recognition using quantized LSP parameters and their transformations in digital communication,” Speech Communication, vol. 30, pp. 223-233, Apr. 2000.
[17] R. Lyon and S. Shamma, “Auditory representation of timbre and pitch,” Journal of Auditory Computations, pp 221–270.
[18] W. C. Lin, H. T. Fan and J. W. Hung, “DCT-based processing of dynamic features for robust speech recognition,” in Proc. IEEE International Symposium on Chinese Spoken Language Processing, Nov. 2010, pp.12-17.
[19] V. F. S. Alencar and A. Alcaim, “LSF and LPC - derived features for large vocabulary distributed continuous speech recognition in Brazilian Portuguese,” in Proc. Asilomar Conference on Signals, Systems and Computers, Oct. 2008, pp. 1237-1241.
[20] M. A. Ferrer, I. G. Alonso, and C. M. Travieso, “Influence of initialisation and stop criteria on HMM based recognisers,” Electronics Letters, vol.36, no.13, pp.1165-1166, Jun. 2000.
[21] D. Sarkar, “Randomness in generalization ability: a source to improve it,” IEEE Transactions on Neural Networks, vol.7, no.3, pp.676-685, May 1996. |