參考文獻 |
[1] X. Huang, A. Acero and H. W. Hon, Spoken Language Processing, Prentice Hall, 2001.
[2] G. R. Doddington, “Speaker recognition—Identifying people by their voices,” Processing of the IEEE, vol. 73, pp. 1651-1664, 1985.
[3] D. O'Shaughnessy, “Speaker recognition,” ASSP Magazine , IEEE, vol. 3, pp. 4-17, 1986.
[4] J. T. Tou and R. C. Gonzalez, Pattern Recognition Principles, Addison Wesley, 1974.
[5] L. S. Lee and Y. Lee, “Voice Access of Global Information for Broad-Band Wireless: Technologies of Today and Challenges of Tomorrow,” Proceedings of the IEEE, vol. 89, no. 1, pp. 41-57, January 2001.
[6] L. Zao, A. Alcaim and R. Coelho, “Robust Access based on Speaker Identification for Optical Communications Security,” Digital Signal Processing, 2009 16th International Conferenxe on, pp. 1-5, 2009.
[7] Wahyudi, W. Astuti and S. Mohamed, “A Comparison of Gaussian Mixture and Artificaial Neural Network Models for Voiced-based Access Control System of Building Security,” Information Technology, 2008. ITSim 2008. International Symposium on, vol. 3, pp. 1-8, 2008.
[8] 蔡仲齡,“含語者驗證之小型場所人臉辨識門禁系統的研發",國立成功大學碩士論文,中華民國九十七年七月。
[9] X. D. Huang and K. F. Lee, “On Speaker-Independent, Speaker-Dependent, and Speaker-Adaptive Speech Recognition,” Speech and Audio Processing, IEEE Transactions on, vol. 1, pp 150-157, 1993.
[10] H. Sakoe and S. Chiba, “Dynamic Programming Algorithm Optimization for Spoken Word Recognition,” Acoustics, Speech and Signal Processing, IEEE Transactions on, vol. 26, pp. 43-49, 1978.
[11] C. Myers, L. Rabiner and A. Rosenberg, “Performance Tradeoffs in Dynamic Time Warping Algorithms for Isolated Word Recognition,” Acoustics, Speech and Signal Processing, IEEE Transactions on, vol. 28, pp. 623-635, 1980.
[12] D. P. Morgan and C. L. Scofield, Neural Network and Speech Processing, Kluwer Academic, 1991.
[13] P. Pujol, S. Pol, C. Nadeu and A. Hagen, “Comparison and Combination of Features in a Hybrid HMM/MLP and a
HMM/GMM Speech Recognition,” Speech and Audio Processing, IEEE Transactions on, vol. 13, pp. 14-22, 2005.
[14] W. Dong-Liang, W.W.Y. Ng, P.P.K. Chan and D. Hai-Lan, “Access control by RFID and face recognition based on neural network,” ICMLC, 2010 International on, vol. 2, pp. 675-680, 2010.
[15] S. Jieun and K. Howon, “The RFID Middleware System Supporting Context-Aware Access Control Service,” ICACT 2006, vol. 1, pp. 863-866, 2006.
[16] Y. Gizatdinova and V. Surakka, “Feature-Based Detection of Facial Landmarks from Neutral and Expressive Facial Images,” Pattern Analysis and Machine Intelliqence, IEEE Transactions on, vol. 28, pp. 135-139, 2006.
[17] P. N. Belhumeur, J. P. Hespanha and D. J. Kriegman, “Eigenfaces vs. Fisherfaces: Recognition Using Class Specific Linear Projection,” Pattern Analysis and Machine Intelliqence, IEEE Transactions on, vol. 19, pp. 711-720, 1997.
[18] M. J. Er, S. Wu, J. Lu and H. L. Toh, “Face Recognition With Radial Basis Function (RBF) Neural Network,” Neural Networks, IEEE Transactions on, vol. 13, no. 3, pp. 697-710, 2002.
[19] D. A. Reynolds and R. C. Rose, “Robust text-independent speaker identification using Gaussian mixture speaker models,” Speech and Audio Processing, IEEE Transactions on, vol. 3, pp. 72-83, 1995.
[20] L. R. Rabiner and B. H. Juang, Fundamentals of Speech Recognition, Prentice Hall, New Jersey, 1993.
[21] 王小川,“語音訊號處理",全華,民國九十三年三月。
[22] R. Vergin, D. O’Shaughnessy and A. Farhat, “Generalized Mel Frequency Coefficients for Large-Vocabulary Speaker-Independent Continuous-Speech Recognition,” IEEE Trans. Speech and Audio Processing, vol. 7, no. 5, pp. 525-532, September 1999.
[23] John R. Deller, Jr. , John G Proakis and John H. L. Hansen, Discrete-Time Processing of Speech Signals, 1987.
[24] L. R. Rabiner, “A Tutorial on Hidden Markov Models and Selected Application in Speech Recognition,” Proceedings of the IEEE, vol. 77, no. 2, Feb. 1989.
[25] S. E. Levinson, L. R. Rabiner and M. M. Sondhi, “An Introduction to the Application of the Theory of Probabilistic Function of a Markov Process to Automatic Speech Recognition,” The Bell System Technical Journal, vol. 62, no. 4, April 1983.
[26] Changsheng Ai, Xuan Sun, Honghua Zhao and Xueren Dong, “Pipeline damage and leak sound recognition based on HMM,” Proceedings of the 7th World Congress on Intelligent Control and Automation, pp. 1940-1944, June. 2008.
[27] 蔡永琪,“基於次音節單元之關鍵詞辨識",國立中央大學碩士論文,中華民國八十四年六月。
[28] M.-W. koo, C.-H. Lee and B.-H. Juang, “Speech Recognition and Utterance Verification Based on a Generalized Confidence Score,” IEEE Trans. on Speech and Audio Processing, vol. 9, no. 8, pp. 821-832, Nov. 2001.
[29] J. Zhi-Hua and Y. Zhen, “Voice conversion using Viterbi algorithm based on Gaussian mixture model,” ISPACS 2007, pp. 32-35, Nov. 2007.
[30] 黃國彰,“關鍵詞萃取與確認之研究",國立中央大學碩士論文,中華民國八十五年六月。
[31] 王維邦,“連續國語語音關鍵詞萃取系統之研究與發展",國立中央大學碩士論文,中華民國八十六年六月。
[32] H. Bourlard, B. D’hoore and J. M. Boite, “Optimizing recognition and rejection performance in wordspotting systems,” ICASSP-94, vol. 1, pp. I/373-I/376, 1994.
[33] H. Ney, “The use of a one stage dynamic programming algorithm for connected word recognition,” IEEE Trans. on Acoustic, Speech Signal, Processing, vol. 32, no. 2, pp. 263-271, April 1984.
[34] W. Jhing-Fa, W. Chung-Hsien, H. Chaug-Ching and L. Jau-Yien, “Integrating Neural Nets and One-Stage Dynamic Programming for Speaker Independent Continuous Mandarin Digit Recognition,” Acoustics, Speech, and Signal Processing, 1991, vol. 1, pp. 69-72, Apr 1991.
[35] J. Neyman and E. S. Pearson, “On the problem of the most efficient tests of statistical hypotheses,” phil. Trans. R. Soc. Lond. A, vol. 231, pp. 289-337, 1933.
[36] J. Neyman and E. S. Pearson, “On the use and interpretation of certain test criteria for purpose of statistical inference,” Biometrika, pt I, vol. 20A, pp. 175-240, 1928.
[37] T. Kawahara, C.-H. Lee and B.-H. Juang, “Flexible Speech Understanding Based on Combined Key-Phrase Detection and Verification,” IEEE Trans. on Speech and Audio Processing, vol. 6, no. 6, pp. 558-568, Nov. 1998.
[38] Tatsuya Kawahara, C.-H. Lee and B.-H. Juang, “Combining Key-Phrase Detection and Subword-Based Verification For Flexible Speech Understanding,” Proc IEEE Int. Conf. Acoustic, Speech, Signal Processing, vol. 2, pp. 1159-1162, Munich Germany, May. 1997.
[39] T. Chee-Ming, S.-H. Salleh, T. Tian-Swee and A. K. Ariff, “Text Independent Speaker Identification Using Gaussian Mixture Model,” ICIAS 2007, pp. 194-198, Nov. 2007.
[40] 黃夢晨,“最小錯誤鑑別式應用於語者辨識之競爭語者探討",國立中央大學碩士論文,中華民國九十七年六月。
[41] F. Soong, A. Rosenberg, L. Rabiner and B. Juang, “A vector quantization approach to speaker recognition,” Acoustics, Speech, and Signal Processing, IEEE International Conference on ICASSP ‘85, vol. 10, pp. 387-390, 1985.
[42] Y. Linde, A. Buzo and R. Gray, “An Algorithm for Vector Quantizer Design,” Communications, IEEE Transactions on, vol. 28, no. 1, pp. 84-95, 1980.
[43] T. K. Moon, “The Expectation-Maximization Algorithm,” IEEE Signal Processing Magazine, vol. 13, no. 6, pp. 47-60, November 1996.
[44] S. Z. Selim and M. A. Ismail, “K-Means-Type Algorithms: A Generalized Convergence Theorem and Characterization of Local Optimality,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, pp. 81-87, Jan. 1984.
[45] A. P. Dempster, N. M. Laird and D. B. Rubin, “Maximum likelihood from incomplete data via the EM algorithm,” Journal of the Royal Statistical Society, Series B, vol. 39, no. 1, pp. 1-38, 1977.
[46] D. A. Reynolds, “Comparison of background normalization methods for text-independent speaker verification,” EUROSPEECH ‘97, 5th European Conference on Speech Communication and Technology, pp. 963-966, 1997.
[47] J.-L. Gauvain and L. Chin-Hui, “Maximum a posteriori estimation for multivariate Gaussian mixture observations of Markov chains,” Speech and Audio Processing, IEEE Transactions on, vol. 2, pp. 291-298, 1994.
[48] C. J. Leggetter and P. C. Woodland, “Maximum likelihood linear regression for speaker adaptation of continuous density hidden Markov models, ” Computer Speech & Language, vol. 9, pp. 171-185, 1995.
[49] A. Sankar and L. Chin-Hui, “A Maximum-Likelihood Approach to Stochastic Matching for Robust Speech Recognition, ” Speech and Audio Processing, IEEE Transactions on, vol. 4, pp. 190-202, May 1996.
[50] O. Siohan, C. Chesta and Lee Chin-Hui, “Joint maximum a posteriori adaptation of transformation and HMM parameters,” Speech and Audio Processing, IEEE Transactions on, vol. 9, pp. 417-428, 2001.
[51] D. A. Reynolds, T. F. Quatieri and R. B. Dunn, “Speaker verification using Adapted Gaussian mixture models, ” Digital Signal Processing, vol. 10, pp. 19-41, 2000.
[52] 范世明,“高斯混合模型在語者辨識與國語語音辨認之應用",國立交通大學碩士論文,中華民國九十一年。
[53] 位元文化,“精通MFC視窗程式設計-Visual Studio 2008版",文魁資訊,民國九十七年。
[54] R. F. Raposa,“C++與MFC視窗程式設計",陳智湧、歐世亮和林志偉譯,文魁資訊,民國九十七年。
[55] 溫家誠,“多媒體應用之語音辨識系統",國立中央大學碩士
文,中華民國九十七年六月。
|