摘要(英) |
A telemarketing company relies heavily on its telemarketers to make numerous
calls to customers in order to promote the company products. To prioritize the
potential customers and evaluate the performance of telemarketers, a objectively
mechanism to identify which stage of promotion a call belongs to is crucial to a
telemarketing company. In this thesis, we design an audio call classification system
based on fine-tuned BERT [1] to automatically classify each telemarketer’s call to an
appropriate stage. The five components of the proposed system are data collection,
data pre-processing, pre-trained model fine-tuning, call-level classification, and the
web service. In data collection, the audio calls are converted into the corresponding transcripts via Kaldi speech recognition. In data pre-processing, transcripts
are processed to remove stopwords, split into segments, and assign labels manually. In pre-trained model fine-tuning, four BERT-based models are retrained to
obtain segment-level classification models. In call-level classification, a rule-based
method is performed to obtain the call-level classification (i.e., stage) of a call from
the classification results of the corresponding segments of the call. Finally, a web
service is provided to allow the company access the system easily. The extensive
experiments show that the proposed system reaches 97% Macro-F1 Score for the
call-level classification. |
參考文獻 |
[1] Bert. https://github.com/google-research/bert.
[2] Cocolong. https://www.cocolong.com.tw/zh-tw.
[3] Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Ziqing Yang, Shijin Wang, and
Guoping Hu. Pre-training with whole word masking for chinese bert, 2019.
[4] Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, and Guoping Hu.
Revisiting pre-trained models for chinese natural language processing. Findings of
the Association for Computational Linguistics: EMNLP 2020, 2020.
[5] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer
Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly
optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
[6] Y. Zhao, Y. Qian, and C. Li. Improved knn text classification algorithm with mapreduce implementation. In 2017 4th International Conference on Systems and Informatics (ICSAI), pages 1417–1422, 2017.
[7] S. Wei, J. Guo, Z. Yu, P. Chen, and Y. Xian. The instructional design of chinese text
classification based on svm. In 2013 25th Chinese Control and Decision Conference
(CCDC), pages 5114–5117, 2013.
[8] Q. Jiang, W. Wang, X. Han, S. Zhang, X. Wang, and C. Wang. Deep feature
weighting in naive bayes for chinese text classification. In 2016 4th International
Conference on Cloud Computing and Intelligence Systems (CCIS), pages 160–164,
2016.
27[9] Yoon Kim. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing
(EMNLP), pages 1746–1751, Doha, Qatar, October 2014. Association for Computational Linguistics.
[10] Junmei Zhong and William Li. Predicting customer call intent by analyzing phone
call transcripts based on CNN for multi-class classification. CoRR, abs/1907.03715,
2019.
[11] Changshun Du and Lei Huang. Text classification research with attention-based recurrent neural networks. International Journal of Computers Communications Control, 13:50, 02 2018.
[12] Xuewei Li and Hongyun Ning. Chinese text classification based on hybrid model of
cnn and lstm. In Proceedings of the 3rd International Conference on Data Science
and Information Technology, DSIT 2020, page 129–134, New York, NY, USA, 2020.
Association for Computing Machinery.
[13] Boulianne Gilles Burget Lukas Glembek Ondrej Goel Nagendra Hannemann Mirko
Motlicek Petr Qian Yanmin Schwarz Petr Silovsky Jan Stemmer Georg Vesely Karel
Povey Daniel, Ghoshal Arnab. The kaldi speech recognition toolkit. 2011.
[14] Apache licence v2.0. https://www.apache.org/licenses/LICENSE-2.0.
[15] Openfstlibrary. http://www.openfst.org/twiki/bin/view/FST/WebHome.
[16] Basic linear algebra subprograms. http://www.netlib.org/blas/.
[17] Linear algebra package. http://www.netlib.org/lapack//.
28[18] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N.
Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. CoRR,
abs/1706.03762, 2017.
[19] Guillaume Lample and Alexis Conneau. Cross-lingual language model pretraining,
2019.
[20] Flask. https://github.com/pallets/flask.
[21] Werkzeug. https://github.com/pallets/werkzeug.
[22] Jinja. https://github.com/pallets/jinja.
[23] Stopwords. https://www.overcoded.net/stop-words-lists-removal-195521/.
[24] Pytorch. https://pytorch.org/.
[25] Chi Sun, Xipeng Qiu, Yige Xu, and Xuanjing Huang. How to fine-tune BERT for
text classification? CoRR, abs/1905.05583, 2019. |