博碩士論文 109521127 詳細資訊




以作者查詢圖書館館藏 以作者查詢臺灣博碩士 以作者查詢全國書目 勘誤回報 、線上人數:33 、訪客IP:3.145.153.135
姓名 曾昱翔(Yu-Xiang Zeng)  查詢紙本館藏   畢業系所 電機工程學系
論文名稱 運用句嵌入向量重排序器 增進中文醫療問答系統效能
(Enhancing Chinese Medical Question-Answering Performance with Sentence Embedding Reranker)
相關論文
★ 多重嵌入增強式門控圖序列神經網路之中文健康照護命名實體辨識★ 基於腦電圖小波分析之中風病人癲癇偵測研究
★ 基於條件式生成對抗網路之資料擴增於思覺失調症自動判別★ 標籤圖卷積增強式超圖注意力網路之中文健康照護文本多重分類
★ 運用合成器混合注意力改善BERT模型於科學語言編輯★ 強化領域知識語言模型於中文醫療問題意圖分類
★ 管道式語言轉譯器 之中文健康照護開放資訊擷取★ 利用雙重註釋編碼器於中文健康照護實體連結
★ 聯合詞性與局部語境於中文健康照護實體關係擷取★ 運用異質圖注意力網路於中文醫療答案擷取式摘要
★ 學習使用者意圖於中文醫療問題生成式摘要★ 標籤強化超圖注意力網路模型於精神疾病文本多標籤分類
檔案 [Endnote RIS 格式]    [Bibtex 格式]    [相關文章]   [文章引用]   [完整記錄]   [館藏目錄]   至系統瀏覽論文 (2027-8-24以後開放)
摘要(中) 真實生活中人們遇到醫療問題,經常藉由不同的管道,尋求醫生的建議與解答,而自動問答系統提供一個即時回覆答案的解決方案。本研究的主要目標為建立中文醫療問答系統,將問題輸入問答系統,從醫療問答資料集中,匹配找出最佳的答案返回給使用者。近年來,不同於傳統的詞彙匹配,深度學習的興起帶動了語義匹配的方式,深度語言模型能有效學習文本的語義訊息,並藉此找出相近的文本。許多研究均顯示出語義匹配的方法較傳統的方法得到更好的效果,因此,我們提出句嵌入向量重排序器 (Sentence Embedding Reranker, SER) 模型。
中文問答資料來自於醫聯網 (https://med-net.com/),資料集共有 26,816 筆醫療問答,我們使用 Pooling method 建立系統測試集,從 26,816 筆問題中取 120 筆問題作為測試問題,每個問題分別經過兩個不同的檢索系統 (BM25 以及 Sentence-BERT),返回100 筆答案,並人工標註其答案的正確性,最後取兩系統的聯集作為系統測試集。藉由實驗結果得知,我們提出的 SER 重排序器模型,在 MAP、NDCG 效能指標達到最好的分數,有效增進中文問答系統的檢索效能。
摘要(英) In the digital era, users usually search and browse web content to obtain healthcare related information before making a doctor’s appointment for diagnosis and treatment. The automatic question-answering system can provide a solution to address this need in real-time. Our main
research objective is to design and implement a Chinese medical question answering system.
In such a medical QA system, users issue a question as a query and then obtain relevant doctors’ answers in the ranked list. Different from traditional lexical matching methods, the deep learning-based semantic matching model can effectively learn the semantic features to retrieve
similar texts. Therefore, we propose a Sentence Embedding Reranker (SER) model to enhance the question-answering performance.
The Pooling method was used to combine the top 100 results returned by BM25 and Sentence-BERT retrieve systems for answer relevance annotation. Based on experimental results from these manual-annotated question-answer pairs, our proposed SER re-ranking model achieved the best results in MAP and NDCG, which can enhance the performance of the Chinese medical question-answering system.
關鍵字(中) ★ 醫療問答系統
★ 資訊檢索
★ 預訓練語言模型
★ 語義匹配
關鍵字(英) ★ Medical question-answering
★ information retrieval
★ pre-trained language models
★ semantic search
論文目次 目錄
摘要 i
Abstract ii
致謝 iii
目錄 iv
圖目錄 vi
表目錄 vii
第一章 緒論 1
1-1 研究背景 1
1-2 研究動機與目的 3
1-3 章節概要 4
第二章 相關研究 5
2-1 自動問答系統 5
2-1-1 資料集 5
2-1-2 評估指標 6
2-1-3 相關模型 10
2-2 詞彙匹配演算法BM25 18
2-3 預訓練語言模型 20
2-3-1 BERT 20
2-3-2 Sentence-BERT 21
2-3-3 ALBERT 23
2-3-4 RoBERTa 23
2-3-5 MacBERT 24
2-4 微調語言模型 25
第三章 研究方法 28
3-1 系統架構 28
3-2 檢索器 (Retriever) 30
3-3 重排序器 (Reranker) 32
第四章 實驗結果 34
4-1 資料集與建置 34
4-2 實驗設定 37
4-3 檢索系統性能比較 40
4-4 SER重排序模型比較 41
4-5 SER重排序數量實驗 42
4-6 SER資料擴增微調訓練實驗 44
4-7 案例分析 45
第五章 結論與未來工作 47
參考文獻 48
參考文獻 參考文獻
[1] Mikolov, T., Sutskever, I., Chen, K., Corrado, G. s., and Dean, J. 2013. Distributed
representations of words and phrases and their compositionality. In proceedings of the
26th International Conference on Neural Information Processing Systems (Volume 2).
pages 3111–3119. https://dl.acm.org/doi/10.5555/2999792.2999959.
[2] Pennington, J., Socher, R., and Manning, C. 2014. GloVe: Global vectors for word
representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural
Language Processing (EMNLP), pages 1532–1543, Doha, Qatar. Association for
Computational Linguistics. https://aclanthology.org/D14-1162.
[3] Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. 2019. BERT: Pre-training of deep
bidirectional transformers for language understanding. In Proceedings of the 2019
Conference of the North American Chapter of the Association for Computational
Linguistics: Human Language Technologies (Volume 1:Long and Short Papers).
Minneapolis, Minnesota. Association for Computational Linguistics.
https://aclanthology.org/N19-1423.
[4] Reimers, N., and Gurevych, I. 2019. Sentence-BERT: Sentence embeddings using
siamese BERT-networks. In Proceedings of the 2019 Conference on Empirical Methods
in Natural Language Processing and the 9th International Joint Conference on Natural
Language Processing (EMNLP-IJCNLP). pages 3982–3992, Hong Kong, China.
Association for Computational Linguistics. https://aclanthology.org/D19-1410.
[5] Robertson, S.E., Walker, S., Jones S., Hancock-Beaulieu, M., and Gatford, M. 1995.
Okapi at TREC-3. In: Harman DK (ed) Proceedings of the third Text REtrieval
Conference (TREC-3), pages 109–126.
[6] Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer,
L., and Stoyanov, V. 2019. Roberta: A robustly optimized bert pretraining approach.
arXiv:1907.11692. https://doi.org/10.48550/arXiv.1907.11692.
[7] Wang, M., Smith, N., and Mitamura, T. 2007. What is the jeopardy model? A quasisynchronous grammar for qa. In Proceedings of the 2007 Joint Conference on Empirical
Methods in Natural Language Processing and Computational Natural Language
Learning (EMNLP-CoNLL), pages 22–32, Prague, Czech Republic. Association for
Computational Linguistics. https://aclanthology.org/D07-1003.
[8] Yang, Y., Yih, W., and Meek, C. 2015. WikiQA: A Challenge Dataset for Open-Domain
Question Answering. In Proceedings of the 2015 Conference on Empirical Methods in
Natural Language Processing, pages 2013–2018, Lisbon, Portugal. Association for
Computational Linguistics. https://aclanthology.org/D15-1237.

[9] Bajaj, P., Campos, D., Craswell, N., Deng, L., Gao, J., Liu, X., Majumder, R.,
McNamara, A., Mitra, B., Nguyen, T., Rosenberg, M., Song, X., Stoica, A., Tiwary, S.,
and Wang, T. 2016. Ms marco: A human generated machine reading comprehension
dataset. arXiv:1611.09268. Version 3.
[10] Abacha, A. B., Shivade, C., and Demner-Fushman, D. 2019. Overview of the MEDIQA
2019 shared task on textual inference, question entailment and question answering.
In Proceedings of the 18th BioNLP Workshop and Shared Task, pages 370–379,
Florence, Italy. Association for Computational Linguistics.
https://aclanthology.org/W19-5039.
[11] Demner-Fushman, D., Mrabet, Y., and Abacha, A. B. 2020. Consumer health
information and question answering: helping consumers find answers to their healthrelated information needs. Journal of the American Medical Informatics Association,
27(2):194-201. https://doi.org/10.1093/jamia/ocz152.
[12] Abacha, A. B., Agichtein, E., Pinter, Y., and Demner-Fushman, D. 2017. Overview of
the medical question answering task at TREC 2017 LiveQA. In Proceedings of the Text
Retrieval Conference (TREC).
[13] Garg, S., Vu, T., and Moschitti, A. 2020. Tanda: Transfer and adapt pre-trained
transformer models for answer sentence selection. In Proceedings of the AAAI
Conference on ArtificialIntelligence, 34(5):7780-7788.
https://doi.org/10.1609/aaai.v34i05.6282.
[14] Laskar, M. T. R., Huang, X., and Hoque, E. 2020. Contextualized embeddings based
transformer encoder for sentence similarity modeling in answer selection task.
In Proceedings of the 12th Language Resources and Evaluation Conference, pages
5505–5514, Marseille, France. European Language Resources Association.
https://aclanthology.org/2020.lrec-1.676.
[15] Ren, R., Qu, Y., Liu, J., Zhao, W. X., She, Q., Wu, H., Wang, H., and Wen, J. R. 2021.
RocketQAv2: A joint training method for dense passage retrieval and passage re-ranking.
In Proceedings of the 2021 Conference on Empirical Methods in Natural Language
Processing, pages 2825–2835, Online and Punta Cana, Dominican Republic.
Association for Computational Linguistics. https://aclanthology.org/2021.emnlpmain.224.
[16] Gao, L., Dai, Z., and Callan, J. 2021. COIL: Revisit exact lexical match in information
retrieval with contextualized inverted list. In Proceedings of the 2021 Conference of the
North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, pages 3030–3042, Online. Association for Computational
Linguistics. https://aclanthology.org/2021.naacl-main.241.

[17] Gao, L., and Callan, J. 2021. Unsupervised corpus aware language model pre-training
for dense passage retrieval. In Proceedings of the 60th Annual Meeting of the
Association for Computational Linguistics (Volume 1: Long Papers), pages 2843–2853,
Dublin, Ireland. Association for Computational Linguistics.
https://aclanthology.org/2022.acl-long.203.
[18] Gao, T., Yao, X., and Chen, D. 2021. SimCSE: Simple contrastive learning of sentence
embeddings. In Proceedings of the 2021 Conference on Empirical Methods in Natural
Language Processing, pages 6894–6910, Online and Punta Cana, Dominican Republic.
Association for Computational Linguistics. https://aclanthology.org/2021.emnlpmain.552.
[19] Zhu, W., Zhou, X., Wang, K., Luo, X., Li, X., Ni, Y., and Xie, G. 2019. PANLP at
MEDIQA 2019: Pre-trained language models, transfer learning and knowledge
distillation. In Proceedings of the 18th BioNLP Workshop and Shared Task, pages 380–
388, Florence, Italy. Association for Computational Linguistics.
https://aclanthology.org/W19-5040.
[20] Cheng, Y., Fu, S., Tang, M., and Liu, D. 2019. Multi-task deep neural network (MTDNN) enabled optical performance monitoring from directly detected PDM-QAM
signals. Optics express, 27(13):19062-19074. https://doi.org/10.1364/OE.27.019062.
[21] Hearst, M. A., Dumais, S. T., Osuna, E., Platt, J., and Scholkopf, B. 1998. Support vector
machines. IEEE Intelligent Systems and their Applications, 13(4):18-28.
https://doi.org/10.1109/5254.708428.
[22] Demner-Fushman, D., Rogers, W., and Aronson, A. 2017. MetaMap Lite: an evaluation
of a new Java implementation of MetaMap, Journal of the American Medical
Informatics Association, 24(4):841-844. https://doi.org/10.1093/jamia/ocw177.
[23] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł.,
and Polosukhin, I. 2017. Attention is all you need. In Proceedings of the 31st
International Conference on Neural Information Processing Systems, Long Beach,
California, USA. https://dl.acm.org/doi/10.5555/3295222.3295349.
[24] Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P. and Soricut, R. 2020. ALBERT:
A lite BERT for self-supervised learning of language representations. arXiv:1909.11942.
Version 6.
[25] Cui, Y., Che, W., Liu, T., Qin, B., Wang, S., and Hu, G. 2020. Revisiting pre-trained
models for Chinese natural language processing. In Findings of the Association for
Computational Linguistics: EMNLP 2020, pages 657–668, Online. Association for
Computational Linguistics. https://doi.org/10.18653/v1/2020.findings-emnlp.58.

[26] Kwiatkowski, T., Palomaki, J., Redfield, O., Collins, M., Parikh, A., Alberti, C., Epstein,
D., Polosukhin, I., Devlin, J., Lee, K., Toutanova, K., Jones, L., Kelcey, M., Chang, M.,
M. Dai, A., Uszkoreit, J., Le, Q., and Petrov, S. 2019. Natural questions: a benchmark
for question answering research. Transactions of the Association for Computational
Linguistics, 7:452–466. https://aclanthology.org/Q19-1026.
[27] Wang, A., Singh, A., Michael, J., Hill, F., Levy, O., and Bowman, S. R. 2018. GLUE: A
multi-task benchmark and analysis platform for natural language understanding.
In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and
Interpreting Neural Networks for NLP, pages 353–355, Brussels, Belgium. Association
for Computational Linguistics. https://aclanthology.org/W18-5446.
指導教授 李龍豪(Lung-Hao Lee) 審核日期 2022-8-25
推文 facebook   plurk   twitter   funp   google   live   udn   HD   myshare   reddit   netvibes   friend   youpush   delicious   baidu   
網路書籤 Google bookmarks   del.icio.us   hemidemi   myshare   

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明