博碩士論文 106522063 詳細資訊




以作者查詢圖書館館藏 以作者查詢臺灣博碩士 以作者查詢全國書目 勘誤回報 、線上人數:13 、訪客IP:3.137.171.121
姓名 李哲豪(Che-Hao Li)  查詢紙本館藏   畢業系所 資訊工程學系
論文名稱 使用句對模型在文章中抓取相關資訊用於問題生成
(Using Sentence Pair Model to Capture Relevant Information from Document for Question Generation)
相關論文
★ A Real-time Embedding Increasing for Session-based Recommendation with Graph Neural Networks★ 基於主診斷的訓練目標修改用於出院病摘之十代國際疾病分類任務
★ 混合式心臟疾病危險因子與其病程辨識 於電子病歷之研究★ 基於 PowerDesigner 規範需求分析產出之快速導入方法
★ 社群論壇之問題檢索★ 非監督式歷史文本事件類型識別──以《明實錄》中之衛所事件為例
★ 應用自然語言處理技術分析文學小說角色 之關係:以互動視覺化呈現★ 基於生醫文本擷取功能性層級之生物學表徵語言敘述:由主成分分析發想之K近鄰算法
★ 基於分類系統建立文章表示向量應用於跨語言線上百科連結★ Code-Mixing Language Model for Sentiment Analysis in Code-Mixing Data
★ 藉由加入多重語音辨識結果來改善對話狀態追蹤★ 對話系統應用於中文線上客服助理:以電信領域為例
★ 應用遞歸神經網路於適當的時機回答問題★ 使用多任務學習改善使用者意圖分類
★ 使用轉移學習來改進針對命名實體音譯的樞軸語言方法★ 基於歷史資訊向量與主題專精程度向量應用於尋找社群問答網站中專家
檔案 [Endnote RIS 格式]    [Bibtex 格式]    [相關文章]   [文章引用]   [完整記錄]   [館藏目錄]   至系統瀏覽論文 (2024-6-30以後開放)
摘要(中) 近幾年問題生成的研究發展迅速,過去以句子的語法結構定義規則生成問題,隨著深度學習成熟的技術,現今機器能理解語意並自動產生適當的問題。

問題生成的目標是給定一段文字訊息與答案,產生相對應的問題,與機器閱讀理解任務類似,因此閱讀理解的資料集常被用在問題生成任務中。以往問題生成模型的輸入並非整篇文章,而是包含答案的句子,但有些問題的內容和答案不在同一個句子,可能是依據其他句子資訊產生該答案的問題,於是本論文提出一個新架構,由句對模型和問題生成模型所組成,利用句對模型處理文章結構,將每一句資訊與包含答案的句子進行匹配,計算各自的相關程度並且重新賦予句子權重,接著傳送到問題生成模型產生最終的問題。句對模型主要目的是從整篇文章中自動找尋和答案有關的內容進而產生適合的問題。

實驗結果表示,我們的系統能有效處理文章結構,相比只有問題生成模型的系統,在中文和英文的資料集都有更好的表現。
摘要(英) In recent years, question generation (QG) has developed rapidly. In the past, using rules that are based on syntactic structure to generate questions. Nowadays, the machine can understand semantic and automatically generate appropriate questions with a proven technique of deep learning.

Question generation aims to generate corresponding questions from a given passage and answer. It is similar to machine reading comprehension (RC) task. Therefore, reading comprehension dataset is often used to question generation task. The input of the previous question generation model is the sentence containing the answer rather than the whole article. However, the content of some questions and its answers are not in the same sentence. The question may be based on other information in sentences. Then, our paper proposed a new framework which consists of sentence pair model and question generation model. Using the sentence pair model to process article structure. Its method is matching each sentence and the sentence containing the answer to compute the respective degree of correlation to reweight sentences and then produce questions by question generation model. The main purpose of sentence pair model is to automatically find the content related to the answer from the article.

Experiment results show that our system can handle article structure. In contrast to a system with only question generation model, our system has better performance in Chinese and English dataset.
關鍵字(中) ★ 問題生成
★ 閱讀理解
★ 序列到序列
★ 注意力機制
★ 複製機制
★ 句對模型
★ 深度學習
關鍵字(英) ★ Question Generation
★ Reading Comprehension
★ Sequence to Sequence
★ Attention Mechanism
★ Copy Mechanism
★ Sentence Pair Model
★ Deep Learning
論文目次 摘要ii
Abstract iii
誌謝v
目錄 vii
圖目錄 ix
表目錄 x
一、緒論1
1.1 研究背景.................................................................. 1
1.2 研究動機.................................................................. 3
1.3 章節概要.................................................................. 4
二、相關研究5
2.1 自然語言處理(Natural Language Processing, NLP)............ 5
2.2 深度學習(Deep Learning) ............................................ 5
2.3 自然語言生成(Natural Language Generation, NLG)........... 6
2.4 注意力機制(Attention Mechanism) ................................ 7
2.5 問題生成(Question Generation, QG) .............................. 7
2.6 自然語言推理(Natural language Inference, NLI) ............... 9
三、系統架構10
3.1 資料前處理............................................................... 11
3.2 詞嵌入..................................................................... 11
3.3 句子編碼器............................................................... 11
3.3.1 Attention-Based Convolutional Neural Network ........ 12
3.3.2 Match-LSTM .................................................... 14
3.4 訊息編碼器............................................................... 15
3.5 問題解碼器............................................................... 17
四、實驗方法與討論20
4.1 資料集..................................................................... 20
4.1.1 台達閱讀理解資料集(Delta Reading Comprehension
Dataset, DRCD)................................................... 20
4.1.2 史丹佛問題答案資料集(Stanford Question Answering
Dataset, SQuAD)................................................... 20
4.2 評估方法.................................................................. 21
4.2.1 BLEU ............................................................. 21
4.2.2 F-score ............................................................ 22
4.3 參數設定.................................................................. 22
4.4 實驗結果.................................................................. 23
4.5 討論........................................................................ 26
五、結論與未來工作28
參考文獻29
參考文獻 [1] P. Rajpurkar, J. Zhang, K. Lopyrev, and P. Liang, “Squad: 100,000+ questions for
machine comprehension of text,” arXiv preprint arXiv:1606.05250, 2016.
[2] X. Du, J. Shao, and C. Cardie, “Learning to ask: Neural question generation for
reading comprehension,” arXiv preprint arXiv:1705.00106, 2017.
[3] C. C. Shao, T. Liu, Y. Lai, Y. Tseng, and S. Tsai, “Drcd: a chinese machine reading
comprehension dataset,” arXiv preprint arXiv:1806.00920, 2018.
[4] I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural
networks,” in Advances in neural information processing systems, pp. 3104–3112,
2014.
[5] D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning
to align and translate,” arXiv preprint arXiv:1409.0473, 2014.
[6] M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural networks,” IEEE
Transactions on Signal Processing, vol. 45, no. 11, pp. 2673–2681, 1997.
[7] M. Heilman and N. A. Smith, “Good question! statistical ranking for question generation,”
in Human Language Technologies: The 2010 Annual Conference of the
North American Chapter of the Association for Computational Linguistics, pp. 609–
617, Association for Computational Linguistics, 2010.
[8] Q. Zhou, N. Yang, F. Wei, C. Tan, H. Bao, and M. Zhou, “Neural question generation
from text: A preliminary study,” in National CCF Conference on Natural Language
Processing and Chinese Computing, pp. 662–671, Springer, 2017.
[9] C. Gulcehre, S. Ahn, R. Nallapati, B. Zhou, and Y. Bengio, “Pointing the unknown
words,” arXiv preprint arXiv:1603.08148, 2016.
[10] J. Gu, Z. Lu, H. Li, and V. O. Li, “Incorporating copying mechanism in sequenceto-
sequence learning,” arXiv preprint arXiv:1603.06393, 2016.
[11] L. Song, Z. Wang, W. Hamza, Y. Zhang, and D. Gildea, “Leveraging context information
for natural question generation,” in Proceedings of the 2018 Conference
of the North American Chapter of the Association for Computational Linguistics:
Human Language Technologies, Volume 2 (Short Papers), pp. 569–574, 2018.
[12] Z. Wang, H. Mi, W. Hamza, and R. Florian, “Multi-perspective context matching
for machine comprehension,” arXiv preprint arXiv:1612.04211, 2016.
[13] V. Kumar, G. Ramakrishnan, and Y.-F. Li, “A framework for automatic question
generation from text using deep reinforcement learning,” arXiv preprint arXiv:
1808.04961, 2018.
[14] Z. Tu, Z. Lu, Y. Liu, X. Liu, and H. Li, “Modeling coverage for neural machine
translation,” arXiv preprint arXiv:1601.04811, 2016.
[15] K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, “Bleu: a method for automatic
evaluation of machine translation,” in Proceedings of the 40th annual meeting on
association for computational linguistics, pp. 311–318, Association for Computational
Linguistics, 2002.
[16] C.-Y. Lin, “Rouge: A package for automatic evaluation of summaries,” Text Summarization
Branches Out, 2004.
[17] Y. Kim, H. Lee, J. Shin, and K. Jung, “Improving neural question generation using
answer separation,” arXiv preprint arXiv:1809.02393, 2018.
[18] S. R. Bowman, G. Angeli, C. Potts, and C. D. Manning, “A large annotated corpus
for learning natural language inference,” arXiv preprint arXiv:1508.05326, 2015.
[19] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation,
vol. 9, no. 8, pp. 1735–1780, 1997.
[20] T. Rocktäschel, E. Grefenstette, K. M. Hermann, T. Kočiskỳ, and P. Blunsom, “Reasoning
about entailment with neural attention,” arXiv preprint arXiv:1509.06664,
2015.
[21] W. Yin, H. Schütze, B. Xiang, and B. Zhou, “Abcnn: Attention-based convolutional
neural network for modeling sentence pairs,” Transactions of the Association for
Computational Linguistics, vol. 4, pp. 259–272, 2016.
[22] S. Wang and J. Jiang, “Learning natural language inference with lstm,” arXiv
preprint arXiv:1512.08849, 2015.
[23] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed representations
of words and phrases and their compositionality,” in Advances in neural
information processing systems, pp. 3111–3119, 2013.
[24] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv
preprint arXiv:1412.6980, 2014.
指導教授 蔡宗翰(Tsung-Han Tsai) 審核日期 2019-7-23
推文 facebook   plurk   twitter   funp   google   live   udn   HD   myshare   reddit   netvibes   friend   youpush   delicious   baidu   
網路書籤 Google bookmarks   del.icio.us   hemidemi   myshare   

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明