博碩士論文 108522601 完整後設資料紀錄

DC 欄位 語言
DC.contributor資訊工程學系zh_TW
DC.creator陶玫婉zh_TW
DC.creatorThamolwan Poopradubsilen_US
dc.date.accessioned2021-7-26T07:39:07Z
dc.date.available2021-7-26T07:39:07Z
dc.date.issued2021
dc.identifier.urihttp://ir.lib.ncu.edu.tw:444/thesis/view_etd.asp?URN=108522601
dc.contributor.department資訊工程學系zh_TW
DC.description國立中央大學zh_TW
DC.descriptionNational Central Universityen_US
dc.description.abstract在本論文中,我們研究了數據準備過程的兩個不同任務:問答配對準備 (Question-Answer Pair Preparation) 和 對話行為標註 (Dialogue Act Tagging)。與其他作品不同,我們的數據來自即時通訊(Instant Messaging: IM)平台,參與者更常將長句拆分成短句,分散成多條消息中發送。因此,在準備問答對時,我們還考慮了一個稱為消息合併任務的任務,用以確定這些消息是否需要進行合併以進行回复預測任務。我們提出了一個 CONTEXT-AOA 模型,將上下文(先前的對話)作為除成對消息之外的附加輸入。 其次,在對話行為標註任務,當我們無法獲得更多標註數據時,我們探索了使用域外數據集來處理該任務的可能性。我們對這個任務進行了兩個實驗。第一個實驗是零樣本學習實驗,我們只使用域外數據集訓練模型並在我們的數據集上測試它們,另一個實驗是我們將一些數據集與外部數據一起包含在模型中域數據集並在剩餘數據上測試它們。我們還提出了一個 CONTEXT-BERT-CRF 模型,它利用了 BERT 微調的能力,同時仍然能夠保留對話中的所有話語並將它們全部提供給模型。 我們在問答對準備任務和對話行為標記任務上的實驗顯示,我們提出的模型在大多數實驗中都能夠勝過所有現有模型。為了演示這兩個任務的使用,我們也構建了基於檢索的聊天機器人。 此聊天機器人不僅根據用戶的輸入從前述準備的問答對中選擇回應,同時也應用對話行為標註資訊來幫助選擇答案。zh_TW
dc.description.abstractIn this thesis, we study two different tasks for data preparation process: Question-Answer Pair Preparation and Dialogue Act Tagging. Unlike other works, our data comes from instant messaging (IM) platform which has different characteristic as participants could split long sentences into short utterances and send them in multiple messages. Therefore, in the preparation for question-answer pairs, we also consider a task called message merging task which aims to determine whether those messages need to be merged or not before generating message pairs for reply-to prediction task. We propose a CONTEXT-AOA model to include the context (previous dialogue) as additional input apart from pairwise messages. For dialogue act tagging task, we explore the possibility of using out-of-domain dataset to deal with this task when we are unable to obtain more annotated data. We conduct two experiments on this task. The first experiment is a zero-shot learning experiment where we train the models using only out-of-domain datasets and test them on our dataset, and another experiment is where we include some of our dataset to the the models along with the out-of-domain datasets and test them on the remaining data. We also propose a CONTEXT-BERT-CRF model which utilizes the ability of BERT and still be able to include all of the utterances from the conversation to the model. Our experiments on both question-answer pair preparation task and dialogue act tagging task show that our proposed models are able to outperform all of the existing models in most of the experiments. To demonstrate the use of these two tasks, the retrieval-based IR-based chatbot has been built. The chatbot will select the response from Q&A pairs prepared in question-answer pair preparation task based on the input from user and return the it back to user. We also apply dialogue act tagging task to help with the answer selection.en_US
DC.subject對話系統zh_TW
DC.subject即時通訊zh_TW
DC.subject對話解開zh_TW
DC.subject響應選擇zh_TW
DC.subject文本分類zh_TW
DC.subject對話行為標記zh_TW
DC.subject信息檢索zh_TW
DC.subjectdialog systemen_US
DC.subjectinstant messagingen_US
DC.subjectconversation disentanglementen_US
DC.subjectresponse selectionen_US
DC.subjecttext classificationen_US
DC.subjectdialogue act taggingen_US
DC.subjectinformation retrievalen_US
DC.titleContext-Aware Question-Answer Pairing and Dialogue Act Tagging from Instant Message Chatlogen_US
dc.language.isoen_USen_US
DC.type博碩士論文zh_TW
DC.typethesisen_US
DC.publisherNational Central Universityen_US

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明