中大機構典藏-NCU Institutional Repository-提供博碩士論文、考古題、期刊論文、研究計畫等下載:Item 987654321/92798
English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 80990/80990 (100%)
造訪人次 : 41639346      線上人數 : 1238
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋


    請使用永久網址來引用或連結此文件: http://ir.lib.ncu.edu.tw/handle/987654321/92798


    題名: 利用與摘要相關的文章重點句結合對比學習改進摘要模型的事實一致性;Combining Key Sentences Related to the Abstract with Contrastive Learning to Improve Summarization Factual Inconsistency
    作者: 張景泰;Chang, Ching-Tai
    貢獻者: 軟體工程研究所
    關鍵詞: 抽象式摘要;預訓練模型;對比學習;事實一致性;Abstractive Summarization;Pre-trained Model;Factual Inconsistency;Hallucination;Contrastive Learning
    日期: 2023-02-02
    上傳時間: 2024-09-19 16:19:24 (UTC+8)
    出版者: 國立中央大學
    摘要: 摘要中的事實不一致性代表摘要中的訊息無法從來源文章中獲得驗證,是抽象式摘要中棘手的問題,研究顯示模型產出的摘要有30\%擁有事實不一致的問題,使得抽象式摘要難以應用在生活中,近幾年研究者也開始重視這個問題。

    過去的方法傾向於提供額外的背景知識,將其融入於模型中,或者在模型解碼後對產出的結果進行檢查及更正。

    對比學習是近幾年新的模型訓練方法,它在影像領域取得了卓越的成績,其概念是利用正樣本、負樣本之間的對比性,使得模型學習出來的向量物以類聚,正樣本經過模型得到的向量彼此間的距離會較貼近,負樣本經過模型得到的向量彼此間的距離會較疏遠。如此模型在一定程度上擁有了區分不同事物的能力。

    在我們的研究中,首先對原始文章找出與摘要每一句最相關的句子,接著對編碼器使用了對比學習方法使得編碼過後的向量可以獲得與摘要更加相關的原始文章向量使得解碼器產出的摘要更符合事實一致。;Hallucination, also known as factual inconsistency, is when models generate summaries that contain incorrect information or information not mentioned in source text.

    It is a critical problem in abstractive summarization and makes summaries generated by models hard to use in practice.
    Previous works prefer to add additional information such as background knowledge into the model or use post-correct/rank method after decoding to improve this headache.

    Contrastive learning is a new model-training method and has achieved excellent results in the Image Processing field. The concept is to use the contrast between positive and negative samples to make vectors learned by the model cluster together. Given the anchor point, the distance between the anchor point and the positive samples will be closer, and the distance between the anchor point and the negative samples will be farther. This way, the model has the ability to distinguish positive examples from negative examples to a certain extent.

    We propose a new method to improve factual consistency by separating representation of the most relevant sentences and the least relevant sentences from the source document during the training phase through contrastive learning so that the model can learn how to generate summaries that are more relevant to the main points of the source documents.
    顯示於類別:[軟體工程研究所 ] 博碩士論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML16檢視/開啟


    在NCUIR中所有的資料項目都受到原著作權保護.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明