中大機構典藏-NCU Institutional Repository-提供博碩士論文、考古題、期刊論文、研究計畫等下載:Item 987654321/92798
English  |  正體中文  |  简体中文  |  Items with full text/Total items : 80990/80990 (100%)
Visitors : 41739031      Online Users : 1215
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
Scope Tips:
  • please add "double quotation mark" for query phrases to get precise results
  • please goto advance search for comprehansive author search
  • Adv. Search
    HomeLoginUploadHelpAboutAdminister Goto mobile version


    Please use this identifier to cite or link to this item: http://ir.lib.ncu.edu.tw/handle/987654321/92798


    Title: 利用與摘要相關的文章重點句結合對比學習改進摘要模型的事實一致性;Combining Key Sentences Related to the Abstract with Contrastive Learning to Improve Summarization Factual Inconsistency
    Authors: 張景泰;Chang, Ching-Tai
    Contributors: 軟體工程研究所
    Keywords: 抽象式摘要;預訓練模型;對比學習;事實一致性;Abstractive Summarization;Pre-trained Model;Factual Inconsistency;Hallucination;Contrastive Learning
    Date: 2023-02-02
    Issue Date: 2024-09-19 16:19:24 (UTC+8)
    Publisher: 國立中央大學
    Abstract: 摘要中的事實不一致性代表摘要中的訊息無法從來源文章中獲得驗證,是抽象式摘要中棘手的問題,研究顯示模型產出的摘要有30\%擁有事實不一致的問題,使得抽象式摘要難以應用在生活中,近幾年研究者也開始重視這個問題。

    過去的方法傾向於提供額外的背景知識,將其融入於模型中,或者在模型解碼後對產出的結果進行檢查及更正。

    對比學習是近幾年新的模型訓練方法,它在影像領域取得了卓越的成績,其概念是利用正樣本、負樣本之間的對比性,使得模型學習出來的向量物以類聚,正樣本經過模型得到的向量彼此間的距離會較貼近,負樣本經過模型得到的向量彼此間的距離會較疏遠。如此模型在一定程度上擁有了區分不同事物的能力。

    在我們的研究中,首先對原始文章找出與摘要每一句最相關的句子,接著對編碼器使用了對比學習方法使得編碼過後的向量可以獲得與摘要更加相關的原始文章向量使得解碼器產出的摘要更符合事實一致。;Hallucination, also known as factual inconsistency, is when models generate summaries that contain incorrect information or information not mentioned in source text.

    It is a critical problem in abstractive summarization and makes summaries generated by models hard to use in practice.
    Previous works prefer to add additional information such as background knowledge into the model or use post-correct/rank method after decoding to improve this headache.

    Contrastive learning is a new model-training method and has achieved excellent results in the Image Processing field. The concept is to use the contrast between positive and negative samples to make vectors learned by the model cluster together. Given the anchor point, the distance between the anchor point and the positive samples will be closer, and the distance between the anchor point and the negative samples will be farther. This way, the model has the ability to distinguish positive examples from negative examples to a certain extent.

    We propose a new method to improve factual consistency by separating representation of the most relevant sentences and the least relevant sentences from the source document during the training phase through contrastive learning so that the model can learn how to generate summaries that are more relevant to the main points of the source documents.
    Appears in Collections:[Software Engineer] Electronic Thesis & Dissertation

    Files in This Item:

    File Description SizeFormat
    index.html0KbHTML16View/Open


    All items in NCUIR are protected by copyright, with all rights reserved.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明