博碩士論文 109521068 完整後設資料紀錄

DC 欄位 語言
DC.contributor電機工程學系zh_TW
DC.creator洪滿珍zh_TW
DC.creatorMan-Chen Hungen_US
dc.date.accessioned2022-8-25T07:39:07Z
dc.date.available2022-8-25T07:39:07Z
dc.date.issued2022
dc.identifier.urihttp://ir.lib.ncu.edu.tw:88/thesis/view_etd.asp?URN=109521068
dc.contributor.department電機工程學系zh_TW
DC.description國立中央大學zh_TW
DC.descriptionNational Central Universityen_US
dc.description.abstract詞義消歧是自然語言理解的一項重要且艱難的任務,尤其是對於醫療領域經常有多 種語義含意的詞彙。我們提出一個雙重註釋編碼器 (Dual Gloss Encoders, DGE) 模型, 以 BERT 轉譯器為基礎,將中文句子中健康照護領域的命名實體,連結到多國語言詞彙 語義網路 BabelNet,以實現上下文感知語義理解。消歧目標詞的每個註釋都源自 BabelNet,我們將原句嵌入註釋得到語境化的目標詞嵌入向量,目標詞嵌入向量再和每 個註釋嵌入配對,以計算語義消歧的分數,將分數最高者選為語句中消歧目標詞的註釋 選項。由於在健康照護領域缺乏中文實體連結數據,我們收集了適當的領域單詞並在句 子中手動標記它們的註釋。最後,我們總共有 10,218 個句子,包含 40 個不同的消歧目 標詞和 94 個不同的語義註釋。我們將建構的數據劃分為訓練集 7,109 筆、發展集 979 筆與測試集 2,130 筆。實驗結果表明,我們提出的 DGE 模型的性能優於三個實體連結 模型,即 BERTWSD、GlossBERT 與 BEM,獲得了 F1-Score 97.81%。zh_TW
dc.description.abstractWord sense disambiguation is an important and difficult task for natural language understanding, especially for those lexical words with many semantic meanings in the healthcare domain. We propose a BERT transformer based Dual Gloss Encoder (DGE) model to link Chinese healthcare entities to the multi-lingual lexical network BabelNet for contextaware semantic understanding. The target word along with its context in original sentence is encoded to obtain embedding vector. Each gloss of the target word is originated from BabelNet to encode the gloss embedding. Target word embedding and each gloss embedding will be paired to calculate the scores for sense disambiguation. The gloss with the highest score is returned as predicted gloss for the target word in a given sentence. Due to a lack of Chinese entity linking data in the healthcare domain, we collected proper domain-specific words and manually annotated their glosses in the sentence. Finally, we have a total of 10,218 sentences containing 40 distinct target words with 94 various semantic glosses. Our constructed data was divided into three mutually exclusive datasets, including training set (7,109 sentences), development set (979 sentences), and test set (2,130 sentences). Experimental results indicate that our proposed DGE model performs better than three entity linking models, i.e., BERTWSD, GlossBERT and BEM, obtaining the best F1-score of 97.81%.en_US
DC.subject實體連結zh_TW
DC.subject詞義消歧zh_TW
DC.subject語言轉譯器zh_TW
DC.subject自然語言理解zh_TW
DC.subject健康資訊學zh_TW
DC.subjectentity linkingen_US
DC.subjectword sense disambiguationen_US
DC.subjectlanguage transformersen_US
DC.subjectnatural language understandingen_US
DC.subjecthealth informaticsen_US
DC.title利用雙重註釋編碼器於中文健康照護實體連結zh_TW
dc.language.isozh-TWzh-TW
DC.titleLeveraging Dual Gloss Encoders in Chinese Healthcare Entity Linkingen_US
DC.type博碩士論文zh_TW
DC.typethesisen_US
DC.publisherNational Central Universityen_US

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明