English  |  正體中文  |  简体中文  |  全文筆數/總筆數 : 78818/78818 (100%)
造訪人次 : 34716431      線上人數 : 778
RC Version 7.0 © Powered By DSPACE, MIT. Enhanced by NTU Library IR team.
搜尋範圍 查詢小技巧:
  • 您可在西文檢索詞彙前後加上"雙引號",以獲取較精準的檢索結果
  • 若欲以作者姓名搜尋,建議至進階搜尋限定作者欄位,可獲得較完整資料
  • 進階搜尋


    請使用永久網址來引用或連結此文件: http://ir.lib.ncu.edu.tw/handle/987654321/86623


    題名: 應用深度學習於藥品後市場監督:Twitter文本分類任務
    作者: ?堅洽;Huang, Jian-Cia
    貢獻者: 資訊管理學系
    關鍵詞: 深度學習;自然語言處理;藥物警戒;資訊分類;藥物不良反應;Deep Learning;Natural Language Processing;Pharmacovigilance;Information Classification;Adverse Drug Reaction
    日期: 2021-07-27
    上傳時間: 2021-12-07 13:02:07 (UTC+8)
    出版者: 國立中央大學
    摘要: 處方用藥是醫生每天面對每位患者需要處理的任務,當醫生開立藥品給患者服用時需了解此類藥品潛在的副作用,而一項藥物以正常劑量用於人類疾病的預防、診斷治療或是改變其生理功能,但卻出現有害且非預期的結果即為藥物不良反應(Adverse Drug Reaction, ADR)。根據美國衛生公共福利部表示「藥物不良反應」佔每年住院總人數1/3。現今社群媒體使用者越來越多,Twitter每天發文數高達6,500萬則推文及Facebook用戶現今也已超過5億人在使用。
    本研究針對44個治療過動症患者的品牌仿製藥及其他81種藥物做推文收集,共收錄5,729筆用戶推文,進行前處理後做特徵擷取,並以各單詞於詞嵌入模型中產生的文字向量作為自變數,由專家們對推文內容進行判斷是否含有藥物不良反應之訊號,以其結果作為依變數,搭配深度學習之架構與預訓練詞嵌入模型──BERT (Bidirectional Encoder Representations from Transformers)及其分支模型BioBERT (Biomedical Bidirectional Encoder Representations from Transformers)、Bio + Clinical BERT及RoBERTa進行BiLSTM (Bidirectional Long Short-Term Memory)模型的分類任務訓練。
    進行資料平衡處理後,搭配不同的詞向量合併方式Average及Concat進行BiLSTM模型的訓練,結合Earlystop避免過擬合,找出應用於有關ADR推文最適合之預訓練詞嵌入模型。
    本研究發現BERT、BioBERT、Bio + Clinical BERT及RoBERTa等預訓練詞嵌入模型即使於資料不平衡之資料集中建立ADR預測模型,其模型準確率均可接近55%甚至更高,且BERT預訓練詞嵌入模型以Concat方式合併詞向量,於Random Undersampling或是Random Oversampling進行模型訓練,均獲得更好的ADR預測能力。
    ;Prescription medication is a task that doctors face each patient every day. When doctors prescribe drugs to patients, they need to understand the potential side effects of such drugs. A drug is used in normal doses for the prevention, diagnosis and treatment of human diseases. It is to change its physiological function, but the harmful and unexpected result is named Adverse Drug Reaction (ADR). According to the US Department of Health and Public Welfare, "adverse drug reactions" account for 1/3 of the total number of hospitalizations each year. Nowadays, social media users are becoming more and more developed. Twitter posts up to 65 million tweets every day and more than 500 million Facebook users are using it now.
    In this study, 44 brand-name generic drugs and 81 other drugs used for the treatment of patients with ADHD were collected for tweet dataset. A total of 5,729 tweets were collected for pre-processing and feature extraction to obtain the text generated by each word in the word embedding model was used as an independent variable. Experts judged whether the tweet content contained adverse drug reactions combined with deep learning architecture and pre-trained word embedding model-BERT (Bidirectional Encoder Representations from Transformers) and its branch models BioBERT (Biomedical Bidirectional Encoder Representations from Transformers), Bio + Clinical BERT and RoBERTa for BiLSTM (Bi-directional Long Short -Term Memory) model classification task training.
    After the data is balanced, the BiLSTM model is trained with different word vector merging methods "Average" and "Concat", combined with Earlystop to avoid over-fitting mechanism to find out the most suitable pre-trained word embedding model for ADR tweets.
    This study found that even if pre-trained word embedding models such as BERT, BioBERT, Bio + Clinical BERT, and RoBERTa build an ADR prediction model in a imbalanced dataset ,the precision of the model can be close to 55% or even higher, and BERT pre-trained words The embedding model merges word vectors in ′Concat′ method, and performs model training on Random Undersampling or Random Oversampling, both of which obtain better ADR prediction capabilities.
    顯示於類別:[資訊管理研究所] 博碩士論文

    文件中的檔案:

    檔案 描述 大小格式瀏覽次數
    index.html0KbHTML134檢視/開啟


    在NCUIR中所有的資料項目都受到原著作權保護.

    社群 sharing

    ::: Copyright National Central University. | 國立中央大學圖書館版權所有 | 收藏本站 | 設為首頁 | 最佳瀏覽畫面: 1024*768 | 建站日期:8-24-2009 :::
    DSpace Software Copyright © 2002-2004  MIT &  Hewlett-Packard  /   Enhanced by   NTU Library IR team Copyright ©   - 隱私權政策聲明