博碩士論文 107521604 完整後設資料紀錄

DC 欄位 語言
DC.contributor電機工程學系zh_TW
DC.creator陳登國zh_TW
DC.creatorTran Dang Khoaen_US
dc.date.accessioned2021-1-27T07:39:07Z
dc.date.available2021-1-27T07:39:07Z
dc.date.issued2021
dc.identifier.urihttp://ir.lib.ncu.edu.tw:88/thesis/view_etd.asp?URN=107521604
dc.contributor.department電機工程學系zh_TW
DC.description國立中央大學zh_TW
DC.descriptionNational Central Universityen_US
dc.description.abstract在本文中,我們提出了一種GRU結構的變體,稱為雙序列門控注意單元(DS-GAU),其中計算了x向量基線的每個TDNN層的統計池,並將其通過DS-GAU層傳遞, 在訓練為幀級時從輸入要素的不同時間上下文中聚合更多信息。 我們提出的架構在VoxCeleb2數據集上進行了訓練,其中特徵向量稱為DSGAU-向量。 我們對VoxCeleb1數據集和“野生演說者”(SITW)數據集進行了評估,並將實驗結果與x矢量基線系統進行了比較。 結果表明,相對於VoxCeleb1數據集的x向量基線,我們提出的方法在EER相對改進方面最多可存檔11.6%,7.9%和7.6%.zh_TW
dc.description.abstractIn this thesis, we present a variant of GRU architecture called Dual-Sequences Gated Attention Unit (DS-GAU), in which the statistics pooling from each TDNN layer of the x-vector baseline are computed and passed through the DS-GAU layer, to aggregate more information from the variant temporal context of input features while training as frame-level. Our proposed architecture was trained on the VoxCeleb2 dataset, where the feature vector is referred to as a DSGAU-vector. We made our evaluation on the VoxCeleb1 dataset and the Speakers in the Wild (SITW) dataset and compared the experimental results with the x-vector baseline system. It showed that our proposed method archived up to 11.6%, 7.9%, and 7.6% in EER relative improvements over the x-vector baseline on the VoxCeleb1 dataset.en_US
DC.subject應用於語者驗證之雙序列門控注意力單元架構zh_TW
DC.title應用於語者驗證之雙序列門控注意力單元架構zh_TW
dc.language.isozh-TWzh-TW
DC.titleDual-Sequences Gated Attention Unit Architecture for Speaker Verificationen_US
DC.type博碩士論文zh_TW
DC.typethesisen_US
DC.publisherNational Central Universityen_US

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明