博碩士論文 106423038 完整後設資料紀錄

DC 欄位 語言
DC.contributor資訊管理學系zh_TW
DC.creator宋亞庭zh_TW
DC.creatorYa-Ting Sungen_US
dc.date.accessioned2019-7-1T07:39:07Z
dc.date.available2019-7-1T07:39:07Z
dc.date.issued2019
dc.identifier.urihttp://ir.lib.ncu.edu.tw:88/thesis/view_etd.asp?URN=106423038
dc.contributor.department資訊管理學系zh_TW
DC.description國立中央大學zh_TW
DC.descriptionNational Central Universityen_US
dc.description.abstract真實世界中的資料時常存有品質不佳的問題,像是含有雜訊、不相關的資料、資 料量過大等。若是直接將這些資料建模,恐怕會導致模型的效果和效益不佳,因此必 須先將這些資料進行前處理,其中特徵選取為常見的資料前處理方法,透過特徵選 取,可以將冗餘、不相關的特徵去除,僅留下具代表性的特徵,集成特徵選取是指使 用多種不同的特徵選取演算法,將他們所選取的特徵子集透過不同的方式聚合,透過 集成能夠提升特徵選取的穩健性甚至是提升分類正確率。目前特徵選取的相關研究多 是採用單一特徵選取,較少有研究涉及集成特徵選取,因此本研究欲比較單一特徵選 取和集成特徵選取在高維度資料的表現,找出較佳的特徵選取方法組合。 本研究使用了三種分屬不同類型的特徵選取演算法,分別為基因演算法(Genetic Algorithm, GA)、決策樹 C4.5(Decision Tree C4.5, DT)、主成分分析(Principal Components Analysis, PCA),引用集成學習的中序列式集成和並列式集成的概念形成 序列式集成特徵選取和並列式集成特徵選取,最後利用分類正確率、F1-Score 以及執 行時間來衡量特徵選取方法的優劣。本研究使用 20 個公開資料集,資料集的維度介於 44 到 19993。 根據本研究實驗結果,使用序列式集成特徵方法與並列式集成特徵選取的表現會 優於單一特徵選取,多數資料集的最佳特徵選取方法都是序列式集成特徵方法與並列 式集成特徵選取,序列式集成特徵選取方法中表現最好的方法是 GA+PCA,並列式集成 特徵選取方法中表現最好的方法是 C4.5∪GA。zh_TW
dc.description.abstractData in the real world often have the problem of bad quality, such as noise, irrelevant data and extreme volume. Without considering data pre-processing, the models that are trained by this kind of data are unlikely to be effective. In particular, feature selection is a common data pre-processing method. Through feature selection, redundant and irrelevant features can be removed, leaving only representative features. In ensemble feature selection, it refers to using multiple different feature selection algorithms and combines their selected feature subsets through different aggregation methods. Ensemble feature selection can improve the robustness of single feature selection and even improve the classification accuracy. Currently, the related research on feature selection mostly adopts single feature selection. There are few researches discussing ensemble feature selection. Thus, the aim of this thesis is to compare the performance of single feature selection and ensemble feature selection in high-dimensional data to find a better combination of feature selection methods. In the experiment, three different types of feature selection algorithms are used, which are GA (Genetic Algorithm), DT (Decision Tree Algorithm), and PCA (Principal Components Analysis). For ensemble feature selection, the concept of sequential ensemble and parallel ensemble in ensemble learning are applied to form sequential ensemble feature selection and parallel ensemble feature selection, respectively. Finally, the classification accuracy, f1-score and execution time are examined to evaluate feature selection methods. Based on 20 public datasets with dimensions ranging from 44 to 19993, the experimental results show that sequential ensemble feature selection and parallel ensemble feature selection perform better than single feature selection. The best feature selection methods for most datasets are sequential ensemble feature selection and parallel ensemble feature selection. The best combination in sequential ensemble feature selection is GA+PCA, and the best combination in parallel ensemble feature selection is C4.5∪GA.en_US
DC.subject資料探勘zh_TW
DC.subject特徵選取zh_TW
DC.subject分類zh_TW
DC.subject集成學習zh_TW
DC.subject支援向量機zh_TW
DC.subjectData Miningen_US
DC.subjectFeature Selectionen_US
DC.subjectEnsemble learningen_US
DC.subjectClassificationen_US
DC.subjectSupport Vector Machinesen_US
DC.title單一與集成特徵選取方法於高維度資料之比較zh_TW
dc.language.isozh-TWzh-TW
DC.titleComparison of Single Feature Selection and Ensemble Feature Selection for High-Dimensional Dataen_US
DC.type博碩士論文zh_TW
DC.typethesisen_US
DC.publisherNational Central Universityen_US

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明