參考文獻 |
[1]. W. Fan, and A. Bifet, “Mining big data: current status, and forecast to the future,”
SIGKDD Explor. Newsl., vol. 14, no. 2, pp. 1-5, 2013.
[2]. G. Bello-Orgaz, J. J. Jung, and D. Camacho, “Social big data: Recent achievements and
new challenges,” Information Fusion, vol. 28, pp. 45-59, 2016.
[3]. J. Han, J. Pei, and M. Kamber, Data mining: concepts and techniques: Elsevier, 2011.
[4]. U. Fayyad, G. Piatetsky-Shapiro, and P. Smyth, “From data mining to knowledge
discovery in databases,” AI magazine, vol. 17, no. 3, pp. 37, 1996.
[5]. A. Famili, W.-M. Shen, R. Weber, and E. Simoudis, “Data preprocessing and intelligent
data analysis,” Intelligent data analysis, vol. 1, no. 1, pp. 3-23, 1997.
[6]. S. Kotsiantis, D. Kanellopoulos, and P. Pintelas, “Data preprocessing for supervised
leaning,” International Journal of Computer Science, vol. 1, no. 2, pp. 111-117, 2006.
[7]. O. E. de Noord, “The influence of data preprocessing on the robustness and parsimony of
multivariate calibration models,” Chemometrics and intelligent laboratory systems, vol. 23,
no. 1, pp. 65-70, 1994.
[8]. L. Yu, and H. Liu, “Feature selection for high-dimensional data: a fast correlation-based
filter solution,” in Proceedings of the Twentieth International Conference on International
Conference on Machine Learning, Washington, DC, USA, 2003, pp. 856-863.
[9]. A. B. Patel, M. Birla, and U. Nair, “Addressing big data problem using Hadoop and Map
Reduce,” in 2012 Nirma University International Conference on Engineering (NUiCONE),
2012, pp. 1-5.
[10]. Y. Zhai, Y.-S. Ong, and I. W. Tsang, “The emerging" big dimensionality",” 2014.
[11]. V. Bolón-Canedo, N. Sánchez-Maroño, and A. Alonso-Betanzos, “A review of feature
selection methods on synthetic data,” Knowledge and information systems, vol. 34, no. 3, pp.
483-519, 2013.
[12]. V. Bolón-Canedo, N. Sánchez-Marono, A. Alonso-Betanzos, J. M. Benítez, and F.
Herrera, “A review of microarray datasets and applied feature selection methods,”
Information Sciences, vol. 282, pp. 111-135, 2014.
[13]. I. Guyon, and A. Elisseeff, “An introduction to variable and feature selection,” Journal
of machine learning research, vol. 3, no. Mar, pp. 1157-1182, 2003.
[14]. S. Rayana, W. Zhong, and L. Akoglu, “Sequential ensemble learning for outlier
detection: A bias-variance perspective,” in 2016 IEEE 16th International Conference on Data
Mining (ICDM), 2016, pp. 1167-1172.
[15]. A. L. Blum, and P. Langley, “Selection of relevant features and examples in machine
learning,” Artificial intelligence, vol. 97, no. 1-2, pp. 245-271, 1997.
[16]. H. Liu, and L. Yu, “Toward integrating feature selection algorithms for classification and
clustering,” IEEE Transactions on Knowledge & Data Engineering, no. 4, pp. 491-502, 2005. 74
[17]. Z. M. Hira, and D. F. Gillies, “A Review of Feature Selection and Feature Extraction
Methods Applied on Microarray Data,” Adv Bioinformatics, vol. 2015, pp. 198363, 2015.
[18]. A. Jain, and D. Zongker, “Feature selection: Evaluation, application, and small sample
performance,” IEEE transactions on pattern analysis and machine intelligence, vol. 19, no. 2,
pp. 153-158, 1997.
[19]. M. Dash, and H. Liu, “Feature selection for classification,” Intelligent data analysis, vol.
1, no. 1-4, pp. 131-156, 1997.
[20]. R. Kohavi, and G. H. John, “Wrappers for feature subset selection,” Artificial
intelligence, vol. 97, no. 1-2, pp. 273-324, 1997.
[21]. A. G. Karegowda, M. Jayaram, and A. Manjunath, “Feature subset selection problem
using wrapper approach in supervised learning,” International journal of Computer
applications, vol. 1, no. 7, pp. 13-17, 2010.
[22]. Y. Saeys, I. Inza, and P. Larrañaga, “A review of feature selection techniques in
bioinformatics,” bioinformatics, vol. 23, no. 19, pp. 2507-2517, 2007.
[23]. V. Kumar, “Feature Selection: A literature Review,” The Smart Computing Review, vol.
4, no. 3, 2014.
[24]. J. H. Holland, “Genetic algorithms,” Scientific american, vol. 267, no. 1, pp. 66-73,
1992.
[25]. S. Cateni, M. Vannucci, M. Vannocci, and V. Colla, "Variable selection and feature
extraction through artificial intelligence techniques," Multivariate Analysis in Management,
Engineering and the Sciences: IntechOpen, 2013.
[26]. Y. Chtioui, D. Bertrand, and D. Barba, “Feature selection by a genetic algorithm.
Application to seed discrimination by artificial vision,” Journal of the Science of Food and
Agriculture, vol. 76, no. 1, pp. 77-86, 1998.
[27]. Y. Lu, I. Cohen, X. S. Zhou, and Q. Tian, “Feature selection using principal feature
analysis,” in Proceedings of the 15th ACM international conference on Multimedia, 2007, pp.
301-304.
[28]. L. I. Smith, A tutorial on principal components analysis, 2002.
[29]. J. R. Quinlan, “Induction of decision trees,” Machine learning, vol. 1, no. 1, pp. 81-106,
1986.
[30]. J. R. Quinlan, C4. 5: programs for machine learning: Elsevier, 2014.
[31]. T. G. Dietterich, “Ensemble methods in machine learning,” in International workshop on
multiple classifier systems, 2000, pp. 1-15.
[32]. L. Rokach, “Ensemble-based classifiers,” Artificial Intelligence Review, vol. 33, no. 1-2,
pp. 1-39, 2010.
[33]. D. Opitz, and R. Maclin, “Popular ensemble methods: An empirical study,” Journal of
artificial intelligence research, vol. 11, pp. 169-198, 1999.
[34]. G. Martínez-Muñoz, and A. Suárez, “Using boosting to prune bagging ensembles,” 75
Pattern Recognition Letters, vol. 28, no. 1, pp. 156-165, 2007.
[35]. P. Bühlmann, Bagging, boosting and ensemble methods: Springer, 2012.
[36]. Y. Saeys, T. Abeel, and Y. Van de Peer, “Robust feature selection using ensemble feature
selection techniques,” in Joint European Conference on Machine Learning and Knowledge
Discovery in Databases, 2008, pp. 313-325.
[37]. T. Abeel, T. Helleputte, Y. Van de Peer, P. Dupont, and Y. Saeys, “Robust biomarker
identification for cancer diagnosis with ensemble feature selection methods,” Bioinformatics,
vol. 26, no. 3, pp. 392-398, 2009.
[38]. A. Tsymbal, M. Pechenizkiy, and P. Cunningham, “Diversity in search strategies for
ensemble feature selection,” Information fusion, vol. 6, no. 1, pp. 83-98, 2005.
[39]. B. Seijo-Pardo, I. Porto-Díaz, V. Bolón-Canedo, and A. Alonso-Betanzos, “Ensemble
feature selection: Homogeneous and heterogeneous approaches,” Knowledge-Based Systems,
vol. 118, pp. 124-139, 2017.
[40]. M. Termenon, and M. Graña, “A two stage sequential ensemble applied to the
classification of Alzheimer’s disease based on mri features,” Neural Processing Letters, vol.
35, no. 1, pp. 1-12, 2012.
[41]. C.-F. Tsai, and Y.-C. Hsiao, “Combining multiple feature selection methods for stock
prediction: Union, intersection, and multi-intersection approaches,” Decision Support
Systems, vol. 50, no. 1, pp. 258-269, 2010.
[42]. M. I. Jordan, and T. M. Mitchell, “Machine learning: Trends, perspectives, and
prospects,” Science, vol. 349, no. 6245, pp. 255-260, 2015.
[43]. P. Cunningham, M. Cord, and S. J. Delany, "Supervised learning," Machine learning
techniques for multimedia, pp. 21-49: Springer, 2008.
[44]. C. Cortes, and V. Vapnik, “Support-vector networks,” Machine learning, vol. 20, no. 3,
pp. 273-297, 1995.
[45]. J. J. J. I. T. o. s. Grefenstette, man,, and cybernetics, “Optimization of control parameters
for genetic algorithms,” vol. 16, no. 1, pp. 122-128, 1986.
[46]. A. Venkatachalam, “M-InfoSift: A Graph-based Approach for Multiclass Document
Classification,” 2007.
[47]. M. Sokolova, G. J. I. P. Lapalme, and Management, “A systematic analysis of
performance measures for classification tasks,” vol. 45, no. 4, pp. 427-437, 2009.
[48]. M. Al-Rajab, J. Lu, Q. J. C. m. Xu, and p. i. biomedicine, “Examining applying high
performance genetic data feature selection and classification algorithms for colon cancer
diagnosis,” vol. 146, pp. 11-24, 2017.
[49]. T. Elgamal, and M. J. a. p. a. Hefeeda, “Analysis of PCA algorithms in distributed
environments,” 2015.
[50]. J. Su, and H. Zhang, “A fast decision tree learning algorithm,” in Proceedings of the 21st
national conference on Artificial intelligence - Volume 1, Boston, Massachusetts, 2006, pp.
500-505. |