參考文獻 |
[1] Baker, B., Gupta, O., Naik, N., and Raskar, R., 2016, “Designing neural network architectures using reinforcement learning,” arXiv preprint arXiv:1611.02167
[2] Baker, B., Gupta, O., Raskar, R., and Naik, N., 2017, “Accelerating neural architecture search using performance prediction,” arXiv preprint arXiv:1705.10823
[3] Bergstra, J., and Bengio, Y., 2012, “Random search for hyper-parameter optimization,” Journal of machine learning research, 13(Feb), pp. 281-305
[4] Brock, A., Lim, T., Ritchie, J. M., and Weston, N., 2017, “Smash: one-shot model architecture search through hypernetworks,” arXiv preprint arXiv:1708.05344
[5] Cai, H., Chen, T., Zhang, W., Yu, Y., and Wang, J., 2018, “Efficient architecture search by network transformation,” Thirty-Second AAAI conference on artificial intelligence
[6] Chollet, F., 2017, “Xception: Deep learning with depthwise separable convolutions,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1251-1258
[7] Deng, J., Dong, W., Socher, R., Li, L., Li, K., and L, Fei-Fei., 2009, “Imagenet: A large-scale hierarchical image database,” 2009 IEEE conference on computer vision and pattern recognition, pp. 248-255, IEEE
[8] Domhan, T., Springenberg, J. T., and Hutter, F., 2015, “Speeding up automatic hyperparameter optimization of deep neural networks by extrapolation of learning curves,” Twenty-Fourth International Joint Conference on Artificial Intelligence
[9] Elsken, T., Metzen, J. H., and Hutter, F., 2017, “Simple and efficient architecture search for convolutional neural networks,” arXiv preprint arXiv:1711.04528
[10] Elsken, T., Metzen, J.H., and Hutter, F., 2018, “Neural architecture search: A survey,” arXiv
[11] Floreano, D., Dürr, P., and Mattiussi, C., 2008, “Neuroevolution: from architectures to learning,” Evolutionary Intelligence
[12] Guo, M., Zhong, Z., Wu, W., Lin, D., and Yan, J., 2019, “Irlas: Inverse reinforcement learning for architecture search,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9021-9029
[13] Hu, J., Shen, L., and Sun, G., 2018, “Squeeze-and-excitation networks,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132-7141
[14] Hutter, F., Kotthoff, L., and Vanschoren, J., 2019, “Automatic Machine Learning: Methods, Systems, Challenges,” Springer, available at http://automl.org/book
[15] Krizhevsky, A., Sutskever, I., and Hinton, G. E., 2012, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, pp. 1097-1105
[16] Liu, C., Zoph, B., Neumann, M., Shlens, J., Hua, W., Li, L. J., ... and Murphy, K., 2018, “Progressive neural architecture search,” Proceedings of the European Conference on Computer Vision (ECCV), pp. 19-34
[17] Liu, H., Simonyan, K., and Yang, Y., 2018, “Darts: Differentiable architecture search,” arXiv preprint arXiv:1806.09055
[18] Luo, R., Tian, F., Qin, T., Chen, E., & Liu, T. Y., 2018, “Neural architecture optimization,” Advances in neural information processing systems, pp. 7816-7827
[19] Mendoza, H., Klein, A., Feurer, M., Springenberg, J. T., and Hutter, F., 2016, “Towards automatically-tuned neural networks,” Workshop on Automatic Machine Learning, pp. 58-65
[20] Pham, H., Guan, M. Y., Zoph, B., Le, Q. V., and Dean, J., 2018, “Efficient neural architecture search via parameter sharing,” arXiv preprint arXiv:1802.03268
[21] Real, E., Aggarwal, A., Huang, Y., and Le, Q. V., 2019, “Regularized evolution for image classifier architecture search,” Proceedings of the aaai conference on artificial intelligence, Vol. 33, pp. 4780-4789
[22] Salimans, T., Ho, J., Chen, X., Sidor, S., & Sutskever, I., 2017, “Evolution strategies as a scalable alternative to reinforcement learning,” arXiv preprint arXiv:1703.03864.
[23] Stanley, K. O., and Miikkulainen, R., 2002, “Evolving neural networks through augmenting topologies,” Evolutionary computation, 10(2), pp. 99-127
[24] Stanley, K. O., Bryant, B. D., and Miikkulainen, R., 2005, “Real-time neuroevolution in the NERO video game,” IEEE transactions on evolutionary computation, 9(6), pp. 653-668
[25] Suganuma, M., Shirakawa, S., and Nagao, T., 2017, “A genetic programming approach to designing convolutional neural network architectures,” Proceedings of the Genetic and Evolutionary Computation Conference, pp. 497-504
[26] Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., ... and Rabinovich, A., 2015, “Going deeper with convolutions,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1-9
[27] Vanschoren, J., 2019, “Meta-learning,” In Automated Machine Learning, pp. 35-61, Springer, Cham
[28] Yao, X., 1999, “Evolving artificial neural networks,” IEEE
[29] Zhong, Z., Yan, J., and Liu, C. L., 2017, “Practical network blocks design with q-learning,” arXiv preprint arXiv:1708.05552, 6
[30] Zoph, B., and Le, Q. V., 2016, “Neural architecture search with reinforcement learning,” arXiv preprint arXiv:1611.01578
[31] Zoph, B., Vasudevan, V., Shlens, J., and Le, Q. V., 2018, “Learning transferable architectures for scalable image recognition,” Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8697-8710 |