參考文獻 |
[1] 吳沛儒,曾明德,吳東凌,林良泰,蘇昭銘,王晉元,吳毅成,周家慶,黃啟倡,黃培書,何國豪,倪文哲,郭松庭,2020,示範型強化學習之人工智慧號誌控制,109年中華民國運輸年會論文集第五冊 222-237頁
[2] 胡大瀛,李卓育,2021,深度強化學習下號誌控制設計獎勵機制之探討,110年中華民國運輸年會論文集第五冊 543-566頁
[3] 陳惠國,2022,強化學習應用於交通號誌控制之展望,中華道路季刊第六十一卷第四期 43-54頁
[4] 許添本,程楷祐,2020,以深度強化學習方式建構混合車流之AI最佳化號誌時制計畫,109年中華民國運輸年會論文集 第五冊 331-352頁
[5] 許添本,黃建皓,2022,深度確定性策略梯度法建構幹道即時號誌控制系統,111年中華民國運輸年會論文集第五冊 1-26頁
[6] 許添本,蔡沐軒,2021,示範式深度強化學習應用於號誌時制最佳化之研究,110年中華民國運輸年會論文集第五冊 468-493頁
[7] 臺北市交通管制工程處,2022,常見問答:為促進交通流暢,交通號誌秒數及連鎖如何設定?臺北市交通管制工程,檢自:https://www.bote.gov.taipei/News_Content.aspx?n=20A2BA930381C524&s=6011548EB76C0FD3 (Oct 15, 2022)
[8] BBC News, 2017. Google AI defeats human Go champion. British Broadcasting Corporation. Retrieved from https://www.bbc.com/news/technology-40042581(Feb 20, 2022)
[9] Liang, X., Du, X., Wang, G, & Han, Z., 2019. A Deep Reinforcement Learning Network for Traffic Light Cycle Control. IEEE Transactions on Vehicular Technology, 68(2), 1243–1253.
[10] Lopez, P. A., Behrisch, M, Bieker-Walz, L., Erdmann, J. E., Flotteröd, Y., Hilbrich, R., Lücken, L., Rummel, J., Wagner, P., & Wießner, E., 2018. Microscopic Traffic Simulation using SUMO. Complex Systems, 16, pp. 29-53.
[11] Google, 2022, Google Earth 6.0-version: Taipei City. Retrieved from https://earth.google.com/web/ (April 5, 2022)
[12] Hessel, M., Modayil, J., Hasselt, H., Schaul, T., Ostrovski, G., Dabney, W., Horgan, D., Piot, B., Azar, M., & Silver, D., 2017. Rainbow: Combining Improvements in Deep Reinforcement Learning. arXiv preprint arXiv: 1710.02298. doi: 10.48550/arXiv.1710.02298.
[13] Hester, T., Vecerik, M., Pietquin, O. et al., 2017. Learning from Demonstrations for Real World Reinforcement Learning. arXiv preprint arXiv: 1704.03732.
[14] Hunt, P. B., Robertson, D. I., Bretherton, R. D., & Winton, R. I., 1981. SCOOT-A Traffic Responsive Method of Coordinating Signals. Transport and Road Research Laboratory Report LR 1041
[15] Lillicrap, T., Hunt, J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., & Wierstra, D.,2015. Continuous control with deep reinforcement learning. In Proceedings of the 4th International Conference on Learning Representations.
[16] Luk, J. Y. K., Sims, A. G., & Lowrie, P. R., 1982. SCATS-Application and Field Comparison with a TRANSYT Optimised Fixed Time System. In Proceedings of International Conference on Road Traffic Signaling, London, UK, pp. 77-74.
[17] Mnih, V., Badia, A., Mirza, M., Graves, A., Lillicrap, T., Harley, T., Silver, D., & Kavukcuoglu, K., 2016. Asynchronous Methods for Deep Reinforcement Learning. In Proceedings of the 33rd International Conference on Machine Learning, 48, pp. 1928-1937.
[18] Mnih, V., Kavukcuoglu, K., Silver, D. et al., 2015. Human-level control through deep reinforcement learning. Nature, 518, pp. 529-533.
[19] Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., & Riedmiller, M., 2013. Playing Atari with Deep Reinforcement Learning. arXiv preprint arXiv:1312.5602. doi: 10.48550/arXiv.1312.5602.
[20] Papers with Code, 2022. Trends: Frameworks. Retrieved from https://paperswithcode.com/trends (January 7, 2022)
[21] PTV Group, 2022. PTV Vissim. PTV Group, Germany
[22] Rummery, G., & Niranjan, M., 1994. On-Line Q-Learning Using Connectionist Systems. Technical Report CUED/F-INFENG/TR 166
[23] Schulman, J., Levine, S., Moritz, P., Jordan, M., & Abbeel, P., 2015. Trust Region Policy Optimization. arXiv preprint arXiv: 1502.05477. doi: 10.48550/arXiv.1502.05477.
[24] Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O., 2017. Proximal Policy Optimization Algorithms. arXiv preprint arXiv: 1707.06347. doi: 10.48550/arXiv.1707.06347.
[25] Silver, D., Huang, A., Maddison, C. et al., 2016. Mastering the game of Go with deep neural networks and tree search. Nature, 529, pp. 484-489. doi: 10.1038/nature16961.
[26] Wang, S., Xie, X., Huang, K., Zeng, J., & Cai, Z., 2019. Deep Reinforcement Learning-Based Traffic Signal Control Using High-Resolution Event-Based Data. Entropy 21. doi:10.3390/e21080744
[27] Wei, H., Zheng, G., Yao, H., & Li, Z., 2018. IntelliLight: A Reinforcement Learning Approach for Intelligent Traffic Light Control. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, pp. 2496-2505. doi: 10.1145/3219819.3220096
[28] Wei, H., Xu, N., Zhang, H., Zheng, G., Zang, X., Chen, C., Zhang, W., Zhu, Y., Xu, K., Li, Z., 2019a. CoLight: Learning Network-level Cooperation for Traffic Signal Control. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, New York, USA, pp. 1913–1922. doi: 10.1145/3357384.3357902
[29] Wei, H., Chen, C., Zheng, Z., Wu, K., Gayah V., Xu, K., Li, Z., 2019b. PressLight: Learning Max Pressure Control to Coordinate Traffic Signals in Arterial Network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, USA, pp. 1290-1298. doi: 10.1145/3292500.3330949
[30] Xu, B., Wang, Y., Wang, Z., Jia, H., & Lu, Z., 2021. Hierarchically and Cooperatively Learning Traffic Signal Control. In Proceedings of the AAAI Conference on Artificial Intelligence, 35(1), pp. 669-677. doi: 10.1609/aaai.v35i1.16147
[31] Zang, X., Yao, H., Zheng, G., Xu, N., Xu, K., & Li, Z., 2020. MetaLight: Value-Based Meta-Reinforcement Learning for Traffic Signal Control. In Proceedings of the AAAI Conference on Artificial Intelligence, 34(1), pp. 1153-1160. doi: 10.1609/aaai.v34i01.5467.
[32] Zhang, H., Feng, S., Liu, C., Ding, Y., Zhu, Y., Zhou, Z., Zhang, W., Yu, Y., Jin, H., & Li, Z., 2019. CityFlow: A Multi-Agent Reinforcement Learning Environment for Large Scale City Traffic Scenario. In Proceedings of the World Wide Web Conference, pp.3620-3624. doi:10.1145/3308558.3314139.
[33] Zhao, W., Ye, Y., Ding, J., Wang, T., Wei, T., & Chen, M., 2022. IPDALight: Intensity- and Phase Duration-Aware Traffic Signal Control Based on Reinforcement Learning. Journal of Systems Architecture, 123, pp.102374-102385. doi: 10.1016/j.sysarc.2021.102374.
[34] Zheng, G., Xiong, Y., Zang, X., Feng, J., Wei, H., Zhang, H., Li, Y., Xu, K., & Li, Z., 2019. Learning Phase Competition for Traffic Signal Control. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 1963-1972. doi: 10.1145/3357384.3357900.
[35] Zheng, G., Zang, X., Xu, N. et al., 2019. Diagnosing Reinforcement Learning for Traffic Signal Control. arXiv preprint arXiv: 1905.04716.
|