DC 欄位 |
值 |
語言 |
DC.contributor | 資訊工程學系 | zh_TW |
DC.creator | 陳律宇 | zh_TW |
DC.creator | Lu-Yu Chen | en_US |
dc.date.accessioned | 2006-7-21T07:39:07Z | |
dc.date.available | 2006-7-21T07:39:07Z | |
dc.date.issued | 2006 | |
dc.identifier.uri | http://ir.lib.ncu.edu.tw:88/thesis/view_etd.asp?URN=93522066 | |
dc.contributor.department | 資訊工程學系 | zh_TW |
DC.description | 國立中央大學 | zh_TW |
DC.description | National Central University | en_US |
dc.description.abstract | 所謂的增強式學習法(Reinforcement Learning),就是訓練對象與環境互動的過程中,不藉助監督者提供完整的指令下,可以自行發掘在各種狀態下該採取什麼行動才能獲得最大報酬。而Q-learning 是一種常見的增強式學習法,藉由建立每一個狀態對應每一個動作之Q值的查詢表(look-up table),Q-learning 可以順利的處理存在少量離散狀態與動作空間的問題上。但當處理的問題擁有大量的狀態與動作時,所要建立的查詢表便會十分的巨大,所以此種對於每一個狀態-動作建立查詢表的方法便顯得不可行。本論文提出一個以自我組織特
徵映射網路(Self-Organization Feature Map network, SOM network)為基礎的模糊系統來實作Q-learning,並以此方法來設計控制系統。為了加速訓練的過程,本論文結合任務分解(task decomposition)與自動任務分解的機制來處理複雜的任務。藉由機器人的模擬實驗,可以看出此方法的有效性。 | zh_TW |
dc.description.abstract | In reinforcement learning, there is no supervisor to critically judge the chosen action at each step. The learning is through a trial-and-error procedure interacting with a dynamic environment. Q-learning is one popular approach to reinforcement learning. It is widely applied to problems with discrete states and actions and usually implemented by a look-up table where each item corresponds to a combination of a state
and an action. However, the look-up table plementation of Q-learning fails in problems with continuous state and action space because an
exhaustive enumeration of all state-action pairs is impossible. In this thesis, an implementation of Q-learning for solving problems with continuous state and action space using SOM-based fuzzy systems is proposed. Simulations of training a robot to complete two different tasks
are used to demonstrate the effectiveness of the proposed approach. Reinforcement learning usually is a slow process. In order to accelerate
the learning procedure, a hybrid approach which integrates the advantages of the ideas of hierarchical learning and the progressive learning to decompose a complex task into simple elementary tasks is proposed. | en_US |
DC.subject | 任務分解 | zh_TW |
DC.subject | 連續性Q-learning | zh_TW |
DC.subject | 增強式學習 | zh_TW |
DC.subject | 自我組織特徵映射圖 | zh_TW |
DC.subject | continuous Q-learning | en_US |
DC.subject | task decomposition | en_US |
DC.subject | self-organizing feature map | en_US |
DC.subject | reinforcement learning | en_US |
DC.title | 以自我組織特徵映射圖為基礎之
模糊系統實作連續性Q-learning | zh_TW |
dc.language.iso | zh-TW | zh-TW |
DC.title | A SOM-based Fuzzy Systems Q-learning in Continuous State and Action Space
| en_US |
DC.type | 博碩士論文 | zh_TW |
DC.type | thesis | en_US |
DC.publisher | National Central University | en_US |