於硬體大幅的進步,使得強化學習(Reinforcement learning)可以實作並在AI領域帶來新的革新技術,也成為受人歡迎的領域。他可以應用在難以預測的環境上有良好的效果。先前關於強化學習的研究大部分專注在一個玩家在一維度且小範圍動作空間環境,但是多玩家可以處理空間更大互動更多的環境。強化學習其中一項有挑戰性的困難點在多任務的互動行為。本論文中我們提出了一種新型的model,模型透過多個玩家的合作在多任務和多維度動作空間來得取高分。此外我們提出可行的方式來分解巨大的動作空間及減少硬體內存需求和減少計算時間來提高效果。在StartCraft2平台上進行了性能評估,以證明迷你遊戲的有效性。實驗結果表明,所提出的方法在所有指標上均明顯優於比較模型。;Owing to the great advance of hardware techniques, RL (Reinforcement learning) can implement and become popular technique to interact with unpredictable environment. Most prior studies of successive RL model only focus on interaction between single agent and environment with single task and small action space. However, the multi-agent can solve more problem. One of RL challenge task is interaction in multi task environment. In this paper, we propose a novel RL model to interact with the multi task and multi dimension action space environment well through cooperation of agents. In addition, we propose a feasible way to decompose action space to reduce memory size and calculation, and improve the effect. The performance evaluations are conducted on the StartCraft2 platform to demonstrate the effectiveness of mini-games. The experimental results show that the proposed methods significantly outperform the state-of-the-art models in all metrics.