Alert button
Picture for Deheng Ye

Deheng Ye

Alert button

Which Heroes to Pick? Learning to Draft in MOBA Games with Neural Networks and Tree Search

Dec 18, 2020
Sheng Chen, Menghui Zhu, Deheng Ye, Weinan Zhang, Qiang Fu, Wei Yang

Figure 1 for Which Heroes to Pick? Learning to Draft in MOBA Games with Neural Networks and Tree Search
Figure 2 for Which Heroes to Pick? Learning to Draft in MOBA Games with Neural Networks and Tree Search
Figure 3 for Which Heroes to Pick? Learning to Draft in MOBA Games with Neural Networks and Tree Search
Figure 4 for Which Heroes to Pick? Learning to Draft in MOBA Games with Neural Networks and Tree Search

Hero drafting is essential in MOBA game playing as it builds the team of each side and directly affects the match outcome. State-of-the-art drafting methods fail to consider: 1) drafting efficiency when the hero pool is expanded; 2) the multi-round nature of a MOBA 5v5 match series, i.e., two teams play best-of-N and the same hero is only allowed to be drafted once throughout the series. In this paper, we formulate the drafting process as a multi-round combinatorial game and propose a novel drafting algorithm based on neural networks and Monte-Carlo tree search, named JueWuDraft. Specifically, we design a long-term value estimation mechanism to handle the best-of-N drafting case. Taking Honor of Kings, one of the most popular MOBA games at present, as a running case, we demonstrate the practicality and effectiveness of JueWuDraft when compared to state-of-the-art drafting methods.

Viaarxiv icon

Towards Playing Full MOBA Games with Deep Reinforcement Learning

Nov 26, 2020
Deheng Ye, Guibin Chen, Wen Zhang, Sheng Chen, Bo Yuan, Bo Liu, Jia Chen, Zhao Liu, Fuhao Qiu, Hongsheng Yu, Yinyuting Yin, Bei Shi, Liang Wang, Tengfei Shi, Qiang Fu, Wei Yang, Lanxiao Huang, Wei Liu

Figure 1 for Towards Playing Full MOBA Games with Deep Reinforcement Learning
Figure 2 for Towards Playing Full MOBA Games with Deep Reinforcement Learning
Figure 3 for Towards Playing Full MOBA Games with Deep Reinforcement Learning
Figure 4 for Towards Playing Full MOBA Games with Deep Reinforcement Learning

MOBA games, e.g., Honor of Kings, League of Legends, and Dota 2, pose grand challenges to AI systems such as multi-agent, enormous state-action space, complex action control, etc. Developing AI for playing MOBA games has raised much attention accordingly. However, existing work falls short in handling the raw game complexity caused by the explosion of agent combinations, i.e., lineups, when expanding the hero pool in case that OpenAI's Dota AI limits the play to a pool of only 17 heroes. As a result, full MOBA games without restrictions are far from being mastered by any existing AI system. In this paper, we propose a MOBA AI learning paradigm that methodologically enables playing full MOBA games with deep reinforcement learning. Specifically, we develop a combination of novel and existing learning techniques, including curriculum self-play learning, policy distillation, off-policy adaption, multi-head value estimation, and Monte-Carlo tree-search, in training and playing a large pool of heroes, meanwhile addressing the scalability issue skillfully. Tested on Honor of Kings, a popular MOBA game, we show how to build superhuman AI agents that can defeat top esports players. The superiority of our AI is demonstrated by the first large-scale performance test of MOBA AI agent in the literature.

* NeurIPS 2020 
Viaarxiv icon

Supervised Learning Achieves Human-Level Performance in MOBA Games: A Case Study of Honor of Kings

Nov 25, 2020
Deheng Ye, Guibin Chen, Peilin Zhao, Fuhao Qiu, Bo Yuan, Wen Zhang, Sheng Chen, Mingfei Sun, Xiaoqian Li, Siqin Li, Jing Liang, Zhenjie Lian, Bei Shi, Liang Wang, Tengfei Shi, Qiang Fu, Wei Yang, Lanxiao Huang

Figure 1 for Supervised Learning Achieves Human-Level Performance in MOBA Games: A Case Study of Honor of Kings
Figure 2 for Supervised Learning Achieves Human-Level Performance in MOBA Games: A Case Study of Honor of Kings
Figure 3 for Supervised Learning Achieves Human-Level Performance in MOBA Games: A Case Study of Honor of Kings
Figure 4 for Supervised Learning Achieves Human-Level Performance in MOBA Games: A Case Study of Honor of Kings

We present JueWu-SL, the first supervised-learning-based artificial intelligence (AI) program that achieves human-level performance in playing multiplayer online battle arena (MOBA) games. Unlike prior attempts, we integrate the macro-strategy and the micromanagement of MOBA-game-playing into neural networks in a supervised and end-to-end manner. Tested on Honor of Kings, the most popular MOBA at present, our AI performs competitively at the level of High King players in standard 5v5 games.

* IEEE Transactions on Neural Networks and Learning Systems (TNNLS) 
Viaarxiv icon

Mastering Complex Control in MOBA Games with Deep Reinforcement Learning

Jan 03, 2020
Deheng Ye, Zhao Liu, Mingfei Sun, Bei Shi, Peilin Zhao, Hao Wu, Hongsheng Yu, Shaojie Yang, Xipeng Wu, Qingwei Guo, Qiaobo Chen, Yinyuting Yin, Hao Zhang, Tengfei Shi, Liang Wang, Qiang Fu, Wei Yang, Lanxiao Huang

Figure 1 for Mastering Complex Control in MOBA Games with Deep Reinforcement Learning
Figure 2 for Mastering Complex Control in MOBA Games with Deep Reinforcement Learning
Figure 3 for Mastering Complex Control in MOBA Games with Deep Reinforcement Learning
Figure 4 for Mastering Complex Control in MOBA Games with Deep Reinforcement Learning

We study the reinforcement learning problem of complex action control in the Multi-player Online Battle Arena (MOBA) 1v1 games. This problem involves far more complicated state and action spaces than those of traditional 1v1 games, such as Go and Atari series, which makes it very difficult to search any policies with human-level performance. In this paper, we present a deep reinforcement learning framework to tackle this problem from the perspectives of both system and algorithm. Our system is of low coupling and high scalability, which enables efficient explorations at large scale. Our algorithm includes several novel strategies, including control dependency decoupling, action mask, target attention, and dual-clip PPO, with which our proposed actor-critic network can be effectively trained in our system. Tested on the MOBA game Honor of Kings, the trained AI agents can defeat top professional human players in full 1v1 games.

* accepted as a conference paper by AAAI-2020 
Viaarxiv icon