Opponent modeling in deep reinforcement learning

H He, J Boyd-Graber, K Kwok… - … conference on machine …, 2016 - proceedings.mlr.press
International conference on machine learning, 2016proceedings.mlr.press
Opponent modeling is necessary in multi-agent settings where secondary agents with
competing goals also adapt their strategies, yet it remains challenging because of strategies'
complex interaction and the non-stationary nature. Most previous work focuses on
developing probabilistic models or parameterized strategies for specific applications.
Inspired by the recent success of deep reinforcement learning, we present neural-based
models that jointly learn a policy and the behavior of opponents. Instead of explicitly …
Abstract
Opponent modeling is necessary in multi-agent settings where secondary agents with competing goals also adapt their strategies, yet it remains challenging because of strategies’ complex interaction and the non-stationary nature. Most previous work focuses on developing probabilistic models or parameterized strategies for specific applications. Inspired by the recent success of deep reinforcement learning, we present neural-based models that jointly learn a policy and the behavior of opponents. Instead of explicitly predicting the opponent’s action, we encode observation of the opponents into a deep Q-Network (DQN), while retaining explicit modeling under multitasking. By using a Mixture-of-Experts architecture, our model automatically discovers different strategy patterns of opponents even without extra supervision. We evaluate our models on a simulated soccer game and a popular trivia game, showing superior performance over DQN and its variants.
proceedings.mlr.press
Showing the best result for this search. See all results