Talk Sessions:

Poster Sessions:

June 22, Booth 42

June 24, Booth 29

Multi-Agent Tree Search with Dynamic Reward Shaping

Alvaro Velasquez, Brett Bissey, Lior Barak, Daniel Melcer, Andre Beckus, Ismail Alkhouri and George Atia

Abstract: Sparse rewards and their representation in multi-agent domains remains a challenge for the development of multi-agent planning systems. While techniques from formal methods can be adopted to represent the underlying planning objectives, their use in facilitating and accelerating learning has witnessed limited attention in multi-agent settings. Reward shaping methods that leverage such formal representations in single-agent settings are typically static in the sense that the artificial rewards remain the same throughout the entire learning process. In contrast, we investigate the use of such formal objective representations to define novel reward shaping functions that capture the learned experience of the agents. More specifically, we leverage the automaton representation of the underlying team objectives in mixed cooperative-competitive domains such that each automaton transition is assigned an expected value proportional to the frequency with which it was observed in successful trajectories of past behavior. This form of dynamic reward shaping is proposed within a multi-agent tree search architecture wherein agents can simultaneously reason about the future behavior of other agents as well as their own future behavior.

*This password protected talk video will only be available after it was presented at the conference.