Object-Oriented Reinforcement Learning in Cooperative Multiagent Domains

Main Article Content

Felipe Leno da Silva Ruben Glatt Anna Helena Reali Costa

Abstract

Although Reinforcement Learning methods have successfully been applied to increasingly large problems, scalability remains a central issue. While Object-Oriented Markov Decision Processes (OO-MDP) are used to exploit regularities in a domain, Multiagent System (MAS) methods are used to divide workload amongst multiple agents. In this work we propose a novel combination of OO-MDP and MAS, called Multiagent Object-Oriented Markov Decision Process (MOO-MDP), so as to accrue the benefits of both strategies and be able to better address scalability issues. We present an algorithm to solve deterministic cooperative MOO-MDPs, and prove that it learns optimal policies while reducing the learning space by exploiting state abstractions. We experimentally compare our results with earlier approaches and show advantages with regard to discounted cumulative reward, number of steps to fulfill the task, and Q-table size.

Article Details

How to Cite
LENO DA SILVA, Felipe; GLATT, Ruben; HELENA REALI COSTA, Anna. Object-Oriented Reinforcement Learning in Cooperative Multiagent Domains. BRACIS, [S.l.], july 2017. Available at: <http://143.54.25.88/index.php/bracis/article/view/124>. Date accessed: 19 sep. 2024. doi: https://doi.org/10.1235/bracis.vi.124.
Section
Artigos