The Center for Education and Research in Information Assurance and Security (CERIAS)

The Center for Education and Research in
Information Assurance and Security (CERIAS)

Temporal Abstractions in Multi-Agent Reinforcement Learning

Principal Investigator: Vaneet Aggarwal

Covering option discovery has been developed to improve the exploration of reinforcement learning in single-agent scenarios with sparse reward signals, through connecting the most distant states in the embedding space provided by the Fiedler vector of the state transition graph. However, these option discovery methods cannot be directly extended to multi-agent scenarios, since the joint state space grows exponentially with the number of agents in the system. In order to alleviate this problem, we design efficient approaches to make multi-agent deep covering options scalable. 

The proposed multi-agent exploration approaches can be used for learning how multiple robots can pick up the object together, coordinate to move across doors, without explosion in complexity. Scalable algorithms are provided. 

Representative Publications