
Research Article
Attention-Aware Actor for Cooperative Multi-agent Reinforcement Learning
@INPROCEEDINGS{10.1007/978-3-030-92638-0_22, author={Chenran Zhao and Dianxi Shi and Yaowen Zhang and Yaqianwen Su and Yongjun Zhang and Shaowu Yang}, title={Attention-Aware Actor for Cooperative Multi-agent Reinforcement Learning}, proceedings={Collaborative Computing: Networking, Applications and Worksharing. 17th EAI International Conference, CollaborateCom 2021, Virtual Event, October 16-18, 2021, Proceedings, Part II}, proceedings_a={COLLABORATECOM PART 2}, year={2022}, month={1}, keywords={Multi-agent cooperation Graph Attention mechanism Multi-agent mutual interplay}, doi={10.1007/978-3-030-92638-0_22} }
- Chenran Zhao
Dianxi Shi
Yaowen Zhang
Yaqianwen Su
Yongjun Zhang
Shaowu Yang
Year: 2022
Attention-Aware Actor for Cooperative Multi-agent Reinforcement Learning
COLLABORATECOM PART 2
Springer
DOI: 10.1007/978-3-030-92638-0_22
Abstract
In multi-agent environments, cooperation is crucially important, and the key is to understand the mutual interplay between agents. However, multi-agent environments are highly dynamic, where the complex relationships between agents cause great difficulty for policy learning, and it’s costly to take all coagents into consideration. Besides, agents may not be allowed to share their information with other agents due to communication restrictions or privacy issues, making it more difficult to understand each other. To tackle these difficulties, we propose Attention-Aware Actor (Tri-A), where the graph-based attention mechanism adapts to the dynamics of the mutual interplay of the multi-agent environment. The graph kernels capture the relations between agents, including cooperation and confrontation, within local observation without information exchange between agents or centralized processing, promoting better decision-making of each coagent in a decentralized way. The refined observations produced by attention-aware actors are exploited to learn to focus more on surrounding agents, which makes Tri-A act as a plug for existing multi-agent reinforcement learning (MARL) methods to improve the learning performance. Empirically, we show that our method substantially achieves significant improvement in a variety of algorithms.