QTypeMix: Enhancing multi-agent cooperative strategies through heterogeneous and homogeneous value decomposition.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

In multi-agent cooperative tasks, the presence of heterogeneous agents is familiar. Compared to cooperation among homogeneous agents, collaboration requires considering the best-suited sub-tasks for each agent. However, the operation of multi-agent systems often involves a large amount of complex interaction information, making it more challenging to learn heterogeneous strategies. Related multi-agent reinforcement learning methods sometimes use grouping mechanisms to form smaller cooperative groups or leverage prior domain knowledge to learn strategies for different roles. In contrast, agents should learn deeper role features without relying on additional information. Therefore, we propose QTypeMix, which divides the value decomposition process into homogeneous and heterogeneous stages. QTypeMix learns to extract type features from local historical observations through the TE loss. In addition, we introduce advanced network structures containing attention mechanisms and hypernets to enhance the representation capability and achieve the value decomposition process. The results of testing the proposed method on 14 maps from SMAC and SMACv2 show that QTypeMix achieves state-of-the-art performance in tasks of varying difficulty.

Authors

  • Songchen Fu
    Laboratory of Speech and Intelligent Information Processing, Institute of Acoustics, CAS, Beijing, China; University of Chinese Academy of Sciences, Beijing, China. Electronic address: fusongchen@hccl.ioa.ac.cn.
  • Shaojing Zhao
    Laboratory of Speech and Intelligent Information Processing, Institute of Acoustics, CAS, Beijing, China; University of Chinese Academy of Sciences, Beijing, China. Electronic address: zhaoshaojing@hccl.ioa.ac.cn.
  • Ta Li
    Laboratory of Speech and Intelligent Information Processing, Institute of Acoustics, CAS, Beijing, China; University of Chinese Academy of Sciences, Beijing, China. Electronic address: lita@hccl.ioa.ac.cn.
  • Yonghong Yan
    Key Laboratory of Speech Acoustics and Content Understanding, Institute of Acoustics, Chinese Academy of Sciences, Beijing 100190, China.