Transformer Based Multi-Agent Framework
- Publisher:
- IEEE
- Publication Type:
- Conference Proceeding
- Citation:
- 2021 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2021, 2021, 00, pp. 1-2
- Issue Date:
- 2021-01-01
Closed Access
Filename | Description | Size | |||
---|---|---|---|---|---|
Transformer_Based_Multi-Agent_Framework.pdf | Published version | 103.75 kB |
Copyright Clearance Process
- Recently Added
- In Progress
- Closed Access
This item is closed access and not available.
We present a Transformer-like agent to learn the policy of multi-agent cooperation tasks, which is a breakthrough for traditional RNN-based multi-agent models that need to be retrained for each task. Our model can handle various input and output with strong transferability and can parallel tackle different tasks. Besides, We are the first to successfully utilize transformer into a recurrent architecture, providing insight on stabilizing transformers in recurrent RL tasks.
Please use this identifier to cite or link to this item: