1 Repositories
Python starcraftii Libraries
This is the official implementation of Multi-Agent PPO.
MAPPO Chao Yu*, Akash Velu*, Eugene Vinitsky, Yu Wang, Alexandre Bayen, and Yi Wu. Website: https://sites.google.com/view/mappo This repository implem
653 Jan 6, 2023