DM2: Decentralized Multi-Agent Reinforcement Learning via Distribution Matching
Abstract
Current approaches to multi-agent cooperation rely heavily on centralized mechanisms or explicit communication protocols to ensure convergence. This paper studies the problem of distributed multi-agent learning without resorting to centralized components or explicit communication. It examines the use of distribution matching to facilitate the coordination of independent agents. In the proposed scheme, each agent independently minimizes the distribution mismatch to the corresponding component of a target visitation distribution. The theoretical analysis shows that under certain conditions, each agent minimizing its individual distribution mismatch allows the convergence to the joint policy that generated the target distribution. Further, if the target distribution is from a joint policy that optimizes a cooperative task, the optimal policy for a combination of this task reward and the distribution matching reward is the same joint policy. This insight is used to formulate a practical algorithm (dm2), in which each individual agent matches a target distribution derived from concurrently sampled trajectories from a joint expert policy. Experimental validation on the StarCraft domain shows that combining (1) a task reward, and (2) a distribution matching reward for expert demonstrations for the same task, allows agents to outperform a naive distributed baseline. Additional experiments probe the conditions under which expert demonstrations need to be sampled to obtain the learning benefits.
1 Introduction
Multi-agent reinforcement learning (MARL) (Littman 1994) is a paradigm for learning agent policies that may interact with each other in cooperative or competitive settings (Silver et al. 2017, 2018; Barrett and Stone 2012; Leibo et al. 2017). Training multiple agents at once is challenging, since an agent updating its own strategy induces a nonstationary environment for other agents, potentially leading to training instabilities, and offsetting any theoretical guarantees single agent RL algorithms confer. To overcome these issues, agent policies can be set up as a single, centralized joint policy, be trained together but then deployed individually (Rashid et al. 2018; Foerster et al. 2018), or be coordinated through some form of communication (Lowe et al. 2017; Jaques et al. 2019; Liu et al. 2021).
Fully distributed training of agent policies remains an open problem in MARL. Distributed, or decentralized, training is desirable particularly in situations where parallelism, robustness, flexibility, or scalability is needed. Such settings include where there are a large number of agents, where agents are faced with changing environments (Marinescu, Dusparic, and Clarke 2017), where agents must perform tasks in varying team configurations over their lifetime (Thrun 1998), or where ensuring privacy is a concern (Léauté and Faltings 2013).
This paper considers the setting of cooperative tasks involving agents, where the goal is to learn a high-performing joint policy in a fully distributed fashion. To mitigate the limitations imposed by this setting, we propose distribution matching to a target state-action distribution, as a strategy to induce coordination. We assume that this distribution is associated with the execution of some joint policy, such as demonstrations of expert teams (Song et al. 2018), or from high-performing trajectories from the agents’ past interactions (Hao et al. 2019). One tempting way to utilize this distribution is to assign each agent the distribution associated with its corresponding expert, and have the agent minimize the distribution mismatch to this target distribution over states and actions.
At first glance, this approach is fraught with complications. Since the target distribution over states and actions is based on the execution of some joint policy, a single agent trying to adjust its policy might not make meaningful progress on its own, given that other agents could change their behaviors at the same time. Second, this distributed approach to distribution matching could suffer from the same destabilization that causes distributed MARL to diverge (Hernandez-Leal, Kartal, and Taylor 2019; Yang and Wang 2020).
This paper shows that despite the above complications, individual distribution matching can be combined with maximization of shared task rewards to learn effectively. In particular, the contributions of the paper are:
-
•
Theoretical analysis showing that distributed distribution matching to the target distributions converges to the joint expert policy that generated the demonstrations.
-
•
The dm2 algorithm, a practical method combining the distribution matching reward with the task reward. Experimental validation shows that if demonstrations are aligned with the shared objective, dm2 accelerates learning compared to a decentralized baseline learning with the task reward only.
-
•
Ablations that empirically verify our assumption that the target distribution needs to be induced by demonstrations from coordinated policies, but do not necessarily need to be concurrently sampled.

2 Related Work
This section details related work, separated into work that relates to decentralized learning and that relates to distribution matching.
Cooperation in the Decentralized Setting:
Many algorithms for multi-agent cooperation tasks require some degree of information sharing between agents. Centralized training decentralized execution (CTDE) methods use a single centralized critic that aggregates information during training, but is no longer required at execution time (Lowe et al. 2017; Sunehag et al. 2018; Rashid et al. 2018; Foerster et al. 2018; Yu et al. 2022). In practical implementations, agent networks often share parameters during training as well.
Rather than sharing model components, methods may also explicitly communicate information between agents. Agents may be allowed to directly communicate information to each other (Jaques et al. 2019; Li and He 2020; Konan, Seraj, and Gombolay 2022). There might also be a central network that provides coordinating signals to all agents (He et al. 2020; Liu et al. 2021). Knowledge of other agents’ policies during training may also be assumed to limit the deviation of the joint policy (Wen et al. 2021).
This work studies the fully decentralized setting without communication or shared model components. To our knowledge, relatively few works consider this setting. Early work analyzed simple cases where two agents with similar but distinct goals could cooperate for mutual benefit under a rationality assumption (Rosenschein and Breese 1989; Genesereth, Ginsberg, and Rosenschein 1986). More recently, in the ALAN system for multi-agent navigation (Godoy et al. 2018), agents learn via a multi-armed bandits method that does not require any communication. Jiang and Lu (2021) study the decentralized multi-agent cooperation in the offline setting—in which each agent can only learn from its own data set of pre-collected behavior without communication—and propose a learning technique that relies on value and transition function error correction.
Distribution Matching in MARL:
Ho and Ermon (2016) originally proposed adversarial distribution matching as a way to perform imitation learning in the single agent setting (the gail algorithm). Song et al. (2018) extend gail to the multi-agent setting in certain respects. Their analysis sets up independent imitation learning as searching for a Nash equilibrium, and assumes that a unique equilibrium exists. Their experiments focus on training the agent policies in the CTDE paradigm, rather than the fully distributed setting. This work instead leverages recent single-agent gail convergence theory (Guan, Xu, and Liang 2021) to demonstrate convergence to the joint expert policy, and performs experiments with distributed learning. Wang et al. (2021) study MARL using copula functions to explicitly model the dependence between marginal agent policies for multi-agent imitation learning. Durugkar, Liebman, and Stone (2020) and Radke, Larson, and Brecht (2022) show that balancing individual preferences (such as matching the state-action visitation distribution of some strategies) with the shared task reward can accelerate progress on the shared task. In contrast to these works, the goal of this paper is not to study imitation learning, but rather to study how distribution matching by independent agents can enhance performance in cooperative tasks.
Perhaps most closely related to this work, Hao et al. (2019) use self-imitation learning (sil) (Oh et al. 2018) to encourage agents to repeat actions that led to high returns in the past. The above approach can be considered as a special case of the setting this paper studies, where the target distribution can be non-stationary, and is generated by the agents themselves. This paper further presents a theoretical analysis, showing that in the case where the target distribution is generated by demonstrations (and is therefore stationary), each agent attempting to minimize mismatch to their individual target distributions leads to convergence to the joint target policy. Due to the non-stationary nature of the target distribution in sil, similar guarantees cannot be obtained.
3 Background
This section describes the problem setup for MARL, imitation learning, and distribution matching.
Markov games:
A Markov game (Littman 1994) or a stochastic game (Gardner and Owen 1983) with agents is defined as a tuple , where is the set of states, and is the product of the set of actions available to each agent. The initial state distribution is described by , where indicates a distribution over the corresponding set. The transitions between states are controlled by the transition distribution . Each agent acts according to a parameterized policy , and the joint policy is the vector of the individual agent policies. Occasionally, the policy parameters are omitted for convenience. Note that each agent observes the full state. We use subscript to refer to all agents except , i.e., refers to the agent policies, .
Each agent is also associated with a reward function . The agent aims to maximize its expected return , where is the reward received by agent at time step , and the discount factor specifies how much to discount future rewards. In the cooperative tasks considered by this paper, the task rewards are identical across agents.
In Markov games, the optimal policy of an agent depends on the policies of the other agents. The best response policy is the best policy an agent can adopt, given the other agent’s policies . If no agent can unilaterally change its policy without reducing its return, then the policies are considered to be in a Nash equilibrium. That is, .
The theoretical analysis in Section 4 deals with the above fully observable setting, and assumes a discrete and finite state and action space. However, the experiments are conducted in partially observable MDPs (POMDPs) with continuous states, which can be formalized as Dec-POMDPs in the multi-agent setting (Oliehoek 2012). Dec-POMDPs include two additional elements: the set of observations and each agent’s observation function .
Distribution matching and imitation learning:
Imitation learning (Bakker and Kuniyoshi 1996; Ross, Gordon, and Bagnell 2011; Schaal 1997) is a problem setting where an agent tries to mimic trajectories where each trajectory is demonstrated by an expert policy . Various methods have been proposed to address the imitation learning problem. Behavioral cloning (Bain and Sammut 1995) applies supervised learning to expert demonstrations to recover the maximum likelihood policy. Inverse reinforcement learning (IRL) (Ng, Russell et al. 2000) recovers a reward function which can then be used to learn the expert policy using reinforcement learning. To do so, aims to recover a reward function under which the trajectories demonstrated by are optimal.
Ho and Ermon (2016) formulate imitation learning as a distribution matching problem and propose the gail algorithm. Let the state-action visitation distribution of a joint policy be:
In a multi-agent setting, for agent ,
refers to the marginal state-action visitation distribution of agent ’s policy , given the other agents’ policies . In the single agent setting, a policy that minimizes the mismatch of its state-action visitation distribution to the one induced by the expert’s trajectories and maximizes its causal entropy is a solution to the problem (Ho and Ermon 2016). That is, distribution matching is a solution to the imitation learning problem.
Guan, Xu, and Liang (2021) showed that in the single-agent case, the gail algorithm converges to the expert policy under a variety of policy gradient techniques, including TRPO (Schulman et al. 2015). Let be a reward function (based on a discriminator) parameterized by , and let be a convex regularizer. Guan, Xu, and Liang (2021) formulate the gail problem as the following min-max problem:
(1) | ||||
where is the expected return from some start state when following policy and using reward function .
In the multi-agent setting, imitation learning has the added complexity that the expert trajectories are generated by the interaction of multiple expert policies . Successful imitation in this setting thus involves the coordination of all agents’ policies.
4 Theoretical Analysis
This section provides theoretical grounding for the core proposition of this paper. The target distribution is assumed to be the empirical distribution of demonstrations from a set of “expert” agents in order to ensure that it is achievable by the agents. Under the conditions stated below, the analysis shows that if agents independently minimize the distribution mismatch to their respective demonstrations in a turn-by-turn fashion, then agent policies will converge to the joint expert policy.
The three conditions are as follows. First, this joint expert policy needs to be coordinated,111Condition is made concrete in Section 5. but does not have to be a Nash equilibrium with respect to any particular task. Second, for every policy considered, there is a minimal probability of visiting each state. Third, each agent learns via a single-agent imitation learning algorithm such that it improves its distribution matching reward at each step.
Next, we establish that if the agents are learning to maximize the mixture of an extrinsic task reward and a distribution matching reward, then the agent policies will converge to a Nash equilibrium with respect to the joint reward.
Convergence of Independent gail Learners
This analysis considers the setting where each agent performs independent learning updates according to the gail algorithm, to match the visitation distribution of the expert. It proposes a condition on an individual agent’s gail objective improvement. If this condition is satisfied, it shows that a lower bound on a joint distribution matching objective is improved. Further, the lower bound objective converges, demonstrating the convergence of independent gail.
Let the parameterized (discriminator) reward of agent be , for . At each agent’s update, all the other agent policies are held fixed, and the corresponding discriminator has converged to , where . Guan, Xu, and Liang (2021) showed that the learning process of a single agent repeatedly updating converges to . The update scheme we consider for theoretical purposes is specified in Algorithm 2, located in Appendix A. 222A technical appendix is included in the arXiv version of this paper, https://arxiv.org/abs/2206.00233.
Define the per-agent gail loss as follows:
where is the discounted state visitation distribution.
Consider the random variable that is the indicator function for the event that at state , agent would take an action that matched expert ’s action. Note that the expectation of this indicator is the probability of matching the expert’s action333For the purpose of exposition, assume that the expert policy is deterministic. The theory in this section can be extended to the case where is stochastic by comparing the distributions over actions.. Define the joint action-matching objective as the probability that agent actions match their corresponding experts (plus a constant), weighted by the probability of visiting states:
(2) |
where indicates the event that all agents take actions that match their corresponding experts. Maximizing precisely corresponds to solving the multi-agent imitation learning problem because the joint expert policy is the unique maximizer of (Lemma 5, Appendix A).
Theorem 1 (action-matching objective).
The joint action-matching objective is lower bounded by the following sum over individual action-matching rewards :
(3) |
When an agent updates its policy to optimize its component of , the state visitation distribution might change such that the expected action rewards for other agents decrease. The next corollary introduces a lower bound on that is independent of the state visitation distribution .
Corollary 1 (lower bound).
Let be the minimum probability of visiting any state. For all , is lower bounded by :
(4) |
With the lower bound, we make the following assumption to relate our action-matching reward to the GAIL discriminator reward that is improved by the GAIL algorithm.
Assumption 1 (action-matching reward).
For all agents and all states , an increase in the expected converged GAIL discriminator reward implies an increase to the expected action-matching reward function:
If the reward , as it is in gail, then the assumption above is valid (see Appendix A).
Assumption 1 ensures each agent updating its policy leads to improvement in . This is a lower bound on the actual objective of interest —by Theorem 1 and Corollary 1. Further, has a unique global maximizer, which is (Lemma 7). Thus, while the action reward for the other agents might decrease in the short term, the joint action matching objective across all agents will increase as the learning process continues. Since is bounded from above (Lemma 4), this process of improving the lower bound will converge to the optimal policy for this objective—the joint expert policy.
Theorem 2 (convergence).
Each agent maximizing its individual return over the individual action rewards will converge to the joint expert policy .
Multi-agent Learning with Mixed Task and Imitation Reward
Lemma 5 in Appendix A shows that the joint expert policy uniquely maximizes the joint imitation learning objective. Let be the optimal discriminator parameters for the expert, . From a game theoretic perspective, this lemma implies that these expert policies are a Nash equilibrium for the imitating agents with respect to .
Next, note that in imitation learning, it is typically not necessary for the agents to know what the demonstration actor’s task reward is. However, suppose that the agents have access to both demonstrations from policies optimal at task , and the corresponding reward function .
Let , and let the expert policies maximize . The expert policies that maximize are in a Nash equilibrium with respect to . Theorem 3 states that if the agents are trained to maximize a reward function that is a linear combination of the task reward and , then the converged agent policies are also in a Nash equilibrium with respect to .
Theorem 3.
Let be the reward function used to train the expert policies , and let the expert policies have converged with respect to (i.e., they are in a Nash equilibrium with respect to reward ). Then are a Nash equilibrium for reward functions of the form, , for any .
Theorem 2 does not require that the demonstrations originate from optimal policies for some task. However, Theorem 3 implies that if the demonstrations do maximize the reward of a desired task, then the task reward and distribution matching reward can be combined to optimize the same task. The proposed algorithm, dm2, takes this approach.
5 Methods
This section discusses practical considerations of fully distributed multi-agent distribution matching, and proposes dm2, an algorithm whose performance is analyzed in Section 6.
Generating Expert Demonstrations
Section 4 shows that agents individually following demonstrations from an existing joint policy can converge to said joint policy without centralized training or communication. In practice, these demonstrations should imply an achievable joint expert policy.
For illustration, consider a four tile gridworld, where only one agent is allowed on a tile at a time. Let one of the tiles be labelled, “A”. Suppose there are two agents, and each agent is provided with a separate target state-action distribution, consisting of agent occupying a tile “A”, and the other agent occupying one of the three remaining tiles. If both agents simultaneously attempt to match their provided target distributions, then both agents will attempt to occupy tile “A”. With these demonstrations, it is impossible for both agents to fully match their desired distributions.
The example above shows that for each agent to completely match its desired distribution, the state-action distributions for all agents must be compatible in some way. This notion of compatibility is defined below.
Definition 1 (Compatible demonstrations).
State-action visitation distributions from a collection of policies (where are the other agent policies executed with to obtain the state-action visitation distribution ) are compatible if for all , , there exists a joint policy with the joint state-action visitation distribution (Equation 3) such that the marginal state-action visitation distribution for agent is:
Observe that expert policies that are trained in the same environment to perform a task induce compatible individual state-action visitation distributions, providing a practical method to obtain compatible demonstrations.
Practical Multi-Agent Distribution Matching
dm2 is inspired by the theoretical analysis in Section 4, and balances the individual objective of distribution matching with the shared task. To do so, the agents are provided a mixed reward: part cost function for minimizing individual distribution mismatch, part environment reward. This approach has been shown to be effective in balancing individual preferences with shared objectives in MARL (Durugkar, Liebman, and Stone 2020; Cui et al. 2021). The individual agent policies are learned by independently updating each agent’s policy using an on-policy RL algorithm of choice. The experiments here use ppo (Schulman et al. 2017).
In the experiments, the demonstrations used as targets for the distribution matching are compatible, state-only trajectories—i.e., originating from policies trained jointly on the task of interest. The use of state-only demonstrations enables learning purely from observations of other agents (e.g. online videos), and is supported by research in the subarea of imitation from observation alone (Torabi, Warnell, and Stone 2019). Our experiments also validate the effectiveness of the approach in this setting. In Section 6, we show that the demonstrator policies may possess intermediate competency in the task at hand, and that the demonstrations do not need to be jointly sampled for all the agents. The proposed learning scheme for training individual agents is summarized by Algorithm 1 and Figure 1.
6 Experimental Evaluation



This section presents two main experiments. The first experiment evaluates whether dm2 may improve coordination–and therefore efficiency of learning–over a decentralized MARL baseline. A comparison against CTDE algorithms is also performed. The second experiment is an ablation study on the demonstrations that are provided to dm2. These ablations seek to answer the question whether the faster, improved learning above is due to coordination between experts, or due to individual experts being competent.
Additional experiments in Appendix B evaluate the effect of the demonstration quality on learning, analyze the usage of demonstrations for behavioral cloning (Bain and Sammut 1995) instead of distribution matching, and examine the impact of using only distribution matching gail rewards for learning instead of mixing them with the task rewards. The code is provided at https://github.com/carolinewang01/dm2.
Environments:
Experiments were conducted on the StarCraft Multi-Agent Challenge domain (Samvelyan et al. 2019). It features cooperative tasks where a team of controllable allied agents must defeat a team of enemy agents. The enemy agents are controlled by a fixed AI. The battle is won and the episode terminates if the allies can defeat all enemy agents. The allies each receive a team reward every time an enemy agent is killed, and when the battle is won. StarCraft is a partially observable domain, where an allied agent can observe features about itself, as well as allies and enemies within a fixed radius. The specific StarCraft tasks used here (with two additional tasks in Appendix B) are:
-
•
5v6: 5 Marines (allies) and 6 Marines (enemies)
-
•
3sv4z: 3 Stalkers (allies) and 4 Zealots (enemies)
-
•
3sv3z: 3 Stalkers (allies) and 3 Zealots (enemies)
Baselines:
dm2 is compared against a naive decentralized MARL algorithm, independent ppo (Schulman et al. 2017) (ippo), where individual ppo agents directly receive the team environment reward. Although agents trained under the ippo scheme cannot share information and see only local observations, prior work has shown that ippo can be surprisingly competitive with CTDE methods (Yu et al. 2022). We also compare against two widely used CTDE methods, qmix (Rashid et al. 2018) and rmappo (Yu et al. 2022). These CTDE methods have the advantage of a shared critic network that receives the global state during training. Thus, their performance is expected to be better than that of decentralized methods with no communication.
Setup:
dm2 uses the same ippo implementation as the baseline, with the addition of a gail discriminator for each independent agent to generate an imitation reward signal, . The scaled gail reward is added to the environment reward , with scaling coefficient : . Learning curves of all algorithms are the mean of 5 runs executed with independent random seeds, where each run is evaluated for 32 test episodes at regular intervals during training. The shaded regions on the plots show the standard error. The evaluation metric is the mean rate of battles won against enemy teams during test episodes.
The data for the gail discriminator consists of 1000 joint state-only trajectories (no actions). The data is sampled from checkpoints during training runs of baseline ippo with the environment reward, and qmix with the environment reward. In runs of dm2, each agent imitates the marginal observations of the corresponding agent from the dataset (i.e., agent will imitate agent ’s observations from the dataset) 444The allied agent teams in our experiments have the same state/action spaces. Thus, the mapping of agents to demonstration trajectories does not matter, as long as it is fixed. For each task, demonstrations are sampled from ippo and qmix-trained joint expert policies, executed stochastically for ippo and with an -greedy sampling for qmix. The win rates achieved by the demonstration policies are plotted as horizontal lines on the graphs. Additionally, sil(Hao et al. 2019) can be seen as a variation of dm2, with the target demonstrations being the agent’s most successful prior trajectories. Experimental details such as hyperparameters are specified in Appendix C.



Main Results
Figure 2 shows that in all three tasks, dm2 significantly improves learning speed over ippo (the decentralized baseline). qmix and rmappo (the CTDE baselines) learn faster than dm2 and ippo on both tasks, illustrating the challenging nature of the decentralized cooperation problem. However, on 5v6 and 3sv3z, all methods converge to a similar win rate towards the end of training. For the demonstrations from ippo experts, dm2 surpasses the win rate of the demonstrations. Despite the significant variance in win rates among the demonstrations for each task, dm2 performs similarly. Similar robustness to demonstration quality is seen even with 10 different demonstration qualities (Figure 7 in Appendix B). The relative invariance to demonstration quality suggests that the demonstrations provide a useful coordinating signal, enabling agents to discover higher-return behaviors than those portrayed in the demonstrations.
On the other hand, using trajectories from earlier in training as demonstrations (sil) shows inconsistent performance. Its performance is comparable to rmappo in 5v6, but it learns more slowly in the other domains. This may occur because sil requires examples of successful coordination by the agents, which may be rare in certain tasks. Using past trajectories as demonstrations also leads to a nonstationary target distribution for imitation, potentially negatively impacting the learning procedure.
Ablation Study
This section presents an ablation study on the demonstrations. It investigates whether the coordination of expert agents or their individual competency is more important to the success of dm2 in the main experiment. This comparison is done by considering demonstrations that vary in two dimensions: whether they were sampled from expert teams that form a joint policy (co-trained), or whether they were sampled simultaneously (concurrently sampled).
The first dimension tests the requirement that agents are trained to coordinate with each other, while the second tests whether agents must act together when generating demonstrations. Concurrently sampled demonstrations of agents that were not co-trained, gives us examples of individually competent agents acting in the multi-agent setting.
The experiments apply dm2 to four possible demonstration styles that vary in the aforementioned two dimensions. A detailed explanation of how these four demonstration styles were constructed is provided in Appendix C. The study is performed on the 5v6 task, with the same hyperparameters used in the experiments of the previous section.
Figure 3 shows the learning curves of the four combinations. The axis that appears to make the greatest difference in learning is whether the demonstrations originate from expert policies that were co-trained, and were thus coordinated. Whether the agent demonstrations were concurrently sampled does not appear to significantly impact learning. Similar trends are observed when dm2 is trained with the lower quality demonstration (Figure 3, left).
7 Discussion and Future Work
This paper studies distributed MARL for cooperative tasks without communication or explicit coordination mechanisms. Fully distributed MARL is challenging, since simultaneous updates to different agents’ policies can cause them to diverge. The benefits of distributed MARL are abundant. Decentralized training could make agents more robust to the presence of agents they were not trained with (e.g. humans). Decentralized training could also enable coordination while preserving the privacy of each agent.
The theoretical analysis of this paper shows that individual agents updating their policies turn-by-turn to reduce their distribution mismatch to corresponding expert distributions improves a lower bound to the joint action-matching objective against the joint expert policy. Fully maximizing the lower bound corresponds to recovering the joint expert policy. The experiments verify that mixing the task reward with the distribution matching reward accelerates cooperative task learning, compared to learning without the distribution matching objective. The ablation experiments show that expert demonstrations should be from policies that were trained together, but not necessarily concurrently sampled.
While this work is a meaningful step towards fully distributed multi-agent learning via distribution matching, some open questions remain. Future work could consider whether demonstrations sampled from expert policies with other properties, such as those trained with reward signals corresponding to different tasks, could be beneficial for distributed learning. The method proposed in this paper could also be leveraged to combine human demonstrations with a task reward for applications of MARL ranging from expert decision making (similar to that done by Gombolay et al. (2018) in the context of medical recommendation) or in the context of complex multi-agent traffic navigation (Behbahani et al. 2019). Another potential path forward would be considering human in the loop settings such as the TAMER architecture (Knox and Stone 2009), but in a fully distributed multi-agent setting.
Acknowledgements
This work has taken place in the Learning Agents Research Group (LARG) at the Artificial Intelligence Laboratory, The University of Texas at Austin, and at SparkCognition Applied Research. LARG research is supported in part by the National Science Foundation (CPS-1739964, IIS-1724157, FAIN-2019844), the Office of Naval Research (N00014-18-2243), Army Research Office (W911NF-19-2-0333), DARPA, General Motors, Bosch, and Good Systems, a research grand challenge at the University of Texas at Austin. The views and conclusions contained in this document are those of the authors alone. Peter Stone serves as the Executive Director of Sony AI America and receives financial compensation for this work. The terms of this arrangement have been reviewed and approved by the University of Texas at Austin in accordance with its policy on objectivity in research.
References
- Bain and Sammut (1995) Bain, M.; and Sammut, C. 1995. A Framework for Behavioural Cloning. In Machine Intelligence 15, 103–129.
- Bakker and Kuniyoshi (1996) Bakker, P.; and Kuniyoshi, Y. 1996. Robot see, robot do: An overview of robot imitation. In AISB96 Workshop on Learning in Robots and Animals, 3–11.
- Barrett and Stone (2012) Barrett, S.; and Stone, P. 2012. An analysis framework for ad hoc teamwork tasks. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems-Volume 1, 357–364.
- Behbahani et al. (2019) Behbahani, F.; Shiarlis, K.; Chen, X.; Kurin, V.; Kasewa, S.; Stirbu, C.; Gomes, J.; Paul, S.; Oliehoek, F. A.; Messias, J.; et al. 2019. Learning from demonstration in the wild. In 2019 International Conference on Robotics and Automation (ICRA), 775–781. IEEE.
- Cui et al. (2021) Cui, J.; Macke, W.; Yedidsion, H.; Goyal, A.; Urielli, D.; and Stone, P. 2021. Scalable Multiagent Driving Policies For Reducing Traffic Congestion. In Proceedings of the 20th International Conference on Autonomous Agents and Multi Agent Systems (AAMAS).
- Durugkar, Liebman, and Stone (2020) Durugkar, I.; Liebman, E.; and Stone, P. 2020. Balancing individual preferences and shared objectives in multiagent reinforcement learning. In Proceedings of the 29th International Joint Conference on Artificial Intelligence (IJCAI).
- Foerster et al. (2018) Foerster, J.; Farquhar, G.; Afouras, T.; Nardelli, N.; and Whiteson, S. 2018. Counterfactual Multi-Agent Policy Gradients. Proceedings of the AAAI Conference on Artificial Intelligence, 32.
- Gardner and Owen (1983) Gardner, R.; and Owen, G. 1983. Game Theory (2nd Ed.). Journal of the American Statistical Association, 78: 502.
- Genesereth, Ginsberg, and Rosenschein (1986) Genesereth, M. R.; Ginsberg, M. L.; and Rosenschein, J. S. 1986. Cooperation without Communication. In Proceedings of the AAAI Conference on Artificial Intelligence.
- Godoy et al. (2018) Godoy, J.; Chen, T.; Guy, S. J.; Karamouzas, I.; and Gini, M. L. 2018. ALAN: adaptive learning for multi-agent navigation. Autonomous Robots, 42: 1543–1562.
- Gombolay et al. (2018) Gombolay, M.; Yang, X. J.; Hayes, B.; Seo, N.; Liu, Z.; Wadhwania, S.; Yu, T.; Shah, N.; Golen, T.; and Shah, J. 2018. Robotic assistance in the coordination of patient care. The International Journal of Robotics Research, 37(10): 1300–1316.
- Guan, Xu, and Liang (2021) Guan, Z.; Xu, T.; and Liang, Y. 2021. When Will Generative Adversarial Imitation Learning Algorithms Attain Global Convergence. In AISTATS.
- Hao et al. (2019) Hao, X.; Wang, W.; Hao, J.; and Yang, Y. 2019. Independent Generative Adversarial Self-Imitation Learning in Cooperative Multiagent Systems. In Proceedings of the 18th International Conference on Autonomous Agents and Multi Agent Systems, 1315–1323.
- He et al. (2020) He, X.; An, B.; Li, Y.; Chen, H.; Wang, R.; Wang, X.; Yu, R.; Li, X.; and Wang, Z. 2020. Learning to Collaborate in Multi-Module Recommendation via Multi-Agent Reinforcement Learning without Communication. Fourteenth ACM Conference on Recommender Systems.
- Hernandez-Leal, Kartal, and Taylor (2019) Hernandez-Leal, P.; Kartal, B.; and Taylor, M. E. 2019. A survey and critique of multiagent deep reinforcement learning. Autonomous Agents and Multi-Agent Systems, 33: 750 – 797.
- Ho and Ermon (2016) Ho, J.; and Ermon, S. 2016. Generative adversarial imitation learning. Advances in neural information processing systems, 29: 4565–4573.
- Jaques et al. (2019) Jaques, N.; Lazaridou, A.; Hughes, E.; Gulcehre, C.; Ortega, P.; Strouse, D.; Leibo, J. Z.; and De Freitas, N. 2019. Social Influence as Intrinsic Motivation for Multi-Agent Deep Reinforcement Learning. In Chaudhuri, K.; and Salakhutdinov, R., eds., Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research. PMLR.
- Jiang and Lu (2021) Jiang, J.; and Lu, Z. 2021. Offline Decentralized Multi-Agent Reinforcement Learning. ArXiv, abs/2108.01832.
- Knox and Stone (2009) Knox, W. B.; and Stone, P. 2009. Interactively shaping agents via human reinforcement: The TAMER framework. In Proceedings of the fifth international conference on Knowledge capture, 9–16.
- Konan, Seraj, and Gombolay (2022) Konan, S.; Seraj, E.; and Gombolay, M. 2022. Iterated Reasoning with Mutual Information in Cooperative and Byzantine Decentralized Teaming. In ICLR.
- Léauté and Faltings (2013) Léauté, T.; and Faltings, B. 2013. Protecting privacy through distributed computation in multi-agent decision making. Journal of Artificial Intelligence Research, 47: 649–695.
- Leibo et al. (2017) Leibo, J. Z.; Zambaldi, V.; Lanctot, M.; Marecki, J.; and Graepel, T. 2017. Multi-agent Reinforcement Learning in Sequential Social Dilemmas. In Proceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems, 464–473.
- Li and He (2020) Li, H.; and He, H. 2020. Multi-Agent Trust Region Policy Optimization. CoRR, abs/2010.07916.
- Littman (1994) Littman, M. L. 1994. Markov games as a framework for multi-agent reinforcement learning. In Machine learning proceedings 1994, 157–163. Elsevier.
- Liu et al. (2021) Liu, B.; Liu, Q.; Stone, P.; Garg, A.; Zhu, Y.; and Anandkumar, A. 2021. Coach-Player Multi-Agent Reinforcement Learning for Dynamic Team Composition. In International Conference on Machine Learning.
- Lowe et al. (2017) Lowe, R.; Wu, Y.; Tamar, A.; Harb, J.; Abbeel, P.; and Mordatch, I. 2017. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. In NeurIPS.
- Marinescu, Dusparic, and Clarke (2017) Marinescu, A.; Dusparic, I.; and Clarke, S. 2017. Prediction-based multi-agent reinforcement learning in inherently non-stationary environments. ACM Transactions on Autonomous and Adaptive Systems (TAAS), 12(2): 1–23.
- Ng, Russell et al. (2000) Ng, A. Y.; Russell, S. J.; et al. 2000. Algorithms for inverse reinforcement learning. In Icml, volume 1, 663–670.
- Oh et al. (2018) Oh, J.; Guo, Y.; Singh, S.; and Lee, H. 2018. Self-imitation learning. In International Conference on Machine Learning, 3878–3887. PMLR.
- Oliehoek (2012) Oliehoek, F. A. 2012. Decentralized POMDPs, 471–503. Berlin, Heidelberg: Springer Berlin Heidelberg. ISBN 978-3-642-27645-3.
- Radke, Larson, and Brecht (2022) Radke, D.; Larson, K.; and Brecht, T. B. 2022. Exploring the Benefits of Teams in Multiagent Learning. ArXiv, abs/2205.02328.
- Rashid et al. (2018) Rashid, T.; Samvelyan, M.; Schroeder, C.; Farquhar, G.; Foerster, J.; and Whiteson, S. 2018. QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning. In Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research. PMLR.
- Rosenschein and Breese (1989) Rosenschein, J. S.; and Breese, J. S. 1989. Communication-Free Interactions among Rational Agents: A Probabilistic Approach. In Distributed Artificial Intelligence.
- Ross, Gordon, and Bagnell (2011) Ross, S.; Gordon, G.; and Bagnell, D. 2011. A reduction of imitation learning and structured prediction to no-regret online learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, 627–635.
- Samvelyan et al. (2019) Samvelyan, M.; Rashid, T.; de Witt, C. S.; Farquhar, G.; Nardelli, N.; Rudner, T. G. J.; Hung, C.-M.; Torr, P. H. S.; Foerster, J.; and Whiteson, S. 2019. The StarCraft Multi-Agent Challenge. CoRR, abs/1902.04043.
- Schaal (1997) Schaal, S. 1997. Learning from demonstration. In Advances in neural information processing systems, 1040–1046.
- Schulman et al. (2015) Schulman, J.; Levine, S.; Abbeel, P.; Jordan, M.; and Moritz, P. 2015. Trust region policy optimization. In International conference on machine learning, 1889–1897. PMLR.
- Schulman et al. (2017) Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; and Klimov, O. 2017. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347.
- Silver et al. (2018) Silver, D.; Hubert, T.; Schrittwieser, J.; Antonoglou, I.; Lai, M.; Guez, A.; Lanctot, M.; Sifre, L.; Kumaran, D.; Graepel, T.; et al. 2018. A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science, 362(6419): 1140–1144.
- Silver et al. (2017) Silver, D.; Schrittwieser, J.; Simonyan, K.; Antonoglou, I.; Huang, A.; Guez, A.; Hubert, T.; Baker, L.; Lai, M.; Bolton, A.; et al. 2017. Mastering the game of go without human knowledge. Nature, 550(7676): 354–359.
- Song et al. (2018) Song, J.; Ren, H.; Sadigh, D.; and Ermon, S. 2018. Multi-Agent Generative Adversarial Imitation Learning. In Advances in Neural Information Processing Systems, volume 31.
- Sunehag et al. (2018) Sunehag, P.; Lever, G.; Gruslys, A.; Czarnecki, W. M.; Zambaldi, V.; Jaderberg, M.; Lanctot, M.; Sonnerat, N.; Leibo, J. Z.; Tuyls, K.; and Graepel, T. 2018. Value-Decomposition Networks For Cooperative Multi-Agent Learning Based On Team Reward. In Proceedings of the 17th International Conference on Autonomous Agents and Multi Agent Systems, AAMAS ’18.
- Thrun (1998) Thrun, S. 1998. Lifelong learning algorithms. In Learning to learn, 181–209. Springer.
- Torabi, Warnell, and Stone (2019) Torabi, F.; Warnell, G.; and Stone, P. 2019. Generative Adversarial Imitation from Observation. arXiv:1807.06158 [cs, stat]. ArXiv: 1807.06158.
- Wang et al. (2021) Wang, H.; Yu, L.; Cao, Z.; and Ermon, S. 2021. Multi-agent Imitation Learning with Copulas. In Machine Learning and Knowledge Discovery in Databases. Research Track, 139–156.
- Wen et al. (2021) Wen, Y.; Chen, H.; Yang, Y.; Tian, Z.; Li, M.; Chen, X.; and Wang, J. 2021. A Game Theoretic Approach to Multi-Agent Trust Region Optimization.
- Yang and Wang (2020) Yang, Y.; and Wang, J. 2020. An overview of multi-agent reinforcement learning from game theoretical perspective. ArXiv, abs/2011.00583.
- Yu et al. (2022) Yu, C.; Velu, A.; Vinitsky, E.; Wang, Y.; Bayen, A.; and Wu, Y. 2022. The Surprising Effectiveness of MAPPO in Cooperative Multi-Agent Games. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks.
Appendix
Appendix A Convergence Proof Details
In Section 4, we lay out some of the conditions for our theoretical analysis. One of these conditions, that for every policy considered there is a minimal probability of visiting each state, is formalized below.
Condition 1.
Let , and let be the state visitation distribution induced by any joint policy during training. For all agents and for all , suppose that .
Recall that the joint action-matching objective is defined over the expected state visitation as the probability that all agent actions match their corresponding experts (plus a constant):
where indicates the event that all the agents took actions that matched their corresponding experts.
We first prove some properties of .
Lemma 4.
The objective is bounded by
Proof.
As shorthand, define . Then,
First, note that because it is a weighted sum of expectations over indicator functions, where all weights are non-negative, and it is precisely if for all states, the joint agent policy does not match the correct expert actions. Thus, is lower bounded by . For the upper bound, notice that the outer summation is equivalent to the expectation under state visitation distribution . Inside, each expectation over the indicators can be at most 1, implying that is at most . Thus, . ∎
Lemma 5.
Suppose the joint expert policy is deterministic. Then is the unique maximizer of .
Proof.
(Maximization) Since is deterministic, by definition, each term
Thus, , which means that achieves the upper bound of .
(Uniqueness) Suppose there exists another policy that also achieves the upper bound, i.e. . Let . Then there must be an agent such that with positive probability, such that . Then it is immediate that at state , . Combined with the non-zero probability of visiting every state (Condition 1), this inequality then implies . By contradiction, is the unique optimizer for . ∎
Next, we establish that individual agents performing GAIL updates maximizes a lower bound on . We leverage the single-agent GAIL convergence result by (Guan, Xu, and Liang 2021) to show this result.
Lemma 6.
Let be the time step at which agent ’s policy is updated. For all agents , denote the optimal discriminators of Equation 1 as . Suppose agent updates its policy parameters from to such that . This decrease in loss is equivalent to increasing the agent ’s expected discriminator reward.
Proof.
First, note that updating a single agent policy while keeping all discriminators fixed does not alter the expert value term or the regularizer term in the loss definition . Thus, the condition that agent ’s loss has decreased is equivalent to the value of agent increasing:
(5) |
For convenience of notation, agent ’s policy at time t will be written as , and the th discriminator implicitly indicated by the action subscript, . Similarly, we will write the state visitation distribution induced by the policies by , and the distribution induced by as .
Rewriting Equation 5 in terms of visitation distributions:
(6) |
∎
We next show that is lower bounded by the sum of the individual action-matching rewards for all agents , over all states.
See 1
Proof.
Let us begin by rewriting in terms of action mismatches.
This formulation allows us to apply the Union Bound:
∎
Theorem 1 relates the multi-agent imitation learning objective to a sum over single-agent imitation learning objectives. This is important because each agent updates independently to improves its own learning objective in our setting. Note also that the additive form of is similar to the value factorization assumptions made by algorithms like VDN (Sunehag et al. 2018).
It is difficult to say anything directly about the expected action-matching reward of agent , , as may be a state distribution over which agent makes more mistakes (i.e. taking actions that don’t match the expert’s). While in general agent ’s expected reward may decrease due to agent ’s update, we show that under Assumptions 1 and 1, agent ’s update increases a lower bound to that is independent of the state distribution.
See 1
Proof.
The proof follows from the definition of and that for all , and all encountered in training, . ∎
Observation 1.
is bounded by
Lemma 7.
Suppose the joint expert policy is deterministic. Then is the unique maximizer of .
Proof.
Proof for this Lemma follows closely the proof for Lemma 5.
(Maximization) Since is deterministic, by definition, for each agent , each term
Summing over all agents and taking the sum weighted by over all states, , which means that achieves the upper bound of .
(Uniqueness) Suppose there is at least one agent policy such that the joint policy also achieves the upper bound, i.e. . Then there must be a state such that with non-zero probability, such that . It follows that , meaning . By contradiction, is the unique optimizer of . ∎
Lemma 6 states that a single agent updating its policy to improve its own GAIL loss is equivalent to increasing the expected action reward , where the expectation over states is with respect to some updated state visitation distribution (Equation 6). Next, we make an assumption to relate our action-matching reward to the GAIL discriminator reward that is improved by the GAIL algorithm.
See 1
This assumption is not as strong as it may first appear.
First, note that the converged GAIL reward consists of the negative discriminator prediction, .
The discriminator predicts the likelihood ratio between the target visitation and the mix of target and agent visitation.
The minimal state visitation assumption states that for all , allowing us to relate to a state-visitation independent reward, . As we will argue next, is a similar quantity to the action-matching indicator reward.
To see this, first suppose , which would imply that the action-matching indicator function is 0. Since the expert policy is assumed to be deterministic, , implying that as well. If , then is not zero, and the only way for agent to increase is to increase — the probability of matching the expert’s action. Thus, an increase in implies an increase in , which behaves similarly to the action-matching indicator function.
Thus far, we have established that each individual agent’s policy improvement under the GAIL reward improves a lower bound to the joint action-matching objective, . The following shows that within finite updates, each agent will be able to independently improve its value function until it converges to the expert policy.
Condition 2.
Let denote the value of an agent following a single-agent imitation learning algorithm. is then the optimality gap at update of the agent. Suppose that , where as , .
This condition says that the single-agent imitation learning process should converge to the optimal (expert) policy with convergence rate dictated by . For our setting, Guan, Xu, and Liang (2021) shows that the single-agent GAIL algorithm converges (Theorem 3 and 4).
The next corollary shows that convergence of the single-agent imitation learning process is sufficient to guarantee the convergence of the multi-agent imitation learning scheme discussed in the main paper.
Corollary 2.
There exists such that within updates, agent is able to improve its policy such that it increases the probability of matching the expert’s action, summed over all states:
Proof.
For a single agent , define . Rewrite as follows:
Define the first term as , and the second term as . Note that is the quantity of interest in the corollary.
The properties of the max operation directly imply that
The expert policy should maximize for any single-agent imitation learning algorithm. Note also that maximizes and . Thus in our setting, the inequality in the above set of equations is actually an equality for the expert policy:
Assume that there is a policy such that , and that cannot be improved within finite updates. This contradicts Condition 2, which establishes that the single-agent imitation learning algorithm should be able to improve the agent policy until its value converges to the value of the expert policy. ∎
Corollary 2 implies that within a finite number of policy updates by agent , the quantity increases, because the other terms corresponding to agents are unchanged. By Theorem 2 and Lemma 7, is upper bounded by the constant . Thus, by the monotone convergence theorem, the objective converges. Further, the expert policy is a maximizer of , , and (Lemma 5 and Lemma 7).
Mixed Task and Imitation Reward
See 3
Proof.
Let . The following reasoning is on a per-agent basis, so we drop the from and for convenience. For to not be a Nash equilibrium with respect to there needs to exist a policy such that
That implies
But by definition, for all ,
and
which is a contradiction. ∎


Appendix B Supplemental Experimental Results
This section provides further commentary on results in the main paper and presents additional, supporting experiments. Figure 4 is an extension of our core results on two more maps in the Starcraft domain (3sv5z and 8v9), presented here due to space constraints.
Demonstration Styles of Ablation Study.
The ablation study examines expert demonstrations that vary in two dimensions: co-trained versus concurrently sampled. For co-trained agents with demonstrations sampled non-concurrently, the demonstrations may be sampled from co-trained expert policies, but each agent’s demonstrations originate from disjoint episodes. However, for agents that were not trained together but whose demonstrations are sampled concurrently, demonstrations could be obtained from expert policies that were each trained in separate teams, but executed together in the same environment. To ensure that each expert policy is of similar quality—despite not being trained together—the joint expert policies are trained with different seeds of the same algorithm.
Behavioral Cloning Pretraining.
Distribution matching is not the only method to use demonstrations. A more naive approach to utilize demonstrations is to use behavioral cloning (BC) (Bain and Sammut 1995) (a form of supervised learning) on the dataset of state-action pairs . BC is accomplished by learning to predict the maximum likelihood action according to the dataset on the states present in the dataset. In practice, this prediction is learned by minimizing the negative log likelihood of the expert action on these states. This experiment pre-trains the agent policy with BC before the agents interact with the environment, after which point they learn using IPPO.
BC typically suffers from a distribution mismatch problem (also known as the covariate shift problem), where the agent’s state visitation when interacting with the environment differs from the expert data distribution, leading to poor imitation even in single-agent settings. Behavioral cloning also requires a dataset of a size that increases quadratically with the horizon of the problem to learn successful policies (Ross, Gordon, and Bagnell 2011). These issues are likely to be exacerbated when dealing with multiple agents. The agents might minimize the supervised learning loss, but it is unlikely that the agents would learn to coordinate effectively. A second issue with using the supervised learning loss to pre-coordinate agent policies is that such coordination is unlikely to last once training with IPPO proceeds.
The result for this alternative usage of the expert demonstrations is presented in Figure 5. As detailed above, BC does not learn to imitate the demonstrations well enough to recover their performance (indicated by the dashed lines), and as training proceeds, the benefits of BC vanish as IPPO training proceeds.


Effect of GAIL Reward.
DM2 consists of IPPO trained with , which is a mixture of the environment and GAIL reward signal. Here, we present a brief ablation on the mixed reward (Figure 6). For both IPPO demonstration qualities, IPPO trained on the GAIL reward achieves a far lower win rate than IPPO trained on the environment reward (labelled as IPPO baseline) or DM2. This result provides further evidence that the primary benefit derived by DM2 comes from the coordination shown in the demonstrations, rather than from individual imitation of expert behaviors.


Sensitivity to Demonstration Quality.
One finding in the experimental section of the paper is that DM2 is relatively insensitive to demonstration quality beyond a certain baseline level of competence. To further support this claim, we train DM2 with more demonstration qualities (where demonstrations are sampled from IPPO policies). Figure 7 shows that the algorithm improves monotonically as the demonstration quality improves, but quickly saturates. This result indicates it is important to supply good demonstrations, but not necessary to supply optimal demonstrations.

Appendix C Experimental Details

Implementation Details.
The algorithm implementations are based on the multi-agent PPO implementations provided by Yu et al. (2022) (MIT license) and the PyMARL code base (Samvelyan et al. 2019) (Apache license). The StarCraft environment is also provided by Samvelyan et al. (2019) (MIT license).
All decentralized MARL implementations in this paper have fully separate policy/critic networks and optimizers per agent. That is, there is no parameter sharing amongst agents. For all IPPO agents, the policy architecture is two fully connected layers, followed by an RNN (GRU) layer. Each layer has 64 neurons with ReLU activation units. The critic architecture is the same as the policy architecture. For DM2 agents, the policy and critic architectures are identical to IPPO. The discriminator architecture consists of two fully connected layers with tanh activation functions.
The centralized MARL algorithms implement agent policy networks with parameter sharing, where agents have a centralized value network. For QMIX agents, the policy architecture is the same except there is only a single fully connected layer before the RNN layer 555This is the architecture used in Rashid et al. (2018). We attempted running QMIX with the the IPPO agent architecture, but found that the performance of QMIX significantly suffered (Figure 8 on 5v6). Thus, for the QMIX experiments in the main body of the paper, the better-performing policy architecture was applied. RMAPPO agents were trained directly using the code published by Yu et al. (2022).
Hyperparameters.
For QMIX, the default parameters specified in Rashid et al. (2018) are used for both tasks. For IPPO, and the IPPO component of DM2, mostly default parameters (as specified in (Rashid et al. 2018; Yu et al. 2022)) were used. Algorithm hyperparameters that varied between tasks or were tuned are provided in Table 2. The remaining hyperparameters may be viewed with the code repository.
We found that for DM2 to learn successfully from QMIX demonstrations, it was sometimes necessary to inject a small amount of random noise into the demonstration sampling process, so that the demonstrations did not constrain the exploration of the learning policies. Specifically, the demonstrations from QMIX were sampled from -greedy QMIX policies, where was chosen so that the win rate did not fall more than . QMIX values are provided in Table 2.
We conducted a hyperparameter search over the following GAIL parameters: the GAIL reward coefficient, the number of epochs that the discriminator was trained for each IPPO update, the buffer size, and the batch size. The final selected values are given in Table 2.
Computing Architecture.
All IPPO, QMIX, DM2 experiments were performed without parallelized training; RMAPPO experiments were performed with parallelized training (as is the default in the RMAPPO codebase). The servers used in our experiments ran Ubuntu 18.04 with the following configurations:
-
•
Intel Xeon CPU E5-2698 v4; Nvidia Tesla V100-SXM2 GPU.
-
•
Intel Xeon CPU E5-2630 v4; Nvidia Titan V GPU.
-
•
Intel Xeon Gold 6342 CPU; Nvidia A40 GPU.
5v6 | 3sv4z | 3sv3z | |
epochs | 10 | 15 | 15 |
buffer size | 1024 | 1024 | 1024 |
gain | 0.01 | 0.01 | 0.01 |
clip | 0.05 | 0.2 | 0.2 |
qmix epsilon | 0 | 0 | 0.1 |
|