HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Reducing Overestimation Bias in Multi-Agent Domains Using Double Centralized Critics

Johannes Ackermann Volker Gabler Takayuki Osa Masashi Sugiyama

Reducing Overestimation Bias in Multi-Agent Domains Using Double Centralized Critics

Abstract

Many real world tasks require multiple agents to work together. Multi-agent reinforcement learning (RL) methods have been proposed in recent years to solve these tasks, but current methods often fail to efficiently learn policies. We thus investigate the presence of a common weakness in single-agent RL, namely value function overestimation bias, in the multi-agent setting. Based on our findings, we propose an approach that reduces this bias by using double centralized critics. We evaluate it on six mixed cooperative-competitive tasks, showing a significant advantage over current methods. Finally, we investigate the application of multi-agent methods to high-dimensional robotic tasks and show that our approach can be used to learn decentralized policies in this domain.

Code Repositories

fdcl-gwu/gym-rotor
pytorch
Mentioned in GitHub
JohannesAck/MATD3implementation
Official
tf
Mentioned in GitHub
JohannesAck/tf2multiagentrl
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
multi-agent-reinforcement-learning-onMATD3
final agent reward: -14

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Reducing Overestimation Bias in Multi-Agent Domains Using Double Centralized Critics | Papers | HyperAI