Xueguang Lyu: Understanding Centralized Critics

  Рет қаралды 193

Multi-Agent Learning Seminar

Multi-Agent Learning Seminar

2 ай бұрын

Abstract: Centralized Training for Decentralized Execution, where agents are trained offline in a centralized fashion and execute online in a decentralized manner, has become a popular approach in Multi-Agent Reinforcement Learning (MARL). In particular, it has become popular to develop actor-critic methods that train decentralized actors with a centralized critic where the centralized critic is allowed access global information of the entire system, including the true system state. Such centralized critics are possible given offline information and are not used for online execution. While these methods perform well in a number of domains and have become a de facto standard in MARL, using a centralized critic in this context has yet to be sufficiently understood theoretically or empirically. I will present recent works where we formally analyze centralized and decentralized critic approaches, and analyze the effect of using state-based critics in partially observable environments. We derive theories contrary to the common intuition: critic centralization is not strictly beneficial, and using state values can be harmful. We also prove that state-based critics, in particular, can introduce unexpected bias and variance compared to history-based critics. In addition, I will mention how the theory applies in practice by comparing different forms of critics on a wide range of common multi-agent benchmarks, where we surface practical issues such as the difficulty of representation learning with partial observability, and highlight why the theoretical problems are often overlooked in the literature.
Bio: Xueguang Lyu is a Ph.D. candidate in Computer Science at Northeastern University's Khoury College of Computer Sciences, advised by Christopher Amato. He completed his B.S. in Informatics from the University of California, Irvine. He also spent time at Amazon Robotics working on simulation-free on-robot reinforcement learning. Xueguang’s research interests lie in artificial intelligence, reinforcement learning and multi-agent systems, with a particular interest in multi-agent exploration and cooperation in theoretical and real-world settings, seeking to develop algorithms that allow autonomous agents to learn to cooperate (efficiently and optimally) from their experience.

Пікірлер
InstaDeep MARL Team: Growing the MARL software ecosystem in JAX
46:02
Multi-Agent Learning Seminar
Рет қаралды 152
Marc Lanctot:  Evaluating Agents using Social Choice Theory
57:59
Multi-Agent Learning Seminar
Рет қаралды 157
Они убрались очень быстро!
00:40
Аришнев
Рет қаралды 2,1 МЛН
Miracle Doctor Saves Blind Girl ❤️
00:59
Alan Chikin Chow
Рет қаралды 69 МЛН
когда достали одноклассники!
00:49
БРУНО
Рет қаралды 4,2 МЛН
Unleashing WebAssembly in Kubernetes with Kwasm
15:16
DevOps Toolkit
Рет қаралды 801
Tom Zahavy: Diversifying AI  Towards Creative Chess with AlphaZero
54:02
Multi-Agent Learning Seminar
Рет қаралды 141
Andrei Lupu and Brandon Cui: Adversarial Diversity in Multi Agent Coordination
51:56
Multi-Agent Learning Seminar
Рет қаралды 383
Yi Wu: Diversity-Driven Reinforcement Learning
47:42
Multi-Agent Learning Seminar
Рет қаралды 107
Niklas Lauffer: Who Needs to Know - Minimal Knowledge for Optimal Coordination
57:21
Multi-Agent Learning Seminar
Рет қаралды 197
Siqi Liu:  NfgTransformer  Equivariant Representation Learning of Normal Form Games
47:30
Robert Lange: Discovering Black Box Optimizers via Evolutionary Meta Learning
56:05
Multi-Agent Learning Seminar
Рет қаралды 196
Карточка Зарядка 📱 ( @ArshSoni )
0:23
EpicShortsRussia
Рет қаралды 714 М.
iPhone 15 Unboxing Paper diy
0:57
Cute Fay
Рет қаралды 2,6 МЛН
ВСЕ МОИ ТЕЛЕФОНЫ
14:31
DimaViper Live
Рет қаралды 40 М.
Best Beast Sounds Handsfree For Multi Phone
0:42
MUN HD
Рет қаралды 339 М.
📱 SAMSUNG, ЧТО С ЛИЦОМ? 🤡
0:46
Яблочный Маньяк
Рет қаралды 2 МЛН
Iphone or nokia
0:15
rishton vines😇
Рет қаралды 1,3 МЛН