Title:
Cooperation in Multi-Agent Reinforcement Learning

dc.contributor.advisor Zha, Hongyuan
dc.contributor.advisor Zhao, Tuo
dc.contributor.author Yang, Jiachen
dc.contributor.committeeMember Isbell, Charles L.
dc.contributor.committeeMember Gombolay, Matthew
dc.contributor.committeeMember Faissol, Daniel M.
dc.contributor.department Computational Science and Engineering
dc.date.accessioned 2022-01-14T16:11:34Z
dc.date.available 2022-01-14T16:11:34Z
dc.date.created 2021-12
dc.date.issued 2021-12-13
dc.date.submitted December 2021
dc.date.updated 2022-01-14T16:11:34Z
dc.description.abstract As progress in reinforcement learning (RL) gives rise to increasingly general and powerful artificial intelligence, society needs to anticipate a possible future in which multiple RL agents must learn and interact in a shared multi-agent environment. When a single principal has oversight of the multi-agent system, how should agents learn to cooperate via centralized training to achieve individual and global objectives? When agents belong to self-interested principals with imperfectly-aligned objectives, how can cooperation emerge from fully-decentralized learning? This dissertation addresses both questions by proposing novel methods for multi-agent reinforcement learning (MARL) and demonstrating the empirical effectiveness of these methods in high-dimensional simulated environments. To address the first case, we propose new algorithms for fully-cooperative MARL in the paradigm of centralized training with decentralized execution. Firstly, we propose a method based on multi-agent curriculum learning and multi-agent credit assignment to address the setting where global optimality is defined as the attainment of all individual goals. Secondly, we propose a hierarchical MARL algorithm to discover and learn interpretable and useful skills for a multi-agent team to optimize a single team objective. Extensive experiments with ablations show the strengths of our approaches over state-of-the-art baselines. To address the second case, we propose learning algorithms to attain cooperation within a population of self-interested RL agents. We propose the design of a new agent who is equipped with the new ability to incentivize other RL agents and explicitly account for the other agents' learning process. This agent overcomes the challenging limitation of fully-decentralized training and generates emergent cooperation in difficult social dilemmas. Then, we extend and apply this technique to the problem of incentive design, where a central incentive designer explicitly optimizes a global objective only by intervening on the rewards of a population of independent RL agents. Experiments on the problem of optimal taxation in a simulated market economy demonstrate the effectiveness of this approach.
dc.description.degree Ph.D.
dc.format.mimetype application/pdf
dc.identifier.uri http://hdl.handle.net/1853/66146
dc.language.iso en_US
dc.publisher Georgia Institute of Technology
dc.subject multi-agent reinforcement learning
dc.subject multi-agent systems
dc.subject reinforcement learning
dc.subject deep reinforcement learning
dc.title Cooperation in Multi-Agent Reinforcement Learning
dc.type Text
dc.type.genre Dissertation
dspace.entity.type Publication
local.contributor.advisor Zhao, Tuo
local.contributor.corporatename College of Computing
local.contributor.corporatename School of Computational Science and Engineering
relation.isAdvisorOfPublication f1ca0ec4-da94-4fab-8a2f-3e6ae1f9b90b
relation.isOrgUnitOfPublication c8892b3c-8db6-4b7b-a33a-1b67f7db2021
relation.isOrgUnitOfPublication 01ab2ef1-c6da-49c9-be98-fbd1d840d2b1
thesis.degree.level Doctoral
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
YANG-DISSERTATION-2021.pdf
Size:
13.6 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
LICENSE.txt
Size:
3.86 KB
Format:
Plain Text
Description: