Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Cooperative Multi-Agent Reinforcement Learning With Approximate Model Learning

Full metadata record
DC Field Value Language
dc.contributor.authorPark, Young Joon-
dc.contributor.authorLee, Young Jae-
dc.contributor.authorKim, Seoung Bum-
dc.date.accessioned2021-08-31T16:05:30Z-
dc.date.available2021-08-31T16:05:30Z-
dc.date.created2021-06-19-
dc.date.issued2020-
dc.identifier.issn2169-3536-
dc.identifier.urihttps://scholar.korea.ac.kr/handle/2021.sw.korea/58989-
dc.description.abstractIn multi-agent reinforcement learning, it is essential for agents to learn communication protocol to optimize collaboration policies and to solve unstable learning problems. Existing methods based on actor-critic networks solve the communication problem among agents. However, these methods have difficulty in improving sample efficiency and learning robust policies because it is not easy to understand the dynamics and nonstationary of the environment as the policies of other agents change. We propose a method for learning cooperative policies in multi-agent environments by considering the communications among agents. The proposed method consists of recurrent neural network-based actor-critic networks and deterministic policy gradients to centrally train decentralized policies. The actor networks cause the agents to communicate using forward and backward paths and to determine subsequent actions. The critic network helps to train the actor networks by sending gradient signals to the actors according to their contribution to the global reward. To address issues with partial observability and unstable learning, we propose using auxiliary prediction networks to approximate state transitions and the reward function. We used multi-agent environments to demonstrate the usefulness and superiority of the proposed method by comparing it with existing multi-agent reinforcement learning methods, in terms of both learning efficiency and goal achievements in the test phase. The results demonstrate that the proposed method outperformed other alternatives.-
dc.languageEnglish-
dc.language.isoen-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.subjectDYNAMICS-
dc.subjectPERFORMANCE-
dc.subjectFRAMEWORK-
dc.titleCooperative Multi-Agent Reinforcement Learning With Approximate Model Learning-
dc.typeArticle-
dc.contributor.affiliatedAuthorKim, Seoung Bum-
dc.identifier.doi10.1109/ACCESS.2020.3007219-
dc.identifier.scopusid2-s2.0-85088690432-
dc.identifier.wosid000554543000001-
dc.identifier.bibliographicCitationIEEE ACCESS, v.8, pp.125389 - 125400-
dc.relation.isPartOfIEEE ACCESS-
dc.citation.titleIEEE ACCESS-
dc.citation.volume8-
dc.citation.startPage125389-
dc.citation.endPage125400-
dc.type.rimsART-
dc.type.docTypeArticle-
dc.description.journalClass1-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaTelecommunications-
dc.relation.journalWebOfScienceCategoryComputer Science, Information Systems-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryTelecommunications-
dc.subject.keywordPlusDYNAMICS-
dc.subject.keywordPlusPERFORMANCE-
dc.subject.keywordPlusFRAMEWORK-
dc.subject.keywordAuthorReinforcement learning-
dc.subject.keywordAuthormodel-free method-
dc.subject.keywordAuthormulti-agent system-
dc.subject.keywordAuthormulti-agent cooperation-
dc.subject.keywordAuthoractor-critic method-
dc.subject.keywordAuthordeterministic policy gradient-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > School of Industrial and Management Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher KIM, Seoung Bum photo

KIM, Seoung Bum
공과대학 (산업경영공학부)
Read more

Altmetrics

Total Views & Downloads

BROWSE