UM  > Faculty of Science and Technology
Residential Collegefalse
Status已發表Published
TieComm: Learning a Hierarchical Communication Topology Based on Tie Theory
Yang,Ming1; Dong,Renzhi1; Wang,Yiming1; Liu,Furui4; Du,Yali5; Zhou,Mingliang6; U, Leong Hou1,2,3
2023
Conference NameInternational Conference on Database Systems for Advanced Applications 2022
Source PublicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13943 LNCS
Pages604-613
Conference Date2023 April 17-20
Conference PlaceTianjin, China
PublisherSpringer Science and Business Media Deutschland GmbH
Abstract

Communication plays an important role in Internet of Things that assists cooperation between devices for better resource management. This work considers the problem of learning cooperative policies using communications in Multi-Agent Reinforcement Learning (MARL), which plays an important role to stabilize agent training and improve the policy learned by enabling the agent to capture more information in partially observable environments. Existing studies either adopt a prior topology by experts or learn a communication topology through a costly process. In this work, we optimize the communication mechanism by exploiting both local agent communications and distant agent communications. Our solution is motivated by tie theory in social networks, where strong ties (close friends) communicate differently with weak ties (distant friends). The proposed novel multi-agent reinforcement learning framework named TieComm, learns a dynamic communication topology consisting of inter- and intra-group communication for efficient policy learning. We factorize the joint multi-agent policy into a centralized tie reasoning policy and decentralized conditional action policies of agents, based on which we propose an alternative updating schema to achieve efficient optimization. Experimental results on Level-Based Foraging and Blind-particle Spread demonstrate the effectiveness of our tie theory based RL framework.

KeywordCommunication Topology Cooperation Multi-agent System Reinforcement Learning Social Welfare
DOI10.1007/978-3-031-30637-2_40
URLView the original
Language英語English
Scopus ID2-s2.0-85161704159
Fulltext Access
Citation statistics
Document TypeConference paper
CollectionFaculty of Science and Technology
Corresponding AuthorU, Leong Hou
Affiliation1.SKL of Internet of Things for Smart City,University of Macau,Zhuhai,China
2.Department of Computer Information Science,University of Macau,Zhuhai,China
3.Centre for Data Science,University of Macau,Zhuhai,China
4.Zhejiang Lab,Hangzhou,China
5.King’s College London,London,United Kingdom
6.College of Computer Science,Chongqing University,Chongqing,China
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Yang,Ming,Dong,Renzhi,Wang,Yiming,et al. TieComm: Learning a Hierarchical Communication Topology Based on Tie Theory[C]:Springer Science and Business Media Deutschland GmbH, 2023, 604-613.
APA Yang,Ming., Dong,Renzhi., Wang,Yiming., Liu,Furui., Du,Yali., Zhou,Mingliang., & U, Leong Hou (2023). TieComm: Learning a Hierarchical Communication Topology Based on Tie Theory. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 13943 LNCS, 604-613.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Yang,Ming]'s Articles
[Dong,Renzhi]'s Articles
[Wang,Yiming]'s Articles
Baidu academic
Similar articles in Baidu academic
[Yang,Ming]'s Articles
[Dong,Renzhi]'s Articles
[Wang,Yiming]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Yang,Ming]'s Articles
[Dong,Renzhi]'s Articles
[Wang,Yiming]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.