UM  > Faculty of Science and Technology  > DEPARTMENT OF ELECTROMECHANICAL ENGINEERING
Residential Collegefalse
Status已發表Published
Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation
Yao, Liang1,2; Zhao, Baoliang2; Xu, Xin3; Wang, Ziwen2; Wong, Pak Kin1; Hu, Ying2
2024-01
Source PublicationIEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS
ABS Journal Level3
ISSN2168-2216
Volume54Issue:1Pages:156-169
Abstract

Offline reinforcement learning (ORL) has been getting increasing attention in robot learning, benefiting from its ability to avoid hazardous exploration and learn policies directly from precollected samples. Approximate policy iteration (API) is one of the most commonly investigated ORL approaches in robotics, due to its linear representation of policies, which makes it fairly transparent in both theoretical and engineering analysis. One open problem of API is how to design efficient and effective basis functions. The broad learning system (BLS) has been extensively studied in supervised and unsupervised learning in various applications. However, few investigations have been conducted on ORL. In this article, a novel incremental ORL approach with sparse broad critic approximation (BORL) is proposed with the advantages of BLS, which approximates the critic function in a linear manner with randomly projected sparse and compact features and dynamically expands its broad structure. The BORL is the first extension of API with BLS in the field of robotics and ORL. The approximation ability and convergence performance of BORL are also analyzed. Comprehensive simulation studies are then conducted on two benchmarks, and the results demonstrate that the proposed BORL can obtain comparable or better performance than conventional API methods without laborious hyperparameter fine-tuning work. To further demonstrate the effectiveness of BORL in practical robotic applications, a variable force tracking problem in robotic ultrasound scanning (RUSS) is investigated, and a learning-based adaptive impedance control (LAIC) algorithm is proposed based on BORL. The experimental results demonstrate the advantages of LAIC compared with conventional force tracking methods.

KeywordBroad Learning System (Bls) Incremental Critic Design Linear Critic Approximation (Lca) Offline Reinforcement Learning (Orl) Variable Force Tracking
DOI10.1109/TSMC.2023.3305498
URLView the original
Indexed BySCIE
Language英語English
WOS Research AreaAutomation & Control Systems ; Computer Science
WOS SubjectAutomation & Control Systems ; Computer Science, Cybernetics
WOS IDWOS:001064568500001
PublisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141
Scopus ID2-s2.0-85169708047
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionDEPARTMENT OF ELECTROMECHANICAL ENGINEERING
Corresponding AuthorZhao, Baoliang; Wong, Pak Kin
Affiliation1.Department of Electromechanical Engineering, University of Macau
2.Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences
3.National University of Defense Technology, China
First Author AffilicationUniversity of Macau
Corresponding Author AffilicationUniversity of Macau
Recommended Citation
GB/T 7714
Yao, Liang,Zhao, Baoliang,Xu, Xin,et al. Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation[J]. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS, 2024, 54(1), 156-169.
APA Yao, Liang., Zhao, Baoliang., Xu, Xin., Wang, Ziwen., Wong, Pak Kin., & Hu, Ying (2024). Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS, 54(1), 156-169.
MLA Yao, Liang,et al."Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation".IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS 54.1(2024):156-169.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Yao, Liang]'s Articles
[Zhao, Baoliang]'s Articles
[Xu, Xin]'s Articles
Baidu academic
Similar articles in Baidu academic
[Yao, Liang]'s Articles
[Zhao, Baoliang]'s Articles
[Xu, Xin]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Yao, Liang]'s Articles
[Zhao, Baoliang]'s Articles
[Xu, Xin]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.