Residential College | false |
Status | 已發表Published |
Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation | |
Yao, Liang1,2; Zhao, Baoliang2; Xu, Xin3; Wang, Ziwen2; Wong, Pak Kin1; Hu, Ying2 | |
2024-01 | |
Source Publication | IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS |
ABS Journal Level | 3 |
ISSN | 2168-2216 |
Volume | 54Issue:1Pages:156-169 |
Abstract | Offline reinforcement learning (ORL) has been getting increasing attention in robot learning, benefiting from its ability to avoid hazardous exploration and learn policies directly from precollected samples. Approximate policy iteration (API) is one of the most commonly investigated ORL approaches in robotics, due to its linear representation of policies, which makes it fairly transparent in both theoretical and engineering analysis. One open problem of API is how to design efficient and effective basis functions. The broad learning system (BLS) has been extensively studied in supervised and unsupervised learning in various applications. However, few investigations have been conducted on ORL. In this article, a novel incremental ORL approach with sparse broad critic approximation (BORL) is proposed with the advantages of BLS, which approximates the critic function in a linear manner with randomly projected sparse and compact features and dynamically expands its broad structure. The BORL is the first extension of API with BLS in the field of robotics and ORL. The approximation ability and convergence performance of BORL are also analyzed. Comprehensive simulation studies are then conducted on two benchmarks, and the results demonstrate that the proposed BORL can obtain comparable or better performance than conventional API methods without laborious hyperparameter fine-tuning work. To further demonstrate the effectiveness of BORL in practical robotic applications, a variable force tracking problem in robotic ultrasound scanning (RUSS) is investigated, and a learning-based adaptive impedance control (LAIC) algorithm is proposed based on BORL. The experimental results demonstrate the advantages of LAIC compared with conventional force tracking methods. |
Keyword | Broad Learning System (Bls) Incremental Critic Design Linear Critic Approximation (Lca) Offline Reinforcement Learning (Orl) Variable Force Tracking |
DOI | 10.1109/TSMC.2023.3305498 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Automation & Control Systems ; Computer Science |
WOS Subject | Automation & Control Systems ; Computer Science, Cybernetics |
WOS ID | WOS:001064568500001 |
Publisher | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC, 445 HOES LANE, PISCATAWAY, NJ 08855-4141 |
Scopus ID | 2-s2.0-85169708047 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | DEPARTMENT OF ELECTROMECHANICAL ENGINEERING |
Corresponding Author | Zhao, Baoliang; Wong, Pak Kin |
Affiliation | 1.Department of Electromechanical Engineering, University of Macau 2.Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences 3.National University of Defense Technology, China |
First Author Affilication | University of Macau |
Corresponding Author Affilication | University of Macau |
Recommended Citation GB/T 7714 | Yao, Liang,Zhao, Baoliang,Xu, Xin,et al. Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation[J]. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS, 2024, 54(1), 156-169. |
APA | Yao, Liang., Zhao, Baoliang., Xu, Xin., Wang, Ziwen., Wong, Pak Kin., & Hu, Ying (2024). Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation. IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS, 54(1), 156-169. |
MLA | Yao, Liang,et al."Efficient Incremental Offline Reinforcement Learning with Sparse Broad Critic Approximation".IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: SYSTEMS 54.1(2024):156-169. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment