Intelligent robot manipulation with deep learning

The application of reinforcement learning (RL) in robotics has seen significant advancements across various sectors, yet a critical challenge persists: the simulation-to-reality (sim-to-real) gap. In real-world scenarios, robots frequently underperform due to their inability to access or accurate...

全面介紹

Saved in:
書目詳細資料
主要作者: Tan, Jun Aun
其他作者: Lin Zhiping
格式: Final Year Project
語言:English
出版: Nanyang Technological University 2024
主題:
在線閱讀:https://hdl.handle.net/10356/176388
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
機構: Nanyang Technological University
語言: English
實物特徵
總結:The application of reinforcement learning (RL) in robotics has seen significant advancements across various sectors, yet a critical challenge persists: the simulation-to-reality (sim-to-real) gap. In real-world scenarios, robots frequently underperform due to their inability to access or accurately interpret all observable states, such as the precise state of objects in their environment. This gap often results in discrepancies between expected and actual robot behavior, hindering the effective translation of learned skills from simulated environments to practical applications. This paper presents an innovative approach to bridge this gap through a combination of reinforcement learning (RL) and imitation learning (IL). We introduce a novel teacher-student framework designed to enhance the performance of robotic systems. In this framework, the teacher possesses complete access to all states, including both the environment and the robot, whereas the student is limited to observing only the robot's state and visual input. By implementing an imitation loss in conjunction with the Proximal Policy Optimization (PPO) policy loss during the training of the student's policy, and leveraging expert knowledge transferred from the teacher, we demonstrate a significant improvement in the student's learning efficiency and performance. Our results reveal that the student model, when trained under this hybrid learning paradigm, converges more swiftly, and outperforms models trained solely on RL. The experiments were conducted across multiple environments, each selected for their unique challenges and the ability to test the algorithm's adaptability and performance in varied conditions. This study not only showcases the potential of combining RL with imitation learning to mitigate the knowledge gap but also establishes a foundational framework for future research in enhancing the adaptability and efficiency of robotic systems in real-world applications.