Learning a closed-loop policy for scooping task

Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy has...

Full description

Saved in:
Bibliographic Details
Main Author: Lau, Wei Quan
Other Authors: Ang Wei Tech
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/177105
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-177105
record_format dspace
spelling sg-ntu-dr.10356-1771052024-05-25T16:50:16Z Learning a closed-loop policy for scooping task Lau, Wei Quan Ang Wei Tech School of Mechanical and Aerospace Engineering Rehabilitation Research Institute of Singapore (RRIS) Neha Priyadarshini Garg J-Anne Yow WTAng@ntu.edu.sg Engineering Assistive feeding Scooping Behavioural cloning Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy has been developed at Rehabilitation Research Institute Singapore (RRIS) for food scooping. Existing assistive feeding systems including the one developed at RRIS mostly use an openloop policy for scooping. This makes it hard to control the amount of food to be scooped which is important as many times therapists have requirements for people with disabilities to have a specific bite-size. Having a closed-loop policy which can adapt the scooping trajectory based on the sensor feedback such as camera, force torque sensors can address this limitation. Developing a closed-loop policy for scooping task which involves dealing with deformable objects like food is not trivial as there is a lack of physics models, which makes classical planning approaches difficult. Hence, data-driven approaches are used which collect training data through demonstrations and learn a policy using that data. However, for learning a policy using data-driven approaches, data specific to a given robot and task needs to be collected. The data can be collected in simulation when a large amount of data is needed or on the real arm if a small amount of data is needed. Therefore, in this work, we aim to learn a closed-loop scooping policy and execute it on the xArm6 at RRIS. It is to noted that learning a closed-loop policy that can scoop a specified amount of food is out of scope of this project as it requires advanced Reinforcement Learning and will be explored in the future by RRIS researchers. The scope of this project is to develop a framework for learning a closed-loop policy and demonstrate that the framework works by learning a simple Behavioural Cloning based policy using that framework. This ensures that this framework can be used in future by RRIS researchers for developing more advanced closed-loop policies. We create the framework for both the simulated and real arm. For simulation, we survey various simulators and recommend one that is compatible with xArm6 and deformable object manipulation. From our experiments, we demonstrate that the policy learned in simulation can generate a scooping trajectory on simulation and the policy learned from the real arm can scoop the food up on the real arm. However, there is a sim-to-real gap when using the learnt policy from simulated data on the real arm. Bachelor's degree 2024-05-21T06:57:17Z 2024-05-21T06:57:17Z 2024 Final Year Project (FYP) Lau, W. Q. (2024). Learning a closed-loop policy for scooping task. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177105 https://hdl.handle.net/10356/177105 en C004 application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Assistive feeding
Scooping
Behavioural cloning
spellingShingle Engineering
Assistive feeding
Scooping
Behavioural cloning
Lau, Wei Quan
Learning a closed-loop policy for scooping task
description Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy has been developed at Rehabilitation Research Institute Singapore (RRIS) for food scooping. Existing assistive feeding systems including the one developed at RRIS mostly use an openloop policy for scooping. This makes it hard to control the amount of food to be scooped which is important as many times therapists have requirements for people with disabilities to have a specific bite-size. Having a closed-loop policy which can adapt the scooping trajectory based on the sensor feedback such as camera, force torque sensors can address this limitation. Developing a closed-loop policy for scooping task which involves dealing with deformable objects like food is not trivial as there is a lack of physics models, which makes classical planning approaches difficult. Hence, data-driven approaches are used which collect training data through demonstrations and learn a policy using that data. However, for learning a policy using data-driven approaches, data specific to a given robot and task needs to be collected. The data can be collected in simulation when a large amount of data is needed or on the real arm if a small amount of data is needed. Therefore, in this work, we aim to learn a closed-loop scooping policy and execute it on the xArm6 at RRIS. It is to noted that learning a closed-loop policy that can scoop a specified amount of food is out of scope of this project as it requires advanced Reinforcement Learning and will be explored in the future by RRIS researchers. The scope of this project is to develop a framework for learning a closed-loop policy and demonstrate that the framework works by learning a simple Behavioural Cloning based policy using that framework. This ensures that this framework can be used in future by RRIS researchers for developing more advanced closed-loop policies. We create the framework for both the simulated and real arm. For simulation, we survey various simulators and recommend one that is compatible with xArm6 and deformable object manipulation. From our experiments, we demonstrate that the policy learned in simulation can generate a scooping trajectory on simulation and the policy learned from the real arm can scoop the food up on the real arm. However, there is a sim-to-real gap when using the learnt policy from simulated data on the real arm.
author2 Ang Wei Tech
author_facet Ang Wei Tech
Lau, Wei Quan
format Final Year Project
author Lau, Wei Quan
author_sort Lau, Wei Quan
title Learning a closed-loop policy for scooping task
title_short Learning a closed-loop policy for scooping task
title_full Learning a closed-loop policy for scooping task
title_fullStr Learning a closed-loop policy for scooping task
title_full_unstemmed Learning a closed-loop policy for scooping task
title_sort learning a closed-loop policy for scooping task
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/177105
_version_ 1814047316434223104