Learning a closed-loop policy for scooping task
Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy has...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177105 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-177105 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1771052024-05-25T16:50:16Z Learning a closed-loop policy for scooping task Lau, Wei Quan Ang Wei Tech School of Mechanical and Aerospace Engineering Rehabilitation Research Institute of Singapore (RRIS) Neha Priyadarshini Garg J-Anne Yow WTAng@ntu.edu.sg Engineering Assistive feeding Scooping Behavioural cloning Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy has been developed at Rehabilitation Research Institute Singapore (RRIS) for food scooping. Existing assistive feeding systems including the one developed at RRIS mostly use an openloop policy for scooping. This makes it hard to control the amount of food to be scooped which is important as many times therapists have requirements for people with disabilities to have a specific bite-size. Having a closed-loop policy which can adapt the scooping trajectory based on the sensor feedback such as camera, force torque sensors can address this limitation. Developing a closed-loop policy for scooping task which involves dealing with deformable objects like food is not trivial as there is a lack of physics models, which makes classical planning approaches difficult. Hence, data-driven approaches are used which collect training data through demonstrations and learn a policy using that data. However, for learning a policy using data-driven approaches, data specific to a given robot and task needs to be collected. The data can be collected in simulation when a large amount of data is needed or on the real arm if a small amount of data is needed. Therefore, in this work, we aim to learn a closed-loop scooping policy and execute it on the xArm6 at RRIS. It is to noted that learning a closed-loop policy that can scoop a specified amount of food is out of scope of this project as it requires advanced Reinforcement Learning and will be explored in the future by RRIS researchers. The scope of this project is to develop a framework for learning a closed-loop policy and demonstrate that the framework works by learning a simple Behavioural Cloning based policy using that framework. This ensures that this framework can be used in future by RRIS researchers for developing more advanced closed-loop policies. We create the framework for both the simulated and real arm. For simulation, we survey various simulators and recommend one that is compatible with xArm6 and deformable object manipulation. From our experiments, we demonstrate that the policy learned in simulation can generate a scooping trajectory on simulation and the policy learned from the real arm can scoop the food up on the real arm. However, there is a sim-to-real gap when using the learnt policy from simulated data on the real arm. Bachelor's degree 2024-05-21T06:57:17Z 2024-05-21T06:57:17Z 2024 Final Year Project (FYP) Lau, W. Q. (2024). Learning a closed-loop policy for scooping task. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177105 https://hdl.handle.net/10356/177105 en C004 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Assistive feeding Scooping Behavioural cloning |
spellingShingle |
Engineering Assistive feeding Scooping Behavioural cloning Lau, Wei Quan Learning a closed-loop policy for scooping task |
description |
Assistive feeding technologies can play a crucial role in improving the quality of life for individuals with disabilities or age-related limitations who struggle with self-feeding. Scooping task is
important to enable assistive feeding. Therefore, a Dynamic Motion Primitives (DMP) based policy
has been developed at Rehabilitation Research Institute Singapore (RRIS) for food scooping.
Existing assistive feeding systems including the one developed at RRIS mostly use an openloop policy for scooping. This makes it hard to control the amount of food to be scooped which is
important as many times therapists have requirements for people with disabilities to have a specific
bite-size. Having a closed-loop policy which can adapt the scooping trajectory based on the sensor
feedback such as camera, force torque sensors can address this limitation.
Developing a closed-loop policy for scooping task which involves dealing with deformable objects like food is not trivial as there is a lack of physics models, which makes classical planning
approaches difficult. Hence, data-driven approaches are used which collect training data through
demonstrations and learn a policy using that data. However, for learning a policy using data-driven
approaches, data specific to a given robot and task needs to be collected. The data can be collected
in simulation when a large amount of data is needed or on the real arm if a small amount of data
is needed. Therefore, in this work, we aim to learn a closed-loop scooping policy and execute it on
the xArm6 at RRIS.
It is to noted that learning a closed-loop policy that can scoop a specified amount of food is
out of scope of this project as it requires advanced Reinforcement Learning and will be explored in
the future by RRIS researchers. The scope of this project is to develop a framework for learning
a closed-loop policy and demonstrate that the framework works by learning a simple Behavioural
Cloning based policy using that framework. This ensures that this framework can be used in future
by RRIS researchers for developing more advanced closed-loop policies. We create the framework
for both the simulated and real arm. For simulation, we survey various simulators and recommend
one that is compatible with xArm6 and deformable object manipulation. From our experiments, we
demonstrate that the policy learned in simulation can generate a scooping trajectory on simulation
and the policy learned from the real arm can scoop the food up on the real arm. However, there is
a sim-to-real gap when using the learnt policy from simulated data on the real arm. |
author2 |
Ang Wei Tech |
author_facet |
Ang Wei Tech Lau, Wei Quan |
format |
Final Year Project |
author |
Lau, Wei Quan |
author_sort |
Lau, Wei Quan |
title |
Learning a closed-loop policy for scooping task |
title_short |
Learning a closed-loop policy for scooping task |
title_full |
Learning a closed-loop policy for scooping task |
title_fullStr |
Learning a closed-loop policy for scooping task |
title_full_unstemmed |
Learning a closed-loop policy for scooping task |
title_sort |
learning a closed-loop policy for scooping task |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/177105 |
_version_ |
1814047316434223104 |