Imitation learning from demonstration videos
Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Thesis-Master by Coursework |
Language: | English |
Published: |
Nanyang Technological University
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/177091 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-177091 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1770912024-05-24T15:56:09Z Imitation learning from demonstration videos Zeng, Jingbo Tan Yap Peng School of Electrical and Electronic Engineering EYPTan@ntu.edu.sg Engineering Imitation learning Hand pose estimation Dexterous manipulation Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control rules remains relatively unexplored. In this dissertation, we introduce an imitation learning method for encoding prior knowledge from dexterous manipulation demonstra- tion videos. Instead of adopting behavior cloning or pure RL algorithms, our model considers two online RL algorithms: 1) Demo Augmented Policy Gra- dients (DAPG) and 2) Generative Adversarial Imitation Learning (GAIL). With the requirements for encoding the finger action in demonstrations, we selected MANO as the baseline of hand pose estimation, and designed a SuperPoint- based module to optimize detection results. Quantitative experimental results show that our framework can exploit hand pose estimation on different dataset effectively and use imitation learning to achieve great overall performance on three defined tasks. Moreover, it has good generalization ability when deployed on unseen objects. Some visual results show that the proposed framework can be applied combining with prior knowledge from demonstration videos, which provides a possible solution for robot’s imitating human behaviors. Master's degree 2024-05-22T23:48:05Z 2024-05-22T23:48:05Z 2024 Thesis-Master by Coursework Zeng, J. (2024). Imitation learning from demonstration videos. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177091 https://hdl.handle.net/10356/177091 en application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering Imitation learning Hand pose estimation Dexterous manipulation |
spellingShingle |
Engineering Imitation learning Hand pose estimation Dexterous manipulation Zeng, Jingbo Imitation learning from demonstration videos |
description |
Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control rules remains relatively unexplored. In this dissertation, we introduce an imitation learning method for encoding prior knowledge from dexterous manipulation demonstra- tion videos. Instead of adopting behavior cloning or pure RL algorithms, our model considers two online RL algorithms: 1) Demo Augmented Policy Gra- dients (DAPG) and 2) Generative Adversarial Imitation Learning (GAIL). With the requirements for encoding the finger action in demonstrations, we selected MANO as the baseline of hand pose estimation, and designed a SuperPoint- based module to optimize detection results. Quantitative experimental results show that our framework can exploit hand pose estimation on different dataset effectively and use imitation learning to achieve great overall performance on three defined tasks. Moreover, it has good generalization ability when deployed on unseen objects. Some visual results show that the proposed framework can be applied combining with prior knowledge from demonstration videos, which provides a possible solution for robot’s imitating human behaviors. |
author2 |
Tan Yap Peng |
author_facet |
Tan Yap Peng Zeng, Jingbo |
format |
Thesis-Master by Coursework |
author |
Zeng, Jingbo |
author_sort |
Zeng, Jingbo |
title |
Imitation learning from demonstration videos |
title_short |
Imitation learning from demonstration videos |
title_full |
Imitation learning from demonstration videos |
title_fullStr |
Imitation learning from demonstration videos |
title_full_unstemmed |
Imitation learning from demonstration videos |
title_sort |
imitation learning from demonstration videos |
publisher |
Nanyang Technological University |
publishDate |
2024 |
url |
https://hdl.handle.net/10356/177091 |
_version_ |
1800916425825583104 |