Imitation learning from demonstration videos

Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control...

Full description

Saved in:
Bibliographic Details
Main Author: Zeng, Jingbo
Other Authors: Tan Yap Peng
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2024
Subjects:
Online Access:https://hdl.handle.net/10356/177091
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-177091
record_format dspace
spelling sg-ntu-dr.10356-1770912024-05-24T15:56:09Z Imitation learning from demonstration videos Zeng, Jingbo Tan Yap Peng School of Electrical and Electronic Engineering EYPTan@ntu.edu.sg Engineering Imitation learning Hand pose estimation Dexterous manipulation Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control rules remains relatively unexplored. In this dissertation, we introduce an imitation learning method for encoding prior knowledge from dexterous manipulation demonstra- tion videos. Instead of adopting behavior cloning or pure RL algorithms, our model considers two online RL algorithms: 1) Demo Augmented Policy Gra- dients (DAPG) and 2) Generative Adversarial Imitation Learning (GAIL). With the requirements for encoding the finger action in demonstrations, we selected MANO as the baseline of hand pose estimation, and designed a SuperPoint- based module to optimize detection results. Quantitative experimental results show that our framework can exploit hand pose estimation on different dataset effectively and use imitation learning to achieve great overall performance on three defined tasks. Moreover, it has good generalization ability when deployed on unseen objects. Some visual results show that the proposed framework can be applied combining with prior knowledge from demonstration videos, which provides a possible solution for robot’s imitating human behaviors. Master's degree 2024-05-22T23:48:05Z 2024-05-22T23:48:05Z 2024 Thesis-Master by Coursework Zeng, J. (2024). Imitation learning from demonstration videos. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/177091 https://hdl.handle.net/10356/177091 en application/pdf Nanyang Technological University
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering
Imitation learning
Hand pose estimation
Dexterous manipulation
spellingShingle Engineering
Imitation learning
Hand pose estimation
Dexterous manipulation
Zeng, Jingbo
Imitation learning from demonstration videos
description Imitation learning is a challenging and meaningful task to encode prior knowl- edge to provide a motion control policy for guiding robot movement and trajec- tory autonomously to complete specified assignment with a given current state. However, effective translation from prior knowledge to control rules remains relatively unexplored. In this dissertation, we introduce an imitation learning method for encoding prior knowledge from dexterous manipulation demonstra- tion videos. Instead of adopting behavior cloning or pure RL algorithms, our model considers two online RL algorithms: 1) Demo Augmented Policy Gra- dients (DAPG) and 2) Generative Adversarial Imitation Learning (GAIL). With the requirements for encoding the finger action in demonstrations, we selected MANO as the baseline of hand pose estimation, and designed a SuperPoint- based module to optimize detection results. Quantitative experimental results show that our framework can exploit hand pose estimation on different dataset effectively and use imitation learning to achieve great overall performance on three defined tasks. Moreover, it has good generalization ability when deployed on unseen objects. Some visual results show that the proposed framework can be applied combining with prior knowledge from demonstration videos, which provides a possible solution for robot’s imitating human behaviors.
author2 Tan Yap Peng
author_facet Tan Yap Peng
Zeng, Jingbo
format Thesis-Master by Coursework
author Zeng, Jingbo
author_sort Zeng, Jingbo
title Imitation learning from demonstration videos
title_short Imitation learning from demonstration videos
title_full Imitation learning from demonstration videos
title_fullStr Imitation learning from demonstration videos
title_full_unstemmed Imitation learning from demonstration videos
title_sort imitation learning from demonstration videos
publisher Nanyang Technological University
publishDate 2024
url https://hdl.handle.net/10356/177091
_version_ 1800916425825583104