Object grasping of humanoid robot based on YOLO

This paper presents a system that aims to achieve autonomous grasping for micro-controller based humanoid robots such as the Inmoov robot [1]. The system consists of a visual sensor, a central controller and a manipulator. We modify the open sourced objection detection software YOLO (You Only Look O...

Full description

Saved in:
Bibliographic Details
Main Authors: Tian, Li, Thalmann, Nadia Magnenat, Thalmann, Daniel, Fang, Zhiwen, Zheng, Jianmin
Other Authors: School of Computer Science and Engineering
Format: Conference or Workshop Item
Language:English
Published: 2020
Subjects:
Online Access:https://hdl.handle.net/10356/138923
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-138923
record_format dspace
spelling sg-ntu-dr.10356-1389232020-09-26T21:53:27Z Object grasping of humanoid robot based on YOLO Tian, Li Thalmann, Nadia Magnenat Thalmann, Daniel Fang, Zhiwen Zheng, Jianmin School of Computer Science and Engineering 36th Computer Graphics International Conference Institute for Media Innovation (IMI) Engineering::Computer science and engineering Robotics Vision This paper presents a system that aims to achieve autonomous grasping for micro-controller based humanoid robots such as the Inmoov robot [1]. The system consists of a visual sensor, a central controller and a manipulator. We modify the open sourced objection detection software YOLO (You Only Look Once) v2 [2] and associate it with the visual sensor to make the sensor be able to detect not only the category of the target object but also the location with the help of a depth camera. We also estimate the dimensions (i.e., the height and width) of the target based on the bounding box technique (Fig. 1). After that, we send the information to the central controller (a humanoid robot), which controls the manipulator (customised robotic hand) to grasp the object with the help of inverse kinematics theory. We conduct experiments to test our method with the Inmoov robot. The experiments show that our method is capable of detecting the object and driving the robotic hands to grasp the target object. NRF (Natl Research Foundation, S’pore) Accepted version 2020-05-14T02:54:47Z 2020-05-14T02:54:47Z 2019 Conference Paper Tian, L., Thalmann, N. M., Thalmann, D., Fang, Z., & Zheng, J. (2019). Object grasping of humanoid robot based on YOLO. 36th Computer Graphics International Conference, 476-482. doi:10.1007/978-3-030-22514-8_47 9783030225131 https://hdl.handle.net/10356/138923 10.1007/978-3-030-22514-8_47 2-s2.0-85067642536 476 482 en © 2019 Springer Nature Switzerland AG. All rights reserved. This paper was published in 36th Computer Graphics International Conference and is made available with permission of Springer Nature Switzerland AG. application/pdf
institution Nanyang Technological University
building NTU Library
country Singapore
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Robotics
Vision
spellingShingle Engineering::Computer science and engineering
Robotics
Vision
Tian, Li
Thalmann, Nadia Magnenat
Thalmann, Daniel
Fang, Zhiwen
Zheng, Jianmin
Object grasping of humanoid robot based on YOLO
description This paper presents a system that aims to achieve autonomous grasping for micro-controller based humanoid robots such as the Inmoov robot [1]. The system consists of a visual sensor, a central controller and a manipulator. We modify the open sourced objection detection software YOLO (You Only Look Once) v2 [2] and associate it with the visual sensor to make the sensor be able to detect not only the category of the target object but also the location with the help of a depth camera. We also estimate the dimensions (i.e., the height and width) of the target based on the bounding box technique (Fig. 1). After that, we send the information to the central controller (a humanoid robot), which controls the manipulator (customised robotic hand) to grasp the object with the help of inverse kinematics theory. We conduct experiments to test our method with the Inmoov robot. The experiments show that our method is capable of detecting the object and driving the robotic hands to grasp the target object.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Tian, Li
Thalmann, Nadia Magnenat
Thalmann, Daniel
Fang, Zhiwen
Zheng, Jianmin
format Conference or Workshop Item
author Tian, Li
Thalmann, Nadia Magnenat
Thalmann, Daniel
Fang, Zhiwen
Zheng, Jianmin
author_sort Tian, Li
title Object grasping of humanoid robot based on YOLO
title_short Object grasping of humanoid robot based on YOLO
title_full Object grasping of humanoid robot based on YOLO
title_fullStr Object grasping of humanoid robot based on YOLO
title_full_unstemmed Object grasping of humanoid robot based on YOLO
title_sort object grasping of humanoid robot based on yolo
publishDate 2020
url https://hdl.handle.net/10356/138923
_version_ 1681059481311510528