Modelling human behaviour for shared control wheelchair

Human Robot Interaction Algorithms that take into account input from humans can be challenging to test since it is expensive, time–consuming and difficult to recruit human subjects. This makes fine tuning these algorithms difficult, as it requires several rounds of testing. Automating the generat...

Full description

Saved in:
Bibliographic Details
Main Author: Kabilan, Anbukani
Other Authors: Ang Wei Tech
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2022
Subjects:
Online Access:https://hdl.handle.net/10356/157596
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:Human Robot Interaction Algorithms that take into account input from humans can be challenging to test since it is expensive, time–consuming and difficult to recruit human subjects. This makes fine tuning these algorithms difficult, as it requires several rounds of testing. Automating the generation of human input can help in significantly faster testing of such algorithms. The most common way of automating the generation of human input is through imitation learning. Using human input data from a few trials, we can train a model which can provide human actions for a given situation. Automatic generation of human input is an extremely broad research area with the ultimate aim of creating a Digital Human Twin which can be used for testing various algorithms used for chatbots, healthcare devices etc. In this work we focus on assistive robotic wheelchair application. Using the data collected during human trials at Rehabilitation Research Institute of Singapore (RRIS), where human provides joystick input to control the wheelchair with a shared control algorithm, a model can be learned which can generate human joystick input given the state and the surroundings of the wheelchair. Such a model can then be used to automate the testing of algorithms for shared control of the wheelchair. Since different humans have different ways of controlling the wheelchair depending on their preferences, ability etc., there are variations in the data collected. To enable proper testing, our learned model should be able to generate these variations. Thus, the scope of this project was to develop an Imitation Learning model that can discover and disentangle the latent factors of variations in the demonstrations. In this work, we use Information Maximizing Generative Adversarial Imitation Learning (InfoGAIL), since it considers variations present in the data. We process the data collected in Rosbags and convert it into a format that can be used for training an InfoGAIL model. For training and evaluation of the InfoGAIL model, we integrate the neural network with the shared control system implemented in gazebo simulator using ROS framework. After training the model, we generate the variations and verify that they correspond to the variations present in the data. We also discover the limitations of learning an InfoGAIL model when the trajectories are long and discuss ways to overcome this limitation.