Adaptive behaviours for robotics

Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart re...

Full description

Saved in:
Bibliographic Details
Main Author: Huang, Jiqing.
Other Authors: Lin Zhiping
Format: Final Year Project
Language:English
Published: 2011
Subjects:
Online Access:http://hdl.handle.net/10356/46004
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-46004
record_format dspace
spelling sg-ntu-dr.10356-460042023-07-07T16:01:40Z Adaptive behaviours for robotics Huang, Jiqing. Lin Zhiping School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart reasoning. This is done by implementing the different clustering techniques, reinforcement learning and exploration strategies into the AI Engine which functions like the learning tool of the UGV. Comparisons between the different kinds of techniques and strategies were made using the first simulation. From the simulations, we can conclude that using Q-learning as our reinforcement theory and Boltzmann Exploration as our exploration strategy would obtain better results. It is also noted that using two vectors instead of three to represents the state, action and reward of a UGV would reduce the number of rules generated and hence reduce the time required for selection and learning at each time step. Lastly, updating of the Q-value at the end of the episode has also been proven to have a faster convergence of the Q-value. Both the second and third simulations were designed to test the AI Engine‟s practicability and capability to execute in a real time environment. We have obtained a near optimum result for the second simulation where the AI Engine learned a series of actions in order to obtain a reward. In the third simulation, the AI Engine managed to learn the desired action without going through a number of episodes. All of these simulations have been proven successful. Other kinds of artificial intelligence and data mining techniques could be implemented and compared in the future. Bachelor of Engineering 2011-06-27T07:14:07Z 2011-06-27T07:14:07Z 2011 2011 Final Year Project (FYP) http://hdl.handle.net/10356/46004 en Nanyang Technological University 58 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
spellingShingle DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics
Huang, Jiqing.
Adaptive behaviours for robotics
description Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart reasoning. This is done by implementing the different clustering techniques, reinforcement learning and exploration strategies into the AI Engine which functions like the learning tool of the UGV. Comparisons between the different kinds of techniques and strategies were made using the first simulation. From the simulations, we can conclude that using Q-learning as our reinforcement theory and Boltzmann Exploration as our exploration strategy would obtain better results. It is also noted that using two vectors instead of three to represents the state, action and reward of a UGV would reduce the number of rules generated and hence reduce the time required for selection and learning at each time step. Lastly, updating of the Q-value at the end of the episode has also been proven to have a faster convergence of the Q-value. Both the second and third simulations were designed to test the AI Engine‟s practicability and capability to execute in a real time environment. We have obtained a near optimum result for the second simulation where the AI Engine learned a series of actions in order to obtain a reward. In the third simulation, the AI Engine managed to learn the desired action without going through a number of episodes. All of these simulations have been proven successful. Other kinds of artificial intelligence and data mining techniques could be implemented and compared in the future.
author2 Lin Zhiping
author_facet Lin Zhiping
Huang, Jiqing.
format Final Year Project
author Huang, Jiqing.
author_sort Huang, Jiqing.
title Adaptive behaviours for robotics
title_short Adaptive behaviours for robotics
title_full Adaptive behaviours for robotics
title_fullStr Adaptive behaviours for robotics
title_full_unstemmed Adaptive behaviours for robotics
title_sort adaptive behaviours for robotics
publishDate 2011
url http://hdl.handle.net/10356/46004
_version_ 1772826522424442880