Adaptive behaviours for robotics
Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart re...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2011
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/46004 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-46004 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-460042023-07-07T16:01:40Z Adaptive behaviours for robotics Huang, Jiqing. Lin Zhiping School of Electrical and Electronic Engineering DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart reasoning. This is done by implementing the different clustering techniques, reinforcement learning and exploration strategies into the AI Engine which functions like the learning tool of the UGV. Comparisons between the different kinds of techniques and strategies were made using the first simulation. From the simulations, we can conclude that using Q-learning as our reinforcement theory and Boltzmann Exploration as our exploration strategy would obtain better results. It is also noted that using two vectors instead of three to represents the state, action and reward of a UGV would reduce the number of rules generated and hence reduce the time required for selection and learning at each time step. Lastly, updating of the Q-value at the end of the episode has also been proven to have a faster convergence of the Q-value. Both the second and third simulations were designed to test the AI Engine‟s practicability and capability to execute in a real time environment. We have obtained a near optimum result for the second simulation where the AI Engine learned a series of actions in order to obtain a reward. In the third simulation, the AI Engine managed to learn the desired action without going through a number of episodes. All of these simulations have been proven successful. Other kinds of artificial intelligence and data mining techniques could be implemented and compared in the future. Bachelor of Engineering 2011-06-27T07:14:07Z 2011-06-27T07:14:07Z 2011 2011 Final Year Project (FYP) http://hdl.handle.net/10356/46004 en Nanyang Technological University 58 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics |
spellingShingle |
DRNTU::Engineering::Electrical and electronic engineering::Control and instrumentation::Robotics Huang, Jiqing. Adaptive behaviours for robotics |
description |
Unmanned Ground Vehicles (UGVs) have been playing more and more important roles in both civilian and military world. Most of the current UGV uses an algorithmic approach which is predictable and works only in certain fixed situation. Our main focus is to develop an adaptive type of UGV with smart reasoning. This is done by implementing the different clustering techniques, reinforcement learning and exploration strategies into the AI Engine which functions like the learning tool of the UGV. Comparisons between the different kinds of techniques and strategies were made using the first simulation. From the simulations, we can conclude that using Q-learning as our reinforcement theory and Boltzmann Exploration as our exploration strategy would obtain better results. It is also noted that using two vectors instead of three to represents the state, action and reward of a UGV would reduce the number of rules generated and hence reduce the time required for selection and learning at each time step. Lastly, updating of the Q-value at the end of the episode has also been proven to have a faster convergence of the Q-value. Both the second and third simulations were designed to test the AI Engine‟s practicability and capability to execute in a real time environment. We have obtained a near optimum result for the second simulation where the AI Engine learned a series of actions in order to obtain a reward. In the third simulation, the AI Engine managed to learn the desired action without going through a number of episodes. All of these simulations have been proven successful. Other kinds of artificial intelligence and data mining techniques could be implemented and compared in the future. |
author2 |
Lin Zhiping |
author_facet |
Lin Zhiping Huang, Jiqing. |
format |
Final Year Project |
author |
Huang, Jiqing. |
author_sort |
Huang, Jiqing. |
title |
Adaptive behaviours for robotics |
title_short |
Adaptive behaviours for robotics |
title_full |
Adaptive behaviours for robotics |
title_fullStr |
Adaptive behaviours for robotics |
title_full_unstemmed |
Adaptive behaviours for robotics |
title_sort |
adaptive behaviours for robotics |
publishDate |
2011 |
url |
http://hdl.handle.net/10356/46004 |
_version_ |
1772826522424442880 |