Adaptive goal selection for agents in dynamic environments
In psychology, goal-setting theory, which has been studied by psychologists for over 35 years, reveals that goals play significant roles in incentive, action and performance for human beings. Based on this theory, a goal net model has been proposed to design intelligent agents that can be viewed as...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2013
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/96478 http://hdl.handle.net/10220/18110 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-96478 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-964782020-05-28T07:17:16Z Adaptive goal selection for agents in dynamic environments Zhang, Huiliang Luo, Xudong Shen, Zhiqi You, Jin Miao, Chun Yan School of Computer Engineering School of Electrical and Electronic Engineering DRNTU::Engineering::Computer science and engineering In psychology, goal-setting theory, which has been studied by psychologists for over 35 years, reveals that goals play significant roles in incentive, action and performance for human beings. Based on this theory, a goal net model has been proposed to design intelligent agents that can be viewed as a soft copy of human being somehow. The goal net model has been successfully applied in many agents, specially, non-player-character agents in computer games. Such an agent selects the optimal solution in all possible solutions found by using a recursive algorithm. However, if a goal net is very complex, the time of selection could be too long for the agent to respond quickly when the agent needs to re-select a new solution against the world’s change. Moreover, in some dynamic environments, it is impossible to know the exact outcome of choosing a solution in advance, and so the possible solutions cannot be evaluated precisely. Thus, to address the problem, this paper applies learning algorithm into goal selection in dynamic environments. More specifically, we first develop a reorganization algorithm that can convert a goal net to its equivalent counterpart that a Q-learning algorithm can operate on; then, we define the key component of Q-learning, reward function, according to the feature of goal nets; and finally lots of experiments are conducted to show that, in dynamic environments, the agent with the learning algorithm significantly outperforms the one with the recursive searching algorithm. Therefore, our work suggests an agent model that can effectively be applied in dynamic time-sensitive domain, like computer games and the P2P systems of online movie watching. 2013-12-05T06:36:27Z 2019-12-06T19:31:17Z 2013-12-05T06:36:27Z 2019-12-06T19:31:17Z 2013 2013 Journal Article Zhang, H., Luo, X., Miao, C., Shen, Z., & You, J. (2013). Adaptive goal selection for agents in dynamic environments. Knowledge and information systems, in press. https://hdl.handle.net/10356/96478 http://hdl.handle.net/10220/18110 10.1007/s10115-013-0645-7 en Knowledge and information systems |
institution |
Nanyang Technological University |
building |
NTU Library |
country |
Singapore |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Computer science and engineering |
spellingShingle |
DRNTU::Engineering::Computer science and engineering Zhang, Huiliang Luo, Xudong Shen, Zhiqi You, Jin Miao, Chun Yan Adaptive goal selection for agents in dynamic environments |
description |
In psychology, goal-setting theory, which has been studied by psychologists for over 35 years, reveals that goals play significant roles in incentive, action and performance for human beings. Based on this theory, a goal net model has been proposed to design intelligent agents that can be viewed as a soft copy of human being somehow. The goal net model has been successfully applied in many agents, specially, non-player-character agents in computer games. Such an agent selects the optimal solution in all possible solutions found by using a recursive algorithm. However, if a goal net is very complex, the time of selection could be too long for the agent to respond quickly when the agent needs to re-select a new solution against the world’s change. Moreover, in some dynamic environments, it is impossible to know the exact outcome of choosing a solution in advance, and so the possible solutions cannot be evaluated precisely. Thus, to address the problem, this paper applies learning algorithm into goal selection in dynamic environments. More specifically, we first develop a reorganization algorithm that can convert a goal net to its equivalent counterpart that a Q-learning algorithm can operate on; then, we define the key component of Q-learning, reward function, according to the feature of goal nets; and finally lots of experiments are conducted to show that, in dynamic environments, the agent with the learning algorithm significantly outperforms the one with the recursive searching algorithm. Therefore, our work suggests an agent model that can effectively be applied in dynamic time-sensitive domain, like computer games and the P2P systems of online movie watching. |
author2 |
School of Computer Engineering |
author_facet |
School of Computer Engineering Zhang, Huiliang Luo, Xudong Shen, Zhiqi You, Jin Miao, Chun Yan |
format |
Article |
author |
Zhang, Huiliang Luo, Xudong Shen, Zhiqi You, Jin Miao, Chun Yan |
author_sort |
Zhang, Huiliang |
title |
Adaptive goal selection for agents in dynamic environments |
title_short |
Adaptive goal selection for agents in dynamic environments |
title_full |
Adaptive goal selection for agents in dynamic environments |
title_fullStr |
Adaptive goal selection for agents in dynamic environments |
title_full_unstemmed |
Adaptive goal selection for agents in dynamic environments |
title_sort |
adaptive goal selection for agents in dynamic environments |
publishDate |
2013 |
url |
https://hdl.handle.net/10356/96478 http://hdl.handle.net/10220/18110 |
_version_ |
1681059531134599168 |