A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization

When applying the learning systems to real-world problems, which have a lot of unknown or uncertain things, there are some issues that need to be solved. One of them is the abstraction ability. In reinforcement learning, to complete each task, the agent will learn to find the best policy. Neverthele...

Full description

Saved in:
Bibliographic Details
Main Authors: Ahmad Afif, Mohd Faudzi, Hirotaka, Takano, Junichi, Murata
Format: Conference or Workshop Item
Language:en
Published: 2013
Subjects:
Online Access:http://umpir.ump.edu.my/id/eprint/6961/1/A_study_on_Visual_Abstraction_for_Reinforcement_Learning_Problem_Using_Learning_Vector_Quantization.pdf
http://umpir.ump.edu.my/id/eprint/6961/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1831522076452716544
author Ahmad Afif, Mohd Faudzi
Hirotaka, Takano
Junichi, Murata
author_facet Ahmad Afif, Mohd Faudzi
Hirotaka, Takano
Junichi, Murata
author_sort Ahmad Afif, Mohd Faudzi
building UMPSA Library
collection Institutional Repository
content_provider Universiti Malaysia Pahang Al-Sultan Abdullah
content_source UMPSA Institutional Repository
continent Asia
country Malaysia
description When applying the learning systems to real-world problems, which have a lot of unknown or uncertain things, there are some issues that need to be solved. One of them is the abstraction ability. In reinforcement learning, to complete each task, the agent will learn to find the best policy. Nevertheless, if a different task is given, we cannot know for sure whether the acquired policy is still valid or not. However, if we can make an abstraction by extract some rules from the policy, it will be easier to understand and possible to apply the policy to different tasks. In this paper, we apply the abstraction at a perceptual level. In the first phase, an action policy is learned using Q-learning, and in the second phase, Learning Vector Quantization is used to extract information out of the learned policy. In this paper, it is verified that by applying the proposed abstraction method, a more useful and simpler representation of the learned policy can be achieved.
format Conference or Workshop Item
id my.ump.umpir.6961
institution Universiti Malaysia Pahang
language en
publishDate 2013
record_format eprints
spelling my.ump.umpir.69612015-03-03T09:33:29Z http://umpir.ump.edu.my/id/eprint/6961/ A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization Ahmad Afif, Mohd Faudzi Hirotaka, Takano Junichi, Murata TK Electrical engineering. Electronics Nuclear engineering When applying the learning systems to real-world problems, which have a lot of unknown or uncertain things, there are some issues that need to be solved. One of them is the abstraction ability. In reinforcement learning, to complete each task, the agent will learn to find the best policy. Nevertheless, if a different task is given, we cannot know for sure whether the acquired policy is still valid or not. However, if we can make an abstraction by extract some rules from the policy, it will be easier to understand and possible to apply the policy to different tasks. In this paper, we apply the abstraction at a perceptual level. In the first phase, an action policy is learned using Q-learning, and in the second phase, Learning Vector Quantization is used to extract information out of the learned policy. In this paper, it is verified that by applying the proposed abstraction method, a more useful and simpler representation of the learned policy can be achieved. 2013 Conference or Workshop Item PeerReviewed application/pdf en http://umpir.ump.edu.my/id/eprint/6961/1/A_study_on_Visual_Abstraction_for_Reinforcement_Learning_Problem_Using_Learning_Vector_Quantization.pdf Ahmad Afif, Mohd Faudzi and Hirotaka, Takano and Junichi, Murata (2013) A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization. In: Proceedings of SICE Annual Conference (SICE) , 14-17 Sept. 2013 , Nagoya, Japan. pp. 1326-1331.. (Published)
spellingShingle TK Electrical engineering. Electronics Nuclear engineering
Ahmad Afif, Mohd Faudzi
Hirotaka, Takano
Junichi, Murata
A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title_full A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title_fullStr A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title_full_unstemmed A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title_short A study on Visual Abstraction for Reinforcement Learning Problem Using Learning Vector Quantization
title_sort study on visual abstraction for reinforcement learning problem using learning vector quantization
topic TK Electrical engineering. Electronics Nuclear engineering
url http://umpir.ump.edu.my/id/eprint/6961/1/A_study_on_Visual_Abstraction_for_Reinforcement_Learning_Problem_Using_Learning_Vector_Quantization.pdf
http://umpir.ump.edu.my/id/eprint/6961/
url_provider http://umpir.ump.edu.my/