Re-exploration of ε-greedy in deep reinforcement learning

This paper presents re-exploration as a method for improving the existing method for balancing the exploration/exploitation problem integral to reinforcement learning. The proposed method uses a ε-greedy method called “decreasing epsilon,” which reiterate the method after a certain period of episode...

Full description

Saved in:
Bibliographic Details
Main Authors: Muhamad Amin, Muhamad Ridzuan Radin, Othman, Mohd. Fauzi
Format: Conference or Workshop Item
Published: 2021
Subjects:
Online Access:http://eprints.utm.my/id/eprint/96174/
http://dx.doi.org/10.1007/978-981-16-4803-8_27
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This paper presents re-exploration as a method for improving the existing method for balancing the exploration/exploitation problem integral to reinforcement learning. The proposed method uses a ε-greedy method called “decreasing epsilon,” which reiterate the method after a certain period of episodes in the middle of the learning. The experiment was conducted using Turtlebot3 simulation under the Robot Operating System (ROS) environment. The evaluation involved comparing the existing method, which is pure exploitation (totally greedy), conventional ε-greedy method and proposed method, which is decreasing-epsilon with the re-exploration method. The preliminary results indicate that applying re-exploration method is easier to implement and yet able to improve the reward obtained with in shorter time (episode) compared to the conventional method.