Reinforcement learning based autonomous vehicle for exploration and exploitation of undiscovered track
Abstract
This research focuses on autonomous traversal of land vehicles through exploring
undiscovered tracks and overcoming environmental barriers. Most of the existing
systems can only operate and traverse in a distinctive mapped model especially in a
known area. However, the proposed system which is trained by Deep Reinforcement
Learning can learn by itself to operate autonomously in extreme conditions. The
dynamic double deep Q-learning (DDQN) model enables the proposed system not
to be confined only to known environments. The ambient environmental obstacles
are identified through Faster R-CNN for smooth movement of the autonomous vehicle. The exploration and exploitation strategies of DDQN enables the autonomous
agent to learn proper decisions for various dynamic environments and tracks. The
proposed model is tested in a gaming environment. It shows the overall effectiveness
in traversing of autonomous land vehicles in comparison to the existing models. The
goal is to integrate Deep Reinforcement learning and Faster R-CNN to make the
system effective to traverse through undiscovered paths by detecting obstacles.