KEYWORDS: Data modeling, Visualization, RGB color model, Network architectures, 3D modeling, Detection and tracking algorithms, Monte Carlo methods, Machine learning, Visual process modeling
Deep reinforcement learning has greatly simplified visual navigation by utilizing the end-to-end network training strategy. Unlike previous navigation methods which build upon high-precision maps, deep reinforcement learning-based method enables real-time navigation by only taking one image as input at a time. As such, deep reinforcement learning based navigation methods are applicable to a variety number of applications in robotics/vision communities, thanks to its light-weight computational cost. Despite the advantages, however, these methods still suffer from inefficient data exploration and poor convergence on network training. In this paper, we propose to use inverse reinforcement learning to solve the problem,which can provide more accurate and efficient guidance for decision-making. The proposed method is able to learn a more effective reward function from less training data. Experiments demonstrated that the proposed method achieves a higher success rate of navigation and produces paths that are more similar to the optimal ones compared to the reinforcement learning baselines.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.