Sorry, you need to enable JavaScript to visit this website.

MEMORY-BASED PARAMETERIZED SKILLS LEARNING FOR MAPLESS VISUAL NAVIGATION

Citation Author(s):
Yuyang Liu, Yang Cong and Gan Sun
Submitted by:
yuyang liu
Last updated:
11 September 2019 - 11:06pm
Document Type:
Poster
Document Year:
2019
Event:
Presenters:
Yuyang Liu
Paper Code:
1881
 

The recently-proposed reinforcement learning for mapless visual navigation can generate an optimal policy for searching different targets. However, most state-of-the-art deep reinforcement learning (DRL) models depend on hard rewards to learn the optimal policy, which can lead to the lack of previous diverse experiences. Moreover, these pre-trained DRL models cannot generalize well to un-trained tasks. To overcome these problems above, in this paper, we propose a Memorybased Parameterized Skills Learning (MPSL) model for mapless visual navigation. The parameterized skills in our MPSL are learned to predict critic parameters for un-trained tasks in actor-critic reinforcement learning, which can be achieved by transferring memory sequence knowledge from long short term memory network. In order to generalize into un-trained tasks, MPSL aims to capture more discriminative features by using a scene-specific layer. Finally, experiment results on an indoor photographic simulation framework AI2THOR demonstrate the effectiveness of our proposed MPSL model, and the generalization ability to un-trained tasks.

Contributions:
1.Parameterizing previous knowledge facilitates generalization to un-trained tasks.
2.Policy parameters are changed with task parameters and entered targets.
3.Feature alignment strategy to learn distinguishable features between different states.

up
1 user has voted: yuyang liu