Sorry, you need to enable JavaScript to visit this website.

TEMPORAL ACTION LOCALIZATION WITH TWO-STREAM SEGMENT-BASED RNN

Citation Author(s):
Tianwei Lin, Xu Zhao, Zhaoxuan Fan
Submitted by:
Tianwei Lin
Last updated:
14 September 2017 - 6:44am
Document Type:
Poster
Document Year:
2017
Event:
Presenters:
Tianwei Lin
Paper Code:
1520
 

Temporal Action localization is a more challenging vision task than action recognition because videos to be analyzed
are usually untrimmed and contain multiple action instances. In this paper, we investigate the potential of recurrent neural network, toward three critical aspects for solving this problem, namely, high-performance feature, high-quality temporal segments and effective recurrent neural network architecture. First of all, we introduce the two-stream (spatial and temporal) network for feature extraction. Then, we propose a novel temporal selective search method to generate temporal segments with variable lengths. Finally, we design a twobranch LSTM architecture for category prediction and confi- dence score computation. Our proposed approach to action localization, along with the key components, say, segments generation and classification architecture, are evaluated on the THUMOS’14 dataset and achieve promising performance by comparing with other state-of-the-art methods.

up
0 users have voted: