Sorry, you need to enable JavaScript to visit this website.

Foveated Neural Network: Gaze Prediction On Egocentric Videos

Citation Author(s):
Mengmi Zhang, Keng-Teck Ma, Joo-Hwee Lim, Qi Zhao
Submitted by:
Keng Teck Ma
Last updated:
15 September 2017 - 4:11am
Document Type:
Poster
Document Year:
2017
Event:
Paper Code:
1433
 

A novel deep convolution neural network, named as Foveated Neural Network (FNN), is proposed to predict gaze on current frames in egocentric videos. The retina-like visual inputs from the region of interest on the previous frame get analysed and encoded. The fusion of the hidden representation of the previous frame and the feature maps of the current frame guides the gaze prediction process on the current frame. In order to simulate motions, we also include the dense optical flow between these adjacent frames as additional inputs to FNN. Experimental results show that FNN outperforms the state-of-the-art algorithms in the publicly available egocentric dataset. The analysis of FNN demonstrates that the hidden representations of the foveated visual input from the previous frame as well as the motion information between adjacent frames are efficient in improving gaze prediction performance in egocentric videos.

up
0 users have voted: