Sorry, you need to enable JavaScript to visit this website.

Attention Boosted Deep Networks for Video Classficaition

Citation Author(s):
Junyong You, Jari Korhonen
Submitted by:
Junyong You
Last updated:
2 November 2020 - 10:39am
Document Type:
Presentation Slides
Document Year:
Junyong You
Paper Code:

Video classification can be performed by summarizing image contents of individual frames into one class by deep neural networks, e.g., CNN and LSTM. Human interpretation of video content is influenced by the attention mechanism. In other words, video class can be more attentively decided by certain information than others. In this paper, we propose to integrate the attention mechanism into deep networks for video classification. The proposed framework employs 2D CNN networks with ImageNet pretrained weights to extract features of video frames that are then fed to a bidirectional LSTM network for video classification. An attention block has been developed that can be added after the LSTM network in the proposed framework. Several different 2D CNN architectures have been tested in the experiments. The results with respect to two publicly available datasets have demonstrated that integrating attention can boost the performance of deep networks in video classification compared to not applying the attention block. We also found out that applying attention to the LSTM outputs on the VGG19 architecture provides the highest classification accuracy in the proposed framework.

0 users have voted: