孙亮
开通时间:..
最后更新时间:..
点击次数:
论文类型:期刊论文
发表时间:2019-07-01
发表刊物:MULTIMEDIA TOOLS AND APPLICATIONS
收录刊物:SCIE、EI
卷号:78
期号:14
页面范围:20533-20556
ISSN号:1380-7501
关键字:Attention mechanism; Convolutional LSTM; Spatial transformer; Video action recognition
摘要:As an important issue in video classification, human action recognition is becoming a hot topic in computer vision. The ways of effectively representing the spatial static and temporal dynamic information of videos are important problems in video action recognition. This paper proposes an attention mechanism based convolutional LSTM action recognition algorithm to improve the accuracy of recognition by extracting the salient regions of actions in videos effectively. First, GoogleNet is used to extract the features of video frames. Then, those feature maps are processed by the spatial transformer network for the attention. Finally the sequential information of the features is modeled via the convolutional LSTM to classify the action in the original video. To accelerate the training speed, we adopt the analysis of temporal coherence to reduce the redundant features extracted by GoogleNet with trivial accuracy loss. In comparison with the state-of-the-art algorithms for video action recognition, competitive results are achieved on three widely-used datasets, UCF-11, HMDB-51 and UCF-101. Moreover, by using the analysis of temporal coherence, desirable results are obtained while the training time is reduced.