Abstract
The purpose of this study is learning and classification of video activities using video color and motion information. The video activity labeling is important for many applications such as video content modeling, indexing, and quick access to content. In this study video activity recognition is performed by deep learning. In order to learn visual features of video, Convolutional Neural Network (CNN) layers and a special type of recursive networks, Long-Short Term Memory (LSTM), layers are stacked. Video sequence learning is performed by end-to-end training. Recent works on deep learning employ color end motion information together to improve learning and classification accuracy. In this study, unlike the existing models, video motion content is learned using SIFT flow vectors and motion and color features are fused for activity recognition. Performance tests performed on a commonly used benchmarking data set, UCF 101 which includes activity labeled videos from 101 action categories such as 'Biking', 'Playing Guitar,' demonstrate that SIFT flow vectors allow us to model motion information more accurately than optical flow vectors and increase video motion classification performance.
Translated title of the contribution | Video action classification by deep learning |
---|---|
Original language | Turkish |
Title of host publication | 2017 25th Signal Processing and Communications Applications Conference, SIU 2017 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9781509064946 |
DOIs | |
Publication status | Published - 27 Jun 2017 |
Event | 25th Signal Processing and Communications Applications Conference, SIU 2017 - Antalya, Turkey Duration: 15 May 2017 → 18 May 2017 |
Publication series
Name | 2017 25th Signal Processing and Communications Applications Conference, SIU 2017 |
---|
Conference
Conference | 25th Signal Processing and Communications Applications Conference, SIU 2017 |
---|---|
Country/Territory | Turkey |
City | Antalya |
Period | 15/05/17 → 18/05/17 |
Bibliographical note
Publisher Copyright:© 2017 IEEE.