Abstract
Recently spatio-temporal local features have been proposed as image features to recognize events or human actions in videos. In this paper, we propose yet another local spatio-temporal feature based on the SURF detector, which is a lightweight local feature. Our method consists of two parts: extracting visual features and extracting motion features. First, we select candidate points based on the SURF detector. Next, we calculate motion features at each point with local temporal units divided in order to consider consecutiveness of motions. Since our proposed feature is intended to be robust to rotation, we rotate optical flow vectors to the main direction of extracted SURF features. In the experiments, we evaluate the proposed spatio-temporal local feature with the common dataset containing six kinds of simple human actions. As the result, the accuracy achieves 86%, which is almost equivalent to state-of-the-art. In addition, we make experiments to classify large amounts of Web video clips downloaded from Youtube.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Dollar, P., Cottrell, G., Belongie, S.: Behavior recognition via sparse spatio-temporal features. In: Proc. of Surveillance and Performance Evaluation of Tracking and Surveillance, pp. 65–72 (2005)
Laptev, I., Lindeberg, T.: Local descriptors for spatio-temporal recognition. In: Proc. of IEEE International Conference on Computer Vision (2003)
Csurka, G., Bray, C., Dance, C., Fan, L.: Visual categorization with bags of keypoints. In: Proc. of ECCV Workshop on Statistical Learning in Computer Vision, pp. 59–74 (2004)
Alireza, F., Greg, M.: Action recognition by learning mid-level feature. In: Proc.of IEEE Computer Vision and Pattern Recognition (2008)
Herbert, B., Andreas, E., Tinne, T., Luc, G.: Surf: Speeded up robust features. Computer Vision and Image Understanding, 346–359 (2008)
Lucas, B., Kanade, T.: An iterative image registration technique with an application to stereo vision. In: Proc. of International Joint Conference on Artificial Intelligence, pp. 674–679 (1981)
Lowe, D.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 91–110 (2004)
Fanti, C., Perona, P.: Hybrid models for human motion recognition. In: Proc. of IEEE Computer Vision and Pattern Recognition (2005)
Rao, C., Yilmaz, A., Shah, M.: View-invariant representation and recognition of actions. International Journal of Computer Vision 50(2), 203–226 (2002)
Yacoob, Y., Black, M.J.: Parameterized modeling and recognition of activities. Computer Vision and Image Understanding 72(2), 203–226 (2002)
Konrad, S., Luc, G.: Action snippets: How many frames does human action recognition require? In: Proc. of IEEE Computer Vision and Pattern Recognition (2008)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Noguchi, A., Yanai, K. (2010). Extracting Spatio-temporal Local Features Considering Consecutiveness of Motions. In: Zha, H., Taniguchi, Ri., Maybank, S. (eds) Computer Vision – ACCV 2009. ACCV 2009. Lecture Notes in Computer Science, vol 5995. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-12304-7_43
Download citation
DOI: https://doi.org/10.1007/978-3-642-12304-7_43
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-12303-0
Online ISBN: 978-3-642-12304-7
eBook Packages: Computer ScienceComputer Science (R0)