Skip to main content

Automatic Multi-view Action Recognition with Robust Features

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 10636))

Abstract

This paper proposes view-invariant features to address multi-view action recognition for different actions performed in different views. The view-invariant features are obtained from clouds of varying temporal scale by extracting holistic features, which are modeled to explicitly take advantage of the global, spatial and temporal distribution of interest points. The proposed view-invariant features are highly discriminative and robust for recognizing actions as the view changes. This paper proposes a mechanism for real world application which can follow the actions of a person in a video based on image sequences and can separate these actions according to given training data. Using the proposed mechanism, the beginning and ending of an action sequence can be labeled automatically without the need for manual setting. It is not necessary in the proposed approach to re-train the system if there are changes in scenario, which means the trained database can be applied in a wide variety of environments. The experiment results show that the proposed approach outperforms existing methods on KTH and WEIZMANN datasets.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Efros, A.A., Berg, A.C., Mori, G., Malik, J.: Recognizing action at a distance. In: Ninth IEEE International Conference on Computer Vision, vol. 2, pp. 726–733 (2003)

    Google Scholar 

  2. Fathi, A., Mori, G.: Action recognition by learning midlevel motion features. In: IEEE Conference on Computer Vision and Pattern Recognition (2008)

    Google Scholar 

  3. Rao, C., Shah, M.: View-invariance in action recognition. Comput. Vis. Pattern Recognit. 2, 316–322 (2001)

    Google Scholar 

  4. Ali, A., Aggarwal, J.: Segmentation and recognition of continuous human activity. In: IEEE Workshop on Detection and Recognition of Events in Video, p. 28 (2001)

    Google Scholar 

  5. Ramanan, D., Forsyth, D.A.: Automatic annotation of everyday movements. In: Conference on Neural Information Processing Systems (2003)

    Google Scholar 

  6. Sheikh, Y., Sheikh, M., Shah, M.: Exploring the space of a human action. In: International Conference on Computer Vision (2005)

    Google Scholar 

  7. Gorelick, L., Blank, M., Shechtman, E., Irani, M., Basri, R.: Actions as space-time shapes. Pattern Anal. Mach. Intell. 29(12), 2247–2253 (2007)

    Article  Google Scholar 

  8. Ke, Y., Sukthankar, R., Hebert, M.: Efficient visual event detection using volumetric features. In: IEEE Computer Society, Los Alamitos, CA, USA, vol. 1, pp. 166–173 (2005)

    Google Scholar 

  9. Yilmaz, A., Shah, M.: Actions sketch: a novel action representation. In: Computer Vision and Pattern Recognition, pp. 984–989 (2005)

    Google Scholar 

  10. Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local SVM approach. In: International Conference on Pattern Recognition, pp. 32–36 (2004)

    Google Scholar 

  11. Scovanner, P., Ali, S., Shah, M.: A 3-Dimensional sift descriptor and its application to action recognition. In: International conference on Multimedia, pp. 357–360 (2007)

    Google Scholar 

  12. Lowe, D.: Distinctive image features from scale-invariant key-points. Int. J. Comput. Vision 20, 91–110 (2003)

    Google Scholar 

  13. Willems, G., Tuytelaars, T., Van Gool, L.: An efficient dense and scale-invariant spatio-temporal interest point detector. In: European Conference on Computer Vision, vol. 2, pp. 650–663 (2008)

    Google Scholar 

  14. Dollar, P., Rabaud, V., Cottrell, G., Belongie, S.: Behavior recognition via sparse spatio-temporal features. In: International Conference on Computer Communications and Networks, pp. 65–72 (2005)

    Google Scholar 

  15. Niebles, J.C., Fei-Fei, L.: A hierarchical model of shape and appearance for human action classification. In: Computer Vision and Pattern Recognition (2007)

    Google Scholar 

  16. Singh, S., Velastin, S.A., Ragheb, H.: MuHAVi: a multicamera human action video dataset for the evaluation of action recognition methods. In: 2nd Workshop on Activity Monitoring by Multi-Camera Surveillance Systems (AMMCSS), 29 August, Boston, USA (2010)

    Google Scholar 

  17. Eweiwi, A., Cheema, S., Thurau, C., Bauckhage, C.: Temporal key poses for human action recognition. In: International Conference on Computer Vision Workshops (2011)

    Google Scholar 

  18. Bregonzio, M., Gong, S., Xiang, T.: Recognising action as clouds of space-time interest points. In: Computer Vision and Pattern Recognition, pp. 1948–1955 (2009)

    Google Scholar 

  19. Parker, J.: Algorithms for Image Processing and Computer Vision. Wiley Computer Publishing, New York (1997)

    Google Scholar 

  20. Shorack, G.R., Wellner, J.A.: Empirical Processes with Applications to Statistics. Wiley, New York (1986)

    MATH  Google Scholar 

  21. Gilbert, A., Illingworth, J., Bowden, R.: Scale invariant action recognition using compound features mined from dense spatio-temporal corners. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008. LNCS, vol. 5302, pp. 222–233. Springer, Heidelberg (2008). doi:10.1007/978-3-540-88682-2_18

    Chapter  Google Scholar 

  22. Savarese, S., Pozo, A.D., Niebles, J., Fei-Fei, L.: Spatial temporal correlations for unsupervised action classification. In: IEEE Workshop on Motion and Video Computing (2008)

    Google Scholar 

  23. Zhang, Z., Hu, Y., Chan, S., Chia, L.-T.: Motion context: a new representation for human action recognition. In: Forsyth, D., Torr, P., Zisserman, A. (eds.) ECCV 2008. LNCS, vol. 5305, pp. 817–829. Springer, Heidelberg (2008). doi:10.1007/978-3-540-88693-8_60

    Chapter  Google Scholar 

  24. Nowozin, S., Bakir, G.H., Tsuda, K.: Discriminative sub-sequence mining for action classification. In: International Conference on Computer Vision, pp. 1–8 (2007)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mukesh Prasad .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Chou, KP. et al. (2017). Automatic Multi-view Action Recognition with Robust Features. In: Liu, D., Xie, S., Li, Y., Zhao, D., El-Alfy, ES. (eds) Neural Information Processing. ICONIP 2017. Lecture Notes in Computer Science(), vol 10636. Springer, Cham. https://doi.org/10.1007/978-3-319-70090-8_56

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-70090-8_56

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-70089-2

  • Online ISBN: 978-3-319-70090-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics