ABSTRACT
In recent times, the proliferation of multimedia devices and reduced costs of data storage have enabled people to easily record and collect a large number of home videos; furthermore, this collection is growing with time. With the popularity of participatory media such as YouTube and facebook, problems are encountered when people intend to share their home videos with others. The first problem is that different people might be interested in different video content. Given the numbers of home videos, it is a time-consuming and hard task to manually select proper content for people with different interests. Secondly, as short videos are becoming more and more popular in media sharing applications, people need to manually cut and edit home videos which is again a tedious task. In this paper, we propose a method that employs affective analysis to automatically create video presentations from home videos. Our novel method adaptively creates presentations based on three properties: emotional tone, local main character and global main character. A novel sparsity-based affective labeling method is proposed to identify the emotional content of the videos. The local and global main characters are determined by applying face recognition in each shot. To demonstrate the proposed method, three kinds of presentations are created for family, acquaintance and outsider. Experimental results show that our method is very effective in video sharing and the users are satisfied with the videos generated by our method.
- E. Babbie. The practice of social research. Wadsworth Pub Co, 2007.Google Scholar
- D. Bordwell, K. Thompson, and J. Ashton. Film art: An introduction. McGraw-Hill, 1990.Google Scholar
- M. Bradley. Emotional memory: A dimensional analysis. Emotions: Essays on emotion theory, pages 97--134, 1994.Google Scholar
- C. Bundesen. A theory of visual attention. Psychological review, 97(4):523, 1990.Google ScholarCross Ref
- J. Burgess and J. Green. YouTube: Online video and participatory culture. Polity Press, 2009.Google Scholar
- E. Candès. Compressive sampling. In Proceedings of the International Congress of Mathematicians, volume 3, pages 1433--1452, 2006.Google Scholar
- Z. Cernekova, C. Kotropoulos, and I. Pitas. Video shot segmentation using singular value decomposition. In Proceedings of IEEE International Conference on Multimedia and Expo, volume 2, pages 301--304, 2003. Google ScholarDigital Library
- L. Chen, X. Xie, X. Fan, W. Ma, H. Zhang, and H. Zhou. A visual attention model for adapting images on small displays. Multimedia systems, 9(4):353--364, 2003.Google ScholarDigital Library
- C. Color. CeWe Photobook. Photoworld, 2007.Google Scholar
- O. Demerdash, S. Bergler, L. Kosseim, and P. Langshaw. Generating Adaptive Multimedia Presentations Based on a Semiotic Framework. Advances in Artificial Intelligence, pages 417--421, 2005. Google ScholarDigital Library
- P. Ekman. An argument for basic emotions. Cognition & Emotion, 6(3):169--200, 1992.Google ScholarCross Ref
- A. Girgensohn, J. Boreczky, P. Chiu, J. Doherty, J. Foote, G. Golovchinsky, S. Uchihashi, and L. Wilcox. A semi-automatic approach to home video editing. In Proceedings of ACM symposium on User interface software and technology, pages 81--89, 2000. Google ScholarDigital Library
- E. Goldstein. Sensation and perception. Brooks/Cole Publishing, 1999.Google Scholar
- A. Hanjalic and L. Xu. Affective video content representation and modeling. IEEE Transactions on Multimedia, 7(1):143--154, 2005. Google ScholarDigital Library
- K. Hevner. Experimental studies of the elements of expression in music. The American Journal of Psychology, 48(2):246--268, 1936.Google ScholarCross Ref
- W. James. The principles of psycholog(Vol. 2). New York: Holt, 1890.Google Scholar
- H. Kang. Affective content detection using HMMs. In Proceedings of ACM international conference on Multimedia, pages 259--262, 2003. Google ScholarDigital Library
- A. Kuper and J. Kuper. The social science encyclopedia. Routledge/Thoemms Press, 1985.Google Scholar
- D. Needell and J. Tropp. CoSaMP: Iterative signal recovery from incomplete and inaccurate samples. Applied and Computational Harmonic Analysis, 26(3):301--321, 2009.Google ScholarCross Ref
- D. Olson and D. Delen. Advanced data mining techniques. Springer Verlag, 2008. Google ScholarDigital Library
- M. Ondaatje and W. Murch. The conversations: Walter Murch and the art of editing film. Knopf, 2002.Google Scholar
- M. Rabbath, P. Sandhaus, and S. Boll. Automatic creation of photo books from stories in social media. In Proceedings of ACM SIGMM workshop on Social media, pages 15--20, 2010. Google ScholarDigital Library
- Y. Rui, T. Huang, and S. Chang. Image Retrieval: Current Techniques, Promising Directions, and Open Issues. Journal of visual communication and image representation, 10(1):39--62, 1999.Google Scholar
- J. Russell. A circumplex model of affect. Journal of personality and social psychology, 39(6):1161--1178, 1980.Google Scholar
- G. Salton and M. McGill. Introduction to modern information retrieval, volume 1. McGraw-Hill New York, 1983. Google ScholarDigital Library
- M. Soleymani, J. Kierkels, G. Chanel, and T. Pun. A bayesian framework for video affective representation. In Affective Computing and Intelligent Interaction and Workshops, pages 1--7, 2009.Google ScholarCross Ref
- K. Sun and J. Yu. Video affective content representation and recognition using video affective tree and hidden Markov models. Affective Computing and Intelligent Interaction, pages 594--605, 2007. Google ScholarDigital Library
- K. Sun, J. Yu, Y. Huang, and X. Hu. An improved valence-arousal emotion space for video affective content representation and recognition. In Proceedings of IEEE International Conference on Multimedia and Expo, pages 566--569, 2009. Google ScholarDigital Library
- M. Turk and A. Pentland. Face recognition using eigenfaces. In Computer Vision and Pattern Recognition, pages 586--591, 1991.Google ScholarCross Ref
- S. Zhang, Q. Tian, Q. Huang, W. Gao, and S. Li. Utilizing affective analysis for efficient movie browsing. In Proceedings of IEEE International Conference on Image Processing, pages 1853--1856, 2010. Google ScholarDigital Library
Index Terms
- Affect-based adaptive presentation of home videos
Recommendations
Understanding behaviors and needs for home videos
BCS-HCI '08: Proceedings of the 22nd British HCI Group Annual Conference on People and Computers: Culture, Creativity, Interaction - Volume 2Nowadays, people capture, edit and share more and more home videos. Devices with video capturing capability and services for video sharing and editing are becoming increasingly popular. This study was conducted to chart in more detail what users ...
Adaptive Content Model for Multimedia Presentation
CW '02: Proceedings of the First International Symposium on Cyber Worlds (CW'02)Multimedia presentation systems require flexible supportfor the modeling of multimedia content models. Manypresentation systems provide the synchronized, sequentialor concurrent, and possible interactive, transfer ofstreams multimedia data such as audio,...
Real-time expression of affect through respiration
CASA' 2010 Special IssueAffect has been shown to influence respiration in people. This paper takes this insight and proposes a real-time model to express affect through respiration in virtual humans. Fourteen affective states are explored: excitement, relaxation, focus, pain, ...
Comments