skip to main content
research-article

Simulating Fixations When Looking at Visual Arts

Authors Info & Claims
Published:26 June 2015Publication History
Skip Abstract Section

Abstract

When people look at pictures, they fixate on specific areas. The sequences of such fixations are so characteristic for certain pictures that metrics can be derived that allow successful grouping of similar pieces of visual art. However, determining enough fixation sequences by eye tracking is not practically feasible for large groups of people and pictures. In order to get around this limitation, we present a novel algorithm that simulates eye movements by calculating scan paths for images and time frames in real time. The basis of our algorithm is an attention model that combines and optimizes rectangle features with Adaboost. The model is adapted to the characteristics of the retina, and its input is dependent on a few earlier fixations. This method results in significant improvements compared to previous approaches. Our simulation process delivers the same data structures as an eye tracker, thus can be analyzed by standard eye-tracking software. A comparison with recorded data from eye tracking experiments shows that our algorithm for simulating fixations has a very good prediction quality for the stimulus areas on which many subjects focus. We also compare the results with those from earlier works. Finally, we demonstrate how the presented algorithm can be used to calculate the similarity of pictures in terms of human perception.

References

  1. Mohamed Alaa El-Dien Mahmoud Hussein Aly. 2011. Searching Large-Scale Image Collections. Ph.D. Dissertation. California Institute of Technology, Pasadena, CA.Google ScholarGoogle Scholar
  2. James W. Davis, Alexander M. Morison, and David D. Woods. 2007. An adaptive focus-of-attention model for video surveillance and monitoring. Machine Vision and Applications 18, 1, 41--64. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Feng-GUI. 2014. Feng Shui for Graphic User Interfaces. Retrieved April 23, 2014 from http://feng-gui.com/help#analysis.Google ScholarGoogle Scholar
  4. Benjamin Höferlin, Hermann Pflüger, Markus Höferlin, Gunther Heidemann, and Daniel Weiskopf. 2012. Learning a visual attention model for adaptive fast-forward in video surveillance. In ICPRAM (2)’12. 25--32.Google ScholarGoogle Scholar
  5. James Hoffman and Baskaran Subramaniam. 1995. The role of visual attention in saccadic eye movements. Perception & Psychophysics 57, 6, 787--795.Google ScholarGoogle Scholar
  6. Kenneth Holmqvist, Marcus Nyström, Richard Andersson, Richard Dewhurst, Halszka Jarodzka, and Joost van de Weijer. 2010. Eye Tracking: A Comprehensive Guide to Methods and Measures. Oxford University Press, New York, NY.Google ScholarGoogle Scholar
  7. Laurent Itti. 2005. Quantifying the contribution of low-level saliency to human eye movements in dynamic scenes. Visual Cognition 12, 1093--1123.Google ScholarGoogle ScholarCross RefCross Ref
  8. Laurent Itti and Nitin Dhavale. 2003. Realistic avatar eye and head animation using a neurobiological model of visual attention. In Proceedings of SPIE. SPIE Press, 64--78.Google ScholarGoogle Scholar
  9. Laurent Itti and Christof Koch. 2013. Bottom-Up Visual Attention Home Page; The Interactive Demo. Retrieved April 23, 2015 from http://ilab.usc.edu/bu/javaDemo/ie/demo.html.Google ScholarGoogle Scholar
  10. Laurent Itti, Christof Koch, and Ernst Niebur. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis Machine Intelligence 20, 11, 1254--1259. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Hector Jasso and Jochen Triesch. 2007. Learning to attend—from bottom-up to top-down. In Attention in Cognitive Systems. Theories and Systems from an Interdisciplinary Viewpoint, Lucas Paletta and Erich Rome (Eds.). Lecture Notes in Computer Science, Vol. 4840. Springer, Berlin, 106--122. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Tilke Judd, Krista Ehinger, Frédo Durand, and Antonio Torralba. 2009. Learning to predict where humans look. In Proceedings of the IEEE 12th International Conference on Computer Vision. IEEE, 2106--2113.Google ScholarGoogle ScholarCross RefCross Ref
  13. Wolf Kienzle, Bernhard Schölkopf, Felix A. Wichmann, and Matthias O. Franz. 2007. How to find interesting locations in video: A spatiotemporal interest point detector learned from human eye movements. In Proceedings of the 29th DAGM Conference on Pattern Recognition. Springer-Verlag, Berlin, 405--414. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Krystian Mikolajczyk, Tinne Tuytelaars, Cordelia Schmid, Andrew Zisserman, Jiri Matas, Frederik Schaffalitzky, Timor Kadir, and Luc van Gool. 2005. A comparison of affine region detectors. International Journal of Computer Vision 65, 1/2 43--72. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Frieder Nake. 1974. Ästhetik als Informationsverarbeitung. Springer Verlag, Wien, New York.Google ScholarGoogle Scholar
  16. Sunaad Nataraju, Vineeth Balasubramanian, and Sethuraman Panchanathan. 2009. Learning attention based saliency in videos from human eye movements. In Proceedings of the 2009 International Conference on Motion and Video Computing (WMVC’09). IEEE Computer Society, Washington, DC, 134--139. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Yaqing Niu, Rebecca M. Todd, Matthew Kyan, and Adam K. Anderson. 2012. Visual and emotional salience influence eye movements. ACM Transactions on Applied Perceptions 9, 3, Article 13, 18 pages. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Robert J. Peters and Laurent Itti. 2007. Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention. In IEEE Conference on Computer Vision and Pattern Recognition, 2007 (CVPR’07). 1--8.Google ScholarGoogle Scholar
  19. Claudio M. Privitera and Lawrence W. Stark. 2000. Algorithms for defining visual regions-of-interest: Comparison with eye fixations. IEEE Transactions on Pattern Analysis and Machine Intelligence 22, 9, 970--982. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Umesh Rajashekar, Ian van der Linde, Alan C. Bovik, and Lawrence K. Cormack. 2008. GAFFE: A gaze-attentive fixation finding engine. IEEE Transactions on Image Processing 17, 564--573. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. Robert W. Rodieck. 1998. The First Steps in Seeing. Sinauer Associates, Inc., Sunderland, MA.Google ScholarGoogle Scholar
  22. Jukka Saarinen. 1993. Shifts of visual attention at fixation and away from fixation. Vision Research 33, 8, 1113--1117.Google ScholarGoogle ScholarCross RefCross Ref
  23. Paul Viola and Michael J. Jones. 2004. Robust real-time face detection. International Journal of Computer Vision 57, 2, 137--154. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Jeremy M. Wolfe. 1994. Guided search 2.0: A revised model of visual search. Psychonomic Bulletin & Review 1, 2, 202--238.Google ScholarGoogle ScholarCross RefCross Ref
  25. Qi Zhao and C. Koch. 2011. Learning visual saliency. In Proceedings of the 2011 45th Annual Conference on Information Sciences and Systems (CISS’11). 1--6.Google ScholarGoogle Scholar

Index Terms

  1. Simulating Fixations When Looking at Visual Arts

                Recommendations

                Comments

                Login options

                Check if you have access through your login credentials or your institution to get full access on this article.

                Sign in

                Full Access

                • Published in

                  cover image ACM Transactions on Applied Perception
                  ACM Transactions on Applied Perception  Volume 12, Issue 3
                  July 2015
                  92 pages
                  ISSN:1544-3558
                  EISSN:1544-3965
                  DOI:10.1145/2798084
                  Issue’s Table of Contents

                  Copyright © 2015 ACM

                  Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

                  Publisher

                  Association for Computing Machinery

                  New York, NY, United States

                  Publication History

                  • Published: 26 June 2015
                  • Accepted: 1 February 2015
                  • Revised: 1 September 2014
                  • Received: 1 November 2013
                  Published in tap Volume 12, Issue 3

                  Permissions

                  Request permissions about this article.

                  Request Permissions

                  Check for updates

                  Qualifiers

                  • research-article
                  • Research
                  • Refereed

                PDF Format

                View or Download as a PDF file.

                PDF

                eReader

                View online with eReader.

                eReader