Abstract
When people look at pictures, they fixate on specific areas. The sequences of such fixations are so characteristic for certain pictures that metrics can be derived that allow successful grouping of similar pieces of visual art. However, determining enough fixation sequences by eye tracking is not practically feasible for large groups of people and pictures. In order to get around this limitation, we present a novel algorithm that simulates eye movements by calculating scan paths for images and time frames in real time. The basis of our algorithm is an attention model that combines and optimizes rectangle features with Adaboost. The model is adapted to the characteristics of the retina, and its input is dependent on a few earlier fixations. This method results in significant improvements compared to previous approaches. Our simulation process delivers the same data structures as an eye tracker, thus can be analyzed by standard eye-tracking software. A comparison with recorded data from eye tracking experiments shows that our algorithm for simulating fixations has a very good prediction quality for the stimulus areas on which many subjects focus. We also compare the results with those from earlier works. Finally, we demonstrate how the presented algorithm can be used to calculate the similarity of pictures in terms of human perception.
- Mohamed Alaa El-Dien Mahmoud Hussein Aly. 2011. Searching Large-Scale Image Collections. Ph.D. Dissertation. California Institute of Technology, Pasadena, CA.Google Scholar
- James W. Davis, Alexander M. Morison, and David D. Woods. 2007. An adaptive focus-of-attention model for video surveillance and monitoring. Machine Vision and Applications 18, 1, 41--64. Google ScholarDigital Library
- Feng-GUI. 2014. Feng Shui for Graphic User Interfaces. Retrieved April 23, 2014 from http://feng-gui.com/help#analysis.Google Scholar
- Benjamin Höferlin, Hermann Pflüger, Markus Höferlin, Gunther Heidemann, and Daniel Weiskopf. 2012. Learning a visual attention model for adaptive fast-forward in video surveillance. In ICPRAM (2)’12. 25--32.Google Scholar
- James Hoffman and Baskaran Subramaniam. 1995. The role of visual attention in saccadic eye movements. Perception & Psychophysics 57, 6, 787--795.Google Scholar
- Kenneth Holmqvist, Marcus Nyström, Richard Andersson, Richard Dewhurst, Halszka Jarodzka, and Joost van de Weijer. 2010. Eye Tracking: A Comprehensive Guide to Methods and Measures. Oxford University Press, New York, NY.Google Scholar
- Laurent Itti. 2005. Quantifying the contribution of low-level saliency to human eye movements in dynamic scenes. Visual Cognition 12, 1093--1123.Google ScholarCross Ref
- Laurent Itti and Nitin Dhavale. 2003. Realistic avatar eye and head animation using a neurobiological model of visual attention. In Proceedings of SPIE. SPIE Press, 64--78.Google Scholar
- Laurent Itti and Christof Koch. 2013. Bottom-Up Visual Attention Home Page; The Interactive Demo. Retrieved April 23, 2015 from http://ilab.usc.edu/bu/javaDemo/ie/demo.html.Google Scholar
- Laurent Itti, Christof Koch, and Ernst Niebur. 1998. A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis Machine Intelligence 20, 11, 1254--1259. Google ScholarDigital Library
- Hector Jasso and Jochen Triesch. 2007. Learning to attend—from bottom-up to top-down. In Attention in Cognitive Systems. Theories and Systems from an Interdisciplinary Viewpoint, Lucas Paletta and Erich Rome (Eds.). Lecture Notes in Computer Science, Vol. 4840. Springer, Berlin, 106--122. Google ScholarDigital Library
- Tilke Judd, Krista Ehinger, Frédo Durand, and Antonio Torralba. 2009. Learning to predict where humans look. In Proceedings of the IEEE 12th International Conference on Computer Vision. IEEE, 2106--2113.Google ScholarCross Ref
- Wolf Kienzle, Bernhard Schölkopf, Felix A. Wichmann, and Matthias O. Franz. 2007. How to find interesting locations in video: A spatiotemporal interest point detector learned from human eye movements. In Proceedings of the 29th DAGM Conference on Pattern Recognition. Springer-Verlag, Berlin, 405--414. Google ScholarDigital Library
- Krystian Mikolajczyk, Tinne Tuytelaars, Cordelia Schmid, Andrew Zisserman, Jiri Matas, Frederik Schaffalitzky, Timor Kadir, and Luc van Gool. 2005. A comparison of affine region detectors. International Journal of Computer Vision 65, 1/2 43--72. Google ScholarDigital Library
- Frieder Nake. 1974. Ästhetik als Informationsverarbeitung. Springer Verlag, Wien, New York.Google Scholar
- Sunaad Nataraju, Vineeth Balasubramanian, and Sethuraman Panchanathan. 2009. Learning attention based saliency in videos from human eye movements. In Proceedings of the 2009 International Conference on Motion and Video Computing (WMVC’09). IEEE Computer Society, Washington, DC, 134--139. Google ScholarDigital Library
- Yaqing Niu, Rebecca M. Todd, Matthew Kyan, and Adam K. Anderson. 2012. Visual and emotional salience influence eye movements. ACM Transactions on Applied Perceptions 9, 3, Article 13, 18 pages. Google ScholarDigital Library
- Robert J. Peters and Laurent Itti. 2007. Beyond bottom-up: Incorporating task-dependent influences into a computational model of spatial attention. In IEEE Conference on Computer Vision and Pattern Recognition, 2007 (CVPR’07). 1--8.Google Scholar
- Claudio M. Privitera and Lawrence W. Stark. 2000. Algorithms for defining visual regions-of-interest: Comparison with eye fixations. IEEE Transactions on Pattern Analysis and Machine Intelligence 22, 9, 970--982. Google ScholarDigital Library
- Umesh Rajashekar, Ian van der Linde, Alan C. Bovik, and Lawrence K. Cormack. 2008. GAFFE: A gaze-attentive fixation finding engine. IEEE Transactions on Image Processing 17, 564--573. Google ScholarDigital Library
- Robert W. Rodieck. 1998. The First Steps in Seeing. Sinauer Associates, Inc., Sunderland, MA.Google Scholar
- Jukka Saarinen. 1993. Shifts of visual attention at fixation and away from fixation. Vision Research 33, 8, 1113--1117.Google ScholarCross Ref
- Paul Viola and Michael J. Jones. 2004. Robust real-time face detection. International Journal of Computer Vision 57, 2, 137--154. Google ScholarDigital Library
- Jeremy M. Wolfe. 1994. Guided search 2.0: A revised model of visual search. Psychonomic Bulletin & Review 1, 2, 202--238.Google ScholarCross Ref
- Qi Zhao and C. Koch. 2011. Learning visual saliency. In Proceedings of the 2011 45th Annual Conference on Information Sciences and Systems (CISS’11). 1--6.Google Scholar
Index Terms
- Simulating Fixations When Looking at Visual Arts
Recommendations
Simulating Human Visual System Based on Vision Transformer
SUI '23: Proceedings of the 2023 ACM Symposium on Spatial User InteractionThe human visual system (HVS) is capable of responding in real-time to complex visual environments. During the process of freely observing visual scenes, predicting eye movements and visual fixations is a task known as scanpath prediction, which aims ...
Human Visual Scanpath Prediction Based on RGB-D Saliency
ICIGP '18: Proceedings of the 2018 International Conference on Image and Graphics ProcessingHuman visual perception is considered as a dynamic process of information acquisition, while the visual scanpath can clearly reflect the shift of our eye fixations. In the previous study of visual attention, researchers generally do the saliency ...
Estimation of Visual Attention using Microsaccades in response to Vibrations in the Peripheral Field of Vision
ETRA '21 Short Papers: ACM Symposium on Eye Tracking Research and ApplicationsViewer’s eye movements and behavioural responses were analysed in order to determine the relationship between selective perception and visual attention during a dual detection task in the central and peripheral fields of vision, in order to design a ...
Comments