Results 1 - 10
of
641
Saliency detection: A spectral residual approach
- In IEEE Conference on Computer Vision and Pattern Recognition (CVPR07). IEEE Computer Society
, 2007
"... The ability of human visual system to detect visual saliency is extraordinarily fast and reliable. However, computational modeling of this basic intelligent behavior still remains a challenge. This paper presents a simple method for the visual saliency detection. Our model is independent of features ..."
Abstract
-
Cited by 335 (10 self)
- Add to MetaCart
(Show Context)
The ability of human visual system to detect visual saliency is extraordinarily fast and reliable. However, computational modeling of this basic intelligent behavior still remains a challenge. This paper presents a simple method for the visual saliency detection. Our model is independent of features, categories, or other forms of prior knowledge of the objects. By analyzing the log-spectrum of an input image, we extract the spectral residual of an image in spectral domain, and propose a fast method to construct the corresponding saliency map in spatial domain. We test this model on both natural pictures and artificial images such as psychological patterns. The result indicate fast and robust saliency detection of our method. 1.
Graph-based visual saliency
- ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 19
, 2007
"... A new bottom-up visual saliency model, Graph-Based Visual Saliency (GBVS), is proposed. It consists of two steps: rst forming activation maps on certain feature channels, and then normalizing them in a way which highlights conspicuity and admits combination with other maps. The model is simple, and ..."
Abstract
-
Cited by 277 (6 self)
- Add to MetaCart
(Show Context)
A new bottom-up visual saliency model, Graph-Based Visual Saliency (GBVS), is proposed. It consists of two steps: rst forming activation maps on certain feature channels, and then normalizing them in a way which highlights conspicuity and admits combination with other maps. The model is simple, and biologically plausible insofar as it is naturally parallelized. This model powerfully predicts human xations on 749 variations of 108 natural images, achieving 98 % of the ROC area of a human-based control, whereas the classical algorithms of Itti & Koch ([2], [3], [4]) achieve only 84%.
Learning to Predict Where Humans Look
"... For many applications in graphics, design, and human computer interaction, it is essential to understand where humans look in a scene. Where eye tracking devices are not a viable option, models of saliency can be used to predict fixation locations. Most saliency approaches are based on bottom-up com ..."
Abstract
-
Cited by 211 (4 self)
- Add to MetaCart
(Show Context)
For many applications in graphics, design, and human computer interaction, it is essential to understand where humans look in a scene. Where eye tracking devices are not a viable option, models of saliency can be used to predict fixation locations. Most saliency approaches are based on bottom-up computation that does not consider top-down image semantics and often does not match actual eye movements. To address this problem, we collected eye tracking data of 15 viewers on 1003 images and use this database as training and testing examples to learn a model of saliency based on low, middle and high-level image features. This large database of eye tracking data is publicly available with this paper. 1.
Visual correlates of fixation selection: effects of scale and time
, 2005
"... What distinguishes the locations that we fixate from those that we do not? To answer this question we recorded eye movements while observers viewed natural scenes, and recorded image characteristics centred at the locations that observers fixated. To investigate potential differences in the visual c ..."
Abstract
-
Cited by 146 (5 self)
- Add to MetaCart
What distinguishes the locations that we fixate from those that we do not? To answer this question we recorded eye movements while observers viewed natural scenes, and recorded image characteristics centred at the locations that observers fixated. To investigate potential differences in the visual characteristics of fixated versus non-fixated locations, these images were transformed to make intensity, contrast, colour, and edge content explicit. Signal detection and information theoretic techniques were then used to compare fixated regions to those that were not. The presence of contrast and edge information was more strongly discriminatory than luminance or chromaticity. Fixated locations tended to be more distinctive in the high spatial frequencies. Extremes of low frequency luminance information were avoided. With prolonged viewing, consistency in fixation locations between observers decreased. In contrast to [Parkhurst, D. J., Law, K., & Niebur, E. (2002). Modeling the role of salience in the allocation of overt visual attention. Vision Research, 42 (1), 107–123] we found no change in the involvement of image features over time. We attribute this difference in our results to a systematic bias in their metric. We propose that saccade target selection involves an unchanging intermediate level representation of the scene but that the high-level interpretation of this representation changes over time.
Sun: A Bayesian framework for saliency using natural statistics
- Journal of Vision
, 2008
"... We propose a definition of saliency by considering what the visual system is trying to optimize when directing attention. The resulting model is a Bayesian framework from which bottom-up saliency emerges naturally as the self-information of visual features, and overall saliency (incorporating top-do ..."
Abstract
-
Cited by 143 (4 self)
- Add to MetaCart
(Show Context)
We propose a definition of saliency by considering what the visual system is trying to optimize when directing attention. The resulting model is a Bayesian framework from which bottom-up saliency emerges naturally as the self-information of visual features, and overall saliency (incorporating top-down information with bottom-up saliency) emerges as the pointwise mutual information between the features and the target when searching for a target. An implementation of our framework demonstrates that our model’s bottom-up saliency maps perform as well as or better than existing algorithms in predicting people’s fixations in free viewing. Unlike existing saliency measures, which depend on the statistics of the particular image being viewed, our measure of saliency is derived from natural image statistics, obtained in advance from a collection of natural images. For this reason, we call our model SUN (Saliency Using Natural statistics). A measure of saliency based on natural image statistics, rather than based on a single test image, provides a straightforward explanation for many search asymmetries observed in humans; the statistics of a single test image lead to predictions that are not consistent with these asymmetries. In our model, saliency is computed locally, which is consistent with the neuroanatomy of the early visual system and results in an efficient algorithm with few free parameters.
The jackknife—a review.
- Biometrika
, 1974
"... The Light Beyond, By Raymond A. Moody, Jr. with Paul Perry. New York, NY: Bantam Books, 1988, 161 pp., $18.95 In his foreword to this book, Andrew Greeley, a prominent priest and sociologist, introduces his comments with the following statement: "Raymond Moody has achieved a rare feat in th ..."
Abstract
-
Cited by 104 (0 self)
- Add to MetaCart
(Show Context)
The Light Beyond, By Raymond A. Moody, Jr. with Paul Perry. New York, NY: Bantam Books, 1988, 161 pp., $18.95 In his foreword to this book, Andrew Greeley, a prominent priest and sociologist, introduces his comments with the following statement: "Raymond Moody has achieved a rare feat in the quest for human knowledge; he has created a paradigm." He then refers to Thomas Kuhn, who pointed out in The Structure of Scientific Revolutions that scientific revolutions occur when someone creates a new perspective, a new model, a new approach to reality. Although Greeley acknowledges that Moody did not discover the near-death experience (NDE), he contends that because Moody put a name to it in his previous bestseller Life After Life (1975), he therefore deserves credit for the new para digm that has evolved. Greeley then refers to The Light Beyond as characterized by Moody's "openness, sensitivity and modesty." This he attributes to Moody's acknowledgement that the NDE does not repre sent proof of life after death; rather, it indicates only the existence and widespread prevalence of the NDE. I must question why Greeley does not comment more on the content of the book, and why Moody felt it was appropriate to be credited with creating a new paradigm. During the last fourteen years since Life
Spatiotemporal Sensitivity and Visual Attention for Efficient Rendering of Dynamic Environments
, 2001
"... INTRODUCTION Global illumination is the physically accurate calculation of lighting in an environment. It is computationally expensive for static environments and even more so for dynamic environments. Not only are many images required for an animation, but the calculation involved increases with th ..."
Abstract
-
Cited by 97 (1 self)
- Add to MetaCart
(Show Context)
INTRODUCTION Global illumination is the physically accurate calculation of lighting in an environment. It is computationally expensive for static environments and even more so for dynamic environments. Not only are many images required for an animation, but the calculation involved increases with the presence of moving objects. In static environments, global illumination algorithms can precompute a lighting solution and reuse it whenever the viewpoint changes, but in dynamic environments, any moving object or light potentially affects the illumination of every other object in a scene. To guarantee accuracy, the algorithm has to recompute the entire lighting solution for each frame. This paper describes a perceptually-based technique that can dramatically reduce this computational load. The technique may also be used in image based rendering, geometry level of detail selection, realistic image synthesis, video telephony and video compression. Perceptually-based rendering operat
Changing your mind: On the contributions of top-down and bottom-up guidance in visual search for feature singletons
- Journal of Experimental Psychology: Human Perception and Performance
, 2003
"... Observers, searching for targets among distractor items, guide attention with a mix of top-down information—based on observers ’ knowledge—and bottom-up information—stimulus-based and largely independent of that knowledge. There are 2 types of top-down guidance: explicit information (e.g., verbal de ..."
Abstract
-
Cited by 91 (0 self)
- Add to MetaCart
(Show Context)
Observers, searching for targets among distractor items, guide attention with a mix of top-down information—based on observers ’ knowledge—and bottom-up information—stimulus-based and largely independent of that knowledge. There are 2 types of top-down guidance: explicit information (e.g., verbal description) and implicit priming by preceding targets (top-down because it implies knowledge of previous searches). Experiments 1 and 2 separate bottom-up and top-down contributions to singleton search. Experiment 3 shows that priming effects are based more strongly on target than on distractor identity. Experiments 4 and 5 show that more difficult search for one type of target (color) can impair search for other types (size, orientation). Experiment 6 shows that priming guides attention and does not just modulate response. When you look at Figure 1, your attention is probably attracted to the spiky diamond. It is a salient item, and, all else being equal, salient items that are different from their neighbors tend to attract attention (Egeth, 1977; Julesz, 1986; Moraglia, 1989). The information that guided your attention to that item can be labeled as bottom-up—meaning that it did not depend on the observer’s
Quantifying the contribution of low-level saliency to human eye movements in dynamic scenes
- Visual Cognition
, 2005
"... in dynamic scenes ..."
(Show Context)