• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Spatially coherent latent topic model for concurrent segmentation and classification of objects and scenes. (2007)

by L Cao, L Fei-Fei
Venue:In ICCV 2007,
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 150
Next 10 →

Simultaneous Image Classification and Annotation

by Chong Wang, David Blei, Li Fei-fei
"... Image classification and annotation are important problems in computer vision, but rarely considered together. Intuitively, annotations provide evidence for the class label, and the class label provides evidence for annotations. For example, an image of class highway is more likely annotated with wo ..."
Abstract - Cited by 148 (7 self) - Add to MetaCart
Image classification and annotation are important problems in computer vision, but rarely considered together. Intuitively, annotations provide evidence for the class label, and the class label provides evidence for annotations. For example, an image of class highway is more likely annotated with words “road, ” “car, ” and “traffic ” than words “fish, ” “boat, ” and “scuba. ” In this paper, we develop a new probabilistic model for jointly modeling the image, its class label, and its annotations. Our model treats the class label as a global description of the image, and treats annotation terms as local descriptions of parts of the image. Its underlying probabilistic assumptions naturally integrate these two sources of information. We derive an approximate inference and estimation algorithms based on variational methods, as well as efficient approximations for classifying and annotating new images. We examine the performance of our model on two real-world image data sets, illustrating that a single model provides competitive annotation performance, and superior classification performance. 1.
(Show Context)

Citation Context

...7, 26, 28, 17], and both discriminative and generative techniques have been applied to this problem. Discriminative methods include the work in [7, 30, 29, 16]. Generative methods include the work in =-=[9, 6, 22, 17]-=-. In the work of [5], the authors combine generative models for latent topic discovery [11] and discriminative methods for classification (knearest neighbors). LDA-based image classification was intro...

Weakly supervised discriminative localization and classification: a joint learning process

by Minh Hoai Nguyen, Lorenzo Torresani, Fernando de la Torre, Carsten Rother , 2009
"... ..."
Abstract - Cited by 63 (5 self) - Add to MetaCart
Abstract not found

Localizing Objects with Smart Dictionaries

by Brian Fulkerson, Andrea Vedaldi, Stefano Soatto - Proceedings of European Conference on Computer Vision
"... Abstract. We present an approach to determine the category and location of objects in images. It performs very fast categorization of each pixel in an image, a brute-force approach made feasible by three key developments: First, our method reduces the size of a large generic dictionary (on the order ..."
Abstract - Cited by 62 (3 self) - Add to MetaCart
Abstract. We present an approach to determine the category and location of objects in images. It performs very fast categorization of each pixel in an image, a brute-force approach made feasible by three key developments: First, our method reduces the size of a large generic dictionary (on the order of ten thousand words) to the low hundreds while increasing classification performance compared to k-means. This is achieved by creating a discriminative dictionary tailored to the task by following the information bottleneck principle. Second, we perform feature-based categorization efficiently on a dense grid by extending the concept of integral images to the computation of local histograms. Third, we compute SIFT descriptors densely in linear time. We compare our method to the state of the art and find that it excels in accuracy and simplicity, performing better while assuming less. 1
(Show Context)

Citation Context

...nent into their model. Approaches are varied, but broadly tend to include one of the following: interactions between pairs of features [11,12,13], absolute position of the features [14], segmentation =-=[15,16]-=-, or a learned shape or parts model of the objects [4,17,18]. Our method exploits interaction between groups of features (all features in theLocalizing Objects with Smart Dictionaries 181 window), bu...

An efficient algorithm for Co-segmentation

by Dorit S. Hochbaum, Vikas Singh , 2009
"... This paper is focused on the Co-segmentation problem [1] – where the objective is to segment a similar object from a pair of images. The background in the two images may be arbitrary; therefore, simultaneous segmentation of both images must be performed with a requirement that the appearance of the ..."
Abstract - Cited by 57 (4 self) - Add to MetaCart
This paper is focused on the Co-segmentation problem [1] – where the objective is to segment a similar object from a pair of images. The background in the two images may be arbitrary; therefore, simultaneous segmentation of both images must be performed with a requirement that the appearance of the two sets of foreground pixels in the respective images are consistent. Existing approaches [1, 2] cast this problem as a Markov Random Field (MRF) based segmentation of the image pair with a regularized difference of the two histograms – assuming a Gaussian prior on the foreground appearance [1] or by calculating the sum of squared differences [2]. Both are interesting formulations but lead to difficult optimization problems, due to the presence of the second (histogram difference) term. The model proposed here bypasses measurement of the histogram differences in a direct fashion; we show that this enables obtaining efficient solutions to the underlying optimization model. Our new algorithm is similar to the existing methods in spirit, but differs substantially in that it can be solved to optimality in polynomial time using a maximum flow procedure on an appropriately constructed graph. We discuss our ideas and present promising experimental results.
(Show Context)

Citation Context

... identification of similar objects in more than one image is a fundamental problem in computer vision and has relied on user annotation or construction of models [8, 9]. A number of recent techniques =-=[1, 10, 11, 12]-=-, however, have preferred an unsupervised (or semi-supervised) approach to the problem and obtained good overall performance. Cosegmentation belongs to this second category. The key idea adopted in [1...

Object cosegmentation

by Sara Vicente, Carsten Rother, Vladimir Kolmogorov - In CVPR , 2011
"... Cosegmentation is typically defined as the task of jointly segmenting “something similar ” in a given set of images. Existing methods are too generic and so far have not demonstrated competitive results for any specific task. In this paper we overcome this limitation by adding two new aspects to cos ..."
Abstract - Cited by 56 (0 self) - Add to MetaCart
Cosegmentation is typically defined as the task of jointly segmenting “something similar ” in a given set of images. Existing methods are too generic and so far have not demonstrated competitive results for any specific task. In this paper we overcome this limitation by adding two new aspects to cosegmentation: (1) the “something ” has to be an object, and (2) the “similarity ” measure is learned. In this way, we are able to achieve excellent results on the recently introduced iCoseg dataset, which contains small sets of images of either the same object instance or similar objects of the same class. The challenge of this dataset lies in the extreme changes in viewpoint, lighting, and object deformations within each set. We are able to considerably outperform several competitors. To achieve this performance, we borrow recent ideas from object recognition: the use of powerful features extracted from a pool of candidate objectlike segmentations. We believe that our work will be beneficial to several application areas, such as image retrieval. 1.
(Show Context)

Citation Context

...odels have been proposed for the task of Unsupervised class segmentation. Examples are the LOCUS model [24], which learns a shape model for the class, and the use of a topic model over image segments =-=[10]-=-, assigning segments to topics depending on their visual words. Both [10, 24] model separately what is common in all images (the shape of the object, sift features) and what is image specific (the app...

Unsupervised modeling of object categories using link analysis techniques

by Gunhee Kim, Christos Faloutsos, Martial Hebert - In CVPR , 2008
"... We propose an approach for learning visual models of object categories in an unsupervised manner in which we first build a large-scale complex network which captures the interactions of all unit visual features across the entire training set and we infer information, such as which features are in wh ..."
Abstract - Cited by 54 (5 self) - Add to MetaCart
We propose an approach for learning visual models of object categories in an unsupervised manner in which we first build a large-scale complex network which captures the interactions of all unit visual features across the entire training set and we infer information, such as which features are in which categories, directly from the graph by using link analysis techniques. The link analysis techniques are based on well-established graph mining techniques used in diverse applications such as WWW, bioinformatics, and social networks. The techniques operate directly on the patterns of connections between features in the graph rather than on statistical properties, e.g., from clustering in feature space. We argue that the resulting techniques are simpler, and we show that they perform similarly or better compared to state of the art techniques on common data sets. We also show results on more challenging data sets than those that have been used in prior work on unsupervised modeling.
(Show Context)

Citation Context

...nments to instances of local features are in themselves challenging. For this reason, there is no dominant methods for dictionary formation (e.g. hierarchical agglomerative clustering [19] or k-means =-=[6, 21]-=-), the optimal selection of dictionary sizes, and the assignment of codewords to each feature instance (e.g. soft or hard assignment). In contrast, we do not try to identify each visual entity (i.e. t...

Localizing objects while learning their appearance

by Thomas Deselaers, Bogdan Alexe, Vittorio Ferrari - In ECCV , 2010
"... Abstract. Learning a new object class from cluttered training images is very challenging when the location of object instances is unknown. Previous works generally require objects covering a large portion of the images. We present a novel approach that can cope with extensive clutter as well as larg ..."
Abstract - Cited by 51 (10 self) - Add to MetaCart
Abstract. Learning a new object class from cluttered training images is very challenging when the location of object instances is unknown. Previous works generally require objects covering a large portion of the images. We present a novel approach that can cope with extensive clutter as well as large scale and appearance variations between object instances. To make this possible we propose a conditional random field that starts from generic knowledge and then progressively adapts to the new class. Our approach simultaneously localizes object instances while learning an appearance model specific for the class. We demonstrate this on the challenging Pascal VOC 2007 dataset. Furthermore, our method enables to train any state-of-the-art object detector in a weakly supervised fashion, although it would normally require object location annotations. 1
(Show Context)

Citation Context

...tion, such as partbased [2, 3], segmentation-based [1, 4–6, 11, 15], and others [7, 12, 16]. However, most methods have been demonstrated on datasets such as Caltech4 [1–4,6,7,16] and Weizmann horses =-=[6,10,15]-=-, where objects are rather centered and occupy a large portion of the image, there is little scale/viewpoint variation, and limited background clutter. This is due to the difficulty of spotting the re...

Half-integrality based algorithms for cosegmentation of images

by Lopamudra Mukherjee, Vikas Singh, Charles R. Dyer - In CVPR , 2009
"... We study the cosegmentation problem where the objective is to segment the same object (i.e., region) from a pair of images. The segmentation for each image can be cast using a partitioning/segmentation function with an additional constraint that seeks to make the histograms of the segmented regions ..."
Abstract - Cited by 49 (4 self) - Add to MetaCart
We study the cosegmentation problem where the objective is to segment the same object (i.e., region) from a pair of images. The segmentation for each image can be cast using a partitioning/segmentation function with an additional constraint that seeks to make the histograms of the segmented regions (based on intensity and texture features) similar. Using Markov Random Field (MRF) energy terms for the simultaneous segmentation of the images together with histogram consistency requirements using the squared L2 (rather than L1) distance, after linearization and adjustments, yields an optimization model with some interesting combinatorial properties. We discuss these properties which are closely related to certain relaxation strategies recently introduced in computer vision. Finally, we show experimental results of the proposed approach. 1.
(Show Context)

Citation Context

...multaneously segmenting a person or object of interest from an image pair. The idea has since found applications in segmentation of videos [2] and shown to be useful in several other problems as well =-=[3; 4]-=-. The model [1] nicely captures the setting where a pair of images have very little in common except the foreground. Notice how the calculation of image to image distances (based on the entire image) ...

Spatial Latent Dirichlet Allocation

by Xiaogang Wang, Eric Grimson , 2007
"... In recent years, the language model Latent Dirichlet Allocation (LDA), which clusters co-occurring words into topics, has been widely applied in the computer vision field. However, many of these applications have difficulty with modeling the spatial and temporal structure among visual words, since L ..."
Abstract - Cited by 46 (3 self) - Add to MetaCart
In recent years, the language model Latent Dirichlet Allocation (LDA), which clusters co-occurring words into topics, has been widely applied in the computer vision field. However, many of these applications have difficulty with modeling the spatial and temporal structure among visual words, since LDA assumes that a document is a “bag-of-words”. It is also critical to properly design “words ” and “documents ” when using a language model to solve vision problems. In this paper, we propose a topic model Spatial Latent Dirichlet Allocation (SLDA), which better encodes spatial structures among visual words that are essential for solving many vision problems. The spatial information is not encoded in the values of visual words but in the design of documents. Instead of knowing the partition of words into documents a priori, the word-document assignment becomes a random hidden variable in SLDA. There is a generative procedure, where knowledge of spatial structure can be flexibly added as a prior, grouping visual words which are close in space into the same document. We use SLDA to discover objects from a collection of images, and show it achieves better performance than LDA. 1
(Show Context)

Citation Context

...l which clusters co-occurring words into topics. In recent years, LDA has been widely used to solve computer vision problems. For example, LDA was used to discover objects from a collection of images =-=[2, 3, 4]-=- and to classify images into different scene categories [5]. [6] employed LDA to classify human actions. In visual surveillance, LDA was used to model atomic activities and interactions in a crowded s...

Bicos: A bi-level co-segmentation method for image classification

by Yuning Chai, Victor Lempitsky, Andrew Zisserman - In Proc. ICCV , 2011
"... The objective of this paper is the unsupervised segmentation of image training sets into foreground and background in order to improve image classification performance. To this end we introduce a new scalable, alternation-based algorithm for co-segmentation, BiCoS, which is simpler than many of its ..."
Abstract - Cited by 42 (2 self) - Add to MetaCart
The objective of this paper is the unsupervised segmentation of image training sets into foreground and background in order to improve image classification performance. To this end we introduce a new scalable, alternation-based algorithm for co-segmentation, BiCoS, which is simpler than many of its predecessors, and yet has superior performance on standard benchmark image datasets. We argue that the reason for this success is that the cosegmentation task is represented at the appropriate levels – pixels and color distributions for individual images, and super-pixels with learnable features at the level of sharing across the image set – together with powerful and efficient inference algorithms (GrabCut and SVM) for each level. We assess both the segmentation and classification performance of the algorithm and compare to previous results
(Show Context)

Citation Context

...in their dimensions, generative modeling and classification of foreground and background distributions would be problematic, although one can use naive-Bayes approximation [3] as well as topic models =-=[9, 32]-=-. Implementation details. We use the popular Felzenswalb’s code for superpixel segmentation [12], with sigma=1, k=200, minSize=640. Parameter values were chosen on another, but similar, problem, and a...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University