Results 1 - 10
of
83
A survey of content-based image retrieval with high-level semantics
, 2007
"... In order to improve the retrieval accuracy of content-based image retrieval systems, research focus has been shifted from designing sophisticated low-level feature extraction algorithms to reducing the ‘semantic gap ’ between the visual features and the richness of human semantics. This paper attemp ..."
Abstract
-
Cited by 150 (5 self)
- Add to MetaCart
In order to improve the retrieval accuracy of content-based image retrieval systems, research focus has been shifted from designing sophisticated low-level feature extraction algorithms to reducing the ‘semantic gap ’ between the visual features and the richness of human semantics. This paper attempts to provide a comprehensive survey of the recent technical achievements in high-level semantic-based image retrieval. Major recent publications are included in this survey covering different aspects of the research in this area, including low-level image feature extraction, similarity measurement, and deriving high-level semantic features. We identify five major categories of the state-of-the-art techniques in narrowing down the ‘semantic gap’: (1) using object ontology to define high-level concepts; (2) using machine learning methods to associate low-level features with query concepts; (3) using relevance feedback to learn users’ intention; (4) generating semantic template to support high-level image retrieval; (5) fusing the evidences from HTML text and the visual content of images for WWW image retrieval. In addition, some other related issues such as image test bed and retrieval performance evaluation are also discussed. Finally, based on existing technology and the demand from real-world applications, a few promising future research directions are suggested.
Image-based human age estimation by manifold learning and locally adjusted robust regression
- IEEE Transactions on Image Processing
, 2008
"... Abstract—Estimating human age automatically via facial image analysis has lots of potential real-world applications, such as human computer interaction and multimedia communication. However, it is still a challenging problem for the existing computer vision systems to automatically and effectively e ..."
Abstract
-
Cited by 61 (5 self)
- Add to MetaCart
(Show Context)
Abstract—Estimating human age automatically via facial image analysis has lots of potential real-world applications, such as human computer interaction and multimedia communication. However, it is still a challenging problem for the existing computer vision systems to automatically and effectively estimate human ages. The aging process is determined by not only the person’s gene, but also many external factors, such as health, living style, living location, and weather conditions. Males and females may also age differently. The current age estimation performance is still not good enough for practical use and more effort has to be put into this research direction. In this paper, we introduce the age manifold learning scheme for extracting face aging features and design a locally adjusted robust regressor for learning and prediction of human ages. The novel approach improves the age estimation accuracy significantly over all previous methods. The merit of the proposed approaches for image-based age estimation is shown by extensive experiments on a large internal age database and the public available FG-NET database. Index Terms—Age manifold, human age estimation, locally adjusted robust regression, manifold learning, nonlinear regression, support vector machine (SVM), support vector regression (SVR). I.
Learning From Examples in the Small Sample Case: Face Expression Recognition
, 2005
"... Example-based learning for computer vision can be difficult when a large number of examples to represent each pattern or object class is not available. In such situations, learning from a small number of samples is of practical value. To study this issue, the task of face expression recognition with ..."
Abstract
-
Cited by 36 (2 self)
- Add to MetaCart
Example-based learning for computer vision can be difficult when a large number of examples to represent each pattern or object class is not available. In such situations, learning from a small number of samples is of practical value. To study this issue, the task of face expression recognition with a small number of training images of each expression is considered. A new technique based on linear programming for both feature selection and classifier training is introduced. A pairwise framework for feature selection, instead of using all classes simultaneously, is presented. Experimental results compare the method with three others: a simplified Bayes classifier, support vector machine, and AdaBoost. Finally, each algorithm is analyzed and a new categorization of these algorithms is given, especially for learning from examples in the small sample case.
An Ontology Approach to Object-Based Image Retrieval
- In Proc. IEEE Int. Conf. on Image Processing (ICIP03
, 2003
"... In this paper, an image retrieval methodology suited for search in large collections of heterogeneous images is presented. The proposed approach employs a fully unsupervised segmentation algorithm to divide images into regions. Low-level features describing the color, position, size and shape of the ..."
Abstract
-
Cited by 36 (6 self)
- Add to MetaCart
(Show Context)
In this paper, an image retrieval methodology suited for search in large collections of heterogeneous images is presented. The proposed approach employs a fully unsupervised segmentation algorithm to divide images into regions. Low-level features describing the color, position, size and shape of the resulting regions are extracted and are automatically mapped to appropriate intermediatelevel descriptors forming a simple vocabulary termed object ontology. The object ontology is used to allow the qualitative definition of the high-level concepts the user queries for (semantic objects, each represented by a keyword) in a human-centered fashion. When querying, clearly irrelevant image regions are rejected using the intermediate-level descriptors; following that, a relevance feedback mechanism employing the low-level features is invoked to produce the final query results. The proposed approach bridges the gap between keyword-based approaches, which assume the existence of rich image captions or require manual evaluation and annotation of every image of the collection, and query-by-example approaches, which assume that the user queries for images similar to one that already is at his disposal.
Direct kernel biased discriminant analysis: a new content-based image retrieval relevance feedback algorithm
- IEEE TRANSACTIONS ON MULTIMEDIA 8 (4) 716-727
, 2006
"... ..."
Retrieval of images from artistic repositories using a decision fusion framework
- IEEE Trans. on Image Proc
"... Abstract—The large volumes of artistic visual data available to museums, art galleries, and online collections motivate the need for effective means to retrieve relevant information from such repositories. This paper proposes a decision making framework for content-based retrieval of art images base ..."
Abstract
-
Cited by 18 (0 self)
- Add to MetaCart
Abstract—The large volumes of artistic visual data available to museums, art galleries, and online collections motivate the need for effective means to retrieve relevant information from such repositories. This paper proposes a decision making framework for content-based retrieval of art images based on a combination of lowlevel features. Traditionally, the similarity among two images has been calculated as a weighted distance between two feature vectors. This approach, however, may not be mathematically and computationally appropriate and does not provide enough flexibility in modeling user queries. This paper proposes a framework that generalizes a wide set of previous approaches to similarity calculation including the weighted distance approach. In this framework, image similarities are obtained through a decision making process based on low-level feature distances using fuzzy theory. The analysis and results of this paper indicate that the aggregation technique presented here provides an effective, general, and flexible tool for similarity calculation based on the combination of individual descriptors and features. Index Terms—Content-based image retrieval, feature combination, fuzzy aggregation operators, MPEG-7 visual descriptors, similarity calculations. I.
Which components are important for interactive image searching?
- JOURNAL OF EMERGING TECHNOLOGY AND ADVANCED ENGINEERING WEBSITE: WWW.IJETAE.COM (ISSN 2250-2459, ISO 9001:2008 CERTIFIED JOURNAL, VOLUME 3, ISSUE 3
, 2008
"... With many potential industrial applications, con-tent-based image retrieval (CBIR) has recently gained more attention for image management and web searching. As an im-portant tool to capture usersâ preferences and thus to improve the performance of CBIR systems, a variety of relevance feedback (R ..."
Abstract
-
Cited by 17 (0 self)
- Add to MetaCart
(Show Context)
With many potential industrial applications, con-tent-based image retrieval (CBIR) has recently gained more attention for image management and web searching. As an im-portant tool to capture usersâ preferences and thus to improve the performance of CBIR systems, a variety of relevance feedback (RF) schemes have been developed in recent years. One key issue in RF is: which features (or feature dimensions) can benefit this human-computer iteration procedure? In this paper, we make theoretical and practical comparisons between principal and com-plement components of image features in CBIR RF. Most of the previous RF approaches treat the positive and negative feedbacks equivalently although this assumption is not appropriate since the two groups of training feedbacks have very different properties. That is, all positive feedbacks share a homogeneous concept while negative feedbacks do not. We explore solutions to this important problem by proposing an orthogonal complement component analysis. Experimental results are reported on a real-world image collection to demonstrate that the proposed complement compo-nents method consistently outperforms the conventional principal components method in both linear and kernel spaces when users want to retrieve images with a homogeneous concept.
Unified Framework for Fast Exact and Approximate Search in Dissimilarity Spaces
, 2007
"... In multimedia systems we usually need to retrieve DB objects based on their similarity to a query object, while the similarity assessment is provided by a measure which defines a (dis)similarity score for every pair of DB objects. In most existing applications, the similarity measure is required to ..."
Abstract
-
Cited by 16 (5 self)
- Add to MetaCart
(Show Context)
In multimedia systems we usually need to retrieve DB objects based on their similarity to a query object, while the similarity assessment is provided by a measure which defines a (dis)similarity score for every pair of DB objects. In most existing applications, the similarity measure is required to be a metric, where the triangle inequality is utilized to speedup the search for relevant objects by use of metric access methods (MAMs), e.g. the M-tree. A recent research has shown, however, that non-metric measures are more appropriate for similarity modeling due to their robustness and ease to model a made-to-measure similarity. Unfortunately, due to the lack of triangle inequality, the non-metric measures cannot be directly utilized by MAMs. From another point of view, some sophisticated similarity measures could be available in a black-box non-analytic form (e.g. as an algorithm or even a hardware device), where no information about their topological properties is provided, so we have to consider them as non-metric measures as well. From yet another point of view, the concept of similarity measuring itself is inherently imprecise and we often prefer fast but approximate retrieval over an exact but slower one. To date, the mentioned aspects of similarity retrieval have been solved separately, i.e. exact vs. approximate search or metric vs. non-metric search. In this paper we introduce a similarity retrieval framework which incorporates both of the aspects into a single unified model. Based on the framework, we show that for any dissimilarity measure (either a metric or non-metric) we are able to change the ”amount ” of triangle inequality, and so to obtain an approximate or full metric which can be used for MAM-based retrieval. Due to the varying ”amoun ” of triangle inequality, the measure is modified in a way suitable for either an exact but slower or an approximate but faster retrieval. Additionally, we introduce the TriGen algorithm aimed to construct the desired modification of any black-box distance automatically, using just a small fraction of the database.
Random sampling based SVM for relevance feedback image retrieval
- In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
, 2004
"... Relevance feedback (RF) schemes based on support vector machine (SVM) have been widely used in content-based image retrieval. However, the performance of SVM based RF is often poor when the number of labeled positive feedback samples is small. This is mainly due to three reasons: 1. SVM classifier i ..."
Abstract
-
Cited by 16 (1 self)
- Add to MetaCart
Relevance feedback (RF) schemes based on support vector machine (SVM) have been widely used in content-based image retrieval. However, the performance of SVM based RF is often poor when the number of labeled positive feedback samples is small. This is mainly due to three reasons: 1. SVM classifier is unstable on small size training set; 2. SVM’s optimal hyper-plane may be biased when the positive feedback samples are much less than the negative feedback samples; 3. overfitting due to that the feature dimension is much higher than the size of the training set. In this paper, we try to use random sampling techniques to overcome these problems. To address the first two problems, we propose an asymmetric bagging based SVM. For the third problem, we combine the random subspace method (RSM) and SVM for RF. Finally, by integrating bagging and RSM, we solve all the three problems and further improve the RF performance. 1.