• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Semi-supervised learning literature survey (2008)

by X Zhu
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 783
Next 10 →

A Survey on Transfer Learning

by Sinno Jialin Pan, Qiang Yang
"... A major assumption in many machine learning and data mining algorithms is that the training and future data must be in the same feature space and have the same distribution. However, in many real-world applications, this assumption may not hold. For example, we sometimes have a classification task i ..."
Abstract - Cited by 459 (24 self) - Add to MetaCart
A major assumption in many machine learning and data mining algorithms is that the training and future data must be in the same feature space and have the same distribution. However, in many real-world applications, this assumption may not hold. For example, we sometimes have a classification task in one domain of interest, but we only have sufficient training data in another domain of interest, where the latter data may be in a different feature space or follow a different data distribution. In such cases, knowledge transfer, if done successfully, would greatly improve the performance of learning by avoiding much expensive data labeling efforts. In recent years, transfer learning has emerged as a new learning framework to address this problem. This survey focuses on categorizing and reviewing the current progress on transfer learning for classification, regression and clustering problems. In this survey, we discuss the relationship between transfer learning and other related machine learning techniques such as domain adaptation, multitask learning and sample selection bias, as well as co-variate shift. We also explore some potential future issues in transfer learning research.
(Show Context)

Citation Context

...ing algorithms make predictions on the future data using statistical models that are trained on previously collected labeled or unlabeled training data [11], [12], [13]. Semisupervised classification =-=[14]-=-, [15], [16], [17] addresses the problem that the labeled data may be too few to build a good classifier, by making use of a large amount of unlabeled data and a small amount of labeled data. Variatio...

Active learning literature survey

by Burr Settles , 2010
"... The key idea behind active learning is that a machine learning algorithm can achieve greater accuracy with fewer labeled training instances if it is allowed to choose the data from which is learns. An active learner may ask queries in the form of unlabeled instances to be labeled by an oracle (e.g., ..."
Abstract - Cited by 326 (1 self) - Add to MetaCart
The key idea behind active learning is that a machine learning algorithm can achieve greater accuracy with fewer labeled training instances if it is allowed to choose the data from which is learns. An active learner may ask queries in the form of unlabeled instances to be labeled by an oracle (e.g., a human annotator). Active learning is well-motivated in many modern machine learning problems, where unlabeled data may be abundant but labels are difficult, time-consuming, or expensive to obtain. This report provides a general introduction to active learning and a survey of the literature. This includes a discussion of the scenarios in which queries can be formulated, and an overview of the query strategy frameworks proposed in the literature to date. An analysis of the empirical and theoretical evidence for active learning, a summary of several problem setting variants, and a discussion

Learning from imbalanced data

by Haibo He, Edwardo A. Garcia - IEEE Trans. on Knowledge and Data Engineering , 2009
"... Abstract—With the continuous expansion of data availability in many large-scale, complex, and networked systems, such as surveillance, security, Internet, and finance, it becomes critical to advance the fundamental understanding of knowledge discovery and analysis from raw data to support decision-m ..."
Abstract - Cited by 260 (6 self) - Add to MetaCart
Abstract—With the continuous expansion of data availability in many large-scale, complex, and networked systems, such as surveillance, security, Internet, and finance, it becomes critical to advance the fundamental understanding of knowledge discovery and analysis from raw data to support decision-making processes. Although existing knowledge discovery and data engineering techniques have shown great success in many real-world applications, the problem of learning from imbalanced data (the imbalanced learning problem) is a relatively new challenge that has attracted growing attention from both academia and industry. The imbalanced learning problem is concerned with the performance of learning algorithms in the presence of underrepresented data and severe class distribution skews. Due to the inherent complex characteristics of imbalanced data sets, learning from such data requires new understandings, principles, algorithms, and tools to transform vast amounts of raw data efficiently into information and knowledge representation. In this paper, we provide a comprehensive review of the development of research in learning from imbalanced data. Our focus is to provide a critical review of the nature of the problem, the state-of-the-art technologies, and the current assessment metrics used to evaluate learning performance under the imbalanced learning scenario. Furthermore, in order to stimulate future research in this field, we also highlight the major opportunities and challenges, as well as potential important research directions for learning from imbalanced data. Index Terms—Imbalanced learning, classification, sampling methods, cost-sensitive learning, kernel-based learning, active learning, assessment metrics. Ç
(Show Context)

Citation Context

...eled. The key idea of semisupervised learning is to exploit the unlabeled examples by using the labeled examples to modify, refine, or reprioritize the hypothesis obtained from the labeled data alone =-=[135]-=-. For instance, cotraining works under the assumption of two-viewed sets of feature spaces. Initially, two separate classifiers are trained with the labeled examples on two sufficient and conditionall...

Semi-Supervised On-line Boosting for Robust Tracking

by Helmut Grabner, Christian Leistner, Horst Bischof , 2008
"... Recently, on-line adaptation of binary classifiers for tracking have been investigated. On-line learning allows for simple classifiers since only the current view of the object from its surrounding background needs to be discriminiated. However, on-line adaption faces one key problem: Each update of ..."
Abstract - Cited by 186 (8 self) - Add to MetaCart
Recently, on-line adaptation of binary classifiers for tracking have been investigated. On-line learning allows for simple classifiers since only the current view of the object from its surrounding background needs to be discriminiated. However, on-line adaption faces one key problem: Each update of the tracker may introduce an error which, finally, can lead to tracking failure (drifting). The contribution of this paper is a novel on-line semi-supervised boosting method which significantly alleviates the drifting problem in tracking applications. This allows to limit the drifting problem while still staying adaptive to appearance changes. The main idea is to formulate the update process in a semisupervised fashion as combined decision of a given prior and an on-line classifier. This comes without any parameter tuning. In the experiments, we demonstrate real-time tracking of our SemiBoost tracker on several challenging test sequences where our tracker outperforms other on-line tracking methods.
(Show Context)

Citation Context

..., this has also been investigated by Li et al. [15] for tracking in low-frame rates. However, in order to formulate this problem in a principled manner we use ideas from semi-supervised learning (see =-=[16]-=- for a recent survey). In particular, we use the recently proposed SemiBoost [17, 18] for learning a classifier. Labeled data (or a previously trained model) is used as a prior and the data collected ...

Multi-Manifold Semi-Supervised Learning

by Andrew B. Goldberg
"... We study semi-supervised learning when the data consists of multiple intersecting manifolds. We give a finite sample analysis to quantify the potential gain of using unlabeled data in this multi-manifold setting. We then propose a semi-supervised learning algorithm that separates different manifolds ..."
Abstract - Cited by 147 (9 self) - Add to MetaCart
We study semi-supervised learning when the data consists of multiple intersecting manifolds. We give a finite sample analysis to quantify the potential gain of using unlabeled data in this multi-manifold setting. We then propose a semi-supervised learning algorithm that separates different manifolds into decision sets, and performs supervised learning within each set. Our algorithm involves a novel application of Hellinger distance and size-constrained spectral clustering. Experiments demonstrate the benefit of our multimanifold semi-supervised learning approach. 1
(Show Context)

Citation Context

...ervised learning (SSL) can improve the performance of a supervised learning (SL) task. There have also been many successful practical SSL algorithms as summarized in (Chapelle, Zien & Schölkopf 2006, =-=Zhu 2005-=-). These theoretical analyses and pracAppearing in Proceedings of the 12 th International Conference on Artificial Intelligence and Statistics (AISTATS) 2009, Clearwater Beach, Florida, USA. Volume 5 ...

Sparse Representation For Computer Vision and Pattern Recognition

by John Wright, Yi Ma, Julien Mairal, Guillermo Sapiro, Thomas Huang, Shuicheng Yan , 2009
"... Techniques from sparse signal representation are beginning to see significant impact in computer vision, often on non-traditional applications where the goal is not just to obtain a compact high-fidelity representation of the observed signal, but also to extract semantic information. The choice of ..."
Abstract - Cited by 146 (9 self) - Add to MetaCart
Techniques from sparse signal representation are beginning to see significant impact in computer vision, often on non-traditional applications where the goal is not just to obtain a compact high-fidelity representation of the observed signal, but also to extract semantic information. The choice of dictionary plays a key role in bridging this gap: unconventional dictionaries consisting of, or learned from, the training samples themselves provide the key to obtaining state-of-theart results and to attaching semantic meaning to sparse signal representations. Understanding the good performance of such unconventional dictionaries in turn demands new algorithmic and analytical techniques. This review paper highlights a few representative examples of how the interaction between sparse signal representation and computer vision can enrich both fields, and raises a number of open questions for further study.
(Show Context)

Citation Context

...all be explained within the graph embedding framework [73]. Also, a number of semisupervised learning algorithms are driven by the regularizing graphs constructed over both labeled and unlabeled data =-=[78]-=-. Most of the works described above rely on one of two popular approaches to graph construction: the k-nearest-neighbor method and the ε-ball method. The first assigns edges between each data point an...

Semi-supervised learning in gigantic image collections

by Rob Fergus, Yair Weiss, Antonio Torralba - In Advances in Neural Information Processing Systems 22 , 2009
"... With the advent of the Internet it is now possible to col-lect hundreds of millions of images. These images come with varying degrees of label information. “Clean labels” can be manually obtained on a small fraction, “noisy la-bels ” may be extracted automatically from surrounding text, while for mo ..."
Abstract - Cited by 76 (3 self) - Add to MetaCart
With the advent of the Internet it is now possible to col-lect hundreds of millions of images. These images come with varying degrees of label information. “Clean labels” can be manually obtained on a small fraction, “noisy la-bels ” may be extracted automatically from surrounding text, while for most images there are no labels at all. Semi-supervised learning is a principled framework for combin-ing these different label sources. However, it scales poly-nomially with the number of images, making it impractical for use on gigantic collections with hundreds of millions of images and thousands of classes. In this paper we show how to utilize recent results in ma-chine learning to obtain highly efficient approximations for semi-supervised learning. Specifically, we use the conver-gence of the eigenvectors of the normalized graph Lapla-cian to eigenfunctions of weighted Laplace-Beltrami oper-ators. We combine this with a label sharing framework obtained from Wordnet to propagate label information to classes lacking manual annotations. Our algorithm enables us to apply semi-supervised learning to a database of 80 million images with 74 thousand classes. 1.
(Show Context)

Citation Context

...y. This may be impractical for large datasets.” The large computational complexity of standard graph Laplacian methods has lead to a number of recent papers on efficient semi-supervised learning (see =-=[27]-=- for an overview). Many of these methods (e.g. [18, 12, 29, 25] are based on calculating the Laplacian only for a smaller, backbone, graph which reduces complexity to be cubic in the size of the small...

Learning with hypergraphs: Clustering, classification, and embedding

by Dengyong Zhou, Jiayuan Huang, Bernhard Schölkopf - Advances in Neural Information Processing Systems (NIPS) 19 , 2006
"... We usually endow the investigated objects with pairwise relationships, which can be illustrated as graphs. In many real-world problems, however, relationships among the objects of our interest are more complex than pairwise. Naively squeezing the complex relationships into pairwise ones will inevita ..."
Abstract - Cited by 74 (2 self) - Add to MetaCart
We usually endow the investigated objects with pairwise relationships, which can be illustrated as graphs. In many real-world problems, however, relationships among the objects of our interest are more complex than pairwise. Naively squeezing the complex relationships into pairwise ones will inevitably lead to loss of information which can be expected valuable for our learning tasks however. Therefore we consider using hypergraphs instead to completely represent complex relationships among the objects of our interest, and thus the problem of learning with hypergraphs arises. Our main contribution in this paper is to generalize the powerful methodology of spectral clustering which originally operates on undirected graphs to hypergraphs, and further develop algorithms for hypergraph embedding and transductive classification on the basis of the spectral hypergraph clustering approach. Our experiments on a number of benchmarks showed the advantages of hypergraphs over usual graphs. 1
(Show Context)

Citation Context

...tion function is finally given by f = (I − ξΘ) −1 y, where the elements of y denote the initial labels, and ξ is a parameter in (0, 1). For a survey on transductive inference, we refer the readers to =-=[21]-=-. 8 Experiments All datasets except a particular version of the 20-newsgroup one are from the UCI Machine Learning Depository. They are usually referred to as the so-called categorical data. Specifica...

Domain transfer svm for video concept detection

by Lixin Duan, Ivor W. Tsang, Dong Xu, Stephen J. Maybank - In CVPR , 2009
"... Cross-domain learning methods have shown promising results by leveraging labeled patterns from auxiliary domains to learn a robust classifier for target domain, which has a limited number of labeled samples. To cope with the tremendous change of feature distribution between different domains in vide ..."
Abstract - Cited by 71 (10 self) - Add to MetaCart
Cross-domain learning methods have shown promising results by leveraging labeled patterns from auxiliary domains to learn a robust classifier for target domain, which has a limited number of labeled samples. To cope with the tremendous change of feature distribution between different domains in video concept detection, we propose a new cross-domain kernel learning method. Our method, referred to as Domain Transfer SVM (DTSVM), simultaneously learns a kernel function and a robust SVM classifier by minimizing the both structural risk functional of SVM and distribution mismatch of labeled and unlabeled samples between the auxiliary and target domains. Comprehensive experiments on the challenging TRECVID corpus demonstrate that DTSVM outperforms existing crossdomain learning and multiple kernel learning methods. 1.
(Show Context)

Citation Context

... with re-weighted patterns. However, all these methods [5, 7, 17, 20] did not utilize unlabeled patterns in the target domain. Such patterns can be also used to improve the classification performance =-=[22]-=-. When there are only a few or even no labeled patterns in the target domain (i.e., an extreme case), the classifier can be trained with the auxiliary patterns. Several crossdomain learning methods [6...

Maximum margin clustering made practical.

by K Zhang, I W Tsang, J T Kwok - IEEE Transactions on Neural Networks, , 2009
"... ..."
Abstract - Cited by 70 (11 self) - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...jective. So the determination of takes a total of only time. Hence, both iterSVR and iterLS-SVM are computationally efficient. F. Discussions Our iterative algorithm is similar to self-training [41], =-=[44]-=-, in which a classifier uses its own predictions to help classify the unlabeled data. A small amount of labeled data is usually available and the labels of some highly confident, but previously unlabe...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University