• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Constrained bayesian inference for low rank multitask learning (2013)

by O Koyejo, J Ghosh
Venue:In UAI
Add To MetaCart

Tools

Sorted by:
Results 1 - 5 of 5

Approved: Lawrence Carin, Supervisor

by Xianxing Zhang, David B. Dunson, Arthur Calderbank, Guillermo Sapiro, Katherine Heller, Xianxing Zhang, David B. Dunson, Arthur Calderbank, Guillermo Sapiro, Katherine Heller , 2014
"... ..."
Abstract - Add to MetaCart
Abstract not found
(Show Context)

Citation Context

...ions. Some are known as the minimum relative entropy method [17, 48] for max-margin based discriminative learning tasks, or the posterior regularization method [13] and constrained Bayesian inference =-=[22]-=- for incorporating constraints into posterior inference. 2.3 hecMEM: Accounting for the local bias As discussed at the beginning of Section 2, in the large-scale setting, the dataset is (pre-)partitio...

Big Learning with Bayesian Methods

by Jun Zhu, Jianfei Chen, Wenbo Hu , 2007
"... Explosive growth in data and availability of cheap computing resources have sparked increasing interest in Big learning, an emerging subfield that studies scalable machine learning algorithms, systems, and applications with Big Data. Bayesian methods represent one important class of statistic metho ..."
Abstract - Add to MetaCart
Explosive growth in data and availability of cheap computing resources have sparked increasing interest in Big learning, an emerging subfield that studies scalable machine learning algorithms, systems, and applications with Big Data. Bayesian methods represent one important class of statistic methods for machine learning, with substantial recent developments on adaptive, flexible and scalable Bayesian learning. This article provides a survey of the recent advances in Big learning with Bayesian methods, termed Big Bayesian Learning, including nonparametric Bayesian methods for adaptively inferring model complexity, regularized Bayesian inference for improving the flexibility via posterior regularization, and scalable algorithms and systems based on stochastic subsampling and distributed computing for dealing with large-scale applications.
(Show Context)

Citation Context

...istribution. It can be defined as the large-margin constraints to enforce a good prediction accuracy [206], or the logic constraints to incorporate expert knowledge [124], or the sparsity constraints =-=[103]-=-. Example 3: Max-margin LDA Following the paradigm of ordinary Bayes, a supervised topic model is often defined by augmenting the likelihood model. For example, the supervised LDA (sLDA) [35] has a si...

Constrained Inference for Multi-View Clustering

by Shalmali Joshi, Oluwasanmi Koyejo, Joydeep Ghosh
"... We propose a novel approach for probabilis-tic multi-view clustering that combines view-specific models to improve global coherence. Global incoherence is measured by the differ-ence between view-specific cluster assignment responsibilities. New cluster responsibilities are estimated by optimizing a ..."
Abstract - Add to MetaCart
We propose a novel approach for probabilis-tic multi-view clustering that combines view-specific models to improve global coherence. Global incoherence is measured by the differ-ence between view-specific cluster assignment responsibilities. New cluster responsibilities are estimated by optimizing a cost function that maximizes per-view accuracy subject to a user-specified global coherence threshold. When combined with a parameter estimation step, this modified inference encourages the estimation of model parameters that agree between views. We show that the modified inference remains convex when global coherence constraints are given by the norm of the difference between the respon-sibilities of each model. In addition, the global correction is embarrassingly parallel between ex-amples. The proposed approach is evaluated on a synthetic dataset as well as real data showing im-proved performance as compared to strong base-line methods for multi-view clustering. 1.

A Constrained Matrix-Variate Gaussian Process for Transposable Data

by Oluwasanmi Koyejo, Cheng Lee, Joydeep Ghosh
"... Transposable data represents interactions among two sets of entities, and are typically represented as a matrix containing the known interaction values. Additional side information may consist of feature vectors specific to entities corresponding to the rows and/or columns of such a matrix. Further ..."
Abstract - Add to MetaCart
Transposable data represents interactions among two sets of entities, and are typically represented as a matrix containing the known interaction values. Additional side information may consist of feature vectors specific to entities corresponding to the rows and/or columns of such a matrix. Further information may also be available in the form of interactions or hierarchies among entities along the same mode (axis). We propose a novel approach for modeling transposable data with missing interactions given additional side information. The interactions are modeled as noisy observations from a latent noise free matrix generated from a matrix-variate Gaussian process. The construction of row and column covariances using side information provides a flexible mechanism for specifying a-priori knowledge of the row and column correlations in the data. Further, the use of such a prior combined with the side information enables predictions for new rows and columns not observed in the training data. In this work, we combine the matrix-variate Gaussian process model with low rank constraints. The constrained Gaussian process approach is applied to the prediction of hidden associations between genes and diseases using a small set of observed associations as well as prior covariances induced by gene-gene interaction networks and disease ontologies. The proposed approach is also applied to recommender systems data which involves predicting the item ratings of users using known associations as well as prior covariances induced by social networks. We present experimental results that highlight the performance of constrained matrix-variate Gaussian process as compared to state of the art approaches in each domain. 1

Bayesian Max-margin Multi-Task Learning with Data Augmentation

by Chengtao Li, Jianfei Chen
"... Both max-margin and Bayesian methods have been extensively studied in multi-task learning, but have rarely been considered together. We present Bayesian max-margin multi-task learning, which conjoins the two schools of methods, thus allowing the discriminative max-margin methods to enjoy the great f ..."
Abstract - Add to MetaCart
Both max-margin and Bayesian methods have been extensively studied in multi-task learning, but have rarely been considered together. We present Bayesian max-margin multi-task learning, which conjoins the two schools of methods, thus allowing the discriminative max-margin methods to enjoy the great flexibility of Bayesian methods on incorporating rich prior information as well as performing nonparametric Bayesian feature learning with the latent dimensionality resolved from data. We develop Gibbs sampling algorithms by exploring data augmentation to deal with the non-smooth hinge loss. For nonparametric models, our algorithms do not need to make mean-field assumptions or truncated approximation. Empirical results demonstrate superior performance than competitors in both multi-task classification and regression.
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University