• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Aggregating ordinal labels from crowds by minimax conditional entropy. (2014)

by D Zhou, Q Liu, J Platt, C Meek
Venue:In ICML,
Add To MetaCart

Tools

Sorted by:
Results 1 - 5 of 5

Spectral methods meet EM: A provably optimal algorithm for crowdsourcing. In Advances in neural information processing systems,

by Yuchen Zhang , Dengyong Zhou , Michael I Jordan , 2014
"... Abstract Crowdsourcing is a popular paradigm for effectively collecting labels at low cost. The Dawid-Skene estimator has been widely used for inferring the true labels from the noisy labels provided by non-expert crowdsourcing workers. However, since the estimator maximizes a non-convex log-likeli ..."
Abstract - Cited by 15 (2 self) - Add to MetaCart
Abstract Crowdsourcing is a popular paradigm for effectively collecting labels at low cost. The Dawid-Skene estimator has been widely used for inferring the true labels from the noisy labels provided by non-expert crowdsourcing workers. However, since the estimator maximizes a non-convex log-likelihood function, it is hard to theoretically justify its performance. In this paper, we propose a two-stage efficient algorithm for multi-class crowd labeling problems. The first stage uses the spectral method to obtain an initial estimate of parameters. Then the second stage refines the estimation by optimizing the objective function of the Dawid-Skene estimator via the EM algorithm. We show that our algorithm achieves the optimal convergence rate up to a logarithmic factor. We conduct extensive experiments on synthetic and real datasets. Experimental results demonstrate that the proposed algorithm is comparable to the most accurate empirical approach, while outperforming several other recently proposed methods.
(Show Context)

Citation Context

...intly estimated by maximizing the likelihood of the observed worker labels, where the unobserved true labels are treated as latent variables. Although this EM-based approach has had empirical success =-=[21, 20, 19, 26, 6, 25]-=-, there is as yet no theoretical guarantee for its performance. A recent theoretical study [10] shows that the global optimal solutions of the Dawid-Skene estimator can achieve minimax rates of conver...

Max-Margin Majority Voting for Learning from Crowds

by Tian Tian , Jun Zhu
"... Abstract Learning-from-crowds aims to design proper aggregation strategies to infer the unknown true labels from the noisy labels provided by ordinary web workers. This paper presents max-margin majority voting (M 3 V) to improve the discriminative ability of majority voting and further presents a ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
Abstract Learning-from-crowds aims to design proper aggregation strategies to infer the unknown true labels from the noisy labels provided by ordinary web workers. This paper presents max-margin majority voting (M 3 V) to improve the discriminative ability of majority voting and further presents a Bayesian generalization to incorporate the flexibility of generative methods on modeling noisy observations with worker confusion matrices. We formulate the joint learning as a regularized Bayesian inference problem, where the posterior regularization is derived by maximizing the margin between the aggregated score of a potential true label and that of any alternative label. Our Bayesian model naturally covers the Dawid-Skene estimator and M 3 V. Empirical results demonstrate that our methods are competitive, often achieving better results than state-of-the-art estimators.
(Show Context)

Citation Context

...ple times by different workers, then the redundant labels can provide hints on resolving the true labels. Much progress has been made in designing effective aggregation mechanisms to infer the true labels from noisy observations. From a modeling perspective, existing work includes both generative approaches and discriminative approaches. A generative method builds a flexible probabilistic model for generating the noisy observations conditioned on the unknown true labels and some behavior assumptions, with examples of the Dawid-Skene (DS) estimator [5], the minimax entropy (Entropy) estimator1 [24, 25], and their variants. In contrast, a discriminative approach does not model the observations; it directly identifies the true labels via some aggregation rules. Examples include majority voting and the weighted majority voting that takes worker reliability into consideration [10, 11]. In this paper, we present a max-margin formulation of the most popular majority voting estimator to improve its discriminative ability, and further present a Bayesian generalization that conjoins the advantages of both generative and discriminative approaches. The max-margin majority voting (M3V) directly maximiz...

Meta-Gradient Boosted Decision Tree Model for Weight and Target Learning Gleb Gusev GLEB57@YANDEX-TEAM.RU

by Yury Ustinovskiy , Valentina Fedorova , Pavel Serdyukov
"... Abstract Labeled training data is an essential part of any supervised machine learning framework. In practice, there is a trade-off between the quality of a label and its cost. In this paper, we consider a problem of learning to rank on a large-scale dataset with low-quality relevance labels aiming ..."
Abstract - Add to MetaCart
Abstract Labeled training data is an essential part of any supervised machine learning framework. In practice, there is a trade-off between the quality of a label and its cost. In this paper, we consider a problem of learning to rank on a large-scale dataset with low-quality relevance labels aiming at maximizing the quality of a trained ranker on a small validation dataset with high-quality ground truth relevance labels. Motivated by the classical Gauss-Markov theorem for the linear regression problem, we formulate the problems of (1) reweighting training instances and (2) remapping learning targets. We propose metagradient decision tree learning framework for optimizing weight and target functions by applying gradient-based hyperparameter optimization. Experiments on a large-scale real-world dataset demonstrate that we can significantly improve state-of-the-art machine-learning algorithms by incorporating our framework.
(Show Context)

Citation Context

...sourced labels. Recent rapid take-up of crowdsourcing marketplaces, e.g., Amazon MTurk1, provides a cheap way of collecting large supervised datasets (Muhammadi et al., 2013). However, due to a lack of expertise and presence of spammer workers, such labels often hugely vary in quality. To overcome this issue, employers assign each item to multi1http://www.mturk.com/ MGBDT Model for Weight and Target Learning ple workers. Afterwards, item’s multiple labels are aggregated into a consensus label with the use of a certain consensus algorithm, e.g., majority voting, label averaging, etc (see, e.g. Zhou et al., 2014). Various consensus models are known to significantly improve the precision of raw crowd labels. The examples above share one common property, which is crucial for the algorithm we propose in this paper. Namely, for each item in the training dataset, besides a noisy label itself, we observe several additional variables, which altogether are further referred to as label features. Examples of label features are: fact of a click, click position, and dwell time in the click prediction problem; worker’s experience, outputs of various consensus models in the problem of learning with crowdsourced lab...

Truth Inference in Crowdsourcing: Is the Problem Solved?

by Yudian Zheng , Guoliang Li , # , Yuanbing Li , # , Caihua Shan , Reynold Cheng
"... ABSTRACT Crowdsourcing has emerged as a novel problem-solving paradigm, which facilitates addressing problems that are hard for computers, e.g., entity resolution and sentiment analysis. However, due to the openness of crowdsourcing, workers may yield low-quality answers, and a redundancy-based met ..."
Abstract - Add to MetaCart
ABSTRACT Crowdsourcing has emerged as a novel problem-solving paradigm, which facilitates addressing problems that are hard for computers, e.g., entity resolution and sentiment analysis. However, due to the openness of crowdsourcing, workers may yield low-quality answers, and a redundancy-based method is widely employed, which first assigns each task to multiple workers and then infers the correct answer (called truth) for the task based on the answers of the assigned workers. A fundamental problem in this method is Truth Inference, which decides how to effectively infer the truth. Recently, the database community and data mining community independently study this problem and propose various algorithms. However, these algorithms are not compared extensively under the same framework and it is hard for practitioners to select appropriate algorithms. To alleviate this problem, we provide a detailed survey on 17 existing algorithms and perform a comprehensive evaluation using 5 real datasets. We make all codes and datasets public for future research. Through experiments we find that existing algorithms are not stable across different datasets and there is no algorithm that outperforms others consistently. We believe that the truth inference problem is not fully solved, and identify the limitations of existing algorithms and point out promising research directions.

Globally Optimal Crowdsourcing Quality Management

by Akash Das Sarma, Aditya G. Parameswaran, Jennifer Widom
"... We study crowdsourcing quality management, that is, given worker responses to a set of tasks, our goal is to jointly estimate the true answers for the tasks, as well as the quality of the workers. Prior work on this problem relies primarily on applying Expectation-Maximization (EM) on the underlying ..."
Abstract - Add to MetaCart
We study crowdsourcing quality management, that is, given worker responses to a set of tasks, our goal is to jointly estimate the true answers for the tasks, as well as the quality of the workers. Prior work on this problem relies primarily on applying Expectation-Maximization (EM) on the underlying maximum likelihood problem to estimate true answers as well as worker quality. Unfortunately, EM only provides a locally optimal solution rather than a globally optimal one. Other solutions to the problem (that do not leverage EM) fail to provide global optimality guarantees as well. In this paper, we focus on filtering, where tasks require the evaluation of a yes/no predicate, and rating, where tasks elicit integer scores from a finite domain. We design algorithms for finding the global optimal estimates of correct task answers and worker quality for the underlying maximum likelihood problem, and characterize the complexity of these algorithms. Our algorithms conceptually consider all mappings from tasks to true answers (typically a very large number), leveraging two key ideas to reduce, by several orders of magnitude, the number of mappings under consideration, while preserving optimality. We also demonstrate that these algorithms often find more accurate estimates than EM-based algorithms. This paper makes an important contribution towards under-standing the inherent complexity of globally optimal crowdsourcing quality management.
(Show Context)

Citation Context

... the problem of obtaining confidence bounds on the worker error rates, we consider the problem of finding the maximum likelihood estimates to the item ground truth and worker error rates. Zhou et al. =-=[29, 30]-=- use minimax entropy to perform worker quality estimation as well as inherent item difficulty estimation; here the inherent item difficulty is represented as a vector. Their technique only applies whe...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University