• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 35,081
Next 10 →

Advances in Prospect Theory: Cumulative Representation of Uncertainty

by Amos Tversky, Daniel Kahneman - JOURNAL OF RISK AND UNCERTAINTY, 5:297-323 (1992) , 1992
"... We develop a new version of prospect theory that employs cumulative rather than separable decision weights and extends the theory in several respects. This version, called cumulative prospect theory, applies to uncertain as well as to risky prospects with any number of outcomes, and it allows differ ..."
Abstract - Cited by 1717 (17 self) - Add to MetaCart
distinctive fourfold pattern of risk attitudes: risk aversion for gains and risk seeking for losses of high probability; risk seeking for gains and risk aversion for losses of low probability.

Estimating the Support of a High-Dimensional Distribution

by Bernhard Schölkopf, John C. Platt, John Shawe-taylor, Alex J. Smola, Robert C. Williamson , 1999
"... Suppose you are given some dataset drawn from an underlying probability distribution P and you want to estimate a "simple" subset S of input space such that the probability that a test point drawn from P lies outside of S is bounded by some a priori specified between 0 and 1. We propo ..."
Abstract - Cited by 783 (29 self) - Add to MetaCart
Suppose you are given some dataset drawn from an underlying probability distribution P and you want to estimate a "simple" subset S of input space such that the probability that a test point drawn from P lies outside of S is bounded by some a priori specified between 0 and 1. We

High dimensional graphs and variable selection with the Lasso

by Nicolai Meinshausen, Peter Bühlmann - ANNALS OF STATISTICS , 2006
"... The pattern of zero entries in the inverse covariance matrix of a multivariate normal distribution corresponds to conditional independence restrictions between variables. Covariance selection aims at estimating those structural zeros from data. We show that neighborhood selection with the Lasso is a ..."
Abstract - Cited by 736 (22 self) - Add to MetaCart
show that the proposed neighborhood selection scheme is consistent for sparse high-dimensional graphs. Consistency hinges on the choice of the penalty parameter. The oracle value for optimal prediction does not lead to a consistent neighborhood estimate. Controlling instead the probability of falsely

Optimal Aggregation Algorithms for Middleware

by Ronald Fagin, Amnon Lotem , Moni Naor - IN PODS , 2001
"... Assume that each object in a database has m grades, or scores, one for each of m attributes. For example, an object can have a color grade, that tells how red it is, and a shape grade, that tells how round it is. For each attribute, there is a sorted list, which lists each object and its grade under ..."
Abstract - Cited by 717 (4 self) - Add to MetaCart
must access every object in the database, to find its grade under each attribute. Fagin has given an algorithm (“Fagin’s Algorithm”, or FA) that is much more efficient. For some monotone aggregation functions, FA is optimal with high probability in the worst case. We analyze an elegant and remarkably

High confidence visual recognition of persons by a test of statistical independence

by John G. Daugman - IEEE TRANS. ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE , 1993
"... A method for rapid visual recognition of personal identity is described, based on the failure of a statistical test of independence. The most unique phenotypic feature visible in a person’s face is the detailed texture of each eye’s iris: An estimate of its statistical complexity in a sample of the ..."
Abstract - Cited by 621 (8 self) - Add to MetaCart
to a conditional false accept probability of one in about l0^31.

Raptor codes

by Amin Shokrollahi - IEEE Transactions on Information Theory , 2006
"... LT-Codes are a new class of codes introduced in [1] for the purpose of scalable and fault-tolerant distribution of data over computer networks. In this paper we introduce Raptor Codes, an extension of LT-Codes with linear time encoding and decoding. We will exhibit a class of universal Raptor codes: ..."
Abstract - Cited by 577 (7 self) - Add to MetaCart
: for a given integer k, and any real ε> 0, Raptor codes in this class produce a potentially infinite stream of symbols such that any subset of symbols of size k(1 + ε) is sufficient to recover the original k symbols with high probability. Each output symbol is generated using O(log(1/ε)) operations

Robust Uncertainty Principles: Exact Signal Reconstruction From Highly Incomplete Frequency Information

by Emmanuel J. Candès, Justin Romberg, Terence Tao , 2006
"... This paper considers the model problem of reconstructing an object from incomplete frequency samples. Consider a discrete-time signal and a randomly chosen set of frequencies. Is it possible to reconstruct from the partial knowledge of its Fourier coefficients on the set? A typical result of this pa ..."
Abstract - Cited by 2632 (50 self) - Add to MetaCart
of this paper is as follows. Suppose that is a superposition of spikes @ Aa @ A @ A obeying @�� � A I for some constant H. We do not know the locations of the spikes nor their amplitudes. Then with probability at least I @ A, can be reconstructed exactly as the solution to the I minimization problem I aH @ A s

The strength of weak learnability

by Robert E. Schapire - MACHINE LEARNING , 1990
"... This paper addresses the problem of improving the accuracy of an hypothesis output by a learning algorithm in the distribution-free (PAC) learning model. A concept class is learnable (or strongly learnable) if, given access to a Source of examples of the unknown concept, the learner with high prob ..."
Abstract - Cited by 871 (26 self) - Add to MetaCart
This paper addresses the problem of improving the accuracy of an hypothesis output by a learning algorithm in the distribution-free (PAC) learning model. A concept class is learnable (or strongly learnable) if, given access to a Source of examples of the unknown concept, the learner with high

Distinctive Image Features from Scale-Invariant Keypoints

by David G. Lowe , 2003
"... This paper presents a method for extracting distinctive invariant features from images, which can be used to perform reliable matching between different images of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a a substa ..."
Abstract - Cited by 8955 (21 self) - Add to MetaCart
substantial range of affine distortion, addition of noise, change in 3D viewpoint, and change in illumination. The features are highly distinctive, in the sense that a single feature can be correctly matched with high probability against a large database of features from many images. This paper also

Exact Matrix Completion via Convex Optimization

by Emmanuel J. Candès, Benjamin Recht , 2008
"... We consider a problem of considerable practical interest: the recovery of a data matrix from a sampling of its entries. Suppose that we observe m entries selected uniformly at random from a matrix M. Can we complete the matrix and recover the entries that we have not seen? We show that one can perfe ..."
Abstract - Cited by 873 (26 self) - Add to MetaCart
perfectly recover most low-rank matrices from what appears to be an incomplete set of entries. We prove that if the number m of sampled entries obeys m ≥ C n 1.2 r log n for some positive numerical constant C, then with very high probability, most n × n matrices of rank r can be perfectly recovered
Next 10 →
Results 1 - 10 of 35,081
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University