• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 6,541
Next 10 →

Neural network ensembles, cross validation, and active learning

by Anders Krogh, Jesper Vedelsby - Neural Information Processing Systems 7 , 1995
"... Learning of continuous valued functions using neural network en-sembles (committees) can give improved accuracy, reliable estima-tion of the generalization error, and active learning. The ambiguity is defined as the variation of the output of ensemble members aver-aged over unlabeled data, so it qua ..."
Abstract - Cited by 479 (6 self) - Add to MetaCart
it quantifies the disagreement among the networks. It is discussed how to use the ambiguity in combina-tion with cross-validation to give a reliable estimate of the ensemble generalization error, and how this type of ensemble cross-validation can sometimes improve performance. It is shown how to estimate

A PERFORMANCE EVALUATION OF LOCAL DESCRIPTORS

by Krystian Mikolajczyk, Cordelia Schmid , 2005
"... In this paper we compare the performance of descriptors computed for local interest regions, as for example extracted by the Harris-Affine detector [32]. Many different descriptors have been proposed in the literature. However, it is unclear which descriptors are more appropriate and how their perfo ..."
Abstract - Cited by 1783 (51 self) - Add to MetaCart
In this paper we compare the performance of descriptors computed for local interest regions, as for example extracted by the Harris-Affine detector [32]. Many different descriptors have been proposed in the literature. However, it is unclear which descriptors are more appropriate and how

Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms

by Thomas G. Dietterich , 1998
"... This article reviews five approximate statistical tests for determining whether one learning algorithm outperforms another on a particular learning task. These tests are compared experimentally to determine their probability of incorrectly detecting a difference when no difference exists (type I err ..."
Abstract - Cited by 723 (8 self) - Add to MetaCart
-differences t test based on 10-fold cross-validation, exhibits somewhat elevated probability of type I error. A fourth test, McNemar’s test, is shown to have low type I error. The fifth test is a new test, 5 × 2 cv, based on five iterations of twofold cross-validation. Experiments show that this test also has

Stacked generalization

by David H. Wolpert - NEURAL NETWORKS , 1992
"... This paper introduces stacked generalization, a scheme for minimizing the generalization error rate of one or more generalizers. Stacked generalization works by deducing the biases of the generalizer(s) with respect to a provided learning set. This deduction proceeds by generalizing in a second sp ..."
Abstract - Cited by 731 (9 self) - Add to MetaCart
sophisticated version of cross-validation, exploiting a strategy more sophisticated than cross-validation’s crude winner-takes-all for combining the individual generalizers. When used with a single generalizer, stacked generalization is a scheme for estimating (and then correcting for) the error of a

Image denoising using a scale mixture of Gaussians in the wavelet domain

by Javier Portilla, Vasily Strela, Martin J. Wainwright, Eero P. Simoncelli - IEEE TRANS IMAGE PROCESSING , 2003
"... We describe a method for removing noise from digital images, based on a statistical model of the coefficients of an overcomplete multiscale oriented basis. Neighborhoods of coefficients at adjacent positions and scales are modeled as the product of two independent random variables: a Gaussian vecto ..."
Abstract - Cited by 513 (17 self) - Add to MetaCart
coefficient reduces to a weighted average of the local linear estimates over all possible values of the hidden multiplier variable. We demonstrate through simulations with images contaminated by additive white Gaussian noise that the performance of this method substantially surpasses that of previously

Representing twentieth century space-time climate variability, part 1: development of a 1961-90 mean monthly terrestrial climatology

by Mark New, Mike Hulme, Phil Jones - Journal of Climate , 1999
"... The construction of a 0.58 lat 3 0.58 long surface climatology of global land areas, excluding Antarctica, is described. The climatology represents the period 1961–90 and comprises a suite of nine variables: precipitation, wet-day frequency, mean temperature, diurnal temperature range, vapor pressur ..."
Abstract - Cited by 581 (13 self) - Add to MetaCart
, longitude, and elevation using thin-plate splines. The accuracy of the interpolations are assessed using cross validation and by comparison with other climatologies. This new climatology represents an advance over earlier published global terrestrial climatologies in that it is strictly constrained

VERY HIGH RESOLUTION INTERPOLATED CLIMATE SURFACES FOR GLOBAL LAND AREAS

by Robert J. Hijmans, Susan E. Cameron, Juan L. Parra, Peter G. Jones , Andy Jarvis , 2005
"... We developed interpolated climate surfaces for global land areas (excluding Antarctica) at a spatial resolution of 30 arc s (often referred to as 1-km spatial resolution). The climate elements considered were monthly precipitation and mean, minimum, and maximum temperature. Input data were gathered ..."
Abstract - Cited by 553 (8 self) - Add to MetaCart
arising from the input data and the interpolation by mapping weather station density, elevation bias in the weather stations, and elevation variation within grid cells and through data partitioning and cross validation. Elevation bias tended to be negative (stations lower than expected) at high latitudes

Loopy belief propagation for approximate inference: An empirical study. In:

by Kevin P Murphy , Yair Weiss , Michael I Jordan - Proceedings of Uncertainty in AI, , 1999
"... Abstract Recently, researchers have demonstrated that "loopy belief propagation" -the use of Pearl's polytree algorithm in a Bayesian network with loops -can perform well in the context of error-correcting codes. The most dramatic instance of this is the near Shannon-limit performanc ..."
Abstract - Cited by 676 (15 self) - Add to MetaCart
the real QMR network to converge if the priors were sampled randomly in the range [0, Small priors are not the only thing that causes oscil lation. Small weights can, too. The effect of both The exact marginals are represented by the circles; the ends of the "error bars" represent the loopy

Constructive Incremental Learning From Only Local Information

by Stefan Schaal, Christopher G. Atkeson - NEURAL COMPUTATION
"... We introduce a constructive, incremental learning system for regression problems that models data by means of spatially localized linear models. In contrast to other approaches, the size and shape of the receptive field of each locally linear model as well as the parameters of the locally linear mod ..."
Abstract - Cited by 208 (40 self) - Add to MetaCart
model itself are learned independently, i.e., without the need for competition or any other kind of communication. Independent learning is accomplished by incrementally minimizing a weighted local cross validation error. As a result, we obtain a learning system that can allocate resources as needed

Estimation and Inference in Large Heterogeneous Panels with a Multifactor Error Structure

by M. Hashem Pesaran , 2004
"... This paper presents a new approach to estimation and inference in panel data models with a multifactor error structure where the unobserved common factors are (possibly) correlated with exogenously given individual-specific regressors, and the factor loadings differ over the cross section units. The ..."
Abstract - Cited by 383 (44 self) - Add to MetaCart
This paper presents a new approach to estimation and inference in panel data models with a multifactor error structure where the unobserved common factors are (possibly) correlated with exogenously given individual-specific regressors, and the factor loadings differ over the cross section units
Next 10 →
Results 1 - 10 of 6,541
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University