Results 1  10
of
29,670
Generating Significant Examples for
, 1994
"... This report bases itself on the idea of using concrete examples to verify conceptual schemas, and in particular cardinality constraints. When novice ORM modellers model domains, the selection of proper cardinality constraints for relationship types is quite often prone to errors. In this report w ..."
Abstract
 Add to MetaCart
we propose a mechanism for the generation of significant examples for selected subschemas. The generated examples are significant in the sense that they illustrate the possible combinations of instances that are allowed with respect to the cardinality constraints on the involved relationship
Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories
, 2004
"... Abstract — Current computational approaches to learning visual object categories require thousands of training images, are slow, cannot learn in an incremental manner and cannot incorporate prior information into the learning process. In addition, no algorithm presented in the literature has been te ..."
Abstract

Cited by 784 (16 self)
 Add to MetaCart
learning is significantly faster, making realtime learning feasible. Both Bayesian methods outperform maximum likelihood on small training sets. I.
A Practical Guide to Wavelet Analysis
, 1998
"... A practical stepbystep guide to wavelet analysis is given, with examples taken from time series of the El Nio Southern Oscillation (ENSO). The guide includes a comparison to the windowed Fourier transform, the choice of an appropriate wavelet basis function, edge effects due to finitelength t ..."
Abstract

Cited by 869 (3 self)
 Add to MetaCart
A practical stepbystep guide to wavelet analysis is given, with examples taken from time series of the El Nio Southern Oscillation (ENSO). The guide includes a comparison to the windowed Fourier transform, the choice of an appropriate wavelet basis function, edge effects due to finite
Improving generalization with active learning
 Machine Learning
, 1994
"... Abstract. Active learning differs from "learning from examples " in that the learning algorithm assumes at least some control over what part of the input domain it receives information about. In some situations, active learning is provably more powerful than learning from examples ..."
Abstract

Cited by 544 (1 self)
 Add to MetaCart
Abstract. Active learning differs from "learning from examples " in that the learning algorithm assumes at least some control over what part of the input domain it receives information about. In some situations, active learning is provably more powerful than learning from examples
Distance metric learning for large margin nearest neighbor classification
 In NIPS
, 2006
"... We show how to learn a Mahanalobis distance metric for knearest neighbor (kNN) classification by semidefinite programming. The metric is trained with the goal that the knearest neighbors always belong to the same class while examples from different classes are separated by a large margin. On seven ..."
Abstract

Cited by 695 (14 self)
 Add to MetaCart
. On seven data sets of varying size and difficulty, we find that metrics trained in this way lead to significant improvements in kNN classification—for example, achieving a test error rate of 1.3 % on the MNIST handwritten digits. As in support vector machines (SVMs), the learning problem reduces to a
New results in linear filtering and prediction theory
 TRANS. ASME, SER. D, J. BASIC ENG
, 1961
"... A nonlinear differential equation of the Riccati type is derived for the covariance matrix of the optimal filtering error. The solution of this "variance equation " completely specifies the optimal filter for either finite or infinite smoothing intervals and stationary or nonstationary sta ..."
Abstract

Cited by 607 (0 self)
 Add to MetaCart
statistics. The variance equation is closely related to the Hamiltonian (canonical) differential equations of the calculus of variations. Analytic solutions are available in some cases. The significance of the variance equation is illustrated by examples which duplicate, simplify, or extend earlier results
Sparse MRI: The Application of Compressed Sensing for Rapid MR Imaging
 MAGNETIC RESONANCE IN MEDICINE 58:1182–1195
, 2007
"... The sparsity which is implicit in MR images is exploited to significantly undersample kspace. Some MR images such as angiograms are already sparse in the pixel representation; other, more complicated images have a sparse representation in some transform domain–for example, in terms of spatial finit ..."
Abstract

Cited by 538 (11 self)
 Add to MetaCart
The sparsity which is implicit in MR images is exploited to significantly undersample kspace. Some MR images such as angiograms are already sparse in the pixel representation; other, more complicated images have a sparse representation in some transform domain–for example, in terms of spatial
Hierarchies from Fluxes in String Compactifications
, 2002
"... Warped compactifications with significant warping provide one of the few known mechanisms for naturally generating large hierarchies of physical scales. We demonstrate that this mechanism is realizable in string theory, and give examples involving orientifold compactifications of IIB string theory a ..."
Abstract

Cited by 715 (33 self)
 Add to MetaCart
Warped compactifications with significant warping provide one of the few known mechanisms for naturally generating large hierarchies of physical scales. We demonstrate that this mechanism is realizable in string theory, and give examples involving orientifold compactifications of IIB string theory
Distance metric learning, with application to clustering with sideinformation,”
 in Advances in Neural Information Processing Systems 15,
, 2002
"... Abstract Many algorithms rely critically on being given a good metric over their inputs. For instance, data can often be clustered in many "plausible" ways, and if a clustering algorithm such as Kmeans initially fails to find one that is meaningful to a user, the only recourse may be for ..."
Abstract

Cited by 818 (13 self)
 Add to MetaCart
to provide examples. In this paper, we present an algorithm that, given examples of similar (and, if desired, dissimilar) pairs of points in Ê Ò , learns a distance metric over Ê Ò that respects these relationships. Our method is based on posing metric learning as a convex optimization problem, which allows
Analysis of TCP Performance over Mobile Ad Hoc Networks Part I: Problem Discussion and Analysis of Results
, 1999
"... Mobile ad hoc networks have gained a lot of attention lately as a means of providing continuous network connectivity to mobile computing devices regardless of physical location. Recently, a large amount of research has focused on the routing protocols needed in such an environment. In this twopart ..."
Abstract

Cited by 521 (5 self)
 Add to MetaCart
examples, such as a situation where throughput is zero for a particular connection. We introduce a new metric, expected throughput, for the comparison of throughput in multihop networks, and then use this metric to show how the use of explicit link failure notification (ELFN) techniques can significantly
Results 1  10
of
29,670