Results 1  10
of
38,030
Approximate Riemann Solvers, Parameter Vectors, and Difference Schemes
 J. COMP. PHYS
, 1981
"... Several numerical schemes for the solution of hyperbolic conservation laws are based on exploiting the information obtained by considering a sequence of Riemann problems. It is argued that in existing schemes much of this information is degraded, and that only certain features of the exact solution ..."
Abstract

Cited by 1010 (2 self)
 Add to MetaCart
Several numerical schemes for the solution of hyperbolic conservation laws are based on exploiting the information obtained by considering a sequence of Riemann problems. It is argued that in existing schemes much of this information is degraded, and that only certain features of the exact solution
PARAPRODUCTS IN ONE AND SEVERAL PARAMETERS
, 2005
"... For multiparameter bilinear paraproduct operators B we prove the estimate B: L p × L q ↦ → L r, 1 < p, q ≤ ∞. Here, 1/p + 1/q = 1/r and special attention is paid to the case of 0 < r < 1. (Note that the families of multiparameter paraproducts are much richer than in the one parameter case. ..."
Abstract

Cited by 10 (0 self)
 Add to MetaCart
For multiparameter bilinear paraproduct operators B we prove the estimate B: L p × L q ↦ → L r, 1 < p, q ≤ ∞. Here, 1/p + 1/q = 1/r and special attention is paid to the case of 0 < r < 1. (Note that the families of multiparameter paraproducts are much richer than in the one parameter case
Several parameters of generalized Mycielskians
, 2006
"... The generalized Mycielskians (also known as cones over graphs) are the natural generalization of the Mycielski graphs (which were first introduced by Mycielski in 1955). Given a graph G and any integer m0, one can transform G into a new graph m(G), the generalized Mycielskian of G. This paper invest ..."
Abstract

Cited by 2 (0 self)
 Add to MetaCart
The generalized Mycielskians (also known as cones over graphs) are the natural generalization of the Mycielski graphs (which were first introduced by Mycielski in 1955). Given a graph G and any integer m0, one can transform G into a new graph m(G), the generalized Mycielskian of G. This paper investigates circular clique number, total domination number, open packing number, fractional open packing number, vertex cover number, determinant, spectrum, and biclique partition number of m(G).
General methods for monitoring convergence of iterative simulations
 J. Comput. Graph. Statist
, 1998
"... We generalize the method proposed by Gelman and Rubin (1992a) for monitoring the convergence of iterative simulations by comparing between and within variances of multiple chains, in order to obtain a family of tests for convergence. We review methods of inference from simulations in order to develo ..."
Abstract

Cited by 551 (8 self)
 Add to MetaCart
, for assessing convergence of several parameters simultaneously.
A View Of The Em Algorithm That Justifies Incremental, Sparse, And Other Variants
 Learning in Graphical Models
, 1998
"... . The EM algorithm performs maximum likelihood estimation for data in which some variables are unobserved. We present a function that resembles negative free energy and show that the M step maximizes this function with respect to the model parameters and the E step maximizes it with respect to the d ..."
Abstract

Cited by 993 (18 self)
 Add to MetaCart
. The EM algorithm performs maximum likelihood estimation for data in which some variables are unobserved. We present a function that resembles negative free energy and show that the M step maximizes this function with respect to the model parameters and the E step maximizes it with respect
Training Products of Experts by Minimizing Contrastive Divergence
, 2002
"... It is possible to combine multiple latentvariable models of the same data by multiplying their probability distributions together and then renormalizing. This way of combining individual “expert ” models makes it hard to generate samples from the combined model but easy to infer the values of the l ..."
Abstract

Cited by 850 (75 self)
 Add to MetaCart
derivatives with regard to the parameters can be approximated accurately and efficiently. Examples are presented of contrastive divergence learning using several types of expert on several types of data.
Conditional random fields: Probabilistic models for segmenting and labeling sequence data
, 2001
"... We present conditional random fields, a framework for building probabilistic models to segment and label sequence data. Conditional random fields offer several advantages over hidden Markov models and stochastic grammars for such tasks, including the ability to relax strong independence assumptions ..."
Abstract

Cited by 3485 (85 self)
 Add to MetaCart
We present conditional random fields, a framework for building probabilistic models to segment and label sequence data. Conditional random fields offer several advantages over hidden Markov models and stochastic grammars for such tasks, including the ability to relax strong independence assumptions
Realtime human pose recognition in parts from single depth images
 IN CVPR
, 2011
"... We propose a new method to quickly and accurately predict 3D positions of body joints from a single depth image, using no temporal information. We take an object recognition approach, designing an intermediate body parts representation that maps the difficult pose estimation problem into a simpler p ..."
Abstract

Cited by 568 (17 self)
 Add to MetaCart
local modes. The system runs at 200 frames per second on consumer hardware. Our evaluation shows high accuracy on both synthetic and real test sets, and investigates the effect of several training parameters. We achieve state of the art accuracy in our comparison with related work and demonstrate
The Cache Performance and Optimizations of Blocked Algorithms
 In Proceedings of the Fourth International Conference on Architectural Support for Programming Languages and Operating Systems
, 1991
"... Blocking is a wellknown optimization technique for improving the effectiveness of memory hierarchies. Instead of operating on entire rows or columns of an array, blocked algorithms operate on submatrices or blocks, so that data loaded into the faster levels of the memory hierarchy are reused. This ..."
Abstract

Cited by 574 (5 self)
 Add to MetaCart
. This paper presents cache performance data for blocked programs and evaluates several optimizations to improve this performance. The data is obtained by a theoretical model of data conflicts in the cache, which has been validated by large amounts of simulation. We show that the degree of cache interference
Improved Boosting Algorithms Using Confidencerated Predictions
 MACHINE LEARNING
, 1999
"... We describe several improvements to Freund and Schapire’s AdaBoost boosting algorithm, particularly in a setting in which hypotheses may assign confidences to each of their predictions. We give a simplified analysis of AdaBoost in this setting, and we show how this analysis can be used to find impr ..."
Abstract

Cited by 940 (26 self)
 Add to MetaCart
We describe several improvements to Freund and Schapire’s AdaBoost boosting algorithm, particularly in a setting in which hypotheses may assign confidences to each of their predictions. We give a simplified analysis of AdaBoost in this setting, and we show how this analysis can be used to find
Results 1  10
of
38,030