Results 1  10
of
1,829,849
A Signal Processing Approach To Fair Surface Design
, 1995
"... In this paper we describe a new tool for interactive freeform fair surface design. By generalizing classical discrete Fourier analysis to twodimensional discrete surface signals  functions defined on polyhedral surfaces of arbitrary topology , we reduce the problem of surface smoothing, or fai ..."
Abstract

Cited by 655 (15 self)
 Add to MetaCart
In this paper we describe a new tool for interactive freeform fair surface design. By generalizing classical discrete Fourier analysis to twodimensional discrete surface signals  functions defined on polyhedral surfaces of arbitrary topology , we reduce the problem of surface smoothing
Breaking and Fixing the NeedhamSchroeder PublicKey Protocol using FDR
, 1996
"... In this paper we analyse the well known NeedhamSchroeder PublicKey Protocol using FDR, a refinement checker for CSP. We use FDR to discover an attack upon the protocol, which allows an intruder to impersonate another agent. We adapt the protocol, and then use FDR to show that the new protocol is s ..."
Abstract

Cited by 718 (13 self)
 Add to MetaCart
is secure, at least for a small system. Finally we prove a result which tells us that if this small system is secure, then so is a system of arbitrary size. 1 Introduction In a distributed computer system, it is necessary to have some mechanism whereby a pair of agents can be assured of each other
Multipoint quantitativetrait linkage analysis in general pedigrees
 Am. J. Hum. Genet
, 1998
"... Multipoint linkage analysis of quantitativetrait loci (QTLs) has previously been restricted to sibships and small pedigrees. In this article, we show how variancecomponent linkage methods can be used in pedigrees of arbitrary size and complexity, and we develop a general framework for multipoint i ..."
Abstract

Cited by 567 (60 self)
 Add to MetaCart
Multipoint linkage analysis of quantitativetrait loci (QTLs) has previously been restricted to sibships and small pedigrees. In this article, we show how variancecomponent linkage methods can be used in pedigrees of arbitrary size and complexity, and we develop a general framework for multipoint
Accurate Methods for the Statistics of Surprise and Coincidence
 COMPUTATIONAL LINGUISTICS
, 1993
"... Much work has been done on the statistical analysis of text. In some cases reported in the literature, inappropriate statistical methods have been used, and statistical significance of results have not been addressed. In particular, asymptotic normality assumptions have often been used unjustifiably ..."
Abstract

Cited by 1055 (1 self)
 Add to MetaCart
unjustifiably, leading to flawed results.This assumption of normal distribution limits the ability to analyze rare events. Unfortunately rare events do make up a large fraction of real text.However, more applicable methods based on likelihood ratio tests are available that yield good results with relatively
The Cache Performance and Optimizations of Blocked Algorithms
 In Proceedings of the Fourth International Conference on Architectural Support for Programming Languages and Operating Systems
, 1991
"... Blocking is a wellknown optimization technique for improving the effectiveness of memory hierarchies. Instead of operating on entire rows or columns of an array, blocked algorithms operate on submatrices or blocks, so that data loaded into the faster levels of the memory hierarchy are reused. This ..."
Abstract

Cited by 572 (5 self)
 Add to MetaCart
is highly sensitive to the stride of data accesses and the size of the blocks, and can cause wide variations in machine performance for different matrix sizes. The conventional wisdom of trying to use the entire cache, or even a fixed fraction of the cache, is incorrect. If a fixed block size is used for a
The strength of weak learnability
 MACHINE LEARNING
, 1990
"... This paper addresses the problem of improving the accuracy of an hypothesis output by a learning algorithm in the distributionfree (PAC) learning model. A concept class is learnable (or strongly learnable) if, given access to a Source of examples of the unknown concept, the learner with high prob ..."
Abstract

Cited by 870 (26 self)
 Add to MetaCart
probability is able to output an hypothesis that is correct on all but an arbitrarily small fraction of the instances. The concept class is weakly learnable if the learner can produce an hypothesis that performs only slightly better than random guessing. In this paper, it is shown that these two notions
Decoding by Linear Programming
, 2004
"... This paper considers the classical error correcting problem which is frequently discussed in coding theory. We wish to recover an input vector f ∈ Rn from corrupted measurements y = Af + e. Here, A is an m by n (coding) matrix and e is an arbitrary and unknown vector of errors. Is it possible to rec ..."
Abstract

Cited by 1399 (16 self)
 Add to MetaCart
This paper considers the classical error correcting problem which is frequently discussed in coding theory. We wish to recover an input vector f ∈ Rn from corrupted measurements y = Af + e. Here, A is an m by n (coding) matrix and e is an arbitrary and unknown vector of errors. Is it possible
Quantum complexity theory
 in Proc. 25th Annual ACM Symposium on Theory of Computing, ACM
, 1993
"... Abstract. In this paper we study quantum computation from a complexity theoretic viewpoint. Our first result is the existence of an efficient universal quantum Turing machine in Deutsch’s model of a quantum Turing machine (QTM) [Proc. Roy. Soc. London Ser. A, 400 (1985), pp. 97–117]. This constructi ..."
Abstract

Cited by 574 (5 self)
 Add to MetaCart
be implemented and introduce some new, purely quantum mechanical primitives, such as changing the computational basis and carrying out an arbitrary unitary transformation of polynomially bounded dimension. We also consider the precision to which the transition amplitudes of a quantum Turing machine need
Integration of trade and disintegration of production in the global economy
 Journal of Economic Perspectives
, 1998
"... The last few decades have seen a spectacular integration of the global economy through trade. The rising integration of world markets has brought with it a disintegration of the production process, however, as manufacturing or services activities done abroad are combined with those performed at home ..."
Abstract

Cited by 496 (7 self)
 Add to MetaCart
seemingly small fraction of U.S. GDP. This is not surprising in view of the fact that large economies trade less with others, and more internally. But the modest share of trade in total national income hides the fact that merchandise trade as a share of merchandise valueadded is quite high for the U
Loopy belief propagation for approximate inference: An empirical study. In:
 Proceedings of Uncertainty in AI,
, 1999
"... Abstract Recently, researchers have demonstrated that "loopy belief propagation" the use of Pearl's polytree algorithm in a Bayesian network with loops can perform well in the context of errorcorrecting codes. The most dramatic instance of this is the near Shannonlimit performanc ..."
Abstract

Cited by 674 (15 self)
 Add to MetaCart
. Introduction The task of calculating posterior marginals on nodes in an arbitrary Bayesian network is known to be NP hard In this paper we investigate the approximation performance of "loopy belief propagation". This refers to using the wellknown Pearl polytree algorithm [12] on a Bayesian network
Results 1  10
of
1,829,849