Results 1  10
of
146,498
PVS: A Prototype Verification System
 CADE
, 1992
"... PVS is a prototype system for writing specifications and constructing proofs. Its development has been shaped by our experiences studying or using several other systems and performing a number of rather substantial formal verifications (e.g., [5,6,8]). PVS is fully implemented and freely available. ..."
Abstract

Cited by 654 (16 self)
 Add to MetaCart
. It has been used to construct proofs of nontrivial difficulty with relatively modest amounts of human effort. Here, we describe some of the motivation behind PVS and provide some details of the system. Automated reasoning systems typically fall in one of two classes: those that provide powerful
The process group approach to reliable distributed computing
 Communications of the ACM
, 1993
"... The difficulty of developing reliable distributed softwme is an impediment to applying distributed computing technology in many settings. Expeti _ with the Isis system suggests that a structured approach based on virtually synchronous _ groups yields systems that are substantially easier to develop, ..."
Abstract

Cited by 573 (19 self)
 Add to MetaCart
The difficulty of developing reliable distributed softwme is an impediment to applying distributed computing technology in many settings. Expeti _ with the Isis system suggests that a structured approach based on virtually synchronous _ groups yields systems that are substantially easier to develop
An introduction to Kolmogorov Complexity and its Applications: Preface to the First Edition
, 1997
"... This document has been prepared using the L a T E X system. We thank Donald Knuth for T E X, Leslie Lamport for L a T E X, and Jan van der Steen at CWI for online help. Some figures were prepared by John Tromp using the xpic program. The London Mathematical Society kindly gave permission to reproduc ..."
Abstract

Cited by 2143 (120 self)
 Add to MetaCart
This document has been prepared using the L a T E X system. We thank Donald Knuth for T E X, Leslie Lamport for L a T E X, and Jan van der Steen at CWI for online help. Some figures were prepared by John Tromp using the xpic program. The London Mathematical Society kindly gave permission to reproduce a long extract by A.M. Turing. The Indian Statistical Institute, through the editor of Sankhy¯a, kindly gave permission to quote A.N. Kolmogorov. We gratefully acknowledge the financial support by NSF Grant DCR8606366, ONR Grant N0001485k0445, ARO Grant DAAL0386K0171, the Natural Sciences and Engineering Research Council of Canada through operating grants OGP0036747, OGP046506, and International Scientific Exchange Awards ISE0046203, ISE0125663, and NWO Grant NF 62376. The book was conceived in late Spring 1986 in the Valley of the Moon in Sonoma County, California. The actual writing lasted on and off from autumn 1987 until summer 1993. One of us [PV] gives very special thanks to his lovely wife Pauline for insisting from the outset on the significance of this enterprise. The Aiken Computation Laboratory of Harvard University, Cambridge, Massachusetts, USA; the Computer Science Department of York University, Ontario, Canada; the Computer Science Department of the University xii of Waterloo, Ontario, Canada; and CWI, Amsterdam, the Netherlands provided the working environments in which this book could be written. Preface to the Second Edition
Genetic Programming
, 1997
"... Introduction Genetic programming is a domainindependent problemsolving approach in which computer programs are evolved to solve, or approximately solve, problems. Genetic programming is based on the Darwinian principle of reproduction and survival of the fittest and analogs of naturally occurring ..."
Abstract

Cited by 1051 (12 self)
 Add to MetaCart
Introduction Genetic programming is a domainindependent problemsolving approach in which computer programs are evolved to solve, or approximately solve, problems. Genetic programming is based on the Darwinian principle of reproduction and survival of the fittest and analogs of naturally occurring genetic operations such as crossover (sexual recombination) and mutation. John Holland's pioneering Adaptation in Natural and Artificial Systems (1975) described how an analog of the evolutionary process can be applied to solving mathematical problems and engineering optimization problems using what is now called the genetic algorithm (GA). The genetic algorithm attempts to find a good (or best) solution to the problem by genetically breeding a population of individuals over a series of generations. In the genetic algorithm, each individual in the population represents a candidate solut
Simulating Physics with Computers
 SIAM Journal on Computing
, 1982
"... A digital computer is generally believed to be an efficient universal computing device; that is, it is believed able to simulate any physical computing device with an increase in computation time of at most a polynomial factor. This may not be true when quantum mechanics is taken into consideration. ..."
Abstract

Cited by 601 (1 self)
 Add to MetaCart
A digital computer is generally believed to be an efficient universal computing device; that is, it is believed able to simulate any physical computing device with an increase in computation time of at most a polynomial factor. This may not be true when quantum mechanics is taken into consideration. This paper considers factoring integers and finding discrete logarithms, two problems which are generally thought to be hard on a classical computer and have been used as the basis of several proposed cryptosystems. Efficient randomized algorithms are given for these two problems on a hypothetical quantum computer. These algorithms take a number of steps polynomial in the input size, e.g., the number of digits of the integer to be factored. AMS subject classifications: 82P10, 11Y05, 68Q10. 1 Introduction One of the first results in the mathematics of computation, which underlies the subsequent development of much of theoretical computer science, was the distinction between computable and ...
A New Extension of the Kalman Filter to Nonlinear Systems
, 1997
"... The Kalman filter(KF) is one of the most widely used methods for tracking and estimation due to its simplicity, optimality, tractability and robustness. However, the application of the KF to nonlinear systems can be difficult. The most common approach is to use the Extended Kalman Filter (EKF) which ..."
Abstract

Cited by 747 (6 self)
 Add to MetaCart
The Kalman filter(KF) is one of the most widely used methods for tracking and estimation due to its simplicity, optimality, tractability and robustness. However, the application of the KF to nonlinear systems can be difficult. The most common approach is to use the Extended Kalman Filter (EKF) which simply linearises all nonlinear models so that the traditional linear Kalman filter can be applied. Although the EKF (in its many forms) is a widely used filtering strategy, over thirty years of experience with it has led to a general consensus within the tracking and control community that it is difficult to implement, difficult to tune, and only reliable for systems which are almost linear on the time scale of the update intervals. In this paper a new linear estimator is developed and demonstrated. Using the principle that a set of discretely sampled points can be used to parameterise mean and covariance, the estimator yields performance equivalent to the KF for linear systems yet general...
Graphs over Time: Densification Laws, Shrinking Diameters and Possible Explanations
, 2005
"... How do real graphs evolve over time? What are “normal” growth patterns in social, technological, and information networks? Many studies have discovered patterns in static graphs, identifying properties in a single snapshot of a large network, or in a very small number of snapshots; these include hea ..."
Abstract

Cited by 534 (48 self)
 Add to MetaCart
How do real graphs evolve over time? What are “normal” growth patterns in social, technological, and information networks? Many studies have discovered patterns in static graphs, identifying properties in a single snapshot of a large network, or in a very small number of snapshots; these include heavy tails for in and outdegree distributions, communities, smallworld phenomena, and others. However, given the lack of information about network evolution over long periods, it has been hard to convert these findings into statements about trends over time. Here we study a wide range of real graphs, and we observe some surprising phenomena. First, most of these graphs densify over time, with the number of edges growing superlinearly in the number of nodes. Second, the average distance between nodes often shrinks over time, in contrast to the conventional wisdom that such distance parameters should increase slowly as a function of the number of nodes (like O(log n) orO(log(log n)). Existing graph generation models do not exhibit these types of behavior, even at a qualitative level. We provide a new graph generator, based on a “forest fire” spreading process, that has a simple, intuitive justification, requires very few parameters (like the “flammability” of nodes), and produces graphs exhibiting the full range of properties observed both in prior work and in the present study.
Algorithms for Quantum Computation: Discrete Logarithms and Factoring
, 1994
"... A computer is generally considered to be a universal computational device; i.e., it is believed able to simulate any physical computational device with a increase in computation time of at most a polynomial factor. It is not clear whether this is still true when quantum mechanics is taken into consi ..."
Abstract

Cited by 1103 (7 self)
 Add to MetaCart
A computer is generally considered to be a universal computational device; i.e., it is believed able to simulate any physical computational device with a increase in computation time of at most a polynomial factor. It is not clear whether this is still true when quantum mechanics is taken into consideration. Several researchers, starting with David Deutsch, have developed models for quantum mechanical computers and have investigated their computational properties. This paper gives Las Vegas algorithms for finding discrete logarithms and factoring integers on a quantum computer that take a number of steps which is polynomial in the input size, e.g., the number of digits of the integer to be factored. These two problems are generally considered hard on a classical computer and have been used as the basis of several proposed cryptosystems. (We thus give the first examples of quantum cryptanalysis.) 1 Introduction Since the discovery of quantum mechanics, people have found the behavior of...
Inverse Acoustic and Electromagnetic Scattering Theory, Second Edition
, 1998
"... Abstract. This paper is a survey of the inverse scattering problem for timeharmonic acoustic and electromagnetic waves at fixed frequency. We begin by a discussion of “weak scattering ” and Newtontype methods for solving the inverse scattering problem for acoustic waves, including a brief discussi ..."
Abstract

Cited by 1072 (45 self)
 Add to MetaCart
Abstract. This paper is a survey of the inverse scattering problem for timeharmonic acoustic and electromagnetic waves at fixed frequency. We begin by a discussion of “weak scattering ” and Newtontype methods for solving the inverse scattering problem for acoustic waves, including a brief discussion of Tikhonov’s method for the numerical solution of illposed problems. We then proceed to prove a uniqueness theorem for the inverse obstacle problems for acoustic waves and the linear sampling method for reconstructing the shape of a scattering obstacle from far field data. Included in our discussion is a description of Kirsch’s factorization method for solving this problem. We then turn our attention to uniqueness and reconstruction algorithms for determining the support of an inhomogeneous, anisotropic media from acoustic far field data. Our survey is concluded by a brief discussion of the inverse scattering problem for timeharmonic electromagnetic waves. 1.
Results 1  10
of
146,498