Results 1  10
of
572,525
Molecular Computation Of Solutions To Combinatorial Problems
, 1994
"... The tools of molecular biology are used to solve an instance of the directed Hamiltonian path problem. A small graph is encoded in molecules of DNA and the `operations' of the computation are performed with standard protocols and enzymes. This experiment demonstrates the feasibility of carrying ..."
Abstract

Cited by 763 (6 self)
 Add to MetaCart
The tools of molecular biology are used to solve an instance of the directed Hamiltonian path problem. A small graph is encoded in molecules of DNA and the `operations' of the computation are performed with standard protocols and enzymes. This experiment demonstrates the feasibility
A New Method for Solving Hard Satisfiability Problems
 AAAI
, 1992
"... We introduce a greedy local search procedure called GSAT for solving propositional satisfiability problems. Our experiments show that this procedure can be used to solve hard, randomly generated problems that are an order of magnitude larger than those that can be handled by more traditional approac ..."
Abstract

Cited by 725 (21 self)
 Add to MetaCart
approaches such as the DavisPutnam procedure or resolution. We also show that GSAT can solve structured satisfiability problems quickly. In particular, we solve encodings of graph coloring problems, Nqueens, and Boolean induction. General application strategies and limitations of the approach are also
Pushing the Envelope: Planning, Propositional Logic, and Stochastic Search
, 1996
"... Planning is a notoriously hard combinatorial search problem. In many interesting domains, current planning algorithms fail to scale up gracefully. By combining a general, stochastic search algorithm and appropriate problem encodings based on propositional logic, we are able to solve hard planning pr ..."
Abstract

Cited by 574 (32 self)
 Add to MetaCart
Planning is a notoriously hard combinatorial search problem. In many interesting domains, current planning algorithms fail to scale up gracefully. By combining a general, stochastic search algorithm and appropriate problem encodings based on propositional logic, we are able to solve hard planning
Completely Derandomized SelfAdaptation in Evolution Strategies
 Evolutionary Computation
, 2001
"... This paper puts forward two useful methods for selfadaptation of the mutation distribution  the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the selfadapta ..."
Abstract

Cited by 533 (58 self)
 Add to MetaCart
adaptation of arbitrary (normal) mutation distributions are developed. Applying arbitrary, normal mutation distributions is equivalent to applying a general, linear problem encoding.
Learning the Kernel Matrix with SemiDefinite Programming
, 2002
"... Kernelbased learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract

Cited by 769 (21 self)
 Add to MetaCart
is contained in the socalled kernel matrix, a symmetric and positive definite matrix that encodes the relative positions of all points. Specifying this matrix amounts to specifying the geometry of the embedding space and inducing a notion of similarity in the input spaceclassical model selection
Markov Random Field Models in Computer Vision
, 1994
"... . A variety of computer vision problems can be optimally posed as Bayesian labeling in which the solution of a problem is defined as the maximum a posteriori (MAP) probability estimate of the true labeling. The posterior probability is usually derived from a prior model and a likelihood model. The l ..."
Abstract

Cited by 510 (18 self)
 Add to MetaCart
. The latter relates to how data is observed and is problem domain dependent. The former depends on how various prior constraints are expressed. Markov Random Field Models (MRF) theory is a tool to encode contextual constraints into the prior probability. This paper presents a unified approach for MRF modeling
Text Chunking using TransformationBased Learning
, 1995
"... Eric Brill introduced transformationbased learning and showed that it can do partofspeech tagging with fairly high accuracy. The same method can be applied at a higher level of textual interpretation for locating chunks in the tagged text, including nonrecursive "baseNP" chunks. For ..."
Abstract

Cited by 511 (0 self)
 Add to MetaCart
. For this purpose, it is convenient to view chunking as a tagging problem by encoding the chunk structure in new tags attached to each word. In automatic tests using Treebankderived data, this technique achieved recall and precision rates of roughly 92% for baseNP chunks and 88% for somewhat more complex chunks
Efficient dispersal of information for security, load balancing, and fault tolerance
 Journal of the ACM
, 1989
"... Abstract. An Information Dispersal Algorithm (IDA) is developed that breaks a file F of length L = ( F ( into n pieces F,, 1 5 i 5 n, each of length ( F, 1 = L/m, so that every m pieces suffice for reconstructing F. Dispersal and reconstruction are computationally efficient. The sum of the lengths ..."
Abstract

Cited by 555 (1 self)
 Add to MetaCart
cations between processors in parallel computers. For the latter problem provably timeefftcient and highly faulttolerant routing on the ncube is achieved, using just constant size buffers. Categories and Subject Descriptors: E.4 [Coding and Information Theory]: nonsecret encoding schemes
SplitStream: HighBandwidth Multicast in Cooperative Environments
 SOSP '03
, 2003
"... In treebased multicast systems, a relatively small number of interior nodes carry the load of forwarding multicast messages. This works well when the interior nodes are highly available, d d cated infrastructure routers but it poses a problem for applicationlevel multicast in peertopeer systems. ..."
Abstract

Cited by 568 (17 self)
 Add to MetaCart
In treebased multicast systems, a relatively small number of interior nodes carry the load of forwarding multicast messages. This works well when the interior nodes are highly available, d d cated infrastructure routers but it poses a problem for applicationlevel multicast in peertopeer systems
A learning algorithm for Boltzmann machines
 Cognitive Science
, 1985
"... The computotionol power of massively parallel networks of simple processing elements resides in the communication bandwidth provided by the hardware connections between elements. These connections con allow a significant fraction of the knowledge of the system to be applied to an instance of a probl ..."
Abstract

Cited by 576 (13 self)
 Add to MetaCart
problem in o very short time. One kind of computation for which massively porollel networks appear to be well suited is large constraint satisfaction searches, but to use the connections efficiently two conditions must be met: First, a search technique that is suitable for parallel networks must be found
Results 1  10
of
572,525