• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 7,666
Next 10 →

KEPLER BINARY CASE STUDIES

by Paul A. Mason, Jorge I. Zuluaga, Joni M. Clark, Pablo A. Cuartas-restrepo , 2014
"... ar ..."
Abstract - Add to MetaCart
Abstract not found

Self-adjusting binary search trees

by Daniel Dominic Sleator, Robert Endre Tarjan , 1985
"... The splay tree, a self-adjusting form of binary search tree, is developed and analyzed. The binary search tree is a data structure for representing tables and lists so that accessing, inserting, and deleting items is easy. On an n-node splay tree, all the standard search tree operations have an am ..."
Abstract - Cited by 432 (18 self) - Add to MetaCart
The splay tree, a self-adjusting form of binary search tree, is developed and analyzed. The binary search tree is a data structure for representing tables and lists so that accessing, inserting, and deleting items is easy. On an n-node splay tree, all the standard search tree operations have

Eraser: a dynamic data race detector for multithreaded programs

by Stefan Savage, Michael Burrows, Greg Nelson, Patrick Sobalvarro, Thomas Anderson - ACM Transaction of Computer System , 1997
"... Multi-threaded programming is difficult and error prone. It is easy to make a mistake in synchronization that produces a data race, yet it can be extremely hard to locate this mistake during debugging. This paper describes a new tool, called Eraser, for dynamically detecting data races in lock-based ..."
Abstract - Cited by 688 (2 self) - Add to MetaCart
-based multi-threaded programs. Eraser uses binary rewriting techniques to monitor every shared memory reference and verify that consistent locking behavior is observed. We present several case studies, including undergraduate coursework and a multi-threaded Web search engine, that demonstrate

The Capacity of Low-Density Parity-Check Codes Under Message-Passing Decoding

by Thomas J. Richardson, Rüdiger L. Urbanke , 2001
"... In this paper, we present a general method for determining the capacity of low-density parity-check (LDPC) codes under message-passing decoding when used over any binary-input memoryless channel with discrete or continuous output alphabets. Transmitting at rates below this capacity, a randomly chos ..."
Abstract - Cited by 574 (9 self) - Add to MetaCart
. Our results are based on the observation that the concentration of the performance of the decoder around its average performance, as observed by Luby et al. [1] in the case of a binary-symmetric channel and a binary message-passing algorithm, is a general phenomenon. For the particularly important

Boosting a Weak Learning Algorithm By Majority

by Yoav Freund , 1995
"... We present an algorithm for improving the accuracy of algorithms for learning binary concepts. The improvement is achieved by combining a large number of hypotheses, each of which is generated by training the given learning algorithm on a different set of examples. Our algorithm is based on ideas pr ..."
Abstract - Cited by 516 (16 self) - Add to MetaCart
We present an algorithm for improving the accuracy of algorithms for learning binary concepts. The improvement is achieved by combining a large number of hypotheses, each of which is generated by training the given learning algorithm on a different set of examples. Our algorithm is based on ideas

Proof-Carrying Code

by George C. Necula , 1997
"... This paper describes proof-carrying code (PCC), a mechanism by which a host system can determine with certainty that it is safe to execute a program supplied (possibly in binary form) by an untrusted source. For this to be possible, the untrusted code producer must supply with the code a safety proo ..."
Abstract - Cited by 1240 (27 self) - Add to MetaCart
This paper describes proof-carrying code (PCC), a mechanism by which a host system can determine with certainty that it is safe to execute a program supplied (possibly in binary form) by an untrusted source. For this to be possible, the untrusted code producer must supply with the code a safety

Xen and the art of virtualization

by Paul Barham, Boris Dragovic, Keir Fraser, Steven Hand, Tim Harris, Alex Ho, Rolf Neugebauer, Ian Pratt, Andrew Warfield - IN SOSP , 2003
"... Numerous systems have been designed which use virtualization to subdivide the ample resources of a modern computer. Some require specialized hardware, or cannot support commodity operating systems. Some target 100 % binary compatibility at the expense of performance. Others sacrifice security or fun ..."
Abstract - Cited by 2010 (35 self) - Add to MetaCart
Numerous systems have been designed which use virtualization to subdivide the ample resources of a modern computer. Some require specialized hardware, or cannot support commodity operating systems. Some target 100 % binary compatibility at the expense of performance. Others sacrifice security

Comparing Images Using the Hausdorff Distance

by Daniel P. Huttenlocher, Gregory A. Klanderman, William J. Rucklidge - IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE , 1993
"... The Hausdorff distance measures the extent to which each point of a `model' set lies near some point of an `image' set and vice versa. Thus this distance can be used to determine the degree of resemblance between two objects that are superimposed on one another. In this paper we provide ef ..."
Abstract - Cited by 659 (10 self) - Add to MetaCart
efficient algorithms for computing the Hausdorff distance between all possible relative positions of a binary image and a model. We focus primarily on the case in which the model is only allowed to translate with respect to the image. Then we consider how to extend the techniques to rigid motion

Cumulated Gain-based Evaluation of IR Techniques

by Kalervo Järvelin, Jaana Kekäläinen - ACM Transactions on Information Systems , 2002
"... Modem large retrieval environments tend to overwhelm their users by their large output. Since all documents are not of equal relevance to their users, highly relevant documents should be identified and ranked first for presentation to the users. In order to develop IR techniques to this direction, i ..."
Abstract - Cited by 694 (3 self) - Add to MetaCart
, it is necessary to develop evaluation approaches and methods that credit IR methods for their ability to retrieve highly relevant documents. This can be done by extending traditional evaluation methods, i.e., recall and precision based on binary relevance assessments, to graded relevance assessments

The rate-distortion function for source coding with side information at the decoder

by Aaron D. Wyner, Jacob Ziv - IEEE Trans. Inform. Theory , 1976
"... Abstract-Let {(X,, Y,J}r = 1 be a sequence of independent drawings of a pair of dependent random variables X, Y. Let us say that X takes values in the finite set 6. It is desired to encode the sequence {X,} in blocks of length n into a binary stream*of rate R, which can in turn be decoded as a seque ..."
Abstract - Cited by 1060 (1 self) - Add to MetaCart
Abstract-Let {(X,, Y,J}r = 1 be a sequence of independent drawings of a pair of dependent random variables X, Y. Let us say that X takes values in the finite set 6. It is desired to encode the sequence {X,} in blocks of length n into a binary stream*of rate R, which can in turn be decoded as a
Next 10 →
Results 1 - 10 of 7,666
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University