• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 11,662
Next 10 →

The strength of weak learnability

by Robert E. Schapire - MACHINE LEARNING , 1990
"... This paper addresses the problem of improving the accuracy of an hypothesis output by a learning algorithm in the distribution-free (PAC) learning model. A concept class is learnable (or strongly learnable) if, given access to a Source of examples of the unknown concept, the learner with high prob ..."
Abstract - Cited by 871 (26 self) - Add to MetaCart
probability is able to output an hypothesis that is correct on all but an arbitrarily small fraction of the instances. The concept class is weakly learnable if the learner can produce an hypothesis that performs only slightly better than random guessing. In this paper, it is shown that these two notions

Integration of trade and disintegration of production in the global economy

by Robert C. Feenstra - Journal of Economic Perspectives , 1998
"... The last few decades have seen a spectacular integration of the global economy through trade. The rising integration of world markets has brought with it a disintegration of the production process, however, as manufacturing or services activities done abroad are combined with those performed at home ..."
Abstract - Cited by 496 (7 self) - Add to MetaCart
seemingly small fraction of U.S. GDP. This is not surprising in view of the fact that large economies trade less with others, and more internally. But the modest share of trade in total national income hides the fact that merchandise trade as a share of merchandise value-added is quite high for the U

Secure routing for structured peer-to-peer overlay networks

by Miguel Castro, Peter Druschel, Ayalvadi Ganesh, Antony Rowstron, Dan S. Wallach , 2002
"... Structured peer-to-peer overlay networks provide a sub-strate for the construction of large-scale, decentralized applications, including distributed storage, group com-munication, and content distribution. These overlays are highly resilient; they can route messages correctly even when a large fract ..."
Abstract - Cited by 473 (12 self) - Add to MetaCart
fraction of the nodes crash or the network partitions. But current overlays are not secure; even a small fraction of malicious nodes can prevent correct message delivery throughout the overlay. This prob-lem is particularly serious in open peer-to-peer systems, where many diverse, autonomous parties

The Cache Performance and Optimizations of Blocked Algorithms

by Monica S. Lam, Edward E. Rothberg, Michael E. Wolf - In Proceedings of the Fourth International Conference on Architectural Support for Programming Languages and Operating Systems , 1991
"... Blocking is a well-known optimization technique for improving the effectiveness of memory hierarchies. Instead of operating on entire rows or columns of an array, blocked algorithms operate on submatrices or blocks, so that data loaded into the faster levels of the memory hierarchy are reused. This ..."
Abstract - Cited by 574 (5 self) - Add to MetaCart
is highly sensitive to the stride of data accesses and the size of the blocks, and can cause wide variations in machine performance for different matrix sizes. The conventional wisdom of trying to use the entire cache, or even a fixed fraction of the cache, is incorrect. If a fixed block size is used for a

Accurate Methods for the Statistics of Surprise and Coincidence

by Ted Dunning - COMPUTATIONAL LINGUISTICS , 1993
"... Much work has been done on the statistical analysis of text. In some cases reported in the literature, inappropriate statistical methods have been used, and statistical significance of results have not been addressed. In particular, asymptotic normality assumptions have often been used unjustifiably ..."
Abstract - Cited by 1057 (1 self) - Add to MetaCart
unjustifiably, leading to flawed results.This assumption of normal distribution limits the ability to analyze rare events. Unfortunately rare events do make up a large fraction of real text.However, more applicable methods based on likelihood ratio tests are available that yield good results with relatively

LOWER BOUNDS FOR SMALL FRACTIONAL MOMENTS OF

by Vorrapan Chandee, Xiannan Li
"... ar ..."
Abstract - Add to MetaCart
Abstract not found

Enhanced hypertext categorization using hyperlinks

by Soumen Chakrabarti, Byron Dom, Piotr Indyk , 1998
"... A major challenge in indexing unstructured hypertext databases is to automatically extract meta-data that enables structured search using topic taxonomies, circumvents keyword ambiguity, and improves the quality of search and profile-based routing and filtering. Therefore, an accurate classifier is ..."
Abstract - Cited by 453 (8 self) - Add to MetaCart
. Naive use of terms in the link neighborhood of a document can even degrade accuracy. Our contribution is to propose robust statistical models and a relaxation labeling technique for better classification by exploiting link information in a small neighborhood around documents. Our technique also adapts

A fast and flexible statistical model for large-scale population genotype data: Applications to inferring missing genotypes and haplotype phase

by Paul Scheet, Matthew Stephens - American Journal of Human Genetics , 2005
"... We present a statistical model for patterns of genetic variation in samples of unrelated individuals from natural populations. This model is based on the idea that, over short regions, haplotypes in a population tend to cluster into groups of similar haplotypes. To capture the fact that, because of ..."
Abstract - Cited by 408 (10 self) - Add to MetaCart
Centre d’Etude du Polymorphisme Humain individuals from the HapMap project, switch error was 0.055 for our method vs. 0.051 for PHASE) but require a small fraction of the computational cost. In addition, we demonstrate that the model accurately reflects uncertainty in its estimates, in that probabilities

Combating web spam with trustrank

by Zoltán Gyöngyi, Hector Garcia-molina, Jan Pedersen - In VLDB , 2004
"... Web spam pages use various techniques to achieve higher-than-deserved rankings in a search engine’s results. While human experts can identify spam, it is too expensive to manually evaluate a large number of pages. Instead, we propose techniques to semi-automatically separate reputable, good pages fr ..."
Abstract - Cited by 413 (3 self) - Add to MetaCart
from spam. We first select a small set of seed pages to be evaluated by an expert. Once we manually identify the reputable seed pages, we use the link structure of the web to discover other pages that are likely to be good. In this paper we discuss possible ways to implement the seed selection

The Power of Convex Relaxation: Near-Optimal Matrix Completion

by Emmanuel J. Candès, Terence Tao , 2009
"... This paper is concerned with the problem of recovering an unknown matrix from a small fraction of its entries. This is known as the matrix completion problem, and comes up in a great number of applications, including the famous Netflix Prize and other similar questions in collaborative filtering. In ..."
Abstract - Cited by 359 (7 self) - Add to MetaCart
This paper is concerned with the problem of recovering an unknown matrix from a small fraction of its entries. This is known as the matrix completion problem, and comes up in a great number of applications, including the famous Netflix Prize and other similar questions in collaborative filtering
Next 10 →
Results 1 - 10 of 11,662
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University