• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 33,078
Next 10 →

Example Experiments

by Christoph Koch, Universität Des Saarlandes, Stefanie Scherzinger, Universität Des Saarlandes, Nicole Schweikardt, Bernhard Stegmaier, Technische Universität München, Poster Cătălin Hriţcu, International Max
"... Flux is an extension of the XQuery language, that supports event-based query processing of XML Streams. The main goal is to minimize the amount of buffering, and is achieved by using order constraints from a DTD. We also present an efficient algorithm for automatically translating a significant frag ..."
Abstract - Add to MetaCart
Flux is an extension of the XQuery language, that supports event-based query processing of XML Streams. The main goal is to minimize the amount of buffering, and is achieved by using order constraints from a DTD. We also present an efficient algorithm for automatically translating a significant fragment of XQuery into equivalent FluX queries. FluX is intended as an internal representation format for queries rather than a language for end-users and provides a strong intuition for buffer-conscious query processing on structured data streams.

Discriminative Training Methods for Hidden Markov Models: Theory and Experiments with Perceptron Algorithms

by Michael Collins , 2002
"... We describe new algorithms for training tagging models, as an alternative to maximum-entropy models or conditional random fields (CRFs). The algorithms rely on Viterbi decoding of training examples, combined with simple additive updates. We describe theory justifying the algorithms through a modific ..."
Abstract - Cited by 660 (13 self) - Add to MetaCart
We describe new algorithms for training tagging models, as an alternative to maximum-entropy models or conditional random fields (CRFs). The algorithms rely on Viterbi decoding of training examples, combined with simple additive updates. We describe theory justifying the algorithms through a

»  Intuition and examples »  Experiments

by Copyright Michael R. Roberts, Prof Michael, R. Roberts, Copyright Michael R. Roberts, Single Difference Estimators
"... ..."
Abstract - Add to MetaCart
Abstract not found

SMOTE: Synthetic Minority Over-sampling Technique

by Nitesh V. Chawla, Kevin W. Bowyer, Lawrence O. Hall, W. Philip Kegelmeyer - Journal of Artificial Intelligence Research , 2002
"... An approach to the construction of classifiers from imbalanced datasets is described. A dataset is imbalanced if the classification categories are not approximately equally represented. Often real-world data sets are predominately composed of ``normal'' examples with only a small percentag ..."
Abstract - Cited by 634 (27 self) - Add to MetaCart
-sampling the minority class involves creating synthetic minority class examples. Experiments are performed using C4.5, Ripper and a Naive Bayes classifier. The method is evaluated using the area under the Receiver Operating Characteristic curve (AUC) and the ROC convex hull strategy.

Literate programming

by Donald E. Knuth - THE COMPUTER JOURNAL , 1984
"... The author and his associates have been experimenting for the past several years with a programming language and documentation system called WEB. This paper presents WEB by example, and discusses why the new system appears to be an improvement over previous ones. ..."
Abstract - Cited by 557 (3 self) - Add to MetaCart
The author and his associates have been experimenting for the past several years with a programming language and documentation system called WEB. This paper presents WEB by example, and discusses why the new system appears to be an improvement over previous ones.

Transductive Inference for Text Classification using Support Vector Machines

by Thorsten Joachims , 1999
"... This paper introduces Transductive Support Vector Machines (TSVMs) for text classification. While regular Support Vector Machines (SVMs) try to induce a general decision function for a learning task, Transductive Support Vector Machines take into account a particular test set and try to minimiz ..."
Abstract - Cited by 892 (4 self) - Add to MetaCart
to minimize misclassifications of just those particular examples. The paper presents an analysis of why TSVMs are well suited for text classification. These theoretical findings are supported by experiments on three test collections. The experiments show substantial improvements over inductive methods

Boosting the margin: A new explanation for the effectiveness of voting methods

by Robert E. Schapire, Yoav Freund, Peter Bartlett, Wee Sun Lee - IN PROCEEDINGS INTERNATIONAL CONFERENCE ON MACHINE LEARNING , 1997
"... One of the surprising recurring phenomena observed in experiments with boosting is that the test error of the generated classifier usually does not increase as its size becomes very large, and often is observed to decrease even after the training error reaches zero. In this paper, we show that this ..."
Abstract - Cited by 897 (52 self) - Add to MetaCart
One of the surprising recurring phenomena observed in experiments with boosting is that the test error of the generated classifier usually does not increase as its size becomes very large, and often is observed to decrease even after the training error reaches zero. In this paper, we show

Optimal Brain Damage

by Yann Le Cun, John S. Denker, Sara A. Sola , 1990
"... We have used information-theoretic ideas to derive a class of practical and nearly optimal schemes for adapting the size of a neural network. By removing unimportant weights from a network, several improvements can be expected: better generalization, fewer training examples required, and improved sp ..."
Abstract - Cited by 510 (5 self) - Add to MetaCart
We have used information-theoretic ideas to derive a class of practical and nearly optimal schemes for adapting the size of a neural network. By removing unimportant weights from a network, several improvements can be expected: better generalization, fewer training examples required, and improved

The SPLASH-2 programs: Characterization and methodological considerations

by Steven Cameron Woo, Moriyoshi Ohara, Evan Torrie, Jaswinder Pal Singh, Anoop Gupta - INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE , 1995
"... The SPLASH-2 suite of parallel applications has recently been released to facilitate the study of centralized and distributed shared-address-space multiprocessors. In this context, this paper has two goals. One is to quantitatively characterize the SPLASH-2 programs in terms of fundamental propertie ..."
Abstract - Cited by 1420 (12 self) - Add to MetaCart
scale with problem size and the number of processors. The other, related goal is methodological: to assist people who will use the programs in architectural evaluations to prune the space of application and machine parameters in an informed and meaningful way. For example, by characterizing the working

Optimizing Search Engines using Clickthrough Data

by Thorsten Joachims , 2002
"... This paper presents an approach to automatically optimizing the retrieval quality of search engines using clickthrough data. Intuitively, a good information retrieval system should present relevant documents high in the ranking, with less relevant documents following below. While previous approaches ..."
Abstract - Cited by 1314 (23 self) - Add to MetaCart
approaches to learning retrieval functions from examples exist, they typically require training data generated from relevance judgments by experts. This makes them difficult and expensive to apply. The goal of this paper is to develop a method that utilizes clickthrough data for training, namely the query
Next 10 →
Results 1 - 10 of 33,078
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University