• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

DMCA

How to Use Expert Advice (1997)

Cached

  • Download as a PDF

Download Links

  • [www.cis.upenn.edu]
  • [www.cse.ucsc.edu]
  • [ftp.cse.ucsc.edu]
  • [www.informatik.uni-bonn.de]
  • [ftp.cse.ucsc.edu]
  • [www.cbse.ucsc.edu]
  • [www.cs.ucsc.edu]
  • [www.soe.ucsc.edu]
  • [www.cse.ucsc.edu]
  • [homes.dsi.unimi.it]
  • [mercurio.srv.di.unimi.it]
  • [users.soe.ucsc.edu]
  • [homes.di.unimi.it]
  • [www.stanford.edu]
  • [users.soe.ucsc.edu]
  • [users.soe.ucsc.edu]
  • [mercurio.srv.di.unimi.it]
  • [www.stanford.edu]
  • [homes.di.unimi.it]

  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Nicolò Cesa-Bianchi , Yoav Freund , David Haussler , David P. Helmbold , Robert E. Schapire , Manfred K. Warmuth
Venue:JOURNAL OF THE ASSOCIATION FOR COMPUTING MACHINERY
Citations:377 - 79 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@MISC{Cesa-Bianchi97howto,
    author = {Nicolò Cesa-Bianchi and Yoav Freund and David Haussler and David P. Helmbold and Robert E. Schapire and Manfred K. Warmuth},
    title = { How to Use Expert Advice},
    year = {1997}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

We analyze algorithms that predict a binary value by combining the predictions of several prediction strategies, called experts. Our analysis is for worst-case situations, i.e., we make no assumptions about the way the sequence of bits to be predicted is generated. We measure the performance of the algorithm by the difference between the expected number of mistakes it makes on the bit sequence and the expected number of mistakes made by the best expert on this sequence, where the expectation is taken with respect to the randomization in the predictions. We show that the minimum achievable difference is on the order of the square root of the number of mistakes of the best expert, and we give efficient algorithms that achieve this. Our upper and lower bounds have matching leading constants in most cases. We then show howthis leads to certain kinds of pattern recognition/learning algorithms with performance bounds that improve on the best results currently known in this context. We also compare our analysis to the case in which log loss is used instead of the expected number of mistakes.

Keyphrases

expected number    expert advice    certain kind    binary value    square root    bit sequence    pattern recognition    minimum achievable difference    howthis lead    performance bound    several prediction strategy    efficient algorithm    worst-case situation    log loss   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University