• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

DMCA

Planning and acting in partially observable stochastic domains (1998)

Cached

  • Download as a PDF

Download Links

  • [www.cis.upenn.edu]
  • [www.cs.tufts.edu]
  • [ftp.cs.brown.edu]
  • [www.cs.duke.edu]
  • [www.cis.upenn.edu]
  • [damas.ift.ulaval.ca]
  • [www.damas.ift.ulaval.ca]
  • [www.cs.duke.edu]
  • [www.cs.brown.edu]
  • [msl.cs.uiuc.edu]
  • [www.cs.ubc.ca]
  • [www.ai.mit.edu]
  • [csail.mit.edu]
  • [people.csail.mit.edu]
  • [staff.science.uva.nl]
  • [www.cs.rutgers.edu]
  • [people.csail.mit.edu]
  • [www.cs.rutgers.edu]
  • [classes.engr.oregonstate.edu]
  • [elite.polito.it]

  • Other Repositories/Bibliography

  • DBLP
  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Leslie Pack Kaelbling , Michael L. Littman , Anthony R. Cassandra
Venue:ARTIFICIAL INTELLIGENCE
Citations:1095 - 38 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@ARTICLE{Kaelbling98planningand,
    author = {Leslie Pack Kaelbling and Michael L. Littman and Anthony R. Cassandra},
    title = {Planning and acting in partially observable stochastic domains},
    journal = {ARTIFICIAL INTELLIGENCE},
    year = {1998},
    volume = {101},
    pages = {99--134}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

In this paper, we bring techniques from operations research to bear on the problem of choosing optimal actions in partially observable stochastic domains. We begin by introducing the theory of Markov decision processes (mdps) and partially observable mdps (pomdps). We then outline a novel algorithm for solving pomdps offline and show how, in some cases, a finite-memory controller can be extracted from the solution to a pomdp. We conclude with a discussion of how our approach relates to previous work, the complexity of finding exact solutions to pomdps, and of some possibilities for finding approximate solutions.

Keyphrases

observable stochastic domain    optimal action    pomdps offline    approximate solution    observable mdps    exact solution    previous work    operation research    novel algorithm    markov decision process    finite-memory controller   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University