Newsgroups: comp.ai.jair.announce
Path: cantaloupe.srv.cs.cmu.edu!bb3.andrew.cmu.edu!newsfeed.pitt.edu!godot.cc.duq.edu!news.duke.edu!news.mathworks.com!newsfeed.internetmci.com!news.msfc.nasa.gov!pendragon!ames!eos!kronos.arc.nasa.gov!jair-ed
From: jair-ed@ptolemy.arc.nasa.gov
Subject: New Article, Reinforcement Learning: A Survey ...
Message-ID: <1996May4.232544.22632@ptolemy-ethernet.arc.nasa.gov>
Originator: jair-ed@polya.arc.nasa.gov
Lines: 40
Sender: usenet@ptolemy-ethernet.arc.nasa.gov (usenet@ptolemy.arc.nasa.gov)
Nntp-Posting-Host: polya.arc.nasa.gov
Organization: NASA/ARC Computational Sciences Division
Date: Sat, 4 May 1996 23:25:44 GMT
Approved: jair-ed@ptolemy.arc.nasa.gov

JAIR is pleased to announce the publication of the following article:

Kaelbling, L.P., Littman, M.L., and Moore, A.W. (1996)  "Reinforcement 
   Learning:  A Survey", Volume 4, pages 237-285.

   Available in HTML, Postscript (903K) and compressed Postscript (362K).
   For quick access via your WWW browser, use this URL:
     http://www.cs.washington.edu/research/jair/abstracts/kaelbling96a.html
   More detailed instructions are below.

   Abstract: This paper surveys the field of reinforcement learning from
   a computer-science perspective. It is written to be accessible to
   researchers familiar with machine learning.  Both the historical basis
   of the field and a broad selection of current work are summarized.
   Reinforcement learning is the problem faced by an agent that learns
   behavior through trial-and-error interactions with a dynamic
   environment.  The work described here has a resemblance to work in
   psychology, but differs considerably in the details and in the use of
   the word ``reinforcement.''  The paper discusses central issues of
   reinforcement learning, including trading off exploration and
   exploitation, establishing the foundations of the field via Markov
   decision theory, learning from delayed reinforcement, constructing
   empirical models to accelerate learning, making use of generalization
   and hierarchy, and coping with hidden state.  It concludes with a
   survey of some implemented systems and an assessment of the practical
   utility of current methods for reinforcement learning.

The article is available via:
   
 -- comp.ai.jair.papers (also see comp.ai.jair.announce)

 -- World Wide Web: The URL for our World Wide Web server is
       http://www.cs.washington.edu/research/jair/home.html
    For direct access to this article and related files try:
       http://www.cs.washington.edu/research/jair/abstracts/kaelbling96a.html

 -- Anonymous FTP from either of the two sites below.

    Carnegie-Mellon University (USA):
	ftp://p.gp.cs.cmu.edu/usr/jair/pub/volume4/kaelbling96a.ps
    The University of Genoa (Italy):
	ftp://ftp.mrg.dist.unige.it/pub/jair/pub/volume4/kaelbling96a.ps

    The compressed PostScript file is named kaelbling96a.ps.Z (362K)

 -- automated email. Send mail to jair@cs.cmu.edu or jair@ftp.mrg.dist.unige.it
    with the subject AUTORESPOND and our automailer will respond. To
    get the Postscript file, use the message body GET volume4/kaelbling96a.ps 
    (Note: Your mailer might find this file too large to handle.) 
    Only one can file be requested in each message.

 -- JAIR Gopher server: At p.gp.cs.cmu.edu, port 70. 

For more information about JAIR, visit our WWW or FTP sites, or
send electronic mail to jair@cs.cmu.edu with the subject AUTORESPOND
and the message body HELP, or contact jair-ed@ptolemy.arc.nasa.gov.



