Reinforcement Learning from Scarce Experience Via Policy Search, Leonid Peshkin (9783639088038) — Readings Books

Become a Readings Member to make your shopping experience even easier. Sign in or sign up for free!

Become a Readings Member. Sign in or sign up for free!

Hello Readings Member! Go to the member centre to view your orders, change your details, or view your lists, or sign out.

Hello Readings Member! Go to the member centre or sign out.

Reinforcement Learning from Scarce Experience Via Policy Search
Paperback

Reinforcement Learning from Scarce Experience Via Policy Search

$88.99
Sign in or become a Readings Member to add this title to your wishlist.

This title is printed to order. This book may have been self-published. If so, we cannot guarantee the quality of the content. In the main most books will have gone through the editing process however some may not. We therefore suggest that you be aware of this before ordering this book. If in doubt check either the author or publisher’s details as we are unable to accept any returns unless they are faulty. Please contact us if you have any questions.

Today we live in the world which is very much a man-made or artificial. In such a world there are many systems and environments, both real and virtual, which can be very well described by formal models. This creates an opportunity for developing a synthetic intelligence - artificial systems which cohabit these environments with human beings and carry out some useful function. In this book we address some aspects of this development in the framework of reinforcement learning, learning how to map sensations to actions, by trial and error from feedback. In some challenging cases, actions may affect not only the immediate reward, but also the next sensation and all subsequent rewards. The general task of reinforcement learning stated in a traditional way is unreasonably ambitious for these two characteristics: search by trial-and-error and delayed reward. We investigate general ways of breaking the task of designing a controller down to more feasible sub-tasks which are solved independently. We propose to consider both taking advantage of past experience by reusing parts of other systems, and facilitating the learning phase by employing a bias in initial configuration.

Read More
In Shop
Out of stock
Shipping & Delivery

$9.00 standard shipping within Australia
FREE standard shipping within Australia for orders over $100.00
Express & International shipping calculated at checkout

MORE INFO

Stock availability can be subject to change without notice. We recommend calling the shop or contacting our online team to check availability of low stock items. Please see our Shopping Online page for more details.

Format
Paperback
Publisher
VDM Verlag Dr. Muller Aktiengesellschaft & Co. KG
Country
Germany
Date
5 December 2008
Pages
140
ISBN
9783639088038

This title is printed to order. This book may have been self-published. If so, we cannot guarantee the quality of the content. In the main most books will have gone through the editing process however some may not. We therefore suggest that you be aware of this before ordering this book. If in doubt check either the author or publisher’s details as we are unable to accept any returns unless they are faulty. Please contact us if you have any questions.

Today we live in the world which is very much a man-made or artificial. In such a world there are many systems and environments, both real and virtual, which can be very well described by formal models. This creates an opportunity for developing a synthetic intelligence - artificial systems which cohabit these environments with human beings and carry out some useful function. In this book we address some aspects of this development in the framework of reinforcement learning, learning how to map sensations to actions, by trial and error from feedback. In some challenging cases, actions may affect not only the immediate reward, but also the next sensation and all subsequent rewards. The general task of reinforcement learning stated in a traditional way is unreasonably ambitious for these two characteristics: search by trial-and-error and delayed reward. We investigate general ways of breaking the task of designing a controller down to more feasible sub-tasks which are solved independently. We propose to consider both taking advantage of past experience by reusing parts of other systems, and facilitating the learning phase by employing a bias in initial configuration.

Read More
Format
Paperback
Publisher
VDM Verlag Dr. Muller Aktiengesellschaft & Co. KG
Country
Germany
Date
5 December 2008
Pages
140
ISBN
9783639088038