109,16 €
121,29 €
-10% with code: EXTRA
Reinforcement Learning from Scarce Experience via Policy Search
Reinforcement Learning from Scarce Experience via Policy Search
109,16
121,29 €
  • We will send in 10–14 business days.
Today we live in the world which is very much a man-made or artificial. In such a world there are many systems and environments, both real and virtual, which can be very well described by formal models. This creates an opportunity for developing a "synthetic intelligence" - artificial systems which cohabit these environments with human beings and carry out some useful function. In this book we address some aspects of this development in the framework of reinforcement learning, learning how to m…
121.29
  • Publisher:
  • ISBN-10: 3639088034
  • ISBN-13: 9783639088038
  • Format: 15.2 x 22.9 x 0.8 cm, minkšti viršeliai
  • Language: English
  • SAVE -10% with code: EXTRA

Reinforcement Learning from Scarce Experience via Policy Search (e-book) (used book) | bookbook.eu

Reviews

Description

Today we live in the world which is very much a man-made or artificial. In such a world there are many systems and environments, both real and virtual, which can be very well described by formal models. This creates an opportunity for developing a "synthetic intelligence" - artificial systems which cohabit these environments with human beings and carry out some useful function. In this book we address some aspects of this development in the framework of reinforcement learning, learning how to map sensations to actions, by trial and error from feedback. In some challenging cases, actions may affect not only the immediate reward, but also the next sensation and all subsequent rewards. The general task of reinforcement learning stated in a traditional way is unreasonably ambitious for these two characteristics: search by trial-and-error and delayed reward. We investigate general ways of breaking the task of designing a controller down to more feasible sub-tasks which are solved independently. We propose to consider both taking advantage of past experience by reusing parts of other systems, and facilitating the learning phase by employing a bias in initial configuration.

EXTRA 10 % discount with code: EXTRA

109,16
121,29 €
We will send in 10–14 business days.

The promotion ends in 21d.23:21:47

The discount code is valid when purchasing from 10 €. Discounts do not stack.

Log in and for this item
you will receive 1,21 Book Euros!?
  • Author: Leonid Peshkin
  • Publisher:
  • ISBN-10: 3639088034
  • ISBN-13: 9783639088038
  • Format: 15.2 x 22.9 x 0.8 cm, minkšti viršeliai
  • Language: English English

Today we live in the world which is very much a man-made or artificial. In such a world there are many systems and environments, both real and virtual, which can be very well described by formal models. This creates an opportunity for developing a "synthetic intelligence" - artificial systems which cohabit these environments with human beings and carry out some useful function. In this book we address some aspects of this development in the framework of reinforcement learning, learning how to map sensations to actions, by trial and error from feedback. In some challenging cases, actions may affect not only the immediate reward, but also the next sensation and all subsequent rewards. The general task of reinforcement learning stated in a traditional way is unreasonably ambitious for these two characteristics: search by trial-and-error and delayed reward. We investigate general ways of breaking the task of designing a controller down to more feasible sub-tasks which are solved independently. We propose to consider both taking advantage of past experience by reusing parts of other systems, and facilitating the learning phase by employing a bias in initial configuration.

Reviews

  • No reviews
0 customers have rated this item.
5
0%
4
0%
3
0%
2
0%
1
0%
(will not be displayed)