Transfer of Samples in Batch Reinforcement Learning
published: Aug. 6, 2008, recorded: July 2008, views: 3546
Slides
Related content
Report a problem or upload files
If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Description
The main objective of transfer learning is to reduce the complexity of learning the solution of a target task by effectively reusing the knowledge retained from solving one or more source tasks. In this paper, we introduce a novel algorithm that transfers samples (i.e., experience tuples ) from source to target tasks. Under the assumption that tasks defined on the same environment often have similar transition models and reward functions, we propose a method to select samples from the source tasks that are mostly similar to the target task, and, then, to use them as input for batch reinforcement learning algorithms. As a result, the number of samples that the agent needs to collect from the target task to learn its solution is reduced. We empirically show that, following the proposed approach, the transfer of samples is effective in reducing the learning complexity, even when the source tasks are significantly different from the target task.
Link this page
Would you like to put a link to this lecture on your homepage?Go ahead! Copy the HTML snippet !
Write your own review or comment: