[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.5555/1402821.1402864acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Transfer of task representation in reinforcement learning using policy-based proto-value functions

Published: 12 May 2008 Publication History

Abstract

Reinforcement Learning research is traditionally devoted to solve single-task problems. Therefore, anytime a new task is faced, learning must be restarted from scratch. Recently, several studies have addressed the issue of reusing the knowledge acquired in solving previous related tasks by transferring information about policies and value functions. In this paper, we analyze the use of proto-value functions under the transfer learning perspective. Proto-value functions are effective basis functions for the approximation of value functions defined over the graph obtained by a random walk on the environment. The definition of this graph is a key aspect in transfer transfer problems in which both the reward function and the dynamics change. Therefore, we introduce policy-based proto-value functions, which can be obtained by considering the graph generated by a random walk guided by the optimal policy of one of the tasks at hand. We compare the effectiveness of policy-based and standard proto-value functions, on different transfer problems defined on a simple grid-world environment.

References

[1]
A. T. Bharucha-Reid. Elements of the Theory of Markov Processes and Their Applications. Dover Publications, 1997.
[2]
F. R. Chung. Spectral Graph Theory. Amer Mathematical Society, 1997.
[3]
K. Ferguson and S. Mahadevan. Proto-transfer learning in markov decision processes using spectral methods. In ICML Workshop on Transfer Learning, 2006.
[4]
G. Konidaris and A. G. Barto. Building portable options: Skill transfer in reinforcement learning. In IJCAI, pages 895--900, 2007.
[5]
M. G. Lagoudakis and R. Parr. Least-squares policy iteration. JMLR, 4:1107--1149, 2003.
[6]
S. Mahadevan and M. Maggioni. Proto-value functions: A laplacian framework for learning representation and control in markov decision processes. JMLR, 8:2169--2231, 2007.
[7]
S. Osentoski and S. Mahadevan. Learning state-action basis functions for hierarchical mdps. In ICML '07, pages 705--712, 2007.
[8]
T. J. Perkins and D. Precup. Using options for knowledge transfer in reinforcement learning. Technical report, University of Massachusetts, Amherst, MA, USA, 1999.
[9]
R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. MIT Press, Cambridge, MA, 1998.
[10]
M. E. Taylor, P. Stone, and Y. Liu. Value functions for RL-based behavior transfer: A comparative study. In AAAI, pages 880--885, July 2005.
[11]
M. E. Taylor, P. Stone, and Y. Liu. Transfer learning via inter-task mappings for temporal difference learning. JMLR, 8:2125--2167, 2007.

Cited By

View all
  • (2018)Towards sample efficient reinforcement learningProceedings of the 27th International Joint Conference on Artificial Intelligence10.5555/3304652.3304836(5739-5743)Online publication date: 13-Jul-2018
  • (2016)Theoretically-grounded policy advice from multiple teachers in reinforcement learning settings with applications to negative transferProceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence10.5555/3060832.3060945(2315-2321)Online publication date: 9-Jul-2016
  • (2016)Hidden parameter markov decision processesProceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence10.5555/3060621.3060820(1432-1440)Online publication date: 9-Jul-2016
  • Show More Cited By

Index Terms

  1. Transfer of task representation in reinforcement learning using policy-based proto-value functions

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    AAMAS '08: Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems - Volume 3
    May 2008
    503 pages
    ISBN:9780981738123

    Sponsors

    Publisher

    International Foundation for Autonomous Agents and Multiagent Systems

    Richland, SC

    Publication History

    Published: 12 May 2008

    Check for updates

    Author Tags

    1. proto-value functions
    2. reinforcement learning
    3. spectral graph theory
    4. transfer learning

    Qualifiers

    • Research-article

    Conference

    AAMAS08
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)2
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 01 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2018)Towards sample efficient reinforcement learningProceedings of the 27th International Joint Conference on Artificial Intelligence10.5555/3304652.3304836(5739-5743)Online publication date: 13-Jul-2018
    • (2016)Theoretically-grounded policy advice from multiple teachers in reinforcement learning settings with applications to negative transferProceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence10.5555/3060832.3060945(2315-2321)Online publication date: 9-Jul-2016
    • (2016)Hidden parameter markov decision processesProceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence10.5555/3060621.3060820(1432-1440)Online publication date: 9-Jul-2016
    • (2016)Transfer Learning for User Adaptation in Spoken Dialogue SystemsProceedings of the 2016 International Conference on Autonomous Agents & Multiagent Systems10.5555/2936924.2937067(975-983)Online publication date: 9-May-2016

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media