Feature Reinforcement Learning: Part I. Unstructured MDPs

Type: Article

Publication Date: 2009-01-01

Citations: 62

DOI: https://doi.org/10.2478/v10229-011-0002-8

Abstract

Feature Reinforcement Learning: Part I. Unstructured MDPsGeneral-purpose, intelligent, learning agents cycle through sequences of observations, actions, and rewards that are complex, uncertain, unknown, and non-Markovian. On the other hand, reinforcement learning is well-developed for small finite state Markov decision processes (MDPs). Up to now, extracting the right state representations out of bare observations, that is, reducing the general agent setup to the MDP framework, is an art that involves significant effort by designers. The primary goal of this work is to automate the reduction process and thereby significantly expand the scope of many existing reinforcement learning algorithms and the agents that employ them. Before we can think of mechanizing this search for suitable MDPs, we need a formal objective criterion. The main contribution of this article is to develop such a criterion. I also integrate the various parts into one learning algorithm. Extensions to more realistic dynamic Bayesian networks are developed in Part II (Hutter, 2009c). The role of POMDPs is also considered there.

Locations

  • Journal of Artificial General Intelligence - View - PDF
  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Feature Reinforcement Learning: Part I: Unstructured MDPs 2009 Marcus Hütter
+ Feature Markov Decision Processes 2008 Marcus Hütter
+ Feature Markov Decision Processes 2009 Marcus Hütter
+ A Unifying Framework for Reinforcement Learning and Planning 2020 Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
+ PDF Chat A Unifying Framework for Reinforcement Learning and Planning 2022 Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
+ Model-based Reinforcement Learning: A Survey 2020 Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
+ PDF Chat Model-based Reinforcement Learning: A Survey 2023 Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
+ PDF Chat Model-based Reinforcement Learning: A Survey 2023 Thomas M. Moerland
Joost Broekens
Aske Plaat
Catholijn M. Jonker
+ PDF Chat Generalizing Multi-Step Inverse Models for Representation Learning to Finite-Memory POMDPs 2024 Lili Wu
Ben Evans
Riashat Islam
Raihan Seraj
Yonathan Efroni
Alex Lamb
+ A Framework for Reinforcement Learning and Planning. 2020 Thomas M. Moerland
Joost Broekens
Catholijn M. Jonker
+ PDF Chat Markov Abstractions for PAC Reinforcement Learning in Non-Markov Decision Processes 2022 Alessandro Ronca
Gabriel Paludo Licks
Giuseppe De Giacomo
+ Markov Abstractions for PAC Reinforcement Learning in Non-Markov Decision Processes 2022 Alessandro Ronca
Gabriel Paludo Licks
Giuseppe De Giacomo
+ PDF Chat Reinforcement learning in non-Markovian environments 2024 Siddharth Chandak
Pratik Shah
Vivek S. Borkar
Parth Dodhia
+ Reinforcement Learning with Information-Theoretic Actuation 2021 Elliot Catt
Marcus Hütter
Joel Veness
+ PDF Chat Reinforcement Learning with Options 2024 Ayoub Ghriss
Masashi Sugiyama
Alessandro Lazaric
+ Provably Efficient Reinforcement Learning for Discounted MDPs with Feature Mapping 2020 Dongruo Zhou
Jiafan He
Quanquan Gu
+ Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning 2021 Kai Wang
Sanket Shah
Haipeng Chen
Andrew Perrault
Finale Doshi‐Velez
Milind Tambe
+ Extreme State Aggregation Beyond MDPs 2014 Marcus Hütter
+ Hierarchical Reinforcement Learning with the MAXQ Value Function Decomposition 1999 Thomas G. Dietterich
+ PDF Chat Robust Markov Decision Processes: A Place Where AI and Formal Methods Meet 2024 Marnix Suilen
Thom Badings
Eline M. Bovy
David Parker
Nils Jansen