Integrating Planning, Execution, and Learning to Improve Plan Execution Articles uri icon

publication date

  • February 2013

start page

  • 1

end page

  • 36

issue

  • 1

volume

  • 29

International Standard Serial Number (ISSN)

  • 0824-7935

Electronic International Standard Serial Number (EISSN)

  • 1467-8640

abstract

  • Algorithms for planning under uncertainty require accurate action models that explicitly capture the uncertainty of the environment. Unfortunately, obtaining these models is usually complex. In environments with uncertainty, actions may produce countless outcomes and hence, specifying them and their probability is a hard task. As a consequence, when implementing agents with planning capabilities, practitioners frequently opt for architectures that interleave classical planning and execution monitoring following a replanning when failure paradigm. Though this approach is more practical, it may produce fragile plans that need continuous replanning episodes or even worse, that result in execution dead-ends. In this paper, we propose a new architecture to relieve these shortcomings. The architecture is based on the integration of a relational learning component and the traditional planning and execution monitoring components. The new component allows the architecture to learn probabilistic rules of the success of actions from the execution of plans and to automatically upgrade the planning model with these rules. The upgraded models can be used by any classical planner that handles metric functions or, alternatively, by any probabilistic planner. This architecture proposal is designed to integrate off-the-shelf interchangeable planning and learning components so it can profit from the last advances in both fields without modifying the architecture.

subjects

  • Computer Science

keywords

  • cognitive architectures; relational reinforcement learning; symbolic planning