Abstract

This paper reports on experiments where techniques of supervised machine learning are applied to the problem of planning. The input to the learning algorithm is composed of a description of a planning domain, planning problems in this domain, and solutions for them. The output is an efficient algorithm—a strategy—for solving problems in that domain. We test the strategy on an independent set of planning problems from the same domain, so that success is measured by its ability to solve complete problems. A system, L2A ct, has been developed in order to perform these experiments. We have experimented with the blocks world domain and the logistics transportation domain, using strategies in the form of a generalisation of decision lists. The condition of a rule in the decision list is an existentially quantified first order expression, and each such rule indicates which action to take when the condition is satisfied. The learning algorithm is a variant of Rivest's (1987) algorithm, improved with several techniques that reduce its time complexity. The experiments demonstrate that the approach is feasible, and generalisation is achieved so that unseen problems can be solved by the learned strategies. Moreover, the learned strategies are efficient, the solutions found by them are competitive with those of known heuristics for the domains, and transfer from small planning problems in the examples to larger ones in the test set is exhibited.

Highlights

  • The problem of enabling an arti cial agent to choose its actions in the world so as to achieve some goals has been extensively studied in AI, and largely in the sub eld of planning

  • We considered two domains that have been discussed in the literature; the blocks world, and the logistics domain

  • A planning problem in this domain includes an arrangement of blocks in the current situation, and a list of required goal conditions, that does not necessarily describe a complete situation[7]

Read more

Summary

Introduction

The problem of enabling an arti cial agent to choose its actions in the world so as to achieve some goals has been extensively studied in AI, and largely in the sub eld of planning. The general setup in planning assumes that agent has a model of the dynamics of each action, that is, when it is applicable and what its e ects are, as well as possibly other information about the world. Using this knowledge, in any situation, the agent can decide what to do by projecting forward various possibilities for actions, and choosing among them one that would lead to achieving its goals. Research supported by ONR grant N00014-95-1-0550, and ARO grant DAAL03-92-G-0115

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call