A Framework for Decision-Theoretic Planning 1: Combining the Situation Calculus, Conditional Plans, Probability and Utility*

A Framework for Decision-Theoretic Planning 1: Combining the Situation Calculus, Conditional Plans, Probability and Utility*

436 Poole A Framework for Decision-Theoretic Planning 1: Combining the Situation Calculus, Conditional Plans, Probability and Utility* David Poole Department of Computer Science University of British Columbia Vancouver, B.C., Canada V6T 1Z4 [email protected] http://www.cs.ubc.ca/spider/poole Abstract of steps, perhaps conditional on observations that solves a goaL In decision-theoretic planning, this is generalised to the case where there is uncertainty about the environment This paper shows how we can combine logical and we are concerned, not only with solving a 'goal', but representations of actions and decision theory in what happens under any of the contingencies. Goal solving such a manner that seems natural for both. In par­ is extended to the problem of maximizing the agent's ex­ ticular we assume an axiomatization of the do­ pected utility, where here the utility is an arbitrary function main in terms of situation calculus, using what of the final state. is essentially Reiter's solution to the frameprob­ lem, in terms of the completion of the axioms Recently there have been claims made that Markov deci­ defining the state change. Uncertainty is handled sion processes (MOPs) [Puterman, 1990] are the appropri­ in terms of the independent choice logic, which ate framework for developing decision theoretic planners allows for independent choices and a logic pro­ (e.g., [Boutilier andPuterman, l995;Deanetal., 1993]). In gram that gives the consequences of the choices. MDPs the idea is to construct policies- functions from ob­ As part of the consequences are a specification served state into actions that maximise long run cumulative of the utility of (final) states. The robot adopts (usually discounted) reward. It is wrong to equate decision robot plans, similar to the GOLOG programming theory with creating policies; decision theory can be used to language. Within this logic, we can define the select plans, and policies can be considered independently expected utility of a conditional plan, based on of decision theory [Schoppers, 1987]. Even when solving the axiomatization of the actions, the uncertainty partially observable MOPs (POMDPs), where a policy is a and the utility. The 'planning' problem is to find function from belief states into actions, it is often more con­ the plan with the highest expected utility. This venient to use a policy tree [Kaelbling et al., 1996], which is is related to recent structured representations for much more like a robot plan as developed here- see Sec­ POMDPs; here we use stochastic situation cal­ tion 7. culus rules to specify the state transition func­ Rather than assuming robots have policies [Poole, 1995c], tion and the reward/value function. Finally we we can instead consider robot plans as in GOLOG show that with stochastic frame axioms, actions [Levesque et al., 1996]. These plans consider sequences representations in probabilistic STRIPSare expo­ of steps, with conditions and loops, rather than reactive nentially larger than using the representation pro­ strategies. In this paper we restrict ourselves to conditional posed here. plans; we do not consider loops or nondeterministic choice, although these also could be considered (see Section 6). Unlike GOLOG, and like Levesque [1996], the conditions 1 Introduction in the branching can be 'observations' about the world or values received by sensors The combination of decision theory and planning is very ap­ As in GOLOG, we assume that the effects of actions are rep­ pealing. Since the combination was advocated in [Feldman resented in the situation calculus. In particularwe adopt Re­ and Sproull, 1975], there has been a recent revival of inter­ iter's [1991] solution to the frame problem. Our represen­ est. The general idea of planning is to construct a sequence tation is simpler in that we do not assume that actions have preconditions - all actions can be attempted at any time, This work was supported by Institute for Robotics and Intel­ the effects of these actions can depend on what else is true ligent Systems, Project IC-7 and Natural Sciences and Engineer­ ing Research Council of Canada Operating Grant OGP0044121. in the world. This is important because the agent may not Thanks to Craig Boutilier, Ronen Brafman and Chris Geib for giv­ know whether the preconditions of an action hold, but, for ing me a hard time about this paper. A Framework for Decision-Theoretic Planning 437 example, may be sure enough to want to try the action. r-- - ­ All of the uncertainty in our rules is relegated to in­ l -- ----------- dependent choices as in the independent choice logic :··-----.. -�---··· ·- . [Poole, 1995b] (an extension of probabilistic Horn abduc­ tion [Poole, 1993]). This allows for a clean separation of the completeness assumed by Reiter's solution to the frame problem and the uncertainty we need for decision theory. Before we describe the theory there are some design choices incorporated into the framework: .. -·- ····�r r123-· · • In the deterministic case, the trajectory of actions by the (single) agent up to some time point determines what is true at that point. Thus, the trajectory of ac­ Figure I: The example robot environment tions, as encapsulated by the 'situation' term of the sit­ uation calculus [McCarthy and Hayes, 1969; Reiter, 1991] can be used to denote the state, as is done in the traditional situation calculus. However, when dealing • When mixing logic and probability, one can extend a with uncertainty, the trajectory of an agent's actions up rich logic with probability, and have two sorts of un­ certainty- that uncertainty from the probabilities and to a point, does not uniquely determine what is true from disjunction in the logic [Bacchus, 1990]. An at that point. What random occurrences or exogenous that alternative that is pursued in the independent choice events occurred also determines what is true. We have logic is to have all of the uncertainty in terms of prob­ a choice: we can keep the semantic conception of a sit­ abilities. The logic is restricted so that there is no un­ uation (as a state) and make the syntactic characteriza­ certainty in the logic - every set of sentences has a tion more complicated by perhaps interleaving exoge­ unique model. particular we choose the logic of nous actions, or we can keep the simple syntactic form In of the situation calculus, and use a different notion that acyclic logic programs under the stable model seman­ prescribes truth values. We have chosen the latter, and tics; this seems to be the strongest practical language with the unique model property. All uncertainty is distinguish the 'situation' denoted by the trajectory of handled by what can be seen as independent stochas­ actions, from the 'state' that specifies what is true in the situation. In general there will be a probability dis­ ti� mechanisms, and a deterministic logic program that tribution over states resulting from a set of actions by gives the consequences of the agent's actions and the random outcomes. In this manner we can get a simple the agent. It is this distribution over states, and their corresponding utility, that we seek to model. mix oflogic and Bayesian decision theory (see [Poole, 1995b]). This division means that agent's actions are treated very differently from exogenous actions that can also • Unlike in Markov decision processes, where there is a change what is true. The situation terms define only reward for each state and utilities accumulate, we as­ the agent's actions in reaching that point in time. The sume that an agent carries out a plan, and receives util­ situation calculus tenns indicate only the trajectory, in ity depending on the state it ends up in. This is done to terms of steps, of the agent and essentially just serve to simplify the formalism, and seems natural. This does delimit time points at which we want to be able to say not mean that we cannot model cases where an agent what holds. receives rewards and costs along the way, but the re­ wards accumulated then have to be part of the state. • When building conditional plans, we have to consider what we can condition these plans on. We assume that Note that MOPs also need to make the cumulative re­ the agent has passive sensors, and that it can condition ward part of the state to model non-additive rewards its actions on the output of these sensors. We only have such as an agent receiving or paying interest on its cur­ one sort of action, and these actions only affect 'the rent utility. This also means that we cannot optimize world' (which includes both the robot and the environ­ ongoing processes that never halt- in fact the acyclic ment). All we need to do is to specify how the agent's restriction in the language means that we cannot model sensors depend on the world. This does not mean that such ongoing processes without inventing an arbitrary we cannot model information-producing actions (e.g., stopping criteria (e.g., stop after 3 years). looking in a particular place)- these information pro­ ducing actions produce effects that make the sensor We use the following ongoing example to show the power values correlate with what is true in the world. The of the fonnalism; it is not intended to be realistic. sensors can be noisy- the value they return does not necessarily correspond with what is true in the world Example 1.1 Suppose we have a robot that can travel (of course if there was no correlation with what is true around an office building, pick up keys, unlock doors, and in the world, they would not be very useful sensors). sense whether the key is at the location it is currently at.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    10 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us