1/8/20
CSE 473: Artificial Intelligence
Problem Spaces & Search
Dan Weld
With slides from Dan Klein, Stuart Russell, Andrew Moore, Luke Zettlemoyer, Dana Nau… 1
Outline
§ Search Problems
§ Uninformed Search Methods § Depth-First Search § Breadth-First Search § Iterative Deepening Search § Uniform-Cost Search
§ Heuristic Search Methods § Heuristic Generation
2
1 1/8/20
Agent vs. Environment
§ An agent is an entity that Agent perceives and acts. Sensors Percepts § A rational agent selects Environment actions that maximize its ? utility function.
§ Characteristics of the Actuators percepts, environment, and Actions action space dictate techniques for selecting rational actions.
3
Goal Based Agents
§ Plan ahead § Ask “what if”
§ Decisions based on (hypothesized) consequences of actions § Requires a model of how the world evolves in response to actions
§ Act on how the world WOULD BE
4
2 1/8/20
Search thru a Problem Space (aka State Space) • Input: Functions: States à States § Set of states Aka “Successor Function” § Operators [and costs] § Start state § Goal state [or test]
• Output: • Path: start Þ a state satisfying goal test [May require shortest path] [Sometimes just need a state that passes test]
6
6
Example: Simplified Pac-Man
§ Input: § A state space
§ Operators § Go North “Go North” § Go East § Etc
“Go East” § A start state
§ A goal test
§ Output:
7
3 1/8/20
State Space Defines a Graph
State space graph: § Each node is a state a G § Each edge is an operators b c § Edges may be labeled with cost of operator e d f S h p r q In contrast to other areas of CS… We can rarely build this graph in Ridiculously tiny search graph memory (so we don’t try) for a tiny search problem … just a conceptual tool
8
State Space Sizes?
§ Search Problem: Eat all of the food § Pacman positions: 1010 x 12x 12 = =120 120 § Pacman facing: up, down, left, right § Food configurations: 223030 § Ghost1 positions: 12 § Ghost 2 positions: 1111
120 x 4 x 230 x 12 x 11 = 6.8 x 1013
9
4 1/8/20
Ex: Route Planning: Arad à Bucharest § Input: § Set of states Different operators may be applicable in § Operators [and costs] different states
§ Start state
§ Goal state (test)
§ Output:
10
Ex: Finding optimal repair sequences
11
11
5 1/8/20
Ex: Hardware Verification
Prove system will never make floating point error 12
12
Search Methods § Blind Search • Depth first search • Breadth first search • Iterative deepening search • Uniform cost search § Local Search § Informed Search § Constraint Satisfaction
§ Adversary Search 19
19
6 1/8/20
Search Trees
“Go North” “Go East”
§ A search tree: § Tree’s root node corresponds to the start state § Children correspond to successors (application of operator) § Edges are labeled with actions/operators and costs § Nodes in the tree contain states, correspond to PATH to those states § For most problems, we can never actually build the whole tree
20
Example: Tree Search
State graph:
a G b c
e d f S h p r q
What is the search tree?
21
7 1/8/20
State Graphs vs. Search Trees
a G Each NODE in in the b c search tree denotes e an entire PATH d f through the state S h space graph. p r q S
e p We construct both d on demand – and b c e h r q we construct as little as possible. a a h r p q f
p q f q c G a q c G Why duplicate nodes? a
22
States vs. Nodes
§ Vertices in state space graphs are problem states § Represent an abstracted state of the world § Have successors, can be goal / non-goal, have multiple predecessors § Vertices in search trees (“Nodes”) are plans § Contain a problem state and one parent, a path length, a depth & a cost § Represent a plan (sequence of actions) which results in the node’s state § The same problem state may be achieved by multiple search tree nodes Search Tree Nodes Problem States Parent Depth 5 Action
Node Depth 6
23
8 1/8/20
Building Search Trees
§ Search: § Expand out possible nodes (plans) in the tree § Maintain a fringe of unexpanded nodes § Try to expand as few nodes as possible
24
General Tree Search
Detailed pseudocode is Important ideas: in the book! § Fringe (leaves of partially-expanded tree) § Expansion (adding successors of a leaf node) § Exploration strategy which fringe node to expand next?
25
9 1/8/20
Review: Breadth First Search
a G b c
e Strategy: expand d f shallowest node first S h Implementation: p r q Fringe is a queue - FIFO
26
Review: Breadth First Search
a G Expansion order: b c e (S,d,e,p,b,c,e,h,r,q,a,a d f ,h,r,p,q,f,p,q,f,q,c,G) S h p q r
S
d e p Search b c e h r q Tiers a a h r p q f
p q f q c G
q c G a a 27
10 1/8/20
Review: Depth First Search
a G b c
e Strategy: expand d f deepest node first S h p r Implementation: q Fringe is a stack - LIFO
Wolog: assume we choose children in lexicographic order
28
Review: Depth First Search
a G Expansion ordering: b c e (d,b,a,c,a,e,h,p,q,q,r,f,c,a,G) d f S h p r q
S
d e p
b c e h r q
a a h r p q f
p q f q c G
q c G a a
29
11 1/8/20
Search Algorithm Properties
§ Complete? Guaranteed to find a solution if one exists? § Optimal? Guaranteed to find the least cost path? § Time complexity? § Space complexity? Variables:
n Number of states in the problem b The maximum branching factor B (the maximum number of successors for a state) C* Cost of least cost solution d Depth of the shallowest solution m Max depth of the search tree
30
Depth-First Search
Algorithm Complete Optimal Time Space DFS Depth First N N AX Search No ) O(LMAX)
START
§ Infinite paths make DFS incomplete… a GOAL § How can we fix this? § Check new nodes against path from S b
31
12 1/8/20
Depth-First Search w/cycle check
Algorithm Complete Optimal Time Space DFS Depth First N N AX Search No ) O(LMAX) Yes if finite tree
START § Infinite trees still make DFS incomplete…
a GOAL § But what if space is finite?
b
32
Depth-First Search w/cycle check Assume finite state space
Algorithm Complete Optimal Time Space DFS Depth First N N AX m Search Yes if finite No O()b ) O(LMAX)O(b m) tree
START
a GOAL b = max branching factor d = depth of solution m = max depth of tree b
33
13 1/8/20
BFS Tree Search
Algorithm Complete Optimal Time Space w/ Path if finite if finite DFS N unless N O(bm) O(bm) Checking finite else ∞ else ∞ BFS Y* Y* O(bd) always O(bd) always
1 node b … b nodes d tiers b2 nodes
bd nodes
bm nodes
* Assuming finite branching factor 35
§ When is BFS optimal? Memory a Limitation? § Suppose: • 4 GHz CPU • 32 GB main memory • 100 instructions / expansion • 5 bytes / node
• 40 M expansions / sec • Memory filled in … 3 min
36
36
14 1/8/20
Iterative Deepening Search
§ DFS with limit; incrementally grow limit § Evaluation
a
37
37
Iterative Deepening Search
§ DFS Tree Search with limit; incrementally grow limit § Evaluation
a b
c d
38
38
15 1/8/20
Iterative Deepening Search
§ DFS Tree Search with limit; incrementally grow limit § Evaluation § Complete? a b e § Time Complexity? c f d i § Space Complexity? g h j k l
39
39
Iterative Deepening Search
§ DFS with limit; incrementally grow limit § Evaluation § Complete? a b e Yes * § Time Complexity? d c d O(bd) § Space Complexity?
O(bd)
* Assuming branching factor is finite 40 Important Note: no cycle checking necessary! 40
16 1/8/20
Cost of Iterative Deepening
b ratio ID to DFS 2 3 3 2 5 1.5 10 1.2 25 1.08 100 1.02
41
41
Puzzles
15 puzzle 3x3 Rubics cube
42
42
17 1/8/20
Speed Assuming 10M nodes/sec & sufficient memory BFS Iter. Deep. Nodes Time Nodes Time
8 Puzzle 105 .01 sec 105 .01 sec
2x2x2 Rubik’s 106 .2 sec 106 .2 sec
15 Puzzle 1013 6 days 1Mx 1017 20k yrs
3x3x3 Rubik’s 1019 68k yrs 8x 1020 574k yrs 24 Puzzle 1025 12B yrs 1037 1023 yrs
Why the difference? Rubik has higher branch factor # of duplicates 15 puzzle has greater depth Slide adapted from Richard Korf presentation 43
43
Tree vs Graph Search
In BFS, for example, we shouldn’t bother expanding the circled nodes (why?)
S
d e p
b c e h r q
a a h r p q f
p q f q c G
q c G a a
44
18 1/8/20
Graph Search
§ Very simple fix: never expand a state type twice
45
Some Hints
§ On small problems § Graph search almost always better than tree search
§ Implement your closed list as a dict or set!
§ On many real problems § Storage space is a huge concern § Graph search impractical
46
19 1/8/20
Search Methods § Depth first search (DFS) Blind search § Breadth first search (BFS) § Iterative deepening depth-first search (IDS)
47
47
Search Methods § Depth first search (DFS) § Breadth first search (BFS) § Iterative deepening depth-first search (IDS) § Best first search Heuristic search § Uniform cost search (UCS) § Greedy search § A* § Iterative Deepening A* (IDA*) § Beam search
§ Hill climbing 48
48
20 1/8/20
Blind vs Heuristic Search
§ Costs on Actions
§ Heuristic Guidance
Separable Issues, but usually linked.
49
49
Costs on Actions
GOAL 2 a 2 b c 3 2 1 8 2 e 3 d f 9 8 2 START h 1 4 1 p 4 r 15 q Objective: Path with smallest overall cost
50
21 1/8/20
Costs on Actions
GOAL 2 a 2 b c 3 2 1 8 2 e 3 d f 9 8 2 START h 1 4 1 p 4 r 15 q What will BFS return? … finds the shortest path in terms of number of transitions. It does not find the least-cost path. 51
Best-First Search § Generalization of breadth-first search § Fringe = Priority queue of nodes to be explored § Cost function f(n) applied to each node
52
52
22 1/8/20
Priority Queue Refresher
A priority queue is a data structure in which you can insert and retrieve (key, value) pairs with the following operations:
pq.push(key, value) inserts (key, value) into the queue. pq.pop() returns the key with the lowest value, and removes it from the queue.
§ You can decrease a key’s priority by pushing it again § Unlike a regular queue, insertions aren’t constant time, usually O(log n) § We’ll need priority queues for cost-sensitive search methods
53
Best-First Search § Generalization of breadth-first search § Fringe = Priority queue of nodes to be explored § Cost function f(n) applied to each node
Add initial state to priority queue While queue not empty Node = head(queue) If goal?(node) then return node Add new children of node to queue
“expanding the node” 54
54
23 1/8/20
Old Friend
Breadth First = Best First with f(n) = depth(n)
55
55
Uniform Cost Search
Best first, where f(n) = “cost from start to n”
aka “Dijkstra’s Algorithm”
56
24 1/8/20
Uniform Cost Search
2 a G b 1 8 c Expansion order: 2 2 3 e d 9 1 f 8 S, p, d, b, e, a, r, f, e, G S h 1 1 p r 15 q S 0
d 3 e 9 p 1
b 4 c e 5 h 17 r 11 q 16 11 Cost contours a 6 a h 13 r 7 p q f (not all shown) p q f 8 q c G
q 11 c G 10 a a
57
Uniform Cost Search Algorithm Complete Optimal Time Space DFS w/ Path m Checking Y if finite N O(b ) O(bm) BFS Y Y* O(bd) O(bd) UCS Y* Y O(bC*/e) O(bC*/e)
b …
C*/e tiers
C* = Optimal cost e = Minimum cost of an action * We’ll keep assuming finite branching factor
58
25 1/8/20
Uniform Cost Issues
§ Remember: explores … c £ 1 increasing cost contours c £ 2 c £ 3
§ The good: UCS is complete and optimal!
§ The bad: § Explores options in every “direction” Start Goal § No information about goal location
59
Uniform Cost: Pac-Man
§ Cost of 1 for each action § Explores all of the states, but one
60
26 1/8/20
What is a Heuristic?
§ An estimate of how close a state is to a goal § Designed for a particular search problem
10
5 11.2
Goal
§ Examples: Manhattan distance: 10+5 = 15 Euclidean distance: 11.2
61
What is a Heuristic?
§ An estimate of how close a state is to a goal § Designed for a particular search problem
§ Actual distance to goal: 2+4+2+1+8=
62
27 1/8/20
Greedy Search
Best first with f(n) = heuristic estimate of distance to goal
63
Greedy Search
Expand the node that seems closest…
A
start
B goal
What can go wrong?
64
28 1/8/20
Greedy Search
§ Common case: b § Greedy search takes you … straight to a (suboptimal) goal
§ Worst-case: like a badly- guided DFS § Can explore everything § Can get stuck in loops if no b cycle checking …
§ Like DFS wrt completeness § Complete w/ cycle checking § If finite # states
65
A* Search Hart, Nilsson & Rafael 1968 Best first search with f(n) = g(n) + h(n) § g(n) = sum of costs from start to n § h(n) = (admissible) estimate of lowest cost path from n ® goal note: h(goal) = 0 Nomenclature Confusion: Can view as cross-breed: Is h the heuristic or is f? g(n) ~ uniform cost search h(n) ~ greedy search
Best of both worlds… 66
66
29 1/8/20
A* Search Hart, Nilsson & Rafael 1968 Best first search with f(n) = g(n) + h(n) § g(n) = sum of costs from start to n § h(n) = admissible estimate of lowest cost path from n ® goal h(goal) = 0 If h(n) is admissible and monotonic then A* is optimal {
Underestimates (≤) cost f values never decrease of reaching goal from node } From node to descendants (triangle inequality)
67 67
Is Manhattan distance admissible? § Is h(x) an underestimate?
S
G
68
68
30 1/8/20
Is a heuristic based on Manhattan distance monotonic? § f values increase from node to children § (triangle inequality)
S
G
69
69
Monotonicity
F(a) ≥ F(b) G(a)+H(a) ≥ G(b)+H(b) a
b
c
70
70
31 1/8/20
Euclidean Distance § Admissible? § Monotonic?
S
G
71
71
Admissible Heuristics F(n) = G(n) + H(n)
Admissible Not Admissible Value
State (x) State (x)
True (optimal) cost remaining H(n) = H-estimate of cost remaining 72 Slide credit: Travis Mandel
72
32 1/8/20
Monotonic/Consistent Heuristics F(n) = G(n) + H(n)
Monotonic Not Monotonic (but admissible) Value
State (x) State (x)
True (optimal) cost remaining H(n) = H-estimate of cost remaining 73 Slide credit: Travis Mandel 73
73
Monotonic/Consistent Heuristics F(n) = G(n) + h(n)
Monotonic Not Monotonic (but admissible) Value
State (x) State (x)
True (optimal) cost remaining H(n) Heuristic-estimated cost remaining 74 74 Slide credit: Travis Mandel F(n) Heuristic + cost so far 74
33 1/8/20
Optimality of A* (tree search)
Monotonicity required for proof in graph search version 75
75
Optimality Continued
76
76
34 1/8/20
Demo
http://qiao.github.io/PathFinding.js/visual/
77
77
A* Example
78
78
35 1/8/20
A* Example
79
79
A* Example
80
80
36 1/8/20
A* Example
81
81
A* Example
82
82
37 1/8/20
A* Example
83
83
start European Example
1 3 2 4
5
end
84
84
38 1/8/20
A* Summary
§ Pros Produces optimal cost solution!
Does so quite quickly (focused) § Cons Maintains priority queue…
Which can get exponentially big L
85
85
Iterative-Deepening A* § Like iterative-deepening depth-first, but... § Depth bound modified to be an f-limit § Start with f-limit = h(start) § Prune any node if f(node) > f-limit § Next f-limit = min-cost of any node pruned e FL=21 FL=15 d a f
b
c 86
86
39 1/8/20
Demo
http://qiao.github.io/PathFinding.js/visual/
87
87
IDA* Analysis § Complete & Optimal (ala A*) § Space usage µ depth of solution § Each iteration is DFS - no priority queue! § # nodes expanded relative to A* § Depends on # unique values of heuristic function § In 8 puzzle: few values Þ close to # A* expands § In traveling salesman: each f value is unique Þ 1+2+…+n = O(n2) where n=nodes A* expands if n is too big for main memory, n2 is too long to wait!
88
88
40 1/8/20
Forgetfulness
§ A* used exponential memory § How much does IDA* use? § During a run? § In between runs? § SMA*
© Daniel S. Weld 89
89
Which Algorithm?
§ Uniform cost search (UCS):
93
93
41 1/8/20
Which Algorithm?
§ A*, Manhattan Heuristic:
94
Which Algorithm?
§ Best First / Greedy, Manhattan Heuristic:
95
42