Professional Documents
Culture Documents
Introduction
Introduction
Introduction
Introduction
Probabilities
Bayes rule
Bayes filters
Probabilistic Robotics
Key idea:
Explicit representation of uncertainty
using the calculus of probability theory
2
Axioms of Probability Theory
Pr(A) denotes probability that proposition A is true.
0 Pr( A) 1
3
A Closer Look at Axiom 3
True
A A B B
4
Using the Axioms
5
Discrete Random Variables
p(x)
E.g.
7
Joint and Conditional Probability
P(X=x and Y=y) = P(x,y)
8
Law of Total Probability, Marginals
P( x) 1
x
p( x) dx 1
P ( x ) P ( x, y ) p( x) p( x, y ) dy
y
P( x) P( x | y ) P( y ) p( x) p( x | y ) p( y ) dy
y
9
Bayes Formula
P ( x, y ) P ( x | y ) P ( y ) P ( y | x ) P ( x )
P( y | x) P( x) likelihood prior
P( x y )
P( y ) evidence
10
Normalization
P( y | x) P( x)
P( x y ) P( y | x) P( x)
P( y )
1
P( y ) 1
P( y | x)P( x)
x
Algorithm:
x : aux x| y P( y | x) P( x)
1
aux x| y
x
x : P( x | y ) aux x| y
11
Conditioning
Law of total probability:
P ( x) P ( x, z )dz
P ( x) P ( x | z ) P ( z )dz
P ( x y ) P ( x | y, z ) P ( z | y ) dz
12
Bayes Rule
with Background Knowledge
P( y | x, z ) P( x | z )
P( x | y, z )
P( y | z )
13
Conditioning
Total probability:
P( x) P( x, z )dz
P( x) P( x | z ) P( z )dz
P( x y ) P( x | y, z ) P( z ) dz
14
Conditional Independence
P( x, y z ) P( x | z ) P( y | z )
equivalent to
P( x z ) P( x | z , y )
and
P( y z ) P( y | z , x )
15
Simple Example of State Estimation
16
Causal vs. Diagnostic Reasoning
P(open|z) is diagnostic.
P(z|open) is causal.
Often causal knowledge is easier to
obtain. count frequencies!
Bayes rule allows us to use causal
knowledge:
P( z | open) P(open)
P(open | z )
P( z )
17
Example
P(z|open) = 0.6 P(z|open) = 0.3
P(open) = P(open) = 0.5
P( z | open) P(open)
P(open | z )
P( z | open) p(open) P( z | open) p(open)
0.6 0.5 2
P(open | z ) 0.67
0.6 0.5 0.3 0.5 3
18
Combining Evidence
Suppose our robot obtains another
observation z2.
19
Recursive Bayesian Updating
P( zn | x, z1,, zn 1) P( x | z1,, zn 1)
P( x | z1,, zn)
P( zn | z1,, zn 1)
P( zn | x) P( x | z1, , zn 1)
P( x | z1, , zn)
P( zn | z1,, zn 1)
P( zn | x) P( x | z1,, zn 1)
1... n P( z | x) P( x)
i 1... n
i
20
Example: Second Measurement
0.8
0.7
0.6
p( x | d)
0.5
0.4
0.3
0.2
0.1
0
5 10 15 20 25 30 35 40 45 50
Number of integrations
22
Actions
23
Typical Actions
The robot turns its wheels to move
The robot uses its manipulator to grasp
an object
Plants grow over time
24
Modeling Actions
P(x|u,x)
25
Example: Closing the door
26
State Transitions
0.9
0.1 open closed 1
0
Discrete case:
P( x | u) P( x | u, x' ) P( x' )
28
Example: The Resulting Belief
P(closed | u ) P(closed | u , x' ) P( x' )
P(closed | u , open) P(open)
P(closed | u , closed ) P(closed )
9 5 1 3 15
10 8 1 8 16
P(open | u ) P(open | u , x' ) P( x' )
P(open | u , open) P(open)
P(open | u , closed ) P(closed )
1 5 0 3 1
10 8 1 8 16
1 P (closed | u )
29
Bayes Filters: Framework
Given:
Stream of observations z and action data u:
dt {u1 , z1 , ut , zt }
Sensor model P(z|x).
Action model P(x|u,x).
Prior probability of the system state P(x).
Wanted:
Estimate of the state X of a dynamical system.
The posterior of the state is also called Belief:
Bel ( xt ) P( xt | u1 , z1 , ut , zt )
30
Markov Assumption
Bel ( xt ) P( xt | u1, z1 , ut , zt )
Bayes P( zt | xt , u1, z1, , ut ) P( xt | u1, z1, , ut )
Markov P( zt | xt ) P( xt | u1, z1, , ut )
Total prob. P( zt | xt ) P( xt | u1 , z1 , , ut , xt 1 )
P( xt 1 | u1 , z1 , , ut ) dxt 1
Markov P( zt | xt ) P( xt | ut , xt 1 ) P( xt 1 | u1 , z1 , , ut ) dxt 1
Markov P( zt | xt ) P( xt | ut , xt 1 ) P( xt 1 | u1 , z1 , , zt 1 ) dxt 1
P( zt | xt ) P( xt | ut , xt 1 ) Bel ( xt 1 ) dxt 1
32
Bayes
Bel ( xt ) Filter
P( zt | xt ) Algorithm
P( xt | ut , xt 1 ) Bel ( xt 1 ) dxt 1
33
Bayes Filters are Familiar!
Kalman filters
Particle filters
Hidden Markov models
Dynamic Bayesian networks
Partially Observable Markov Decision
Processes (POMDPs)
34
Summary
Bayes rule allows us to compute
probabilities that are hard to assess
otherwise.
Under the Markov assumption,
recursive Bayesian updating can be
used to efficiently combine evidence.
Bayes filters are a probabilistic tool
for estimating the state of dynamic
systems.
35