Professional Documents
Culture Documents
Slides Basics Learnercomponents Hro
Slides Basics Learnercomponents Hro
Slides Basics Learnercomponents Hro
Learning goals
Know the three components of a
6 learner: Hypothesis space, risk,
θ = ( −1.7 , 1.3 )T
4 optimization
2
0 Remp = 5.88
Understand that defining these
−2 separately is the basic design of
0 2 4 6 8 a learner
Know a variety of choices for all
three components
COMPONENTS OF SUPERVISED LEARNING
Summarizing what we have seen before, many supervised learning
algorithms can be described in terms of three components:
c Introduction to Machine Learning – 1 / 8
COMPONENTS OF SUPERVISED LEARNING
This concept can be extended by the concept of regularization, where
the model complexity is accounted for in the risk:
For now you can just think of the risk as sum of the losses.
While this is a useful framework for most supervised ML problems,
it does not cover all special cases, because some ML methods are
not defined via risk minimization and for some models, it is not
possible (or very hard) to explicitly define the hypothesis space.
c Introduction to Machine Learning – 2 / 8
VARIETY OF LEARNING COMPONENTS
The framework is a good orientation to not get lost here:
Step functions
Linear functions
Sets of rules
Hypothesis Space :
Neural networks
Voronoi tesselations
...
Mean squared error
Misclassification rate
Risk / Loss : Negative log-likelihood
Information gain
...
Analytical solution
Gradient descent
Optimization : Combinatorial optimization
Genetic algorithms
...
c Introduction to Machine Learning – 3 / 8
SUPERVISED LEARNING, FORMALIZED
A learner (or inducer) I is a program or algorithm which
receives a training set D ∈ X × Y , and,
for a given hypothesis space H of models f : X → Rg ,
uses a risk function Remp (f ) to evaluate f ∈ H on D ;
or we use Remp (θ) to evaluate f’s parametrization θ on D
uses an optimization procedure to find
I :D×Λ→H
I :D×Λ→Θ
c Introduction to Machine Learning – 4 / 8
EXAMPLE: LINEAR REGRESSION ON 1D
H = {f (x) = θ0 + θx : θ0 , θ ∈ R}
6
4
2
0
−2
0 2 4 6 8
c Introduction to Machine Learning – 5 / 8
EXAMPLE: LINEAR REGRESSION ON 1D
We might use the squared error as loss function to our risk,
punishing larger distances more severely:
n
X
Remp (θ) = (y (i ) − θ0 − θx(i ) )2
i =1
6
θ = ( 0.3 , 0 )T
4
2
0 Remp = 40.96
−2
0 2 4 6 8
c Introduction to Machine Learning – 6 / 8
EXAMPLE: LINEAR REGRESSION ON 1D
Optimization will usually mean deriving the
ordinary-least-squares (OLS) estimator θ̂ analytically.
100
6 80
θ = ( −1.7 , 1.3 ) T
Remp
4 60
2 40
0 Remp = 5.88 20
−2 2
1 0.0
In 0.5
0 2 4 6 8 te 0
rc 1.0
ep -1 pe
t Slo
-2 1.5
c Introduction to Machine Learning – 8 / 8