Professional Documents
Culture Documents
Definition of Concept Learning
Definition of Concept Learning
Definition of Concept Learning
1
The inductive learning assumption Concept learning as search
Concept learning is a task of searching an hypotheses space
We can at best guarantee that the output hypothesis The representation chosen for hypotheses determines the
search space
fits the target concept over the training data
In the example we have:
Assumption: an hypothesis that approximates well the
3 x 25 = 96 possible instances (6 attributes)
training data will also approximate the target
function over unobserved examples 5.4.4.4.4.4=5120 syntactically distinct hypothesis
i.e. given a significant training set, the output 1 + 4 .3.3.3.3.3= 973 semantically distinct hypothesis
hypothesis is able to make predictions considering that all the hypothesis with some are
semantically equivalent, i.e. inconsistent
Structuring the search space may help in searching more
efficiently
2
Properties of Find-S Candidate elimination algorithm: the idea
Find-S is guaranteed to output the most specific The idea: output a description of the set
hypothesis within H that is consistent with the positive of all hypotheses consistent with the
training examples training examples (correctly classify
The final hypothesis will also be consistent with the training examples).
negative examples
Problems: Version space: a representation of the set
There can be more than one “most specific hypotheses” of hypotheses which are consistent with D
We cannot say if the learner converged to the correct 1. an explicit list of hypotheses (List-Than-
target
Why choose the most specific?
Eliminate)
If the training examples are inconsistent, the algorithm can 2. a compact representation of hypotheses
be misleading: no tolerance to rumor. which exploits the more_general_than
Negative examples are not considered partial ordering (Candidate-Elimination)
VSH,D {h H | Consistent(h, D)} Remove from VersionSpace any hypothesis h for
which h(x) c(x)
An hypothesis h is consistent with a set of training examples D iff 3.Output the list of hypotheses in VersionSpace
Problems
h(x) = c(x) for each example in D The hypothesis space must be finite
Enumeration of all the hypothesis, rather inefficient
Consistent(h, D) ( x, c(x) D) h(x) = c(x))
3
Candidate elimination algorithm Candidate elimination algorithm
S minimally general hypotheses in H, Generalize(S, d): d is positive
G maximally general hypotheses in H For each hypothesis s in S not consistent with d:
Initially any hypothesis is still possible
1. Remove s from S
S0 = , , , , , G0 = ?, ?, ?, ?, ?, ? 2. Add to S all minimal generalizations of s consistent with d and
For each training example d, do: having a generalization in G
If d is a positive example: 3. Remove from S any hypothesis with a more specific h in S
1. Remove from G any h inconsistent with d Specialize(G, d): d is negative
2. Generalize(S, d) For each hypothesis g in G not consistent with d: i.e. g satisfies d,
If d is a negative example: 1. Remove g from G but d is negative
1. Remove from S any h inconsistent with d 2. Add to G all minimal specializations of g consistent with d and
2. Specialize(G, d) having a specialization in S
3. Remove from G any hypothesis having a more general hypothesis
Note: when d =x, No is a negative example, an hypothesis h is in G
inconsistent with d iff h satisfies x
Example:
Example: initially after seeing Sunny,Warm, Normal, Strong, Warm, Same +
G 1, G 2 ?, ?, ?, ?, ?, ?
4
Example:
after seeing Sunny, Warm, High, Strong, Cool Change + Learned Version Space
S3 Sunny, Warm, ?, Strong, Warm, Same
G 4: Sunny, ?, ?, ?, ?, ? ?, Warm, ?, ?, ?, ?
Classified as positive by all hypothesis, since Classified as negative by all hypothesis, since
satisfies any hypothesis in S does not satisfy any hypothesis in G
5
Classifying new examples Classifying new examples
What if H does not contain the target concept? 1 Sunny Warm Normal Strong Cool Change YES
Can we improve the situation by extending the 2 Cloudy Warm Normal Strong Cool Change YES
hypothesis space? 3 Rainy Warm Normal Strong Cool Change NO
Will this influence the ability to generalize?
No hypothesis is consistent with the above three examples with the
These are general questions for inductive inference, assumption that the target is a conjunction of constraints
addressed in the context of Candidate-Elimination ?, Warm, Normal, Strong, Cool, Change is too general
Suppose we include in H every possible hypothesis Target concept exists in a different space H', including disjunction and
… including the ability to represent disjunctive in particular the hypothesis
concepts Sky=Sunny or Sky=Cloudy
The problem is that we have biased the learner to consider only
conjunctive hypothesis space
6
Inductive bias: definition
Given:
No inductive inference without a bias
a concept learning algorithm L for a set of
instances X
A learner that makes no a priori assumptions regarding a concept c defined over X
the identity of the target concept, has no rational basis a set of training examples for c: Dc = {x, c(x)}
for classifying unseen instances L(xi, Dc) outcome of classification of xi after
The inductive bias of a learner are the assumptions learning
that justify its inductive conclusions or the policy Inductive inference ( ≻ ):
adopted for generalization
Dc xi ≻ L(xi, Dc)
Different learners can be characterized by their bias
The inductive bias is defined as a minimal set of
See next for a more formal definition of inductive bias
assumptions B, such that (|− for deduction)
…
(xi X) [ (B Dc xi) |− L(xi, Dc) ]