Chapter 2: Concept Learning and the General-to-Specific Ordering

Definition

Notation

Inductive Learning Hypothesis

Any hypothesis found to approximate the target function well over a sufficiently large set of training examples will also approximate the target function well over other unobserved examples.

Definition

Let hj and hk be boolean-valued functions defined over X. hj is more general than or equal to hk (written hjg hk) if and only if
(∀ x ∈ X) [ (hk(x) = 1) → (hj(x) = 1)]

This is a partial order since it is reflexive, antisymmetric and transitive.

Find-S Algorithm

Outputs a description of the most specific hypothesis consistent with the training examples.

  1. Initialize h to the most specific hypothesis in H
  2. For each positive training instance x
  3. Output hypothesis h

For this particular algorithm, there is a bias that the target concept can be represented by a conjunction of attribute constraints.

Candidate Elimination Algorithm

Outputs a description of the set of all hypotheses consistent with the training examples.

Definition

A hypothesis h is consistent with a set of training examples D if and only if h(x) = c(x) for each example < x, c(x) > in D.
Consistent(h, D) ≡ (∀ < x, c(x) > ∈ D) h(x) = c(x)

Definition

The version space denoted VSH,D with respect to hypothesis space H and training examples D, is the subset of hypotheses from H consistent with the training examples in D.
VSH,D ≡ { h ∈ H | Consistent(h, D) }

Definition

The general boundary G, with respect to hypothesis space H and training data D, is the set of maximally general members of H consistent with D.

Definition

The specific boundary S, with respect to hypothesis space H and training data D, is the set of maximally specific members of H consistent with D.

Version Space Representation

Let X be an arbitrary set of instances and let H be a set of boolean-valued hypotheses defined over X. Let c:X → {0,1} be an arbitrary target concept defined over X, and let D be an arbitrary set of training examples {<x, c(x)>}. For all X, H, c and D such that S and G are well defined,
VSH,D = {h ∈ H | (∃s ∈ S) (∃g ∈ G) (g ≥g h ≥g s)}

Algorithm

  1. Initialize G to the set of maximally general hypotheses in H
  2. Initialize S to the set of maximally specific hypotheses in H
  3. For each positive training example d,
  4. For each negative training example d,

Candidate Elimination Algorithm Issues

Inductive Bias

Sample Exercise

Work exercise 2.4 on page 48.

Valid XHTML 1.0!