Any hypothesis found to approximate the target function well over a sufficiently large set of training examples will also approximate the target function well over other unobserved examples.
Let hj and hk be boolean-valued functions defined
over X. The hj is more general than or equal to
hk (written hj ≥g hk)
if and only if
(∀ x ∈ X) [ (hk(x) = 1)
→ (hj(x) = 1)]
This is a partial order since it is reflexive, antisymmetric and transitive.
Outputs a description of the most specific hypothesis consistent with the training examples.
For this particular algorithm, there is a bias that the target concept can be represented by a conjunction of attribute constraints.
Outputs a description of the set of all hypotheses consistent with the training examples.
A hypothesis h is consistent with a set of training examples
D if and only if h(x) = c(x) for each example < x, c(x) > in D.
Consistent(h, D) ≡ (∀ < x, c(x) > ∈ D)
h(x) = c(x)
The version space denoted VSH,D with respect to
hypothesis space H and training examples D, is the subset of hypotheses
from H consistent with the training examples in D.
VSH,D ≡ { h ∈ H | Consistent(h, D) }
The general boundary G, with respect to hypothesis space H and training data D, is the set of maximally general members of H consistent with D.
The specific boundary S, with respect to hypothesis space H and training data D, is the set of maximally specific members of H consistent with D.
Let X be an arbitrary set of instances and let H be a set of
boolean-valued hypotheses defined over X. Let c:X → {0,1}
be an arbitrary target concept defined over X, and let D be an
arbitrary set of training examples {<x, c(x)>}. For all
X, H, c and D such that S and G are well defined,
VSH,D = {h ∈ H | (∃s ∈ S)
(∃g ∈ G) (g ≥g h ≥g s)}