Chapter 12: Combining Inductive and Analytical Learning
Using Prior Knowledge to Alter the Search Objective
- Make the network fit a combined function of the training
data and domain theory
TangentProp
- A training example must now be accompanied by its classification
and one or more derivatives
(a derivative can be taken with respect to a transformation)
- Take a look at Figure 12.5
- The error term is now modified to be E =
Σ[(target classification - actual classification)2 + μ
Σ(target derivative - actual derivative)2]
- μ is a constant that shows the relative importance of
fitting the classifications vs. fitting the derivatives
- See equation 12.1
- Table 12.4 compares the performance of TangentProp and Backpropagation when
TangentProp is supplied with the fact that the classification
of a digit is invariant with respect to both vertical and horizontal
translations
- A drawback of TangentProp is that it is not robust to errors
in the prior knowledge
EBNN
- Explanation-Based Neural Network
- Computes the training derivative itself by explaining
the example in terms of the domain theory and then
extracting the derivative
- Can vary the value of μ for each example
- Given, training data of the form < xi, f(xi) >
- Given, domain theory represented by a set of previously trained
neural networks
- Produce a new neural network to approximate the target function f
- Take a look at Figure 12.7
- The previously learned neural networks are used to calculate the
derivatives
- Derivatives that have a large magnitude indicate a highly
relevant feature
- If the domain theory explains the example well, μ should have
a higher value
- Thus, EBNN accommodates imperfect domain theory
Using Prior Knowledge to Augment Search Operators
FOCL
- An extension of FOIL (chapter 10)
- Learns a set of first-order Horn clauses
- The domain theory can be used to guide specializations
- Take a look at Figure 12.8
- All rules must consist of operational features