Neural Network Toolbox | Search  Help Desk |
learnhd | Examples See Also |
Hebb with decay weight learning rule
[dW,LS] = learnhd(W,P,Z,N,A,T,E,gW,gA,D,LP,LS)
info = learnhd(code)
learnhd
is the Hebb weight learning function.
learnhd(W,P,Z,N,A,T,E,gW,gA,D,LP,LS)
takes several inputs,
W - S
x R
weight matrix (or S
x 1
bias vector).
P - R
x Q
input vectors (or ones(1,Q)
).
Z - S
x Q
weighted input vectors.
T - S
x Q
layer target vectors.
E - S
x Q
layer error vectors.
gW - S
x R
gradient with respect to performance.
gA - S
x Q
output gradient with respect to performance.
LP -
Learning parameters, none, LP = []
.
LS -
Learning state, initially should be = []
.
learnhd
's learning parameters shown here with default values.
learnhd(code)
returns useful information for each code
string:
'pnames
' - Names of learning parameters.
'pdefaults
' - Default learning parameters.
'needg
' - Returns 1 if this function uses gW
or gA
.
P
, output A
, and weights W
for a layer with a 2-element input and 3 neurons. We also define the decay and learning rates.
p = rand(2,1); a = rand(3,1); w = rand(3,2); lp.dr = 0.05; lp.lr = 0.5;Since
learnhd
only needs these values to calculate a weight change (see algorithm below), we will use them to do so.
dW = learnhd(w,p,[],[],a,[],[],[],[],[],lp,[])To prepare the weights and the bias of layer
i
of a custom network to learn with learnhd
:
.net.trainFcn
to 'trainwb
'. (net.trainParam
will automatically become
trainwb
's default parameters.)
.net.adaptFcn
to 'adaptwb'. (net.adaptParam
will automatically become
trainwb
's default parameters.)
.net.inputWeights{i,j}.learnFcn
to 'learnhd
'. Set each
net.layerWeights{i,j}.learnFcn
to 'learnhd
'. (Each weight learning
parameter property will automatically be set to learnhd's default
parameters.)
.net.trainParam
(net
.adaptParam
) properties to desired values.
.train
(adapt
).
learnhd
calculates the weight change dW
for a given neuron from the neuron's input P
, output A
, decay rate DR
, and learning rate LR
according to the Hebb with decay learning rule:
dw = lr*a*p' - dr*w
learnh
,
adaptwb
,
trainwb
,
adapt
,
train