Design generalized regression neural network
net = newgrnn(P,T,spread)
Generalized regression neural networks (
grnns) are a kind of radial
basis network that is often used for function approximation.
grnns can be
designed very quickly.
net = newgrnn(P,T,spread) takes three inputs,
Spread of radial basis functions (default = 1.0)
and returns a new generalized regression neural network.
The larger the
spread, the smoother the function approximation. To fit
data very closely, use a
spread smaller than the typical distance between
input vectors. To fit the data more smoothly, use a larger
newgrnn creates a two-layer network. The first layer has
radbas neurons, and calculates weighted inputs with
and net input with
netprod. The second layer has
neurons, calculates weighted input with
normprod, and net inputs with
netsum. Only the first layer has biases.
newgrnn sets the first layer weights to
P', and the
first layer biases are all set to
0.8326/spread, resulting in radial basis
functions that cross 0.5 at weighted inputs of +/–
spread. The second layer
W2 are set to
Here you design a radial basis network, given inputs
P and targets
P = [1 2 3]; T = [2.0 4.1 5.9]; net = newgrnn(P,T);
The network is simulated for a new input.
P = 1.5; Y = sim(net,P)
Wasserman, P.D., Advanced Methods in Neural Computing, New York, Van Nostrand Reinhold, 1993, pp. 155–61
Introduced before R2006a