MATLAB and Simulink resources for Arduino, LEGO, and Raspberry Pi

Learn moreOpportunities for recent engineering grads.

Apply Today**New to MATLAB?**

Asked by mangood UK
on 5 Jan 2013

hi.. i want to train my neural network by use matlab command but the problem is that my transfer function in my neural network equal tanh(m*x) the m may 0.01 or 0.001 or 2 the derivative of function is m*(1-tanh(m*x).^2)

can any suggestion !!!!!

thank in advance

Answer by Greg Heath
on 6 Jan 2013

Edited by Greg Heath
on 8 Jan 2013

Accepted answer

What you want to do is totally unnecessary. What is your reason for wanting to do this?

tanh = tansig.

The weight initialization logic in init automatically determines an optimal range for the intial weights. The final weights needed are automatically determined by the training algorithm.

Since your very obsolete version of newff does not automatically normalize the input to the range [-1,1], you can test what happens if you keep the same target but multiply all of the inputs by the same constant (equivalent to using the multiplier in tanh).

[ x0, t ] = simplefit_dataset;

MSE00 = mean(var(t',1)) % Reference MSE

H = 6,

nmax = 10,

Ntrials = 10,

rng(0),

j = 0

for n =1:nmax

j = j+1

x = n*x0;

for i = 1:Ntrials

net = newff( minmax(x), [ H O ] ,{'tansig' 'purelin'});

net.trainParam.goal = 0.01*MSE00;

[net tr ] = train(net,x,t);

time(i,j) = tr.time(end);

Nepochs(i,j) = tr.epoch(end);

MSE = tr.perf(end);

R2(i,j) = 1-MSE/MSE00;

end

end

n = (1:nmax)

time = time

Nepochs = Nepochs

R2 = R2

Does looking at the summary statistics of time, Nepochs and R2 indicate any advantage to using a multiplier?

Hope this help.

Thank you for formally accepting my answer.

Greg

P.S. If you have 2012, why are you using the very obsolete version of newff?

Does it have anything to do with the fact that the more recent obsolete version of newff (net =newff(x,t,H)) or the current version of fitnet (fitnet(H) ) normalize, by default, the inputs to the range [-1,1]?

Show 2 older comments

Greg Heath
on 8 Jan 2013

I don't follow. The default algorithm already is

h = tanh(IW*x + b1);

y = LW*h + b2;

where IW is optimized to minimize mse(t-y).

Replace the x in the 1st equation by x0.

[ x0, t ] = simplefit_dataset;

Then cut and paste the code into the command line.

The results for 100 designs are tabulated in 3 10X10 matrices.

Rows show the results of varying the multiplier n in x=n*x0.

Columns show the effect of changing the random initial weights.

You can just peruse the rows or calculate the min, mean, median, mean and max of the matrices to see if there is any indication that decreasing the scale of x (equivalent to increasing the scale of IW) is of any importance.

## 2 Comments

## Greg Heath (view profile)

Direct link to this comment:http://www.mathworks.com/matlabcentral/answers/57998#comment_120813

I don't understand your problem.

What version of the NNTBX are you using?

Details?

Code??

## mangood UK (view profile)

Direct link to this comment:http://www.mathworks.com/matlabcentral/answers/57998#comment_120814

dear Greg i have matlab 2012. i have this neural network, net = newff([0 1;0 1 ;0 1],[4 1],{'tansig','purelin'})

the problem is i want to replace the tansig with tanh(3*X) and train network using net=train(net,I,O);

can we do this in matlab!!!! any suggestion to solve the problem