Stopping the neural network by tr.gradient

In training an ANN using FITNET , I noticed , the tr.gradient gives a row matrice that the number of columns are the number of iterations , and the last column is the gradient reported on the train window
I tried doing :
for h=Hmin:dH:Hmax
j = j+1
net = fitnet(10);
net = init(net); % Improving Results since we use patternet we should use init
[ net tr y ] = train( net, x, t );
e = gsubtract(t,y);
performance = perform(net,t,y)
if tr.gradient(end) < 0.05
tr.stop
end
but it only stops the Validation test , not the actual training test , is there a way to do this ? and also when I retrain after a gradient like 0.503 and I get a smaller gradient , if from my outputs one is calculated not so precisely , the only thing happens is that , another output will be unprecise.
I have 8 inputs and 3 outputs

 Accepted Answer

Maybe you are looking for the property “trainParam.min_grad”.
net = fitnet(10);
net.trainParam.min_grad % default 1e-7
net.trainParam.min_grad = 1e-5;
net.trainParam.min_grad % changed to 1e-5

7 Comments

Yes exactly , but how should I stop the training after comparison ?
It stops automatically.
Why are you not using the default values?
well I didn't know that the default value is 1e-7 , it seemed to me that 1e-5 is the smallest number I have seen on the training while I was following.
is there any way to increase the accuracy of the loop ? only by changing H ? I changed it from 40 to 80, but Hub is 139
Dear Professor , still , with Ntrials = 20 , and as you have mentioned that MATLAB will stop in a very low gradient. I wonder that when I manually check some of my inputs , there is an error as big as integers. sometimes like 8 or 9.
I have found the following usually sufficient
MSEgoal = 0.01*mean(var(target',1))
MinGrad = MSEgoal/100
net.trainParam.goal = MSEgoal;
net.trainParam.min_grad = MinGrad;
Hope this helps.
^Thank you for formally accepting my answer
Greg
Thank you very much dear professor
I wish I could accept ,but it was a comment
Not a problem.
Good Luck
Greg

Sign in to comment.

More Answers (0)

Categories

Find more on Deep Learning Toolbox in Help Center and File Exchange

Asked:

on 6 Mar 2015

Commented:

on 15 Mar 2015

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!