Enhancement Processing Time and Accuracy Training via Significant Parameters in the Batch BP Algorithm
Автор: Mohammed Sarhan Al_Duais, Fatma Susilawati Mohamad, Mumtazimah Mohamad, Mohd Nizam Husen
Журнал: International Journal of Intelligent Systems and Applications @ijisa
Статья в выпуске: 1 vol.12, 2020 года.
Бесплатный доступ
The batch back prorogation algorithm is anew style for weight updating. The drawback of the BBP algorithm is its slow learning rate and easy convergence to the local minimum. The learning rate and momentum factor are the are the most significant parameter for increasing the efficiency of the BBP algorithm. We created the dynamic learning rate and dynamic momentum factor for increasing the efficiency of the algorithm. We used several data set for testing the effects of the dynamic learning rate and dynamic momentum factor that we created in this paper. All the experiments for both algorithms were performed on Matlab 2016 a. The stop training was determined ten power -5. The average accuracy training is 0.9909 and average processing time improved of dynamic algorithm is 430 times faster than the BBP algorithm. From the experimental results, the dynamic algorithm provides superior performance in terms of faster training with highest accuracy training compared to the manual algorithm. The dynamic parameters which created in this paper helped the algorithm to escape the local minimum and eliminate training saturation, thereby reducing training time and the number of epochs. The dynamic algorithm was achieving a superior level of performance compared with existing works (latest studies).
Enhancement processing time, accuracy Training, Dynamic momentum factor, Dynamic learning rate, Batch Back-propagation algorithm
Короткий адрес: https://sciup.org/15017123
IDR: 15017123 | DOI: 10.5815/ijisa.2020.01.05
Текст научной статьи Enhancement Processing Time and Accuracy Training via Significant Parameters in the Batch BP Algorithm
Published Online February 2020 in MECS
An artificial neural network (ANN) is a mathematical model inspired by biological node systems [1,2]. This is a description that offers great flexibility in modeling quantitative methods. The ANN has a strong ability to complete tasks many and is considered an active tool for training and classification[3-6]. An ANN provides a supervised learning algorithm which implements a nonlinear model within [0, 1] or [-1, 1], depending on the activation function [ 7, 8]. An ANN involves parallel processing, which consists of several parameters which need to be adjusted to minimize error training
Heuristic techniques are the most current methods for improving the training of the BBP algorithm. These methods include learning rate and momentum factor; these are significant parameters for adjusting the weights in training [ 9-10] . In general, the values of learning rate and momentum factor are fixed and are chosen from the interval [0, 1]. Manually set values for each training rate and momentum factor do not help the BBP algorithm escape from local minima or to meet the requirements of speeding up the training of the BBP algorithm.
Generally, there are two techniques for selecting the values for each learning rate and momentum factor. The first is set to be a small constant value from interval [ 0,1], the second the selected series value from[0,1], [11,12]. Therefore, one of the requirements for speeding up the BBP algorithm is adaptive learning rate and momentum factor together [13,14]. The learning rate should be sufficiently large to allow for escaping the local minimum to facilitate fast convergence to minimize error training. But the biggest value of the training rate leads to fast training with oscillation error training [15,16] .On the contrary, the small value of learning rate leads to reflex the weight that is lead to flat spot which makes BBP algorithm slow training [17,18]. The big values or small values are not suitable for smooth training. It is difficult to select the best or suitable values in training [19, 20].
The heuristic method is a current method for improving the BP algorithm, which covers two parameter training rate and momentum factor. The weaknesses, the values of training, rate and momentum are selected manual values. On the other hand, improves BP algorithm through creating dynamic training rate. But the weaknesses of these dynamic training depend on the initial value of training rate. Manually set values for learning rate and momentum factor do not help the BBP algorithm avoid local minima or meet the requirements of increasing training speed and difficult to choose a suitable value for learning rate and momentum factor. One way of avoiding localminima and speeding up the training of the BBP algorithm is to use a large value for training rate.
However, a small adjustment to the training weight slows the training of the BBP algorithm, while large adjustment to the weight results in unsmooth training. Most previous studies have tried to escape the local minimum through adaptive learning rates or momentum factors to improve the BBP algorithm. The persisting major problem, though, is the accumulation of weight in the BBP, which decreases accuracy and increases the time of training.
To fill the gap, we need to avoid the gross weight training, through creating dynamic parameter each learning rate and momentum factor with boundary to control the weight updated.
The remainder of this paper is organized as follows: Section II, literature review ; Section III, theoretically of training batch of back-Propagation algorithm; Section IV materials and method; Section V Created the dynamic learning rate and momentum factor, Section VI Experimental results; Section VII Discussion to validate the performance of improve algorithm; Section VIII Evaluation of the Performance of Improved Batch BP algorithm Section XI Conlusion.
-
II. Related Works
Currently, works have been carried out, such as author, In [21] Improved back propagation Algorithm by adaptive the momentum factor by dynamic function. The dynamic function which proposed is depend on the two manual factors parameters. The learning rate selected by manual value to control the weight updated. Each factor and learning rate selected different value for each data set. To validate, this study through compares the performance of improved algorithm with BP algorithm.The proposed algorithm gave superior training algorithm than other with whole data set. In [22] improved the BP algorithm through two techniques, the training rate and momentum factor, values of training rate were fixed at different values. The idea of this study is to set the value of training, the rate to be large initially, and then to look at the value of error training after iteration. If the error (e) training is increased, the fit produced changes the value of training, rate multiplied by less than one and then recalculated in the original direction. If the iteration error can be reduced, this integration is effective. Therefore, by changing the training rate multiplied by a constant greater than one, the next iteration is calculated continuously. In[23] compared several techniques such as BP with momentum, BP with the adaptive learning rate, BP with adaptive learning rate and momentum, Polak-Ribikre CGA, Powell-Beale CGA, scaled CGA, resilient BP (RBP) conjugate gradient algorithm (CGA), and Fletcher-Reeves. The epochs of training are fixed with different values. The BP algorithm with adaptive learning rate and momentum gave superior accuracy training at 1000 epochs.
-
III. Theoretically of Training Batch of Back-Propagation Algorithm
In this section we will foucas on the theoretivall of BBP algorithm which cover the mathematical framework with significant parameters such as learning rate and momentum factor, and the gradient descent method. The BBP algorithm was designed to train models to achieve a balance between the actual data and the target correctly in the input patterns; this is done using the gradient descent method to minimize the error training of the BBP algorithm.
-
A. Mathematical Framework BBP Algorithm with γ and µ
The training BP algorithm is consist of three stages , namely ,forward propagation , in the feed forward phase , each input unit x receives an input signal x and broadcasts this signal to the next layer until the end layer in the system. The backward propagation this step is starting when the output of the last layer reach to end step then the start feedback. The end steep is update the weight, in the batch BP algorithm the weight adjustment stage for all layer adjusted simultaneously. The training of BP algorithm shown in the figure below

Fig.1. Batch BP algorithm stractur
The BP Algorithm with Training rate and Momentum term. The weight updates between the neuron k from output layer and neuron j from hidden layer as follows All steps training of the BP algorithm with training rate and momentum. Any neuron k from output layer and neuron j from hidden layer, the weight is adjusted between them as follows:
∆ wjr ( t + 1) = w ( t ) - γ ∂ E + µ ∆ wjr ( t - 1) (1)
wjk ( t )
All steps training of the BP algorithm with training rate and momentum.
Forward Pass Phase: In this stage, the weight calculates layer by layer until reaching the end layer or output layer of the end layer under affected the activation function. Every variable f ′ ( x ) receives an input with initial weight and broadcasts to next the layer L ,..., L . The input signal L for each hidden unit is the sum of each input signal x . Each input layer as follows:
L - inh
k
= u 0 h + ∑ uikxi i = 1
Compute the output of first hidden layer of L
Lh = f (L-inh)(2)
Send the result to LL ( LL j = 1, 2,..., p ) and Calculate the input signal
LLj =v0h +∑vhjLh i=1
Calculate the output layer of second hidden LL
LLj = f (LL-inh)(4)
Send the output LL to input Y and then get the input layer of Y - inr
k
Y-inh =w0h +∑LLJwjr(5)
i = 1
Compute theoutput layer of Y
Yr = f ( Y - inr )
Backward Pass Phase: when the weight reaches to the output of the last hidden layer to end step then the start all steps as follow
n er =∑(tr-Yr)(6)
r = 1
The local gradient of output layer Y at neuron r is defined by below equation
δr =erf(Y-inr), f'(Y-inr)=Y-inrf(1-Y-inr)(7)
Compute weight correction term (update later)
Δwjr = -γδrLLj + µ∆wjr(t-1)(8)
Calculate, bias correction term (update w г later
Δw0r = -γδ0r + µ∆w0r(t-1)(9)
Send δ to hidden layer , each hidden unite
LLj , j = 1,..., p )
The single error or error back propagation from units in layer above to given as below
m
δ - inj = ∑ δ r W jr (10)
r = 1
The local gradient of hidden layer LL at neuron j is defined by below equation
δj =-δ-inj f′(LL-inJ)(11)
Compute weight correction term (update v later)
∆vhj(t+1)=-γδjLh+µ∆whj(t-1)(12)
Calculate, bias correction (update v later)
∆vjr(t+1) = -γδj + µ∆v0r(t-1)(13)
Send δ to first layer h=1,…,q).
The single error or error backpropagation is given as below
b
δ - inh = ∑ δ j v hl (14)
j = 1
The local gradient of hidden layer L compute as follow '
h - inh - inh , - inr - inr - inr
Compute weight correction (update u later)
∆uih = -γδhxi + µ∆uih(t-1)(16)
Calculate bias weight corrective (used to update u later)
∆u0h = -γδh+ µ∆u0h(t-1)(17)
Update Weight Phase: the weight is update of the output layer j =(0,1,2,..., p;r = 1..., m) as follows wjr(t+1) =w(t) -γ∆wjr(t) + µ∆wjr(t-1)(18)
For Bias w0r(t+1) =w(t) -γδj + µ∆wjr(t-1)(19)
For each hidden layer ( LL h = 0,..., q ; j = 1,... p )
vhj(t+1) =vhj(t) -γ∆whj(t) + µ∆whj(t-1)(20)
For Bias v0j(t+1)=v0j(t)-γ∆w0j(t)+µ∆w0j(t-1)(21)
For each hidden layer L ( i = 0,..., n ; h = 1,..., q )
uih(t+1) =uih(t) -γ∆uih(t) + µ∆uih(t-1)(22)
For Bias u0h(t+1) =u0h(t) -γ∆u0h(t) + µ∆u0h(t-1)(23)
The flowchart of the BP algorithm with trail value for each training rate and
-
B. Gradient Descent Backpropagation
Algorithm
The gradient descent method is one of the most popular methods used to adjust the weight for minimizing the training error in the BP algorithm [24, 25] Unfortunately, the gradient descent method is not guaranteed to find the global error [26, 27] .It may, therefore, result in leading to a local minimum, as the gradient descent method is not powerful enough to adjust for the weight, apart from limited changes [28] .The batch gradient descent method computes gradients using pattern training and updates the weight in the final stages of training [29]. This method focuses on modifying one parameter or more than one as adaptive of activation function or modifying the gain. The value of the gain is seen in the effect of the behavior of the convergence algorithm [30]. The author [31] proposed a new algorithm through the modified gain in the activation function. The proposed algorithm gives superior training compared with the BBP an algorithm.
Gradient descent ∆ E taken partial of the differentiate of error with respect the weight w in each layer. ANN which is consist one input, one hidden layer and one output also focus for one neuron as k from output layer and one neuron j from hidden, as follows

Fig.2. Gradient Descent for one Hidden Layer
1 k
Let the error training is given by E = ∑ ( t - o )2 , 2 k = 1
activation function of training is sigmoid function
f ( x ) = . When the gain equals is one. The
-
1 + Exp ( - x )
weight update between the neuron K , from output and neuron j from hidden layer as follow
A w k ( new ) = A w k ( old ) - / А w k + ^A v hj (t - 1) (24) |
d E = 8 E 8 net k d oj d netk 8 oj |
The gradient between neuron K from output layer and neuron j from hidden layer |
Insert equation (35) into equation (34) it becomes as below |
d E 8 E 8 ojk 8 netk A w ,k = = ---- (25) d W jk d o k 8 net k wjk |
8 E 8 E 8 netk 8 o -----------— (36) d net,. 8 net, 8 o, 8 net, j kj j |
d E 1 d , 2 a = i a ( tk ok ) 8 o k 2 8 o k |
A 8E 8 E 8 netk 8 o j d net j A w =--------------- (37) 5 w„ 5 net, d o, 8 net, 8 w„ ij k j j ij |
I E" = - ( t k - o k ) (26) d o k |
8 E _ 8E 8 ok d netk d ok d netk |
d o jk ; 8 1 d netk 8 netk 1 + e - x |
8E — = - ( tt - o k ) o k (1 - o k ) (38) d netk |
8 oik -=- = ok (1 - ok ) (27) d netk |
Put 5 k = ( t t - o k ) o k (1 - o k ) then get |
■' • = x, (28) w jk |
= 5 k (39) d netk |
Insert equation 25, 26, 27 into equation (25) it becomes as follow |
dnetk 8 .j ... —- = —(/ o,w„ + 9, ) 8 o j 8 o j ^ j ij j |
A w jk = - ( t k - o k ) o k (1 - o k ) x j (29) |
d netk k = W y (40) 8oj |
Insert equation (29) into equation (24) to get a new equation for update the weight as below |
.5 o j =_8 L *„ 1 |
A w jk ( new ) = A w jk ( old ) - / ( t , - o k ) o k (1 - o k ) x j (30) |
d nety d nety ( 1 + e j ) |
Put the 5. = ( tk - ok ) ok (1 - Ok ) then get the equation |
6 o, -=- = f '( netj ) = Gy (1 - Gy ) (41) 6 netj |
A w k ( new ) = w jk ( old ) + n§ k x j (31) |
8 net 8 ( к A |
For calculate the local gradient between neuron j from hidden layer and neuron i from input layer. In this case the weight update as follow |
------=---- / XiWii + 9, 8 W ,j 8 W j ( k = 1 J 8 net, --x- = x (42) |
A W j ( t + 1) = W j ( t ) - yA W j ( t ) (32) |
d Wj |
The gradient between neuron j from hidden layer and neuron i from input layer given as follow |
Insert equation (26, 27, 28) into equation (25) it becomes as below |
BE 8E 8 net, — = — ---(33) 5 W 5 nety d Wy |
A W = = - ^ ( t t - o k o (1 - o k > w yo^ (1 - o j > x d W lj k = 1 |
5=dE .j^ (34) 8 nety 8 oj 8 nety |
From equation (29) replace ( tt - ok ) ok (1 - ok ) by 5 |
∂ E
∆w = ij ∂Wij
-
∑ δ kwijoi (1 - oj ) xi k = 1
Put δ j = ∑ j = i δ k w ij o i (1 - o j ) in (42) to get j = 1
∂ E
∆w== ji ∂W
Insert (43) into equation (32), then the weight update between neuron j from hidden layer and neuron i from input layer as below
∆W (new) = w (old) +γδ x(45)
-
IV. Materials and Method
This kind of this research belongs to the heuristic method. This method includes the learning rate and momentum factor. To investigate the aims of this study there are many steps as follows:
-
A. Data set
The data set is very important for verification of improved DBBPLM algorithm. In this study, all data are taken from UCI Machine Learning Repository through the link All real data set to change to become nomination data set between [0,1] and divided into two set training set and testing set.
-
B. Neural Network Model
We propose an ANN model, which is a three-layer neural network that has an input layer, hidden layer, and output layer. The input layer is considered to be { x 1, x 2, , x }, which represents the nodes; the nodes depend on the types or attributes of the data. The hidden layer is made of two layers with four nodes. The output layer Y is made of one layer with one neuron. Three basis, two of them are used in the hidden layers and one in the output layer. Finally, the sigmoid function is employed as an activation function
-
V. Created the Dynamic the Prameters Learning Rate and Momentum Factor
The weight update between neuron k from the output layer and neuron j from the hidden layer ( w ) in equqtion (1) which recorded as above.
To enhance the BBP algorithm, which is given by the Equation above (1) to avoid the local minima and to avoid saturation training. The learning rate should be sufficiently large to allow escape from the local minimum and to facilitate fast convergence to minimize error training . However, the value of the learning rate which is too large leads to fast training with an oscillation in error training. To ensure a stable learning BP algorithm, the learning rate must be small. Depend on above, we will create the dynamic training rate with boundary to keep the weight as moderate as follows:
sec h ( sh (1 - sh ) γ dmic = + exp( sh . e ) (46)
Where the sech θ is Hyperbolic functions, θ is angle, and the sech θ = 2/( e θ + e - θ ) . The sh(1-sh) is slop the first derivative of the activation function of the out put of second hidden layer and (e) is the error training. We can see the value of ( sh ) is located between 0 and 1, the value of ( sh ) is a boundary. In the same way, the training error ( e ) is boundary value 0 ≤ e≤ 1. Each sh and e boundary function, that is lead to sh×e is boundary function, 0 ≤sh×e≤1. Depend on the above we can find the 1≤ ≤2.78 . the dynamic γ is boundary funcation.
The second significant parameter is momentum. To get smooth training and avoid inflation in the gross weight of the added values for momentum factor, the fitting producer through creating dynamic momentum factor and implicate the Depend above we can create the dynamic momentum factor µ as follow sech(sh(1 - sh)
+ exp( sh . e )
We Insert the equation (46) and (47) into equation (1), then the weight is updated between any layer as below sech(sh(1-sh) ∂E wjk(t+1)=wjk(t)-[ + exp(sh.e)]
2 wjk ( t )
+
sec h ( sh (1 - sh )
+ exp( sh . e )
∆ wjk ( t - 1)
For the equation (48) the dynamic algorithm will be updated the weight under affected the dynamic learning rate and dynamic momentum factor.
-
VI. Experimental Results
The accuracy of training is calculated as follows [32,33]
1-absolut(T -O )
Accuracy (%) = i i ∗ 100 where UP and LW are the upper bound and lower bound of the activation function.
-
A. Experimental Results for the DBBPLM Algorithm Using the XOR Problem
Ten experiments were carried out using Matlab, and the average (AV) was taken for several criteria used for measurement of the training performance. Ten experimental has been done for DBBPLM with XOR and then taken the average and S.D.The average and S.D of experimental results are tabulated in Table 1.
Table 1. Average the Performance of DBBPLM algorithm with XOR
Time-sec |
Epoch |
MSE |
Accuracy |
|
Average |
2.2463 |
1681 |
9.998E-05 |
0.9859 |
S.D |
0.36127166 |
0 |
0 |
2.22E-16 |
From Table 1., the formulas proposed in equations (46 and 47) help the back-propagation algorithm to enhance the performance of the training. Whereas the average time training is t = 2.2463 seconds and the epoch is 1681 epoch. The smallest value of standard deviation S.D indicates the scatter of time or error training is very close for every experiment, whereas the S.D = 0.36127166 second. The accuracy rate is 0.9859.
-
B. Experiment result of the BBP algorithm with XORProblem
For the dynamic DBBPLM algorithm we run the dynamic algorithm 10 time with XOR problem and then take the averge for time , epoch and accuracy rate as follow
Table 2. Performance of BBP algorithm with XOR
Value of |
times |
MSE |
Epoch |
|
γ |
µ |
|||
0.1 |
0.004 |
52.4760 |
9.999e-05 |
67490 |
0.0001 |
0.0001 |
6530 |
0.5011 |
194345 |
0.004 |
0.95 |
5670 |
1.635e-04 |
153329 |
0.1 |
0.1 |
55.9070 |
9.999e-05 |
76190 |
0.2 |
0.2 |
31.0090 |
9.997e-05 |
37839 |
Average |
2467.8784 |
0.1003127 |
105838.6 |
|
S.D |
2978.0706 |
0.2003937 |
58416.301 |
From Table 2., the range of the training time is 31.0090.0090≤ t≤ 6530 seconds. We consider the value 31.0090second is as minimum training time and the value6530 second is as maximum training time. The best performance of the BP algorithm is achieved at γ = 0.2 and at µ =0.2 whereas the training time is 21.0090 seconds, the worst performance of the BP algorithm is achieved at γ = 0.0001 and µ =0.0001 for first, whereas the training time is 65300 second.
-
C. Experiments result of the DBBPLM algorithm with Balance- Train set
We implement the DBBPLM algorithm ten time with the balance data training set. and then the average was taken for several criteria. The experiment's result is tabulated in the Table 3.
Table 3. Average the Performance of DBBPLM algorithm with Iris- Train set
Time-sec |
Epoch |
MSE |
Accuracy |
|
Average |
3.3663 |
61 |
9.887E-05 |
0.992 |
S.D |
1.55506733 |
13.169662 |
9.139E-07 |
0.0016697 |
From Table 3., the dynamic approach for training rate and momentum term reduce the time required for training and enhance the convergence of the time training. The average training time is 3.3663 seconds at an average epoch is 61 epochs. The average accuracy training is 0.992
-
D. Experiments of the BBP Algorithm with BalanceTrain set
We tested the BBP algorithm with several optimum values. The experiments result recorded as table V with iris –training set
Table 4. Performance of BBP bbp algorithm with Iris- Train set
Value of |
times |
MSE |
Epoch |
|
γ |
µ |
|||
0.1 |
0.004 |
67.5020 |
9.7199e-05 |
323 |
0.01 |
0.25 |
61.5210 |
9.8827e-05 |
1129 |
0.004 |
0.95 |
220.0530 |
9.9736e-05 |
4361 |
0.005 |
0.09 |
153.8200 |
9.9950e-05 |
2993 |
0.005 |
0.001 |
179.8890 |
9.9960e-05 |
3535 |
Average |
129.157 |
9.913E-05 |
2468.2 |
|
S.D |
71.7656181 |
1.053E-06 |
1509.4044 |
From Table 4., the best performance of the BP algorithm was at γ = 0.01 and µ =0.25 whereas the time training is 61.5210 ≈ 52 second while the best performance of the BP algorithm is achieved at γ = 0.004 and µ =0.95 whereas the time training is 220.0530 ≈ 220 seconds. In addition, the range of the average training time is 61.5210≤ t ≤ 220.0530seconds.
-
E. Experiments result of the DBBPLM Algorithm with Balance- Testing set
Ten experiments has been done with balance testing and then taken the average of several criteria. The experiment result is tabulated in the Table 5.
Table 5. Performance of DBBPLM with Iris- Testing set
Time-sec |
Epoch |
MSE |
Accuracy |
|
Average |
3.6801 |
88 |
9.821E-05 |
0.9906 |
S.D |
0.98730577 |
20.813697 |
1.143E-06 |
0.0019355 |
From Table 5., the average training time is 3.6801 seconds at an average epoch is 88 epoch. The average S.D of time is 0.98730577. The accuracy rate is very highest about 0.9887.
-
F. Experiments result of the BBP Algorithm with Balance- Testing set
We tested the BBP algorithm with several optimum values. The experiments result recorded in the Table 6., with iris –testing set
From Table 6., the best performance of the BP algorithm was at γ = 0.1 and µ = 0.1, whereas the time training is 79.9910 ≈ 80 seconds, while the worst performance of the BBP algorithm is achieved at γ =0.0001 and µ =0.0001 whereas the time training is 5600 second. The range of the average training time is 79.9910≤ t ≤ 5600 seconds. The average S.D of time is greater than one.
Table 6. Performance of BBP Algorithms with Iris- Testing set
Value of |
times |
MSE |
Epoch |
|
γ |
µ |
|||
0.1 |
0.004 |
79.9910 |
5.5408e-05 |
2867 |
0.0001 |
0.0001 |
5600 |
0.2049 |
96800 |
0.005 |
0.09 |
4868 |
0.8193 |
7352 |
0.005 |
0.001 |
1.422e+3 |
5.0238e-05 |
39854 |
0.01 |
0.25 |
591.4530 |
5.0937e-05 |
14140 |
0.1 |
0.1 |
70.0050 |
5.6064e-05 |
1840 |
Average |
2105.2415 |
0.1707354 |
27142.167 |
|
S.D |
2267.46456 |
0.2995366 |
33674.904 |
-
G. Experiments result of the DBBPLM Algorithm with Iris.- Train set
Table 7. Average the Performance of DBBPLM Algorithm with Balance - Train set
Time-sec |
Epoch |
MSE |
Accuracy |
|
Average |
9.9023 |
77 |
2.9658e-04 |
100 |
S.D |
0.52140945 |
0 |
0 |
0 |
From Table 7., the experiments result indicates the dynamic learning rate and momentum factor helps the BBP algorithm for reducing the time training. The average time is 9.9023 ≈ 10 seconds with 77 epoch. The average S.D of time is 0.52140945. .
-
H. Experiments result of the BP Algorithm with Balance - Train set
We test the manual algorithm at several values for each learning rate and momentum factor. The result are recoded in the Table 8.
Table 8. Performance of DBBPLM Algorithm with Balance - Train set
Value of |
times |
MSE |
Epoch |
|
γ |
µ |
|||
0.1 |
0.004 |
105.5750 |
9.996e-05 |
6882 |
0.004 |
0.95 |
96.5370 |
1.0000e-04 |
6176 |
0.005 |
0.09 |
1.85e+3 |
9.99e-05 |
124715 |
0.005 |
0.001 |
4963 |
1.185e-04 |
117460 |
0.1 |
0.1 |
93.0560 |
9.98e-05 |
6289 |
0.2 |
0.2 |
76.7300 |
9.91e-05 |
9554 |
Average |
1197.483 |
0.0001029 |
45179.333 |
|
S.D |
1197.483 |
0.0001029 |
45179.333 |
From Table 8., the range of the training time is 76.7300≤ t ≤ 105.5750 seconds, the range of training time is very widely.
-
I. Experiments result of the DBBPLM Algorithm with iris.- Testing set
Table 9. Average the Performance of DBBPLM Algorithm with Balance - Testing set
Time-sec |
Epoch |
Accuracy |
|
Average |
6.482 |
76 |
0.9860 |
S.D |
0.3571277 |
0 |
0 |
From Table 9., the average time is 6.482 6 seconds with 76 epochs. The accuracy rate is 0.9860, the accuracy rate is the nearest one
-
J. Experiments result of the manually BBP Algorithm with Iris.- Testing set
We test the manual algorithm at several optimum values for each learning rate and momentum factor . The result are recoded in the Table 10.
Table 10. Performance of BBP Algorithm with Balance - Testing set
Value of |
Time-S |
MSE |
Epoch |
|
γ |
µ |
|||
0.1 |
0.004 |
81.9550 |
2.772e-4 |
11490 |
.0001 |
0.0001 |
4976 |
0.0077 |
369071 |
0.004 |
0.95 |
3693 |
0.0077 |
152912 |
0.005 |
0.09 |
3546 |
5.35e-04 |
129706 |
0.005 |
0.001 |
1.97e+03 |
2.68e-04 |
273293 |
0.1 |
0.1 |
94.9780 |
2.76e-04 |
9833 |
0.2 |
0.2 |
75.6650 |
2.979e-04 |
5204 |
Average |
2062.514 |
0.0024363 |
135929 |
|
S.D |
1893.50603 |
0.0033302 |
132031 |
From Table 10., the range of the average training time is in the interval 75.6650 ≤ t ≤ 4976 seconds this means the range of time training is wide. The best performance of the BP algorithm is achieved at у = 0.2 and Ц =0.2 the worst performance of the BP algorithm is achieved at / = 0.0001 and µ =0.0001. The BBP algorithm suffers the highest saturation at γ = 0.0001 and µ =0.0001.
-
VII. Discussion to Validate the Performance of Improved Algorithm
To validate the efficiency of the improved algorithm, through comparing the performance of the DBBPML algorithm with the performance of the batch BP algorithm based on certain criteria[34, 35].We calculate the speed up training using the following formula [36].
Proccessing Time
Execution time of BBP algorithm
Execution time of DBBPLM algorithm
In this part we will compare the performance for each DBBPLM dynamic algorithm which listed in the from Table.1, 3,5,7,9 with the performance of manually algorithm wich listed in the Tables 2,4,6,8,10. The comparission result are tabulated in Table 11.
Table 11. Processing Time Improved the DBBPLM Algorithm versus BP with first structure
Dynamic (DBBPLM)algorithm |
A manually BBP algorithm |
||||
AV Time |
AV Epoch |
AV Time |
AV Epoch |
Processing Time Improved |
|
XOR |
2.2463 |
1681 |
2467.8784 |
105839 |
1098.641499 |
Balance Training |
9.9023 |
77 |
1197.483 |
45179 |
120.929784 |
Balance Testing |
6.482 |
76 |
2062.514 |
135930 |
318.1909904 |
Iris Training Iris |
3.3663 |
61 |
129.157 |
2468 |
38.367644 |
Testing |
3.6801 |
88 |
2105.2415 |
27142 |
572.0609494 |
From Table 11., it is evident that the dynamic algorithm provides superior performance over the Manually algorithm for all datasets. The range of the training time of the Dynamic algorithm is 2.2463 ≤ t ≤ 9.9023 s; this is a narrow interval, meaning that the dynamic parameters which created help the algorithm to reaches the global minimum in a short time and with few epochs. The range of training times of the BBP algorithm is 129.157 ≤ t ≤ 2105 s; this is a wide interval, meaning that the BBP algorithm has a long training time and a high level of training saturation. The dynamic algorithm is 2360.166 ≈ 2360 times faster than the BBP algorithm at its maximum.
Easily we can see in the end Column in the table above, dynamic algorithm which created in this paper gave faster training than manual algorithm. For example the dynamic algorithm is 1098.641499se 1099 time faster than manual algorithm at its maximum time training. Also dynamic algorithm which created in this paper gave faster training than manually algorithm. For example the dynamic algorithm is 38.367644se 38 time faster than manual algorithm at its minimum time training
From Figure 3 Obviously, we can see the dynamic algorithm taken a short time to reach the glob minimum, for all data set while the Manual algorithm takes a long time to reach a global minimum. The dynamic algorithm created in this study is able to escape the local minimum and remove the saturation training.

Fig.3. The performance of Dynamic algorithm versus Manually algorithm for time

■ Dynamic Algorithm (Average Epoch)
■ Manually algorithm (Average Epoch)
Fig.4. The performance of Dynamic algorithm versus Manually algorithm for epoch
From Figure 4 it is evident that the dynamic algorithm which created in this study provides superior performance over the Manually algorithm for all datasets. The dynamic algorithm taken few epochs to reach to the global minimum while the Manual algorithm takes a big value of epochs to reach the global minimum.
-
VIII. Evaluation of the Performance of Improved Batch BP Algorithm
To evaluated the performances of the improved batch BP algorithm for speeding up training the performances of the improved batch BP algorithm are compared to previous research works such, Hameed et al (2016) and Azami, (2015), the stop training for both study determined by number of iteration, the number iteration set by 2000 iteration and 1000 iteration respectively. For the purpose of the comparison between the results of this study and the previous works, the fit is re-run again with different stop training values according to previous work. The performance of the improved algorithm which proposed in this study give superior performance than exists works.
-
IX. Conclusion
We have introduced the dynamic DBBPLM algorithm, which trains by creating a dynamic function for each the learning rate and momentum factor. Sevral dataset were used, all data are taken from the UCI Machine Learning Repository through the following link: All datasets should be changed to become normalized in the interval [0, 1]. The dataset was divided into two sets, a training set and a testing set for validating the performance of the contribution. This function influences for the weight. Of the DBBPLM algorithm were carried out on Matlab 2012a with same goal error or limited error training. One of the main advantages of the dynamic training and dynamic momentum factor are reduces the training time, the training error, number of epochs and enhancement the accuracy of the training. The performance of dynamic DBBPLM algorithm which presented in this study gave superior performance compared with exists work.
Список литературы Enhancement Processing Time and Accuracy Training via Significant Parameters in the Batch BP Algorithm
- M. S. Al_Duais , F. S. Mohamad,” Improved Time Training with Accuracy of Batch Back Propagation Algorithm Via Dynamic Learning Rate and Dynamic Momentum Factor,” IAES International Journal of Artificial Intelligence , Vol. 7, No. 4, pp. 170~178, 2018.
- F.Ortega-Zamorano, J.M. Jerez, I. Gómez, & L.Franco,”Layer multiplexing FPGA implementation for deep back-propagation learning”. Integrated Computer Aided Engineering , 24(2), PP.171-185,2017
- Y.Hou, & H.Zhao, “Handwritten Digit Recognition Based on Improved BP Neural Network,” Proceedings of 2017 Chinese Intelligent Systems Conference.. CISC 2017 ,pp. 63-70, 2018. E. Ibrahim, Hussien, & Z.E.Mohamed, “Improving Error Back Propagation Algorithm by using Cross Entropy Error Function and Adaptive Learning rate,” International Journal of Computer Applications, 161(8), PP.5-9, 2017.
- C.B.Khadse, M.A Chaudhari, & V.B.Borghate, “Conjugate gradient backpropagation based artificial neural network for real time power quality assessment, “ Electrical Power and Energy Systems, 82, PP.197–206 ,2016.
- M. S. Al_Duais , F. S Mohamad & M.Mohamad,” Improved the Speed Up Time and Accuracy Training in the Batch Back Propagation Algorithm via Significant Parameter”. International Journal of Engineering & Technology, 7 (3.28) , pp. 124-130, 2018.
- Z. X Yang, G. S. Zhao, H. J. Rong & J.Yang, “Adaptive backstepping control for magnetic bearing system via feed forward networks with random hidden nodes,” Neurocomputing, 174, PP.109–120, 2016.
- Li, Jie, R.Wang , T.Zhang , & X.Zhang, “Predication Photovoltaic Power Generation Using an Improved Hybrid Heuristic Method. Sixth International conference on information Science and Technology ,pp. 383-387, 2016.
- M.Sheikhan, M.A Arabi, &D. Gharavian, “Structure and weights optimization of a modified Elman network emotion classifier using hybrid computational intelligence algorithms: a comparative study,” Connection Science, 27(4), PP.340-357, 2015, doi:10.1080/09540091.2015.1080224
- R. Karimi, F.Yousefi, M.Ghaedi & K.Dashtian,” Back propagation artificial neural network and central composite design,” Chemometrics and Intelligent Laboratory Systems, 159, 127–137,PP.2016
- R.Kalaivani, K.Sudhagar K, Lakshmi P, “Neural Network based Vibration Control for Vehicle Active Suspension System”, Indian Journal of Science and Technology.9(1), 2016.
- S. X. Wu, D. L.Luo, Z. W. Zhou, J. H.Cai, & Y. X. Shi, “A kind of BP neural network algorithm based on grey interval,” ,International Journal of Systems Science, 42(3) ,PP.389-96,2011.
- H.. Mo, J.Wang ,H. Niu, “Exponent back propagation neural network forecasting for financial cross-correlation relationship,” Expert Systems with Applications , 53 , PP.106-1016 , 2016.
- l.v Kamble, ;D.R Pangavhane, & T.P Singh, “ Improving the Performance of Back-Propagation Training Algorithm by Using ANN,”International Journal of Computer, Electrical, Automation, Control and Information Engineering, 9(1), PP.187- 192, 2015.
- J.M.Rizwan, PN.Krishnan, R.Karthikeyan, SR.Kumar, “ Multi layer perception type artificial neural network based traffic control,” Indian Journal of Science and Technology, 9(5), 2016.
- X.Xue, Y.Pan, R. Jiang, & Y. Liu, “Optimizing Neural Network Classification by Using the Cuckoo Algorithm,” 11 the International Conference on Natural Computation (ICNC) , PP.24-30,2015.
- L. He , Y.Bo, & G.Zhao,” Speech-oriented negate-motion recognition,” Proceedings of the 34th Chinese Control Conference ,PP.3553 -3558 , 2015.
- Y.Zhang, S.Zhao, & L.Tang, “Energy Consumption Prediction for Steelmaking Production Using PSO-based BP Neural Network,”Congress on Evolutionary Computation (CEC) PP.3207- 3214, 2016.
- Q.Abbas, F. Ahmad, & M. Imran, “Variable Learning rate Based Modification in Backpropagation algorithm (MBPA) of Artificial neural Network for Data Classification,” Science International, 28(3), 2016.
- V. P. S. Kirar, “Improving the Performance of Back-Propagation Training Algorithm by Using ANN,” World Academy of Science, Engineering, and Technology, International Journal of Computer, Electrical, Automation, Control and Information Engineering, 9(1) ,PP.187-192,2015
- A. A. Hameed , B. Karlik, & M. S. Salman, “Back-propagation algorithm with variable adaptive momentum”, Knowle dge-Base d Systems, 114,PP.79–87, 2016.
- W. Zhang, Z. Li, W .Xu, H.Zhou , “A classifier of satellite signals based on the back-propagation neural network,” In 8th International Congress on Image and Signal Processing (CISP), PP.1353-1357,2015.
- H. Azami , & J.Escudero, “A comparative study of breast cancer diagnosis based on neural network ensemble via improved training algorithms,” Proceedings of 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) ,PP.2836-2839,2015.
- Leema, N., Nehemiah, H. K., & Kannan, A. (2016). Neural Network Classifier Optimization using Differential Evolution with Global Information and Back Propagation Algorithm for Clinical Datasets. Applied Soft Computing. doi:http://dx.doi.org/doi:10.1016/j.asoc.2016.08.001
- Kamble, L V; Pangavhane, D R; Singh, T P. (2015). Improving the Performance of Back-Propagation Training Algorithm by Using ANN. International Journal of Computer, Electrical, Automation, Control and Information Engineering, 9(1), 187- 192.
- Liew, S. S., Hani, M. K., & Bakhteri, R. (2016). An Optimized Second Order Stochastic Learning Algorithm for Neural Network Training.Neurocomputing. Neurocomputing, 186, 74 – 89.
- Andrade, A., Costa, M., Paolucci,, L., Braga, A., Pires, F., Ugrinowitsch, H., & Menzel, H.-J. (2015). A new training algorithm using artificial neural networks to classify gender-specific dynamic gait patterns. Computer Methods in Biomechanics and Biomedical Engineering, 4, 382-390. doi:10.1080/10255842.2013.803081
- Kirar, V. P. (2015). Improving the Performance of Back-Propagation Training Algorithm by Using. International Journal of, 9(1), 187-192.
- Yui, M., & Kojima, I. (2013). A Database-Hadoop Hybrid Approach to Scalable Machine Learning. IEEE International Congress on Big Data (pp. 1-8). IEEE.
- Hamid, N. A., Nawi, N. M., & Ghazali, R. (2012). The effect of adaptive gain and adaptive momentum in improving training time of gradient descent back propagation algorithm on classifation problems. International Scientific Conferenc. 15. ISC.
- Nawi, N. M., Khan, A., & Rehman, M. Z. (2013). A new levenberg marquardt based back propagation algorithm training with cuckoo search. The 4th International Conference on Electrical Engineering and Informatics (pp. 18-23). ICEEI.
- N.M. Nawi, N. A Hamid, R. S , R.Ransing, Ghazali & M. N. Salleh, “Enhancing Back Propagation NeuralNetwork Algorithm with Adaptive Gain on Classification Problems,” networks, 4(2) ,2011.
- T.N .Bui & H. Hasegawa, “Training Artificial Neural Network using modification of Differential Evolution Algorithm,” International journal of Machine Learning and computing, 5(1) , PP.1-6, 2015.
- M. S. Al_Duais & F. S. Mohamad, A Review on Enhancements to Speed up Training of the Batch Back Propagation Algorithm. Indian Journal of Science and Technology, 9(46), 2016.
- M. S. Al Duais,& F.S.Mohamad, “Dynamically-adaptive Weight in Batch Back Propagation Algorithm via Dynamic Training Rate for Speedup and Accuracy Training. Training,”Journal of Telecommunications and Information Technology. 4 , 2017, doi.org/10.26636/jtit.2017: 113017
- S. Scanzio , S.Cumani, R.Gemello, R., F.Mana &F.PLaface, “Parallel implementation of Neural Network training for speech recognition,” Pattern Recognition Letters. Vol..1, no.11, PP.1302–1309,2010.