Download as pdf or txt
Download as pdf or txt
You are on page 1of 2

Mathematical Problems in Engineering 3

(i) It has high first hidden layer and small second hidden √𝑁𝑖 𝑁0 where 𝑁𝑖 is the input neuron and 𝑁0 is the output
layer. neuron. Since neural network has several assumptions which
(ii) Weights connecting the input to first hidden layer can are given before starting the discussion to prevent divergent.
be prefixed with most of the weights connecting the In unstable models, number of hidden neurons becomes too
first hidden layer and second hidden layer that can be large or too small. A tradeoff is formed that if the number
determined analytically. of hidden neurons becomes too large, output of neurons
becomes unstable, and if the number of hidden neurons
(iii) It may be trained only by adjusting weights and becomes too small, the hidden neurons becomes unstable
quantization factors to optimize the generalization again.
performance. In 2010, Doukim et al. [21] proposed a technique to find
(iv) It may be able to overfit the sample with any arbitrary the number of hidden neurons in MLP network using coarse-
small error. to-fine search technique which is applied in skin detection.
In 2006, Choi et al. [16] developed a separate learn- This technique includes binary search and sequential search.
ing algorithm which includes a deterministic and heuristic This implementation is trained by 30 networks and searched
approach. In this algorithm, hidden-to-output and input-to- for lowest mean squared error. The sequential search is
hidden nodes are separately trained. It solved the local min- performed in order to find the best number of hidden
ima in two-layered feedforward network. The achievement is neurons. Yuan et al. [22] proposed a method for estimation
best convergence speed. In 2008, Jiang et al. [17] presented of hidden neuron based on information entropy. This method
the lower bound on the number of hidden neurons. The is based on decision tree algorithm. The goal is to avoid the
number of hidden neurons is 𝑁ℎ = 𝑞𝑛 where 𝑞 is valued overlearning problem because of exceeding numbers of the
upper bound function. The calculated values represent that hidden neurons and to avoid the shortage of capacity because
the lower bound is tighter than the ones that has existed. The of few hidden neurons. The number of hidden neurons of
lower and upper bound on the number of hidden neurons feedforward neural network is generally decided on the basis
help to design constructive learning algorithms. The lower of experience. In 2010, Wu and Hong [23] proposed the
bound can accelerate the learning speed, and the upper learning algorithms for determination of number of hidden
bound gives the stopping condition of constructive learning neurons. In 2011, Panchal et al. [24] proposed a methodology
algorithms. It can be applied to the design of constructive to analyze the behavior of MLP. The number of hidden layers
learning algorithm with training set 𝑁 numbers. In the same is inversely proportional to the minimal error.
year, Jinchuan and Xinzhe [3] investigated a formula tested In 2012, Hunter et al. [2] developed a method used in
proper NN architectures. The advantages are the absence
on 40 cases: 𝑁ℎ = (𝑁in + √𝑁𝑝 )/𝐿 where 𝐿 is the number of trial-and-error method and preservation of the general-
of hidden layer, 𝑁in is the number of input neuron and ization ability. The three networks MLP, bridged MLP, and
𝑁𝑝 is the number of input sample. The optimum number of fully connected cascaded network are used. The implemented
hidden layers and hidden units depends on the complexity of formula: as follows, 𝑁ℎ = 𝑁 + 1 for MLP Network, 𝑁ℎ =
network architecture, the number of input and output units, 2𝑁 + 1 for bridged MLP Network and 𝑁ℎ = 2𝑛 − 1 for fully
the number of training samples, the degree of the noise in the connected cascade NN. The experimental results show that
sample data set, and the training algorithm. the successive rate decreases with increasing parity number.
The quality of prediction made by the network is mea- The successive rate increases with number of neurons used.
sured in terms of the generalization error. Generalization The result is obtained with 85% accuracy.
performance varies over time as the network adapts during The other algorithm used to fix the hidden neuron is
training. The necessary numbers of hidden neurons approx- the data structure preserving (DSP) algorithm [25]. It is an
imated in hidden layer using multilayer perceptron (MLP) unsupervised neuron selection algorithm. The data structure
were found by Trenn [18]. The key points are simplicity, denotes relative location of samples in high dimensional
scalability, and adaptivity. The number of hidden neurons space. The key point is retaining the separate margin under-
is 𝑁ℎ = 𝑛 + 𝑛0 − 1/2 where 𝑛 is the number of inputs lying the full set of neuron. The optimum number of hidden
and 𝑛0 is the number of outputs. In 2008, Xu and Chen nodes is found out by trial-and-error approach [26]. The
[19] developed a novel approach for determining optimum advantages are the improvement in learning and classifica-
number of hidden neurons in data mining. The best number tion, cavitations signal using Elman neural network. The
of hidden neurons leads to minimum root means Squared simulation results show that error gradient and 𝑁ℎ selection
Error. The implemented formula is 𝑁ℎ = 𝐶𝑓 (𝑁/𝑑 log 𝑁)1/2 , scheme work well. Another approach is to fix hidden neuron
where 𝑁 is number of training pairs, 𝑑 is input dimension, based on information entropy which uses decision tree
and 𝐶𝑓 is first absolute moment of Fourier magnitude distri- algorithm [27]. The 𝑁ℎ is generally decided based on the
bution of target function. experience. Initially 𝑁ℎ should be trained. The activation
In 2009, Shibata and Ikeda [20] investigated the effect values of hidden neuron should be calculated by inputting the
of learning stability and hidden neuron in neural network. training sample. Finally information is calculated. To select
The simulation results show that the hidden output connec- the hidden neurons, SVM stepwise algorithm is used. In this
tion weight becomes small as number of hidden neurons algorithm, linear programming SVM is employed to preselect
𝑁ℎ becomes large. This is implemented in random number the number of hidden neurons [28]. The performance is
mapping problems. The formula for hidden nodes is 𝑁ℎ = evaluated by RMSE (root means square error). The advantage
4 Mathematical Problems in Engineering

is improved computation time. The hidden neuron is selected unstable again. The problems in the fixation of hidden
empirically such as 2, 4, 6, 12, 24, and it is applied in neurons still exist. The properties of neural networks are
sonar target classification problem [29]. It is close to Bayes convergence and stability to be verified by the performance
classifier. The analysis of variance is done on the result of the analysis. The problem of wind speed prediction is closely
aspect angle dependent test experiments. The improvement linked to intermittency nature of wind. The characteristics of
of performance is by 10%. wind involve uncertainty.
Another approach to fix hidden neuron is the sequential The input and output neuron is to be modeled,
orthogonal approach (SOA). This approach [30] is about while 𝑁ℎ should be fixed properly in order to provide
adding hidden neurons one by one. Initially, increase 𝑁ℎ good generalization capabilities for the prediction. ANN
sequentially until error is sufficiently small. This selection comprising deficient number of hidden neuron may not
problem can be approached statistically by generalizing be feasible to dynamic system. During the last couple
Akaike’s information criterion (AIC) to be applied in unre- decades, various methods were developed to fix hidden
alizable model under general loss function including regu- neurons. Nowadays most predicting research fields have
larization. The other existing methods are trial and error, been heuristic in nature. There is no generally accepted
thump rule, and so forth. In thump rule, 𝑁ℎ is between size theory to determine how many hidden neurons are needed
of number of input neurons and number of output neurons. to approximate any given function in single hidden layer. If
Another rule is equal to 2/3 size of input layer and output layer it has few numbers of hidden neurons, it might have a large
[19]. The other 𝑁ℎ is less than twice the size of input layer. training error due to underfitting. If it has more numbers
The number of hidden neuron depends on number of inputs, of hidden neurons, might have a large training error due to
outputs, architectures, activations, training sets, algorithms, overfitting. An exceeding number of hidden neurons made
and noises. The demerit is higher training time. The other on the network deepen the local minima problem [30]. The
existing techniques are network growing and network prun- proposed method shows that there is a stable performance
ing [31, 32]. The growing algorithm allows the adaptation of on training despite of the large number of hidden neurons
network structure. This starts with undersized 𝑁ℎ and adds in Elman network. The objective is to select hidden neurons
neurons to number of hidden neurons. The disadvantages to design the Elman network and minimize the error for
are time consuming and no guarantee of fixing the hidden wind speed prediction in renewable energy systems. Thus,
neuron. research is being carried out for fixing hidden neuron in
The researchers have been implemented various methods neural networks. The optimal number of hidden neurons
for selecting the hidden neuron. The researchers are aimed based on the following error criteria. The error criteria such
at improving factors like faster computing process, more as mean square error (MSE), mean relative error (MRE), and
efficiency and accuracy and less errors. The proper selection Mean Absolute Error (MAE) are assessed on the proposed
of hidden neuron is important for the design of neural model performance. The fixing of the number of hidden
network. neurons in Elman network is based on minimal error
performance. The formulas of error criteria are as follows;
2
3. Problem Description 𝑁 (𝑌𝑖󸀠 − 𝑌𝑖 )
MSE = ∑ ,
The proper selection of number of hidden neurons has been 𝑖=1 𝑁
analyzed for Elman neural network. To select hidden neurons 󵄨󵄨 󸀠 󵄨󵄨
in order to to solve a specific task has been an important 1 𝑁 󵄨󵄨󵄨 (𝑌𝑖 − 𝑌𝑖 ) 󵄨󵄨󵄨
MRE = ∑ 󵄨󵄨 󵄨󵄨 , (1)
problem. With few hidden neurons, the network may not be 𝑁 𝑖=1 󵄨󵄨󵄨 𝑌𝑖 󵄨󵄨
󵄨󵄨
powerful enough to meet the desired requirements including 󵄨
capacity and error precision. In the design of neural network, 1 𝑁 󸀠
an issue called overtraining has occurred. Over training is MAE = ∑ (𝑌 − 𝑌𝑖 ) ,
𝑁 𝑖=1 𝑖
akin to the problem of overfitting data. So fixing the number
of a hidden neuron is important for a given problem. An
important but difficult task is to determine the optimal num- where 𝑌𝑖 is predicted output, 𝑌𝑖󸀠 is actual output, 𝑌𝑖 is average
ber of parameters. In other words, it needs to measure the actual output, and 𝑁 is number of samples. The process
discrepancy between neural network and an actual system. In of designing the network plays an important role in the
order to tackle this, most researches have mainly focused on performance of network.
improving the performance. There is no way to find hidden
neuron in neural network without trying and testing during 4. Proposed Architecture
the training and computing the generalization error. The
hidden output connection weights becomes small as number There exists various heuristics in the literature; amalgamating
of hidden neurons become large, and also the tradeoff in the knowledge gained from previous experiments where a
stability between input and hidden output connection exists. near optimal topology might exist [33–35]. The objective is
A tradeoff is formed that if the 𝑁ℎ becomes too large, the to devise the criteria that estimate the number of hidden
output neurons becomes unstable, and if the number of hid- neurons as a function of input neurons (𝑛) and to develop
den neuron becomes too small, the hidden neuron becomes the model for wind speed prediction in renewable energy

You might also like