Professional Documents
Culture Documents
An Improved Brain Emotional Learning Algorithm For Accurate and Efficient Data Analysis
An Improved Brain Emotional Learning Algorithm For Accurate and Efficient Data Analysis
1. School of Information Science and Engineering, Central South University, Changsha 410083, China;
2. Electrical and Information Engineering College, Hunan University of Arts and Science,
Changde 415000, China
© Central South University Press and Springer-Verlag GmbH Germany, part of Springer Nature 2018
Abstract: To overcome the deficiencies of high computational complexity and low convergence speed in traditional
neural networks, a novel bio-inspired machine learning algorithm named brain emotional learning (BEL) is introduced.
BEL mimics the emotional learning mechanism in brain which has the superior features of fast learning and quick
reacting. To further improve the performance of BEL in data analysis, genetic algorithm (GA) is adopted for optimally
tuning the weights and biases of amygdala and orbitofrontal cortex in BEL neural network. The integrated algorithm
named GA-BEL combines the advantages of the fast learning of BEL, and the global optimum solution of GA.
GA-BEL has been tested on a real-world chaotic time series of geomagnetic activity index for prediction, eight
benchmark datasets of university California at Irvine (UCI) and a functional magnetic resonance imaging (fMRI)
dataset for classifications. The comparisons of experimental results have shown that the proposed GA-BEL algorithm is
more accurate than the original BEL in prediction, and more effective when dealing with large-scale classification
problems. Further, it outperforms most other traditional algorithms in terms of accuracy and execution speed in both
prediction and classification applications.
Cite this article as: MEI Ying, TAN Guan-zheng. An improved brain emotional learning algorithm for accurate and
efficient data analysis [J]. Journal of Central South University, 2018, 25(5): 1084–1098. DOI: https://doi.org/10.1007/
s11771-018-3808-6.
recent years, and it has attracted an increasing bee colony algorithm (ABC) [15], differential
interest around the world. Several bio-inspired brain evolution (DE) [16] and genetic algorithm (GA)
emotional learning (BEL) models have been [17]. However, the standard ABC suffers from a
proposed and successfully applied in intelligent slow convergence speed for its solution search
engineering applications [4, 5]. These BEL models equation [15]. Although DE is relatively simple and
are based on a computational model called easily to converge, it may easily fall into local
amygdala-orbitofrontal model [6], which was minimum in the searching process [18]. GA tends to
inspired by LEDOUX’s anatomical findings [7] of efficiently explore various regions of the decision
emotional learning mechanism in mammalian brain. space, and find an optimal solution with a high
BEL-based models mimic the high speed of probability [19]. It has been demonstrated that the
emotional learning in brain, which have the performance of neural network can be substantially
superior features of fast learning and quick reacting. improved by optimizing the weights with GA
Thus, they are also widely used in classification, [20–22] and the learning process of network is
prediction and control applications [8–10]. regarded as the process of searching for optimum in
In amygdala-orbitofrontal model [6], the the weight space. Therefore, the present work
reward signal plays an important role to adjust the intends to integrate GA with the novel BEL neural
weights of amygdala and orbitofrontal cortex in network to determine properly weights of the
emotional learning process, but it is not clearly network. The integrated algorithm named GA-BEL
defined so far. Many researchers proposed different takes advantage of the fast learning and low
versions of BEL models based on amygdala- computational complexity of BEL, as well as the
orbitofrontal model, as well as different reward global optimum solution of GA. Thus, GA-BEL is
signal determinations. LUCAS et al [11] proposed expected to achieve better performance than the
the BEL-based intelligent controller (BELBIC) original BEL in data analysis. The GA-BEL has
which had been successfully applied in intelligent been tested on a chaotic time series of geomagnetic
engineering applications. ABDI et al [12] applied a index for real-world prediction, eight datasets of
modified BEL model to predict short-term traffic benchmark university California at Irvine (UCI)
flow and defined the reward signal as the and a functional magnetic resonance imaging
multiplication of some related weights. CHEN et al (fMRI) dataset for classifications. The comparisons
[13] presented a BEL-based controller for a four- of experimental results indicate the superiority of
wheel drive robot. Although these BEL-based the proposed GA-BEL in terms of accuracy and
models achieved success in the applications, they execution speed.
were based on reinforcement learning to adjust the
weights of Amygdala and Orbitofrontal cortex. 2 Related works
These methods are model-sensitive and cannot be
generalized to other issues. LOTFI et al [14] 2.1 Anatomical foundation
proposed a novel BEL-based pattern recognizer The limbic system theory [23] is the neural
(BELPR), instead of employing reward-based basis for emotional brain studies. Figure 1(a) [14]
reinforcement learning, it employed activation shows the limbic system in the brain and its
functions and target values to update the weights of components, including the sensory cortex, thalamus,
BEL network in the learning phase. The BELPR amygdala, orbitofrontal cortex etc. There are two
can be learned by using pattern-target examples and main parts among these components. One is
it is model-free and time-saving. However, the amygdala, which is properly situated to reach the
pattern-target learning method reduced the emotional stimulus and plays a pivotal role in
precision of the process, and the performance of emotional learning process. Another is orbitofrontal
accuracy in data analysis needs to be further cortex, which assists amygdala to process emotional
improved. stimulus. LEDOUX [7] stated that emotional
In this study, we aim to optimize the BEL stimuli can reach amygdala by two different paths,
network and make it more accurate. Recently, as shown in Figure 1(b). One is long and precise
several evolutionary algorithms have been available path, coming from the sensory cortex,
to apply in optimization problems, such as artificial and another is short but imprecise path, coming
1086 J. Cent. South Univ. (2018) 25: 1084–1098
responsible for processing emotional stimulus.
In amygdala-orbitofrontal model, Si is sensory
input; Aj is the internal output of the amygdala; Oj is
the internal output of the orbitofrontal cortex. The
reward signal Rew is used to update weights of
amygdala and orbitofrontal cortex in emotional
learning, and the learning rules are expressed as
follows [6]:
vi ( Si max(0, Rew A j )) (1)
j
Motivated by LEDOUX’s anatomical findings where rj stands for the factor of the reinforcement
[7] in the mammalian brain, MORÉN and agent and wj represents the related weight.
BALKENIUS [6] firstly proposed the amygdala- CHEN et al [13] proposed a BEL controller for
orbitofrontal model in 2000 and the framework of a four-wheel drive robot and the reward signal Rew
amygdala-orbitofrontal model is shown in Figure 2. is defined as:
In amygdala-orbitofrontal model, amygdala and
orbitofrontal cortex are two crucial parts of de
Rew =r1e+ r2 edt r3 (4)
emotional learning and reacting. Amygdala receives dt
emotional stimuli from the Sensory cortex and where r1, r2, r3 stand for the factors of weights; e
Thalamus as well as the external reward signal, and represents the error.
it interacts with the orbitofrontal cortex and reacts Although these BEL models achieved success
to the emotional stimuli based on the reward signal. in the special applications, they applied reward-
The orbitofrontal cortex receives sensory input based reinforcement learning to adjust the weights
from the sensory cortex and evaluates the of amygdala and orbitofrontal cortex, and they were
amygdala’s response to prevent inappropriate model-sensitive and cannot be generalized to other
learning connections. They interact frequently to issues.
mimic the functionality of the emotional brain LOTFI et al [14] firstly proposed the model-
free version of BEL-based pattern recognizer,
which employed the target value (T) of input pattern
in the learning phase and the reward signal Rew is
defined as follows:
Rew = T (5)
Figure 2 Framework of amygdala-orbitofrontal model Thus, the supervised learning rules in BEL
J. Cent. South Univ. (2018) 25: 1084–1098 1087
model are described as follows: neural network is shown in Figure 3. Similar to the
amygdala-orbitofrontal model, it consists of four
v kj 1 (1 )v kj max(T k Eak , 0) Pjk ,
common components including thalamus, sensory
j 1, 2, , n 1 (6) cortex, orbitofrontal cortex and amygdala.
wkj 1 wkj ( E k T k ) Pjk , j 1, 2, , n (7) Amygdala and orbitofrontal cortex are the two main
parts, which are mainly responsible for emotional
where vkj and wkj represent the weights of amygdala learning and response.
and orbitofrontal cortex, respectively; Tk is the The model is presented as multi-input single-
target value associated with the kth input pattern Pk; output architecture, amygdala receives m input
Eka is the internal output of amygdala; Ek is the final patterns S=[S1, S2, …, Sm] from the Sensory cortex
output of the BEL model; γ is the decay rate in the and Ath from the thalamus. Ath is calculated by
amygdala learning rule; α and β are learning rates. Ath max( S1 , S2 , , Sm ) (8)
The model can be employed to learn the
pattern-target relationship of an application by As shown in Figure 3, vi is the amygdala
using BEL algorithm, but this method reduces the weight and ba is the bias of amygdala neuron. For
precision of the process. Thus, we employed GA to each sensory input, there is an amygdala node Ai to
optimally tune the weights of BEL neural network. receive. EA is the internal output of the amygdala
At the same time, we improved BEL neural network and it is calculated by
for prediction and classification applications.
Ai Si vi , i 1, 2, , m
(9)
3 Implementation Am 1 Ath vm 1
m 1 m
Finally, the final output is simply calculated by m represents the number of input features; n
E E A EO (13) represents the number of sample classes; each
amygdala-orbitofrontal cortex unit interacts
where E is the final output that represents the separately in the learning process.
correct response of amygdala. In this study, the weights and biases of
In the single-output BEL neural network, the amygdala and orbitofrontal cortex in BEL neural
number of input patterns determines the neurons network is optimized by GA. The data analysis with
number in the thalamus and sensory cortex units. GA-BEL can be divided into three parts: BEL
The BEL neural network can be learned by neural network structure determination, parameters
pattern-target examples and it is model free and can optimization and prediction or classification output
be utilized in prediction application. of BEL neural network.
3.2 Multiple-output BEL neural network for 3.3 Optimizing BEL neural network with GA
classification 1) Chromosome representation
The single-output BEL neural network can be For the advantage of high precision, real
extended to multiple-output network for encoding is adopted to acquire optimal results. One
classification. The number of sample classes real number string represents one chromosome,
determines the number of orbitofrontal cortex and which consists of connection weights and biases of
amygdala units. Thus, the extended BEL model can orbitofrontal cortex and amygdala. According to the
be applied in binary and multiple classifications. In structure of BEL neural network, each chromosome
the proposed m–n architecture as shown in Figure 4, is initialized as follows [9]:
Figure 5 Prediction results of Dst index with GA-BEL: Figure 6 Prediction results of Dst index with BEL:
(a) Mean squared error; (b) Linear correlation analysis of (a) Mean squared error; (b) Linear correlation analysis of
training samples; (c) Linear correlation analysis of training samples; (c) Linear correlation analysis of
testing samples testing samples
actual output of the GA-BEL model. The linear As illustrated in Figure 6(a), the best MSE is
relationship between target and output is also given obtained at epoch 2. After the process is in steady
on the left side. As illustrated in Figures 5(b) and state, Figure 6(b) illustrates the predicted versus
(c), the values of R are 0.95897 and 0.97128 in the desired output of the Dst index. As illustrated in
training and testing samples, respectively. It Figure 6(b), the BEL obtained R=0.94462 in testing
indicates that the linear correlation between the samples, which is much less than R=0.97128 of
output and the desired value is very good in the GA-BEL in testing samples.
testing samples. We also compared GA-BEL with the
For comparison, we apply the original BEL to traditional multilayer perceptron (MLP) [26].
test on the same Dst index dataset. Figure 6 Table 1 presents the detailed MSE and R obtained
presents the prediction results in steady states. from the GA-BEL, BEL and LM-BP in the testing
1092 J. Cent. South Univ. (2018) 25: 1084–1098
samples. It is obvious that fast training is the main information for each subject, including age, gender,
feature of BEL-based algorithms compared to verbal IQ, performance IQ, etc. There are several
LM-BP algorithm [26]. Although the original BEL imaging sites: NeuroImage (NI), New York
needs less epochs to reach the steady state, University Medical Center (NYU), Peking
GA-BEL achieves higher correlation and lower University (Peking), etc. We adopted the widely
mean square error, which implies that GA-BEL is used ADHD-200 dataset provided by Peking
more accurate than BEL in prediction. University and the description is shown in Table 3.
Table 1 Comparisons of GA-BEL, BEL and LM-BP on Table 3 ADHD-200 dataset description
Dst prediction Control subject ADHD subject
Total No.
Model Learning Epoch MSE R of image Female Male Average Female Male Average
Age Age
ENN BEL 2 0.2086 0.94462 194 46 68 11 10 70 12
ENN GA-BEL 100 0.0014 0.97128
MLP LM-BP 783 4.1075 0.92183
We preprocessed the ADHD-200 dataset using
SPM8 toolbox [29], and chose 125 samples (68
4.2 Case 2: Experimental results on classification
controls, 57 ADHD) for this study. After feature
4.2.1 Datasets description
extraction based on wavelet transform and
UCI [27] datasets are usually used to evaluate
reduction [30], the feature vectors are sent to the
the performance of the algorithm. In this study, we
chose eight benchmark datasets for testing, proposed GA-BEL for classification. By combining
including binary and multiclass datasets, which are all ADHD subtypes in one category, the
of relatively high or low dimensions, large or small classification on ADHD-200 dataset is posed as a
sizes, and the details are summarized in Table 2. two-class classification, that is, ADHD’s subjects
and normal controls.
Table 2 Datasets description 4.2.2 Performance evaluation
Attribute Dataset Sample Feature Class Classification performance can be evaluated
by the confusion matrix as described in Figure 7, in
Iris 150 4 3
Low dims, which measures such as accuracy, precision and
small size Breast
699 9 2
Cancer recall are commonly used to assess the performance
Low dims, Banana 5300 2 2 of bankruptcy classification systems. Among them,
large size SVMguide1 7089 4 2 the classification accuracy (row 3-column 3, blue
High dims, Heart 270 13 2 area) is the main evaluation index, which is
small size Wine 178 13 3 calculated as follows [30, 31]:
5 Conclusions
Table 6 Classification results obtained with proposed method and previous methods
Dataset Sample Study Algorithm Accuracy/% Time/s
Breast ZHANG et al [33] o-GSCCA 62.9 1.64×10−2
699
Cancer This study GA-BEL 97.5 7.51×10−3
LUO et al [34] L1-L2-ELM 98.0 9.72×10−1
Iris 150
This study GA-BEL 98.3 2.91×10−2
HUANG et al [35] ELM( Gaussian) 89.8 4.69×10−2
Banana 5300
This study GA-BEL 91.5 6.31×10−2
BAI et al [36] SELM Gaussian) 84.4 4.20×10−3
Heart 270
This study GA-BEL 88.3 2.93×10−3
LUO et al [34] L1-L2-ELM 98.3 1.89
Wine 178
This study GA-BEL 97.2 3.69×10−2
BAI et al [36] SELM (Gaussian) 90.1 2.41
Satimage 6435
This study GA-BEL 91.7 2.27
RIAZ et al [37] SVM 64.7 4.13
ADHD-200 125
This study GA-BEL 79.3 2.73×10−2
This study introduces emotional intelligence radiation necrosis based on texture analysis in MRI [J].
into artificial intelligence, which opens a lot of Journal of Magnetic Resonance Imaging, 2015, 42(5):
1362–1368.
research gates for the bio-inspired research. In
[2] YAMASHITA Y, WAKAHARA T. Affine-transformation and
future work, other optimization method can be 2D-projection invariant k-NN classification of handwritten
employed, e.g., particle swarm optimization (PSO), characters via a new matching measure [J]. Pattern
which employs different strategies and Recognition, 2016, 52(C): 459–470.
computational effort to find a solution. Moreover, it [3] SHI Tian, KONG Jian-yi, WANG Xing-dong, LIU Zhao,
is appropriate to combine GA with PSO to further ZHENG Guo. Improved Sobel algorithm for defect detection
improve the performance of the BEL network in the of rail surfaces with enhanced efficiency and accuracy [J].
Journal of Central South University, 2016, 23(11):
future investigation, and apply them in real-time
2867–2875.
applications.
[4] KHOOBAN M H, JAVIDAN R. A novel control strategy for
DVR: Optimal bi-objective structure emotional learning [J].
References International Journal of Electrical Power & Energy Systems,
2016, 83: 259–269.
[1] LARROZA A, MORATAL D, PAREDES-SÁNCHEZ A. [5] SHARMA M K, KUMAR A. Performance comparison of
Support vector machine classification of brain metastasis and brain emotional learning-based intelligent controller
J. Cent. South Univ. (2018) 25: 1084–1098 1097
(BELBIC) and PI controller for continually stirred tank algorithm for neural network ensemble [J]. Lecture Notes in
heater (CSTH) [J]. Lecture Notes in Electrical Engineering, Computer Science, 2004, 3173: 323–331.
2015, 335: 293–301. [21] WU Jian-shen, LONG Jin, LIU Ming-zhe. Evolving RBF
[6] MORÉN J, BALKENIUS C. A computational model of neural networks for rainfall prediction using hybrid particle
emotional learning in the Amygdala [C]// Proceedings of the swarm optimization and genetic algorithm [J].
6th International Conference on the Simulation of Adaptive Neurocomputing, 2015, 148(2): 136–142.
Behaviour. MIT Press, 2000: 115–124. [22] HOSSEINI Z, NAKHAE I M. Estimation of groundwater
[7] LEDOUX J E. Emotion circuits in the brain [J]. Annual level using a hybrid genetic algorithm-neural network [J].
Review of Neuroscience, 2000, 23: 155–184. Pollution, 2015, 1(1): 9–21.
[8] SHARAFI Y, SETAYESHI S, FALAHIAZAR A. An [23] LEDOUX J E. Emotion and the limbic system concept [J].
improved model of brain emotional learning algorithm based Concepts in Neuroscience, 1991, 2: 169–199.
on interval knowledge [J]. Journal of Mathematics and [24] SRINIVAS M, PATTANAIK L M. Genetic algorithms: A
Computer Science, 2015, 14: 42–53. survey [J]. Computer, 1994, 27(6): 17–27
[9] LOTFI E. Wind power forecasting using emotional neural [25] LOTFI E, AKBARZADEH-T M R. Adaptive brain
networks [C]// Proceedings of the IEEE International emotional decayed learning for online prediction of
Conference on Systems, Man and Cybernetics. San Diego, geomagnetic activity indices [J]. Neurocomputing, 2014,
USA: MIT Press, 2014: 311–316. 126(3): 188–196.
[10] SHARBAFI M A, LUCAS C, DANESHVAR R. Motion [26] HAGAN M T, DEMUTH H B, BEALE M. Neural network
control of omni-directional three-wheel robots by brain- design [M]. Beijing: China Machine Press, 2002: 357.
emotional-learning-based intelligent controller [J]. IEEE [27] UCI machine learning repository [EB/OL]. [2017–03–02].
Transactions on Systems Man & Cybernetics Part C, 2010, http://archive.ics. uci.edu/ml.
40(6): 630–638. [28] ADHD-200 database. [2017–03–02]. http://fcon_1000.projects.
[11] LUCAS C, DANIAL S, NIMA S. Introducing Belbic: Brain nitrc.org/indi/adhd200/.
emotional learning based intelligent controller [J]. Intelligent [29] SPM toolbox [2017–03–02]. http://www.fil.ion.ucl.ac.uk/
Automation & Soft Computing, 2004, 10(1): 11–21. spm/.
[12] ABDI J, MOSHIRI B, ABDULHAI B, SEDIGH A K. [30] TAN Ying, ZHANG Tao, TAN Rui, SHEN Xiao-tao, XIAO
Forecasting of short-term traffic-flow based on improved Jing-zhong. Classification based Wavelet Translate and SVM
neuro-fuzzy models via emotional temporal difference in the ADHD [J]. Journal of University of Electronic Science
learning algorithm [J]. Engineering Applications of Artificial and Technology of China, 2015, 44(5): 789–794.
Intelligence, 2012, 25(5): 1022–1042. [31] ZUO Wan-li, WANG Zhi-yan, LIU Tong, CHEN Hui-ting.
[13] CHEN Jian-ping, WANG Jian-bin, YANG Yi-min. Velocity Effective detection of Parkinson’s disease using an adaptive
compensation control for a four-wheel drive robot based on fuzzy k-nearest neighbor approach [J]. Biomedical Signal
brain emotional learning [J]. CAAI Transactions on Processing & Control, 2013, 8(4): 364–373.
Intelligent Systems, 2013, 8(4): 361–366. [32] LIBSVM: A library for support vector machines [EB/OL].
[14] LOTFI E, AKBARZADEH T M R. Brain emotional [2017–03-05]. http://www.csie.ntu.edu.tw/~cjlin/libsvm.
learning-based pattern recognizer [J]. Cybernetics & Systems, [33] ZHANG Zhao, ZHAO Ming-bo. Binary- and multi-class
2013, 44(5): 402–421. group sparse canonical correlation analysis for feature
[15] CUI Lai-zhong, LI Geng-hui , LIN Qiu-zhen , DU Zhi-hua, extraction and classification [J]. IEEE Transactions on
GAO Wei-feng, CHEN Jian-yong, LU Nan. A novel artificial Knowledge & Data Engineering, 2013, 25(10): 2192–2205.
bee colony algorithm with depth-first search framework and [34] LUO Xiong, CHANG Xiao-hui, BAN Xiao-juan. Regression
elite-guided search equation [J]. Information Science, 2016 and classification using extreme learning machine based on
(367, 368): 1012–1044. L1-norm and L2-norm [J]. Neurocomputing, 2016, 174:
[16] CUI Lai-zhong, LI Geng-hui, LIN Qiu-zhen, CHEN 179–186.
Jian-yong, LU Nan. Adaptive differential evolution [35] HUANG Guang-bin, ZHOU Hong-ming, DING Xiao-jian,
algorithm with novel mutation strategies in multiple ZHANG Rui. Extreme learning machine for regression and
sub-populations [J]. Computers & Operations Research, multiclass classification [J]. IEEE Transactions on Systems
2016, 67: 155–173. Man & Cybernetics: Part B, Cybernetics, 2012, 42(2):
[17] HOLLAND J H. Adaptation in Natural and Artificial 513–529.
Systems [M]. Cambridge, UK: MIT Press, 1992. [36] BAI Zuo, HUANG Guang-Bin, WANG Dan-wei. Sparse
[18] DAS S, ABRAHAM A, KONAR A. Automatic clustering extreme learning machine for classification [J]. IEEE
using an improved differential evolution algorithm [J]. IEEE Transactions on Cybernetics, 2014, 44(10): 1858–1870.
Transactions on Systems, Man, and Cybernetics-Part A: [37] RIAZ A, ALONSO E, SLABAUGH G. Phenotypic
Systems and Humans, 2008, 38(1): 218–237. integrated framework for classification of ADHD using
[19] COOK D F, RAGSDALE C T, MAJOR R L. Combining a fMRI [C]// Proceedings of the 13th International Conference
neural network with a genetic algorithm for process Image Analysis and Recognition. Springer International
parameter optimization [J]. Engineering Applications of Publishing, 2016: 217–225.
Artificial Intelligence, 2000, 13(4): 391–396. (Edited by YANG Hua)
[20] SHEN Z Q, KONG F S. Optimizing weights by genetic
1098 J. Cent. South Univ. (2018) 25: 1084–1098
中文导读
基于改进大脑情感学习算法的有效数据分类
摘要:提出了采用遗传算法优化大脑情感学习模型的方法。大脑情感学习(BEL)模型是一种计算模型,
由 Morén 等人于 2000 年根据神经生理学上的发现提出。该模型根据大脑中杏仁体和眶额皮质之间的
情感学习机制建立,不完全地模拟了情感刺激在大脑反射通路中的信息处理过程。大脑情感学习模型
具有结构简单、计算复杂度低、运算速度快的特点。为了进一步提高模型的精度,采用遗传算法优化
调整大脑情感学习模型的权值,构造具有强泛化能力的大脑情感学习数据分析模型,并用于数据预测
与数据分类两方面。在数据预测方面,采用典型的磁暴环电流指数 Dst 时间序列作为测试数据。实验
结果表明,从均方差 MSE 和线性相关性 R 指标来看,GA-BEL 算法的误差小、相关度高,说明该算法用
于预测的有效性。在分类方面,采用 8 个典型的 UCI 数据集和一个典型的头部磁共振图像数据集(fMRI)
作为测试数据。分类实验结果表明,GA-BEL 算法的分类正确率高,运算速度快于传统算法,说明该算
法用于分类的有效性。
关键词:预测;分类;大脑情感学习;遗传算法