Professional Documents
Culture Documents
A Separated Feature Learning based DBN Structure for
A Separated Feature Learning based DBN Structure for
A Separated Feature Learning based DBN Structure for
TABLE I. ALIGNMENT OF TARGET POSITIONS Contrastive divergence (CD) proposed by Hinton is used
in learning of RBM [13]. First, a training vector is input to the
Position Frequency visible layer. Then using (5) to compute the corresponding
Left 8.57 Hz states of the hidden units. Finally, a “reconstruction" of visible
Right 10 Hz layer is given by sampling with the probability computed by
Up 12 Hz (6). The learning rule is:
Down 15 Hz
∆𝑤𝑖𝑗 = 𝜖[(𝑣𝑖 ℎ𝑗 )𝑑𝑎𝑡𝑎 − (𝑣𝑖 ℎ𝑗 )𝑟𝑒𝑐𝑜𝑛 ]
∆𝑎𝑖 = 𝜖[(𝑣𝑖 )𝑑𝑎𝑡𝑎 − (𝑣𝑖 )𝑟𝑒𝑐𝑜𝑛 ]
B. Using RBM to Initialize the DBN
RBM is a probabilistic graphical model, and stochastic ∆𝑏𝑗 = 𝜖[(ℎ𝑗 )𝑑𝑎𝑡𝑎 − (ℎ𝑗 )𝑟𝑒𝑐𝑜𝑛 ]
neural network can be used to interpret RBM. It is very good
at modeling high dimensionality data. The structure of RBM is It has been proved by Hinton that using pretrained RBM to
shown in Fig. 2. initialize ANNs can avoid local minimum effectively [14]. We
used mini-batch gradient descent to pretrain the RBM in this
study, and the pretraining is greedy and unsupervised.
C. The Novel Structure of DBN
A novel structure of DBN was designed by a plurality of
RBM. The structure of the DBN in this study is shown in Fig.
3. In this novel structure, the local feature of data is extracted
from each channel separately. Then in the next layer, we
3357
combined the three local features in a whole layer. Finally, the Given K stimulus frequencies f1, … , fK, a set of standard
EEG data were fitted to a five-layer DBN. Every layer is a stimulus signals Yi, composed of cosinusoids and sinusoids at
RBM except the output classification layer, which is a softmax each stimulus frequencies and their harmonics, can be
layer. And they were pretrained one by one before the back constructed as:
propagation (BP) to avoid local minimum.
cos(2𝜋 · 𝑓𝑖 · 𝑡)
output sin(2𝜋 · 𝑓𝑖 · 𝑡)
·
1 𝑆
Softmax for classification(4 units) 𝑌𝑖 = · ,𝑡 = ,…,
𝐹𝑠 𝐹𝑠
·
cos(2𝜋 · 𝐻𝑓𝑖 · 𝑡)
More abstract feature of the 3 channels(1000 units) ( sin(2𝜋 · 𝐻𝑓𝑖 · 𝑡))
where Fs is the sampling rate, H is the number of
harmonics, and S is the number of sample points, which is also
Distributed feature of the 3 channels(2000 units) the time window of sample points of the SSMVEP signals X,
recorded from C channels.
3358
In Fig. 4, accuracies assessed by DBN method multiple models and Dropout, on the generalization of the
successfully exceeded 70% accuracy level even under a very DBN network to improve the detection accuracy. And
short response time (i.e., 2 seconds adopted in our study). Gaussian unit RBM may be another alternative way to more
What’s more, the performance obtained by DBN exceeded practically model the BCI signals.
CCA method for every subject.
One-way analysis of variance (ANOVA) test was used to ACKNOWLEDGMENT
reveal the decrement of the accuracies from DBN to CCA. We want to thank the subjects for participating in these
The results are shown in Table III. experiments and anonymous reviewers for their helpful
comments.
TABLE III. ACCURACY DECREMENT FROM DBN TO CCA
DBN CCA REFERENCES
Sub Decre
Mean Mean F p
ject Accuracy Accuracy ment [1] M Xu, H Qi and B Wan, “A Hybrid BCI Speller Paradigm Combining
P300 Potential And The SSVEP Blocking Feature”, Journal of Neural
S1 70.92% 52.50% 18% 103.3 8.9788×10-10
Engineering, 2013, 10(2): 026001.
S2 72.50% 33.00% 40% 913.3 2.0732×10-19 [2] Y. Lecun, Y. Bengio and G. E. Hinton, “Deep Learning”, Nature 521,
S3 75.50% 57.00% 17% 167.9 8.9525×10-12 436-444(28 May 2015).
[3] Xiu An, Deping Kuang, Xiaojiao Guo, “A Deep Learning Method for
S4 73.54% 57.50% 18% 150.2 2.6580×10-11 Classification of EEG Data Based on Motor Imagery”, ICIC 2014,
S5 74.45% 53.50% 20% 193.8 9.4522×10-12 LNBI 8590, pp. 203–210, 2014.
S6 71.45% 52.55% 18% 114.8 3.0486×10-9 [4] Na Lu, Tengfei Li, Xiaodong Ren, “A Deep Learning Scheme for
Motor Imagery Classification based on Restricted Boltzmann
S7 71.95% 52.95% 19% 216.7 1.7665×10-11 Machines”, IEEE Transactions on Neural Systems and Rehabilitation
Engineering(17 August 2016), Volume: PP, Issue: 99 | DOI:
The result in Table III indicate that the DBN method 10.1109/TNSRE.2016.2601240.
proposed in our study largely prevailed the traditional CCA [5] Teng Ma, Hui Li, Hao. Yang, “The extraction of motion-onset VEP
method in correct detection accuracy. And across subjects, BCI features based on deep learning and compressed sensing”, Journal
of Neuroscience Methods, Volume 275, 1 January 2017, Pages 80–92.
grand-averaged accuracies significantly declined 22% (F =
[6] Miku Yanagimoto, Chika Sugimoto, “Recognition of persisting
416.4, p = 6.8791×10-46) from DBN method (72.94% ± 3.83) emotional valence from EEG using convolutional neural networks”, in
to CCA method (51.18% ± 8.67). proceedings of 2016 IEEE International Workshop on Computational
Intelligence and Applications (IWCIA), November 2016.
B. Analysis of the inter-subject variability [7] Junhua Li, Zbigniew Struzik, Liqing Zhang, “Feature learning from
incomplete EEG with denoising autoencoder”, Neuro-computing,
The statistical variability of SSVEP responses is very Volume 165, 1 October 2015, Pages 23–31.
critical for the applicability of BCI applications. The [8] Tabar YR, Halici U, “A novel deep learning approach for classification
inter-subject variability, which was assessed by one-way of EEG motor imagery signals.”, Journal of Neural Engineering, 2017
ANOVA with Bonferroni-corrected multiple comparisons, Feb, 14(1):016003.
revealed that for the accuracies assessed by CCA method, [9] Yousef Rezaei Tabar and Ugur Halici, “A novel deep learning
approach for classification of EEG motor imagery signals”, Journal of
there were significant differences between individual subjects Neural Engineering, 30 November 2016, Volume 14, Number 1.
(F = 60.93, p = 2.3634×10-24), where subject S2 exhibited [10] Xie J, Xu G, Wang J, Zhang F, Zhang Y (2012), “Steady-State Motion
significant much lower accuracies while subject S3 and S4 Visual Evoked Potentials Produced by Oscillating Newton’s Rings:
showed significant higher accuracies, inferring that there is a Implications for Brain-Computer Interfaces”. PLoS ONE 7(6): e39707.
[11] M. Welling, M. Rosen-Zvi, and G. E. Hinton, (2005). “Exponential
large variability between subjects. This may be due to the Family Harmoniums with An Application to Information Retrieval”,
large response variability across different subjects. But for the Advances in Neural Information Processing Systems, pages 1481-1488,
accuracies assessed by DBN method proposed in our study, Cambridge, MA. MIT Press.
there is no significant accuracy differences (F = 2.19, p > [12] FB Vialatte, M Maurice and J Dauwels, “Steady-state visually evoked
0.05; one-way ANOVA with Bonferroni-corrected multiple potentials: Focus on essential paradigms and future perspectives”,
Progress in Neuralbiology, 2010, 90(4): 418-438.
comparisons) among different subjects, which reflected a [13] Hinton, G. E. (2002). “Training products of experts by minimizing
facilitation in the alleviation of large inter-subject variability. contrastive divergence”, Neural Computation, 14(8):1711-1800.
[14] G. E. Hinton and R. R. Salakhutdinov, “Reducing The Dimensionality
IV. CONCLUSION of Data with Neural Networks”, Science 28 July 2006: vol. 313, Issue
5786, pp.504-507.
In this study, a separated feature learning based DBN
structure was proposed for the classification of SSMVEP
signals. Results indicated that this method is more robust and
can achieve higher SSMVEP discriminative accuracies and
lower inter-subject variability in short response time when
compared to traditional CCA method, which proved the
capability of the proposed DBN structure in the application of
SSVEP detection. But we can also find that the achieved
accuracy is still not high enough and also did not exceed the
common accuracy level of 80%. Further research work will
be focused on a variety of promising ways, such as mixture of
3359