Unsupervised Learning in Reservoir Computing For EEG-based Emotion Recognition
Unsupervised Learning in Reservoir Computing For EEG-based Emotion Recognition
Abstract—In real-world applications such as emotion recognition from recorded brain activity, data are captured from electrodes over
time. These signals constitute a multidimensional time series. In this paper, Echo State Network (ESN), a recurrent neural network with
a great success in time series prediction and classification, is optimized with different neural plasticity rules for classification of
emotions based on electroencephalogram (EEG) time series. Actually, the neural plasticity rules are a kind of unsupervised learning
arXiv:1811.07516v2 [cs.CV] 23 Nov 2018
adapted for the reservoir, i.e. the hidden layer of ESN. More specifically, an investigation of Oja’s rule, BCM rule and gaussian intrinsic
plasticity rule was carried out in the context of EEG-based emotion recognition. The study, also, includes a comparison of the offline
and online training of the ESN. When testing on the well-known affective benchmark ”DEAP dataset” which contains EEG signals from
32 subjects, we find that pretraining ESN with gaussian intrinsic plasticity enhanced the classification accuracy and outperformed the
results achieved with an ESN pretrained with synaptic plasticity. Four classification problems were conducted in which the system
complexity is increased and the discrimination is more challenging, i.e. inter-subject emotion discrimination. Our proposed method
achieves higher performance over the state of the art methods.
Index Terms—Emotion recognition, Electroencephalogram, Echo state network, synaptic plasticity, intrinsic plasticity.
1 I NTRODUCTION
TABLE 2
Previous works on DEAP dataset
Subject-
Study Year Input #Channels Classifier Affective states Performance (%)
Independent
[23] 2019 Yes DWT coefficients 1 MLP Happy/Sad 58.50
LAHA 82.77
[24] 2018 No Statistical features 32 k-NN
LVHV 82.76
10 89.81±0.46
14 92.24±0.33
LAHA
18 93.69±0.30
32 95.69±0.21
[25] 2018 Yes Entropy and energy features of 4s k-NN
10 89.54±0.81
14 92.28±0.62
LVHV
18 93.72±0.48
32 95.70±0.62
[20] 2017 Yes Entropy features 32 SVM Stress and Calm 81.31
8 71.99
LAHA
32 72.10
[26] 2017 No IMF features of 5s SVM
8 69.10
LVHV
32 70.41
LAHA 68.28
[27] 2017 Yes Raw EEG 32 ESN LVHV 71.03
8 emotions 68.79
LALV, HALV,
[21] 2017 Yes Differential entropy of 1s 32 GELM LAHV, and 69.67
HAHV
LAHA 85.65
[28] 2017 No Raw EEG segment 32 LSTM
LVHV 85.45
LAHA 64.30
PSD of 1s
LVHV 58.20
[29] 2015 No 32 SVM
LAHA 64.20
DBN output from 1s raw EEG
LVHV 58.40
LAHA 55.00±4.5
[30] 2014 Yes Raw EEG 32 HMM
LVHV 58.75±3.8
Bandpower features 10 64.90
32 62.90
LAHA
PSD features 10 63.00
32 63.40
[31] 2014 Yes SVM
Bandpower features 10 64.90
32 62.30
LVHV
PSD features 10 56.40
32 60.00
16 65.63
[22] 2013 Yes Fractal dimension features SVM 8 emotions
32 69.53
a model trained and tested with data from independent Wavelet Transform (DWT) [23] or Intrinsic Mode Functions
users will lead to an easy application to new subject, i. e. (IMF) [26]. Most of the work done on EEG-based ER have
no requirement to design a new model for the new subject. suffered from finding informative features from EEG data.
To achieve this merit, several works focused on the These findings have reshaped scientific understanding
feature extraction step aiming to find the most relevant of EEG signals and inspired following works to analyze
features for EEG-based ER. Power spectral features (PSD) them directly instead of performing the feature extraction
using Short-Time Fourier Transform as in [29] and [31] are step. For example, [30] classified the EEG preprocessed
considered the baseline features and the most used. Fractal signals using the Hidden Markov Model (HMM). Likewise,
Dimension (FD) [22], entropy and energy features [25] have feature learning was performed by feeding the raw channel
shown acceptable results as depicted in Table 2. Other works data to the Deep Belief Network (DBN) [29]. The new
tested features of decomposition techniques such as Discrete representation obtained from DBN is then fed to Support
4
(a) (b)
Fig. 2. Unsupervised learning of Reservoir layer using (a) Synaptic plasticity rule and offline mode and (b) Intrinsic plasticity rule and online mode.
wavelet function corresponds to the high frequency com- reservoir and the inner weights of reservoir are randomly
ponents of the original signal while the stretched version initialized and remain unchanged during training phase.
corresponds to the low frequency components; and the The simplicity of ESN turns out in the need of training only
translation variable provides analysis in the time domain. the weights of the readout layer which is a linear regression.
The decomposition of DWT leads to coefficients of high Consider a topology of I input neurons, R internal neurons
frequency which are ”detail” and coefficients of coarse and O output neurons. The first step in training ESN is
approximation of original signal in time domain which are to collect the matrix of activation states of each neuron in
”approximation”. After decomposition, power features from the reservoir. The activation equation at each time step is
all bands are extracted. Note that, power bands features expressed in (3):
are the most popular features in the context of EEG-based
x(t) = fres W in u(t) + W res x(t − 1)
emotion recognition. The Definition of EEG frequency bands (3)
differs slightly between studies. Finally, a feature vector is
where W in and W res are the weights of the input and the
formed through the concatenation of all power features from
reservoir layers respectively. fres is the non-linear activation
all channels of the same EEG trial. Thus, the size of the
function, usually, a sigmoid. The second step consists in
feature vector is equal to the number of the bands multiplied
calculating the output weights. Here, we distinguish two
by the number of channels. Fig. 1 illustrates the feature
modes: offline and online. Linear regression is calculated as
extraction process adapted in our methodology.
follows in (4) and it is considered as an offline mode:
−1
W out = Ytarget ∗ X T X (4)
3.2 ESN Model
ESNs are a type of recurrent neural networks proposed by The training in the online mode is ensured through the
Jaeger [38]. It is composed of input layer, a reservoir and an minimization between the target output and the produced
output layer often called readout layer as depicted in Fig. output. The presentation of the training samples is in the
2a. The recurrence in ESNs is handled through the recurrent sequential form.
connections between hidden units and the possible feedback Originally proposed for MLP model, the delta rule is a
connection from the output layer to the reservoir. Direct stochastic gradient descent method used for the update of
connection from input to output layer can be added. one layer of MLP neural network [41]. It is expressed as
The basic structure proposed for solving computational follows:
∆W = η y desired (t) − y(t) x(t)
intelligence problems was the FFNN. However, for specific (5)
problems which have dynamic and temporal nature, FFNNs
are not able to handle complex temporal machine learning W out = W out + ∆W (6)
problems. As a solution, recurrent connections are added
to the structure, giving birth to RNNs for handling several where η is the learning rate and t is the time step of the
problems such as SVM leaning process [39]. Training RNNs learning iterations, t = 1, 2, , T . x(t) is the vector of neuron
was done by extending the canonical backpropagation al- firing activation states of x at time step t. This mechanism
gorithm to be Back-Propagation-Through-Time (BPTT) [40]. computes the incremental adaptation of readout weights.
One of the major limitations of BBTT is its high computa- Then, the output of the ESN can be generated according
tional cost and its slow convergence. Hence, Jaeger et al. [38] to (7) using linear regression if fout is a linear function:
proposed the reservoir computing approach to enhance the y(t) = f out W out ∗ x(t)
(7)
learning process.
The reservoir is composed of sparsely connected neu- For a specific input sample, the neuron having the highest
rons with cycled connections. The weights from input to activation score wins, while the other neurons in the same
6
layer are inhibited. This mechanism is called winner-take- synapses. The threshold varies as a nonlinear function of
all. Assuming that a trial is composed of N channels, the the average output of the postsynaptic neuron, which is the
ESN is fed sequentially with these N signals one by one. main concept of the present BCM model providing stability
As a matter of fact, the ESN output is a channel label. Each properties.
successive N channels labels are combined through majority The BCM rule has several variants and here the one
voting and the class label receiving the most number of suggested in [44] is adopted as given by (9) and (10):
votes is regarded as the trial label.
∆Wkj (t) = yk (yk − θM )xj /θM (9)
X
3.3 Plasticity Rules for Unsupervised Learning of θM = E yk2 = pk yk2 (10)
Reservoir Layer
where θM is the modification threshold of the postsynaptic
The generation of fixed random reservoir makes the RNN
neuron, yk , and pk is the probability of choosing vector
training fast. Meanwhile, other studies in neuroscience have
of from the dataset, E [] is the temporal average, Wkj is
reported that the modification of the connection strength
the adjustment of the synaptic weight between postsynaptic
endows neural networks with a powerful learning ability.
neuron, yk and presynaptic neuron, xj , at time t.
This mechanism is called Synaptic plasticity illustrated in
Fig. 2a. Here we detail the most used existing ones, to know,
3.3.3 Intrinsic plasticity rule
the Oja’s rule and the BCM rule. Recent findings showed
that the neuron is able to change its intrinsic excitability to From a biological point of view, Triesch [45] stated that
fit the distribution of the input as shown in Fig. 2b. We also the biological neuron does not adapt its synapses, rather
explain the gaussian intrinsic plasticity mechanism. it adapts its intrinsic excitability. While traditional learning
algorithms update the weights of connections between neu-
3.3.1 Oja’s rule rons, the IP rule algorithm update the activation function of
the neuron. Particularly, Triesch derived the IP rule for fermi
Oja learning rule, proposed by Erkki Oja [42], is a model
activation function and for an exponential desired distribu-
of how neurons in the brain or in artificial neural networks
tion. In the same manner, Schrauwen et al. [46] extended IP
alter the strength of connections, or learn, over time. Oja’s
considering also hyperbolic tangent as activation function
rule is an extension of hebbian learning rule [43] proposed
with Gaussian desired distribution.
in Hebb book ”The Organization of Behaviour”. In classical
IP rule is local, that means it is applied on each single
hebbian learning rule the update scheme for weights may
neuron to maximize information about its input. From
result in very large weights when the number of iterations
information theory, entropy measure allows us to realize
is large.
the maximization of information. Equation (11) measures
Oja’s rule is based on normalized weights, the weights
the distance between the actual probability density of the
are normally normalized to unit length. This simple change
neurons output and the targeted probability density using
results in a totally different but more general and stable
the Kullback-Leiber divergence metric.
weight update scheme compared to classical hebbian learn-
ing scheme. It can also be extended to non-linear neurons Z
p(x))
as well. It has been mathematically proven that when the DKL (p(x), pd (x))) = p(x)log (11)
pd (x)
input data is centered at the origin and when Oja’s update
rule converges it results in the neuron learning the first The Kullback-Leiber divergence can be developed into (12)
principal component of the training data, hence Oja’s rule for a Gaussian distribution with a mean µ and a standard
is important in feature based recognition systems [42]. deviation σ .
The Oja learning rule can be described as follows in (8): 1
2
DKL (p(x), pd (x))) = −H(x) + 2 E (x − µ)
∆Wkj (t) = ξyk (t) [xj (t) − yk (t)Wkj (t)]) (8) 2σ
1
+ log √ (12)
where Wkj is the change of the synaptic weight between the σ 2π
postsynaptic neuron, yk , and the presynaptic neuron,xj , at
time, t. ξ is the learning rate. Note that Eq. (8) is a modified A balance is achieved between the maximization of the
version of the anti-Hebbian rule by adding a forgetting actual entropy H and the minimization of the expected
factor to limit the growth of the synaptic weight to avoid entropy E . The update of the gain a and bias b is handled
the saturation of Wkj . using (13) and (14).
η
∆a = + ∆b W in u + W res x
3.3.2 BCM rule (13)
a
The BCM rule [44], named for Bienenstock, Cooper and µ x
Munro, follows the Hebbian learning principle, with a slid- ∆b = −η − 2 + 2 2σ 2 + 1 − x2 + µx (14)
ing threshold as a stabilizer function to control the synaptic
σ σ
alteration. Each neuron in the reservoir layer will be activated using
The main idea was that the sign of weight modification (15).
should be based on whether the postsynaptic response is x(t) = fres diag(a) W in u(t) + W res x(t − 1) + b (15)
above or below a threshold. Responses above the threshold
should lead to strengthening of the active synapses, re- Schrauwen et al. [46] showed that updating IP param-
sponses below the threshold lead to weakening of the active eters in a gaussian distribution with hyperbolic tangent
7
TABLE 3 TABLE 4
Wavelet Decomposition OF EEG channel signal Arousal Discrimination Results
Bandwith (Hz) Frequency Band Decomposition level System Input type Offline Online Hybrid
64-128 Hz Noise D1 Feature 59.77% 49.22% 60.01%
ESN-Oja rule
32-64 Hz Gamma D2 Signal 54.30% 58.98% 60.29%
16-32 Hz Beta D3 Feature 61.72% 54.14% 62.17%
ESN-BCM rule
8-16 Hz Alpha D4 Signal 56.25% 50.00% 59.34%
4-8 Hz Theta D5 Feature 61.21% 59.11% 62.39%
ESN-IP rule
1-4 Hz Delta A5 Signal 68.28% [27] 62.98% 69.23%
SVM with PSD Feature 63.40%
features [31]
activation function is similar to the update in an exponential
HMM [30] Signal 55.00±4.5%
distribution with fermi activation function. Hence, there is
no dependence between the chosen non-linearity function
and the targeted distribution. TABLE 5
Valence Discrimination Results
A very interesting unsupervised rule, the IP, is able to
make reservoir computing more robust in a fashion that System Input type Offline Online Hybrid
its internal dynamics can autonomously tune themselves
Feature 59.77% 52.73% 60.81%
independently of the randomly generated weights or the ESN-Oja rule
Signal 61.26% 54.92% 62.13%
scaling of input to the optimal dynamic regime for emotion
recognition. For reservoir enhancing purpose, we follow Feature 57.42% 46.88% 58.26%
ESN-BCM rule
previous works in which they used the IP rule [45], [46] Signal 56.25% 41.67% 59.31%
and [47], [48]. Feature 53.52% 55.86% 57.94%
ESN-IP rule
Signal 71.03% [27] 66.23% 71.25%
ACKNOWLEDGMENTS
The research leading to these results has received fund-
ing from the Ministry of Higher Education and Scientific
Research of Tunisia under the grant agreement number
LR11ES48.
R EFERENCES
[1] S. Poria, E. Cambria, R. Bajpai, and A. Hussain, “A review of affec-
tive computing: From unimodal analysis to multimodal fusion,”
Information Fusion, vol. 37, pp. 98–125, 2017.
[2] J. Tao and T. Tan, “Affective computing: A review,” in Interna-
Fig. 5. Impact of reservoir size on the performance. tional Conference on Affective computing and intelligent interaction.
Springer, 2005, pp. 981–995.
[3] R. Picard, Affective Computing, ser. Inteligencia artificial. The MIT
that the reservoir should satisfy the ESP. It relates asymp- Press, 1997. [Online]. Available: https://books.google.es/books?
totic properties of the excited reservoir dynamics to the id=N1qqQgAACAAJ
[4] P. Ekman, “Basic emotions in handbook of cognition and emotions
driving signal. (t. dalgleish and m. power, eds.),” John Wiley& Sons Ltd, 1999.
Intuitively, the ESP states that the reservoir will asymp- [5] J. A. Russell, “A circumplex model of affect.” Journal of personality
totically wash out any information from initial conditions. and social psychology, vol. 39, no. 6, p. 1161, 1980.
[6] A. Mehrabian, “Framework for a comprehensive description and
The ESP is granted for any input if this spectral radius is
measurement of emotional states.” Genetic, social, and general psy-
smaller than unity. In Fig. 4, it is shown that the best value chology monographs, 1995.
of spectral radius is 0.85 for our classification problems. In [7] ——, “Pleasure-arousal-dominance: A general framework for de-
order to evaluate the sensitivity of our proposed system, we scribing and measuring individual differences in temperament,”
Current Psychology, vol. 14, no. 4, pp. 261–292, 1996.
also varied the reservoir parameter from 500 to 5000 neurons [8] K. Sun, J. Yu, Y. Huang, and X. Hu, “An improved valence-
as depicted in 5. The choice was not arbitrary, rather we tried arousal emotion space for video affective content representation
to take into consideration the channel data length which is and recognition,” in Multimedia and Expo, 2009. ICME 2009. IEEE
8064. The role of the reservoir here is to compress the input International Conference on. IEEE, 2009, pp. 566–569.
[9] D. Mellouli, T. M. Hamdani, and A. M. Alimi, “Deep neural net-
data and to provide a new representation for classification work with rbf and sparse auto-encoders for numeral recognition,”
purpose. in Intelligent Systems Design and Applications (ISDA), 2015 15th
The smaller size 500 does not result in a good repre- International Conference on. IEEE, 2015, pp. 468–472.
[10] N. Chouikhi, B. Ammar, N. Rokbani, A. M. Alimi, and A. Abra-
sentation of channel data leading to low accuracy results. ham, “A hybrid approach based on particle swarm optimization
For a medium size 1500, the performance is acceptable for echo state network initialization,” in Systems, Man, and Cyber-
for both emotional states and stress/calm discrimination. netics (SMC), 2015 IEEE International Conference on. IEEE, 2015,
But, for arousal and valence discrimination the performance pp. 2896–2901.
[11] N. Chouikhi, R. Fdhila, B. Ammar, N. Rokbani, and A. M. Alimi,
is very low. Using 2500 as a reservoir size enhances the “Single-and multi-objective particle swarm optimization of reser-
performance for classification problems. That means the voir structure in echo state network,” in Neural Networks (IJCNN),
reservoir is able to make a new representation of the raw 2016 International Joint Conference on. IEEE, 2016, pp. 440–447.
[12] N. Chouikhi, B. Ammar, N. Rokbani, and A. M. Alimi, “Pso-
EEG signal making the new feature vector more discrimi-
based analysis of echo state network parameters for time series
native. When increasing the reservoir size to 5000 neurons, forecasting,” Applied Soft Computing, vol. 55, pp. 211–225, 2017.
the performance is slightly lower. This is mainly due to the [13] N. Slama, W. Elloumi, and A. M. Alimi, “Distributed recurrent
increased complexity of the system. It can also include that neural network learning via metropolis-weights consensus,” in
International Conference on Neural Information Processing. Springer,
the reservoir is stable with 2500 neurons. 2017, pp. 108–119.
[14] B. Ammar, N. Chouikhi, A. M. Alimi, F. Chérif, N. Rezzoug, and
P. Gorce, “Learning to walk using a recurrent neural network with
5 C ONCLUSION time delay,” in International Conference on Artificial Neural Networks.
Springer, 2013, pp. 511–518.
This paper presented the prominent computational models [15] M. Soleymani, J. Lichtenauer, T. Pun, and M. Pantic, “A multi-
of plasticity and their current applicability to empirical im- modal database for affect recognition and implicit tagging,” IEEE
provements as neural network adaptation mechanisms. The Transactions on Affective Computing, vol. 3, no. 1, pp. 42–55, 2012.
[16] S. Koelstra, C. Muhl, M. Soleymani, J.-S. Lee, A. Yazdani,
presented forms of plasticity are applied to randomly con- T. Ebrahimi, T. Pun, A. Nijholt, and I. Patras, “Deap: A database for
nected recurrent reservoirs to learn the structural informa- emotion analysis; using physiological signals,” IEEE Transactions
tion in EEG signals, achieve sparsity in neural connectivity on Affective Computing, vol. 3, no. 1, pp. 18–31, 2012.
and enhance learning performance of emotion recognition. [17] W.-L. Zheng and B.-L. Lu, “Investigating critical frequency bands
and channels for eeg-based emotion recognition with deep neural
We advocate the use of intrinsic plasticity for reservoir networks,” IEEE Transactions on Autonomous Mental Development,
pretraining since it achieves best results in comparison with vol. 7, no. 3, pp. 162–175, 2015.
10
[18] S. Katsigiannis and N. Ramzan, “Dreamer: a database for emotion [40] P. J. Werbos, “Backpropagation through time: what it does and
recognition through eeg and ecg signals from wireless low-cost how to do it,” Proceedings of the IEEE, vol. 78, no. 10, pp. 1550–
off-the-shelf devices,” IEEE journal of biomedical and health informat- 1560, 1990.
ics, vol. 22, no. 1, pp. 98–107, 2018. [41] J. C. Pemberton and J. J. Vidal, “When is the generalized delta
[19] H. Becker, J. Fleureau, P. Guillotel, F. Wendling, I. Merlet, and rule a learning rule? a physical analogy,” in IEEE International
L. Albera, “Emotion recognition based on high-resolution eeg Conference on Neural Network, vol. 1, 1988, pp. 309–315.
recordings and reconstructed brain sources,” IEEE Transactions on [42] E. Oja, “Simplified neuron model as a principal component ana-
Affective Computing, no. 1, pp. 1–1, 2017. lyzer,” Journal of mathematical biology, vol. 15, no. 3, pp. 267–273,
[20] B. Garcı́a-Martı́nez, A. Martı́nez-Rodrigo, R. Zangróniz, J. M. Pas- 1982.
tor, and R. Alcaraz, “Symbolic analysis of brain dynamics detects [43] D. Hebb, J. Martinez, and S. Glickman, “The organization of
negative stress,” Entropy, vol. 19, no. 5, p. 196, 2017. behavior-a neuropsychological theory-hebb, do,” 1994.
[21] W.-L. Zheng, J.-Y. Zhu, and B.-L. Lu, “Identifying stable patterns [44] G. Castellani, N. Intrator, H. Shouval, and L. Cooper, “Solutions of
over time for emotion recognition from eeg,” IEEE Transactions on the bcm learning rule in a network of lateral interacting nonlinear
Affective Computing, 2017. neurons,” Network: Computation in Neural Systems, vol. 10, no. 2,
pp. 111–121, 1999.
[22] Y. Liu and O. Sourina, “Eeg databases for emotion recognition,” [45] J. Triesch, “A gradient rule for the plasticity of a neurons in-
in Cyberworlds (CW), 2013 International Conference on. IEEE, 2013, trinsic excitability,” in International Conference on Artificial Neural
pp. 302–309. Networks. Springer, 2005, pp. 65–70.
[23] P. Pandey and K. Seeja, “Emotional state recognition with eeg [46] B. Schrauwen, M. Wardermann, D. Verstraeten, J. J. Steil, and
signals using subject independent approach,” in Data Science and D. Stroobandt, “Improving reservoirs using intrinsic plasticity,”
Big Data Analytics. Springer, 2019, pp. 117–124. Neurocomputing, vol. 71, no. 7-9, pp. 1159–1171, 2008.
[24] L. Piho and T. Tjahjadi, “A mutual information based adaptive [47] J. J. Steil, “Online reservoir adaptation by intrinsic plasticity for
windowing of informative eeg for emotion recognition,” IEEE backpropagation–decorrelation and echo state learning,” Neural
Transactions on Affective Computing, 2018. Networks, vol. 20, no. 3, pp. 353–364, 2007.
[25] M. Li, H. Xu, X. Liu, and S. Lu, “Emotion recognition from [48] M. Wardermann and J. J. Steil, “Intrinsic plasticity for reservoir
multichannel eeg signals using k-nearest neighbor classification,” learning algorithms.” in ESANN, 2007, pp. 513–518.
Technology and Health Care, no. Preprint, pp. 1–11, 2018.
[26] N. Zhuang, Y. Zeng, L. Tong, C. Zhang, H. Zhang, and B. Yan,
“Emotion recognition from eeg signals using multidimensional
information in emd domain,” BioMed research international, vol.
2017, 2017.
[27] R. Fourati, B. Ammar, C. Aouiti, J. Sanchez-Medina, and A. M.
Alimi, “Optimized echo state network with intrinsic plasticity
for eeg-based emotion recognition,” in International Conference on
Neural Information Processing. Springer, 2017, pp. 718–727.
[28] S. Alhagry, A. A. Fahmy, and R. A. El-Khoribi, “Emotion recogni- Rahma Fourati (IEEE Graduated Student Mem-
tion based on eeg using lstm recurrent neural network,” Emotion, ber 15). She was born in Sfax, she is a PhD
vol. 8, no. 10, 2017. student in Computer Systems Engineering at
the National Engineering School of Sfax (ENIS),
[29] X. Li, P. Zhang, D. Song, G. Yu, Y. Hou, and B. Hu, “Eeg based
since October 2015. She received the M.S. de-
emotion identification using unsupervised deep feature learning,”
gree in 2011 from the Faculty of Economic Sci-
2015.
ences and Management of Sfax (FSEGS). She
[30] C. A. Torres-Valencia, H. F. Garcia-Arias, M. A. A. Lopez, and is currently a member of the REsearch Group
A. A. Orozco-Gutiérrez, “Comparative analysis of physiological in Intelligent Machines (REGIM).Her research
signals and electroencephalogram (eeg) for multimodal emotion interests include Recurrent Neural Networks, Af-
recognition using generative models,” in Image, Signal Processing fective computing, EEG signals analysis, Sup-
and Artificial Vision (STSIVA), 2014 XIX Symposium on. IEEE, 2014, port Vector Machines, Simulink Modeling.
pp. 1–5.
[31] I. Wichakam and P. Vateekul, “An evaluation of feature extraction
in eeg-based emotion prediction with support vector machines,”
in Computer science and software engineering (JCSSE), 2014 11th
international joint conference on. IEEE, 2014, pp. 106–110.
[32] L. Bozhkov, P. Koprinkova-Hristova, and P. Georgieva, “Learning
to decode human emotions with echo state networks,” Neural
Networks, vol. 78, pp. 112–119, 2016.
[33] ——, “Reservoir computing for emotion valence discrimination
from eeg signals,” Neurocomputing, vol. 231, pp. 28–40, 2017. Boudour Ammar (IEEE Student Member08,
[34] F. Ren, Y. Dong, and W. Wang, “Emotion recognition based on Member13, Senior Member’17) was born in
physiological signals using brain asymmetry index and echo state Sfax, Tunisia. She graduated in computer sci-
network,” Neural Computing and Applications, pp. 1–11, 2018. ence in 2005. She received the Master degree
in automatic and industrial computing from the
[35] M.-H. Yusoff, J. Chrol-Cannon, and Y. Jin, “Modeling neural
National School of Engineers of Sfax, University
plasticity in echo state networks for classification and regression,”
of Sfax in 2006. She obtained a PhD degree
Information Sciences, vol. 364, pp. 184–196, 2016.
in recurrent neural network learning model for
[36] C. B. Chaabane, D. Mellouli, T. M. Hamdani, A. M. Alimi, and a biped walking simulator with the Research
A. Abraham, “Wavelet convolutional neural networks for hand- Group on Intelligent Machines (REGIM), Univer-
written digits recognition,” in International Conference on Health sity of Sfax, since February 2014. She is cur-
Information Science. Springer, 2017, pp. 305–310. rently an assistant professor with the Department of Computer En-
[37] P. S. Addison, The illustrated wavelet transform handbook: introductory gineering and Applied Mathematics at National Engineering School
theory and applications in science, engineering, medicine and finance. of Sfax (ENIS). Her research interests include iBrain (Artificial neural
CRC press, 2017. networks, Machine learning, Recurrent neural Network) and i-health
[38] H. Jaeger, Tutorial on training recurrent neural networks, covering (Autonomous Robots, Intelligent Control, Embedded Systems, medical
BPPT, RTRL, EKF and the” echo state network” approach. GMD- applications, EEG, ECG).
Forschungszentrum Informationstechnik Bonn, 2002, vol. 5.
[39] R. Fourati, C. Aouiti, and A. M. Alimi, “Improved recurrent neu-
ral network architecture for svm learning,” in Intelligent Systems
Design and Applications (ISDA), 2015 15th International Conference
on. IEEE, 2015, pp. 178–182.
11