Online Learning For Wearable EEG-Based Emotion Cla
Online Learning For Wearable EEG-Based Emotion Cla
Article
Online Learning for Wearable EEG-Based Emotion Classification
Sidratul Moontaha *,† , Franziska Elisabeth Friederike Schumann † and Bert Arnrich *
Abstract: Giving emotional intelligence to machines can facilitate the early detection and prediction
of mental diseases and symptoms. Electroencephalography (EEG)-based emotion recognition is
widely applied because it measures electrical correlates directly from the brain rather than indirect
measurement of other physiological responses initiated by the brain. Therefore, we used non-invasive
and portable EEG sensors to develop a real-time emotion classification pipeline. The pipeline trains
different binary classifiers for Valence and Arousal dimensions from an incoming EEG data stream
achieving a 23.9% (Arousal) and 25.8% (Valence) higher F1-Score on the state-of-art AMIGOS dataset
than previous work. Afterward, the pipeline was applied to the curated dataset from 15 participants
using two consumer-grade EEG devices while watching 16 short emotional videos in a controlled
environment. Mean F1-Scores of 87% (Arousal) and 82% (Valence) were achieved for an immediate
label setting. Additionally, the pipeline proved to be fast enough to achieve predictions in real-time
in a live scenario with delayed labels while continuously being updated. The significant discrepancy
from the readily available labels on the classification scores leads to future work to include more data.
Thereafter, the pipeline is ready to be used for real-time applications of emotion classification.
Keywords: online learning; real-time; emotion classification; AMIGOS dataset; wearable EEG (muse
and neurosity crown); psychopy experiments
improve human–computer interactions (HCI) and enhance their ability to make decisions
by adapting their reactions accordingly.
Since emotional reactions are seemingly subjective experiences, neurophysiological
biomarkers, such as heart rate, respiration, or brain activity [10,11], are inevitable. Ad-
ditionally, for continuous monitoring of affective states and thus detecting or predicting
stress-related events reliably, low-cost, consumer-grade devices rather than expensive and
immobile hospital equipment would be more meaningful [12]. This is an important area of
interest in cognitive science and affective computing, with use cases varying from design-
ing brain–computer interfaces [13,14] to improving healthcare for patients suffering from
neurological disorders [15,16]. Among these, electroencephalography (EEG) has proven to
be an accurate and reliable modality without needing external annotation [17,18]. Since
clinical EEG is the gold standard for epileptic seizure detection [19], utilizing EEG-based
emotion classification in detection systems could potentially predict seizures by knowing
the affective states. Moreover, with recent advancements in wearable technology, consumer-
grade EEG devices have become more accessible and reliable, opening possibilities for
countless real-life applications. Wearable EEG devices such as the Emotiv EPOC headset
or the Muse S headband have become quite popular tools in emotion recognition [20–22].
The Muse S headband has also been used for event-related potential (ERP) research [12]
and for the challenge of affect recognition in particular. More specifically, Muse S has
already been used in experimental setups to obtain EEG data from which the mental state
(relaxed/concentrated/neutral) [13] and the emotional state (using the valence-arousal
space) [23] could be reliably inferred through the use of a properly trained classifier.
paper uses fewer electrodes and multiple algorithms and develops and publishes the data
collection framework.
Very recently, Bajada et al. [36] built an emotion classification pipeline incorporating
discrete wavelet transforms (DWT) features into a 3D convolutional neural network (3D
CNN) and support vector machine (SVM) classifiers from pre-recorded and pre-trained data
from the state-of-the-art DEAP [37] dataset. They proposed using the proposed algorithm
for real-time monitoring because it maintains a high accuracy of up to 92%, reducing from
32 channels to 5 channels. Li et al. [38] addressed the challenge of when a model can see
the data only once by leveraging cross-subject and cross-session data by implementing the
Fast Online Instance Transfer (FOIT) algorithm. They validated their methodology on two
state-of-the-art datasets: SEED [39] and SEED-IV [40]. However, these mentioned studies
need to apply live incoming data streams similar to our approach, which pose challenges
such as velocity, veracity, and concept drift. To the best of our knowledge, only Nandi
et al. [41] have employed online learning to classify emotions from an EEG data stream
from the DEAP dataset and proposed an application scenario in e-learning, but have yet to
report undertaking any such live experiments. They compared the performance of different
state-of-the-art online classifiers, such as adaptive random forest (ARF) [26] and Hoeffding
adaptive tree (HAT) [42], against their own real-time emotion classification system (RECS)
on the DEAP dataset.
Indeed, more research is needed on using online machine learning for emotion recognition.
Moreover, multi-modal labeled data for the classification of affective states have
been made freely available through annotated affective databases, such as DEAP [37],
DREAMER [43], ASCERTAIN [44], SAFE [33], and AMIGOS [22], which play a signifi-
cant role in further enhancing the research of this field. They include diverse data from
experimental setups using differing emotional stimuli such as music, videos, pictures, or
cognitive load tasks in an isolated or social setting. Such databases enable the development
and improvement of frameworks and model architectures with existing data of ensured
quality. However, none of these published datasets include the data collection framework
to be reused in curating the data from wearable EEG devices in live settings.
To validate the established pipeline in daily life and live setup, two different datasets
with individual experimental protocols, named Dataset II and Dataset III, were curated with
the description of participants, data acquisition, and experimental protocols mentioned in
the following sections.
2.2. Participants
For Dataset II, eleven participants were recruited (six females and five males) between
the ages of 25 and 42 (µ = 29.27, σ = 5.41 years). Data from two participants had to be
discarded for further analysis. Dataset III was made with data from four participants (one
female and three males) between the ages of 25 and 34 (µ = 28.5, σ = 3.5 years). People
who were pregnant, older than 65 years, and had taken part in one of the experiments were
excluded from participation. All participants had normal or corrected vision and reported
no history of neurological or mental illnesses or head injuries.
According to the international10–20 system [47], the channels on the Muse S Headband
correspond to AF7, AF8, TP9, and TP10 (see Figure 1b), with a reference electrode at
Fpz [12]. The channel locations of Neurosity Crown are C3, C4, CP3, CP4, F5, F6, PO3,
and PO4, with reference sensors located at T7 and T8, as shown in Figure 1c. Using the
Mind Monitor App (https://mind-monitor.com/ (accessed on 20 February 2023)), the
raw EEG data were streamed from Muse to a phone via Bluetooth. The app sends the
Sensors 2023, 23, 2387 6 of 23
data to a laptop via the open sound control (OSC) protocol and the python-osc library
(https://pypi.org/project/python-osc/ (accessed on 20 February 2023)) on the receiving
end. As the incoming data tuples from the Muse Monitor App did not include timestamps,
they were added by the pipeline upon arrival of each sample. Similarly, the Crown uses
the python-osc library to stream the raw EEG data to a laptop without enabling any pre-
processing settings. In contrast to the Muse Headband, the Crown includes a timestamp
when sending data.
Software: In this paper, the experiment was implemented using the software PsychoPy
(v 2021.2.3) [48] in a way that guided the participants through instructions, questionnaires,
and stimuli. The participants were allowed to go at their own pace by clicking on the “Next”
(“Weiter” in German) button, as shown in the screenshots of PsychoPy in Figure 3.
Table 1. The source movies of the videos used are listed per quadrant in the valence–arousal space.
Video IDs are stated in parentheses, sources marked with a † were taken from the MAHNOB-
HCI dataset [55]; all the others stem from DECAF [54]. In the category column, H, L, A, and
V stand for high, low, arousal, and valence, respectively. This table has been adapted from
Miranda-Correa et al. [22].
experiments, they were asked to answer the PANAS once at the beginning and once again
at the end. For the German version of the PANAS questionnaire, the translation of Breyer
and Bluemke [58] was used.
Affect Self-Assessment: The Affective Slider (AS) [59] was used in the experiment to
capture participants’ emotional self-assessment after presenting each stimulus, as depicted
in the screenshot in Figure 3b (https://github.com/albertobeta/AffectiveSlider (accessed
on 20 February 2023)). The AS is a digital self-reporting tool composed of two slider
controls for the quick assessment of pleasure and arousal. The two sliders show emoticons
at their ends to represent the extreme points of their respective scales, i.e., unhappy/happy
for pleasure (valence) and sleepy/wide awake for arousal. For the experiments, AS was
designed in a continuous normalized scale with a step size of 0.01 (i.e., a resolution of 100),
and the order of the two sliders was randomized each time.
Familiarity: The participants were asked to indicate their familiarity with each video on a
discrete 5-point-scale ranging from “Have never seen this video before” (1) to “Know the
video very well” (5). The PsychoPy slide with this question was always shown after the AS.
2.6. Dataset II
Briefing Session: In the beginning, each participant went through a pre-experimental brief-
ing where the experimenter explained the study procedure and informed the participant
of the experiment’s duration, i.e., two parts of approximately 20 min each with a small
intermediate break. The participant then received and read the data information sheet,
filled out the personal information sheet, and signed the consent to participate. Personal
information included age, nationality, biological sex, handedness (left- or right-handed),
education level, and neurological- or mental health-related problems. The documents and
the study platform (i.e., PsychoPy) were provided according to the participant’s choice
of study language between English and German. Afterward, the experimenter explained
the three scales mentioned in and allowed the participant to accustom to the PsychoPy
platform. This ensured the understanding of the different terms and scales used for the
experiment without having to interrupt the experiment afterwards. The participant could
refrain from participating at any moment during the experiment.
Data Collection: After the briefing, the experimenter put either the Muse headband or
the Crown on the participant by a random choice. Putting headphones over the device,
the participant was asked to refrain from strong movements, especially of the head. The
experimenter then checked the incoming EEG data and let the participant begin the ex-
periment. After greeting the participant with a welcome screen, a relaxation video was
shown to the participant (https://www.youtube.com/watch?v=S6jCd2hSVKA (accessed
on 20 February 2023)) for 3 min. They answered the PANAS questionnaire to rate their
current mood and closed eyes for half a minute to get a baseline measure of EEG data.
Afterwards, they were asked to initially rate the valence and arousal state with the AS.
Sensors 2023, 23, 2387 8 of 23
Following this, an instruction about watching eight short videos was provided. Each of
those was preceded by a video counter and followed by two questionnaires: the AS and the
familiarity. The order of the videos and the order of the two sliders of AS were randomized
over both parts of the experiments, fulfilling the condition that the labels of the videos are
balanced. The first part of the experiment ended after watching eight videos and answering
the corresponding questionnaire. The participant was allowed a short break after taking
the EEG device and the headphones off.
In the second part of the experiment, the experimenter put the device that had not been
used in the first part (Muse or Crown, respectively) and the headphones on the participant.
Subsequently, the experimenter started the second part of the experiment, again after
ensuring that the data collection was running smoothly. The participant followed the exact
same protocol: watching the relaxation video, closing their eyes, and watching eight more
movie scenes with the AS and familiarity questions in between. Lastly, they were asked
for a final mood self-assessment via a second PANAS questionnaire to capture differences
before and after the experiment. The experimental setup for curating Dataset II is depicted
in Figure 4.
In this experiment, one PC (with a 2.4 to 3.0 GHz Dual Intel Core i5-6300U and
12 GB RAM) was used to present the stimuli and store EEG data to be used only after the
experiment session.
~20 min
3 min 30 s
~20 min
3 min 30 s
Figure 4. Experimental setup for curating Dataset II. The participants watched a relaxation video
at the beginning and eight videos, two of each dimension category wearing one of the two devices.
Between the eight videos, they answered AS slider and familiarity with the video.
parallel. Figure 6 displays the initialized model in the bottom gray rectangle that performed
live emotion classification on the incoming EEG data stream. However, the prediction
results were only displayed to the experimenter to avoid additional bias. Since the objective
of this experiment was live online learning and classification, the data were coming in an
online stream; however, the data were also stored for later evaluation and reproducibility.
Neurosity ~20 min
Crown 3 min 30 s
Offline session
or
Muse
AS Familiarity AS
Relaxation Baseline eye
PANAS Video clips slider label
video closing
8 videos two per category
10 min break
Pre-processing
Model training
~20 min
3 min 30 s
Same device
Live session
from session 1
AS AS
Relaxation Baseline eye Familiarity
Video clips slider label PANAS
video closing
8 videos two per category
Delayed
labels
Figure 5. Experimental setup for curating Dataset III. In the first session, the participants watched a
relaxation video at the beginning and eight videos, two of each dimension category wearing one of
the two devices. Between the eight videos, they answered AS slider, familiarity with the video, and
had seen the actual AS label. In the second session, they watched the same set of videos while the
prediction was available to the experimenter before the delayed label arrived.
per channel
Immediate Labels
Part 1 Notch Filter Bandpass Filter Average 5 Bands PSD & Power Ratio True Label yi
50Hz 0.5Hz-45Hz Referencing total Power between Bands
2 update
Feature
Vector Fi
Stream Feature
Recorded Data
Preprocessing Shuffling Model Label ŷi
Extraction
1 predict BREA
Live
predict
Incoming Feature
Preprocessing Model Video Label Ŷj
Data Stream Extraction
update
Figure 6. Overview of pipeline steps for affect classification. The top gray rectangle shows the
pipeline steps employed in an immediate label setting with prerecorded data. For each extracted
feature vector the model (1) first classifies its label before (2) being updated with the true label for
that sample. In the live setting, the model is not updated after every prediction, as the true label of a
video only becomes available after the stimulus has ended. The timestamp of the video is matched to
the samples’ timestamps to find all samples that fall into the corresponding time frame and update
the model with their true labels (shown in dotted lines).
Sensors 2023, 23, 2387 10 of 23
In this experiment, the same PC from the previous experiment was again used to
present the stimuli. Additionally, the AS label was sent to a second machine (MacBook Pro
(2019) with a 2.8 GHz Quad-Core (Intel Core i7) and 16 GB). It also received EEG data, and
performed data pre-processing, online model training, and live emotion classification.
… …
Extracted … …
Fi-k Fi Fi+1 Fi+4 Fi+m
Feature Vectors
Figure 7. The incoming data stream is processed in tumbling windows (gray rectangles). One window
includes all samples xi , xi+1 , . . . arriving during a specified time period, e.g., 1 s. The pipeline extracts
one feature vector, Fi , per window. Windows during a stimulus (video) are marked in dark gray.
Participants rated each video with one label per affect dimension, Yj . All feature vectors extracted
from windows that fall into the time frame of a video (between tstart and tend of that video) receive a
label yi corresponding to the reported label, Yj , of that video. If possible, the windows are aligned
with the end of the stimulus; otherwise, all windows that lie completely inside a video’s time range
are considered.
Sensors 2023, 23, 2387 11 of 23
Figure 7 shows that a lot of samples in a row received the same label of the duration
of each video up to several minutes. Internal testing implied that training a model by
streaming the data resulted in classifiers that did not learn from features but only returned
the same class until seeing a different one. Therefore, the time windows were shuffled
among one another with the corresponding labels. Since shuffling needs all data and
labels to be present before feature extraction, it was not performed during live training
and classification.
Table 2. Number of channels and derived features for each device: Muse Headband: 64 features;
Neurosity Crown: 128 features; Emotiv EPOC: 224 features.
3.4. Labeling
During the live streaming of Dataset III, labels had to be mapped to their corresponding
sample. Therefore, the labels were sent in a stream of tuples: L1, A , L1, V , L2, A , L2, V , . . .,
where
L j, dimension = (Yj, dimension , tstart , tend ) (1)
A and V stand for arousal and valence, respectively, and Yj, dimension represents the AS label
given by the participant after each video of two timestamps, tstart and tend . One labeled
tuple L j, dimension per video and dimension was sent from the PC running the PsychoPy
experiment to the PC training the classification model. The included timestamps were
used to match the incoming ratings, Yj, dimension , as labels to the samples that the model had
classified before. This was performed in a way that all the samples that fell into the time
period between tstart and tend received the respective class label for each dimension. The
model could then be updated with these labels.
3.5. Evaluation
Online Learning and Progressive Validation: This paper aims at building a classification
pipeline from evolving data streams. Therefore, the static data from Dataset I and Dataset II
were streamed using a library for online learning: river [66]. Progressive validation, also called
test-then-train evaluation [67], was used for model evaluation in the supervised immediate
label setting, when the labels for all samples were present at processing time [68]. Figure 8a
shows the training process of an online classifier including progressive validation. Every
time the model sees a new sample xi , it first classifies this sample as the test step of the
test-then-train procedure. In the training process, the model will calculate the loss by
comparing the true label, yi , which might come from a different data source than the
samples. The updated model will go on to classify the next incoming sample, xi+1 , before
seeing its label, yi+1 , and, again, execute the training and performance metric updating step.
This continues as long as data are streamed to the model. In this way, all samples can be
used for training as well as for validation without corrupting the performance evaluation.
Sensors 2023, 23, 2387 12 of 23
(a) (b)
Figure 8. (a) Progressive validation incorporated into the basic flow of the training process (‘test-
then-train’) of an online classifier in an immediate label setting. (xi , yi ) represents an input feature
vector and its corresponding label. (b) Evaluation incorporated into the basic flow of the training
process of an online classifier when labels arrive delayed (i ≥ j).
In the experimental setup acquiring Dataset III, the labels were available after the pre-
diction, in contrast to the immediate labeling setting and progressive validation. Therefore,
a delayed progressive validation was performed with the delayed labels, which is mostly the
case for real-life scenarios. Figure 8b depicts the delayed progressive validation procedure,
where the samples are classified by the model until unseen labels are available. However,
the model can be updated as in the immediate label setting. Whenever new labels become
available, the performance metric is updated without any further calculations [69]. Once
the model has been updated with all available labels, the classification of further samples
continues with the new updated model. This can, of course, be implemented in parallel as
well. These steps continue as long as there are incoming data.
TP + TN
Accuracy = (2)
TP + TN + FP + FN
TP
F1-Score = 1
, (3)
TP + 2 (FP + FN)
Sensors 2023, 23, 2387 13 of 23
where TP and TN denote the number of true positives and true negatives classified by the
model, respectively. Accordingly, FP and FN stand for the number of false positives and
false negatives classified by the model, respectively. “The higher, the better”, can be said
for both accuracy and F1-Score, i.e., a perfect model has an accuracy of one (100%) and an
F1-Score of one.
To determine whether the performance differences between the different setups were
significant, two-sided t-tests with a significance level of α = 0.05 were conducted on the
respective dataset. When important, the results of these tests will be reported by either a
p > 0.05, meaning that no significant differences could be determined at this significance
level, or by a p < 0.05, denoting that the test showed the results of the two compared
groups to be significantly different under this test setup.
4. Results
4.1. Immediate Label Setting
In this paper, at first the real-time emotion classification pipeline with immediate label
setting was applied to Dataset I and Dataset II. The data were streamed to pre-process
and to extract features from tumbling windows with a window length of 1 s. To perform
binary classification for both dimensions of AS (valence and arousal), the self-rating of the
participant was used by applying a threshold at 0.5 and defining high and low classes. As
mentioned earlier, ARF, SRP, and LR classifiers were employed for evaluation. The setting
of five trees and four trees for SRP worked best for Dataset I and for Dataset II, respectively.
ARF included five trees for both data sets. A subject-dependent model was trained with
10-fold cross-validation, and the performances were evaluated with progressive validation.
Figure 9 presents the overall model performances of Dataset I by showing the subject-
wise distribution of the evaluation matrix. The mean F1-Score for the positive and negative
classes of valence and arousal recognition, respectively, are shown only for ARF and SRP
classifiers. The LR showed a comparatively poor performance since it is not an online
model but trained on mini-batches. On the contrary, the ensemble models (ARF and SRP)
show consistently higher F1-Scores, mostly between 0.7 and 0.95, with two outliers, which
validates the emotion classification pipeline built in this paper. Two outliers are visible
from subjects 11 and 30 and might be due to a label imbalance (high/low) in the data or
insufficient data quality.
Figure 9. F1-Score for Valence and Arousal classification achieved by ARF and SRP per subject from
Dataset I.
Furthermore, the means of F1-Score and accuracy over all the subjects from Dataset I
are presented in Table 3. As depicted in “gray”, both evaluation matrices reach more
than 80% for both the ensemble models (ARF and SRP), whereas the performance of LR is
relatively poor. Additionally, Table 3 also shows the comparison to the evaluation of the
baseline results by Miranda-Correa et al. [22] with a reported approximately 50% F1-Score
Sensors 2023, 23, 2387 14 of 23
with no accuracy score reported. Siddharth et al. [73] reports a more than 70% accuracy and
F1-Score, and Topic et al. [74] achieved the current benchmark for this dataset by reporting
a 90% accuracy. However, all the related work mentioned was obtained using a hold-out or
k-fold cross-validation with computation-heavy offline classifiers, making it inadequate
for real-time classification. Consequently, the proposed online classifiers with ensemble
models can contribute to daily life setups.
Table 3. Comparison of mean F1-Scores and accuracy of Valence and Arousal recognition on Dataset I.
Gray represents the results from this paper. NR stands for not reported.
F1-Score Accuracy
Study or Classifier
Valence Arousal Valence Arousal
LR 0.669 0.65 0.702 0.688
ARF 0.825 0.826 0.82 0.846
SRP 0.834 0.831 0.826 0.847
Miranda-Correa et al. [22] 0.576 0.592 NR NR
Siddharth et al. [73] 0.8 0.74 0.83 0.791
Topic et al. [74] NR NR 0.874 0.905
Table 4. Comparison of mean F1-Scores of Arousal and Valence recognition per participant and
device from Dataset I with three classifiers using progressive validation. Bold values indicate the
best-performing model per participant and dimension. The mean total represents the calculated
average of all models’ F1-Scores.
ARF SRP LR
Subject ID
Crown Muse Crown Muse Crown Muse
3 0.902 0.885 0.895 0.898 0.8 0.785
4 0.836 0.794 0.838 0.845 0.793 0.604
5 0.651 0.812 0.699 0.827 0.764 0.682
6 0.836 0.843 0.863 0.889 0.771 0.62
7 0.958 0.833 0.933 0.878 0.841 0.725
Arousal
Table 4. Cont.
ARF SRP LR
Subject ID
Crown Muse Crown Muse Crown Muse
3 0.837 0.887 0.811 0.876 0.716 0.712
4 0.841 0.69 0.773 0.859 0.804 0.524
5 0.546 0.734 0.639 0.748 0.781 0.58
6 0.713 0.687 0.785 0.778 0.73 0.393
7 0.935 0.666 0.926 0.757 0.776 0.616
Valence
Table 5. F1-Score and accuracy for the live affect classification (Dataset III, part 2). Subjects 14 and 17
wore Muse, while subjects 15 and 16 wore the Crown for data collection. The highest scores across all
participants for each evaluation matrix are marked in bold.
F1-Score Accuracy
Subject ID
Valence Arousal Valence Arousal
14 0.521 0.357 0.562 0.385
15 0.601 0.64 0.609 0.575
16 0.353 0.73 0.502 0.575
17 0.512 0.383 0.533 0.24
Furthermore, the prediction results are presented in the form of confusion matrices in
Figure 11, showing that many of the samples have been misclassified. For three out of the
four subjects, a low valence could be classified with a recall of at least 0.567, while for the
arousal dimension, a low class was misclassified most often with a low recall. Though the
proposed prediction pipeline was able to classify valence for over half the samples, it was
unsuccessful in reliably predicting accurate classes for both affect dimensions. However,
the findings from the live experiment show the importance of testing an application with
settings as close as possible to the expected production scenario in order to get a reasonable
performance estimate. Moreover, the results display the potential of emotion classification
for live applications and motivate us to further investigate frameworks for real use cases
with delayed labels instead of solely focusing on the immediate label setting.
Additionally, the lower predictive performance led us to further investigate the de-
layed labels. To imitate production settings, we induced a delay into the pipeline and
applied progressive delayed validation on Dataset II. Therefore, the subject-dependent
model was updated with the true label after 86 samples, i.e., the mean length of the video
stimuli was 86 s. Table 6 displays the F1-Scores of both the models for valence and arousal
recognition with a label delay of 86 s using an ARF with four trees and a window length of
1 s. The F1-Score for individual participants reached 77% for valence and 78% for arousal.
However, the mean F1-Score across all participants achieved 63% for arousal and did not
reach chance level for the valence classification. The performance declines significantly
compared to Table 4, when a delay is induced.
Sensors 2023, 23, 2387 17 of 23
Subject 14, Valence. recall: 0.567 Subject 14, Arousal. recall: 0.387
Subject 15, Valence. recall: 0.612 Subject 15, Arousal. recall: 0.432
Subject 16, Valence. recall: 0.582 Subject 16, Arousal. recall: 0.0
Subject 17, Valence. recall: 0.489 Subject 17, Arousal. recall: 0.008
Figure 11. Confusion matrices for the live affect classification (Dataset III, part 2). Employed model:
ARF (four trees), window length = 1 s. Recall was calculated only for a low class for both the models.
However, the findings justify the poor performance in the live settings and validate the
pipeline as a useful one with the possibility of modifications in future work. Furthermore,
the binary arousal classification with the induced label delay outperforms the baseline
results obtained by Miranda-Correa et al. [22] by 4.5% with an immediate label setting.
However, the results reported by Siddharth et al. [73] and Topic et al. [74] outperform our
work when used with immediate labels.
Sensors 2023, 23, 2387 18 of 23
Table 6. Mean F1-Scores for Valence and Arousal recognition of Dataset II, relayed per participant
and device. Obtained using ARF (with four trees), a window length of 1 s, and progressive delayed
validation with a label delay of 86 s. The bold values represent the highest scores, and the last row
shows the mean F1-Score of all participants.
Valence Arousal
Participant ID
Crown Muse Crown Muse
3 0.338 0.584 0.614 0.718
4 0.674 0.429 0.551 0.575
5 0.282 0.554 0.355 0.69
6 0.357 0.27 0.608 0.619
7 0.568 0.574 0.698 0.769
8 0.266 0.286 0.561 0.574
9 0.553 0.53 0.719 0.749
10 0.767 0.561 0.784 0.691
11 0.469 0.207 0.676 0.418
12 0.443 0.51 0.575 0.679
13 0.335 0.451 0.646 0.711
Mean 0.476 0.46 0.637 0.637
5. Discussion
In this paper, firstly, a real-time emotion classification pipeline was built for binary clas-
sification (high/low) of the two affect dimensions: Valence and Arousal. Adaptive random
forest (ARF), streaming random patches (SRP), and logistic regression (LR) classifiers with
10-fold cross-validation were applied to the EEG data stream. The subject-dependent mod-
els were evaluated with progressive and delayed validation, respectively, when immediate
and delayed labels were available. The pipeline was validated on existing data of ensured
quality from the state-of-the-art AMIGOS [22] dataset. By streaming the recorded data to
the pipeline, the mean F1-Scores were more than 80% for both ARF and SRP models. The
results outperform the authors’ baseline results by approximately 25% and are also slightly
better than the work reported in [73] using the same dataset. The results of Topic et al. [74]
showed a better performance; however, due to the reported complex setup and computa-
tionally expensive methods, the system is unsuitable for real-time emotion classification.
Nevertheless, the results mentioned in the related work apply offline classifiers with a
hold-out or a k-fold cross-validation technique. In contrast, our pipeline applies an online
classifier by employing progressive validation. To the best of our knowledge, no other
work has tested and outperformed our online EEG-based emotion classification framework
on the published AMIGOS dataset.
Secondly, a similar framework to the AMIGOS dataset, Dataset II, was established
within this paper, which can collect neurophysiological data from a wide range of neuro-
physiological sensors. In this paper, two consumer-grade EEG devices were used to collect
data from 15 participants while watching 16 emotional videos. The framework available in
the mentioned repository can be adapted for similar experiments.
Thirdly, and most importantly, we curated data in two experiments to validate our
classification pipeline using the mentioned framework. Eleven participants took part in
acquiring th data for Dataset II, where EEG data were recorded while watching 16 emotion
elicitation videos. The pre-recorded data were streamed to the pipeline and showed a
mean F1-Score of more than 82% with ARF and SRP classifiers using progressive validation.
This finding validates the competence of the pipeline on the challenging dataset from
consumer-grade EEG devices. Additionally, the online classifiers consistently showed bet-
ter performance for ARF and SRP than LR on all compared modalities. However, internal
testing verified that the run-time of the training step of the pipeline of ARF is less than that
of SRP, concluding that ARF should be used in live prediction. The analysis on window
length shows a clear trend of increasing performance scores with decreasing window
length; therefore, a window length of 1 s was chosen for live prediction. Although the two
Sensors 2023, 23, 2387 19 of 23
Author Contributions: Conceptualization, S.M., F.E.F.S. and B.A.; data curation, S.M. and F.E.F.S.;
formal analysis, F.E.F.S.; resources, S.M.; visualization, S.M. and F.E.F.S.; writing—original draft
preparation, S.M.; writing—review and editing, S.M., F.E.F.S. and B.A.; supervision, B.A. All authors
have read and agreed to the published version of the manuscript.
Funding: This research was (partially) funded by the Hasso-Plattner Institute Research School on
Data Science and Engineering. The publication costs were covered by the Deutsche Forschungsge-
meinschaft (DFG, German Research Foundation)—Project number 491466077.
Institutional Review Board Statement: The study was conducted in accordance with the Ethics
Committee of University of Potsdam (44/2022).
Informed Consent Statement: Informed consent was obtained from all participants involved in
the study.
Data Availability Statement: The dataset is available from the authors upon request for scientific
purposes at https://doi.org/10.5281/zenodo.7398263 (accessed on 18 February 2023). The source
code used for analysis in this study can be found at https://github.com/HPI-CH/EEGEMO (accessed
on 18 February 2023).
Acknowledgments: We appreciate the contribution of all the participants who took part in the study
for the sake of science. We also acknowledge the researchers of the AMIGOS dataset for making it
available to the researchers.
Conflicts of Interest: The authors declare no conflict of interest.
Sensors 2023, 23, 2387 20 of 23
Abbreviations
The following abbreviations are used in this manuscript:
Appendix A
References
1. Picard, R.W. Affective Computing; The MIT Press: Cambridge, MA, USA, 2000. [CrossRef]
2. Cowie, R.; Douglas-Cowie, E.; Tsapatsoulis, N.; Votsis, G.; Kollias, S.; Fellenz, W.; Taylor, J. Emotion recognition in human–
computer interaction. IEEE Signal Process. Mag. 2001, 18, 32–80. [CrossRef]
3. Haut, S.R.; Hall, C.B.; Borkowski, T.; Tennen, H.; Lipton, R.B. Clinical features of the pre-ictal state: Mood changes and
premonitory symptoms. Epilepsy Behav. 2012, 23, 415–421. [CrossRef] [PubMed]
4. Kocielnik, R.; Sidorova, N.; Maggi, F.M.; Ouwerkerk, M.; Westerink, J.H.D.M. Smart technologies for long-term stress monitoring
at work. In Proceedings of the 26th IEEE International Symposium on Computer-Based Medical Systems (CBMS), Porto, Portugal,
20–22 June 2013; pp. 53–58. [CrossRef]
5. Schulze-Bonhage, A.; Kurth, C.; Carius, A.; Steinhoff, B.J.; Mayer, T. Seizure anticipation by patients with focal and generalized
epilepsy: A multicentre assessment of premonitory symptoms. Epilepsy Res. 2006, 70, 83–88. [CrossRef]
6. Privitera, M.; Haut, S.R.; Lipton, R.B.; McGinley, J.S.; Cornes, S. Seizure self-prediction in a randomized controlled trial of stress
management. Neurology 2019, 93, e2021–e2031. [CrossRef] [PubMed]
7. Kotwas, I.; McGonigal, A.; Trebuchon, A.; Bastien-Toniazzo, M.; Nagai, Y.; Bartolomei, F.; Micoulaud-Franchi, J.A. Self-control of
epileptic seizures by nonpharmacological strategies. Epilepsy Behav. 2016, 55, 157–164. [CrossRef] [PubMed]
8. Scaramelli, A.; Braga, P.; Avellanal, A.; Bogacz, A.; Camejo, C.; Rega, I.; Messano, T.; Arciere, B. Prodromal symptoms in epileptic
patients: Clinical characterization of the pre-ictal phase. Seizure 2009, 18, 246–250. [CrossRef]
9. Moontaha, S.; Steckhan, N.; Kappattanavar, A.; Surges, R.; Arnrich, B. Self-prediction of seizures in drug resistance epilepsy
using digital phenotyping: A concept study. In Proceedings of the 14th EAI International Conference on Pervasive Computing
Technologies for Healthcare, PervasiveHealth ’20, Atlanta, GA, USA, 18–20 May 2020; Association for Computing Machinery:
New York, NY, USA, 2021; pp. 384–387. [CrossRef]
10. Levenson, R.; Lwi, S.; Brown, C.; Ford, B.; Otero, M.; Verstaen, A. Emotion. In Handbook of Psychophysiology, 4th ed.; Cambridge
University Press: Cambridge, UK, 2016; pp. 444–464. [CrossRef]
11. Liu, H.; Zhang, Y.; Li, Y.; Kong, X. Review on Emotion Recognition Based on Electroencephalography. Front. Comput. Neurosci.
2021, 15, 758212. [CrossRef] [PubMed]
12. Krigolson, O.E.; Williams, C.C.; Norton, A.; Hassall, C.D.; Colino, F.L. Choosing MUSE: Validation of a Low-Cost, Portable EEG
System for ERP Research. Front. Neurosci. 2017, 11, 109. [CrossRef]
13. Bird, J.J.; Manso, L.J.; Ribeiro, E.P.; Ekárt, A.; Faria, D.R. A Study on Mental State Classification using EEG-based Brain–Machine
Interface. In Proceedings of the 2018 International Conference on Intelligent Systems (IS), Funchal, Portugal, 25–27 September
2018; pp. 795–800. [CrossRef]
14. Teo, J.; Chia, J.T. Deep Neural Classifiers For Eeg-Based Emotion Recognition In Immersive Environments. In Proceedings of the
2018 International Conference on Smart Computing and Electronic Enterprise (ICSCEE), Shah Alam, Malaysia, 11–12 July 2018;
pp. 1–6. [CrossRef]
15. Gonzalez, H.A.; Yoo, J.; Elfadel, I.M. EEG-based Emotion Detection Using Unsupervised Transfer Learning. In Proceedings of the
2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany,
23–27 July 2019; pp. 694–697. [CrossRef]
16. Hasnul, M.A.; Aziz, N.A.A.; Alelyani, S.; Mohana, M.; Aziz, A.A. Electrocardiogram-Based Emotion Recognition Systems and
Their Applications in Healthcare—A Review. Sensors 2021, 21, 5015. [CrossRef]
17. Huang, X.; Kortelainen, J.; Zhao, G.; Li, X.; Moilanen, A.; Seppänen, T.; Pietikäinen, M. Multi-modal emotion analysis from facial
expressions and electroencephalogram. Comput. Vis. Image Underst. 2016, 147, 114–124. [CrossRef]
18. Li, J.; Qiu, S.; Shen, Y.Y.; Liu, C.L.; He, H. Multisource Transfer Learning for Cross-Subject EEG Emotion Recognition. IEEE Trans.
Cybern. 2019, 50, 3281–3293. [CrossRef] [PubMed]
19. Hasan, T.F.; Tatum, W.O. Ambulatory EEG Usefulness in Epilepsy Management. J. Clin. Neurophysiol. 2021, 38, 101–111.
[CrossRef] [PubMed]
20. Bota, P.J.; Wang, C.; Fred, A.L.N.; Plácido Da Silva, H. A Review, Current Challenges, and Future Possibilities on Emotion
Recognition Using Machine Learning and Physiological Signals. IEEE Access 2019, 7, 140990–141020. [CrossRef]
21. Horvat, M.; Dobrinic, M.; Novosel, M.; Jercic, P. Assessing emotional responses induced in virtual reality using a consumer EEG
headset: A preliminary report. In Proceedings of the 2018 41st International Convention on Information and Communication
Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 21–25 May 2018; pp. 1006–1010. [CrossRef]
22. Miranda-Correa, J.A.; Abadi, M.K.; Sebe, N.; Patras, I. AMIGOS: A Dataset for Affect, Personality and Mood Research on
Individuals and Groups. IEEE Trans. Affect. Comput. 2021, 12, 479–493. [CrossRef]
23. Laureanti, R.; Bilucaglia, M.; Zito, M.; Circi, R.; Fici, A.; Rivetti, F.; Valesi, R.; Oldrini, C.; Mainardi, L.T.; Russo, V. Emotion
assessment using Machine Learning and low-cost wearable devices. In Proceedings of the 2020 42nd Annual International
Conference of the IEEE Engineering in Medicine Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 576–579.
[CrossRef]
24. Dadebayev, D.; Goh, W.W.; Tan, E.X. EEG-based emotion recognition: Review of commercial EEG devices and machine learning
techniques. J. King Saud-Univ.-Comput. Inf. Sci. 2022, 34, 4385–4401. [CrossRef]
25. Suhaimi, N.S.; Mountstephens, J.; Teo, J. EEG-Based Emotion Recognition: A State-of-the-Art Review of Current Trends and
Opportunities. Comput. Intell. Neurosci. 2020, 2020, 8875426. [CrossRef]
Sensors 2023, 23, 2387 22 of 23
26. Gomes, H.M.; Bifet, A.; Read, J.; Barddal, J.P.; Enembreck, F.; Pfahringer, B.; Holmes, G.; Abdessalem, T. Adaptive random forests
for evolving data stream classification. Mach. Learn. 2017, 106, 1469–1495. [CrossRef]
27. Müller, K.R.; Tangermann, M.; Dornhege, G.; Krauledat, M.; Curio, G.; Blankertz, B. Machine learning for real-time single-trial
EEG-analysis: From brain–computer interfacing to mental state monitoring. J. Neurosci. Methods 2008, 167, 82–90. [CrossRef]
28. Liu, Y.; Sourina, O.; Nguyen, M.K. Real-Time EEG-Based Human Emotion Recognition and Visualization. In Proceedings of the
2010 International Conference on Cyberworlds CW, Singapore, 20–22 October 2010; pp. 262–269. [CrossRef]
29. Liu, Y.; Sourina, O. EEG-based subject-dependent emotion recognition algorithm using fractal dimension. In Proceedings
of the 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC), San Diego, CA, USA, 5–8 October 2014;
pp. 3166–3171. [CrossRef]
30. Lan, Z.; Sourina, O.; Wang, L.; Liu, Y. Real-time EEG-based emotion monitoring using stable features. Vis. Comput. 2016,
32, 347–358. [CrossRef]
31. Lan, Z. EEG-Based Emotion Recognition Using Machine Learning Techniques. Ph.D. Thesis, Nanyang Technological University,
Singapore, 2018. [CrossRef]
32. Hou, X.; Liu, Y.; Sourina, O.; Mueller-Wittig, W. CogniMeter: EEG-based Emotion, Mental Workload and Stress Visual Monitoring.
In Proceedings of the 2015 International Conference on Cyberworlds (CW), Visby, Sweden, 7–9 October 2015; pp. 153–160. [CrossRef]
33. Lan, Z.; Liu, Y.; Sourina, O.; Wang, L.; Scherer, R.; Müller-Putz, G. SAFE: An EEG dataset for stable affective feature selection.
Adv. Eng. Inform. 2020, 44, 101047. [CrossRef]
34. Javaid, M.M.; Yousaf, M.A.; Sheikh, Q.Z.; Awais, M.M.; Saleem, S.; Khalid, M. Real-Time EEG-Based Human Emotion Recognition.
In Neural Information Processing; Arik, S., Huang, T., Lai, W.K., Liu, Q., Eds.; Springer International Publishing: Cham, Switzerland,
2015; Volume 9492, pp. 182–190. [CrossRef]
35. Sarno, R.; Munawar, M.N.; Nugraha, B.T. Real-Time Electroencephalography-Based Emotion Recognition System. Int. Rev.
Comput. Softw. IRECOS 2016, 11, 456. [CrossRef]
36. Bajada, J.; Bonello, F.B. Real-time EEG-based Emotion Recognition using Discrete Wavelet Transforms on Full and Reduced
Channel Signals. arXiv 2021, arXiv:2110.05635. [CrossRef]
37. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for
Emotion Analysis ;Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [CrossRef]
38. Li, J.; Chen, H.; Cai, T. FOIT: Fast Online Instance Transfer for Improved EEG Emotion Recognition. In Proceedings of the
2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 16–19 December 2020;
pp. 2618–2625. [CrossRef]
39. Zheng, W.-L.; Lu, B.-L. Investigating Critical Frequency Bands and Channels for EEG-Based Emotion Recognition with Deep
Neural Networks. IEEE Trans. Auton. Mental Dev. 2015, 7, 162–175. [CrossRef]
40. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. EmotionMeter: A Multimodal Framework for Recognizing Human Emotions.
IEEE Trans. Cybern. 2019, 49, 1110–1122. [CrossRef]
41. Nandi, A.; Xhafa, F.; Subirats, L.; Fort, S. Real-Time Emotion Classification Using EEG Data Stream in E-Learning Contexts.
Sensors 2021, 21, 1589. [CrossRef]
42. Bifet, A.; Gavaldà, R. Adaptive Learning from Evolving Data Streams. In Proceedings of the 8th International Symposium
on Intelligent Data Analysis: Advances in Intelligent Data Analysis VIII, IDA ’09, Lyon, France, 31 August–2 September 2009;
Springer: Berlin/Heidelberg, Germany, 2009; pp. 249–260. [CrossRef]
43. Katsigiannis, S.; Ramzan, N. DREAMER: A Database for Emotion Recognition Through EEG and ECG Signals from Wireless
Low-cost Off-the-Shelf Devices. IEEE J. Biomed. Health Inform. JBHI 2018, 22, 98–107. [CrossRef]
44. Subramanian, R.; Wache, J.; Abadi, M.K.; Vieriu, R.L.; Winkler, S.; Sebe, N. ASCERTAIN: Emotion and Personality Recognition
Using Commercial Sensors. IEEE Trans. Affect. Comput. 2018, 9, 147–160. [CrossRef]
45. Bradley, M.M.; Lang, P.J. Measuring emotion: The self-assessment manikin and the semantic differential. J. Behav. Ther. Exp.
Psychiatry 1994, 25, 49–59. [CrossRef]
46. Watson, D.; Clark, L.A.; Tellegen, A. Development and validation of brief measures of positive and negative affect: The PANAS
scales. J. Personal. Soc. Psychol. 1988, 54, 1063–1070. [CrossRef]
47. Towle, V.L.; Bolaños, J.; Suarez, D.; Tan, K.B.; Grzeszczuk, R.P.; Levin, D.N.; Cakmur, R.; Frank, S.A.; Spire, J.P. The spatial location
of EEG electrodes: Locating the best-fitting sphere relative to cortical anatomy. Electroencephalogr. Clin. Neurophysiol. 1993, 86, 1–6.
[CrossRef]
48. Peirce, J.; Gray, J.; Simpson, S.; MacAskill, M.; Höchenberger, R.; Sogo, H.; Kastman, E.; Lindeløv, J. PsychoPy2: Experiments in
behavior made easy. Behav. Res. Methods 2019, 51, 195–203. [CrossRef] [PubMed]
49. Dan-Glauser, E.S.; Scherer, K.R. The Geneva affective picture database (GAPED): A new 730-picture database focusing on valence
and normative significance. Behav. Res. Methods 2011, 43, 468–477. [CrossRef] [PubMed]
50. Kurdi, B.; Lozano, S.; Banaji, M. Introducing the Open Affective Standardized Image Set (OASIS). Behav. Res. Methods 2017,
49, 457–470. [CrossRef] [PubMed]
51. Lang, P.J.; Bradley, M.M.; Cuthbert, B.N. International Affective Picture System (IAPS): Affective Ratings of Pictures and Instruction
Manual; Technical Report; Springer: Berlin/Heidelberg, Germany, 2008. [CrossRef]
52. Panda, R.; Malheiro, R.; Paiva, R.P. Novel Audio Features for Music Emotion Recognition. IEEE Trans. Affect. Comput. 2018,
11, 614–626. [CrossRef]
Sensors 2023, 23, 2387 23 of 23
53. Zhang, K.; Zhang, H.; Li, S.; Yang, C.; Sun, L. The PMEmo Dataset for Music Emotion Recognition. In Proceedings of the 2018
ACM on International Conference on Multimedia Retrieval, ICMR ’18, Yokohama, Japan, 11–14 June 2018; ACM: New York, NY,
USA, 2018; pp. 135–142. [CrossRef]
54. Abadi, M.K.; Subramanian, R.; Kia, S.M.; Avesani, P.; Patras, I.; Sebe, N. DECAF: MEG-Based Multimodal Database for Decoding
Affective Physiological Responses. IEEE Trans. Affect. Comput. 2015, 6, 209–222. [CrossRef]
55. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A Multi-Modal Affective Database for Affect Recognition and Implicit Tagging.
Affect. Comput. IEEE Trans. 2012, 3, 42–55. [CrossRef]
56. Verma, G.; Dhekane, E.G.; Guha, T. Learning Affective Correspondence between Music and Image. In Proceedings of the
2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019;
pp. 3975–3979. [CrossRef]
57. Russell, J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980, 39, 1161. [CrossRef]
58. Breyer, B.; Bluemke, M. Deutsche Version der Positive and Negative Affect Schedule PANAS (GESIS Panel). In Zusammenstellung
Sozialwissenschaftlicher Items und Skalen; Social Science Open Access Repository (SSOAR): Mannheim, Germany, 2016. [CrossRef]
59. Betella, A.; Verschure, P. The Affective Slider: A Digital Self-Assessment Scale for the Measurement of Human Emotions. PLoS
ONE 2016, 11, e0148037. [CrossRef]
60. Jiang, X.; Bian, G.B.; Tian, Z. Removal of Artifacts from EEG Signals: A Review. Sensors 2019, 19, 987. [CrossRef]
61. Sörnmo, L.; Laguna, P. Chapter 3—EEG Signal Processing. In Bioelectrical Signal Processing in Cardiac and Neurological Applications;
Biomedical Engineering; Academic Press: Burlington, NJ, USA, 2005; pp. 55–179. [CrossRef]
62. Akwei-Sekyere, S. Powerline noise elimination in neural signals via blind source separation and wavelet analysis. PeerJ PrePrints
2014, 3. [CrossRef]
63. Sweeney, K.; Ward, T.; Mcloone, S. Artifact Removal in Physiological Signals-Practices and Possibilities. IEEE Trans. Inf. Technol.
Biomed. Publ. IEEE Eng. Med. Biol. Soc. 2012, 16, 488–500. [CrossRef] [PubMed]
64. Yao, D.; Qin, Y.; Hu, S.; Dong, l.; Vega, M.; Sosa, P. Which Reference Should We Use for EEG and ERP practice? Brain Topogr. 2019,
32, 530–549. [CrossRef] [PubMed]
65. Welch, P. The use of fast Fourier transform for the estimation of power spectra: A method based on time averaging over short,
modified periodograms. IEEE Trans. Audio Electroacoust. 1967, 15, 70–73. [CrossRef]
66. Montiel, J.; Halford, M.; Mastelini, S.M.; Bolmier, G.; Sourty, R.; Vaysse, R.; Zouitine, A.; Gomes, H.M.; Read, J.; Abdessalem, T.;
et al. River: Machine learning for streaming data in Python. J. Mach. Learn. Res. 2021, 22, 4945–4952. [CrossRef]
67. Grzenda, M.; Gomes, H.M.; Bifet, A. Delayed labelling evaluation for data streams. Data Min. Knowl. Discov. 2020, 34, 1237–1266.
[CrossRef]
68. Blum, A.; Kalai, A.T.; Langford, J. Beating the hold-out: Bounds for K-fold and progressive cross-validation. In Proceedings of
the Twelfth Annual Conference on Computational Learning Theory COLT ’99, Santa Cruz, CA, USA, 7–9 July 1999.
69. McMahan, H.B.; Holt, G.; Sculley, D.; Young, M.; Ebner, D.; Grady, J.; Nie, L.; Phillips, T.; Davydov, E.; Golovin, D.; et al. Ad
Click Prediction: A View from the Trenches. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge
Discovery and Data Mining (KDD), Chicago, IL, USA, 11–14 August 2013; pp. 1222–1230. [CrossRef]
70. Gomes, H.M.; Read, J.; Bifet, A. Streaming Random Patches for Evolving Data Stream Classification. In Proceedings of the 2019
IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 240–249. [CrossRef]
71. Parker, B.; Khan, L. Detecting and Tracking Concept Class Drift and Emergence in Non-Stationary Fast Data Streams. In Proceed-
ings of the AAAI, Austin, TX, USA, 25–30 January 2015. [CrossRef]
72. Aggarwal, C.C. Data Classification: Algorithms and Applications, 1st ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2014;
pp. 636–638.
73. Siddharth, S.; Jung, T.P.; Sejnowski, T.J. Utilizing Deep Learning Towards Multi-Modal Bio-Sensing and Vision-Based Affective
Computing. IEEE Trans. Affect. Comput. 2019, 13, 96–107. [CrossRef]
74. Topic, A.; Russo, M. Emotion recognition based on EEG feature maps through deep learning network. Eng. Sci. Technol. Int. J.
JESTECH 2021, 24, 1442–1454. [CrossRef]
75. Liu, Y.J.; Yu, M.; Zhao, G.; Song, J.; Ge, Y.; Shi, Y. Real-Time Movie-Induced Discrete Emotion Recognition from EEG Signals.
IEEE Trans. Affect. Comput. 2018, 9, 550–562. [CrossRef]
76. Ekman, P.; Friesen, W. Unmasking the Face: A Guide to Recognizingemotions from Facial Clues; Prentice-Hall: Oxford, UK, 1975.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.