Brainsci 13 00770 v2
Brainsci 13 00770 v2
sciences
Article
Machine Learning on Visibility Graph Features Discriminates
the Cognitive Event-Related Potentials of Patients with Early
Alzheimer’s Disease from Healthy Aging
Jesse Zhang 1 , Jiangyi Xia 2 , Xin Liu 3 and John Olichney 2, *
1 Computer Science Department, University of Southern California, Los Angeles, CA 90089, USA;
jessez@usc.edu
2 UC Davis Center for Mind and Brain, Davis, CA 95618, USA; jixia@ucdavis.edu
3 UC Davis Computer Science Department, Davis, CA 95616, USA; xinliu@ucdavis.edu
* Correspondence: jmolichney@ucdavis.edu
(9–11 Hz) that is attenuated for repeated relative to new words in healthy elderly [10]. This
alpha word repetition effect is also compromised in amnestic MCI and correlated with
verbal memory measures [10].
A limitation of traditional ERP/oscillatory analyses is that they usually focus on the
timing and the magnitude of pre-defined components at the expense of the overall pattern
and complexity of EEG data. Some prior works convert EEG signals to visibility graphs
(VGs) [14], which preserve many features of the original EEG signal. Converting resting
state EEG signals to VGs allows for discriminative graph features to be discovered [15] and
utilized in high-accuracy neural network based classification (98%) between AD patients
and normal elderly [16].
Other studies which have applied neural networks or other machine learning algo-
rithms to resting state EEG in AD include Morabito et al. [17], who used convolutional
neural networks on 19 channel EEG and achieved a three-class AD/MCI/cognitively
normal (CN) classification accuracy of 82% [17,18]. Zhao and He [19] combined deep
belief networks with support vector machines on 16 channel EEG signals and achieved
92% accuracy classifying AD vs. CN [18,19]. Duan et al. [20] quantified between-channel
connectivity of resting-state EEG signals in MCI and mild AD patients using coherence
measures; they used the Resnet-18 model [21] to classify between MCI and controls, and
AD and controls with an average 93% and 98.5% accuracy, respectively.
Despite the promise of the above studies and other machine learning algorithms which
have used biomarkers of AD to improve diagnostic accuracy [22], there are still to date
no widely used machine learning algorithms for the clinical diagnosis of AD. Historically,
clinical diagnosis of possible and probable AD (generally found to be between 80 and 90%
accurate in clinicopathological studies) was based on recognizing the typical cognitive
and behavioral symptoms of this dementia and the exclusion of other possible causes of
dementia, whereas a “definite AD” diagnosis was only possible via invasive brain measures
from a biopsy or autopsy providing histopathological evidence of AD [23]. Currently, the
International Working Group (IWG) recommends that the clinical diagnosis of AD be
restricted to those with positive biomarkers together with specific AD phenotypes [24].
While purely biological definitions of AD (e.g., [25]) have become more widely used for
research purposes in recent years, the IWG considers the present limitations of biomarkers
sufficient that they should not be used for the diagnosis of disease in the cognitively
unimpaired [24]. Thus, the “gold standard” for the clinical diagnosis of AD is criteria
(e.g., [23,26]) which incorporate multiple biomarkers (including markers of amyloid-β
(Aβ) and tau pathology, neuronal injury and neurodegeneration) along with the clinical
phenotype. With the rapid emergence of machine learning algorithms into medical research,
this could, however, change rapidly in upcoming years [27].
Our hypothesis is that word repetition tasks, which have been shown sensitive to
detect MCI-to-AD conversion and even preclinical AD using ERPs [7,8,11], can also be used
to discriminate AD from normal elderly with high accuracy using a VG-based machine
learning approach. Compared to resting state EEG, word repetition task signals are expected
to yield better discriminative features given that verbal memory impairments are the best
predictors of MCI to AD conversion [28]. Combining these two lines of past work, we
converted EEG signals recorded during word repetition experiments to visibility graphs.
We operated under the assumption that the ERP components of interest will be preserved
after conversion to graphs and features extracted from these graphs will encode the ERP
components while reducing variance across subject data for better downstream machine
learning classification performance.
Therefore, this work focuses on the analysis of EEG signals and extracting features
from them that are useful in discriminating between AD and RNE in a variety of machine
learning algorithms. To demonstrate the generalizability of those features, we tested
whether they can also effectively discriminate between prodromal Alzheimer’s (pAD, MCI
patients who converted to Alzheimer’s Dementia within 3 years) and robust normal elderly
(RNE, normal elderly persons who have remained cognitively normal for the duration of
Brain Sci. 2023, 13, 770 3 of 21
follow-up). We apply a similar approach to that of Ahmadlou et al. [16], although extracting
many more features (including many novel ones in this context) from word repetition task
EEG signals (instead of resting state as in Ahmadlou et al. [16]).
In our framework, pictured in Figure 1, we first collect EEG data from word repetition
tasks. We then perform pre-processing of this data and then convert the EEG signals to
visibility graphs. From these VGs we extract 12 features and perform statistical tests for
feature selection, keeping the discovered statistically significant predictors as inputs for
machine learning algorithms. Finally, the dimensionality of this feature space is reduced
with principal component analysis and we use the resulting reduced feature space as inputs
to machine learning algorithms.
2. Methods
Figure 1 details the framework. Open-sourced code for our method is available
online at https://github.com/jesbu1/ML-Visibility-Graphs-for-Alzheimers (accessed on
19 March 2023).
2.1. Participants
EEG and behavioral data were taken from 15 patients with probable AD (mean age
78.5 years, range 67–91) [23] recruited primarily through the Alzheimer’s Disease Research
Centers at the University of California, San Diego and the University of California, Davis.
Additional data were taken from 15 patients with amnestic MCI (mean age 74.6 years,
range 60–84) [29] who later converted to dementia and 11 healthy elderly controls (mean
age 74.1 years, range 57–79) who were recruited in a previous published longitudinal
study [8]. See Table 1 for participant details. All participants were screened for treatable
causes of cognitive impairments such as vitamin B12 deficiency and thyroid dysfunction,
and underwent a brain scan (generally MRI) prior to enrollment. The exclusion criteria
included stroke, epilepsy and psychiatric conditions, as well as several classes of central
nervous system (CNS) active medications.
Brain Sci. 2023, 13, 770 4 of 21
Table 1. Mean ± SD values of demographics and MMSE (Mini Mental State Examination) scores in
the three groups. Note: 3 AD patients had no MMSE scores; Montreal Cognitive Assessment (MoCA)
scores were converted to MMSE [30] for these 3 patients.
RNE pAD AD
N 11 15 11
Age (yrs) 74.1 ± 6.8 74.6 ± 6.9 78.5 ± 7.5
Sex 7F, 4M 5F, 10M 4F, 7M
Education (yrs) 15.8 ± 2.8 16.8 ± 2.8 14.6 ± 2.6
MMSE 29.7 ± 0.5 26.9 ± 2.0 * 22.9 ± 2.8 #
* p < 0.05: RNE vs. pAD, # p < 0.05: pAD vs. AD.
The patients were tested with an EEG word repetition paradigm and clinical assess-
ments. At the initial baseline recording session, the 15 MCI patients all met Petersen Criteria
for amnestic MCI [31] but not for dementia [32]. Probable AD was diagnosed according to
criteria set out by the National Institute of Neurological and Communicative Disorders and
Stroke–Alzheimer’s Disease and Related Disorders Association [23]. The 15 MCI patients
subsequently converted to AD within 3 years of their initial baseline session (mean number
of years 1.62 ± 0.7). In the present study we focus on the initial baseline ERP data in order
to investigate neural activity associated with AD and prodromal AD (pAD, MCI to AD
conversion within 3 years). For more information about participant demographics and
their neurocognitive test results please refer to [7,10].
θ 4–8 Hz, α 8–13 Hz, β 13–30 Hz, γ 30–45 Hz), using zero-phase Hamming-windowed
sync finite impulse response filters as implemented in the EEGLAB (pop_eegfiltnew).
This function automatically selects the optimal filter order and transition bandwidth to
minimize filter distortions and maximize time precision. For each of the five frequency
bands of interest, a high-pass filter was first applied and then a low-pass filter. Transition
bandwidths were set to be 25% of the passband edge for passband edges >4 Hz, with
−6 dB cutoff frequency at the center of the transition band. For the 4 Hz passband we used
a transition bandwidth of 2 Hz and for the 1 Hz passband (δ band) we used a transition
bandwidth of 1 Hz. Finally, raw and bandpass filtered EEG segments were extracted 1 s
before and 2 s after the word onset for further analyses.
n − (m + j)
xm+ j < xn + ( )( xm − xn ) ∀ j ∈ Z+ : j < n − m
n−m
Figure 2 demonstrates the creation of a VG. The top graph represents the original time
series, while the graph underneath represents the corresponding nodes and edges of the
visibility graph. There is a line connecting points in the time series (and an undirected
edge in the corresponding VG) if and only if those two points are visible from each other.
Visibility graphs allow for features to be extracted which can encode temporal locality
(as a node is always connected to its direct neighbors in the original EEG signal) but also
features which capture information from nodes that are farther away, as nodes that are
visible from each other will be connected, even if they are far away in time in the original
signal. In general, VGs are biased towards creating local edges that capture information
about the signal over short periods of time, with the exception of peaks in the signal. VGs
can also only be extracted per electrode; however, we compensate for this by also extracting
a cross-channel feature, as detailed below.
Brain Sci. 2023, 13, 770 6 of 21
Figure 2. The top graph represents a time series and the edges between points signify which points
can see each other. The bottom graph represents the VG of the time series, with nodes corresponding
to timepoints and edges corresponding to lines of visibility. χi EEG Voltage at time i, ai VG node for
timepoint i.
2| Ei |
Ci = (2)
|Ki |(|Ki | − 1)
where | Ei | denotes the number of edges of the neighbors of a node i, |Ki | indicates the
|K |(|K |−1)
number of neighbors of node i and i 2 i represents the number of possible connections
in a complete graph consisting of node i’s neighbors. During t-testing, we compared the
average number of edges per person between the two groups.
1 1
Eglobal = ∑ dij
|V |(|V | − 1) i,j,i
(4)
6= j
It is interpreted as sum of all inverse shortest path distances divided by the number of
shortest path distances counted. A higher global efficiency corresponds to a network that
is more efficient at transmitting/combining information and relates to the small-worldness
of the network [15,38,47–50]. In context, a higher global efficiency in a VG means that there
are likely more EEG time points that are visible from other points which are relatively
farther away in time.
1 1 1
|V | ∑ ∑
Elocal = (5)
i
| Vgi |(| Vgi | − 1 ) j,k,j6=k
d jk
where |Vgi | represents the number of vertices in the subgraph of vertex i (composed only of
its direct neighbors) and |V | represents the number of vertices of the entire graph [47]. As
each edge in our VG is of weight one, a higher local efficiency corresponds to more direct
edges on average in each subgraph, indicating EEG signals with variations in voltage that
allow for a greater number of direct connections between points close in time.
Brain Sci. 2023, 13, 770 8 of 21
2.6.5. Small-Worldness
Small-worldness is a measure of how much a graph acts like a small-world network.
Small-world networks have the property that the typical distance between any two ran-
domly chosen vertices grows logarithmically in terms of total number of vertices of a
graph [47]. As logarithmic functions grow very slowly, this correlates with low average
shortest path lengths and high global efficiencies and clustering coefficients. A measure of
small-worldness, S was defined by Humphries and Gurney [51] as
C/Cr
S=
L/Lr
where C, Cr are the average clustering coefficients of the graph in question and a random
graph, respectively, and L, Lr are the average shortest paths lengths between all pairs of
vertices in the graph in question and the random graph, respectively. Our random graphs
were generated with the Erdös–Rényi method [52], and the same random graph was used
to compare all VGs.
2.6.9. Density
Graph density is a measure of how close a graph is to having the maximum number of
edges. It is simply the actual number of edges divided by the maximum possible number
of edges [57]. Density, D, for an undirected graph is defined as
2| E |
D= (8)
|V |(|V | − 1)
|V |(|V |−1)
as it can have at most 2 edges. Density can highlight differences in the number of
edges of VGs across groups.
PCA
We select all features with a p-value of less than 0.01. A high number of feature
combinations combined with a false positive rate of 1% lead us to use principal component
analysis (PCA), a method of linearly mapping features from a higher dimensional space
onto a lower dimensional subspace spanned by the eigenvectors that account for the
directions of highest variance. As suggested by Ahmadlou et al. [16], we apply PCA to
reduce the dimensionality of the feature space to about 10% of its original dimensionality.
At a high level, PCA can be interpreted as a way to linearly project the vector onto a
lower-dimensional latent space such that the distance between the original and projected
latent datapoints is minimized.
Specifically, the “principal components” of PCA are calculated by performing an
eigendecomposition of the covariance matrix of the data. Consider an n × d matrix X where
n is the number of datapoints (number of patients in our study) and d is the dimensionality
of the feature space (the statistically significant features discovered by t-testing). The input
Brain Sci. 2023, 13, 770 10 of 21
matrix X is first normalized, and then the principal components and their weights can be
discovered by the following eigendecomposition:
1
XX T = P D PT . (9)
n−1 |{z} |{z} |{z}
components component magnitudes components
Assuming the prinicpal component vectors and their magnitudes are sorted in descending
order of magnitude, then feature reduction to k features is performed by taking the first
k components. These k components thereby intuitively correspond to the k directions of
highest variance in the data. In our study we use PCA to reduce the data dimensionality to
11, selected from cross-validation of values around 10% of the number of original features.
5. Results
We found 72 statistically significant (p < 0.01) features. The total number of features
tested was 5976 (resulting in 60 features that are expected to be false positives). The total
5976 is derived from 15 channels × ((5 bands + 1 raw) × 11 single-channel features × 6
conditions) + ((5 bands + 1 raw) × 1 all-channel feature (CCSS) × 6 conditions)). We utilize
PCA to reduce the number of features down to 11, close to the number of features expected
to be true positives, in order to combat the high number of expected false positives.
combination. The word conditions are defined in the Word Repetition Paradigm subsection.
Additionally, Table 3 displays the number of features produced by each channel for all
conditions.
Table 2. Comparing the number of features produced by each band. Global Table Key: AN: All New,
NC: New Congruous, NI: New Incongruous, AO: All Old, OC: Old Congruous, OI: Old Incongruous.
Category AN NC NI AO OC OI Total
raw 2 2 0 2 3 3 12
delta δ 0 1 5 6 7 2 21
theta θ 2 0 0 0 1 0 3
alpha α 0 11 0 3 4 0 18
beta β 0 0 5 0 4 1 10
gamma γ 0 3 0 4 1 0 8
Total 4 18 5 19 20 6 72
Channel AN NC NI AO OC OI Total
Fz 0 7 0 3 0 0 10
Pz 0 2 0 1 2 2 7
Cz 0 0 0 0 1 1 2
F7 1 0 0 0 1 0 2
F8 0 1 4 0 5 0 10
Bl 0 1 0 0 0 0 1
Br 2 0 1 0 4 0 7
L41 0 0 0 1 0 0 1
R41 0 4 0 1 1 2 8
Wl 0 0 0 6 0 0 6
Wr 1 1 0 1 0 1 4
T5 0 1 0 1 0 0 2
T6 0 0 0 2 1 0 3
O1 0 0 0 1 4 0 5
O2 0 1 0 2 1 0 4
While all sub-bands and the raw signal produced at least one significant result, the
δ and α sub-bands and raw signal seems to be the most effective in discriminating across
groups. As an example, Figure 3 demonstrates the mapping from the average EEG time
series for each group for the raw signal in electrode R41 under the condition Old Congruous
to averaged node degree time series (i.e., the number of other timepoints visible from each
timepoint in the voltage graph).
Finally, we visualize the separation of patient classes by projecting the 72 features
down to 2 dimensions in Figure 4. In the comparison of all patient classes against each
other (top), we see three clear clusters of points for each class. Notably, the pAD group lies
in between the AD and RNE groups in the top subfigure. This may be because the features
extracted from the AD vs. RNE comparison are likely less significant for the pAD patients,
although they are still general enough to create clear separation between the three groups.
In the AD vs. RNE projection plot (bottom right), we see that all classifiers are able to
perfectly separate the two groups, even in two dimensions. The pAD vs. RNE plot (bottom
left) also demonstrates very good separability between the two groups in two dimensions,
although the two sets of points are closer together than in the AD vs. RNE comparison.
The 10 most important features for each two-dimensional PCA projection comparison
from Figure 4 are listed in Table 4. The feature–band–electrode combinations that were
shared across at least two PCA comparisons are bolded and numbered in the tables. The
α and δ bands produced the largest number of these shared combinations, and the most
common features in these were global efficiency, density, TSP and GIC in electrodes F8, Fz,
O1 and R41. In every single shared combination, the value of the feature increased in the
Brain Sci. 2023, 13, 770 12 of 21
groups with dementia. The increase in these feature values generally indicates an increase
in the number of edges between nodes, indicating significantly different ERP structure that
results in the change in their VGs.
Raw Voltage in Electrode wr; Condition: Old Congruous Raw Node Degrees in Electrode wr; Condition: Old Congruous
RNE
AD 1.6
0 pAD
1.8
1
2.0
Degree of Node
2
Voltage ( V)
3 2.2
4 2.4
5 2.6
9000
800
700
600
500
400
300
200
100
00
200
300
400
500
600
700
800
90
1000
1100
1200
1300
1400
1500
1600
1700
1800
1900
2000
00
9000
800
700
600
500
400
300
200
100
00
200
300
400
500
600
700
800
90
1000
1100
1200
1300
1400
1500
1600
1700
1800
1900
2000
00
10
10
10
10
Time past word stimulus (ms) Time past word stimulus (ms)
Figure 3. The raw band averaged voltage time series on the left, corresponding averaged degree
sequences on the right. Each timepoint (represented by a unique node in the VG) in the left graph is
represented by the average EEG voltage for each group; in the right graph it is represented by the
average degree for its associated node in each group.
Figure 4. Two-dimensional PCA projections of data with associated decision boundaries for all
classifiers, datapoints and comparisons. For each plot, the PCA components were computed with
only the data in the plot to see the actual input to the ML algorithms. Top: Comparison of all classes,
using the 72 extracted features from AD vs. RNE. The datapoints from all classes are projected
together down to 2 dimensions for a 3-way comparison. pAD patients are intermediate to the RNE
and AD patients, as expected. Bottom Left: pAD and RNE are also distinctly separated, resulting in
excellent performance in our results by all classifiers. Bottom Right: In two dimensions, we easily
see that AD and RNE are linearly separable with the features we extracted.
Brain Sci. 2023, 13, 770 13 of 21
Table 4. Top 10 PCA Loading Table. The top 10 magnitude features for each of the two components
and comparisons in Figure 4. Any feature–band–electrode combination that was shared across at
least two of these comparisons is bolded and shares the same superscript number in the table. The
α and δ bands exclusively produce combinations shared across multiple PCA components in each
comparison, and the features that appear the most across those shared are global efficiency, density,
TSP and GIC. The conditions and electrodes that produce these shared features are Old Congruous,
New Incongruous and New Congruous in electrodes F8, Fz, O1 and R41.
5.2. Classification
All models are trained and evaluated 100 times, each time randomly splitting the
dataset into a training set of 85% of the patients and a testing set composed of the remaining
15%. We report classification metrics on the 15% testing set, where the metrics are averaged
across all 100 trials for each model. The PCA feature reduction step is performed each
time only on the features for the patients in the training set. The best results across all
classifiers are obtained by reducing the dimension of the feature vector to 11 via PCA.
On the AD vs. RNE comparison, we utilize the features extracted on the full dataset for
classification to test the ability of the features discovered and analyzed in Section 5.1. To
test the generalization of these features, we use only the 72 discovered AD vs. RNE features
for classification of the pAD patient group.
In summary, the accuracy for AD vs. RNE was 100% across all classifiers and the best
discrimination of pAD vs. RNE—using the 72 discovered AD vs. RNE features to measure
generalization—was 92.5% with the ANN. Notably, AD vs. RNE was perfectly solved with
linear classifiers and pAD vs. RNE classification performance with the same classifiers was
also excellent. Logistic regression achieved perfect precision and a very high AUC score
(0.99); however, the ANN provided even higher accuracy (92.5 %) and very similar AUC
and precision scores. Table 5 presents the classification metrics (accuracy, precision, recall,
AUC).
Table 5. Classification Statistics. Mean classification statistics and standard deviations for all classifiers
on both classification tasks. The rounded best performance across each column for each classification
type is bolded. AUC = Area Under (the ROC) Curve.
discriminate AD from RNE using a word repetition task EEG paradigm. Novel features
introduced in this paper have been shown to encode more differences between AD and RNE
in word repetition trials. To minimize type I error, we utilized PCA to reduce the number of
input metrics used for classification. Both novel and previously studied features appeared
in the top two components of our PCA loading table (Table 4). The most common features
were global efficiency, density, TSP cost and GIC. Two of these features, namely TSP and
density, are from the six novel ones we introduced. The presence of these features generally
points to a difference in EEG time series structure between groups, especially with regards
to voltage differences and overall structure differences in the waveforms. We note that min
cut size, max clique size and independence number also appear in Table 4, indicating that
five out of six of the novel features we introduced are important for prediction.
Learned graph features, representing group differences in the morphology of EEG
time series, may reflect AD pathological changes in the neural generators of ERPs, including
N400 and P600. Putative N400 generators have been found in the anterior fusiform gyri and
other temporal cortical regions [67,68]. The primary neural generators of the P600 word
repetition effect were localized by functional MRI to the hippocampus, parahippocampal
gyrus, cingulate, left inferior parietal cortex and inferior frontal gyrus [69,70]. Extended
synaptic failure in these regions due to AD pathology may account for the N400 and P600
abnormalities in AD and prodromal AD patients. For example, abnormal memory-related
P600 may be associated with tau load in the medial temporal lobe (MTL), including the
hippocampus, entorhinal and perirhinal cortices, based on the evidence that early tau
accumulation in these regions correlates with lower memory performance and reductions
in functional connectivity between the MTL and cortical memory systems [71].
Using raw and bandpass filtered EEG data, we find that the δ band produced the
largest number of features, closely followed by the α band. Neural oscillations in differ-
ent frequency bands are thought to carry different spatial and temporal dimensions of
brain integration. Spatially, slow oscillations integrate large neural networks whereas
fast oscillations synchronize local networks [72]. Temporally, slow neural fluctuations are
related to the accumulation of information over long timescales across higher order cortical
regions [73]. In line with these hypotheses, empirical evidence has indicated that slow
oscillations in the delta range are important for higher cognitive functions that require
large-scale information integration (see Güntekin and Başar [74] for a review). Delta activity
has been shown to play important roles in language comprehension such as chunking
words into meaningful syntactic phrases [75]. Slow wave activity (SWA) also facilitates
memory consolidation during sleep by orchestrating fast oscillations across multiple brain
regions [76]. It may therefore be hypothesized that cognitive impairments in AD are re-
lated to alterations in slow oscillatory activity. Accumulating evidence has supported this
hypothesis, showing that decreased delta responses following cognitive stimulation may
serve as a general electrophysiological marker of cognitive dysfunction including MCI and
AD [74]. The present findings add to this line of research showing that the patterns of slow
EEG fluctuations, as characterized by VG features, reflect neural/cognitive abnormalities
in AD. Specific to this word repetition paradigm, Xia et al. [77] has shown that the vast
majority of the memory-related P600 word repetition effect is mediated by slow oscillations
in the delta band. Modulation of alpha band power, in comparison, is associated with se-
mantic processing of congruous and incongruous words. Alpha suppression was found to
be greater for New than for Old words [10]. The P600 (delta activity) and alpha suppression
effects reflect different aspects of verbal memory processing, and each uniquely contributes
to predicting individual verbal memory performance [77].
An interesting finding in the present study is that the Old Congruous condition (words
that are semantically congruous to the preceding category statements on repeated trials)
produces the highest number of features. Our previous ERP studies and many behavioral
studies have shown that old words are processed very differently from new words in
normal elderly, due to their intact memory function, but much less so in AD patients. EEG
channels producing the highest number of features were Fz, F8, R41, Pz, Br, Wl and O1. In
Brain Sci. 2023, 13, 770 16 of 21
the PCA comparision in Figure 4 and Table 4, we see this trend continue across even the
different comparisons (all classes, pAD vs. RNE, AD vs. RNE). Several of these channels
are known to be sensitive to word repetition and congruity manipulations in pAD patients.
For example, the N400 brain potential usually becomes smaller when an incongruous word
is repeated, i.e., the N400 repetition effect, and the effect is typically largest over midline
and right posterior channels including Cz, Pz, Wr, R41 and T6 [7,8,13]. The P600 ERP
usually becomes smaller when a congruous word is repeated, i.e., the P600 congruous
repetition effect, and the effect is widespread and largest over the midline channels with a
peak typically near Pz [7,8,13]. These ERP repetition effects are consistently found to be
reduced or abnormal in MCI patients [7,8], and severely diminished in AD patients [13]
compared to RNE, although they still appear in our comparison. The consistency across
studies in channel locations where group differences were found suggests that the VG
features may capture the underlying brain mechanisms related to the ERP repetition effects.
We now list strengths and limitations of our study. One of the strengths of our study is
our 100% accuracy with all classifiers on AD vs. RNE which demonstrates the effectiveness
of the features our method extracts. Linear separability after PCA implies that, even before
dimension reduction, AD vs. RNE is still a linearly separable comparison; indeed, Figure 4
explicitly demonstrates this. Additionally, classification accuracy of 92.5% on pAD vs.
RNE with non-linear neural networks and similar accuracies with linear classifiers using
only the features extracted from AD vs. RNE highlights how these generalizable features
alone may be sufficient for high-accuracy, near-linear classification of these two groups
that remains competitive with other EEG-based published work which explicitly extract
features for pAD vs. RNE classification [60–66]. This strength also likely comes from
looking at word repetition EEG tasks which have been shown to be sensitive to detecting
MCI-to-AD conversion and preclinical AD using ERPs [7,8,11]. Furthermore, our code is
open source and linked in the paper so that future work can build upon our strong results
and apply it to other datasets and tasks.
A potential limitation of the present study is the down-sampling procedure used
for data reduction. Averaging EEG data across non-overlapping 80 ms time windows is
effectively similar to lowpass filtering the data to 12.5 Hz, which would have reduced
the amount of information in higher frequencies including beta band and above. This
procedure most likely limited our ability to find discriminative VG features in these higher
frequency bands. It is also worth noting that, in the present study, we used EEG time series
averaged across trials for VG conversion. Cross-trial averaging is commonly used in ERP
analyses to increase the signal-to-noise ratio in EEG data and extract activity that is evoked
by, and phase-locked to, experimental stimuli. This averaging procedure, although highly
effective as demonstrated in the present study, ignores EEG activity that is related, but
not phase-locked, to the stimuli. With greater computing power, it would be valuable for
future studies to identify discriminative VG features from higher frequency bands and
non-phase-locked activity.
Another limitation is the small sample size used in our classification tests, feature
extraction and statistical analysis (15 AD, 15 pAD, 11 RNE). We mitigate this issue in two
ways: (1) we report classification scores as an average of 100 trials of training on 85% of
the data and testing on 15% (and only reporting the testing accuracy), and (2) we verify
the feature extraction step by only using AD vs. RNE features to classify pAD patients,
demonstrating generalization of those features. Despite this, further replication of these
results on larger datasets would be beneficial to the field. In such studies, it could be useful
to perform data augmentation, reduce model bias by imposing some penalties during
training (e.g., weight decay, dropout, etc.) or try different network architectures (such
as graph neural networks) to achieve even better generalization results. An additional
limitation is that we did not require amyloid biomarker studies in the definition of our
clinically defined subject groups, who were well-characterized by expert clinicians and
longitudinal cognitive testing.
Brain Sci. 2023, 13, 770 17 of 21
In summary, this paper extends the results of prior studies on the use of visibility
graphs for finding distinguishing features between and classifying Alzheimer’s and RNE
groups [15,16] to word repetition tasks on both AD and pAD with a novel set of features.
Distinguishing between pAD and RNE groups has historically produced poorer classifica-
tion accuracy in the literature; however, this paper provides novel features for this type of
classification that discriminates between pAD and RNE with competitive accuracy on our
dataset (92.5%) simply by generalizing AD vs. RNE features. Although we achieve perfect
100% accuracy on the AD vs. RNE task and demonstrate its generalization, a larger study
with a much larger sample size is still required to verify the efficacy of our framework.
Because all of the code is open source, this experiment can be readily applied to much
larger datasets; future applications could include predictors of conversion in MCI and
discriminate between different dementia pathologies. In future work, we plan to apply
our framework to larger AD and MCI datasets, and also to test similar frameworks in
preclinical AD.
Author Contributions: Conceptualization, J.Z., J.X. and J.O.; methodology, J.Z. and J.X.; software,
J.Z. and J.X; validation, J.Z. and J.X.; formal analysis, J.X., X.L. and J.O.; investigation, J.Z. and J.X.;
resources, J.O.; data curation, J.O.; writing—original draft preparation, J.Z. and J.X.; writing—review
and editing, J.X., X.L. and J.O.; visualization, J.Z.; supervision, J.O.; project administration, J.O.;
funding acquisition, J.O. All authors have read and agreed to the published version of the manuscript.
Funding: This research was funded by National Institute of Health grants RO1-AG048252, RO1-
AG18442, RO1-AG08313, P30 AG10129, and P30 AG062429.
Institutional Review Board Statement: Not applicable as human data was obtained in prior studies.
Informed Consent Statement: Not applicable as human data was obtained in prior studies.
Data Availability Statement: The datasets presented in this article are not readily available because
they may contain identifying information and are used with permissions from the Alzheimer’s
Disease Research Centers at the University of California, San Diego and the University of California,
Davis. Requests to access the datasets should be directed to John Olichney, M.D.
Conflicts of Interest: The authors declare that the research was conducted in the absence of any
commercial or financial relationships that could be construed as a potential conflict of interest.
Table A1. Eight-Fold CV Classification Statistics. Mean classification statistics and standard devi-
ations for all classifiers on both classification tasks. Results are averaged over validation accuracies
across all 8 folds. The rounded best performance across each column for each classification type is
bolded. AUC = Area Under (the ROC) Curve.
Table A2. Eight-Fold Classification Results with only Features from Prior Work. Mean classification
statistics and standard deviations for all classifiers on both classification tasks, with only the six
features from prior work: CCSS, clustering coefficient, global efficiency, local efficiency, small-
worldness and GIC. Results are averaged over validation accuracies across all 8 folds. The rounded
best performance across each column for each classification type is bolded. AUC = Area Under (the
ROC) Curve.
References
1. Selkoe, D.J. Alzheimer’s disease is a synaptic failure. Science 2002, 298, 789–791. [CrossRef] [PubMed]
2. Nunez, P.L.; Srinivasan, R. A theoretical basis for standing and traveling brain waves measured with human EEG with
implications for an integrated consciousness. Clin. Neurophysiol. 2006, 117, 2424–2435. [CrossRef] [PubMed]
3. Weintraub, S.; Wicklund, A.H.; Salmon, D.P. The neuropsychological profile of Alzheimer disease. Cold Spring Harb. Perspect.
Med. 2012, 2, a006171. [CrossRef]
4. Lizio, R.; Vecchio, F.; Frisoni, G.B.; Ferri, R.; Rodriguez, G.; Babiloni, C. Electroencephalographic Rhythms in Alzheimer’s Disease.
Int. J. Alzheimer’S Dis. 2011, 2011, 927573. [CrossRef] [PubMed]
5. Olichney, J.M.; Yang, J.C.; Taylor, J.; Kutas, M. Cognitive event-related potentials: Biomarkers of synaptic dysfunction across the
stages of Alzheimer’s disease. J. Alzheimer’S Dis. 2011, 26, 215–228. [CrossRef]
6. Yener, G.G.; Başar, E. Biomarkers in Alzheimer’s disease with a special emphasis on event-related oscillatory responses. In
Supplements to Clinical neurophysiology; Elsevier: Amsterdam, The Netherlands, 2013; Volume 62, pp. 237–273.
7. Olichney, J.M.; Morris, S.K.; Ochoa, C.; Salmon, D.P.; Thal, L.J.; Kutas, M.; Iragui, V.J. Abnormal verbal event related potentials in
mild cognitive impairment and incipient Alzheimer’s disease. J. Neurol. Neurosurg. Psychiatry 2002, 73, 377–384. [CrossRef]
8. Olichney, J.M.; Taylor, J.R.; Gatherwright, J.; Salmon, D.P.; Bressler, A.J.; Kutas, M.; Iragui-Madoz, V.J. Patients with MCI and
N400 or P600 abnormalities are at very high risk for conversion to dementia. Neurology 2008, 70, 1763–1770. [CrossRef]
9. Olichney, J.M.; Pak, J.; Salmon, D.P.; Yang, J.C.; Gahagan, T.; Nowacki, R.; Hansen, L.; Galasko, D.; Kutas, M.; Iragui-Madoz, V.J.
Abnormal P600 word repetition effect in elderly persons with preclinical Alzheimer’s disease. Cogn. Neurosci. 2013, 4, 143–151.
[CrossRef]
10. Mazaheri, A.; Segaert, K.; Olichney, J.; Yang, J.C.; Niu, Y.Q.; Shapiro, K.; Bowman, H. EEG oscillations during word processing
predict MCI conversion to Alzheimer’s disease. Neuroimage Clin. 2018, 17, 188–197. [CrossRef]
11. Taylor, J.R.; Olichney, J.M. From Amnesia to Dementia: ERP Studies of Memory and Language. Clin. EEG Neurosci. 2007, 38, 8–17.
[CrossRef]
12. Olichney, J.; Petten, C.; Paller, K.; Salmon, D.; Iragui, V.; Kutas, M. Word repetition in amnesia: Electrophysiological measures of
impaired and spared memory. Brain J. Neurol. 2000, 123, 1948–1963. [CrossRef] [PubMed]
Brain Sci. 2023, 13, 770 19 of 21
13. Olichney, J.M.; Iragui, V.J.; Salmon, D.P.; Riggins, B.R.; Morris, S.K.; Kutas, M. Absent event-related potential (ERP) word
repetition effects in mild Alzheimer’s disease. Clin. Neurophysiol. 2006, 117, 1319–1330. [CrossRef] [PubMed]
14. Lacasa, L.; Luque, B.; Ballesteros, F.; Luque, J.; Nuño, J.C. From time series to complex networks: The visibility graph. Proc. Natl.
Acad. Sci. USA 2008, 105, 4972–4975. [CrossRef] [PubMed]
15. Wang, J.; Yang, C.; Wang, R.; Yu, H.; Cao, Y.; Liu, J. Functional brain networks in Alzheimer’s disease: EEG analysis based on
limited penetrable visibility graph and phase space method. Phys. A Stat. Mech. Its Appl. 2016, 460, 174–187. [CrossRef]
16. Ahmadlou, M.; Adeli, H.; Adeli, A. New diagnostic EEG markers of the Alzheimer’s disease using visibility graph. J. Neural
Transm. 2010, 117, 1099–1109. [CrossRef]
17. Morabito, F.C.; Campolo, M.; Ieracitano, C.; Ebadi, J.M.; Bonanno, L.; Bramanti, A.; Desalvo, S.; Mammone, N.; Bramanti, P. Deep
convolutional neural networks for classification of mild cognitive impaired and Alzheimer’s disease patients from scalp EEG
recordings. In Proceedings of the 2016 IEEE 2nd International Forum on Research and Technologies for Society and Industry
Leveraging a better tomorrow (RTSI), Bologna, Italy, 7–9 September 2016; pp. 1–6. [CrossRef]
18. Zhang, X.; Yao, L.; Wang, X.; Monaghan, J.; McAlpine, D.; Zhang, Y. A survey on deep learning-based non-invasive brain signals:
Recent advances and new frontiers. J. Neural Eng. 2021, 18, 031002. [CrossRef]
19. Zhao, Y.; He, L. Deep Learning in the EEG Diagnosis of Alzheimer’s Disease. In Proceedings of the Computer Vision—ACCV 2014
Workshops, Singapore, 1–2 November 2014; Jawahar, C., Shan, S., Eds.; Springer International Publishing: Cham, Switzerland, 2015;
pp. 340–353.
20. Duan, F.; Huang, Z.; Sun, Z.; Zhang, Y.; Zhao, Q.; Cichocki, A.; Yang, Z.; Solé-Casals, J. Topological Network Analysis of Early
Alzheimer’s Disease Based on Resting-State EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 2164–2172. [CrossRef]
21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385.
22. Chang, C.H.; Lin, C.H.; Lane, H.Y. Machine Learning and Novel Biomarkers for the Diagnosis of Alzheimer’s Disease. Int. J. Mol.
Sci. 2021, 22, 2761. [CrossRef]
23. McKhann, G.M.; Knopman, D.S.; Chertkow, H.; Hyman, B.T.; Jack, C.R., Jr.; Kawas, C.H.; Klunk, W.E.; Koroshetz, W.J.; Manly, J.J.;
Mayeux, R.; et al. The diagnosis of dementia due to Alzheimer’s disease: Recommendations from the National Institute
on Aging-Alzheimer’s Association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimer’s Dement. 2011,
7, 263–269. [CrossRef]
24. Dubois, B.; Villain, N.; Frisoni, G.; Rabinovici, G.D.; Sabbagh, M.N.; Cappa, S.; Bejanin, A.; Bombois, S.; Epelbaum, S.;
Teichmann, M.; et al. Recommendations of the International Working Group for the clinical diagnosis of Alzheimer’s disease.
Alzheimer’s Dement. 2021, 17, e057538. [CrossRef]
25. Jack, C.R.; Bennett, D.A.; Blennow, K.; Carrillo, M.C.; Dunn, B.; Haeberlein, S.B.; Holtzman, D.M.; Jagust, W.; Jessen, F.;
Karlawish, J.; et al. NIA-AA Research Framework: Toward a biological definition of Alzheimer’s disease. Alzheimer’s Dement.
2018, 14, 535–562. [CrossRef] [PubMed]
26. Albert, M.S.; DeKosky, S.T.; Dickson, D.; Dubois, B.; Feldman, H.H.; Fox, N.C.; Gamst, A.; Holtzman, D.M.; Jagust, W.J.;
Petersen, R.C.; et al. The diagnosis of mild cognitive impairment due to Alzheimer’s disease: Recommendations from the
National Institute on Aging-Alzheimer’s Association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimer’s
Dement. 2011, 7, 270–279. [CrossRef] [PubMed]
27. Tăuţan, A.M.; Ionescu, B.; Santarnecchi, E. Artificial intelligence in neurodegenerative diseases: A review of available tools with
a focus on machine learning techniques. Artif. Intell. Med. 2021, 117, 102081. [CrossRef] [PubMed]
28. Gainotti, G.; Quaranta, D.; Vita, M.G.; Marra, C. Neuropsychological predictors of conversion from mild cognitive impairment to
Alzheimer’s disease. J. Alzheimer’S Dis. 2014, 38, 481–495. [CrossRef]
29. Petersen, R.C.; Smith, G.E.; Waring, S.C.; Ivnik, R.J.; Tangalos, E.G.; Kokmen, E. Mild Cognitive Impairment: Clinical Characteri-
zation and Outcome. Arch. Neurol. 1999, 56, 303–308. [CrossRef]
30. Roalf, D.R.; Moberg, P.J.; Xie, S.X.; Wolk, D.A.; Moelter, S.T.; Arnold, S.E. Comparative accuracies of two common screening
instruments for classification of Alzheimer’s disease, mild cognitive impairment, and healthy aging. Alzheimer’s Dement. 2013,
9, 529–537. [CrossRef]
31. Petersen, R.C. Mild cognitive impairment as a diagnostic entity. J. Intern. Med. 2004, 256, 183–194. [CrossRef]
32. APA. Diagnostic and Statistical Manual of mental Disorders; American Psychiatric Association: Washington, DC, USA, 2000.
33. Francis, W.N.; Henry, K.; Mackie, A.W. Frequency Analysis of English Usage: Lexicon and Grammar; Houghton Mifflin: Boston, MA,
USA, 1982.
34. MATLAB Optimization Toolbox; The MathWorks: Natick, MA, USA, 2012.
35. Delorme, A.; Makeig, S. {EEGLAB}: {A}n Open Source Toolbox for Analysis of Single-Trial {EEG} Dynamics Including Independent
Component Analysis. J. Neurosci. Methods 2004, 134, 9–21. [CrossRef]
36. Oostenveld, R.; Fries, P.; Maris, E.; Schoffelen, J.M. FieldTrip: Open Source Software for Advanced Analysis of MEG, EEG, and
Invasive Electrophysiological Data. Comput. Intell. Neurosci. 2011, 2011, 156869. [CrossRef]
37. Jung, T.; Makeig, S.; McKeown, M.J.; Bell, A.J.; Lee, T.; Sejnowski, T.J. Imaging brain dynamics using independent component
analysis. Proc. IEEE 2001, 89, 1107–1122. [CrossRef] [PubMed]
38. Kabbara, A.; Falou, W.E.; Khalil, M.; Eid, H.; Hassan, M. A scalp-EEG network-based analysis of Alzheimer’s disease patients at
rest. In Proceedings of the 2017 Fourth International Conference on Advances in Biomedical Engineering (ICABME), Beirut,
Lebanon, 19–21 October 2017; pp. 1–4. [CrossRef]
Brain Sci. 2023, 13, 770 20 of 21
39. de Haan, W.; Pijnenburg, Y.A.L.; Strijers, R.L.M.; van der Made, Y.; van der Flier, W.M.; Scheltens, P.; Stam, C.J. Functional neural
network analysis in frontotemporal dementia and Alzheimer’s disease using EEG and graph theory. BMC Neurosci. 2009, 10, 101.
[CrossRef] [PubMed]
40. Wang, L.; Long, X.; Arends, J.B.; Aarts, R.M. EEG analysis of seizure patterns using visibility graphs for detection of generalized
seizures. J. Neurosci. Methods 2017, 290, 85 – 94. . [CrossRef] [PubMed]
41. Wang, R.; Yang, Z.; Wang, J.; Shi, L. An Improved Visibility Graph Analysis of EEG Signals of Alzheimer Brain. In Proceedings of
the 2018 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI),
Beijing, China, 13–15 October 2018; pp. 1–5. [CrossRef]
42. Oliphant, T. NumPy: A Guide to NumPy; Continuum Press: New York City, NY, USA, 2015.
43. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.;
Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [CrossRef]
44. Hagberg, A.A.; Schult, D.A.; Swart, P.J. Exploring Network Structure, Dynamics, and Function using NetworkX. In Proceedings
of the 7th Python in Science Conference, Pasadena, CA, USA, 19–24 August 2008; Varoquaux, G., Vaught, T., Millman, J., Eds.;
Los Alamos National Lab.(LANL): Los Alamos, NM, USA, 2008; pp. 11–15.
45. Ahmadlou, M.; Adeli, H. Visibility graph similarity: A new measure of generalized synchronization in coupled dynamic systems.
Phys. D Nonlinear Phenom. 2012, 241, 326 – 332. [CrossRef]
46. Watts, D.; Strogatz, S.H. Collective Dynamics of Small World Networks. Nature 1998, 393, 440–2. [CrossRef]
47. Latora, V.; Marchiori, M. Efficient Behavior of Small-World Networks. Phys. Rev. Lett. 2001, 87, 198701. [CrossRef]
48. Rubinov, M.; Sporns, O. Complex network measures of brain connectivity: Uses and interpretations. NeuroImage 2010,
52, 1059–1069. Computational Models of the Brain. [CrossRef]
49. Stam, C.J.; Reijneveld, J.C. Graph theoretical analysis of complex networks in the brain. Nonlinear Biomed. Phys. 2007, 1, 3.
[CrossRef]
50. Harrington, D.L.; Rubinov, M.; Durgerian, S.; Mourany, L.; Reece, C.; Koenig, K.; Bullmore, E.; Long, J.D.; Paulsen, J.S.; for the
PREDICT-HD investigators of the Huntington Study Group; et al. Network topology and functional connectivity disturbances
precede the onset of Huntington’s disease. Brain 2015, 138, 2332–2346. [CrossRef]
51. Humphries, M.D.; Gurney, K. Network ‘Small-World-Ness’: A Quantitative Method for Determining Canonical Network
Equivalence. PLoS ONE 2008, 3, e0002051. [CrossRef] [PubMed]
52. Erdös, P.; Rényi, A. On Random Graphs I. Publ. Math. Debr. 1959, 6, 290. [CrossRef]
53. Kim, J.; Wilhelm, T. What is a complex graph? Phys. A Stat. Mech. Its Appl. 2008, 387, 2637–2652. [CrossRef]
54. Dasgupta, S.; Papadimitrious, C.; Vazirani, U. Algorithms; McGraw-Hill Education: New York, NY, USA, 2006.
55. Boppana, R.; Halldórsson, M.M. Approximating maximum independent sets by excluding subgraphs. BIT Numer. Math. 1992,
32, 180–196. [CrossRef]
56. Papadimitriou, C.H. The Euclidean travelling salesman problem is NP-complete. Theor. Comput. Sci. 1977, 4, 237–244. [CrossRef]
57. van Wijk, B.C.M.; Stam, C.J.; Daffertshofer, A. Comparing Brain Networks of Different Size and Connectivity Density Using
Graph Theory. PLoS ONE 2010, 5, e13701. [CrossRef] [PubMed]
58. Bollobás, B. The Independence Ratio of Regular Graphs. Proc. Am. Math. Soc. 1981, 83, 433–436. [CrossRef]
59. V. Kosowski, A.; Manuszewski, K. Classical coloring of graphs. Contemp. Math. 2004.
60. Cichocki, A.; Shishkin, S.L.; Musha, T.; Leonowicz, Z.; Asada, T.; Kurachi, T. EEG filtering based on blind source separation (BSS)
for early detection of Alzheimer’s disease. Clin. Neurophysiol. 2005, 116, 729–737. [CrossRef]
61. Fiscon, G.; Weitschek, E.; Felici, G.; Bertolazzi, P.; De Salvo, S.; Bramanti, P.; De Cola, M.C. Alzheimer’s disease patients
classification through EEG signals processing. In Proceedings of the 2014 IEEE Symposium on Computational Intelligence and
Data Mining (CIDM), Orlando, FL, USA, 9–12 December 2014; pp. 105–112. [CrossRef]
62. Chupin, M.; Gérardin, E.; Cuingnet, R.; Boutet, C.; Lemieux, L.; Lehéricy, S.; Benali, H.; Garnero, L.; Colliot, O. Fully automatic
hippocampus segmentation and classification in Alzheimer’s disease and mild cognitive impairment applied on data from ADNI.
Hippocampus 2009, 19, 579–587. [CrossRef]
63. Liu, J.; Li, M.; Lan, W.; Wu, F.; Pan, Y.; Wang, J. Classification of Alzheimer’s Disease Using Whole Brain Hierarchical Network.
IEEE/ACM Trans. Comput. Biol. Bioinform. 2018, 15, 624–632. [CrossRef] [PubMed]
64. Li, Q.; Wu, X.; Xu, L.; Chen, K.; Yao, L.; Alzheimer’s Disease Neuroimaging Initiative. Classification of Alzheimer’s Disease, Mild
Cognitive Impairment, and Cognitively Unimpaired Individuals Using Multi-feature Kernel Discriminant Dictionary Learning.
Front. Comput. Neurosci. 2018, 11, 117. [CrossRef]
65. Fiscon, G.; Weitschek, E.; Cialini, A.; Felici, G.; Bertolazzi, P.; De Salvo, S.; Bramanti, A.; Bramanti, P.; De Cola, M.C. Combining
EEG signal processing with supervised methods for Alzheimer’s patients classification. BMC Med. Inform. Decis. Mak. 2018,
18, 35. [CrossRef] [PubMed]
66. Ieracitano, C.; Mammone, N.; Bramanti, A.; Hussain, A.; Morabito, F.C. A Convolutional Neural Network approach for
classification of dementia stages based on 2D-spectral representation of EEG recordings. Neurocomputing 2019, 323, 96–107.
[CrossRef]
67. McCarthy, G.; Nobre, A.; Bentin, S.; Spencer, D. Language-related field potentials in the anterior-medial temporal lobe: I.
Intracranial distribution and neural generators. J. Neurosci. 1995, 15, 1080–1089. [CrossRef] [PubMed]
Brain Sci. 2023, 13, 770 21 of 21
68. Halgren, E.; Dhond, R.P.; Christensen, N.; Van Petten, C.; Marinkovic, K.; Lewine, J.D.; Dale, A.M. N400-like Magnetoen-
cephalography Responses Modulated by Semantic Context, Word Frequency, and Lexical Class in Sentences. NeuroImage 2002,
17, 1101–1116. [CrossRef]
69. Olichney, J.M.; Taylor, J.R.; Hillert, D.G.; hui Chan, S.; Salmon, D.P.; Gatherwright, J.; Iragui, V.J.; Kutas, M. fMRI congruous word
repetition effects reflect memory variability in normal elderly. Neurobiol. Aging 2010, 31, 1975–1990. . [CrossRef]
70. Olichney, J.M.; Taylor, J.R.; Chan, S.; Yang, J.C.; Stringfellow, A.; Hillert, D.G.; Simmons, A.L.; Salmon, D.P.; Iragui-Madoz,
V.; Kutas, M. fMRI responses to words repeated in a congruous semantic context are abnormal in mild Alzheimer’s disease.
Neuropsychologia 2010, 48, 2476–2487. . [CrossRef]
71. Berron, D.; Vogel, J.W.; Insel, P.S.; Pereira, J.B.; Xie, L.; Wisse, L.E.M.; Yushkevich, P.A.; Palmqvist, S.; Mattsson-Carlgren, N.;
Stomrud, E.; et al. Early stages of tau pathology and its associations with functional connectivity, atrophy and memory. Brain
2021, 144, 2771–2783. [CrossRef]
72. Buzsáki, G.; Draguhn, A. Neuronal oscillations in cortical networks. Science 2004, 304, 1926–1929. [CrossRef]
73. Honey, C.J.; Thesen, T.; Donner, T.H.; Silbert, L.J.; Carlson, C.E.; Devinsky, O.; Doyle, W.K.; Rubin, N.; Heeger, D.J.; Hasson, U.
Slow cortical dynamics and the accumulation of information over long timescales. Neuron 2012, 76, 423–434. [CrossRef] [PubMed]
74. Güntekin, B.; Başar, E. Review of evoked and event-related delta responses in the human brain. Int. J. Psychophysiol. 2016,
103, 43–52. [CrossRef] [PubMed]
75. Meyer, L. The neural oscillations of speech processing and language comprehension: State of the art and emerging mechanisms.
Eur. J. Neurosci. 2018, 48, 2609–2621. [CrossRef]
76. Rasch, B.; Born, J. About sleep’s role in memory. Physiol. Rev. 2013, 93, 681–766. [CrossRef] [PubMed]
77. Xia, J.; Mazaheri, A.; Segaert, K.; Salmon, D.; Harvey, D.; Shapiro, K.; Kutas, M.; Olichney, J. Event-related potential and EEG
oscillatory predictors of verbal memory in mild cognitive impairment. Brain Commun. 2020, 2, fcaa213. [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.