2018 Book BiomedicalEngineeringSystemsAn
2018 Book BiomedicalEngineeringSystemsAn
Margarida Silveira
Hesham H. Ali
Carlos Maciel
Egon L. van den Broek (Eds.)
Biomedical Engineering
Systems and Technologies
10th International Joint Conference, BIOSTEC 2017
Porto, Portugal, February 21–23, 2017
Revised Selected Papers
123
Communications
in Computer and Information Science 881
Commenced Publication in 2007
Founding and Former Series Editors:
Alfredo Cuzzocrea, Xiaoyong Du, Orhun Kara, Ting Liu, Dominik Ślęzak,
and Xiaokang Yang
Editorial Board
Simone Diniz Junqueira Barbosa
Pontifical Catholic University of Rio de Janeiro (PUC-Rio),
Rio de Janeiro, Brazil
Phoebe Chen
La Trobe University, Melbourne, Australia
Joaquim Filipe
Polytechnic Institute of Setúbal, Setúbal, Portugal
Igor Kotenko
St. Petersburg Institute for Informatics and Automation of the Russian
Academy of Sciences, St. Petersburg, Russia
Krishna M. Sivalingam
Indian Institute of Technology Madras, Chennai, India
Takashi Washio
Osaka University, Osaka, Japan
Junsong Yuan
University at Buffalo, The State University of New York, Buffalo, USA
Lizhu Zhou
Tsinghua University, Beijing, China
More information about this series at http://www.springer.com/series/7899
Nathalia Peixoto Margarida Silveira
•
Biomedical Engineering
Systems and Technologies
10th International Joint Conference, BIOSTEC 2017
Porto, Portugal, February 21–23, 2017
Revised Selected Papers
123
Editors
Nathalia Peixoto Carlos Maciel
George Mason University University of Sao Paulo
Fairfax, VA Sao Carlos, SP
USA Brazil
Margarida Silveira Egon L. van den Broek
Instituto Superior Técnico (IST) Utrecht University
University of Lisbon Utrecht
Lisbon The Netherlands
Portugal
Hesham H. Ali
University of Nebraska at Omaha
Omaha, NE
USA
This Springer imprint is published by the registered company Springer International Publishing AG
part of Springer Nature
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
The present book includes extended and revised versions of a set of selected papers
from the 10th International Joint Conference on Biomedical Engineering Systems and
Technologies (BIOSTEC 2017), held in Porto, Portugal, during February 21–23, 2017.
BIOSTEC is composed of five co-located conferences, each specialized in a different
knowledge area, namely, BIODEVICES, BIOIMAGING, BIOINFORMATICS, BIO-
SIGNALS, and HEALTHINF.
BIOSTEC 2017 received 297 paper submissions from 56 countries, of which only
6% are included in this book. This reflects our care in selecting contributions. These
papers were selected by the conference chairs on the basis of a number of criteria that
include the classifications and comments provided by the Program Committee mem-
bers, the session chairs’ assessment, and the program chairs’ meta-review of the papers
that were included in the technical program. The authors of selected papers were
invited to submit a revised, extended, and improved version of their conference paper,
including at least 30% new material.
The purpose of the BIOSTEC joint conferences is to bring together researchers and
practitioners, including engineers, biologists, health professionals, and
informatics/computer scientists. Research presented at BIOSTEC included both theo-
retical advances and applications of information systems, artificial intelligence, signal
processing, electronics, and other engineering tools in areas related to advancing
biomedical research and improving health care.
The papers included in this book contribute to the understanding of relevant research
trends in biomedical engineering systems and technologies. As such, they provide an
overview of the field’s current state of the art.
We thank the authors for their contributions and Monica Saramago for process
management. In particular, we express our gratitude to the reviewers, who helped to
ensure the quality of this publication.
Conference Co-chairs
Ana Fred Instituto de Telecomunicações/IST, Portugal
Hugo Gamboa LIBPHYS-UNL/FCT - New University of Lisbon,
Portugal
Mário Vaz INEGI LOME, FEUP, Portugal
Program Co-chairs
BIODEVICES
Nathalia Peixoto Neural Engineering Lab, George Mason University,
USA
BIOIMAGING
Margarida Silveira Instituto Superior Técnico (IST), Portugal
BIOINFORMATICS
Hesham Ali University of Nebraska at Omaha, USA
BIOSIGNALS
Carlos Maciel University of São Paulo, Brazil
HEALTHINF
Egon L. van den Broek Utrecht University, The Netherlands
Invited Speakers
Bart M. ter Haar Romeny Eindhoven University of Technology (TU/e),
The Netherlands
Kristina Höök Royal Institute of Technology, Sweden
Bethany Bracken Charles River Analytics Inc., USA
Hugo Plácido da Silva IT, Institute of Telecommunications, Portugal
Contents
Bioimaging
Health Informatics
Abstract. The median sternotomy can rise in rib and/or sternum micro/
macro-fractures and/or brachial plexus injuries, which can even evolve in
chronic pain with significant impact on patient’s quality life. Post-sternotomy
chronic pain is recognized as a multifactorial complex issue, and it has been
assessed that sternum retraction forces, applied by the surgeons, can be con-
sidered one of these factors. In order to investigate the behavior of these forces,
we developed a reliable and sterilizable system, to monitor the retraction forces
along the hemisternums. Therefore, a Finochietto sternal retractor was instru-
mented by means of ultra-thin force sensors interfaced with ad hoc electronic
circuitry. Two different sets of sensors were adopted, one of which able to
support autoclave operating conditions. In-vitro tests were performed by means
of a made on purpose dummy. The instrumented retractor allows monitoring the
force exerted on both the arms during the opening procedure. Force versus time
patterns were acquired and stored, and so we determined how the forces are
distributed in terms of their mean, maximum and plateaus. Results demonstrate
the reliability of the instrumented retractor in measuring forces, up to 400 N.
Cost-effectiveness and feasibility can be considered further additional values of
the proposed instrumented retractor.
1 Introduction
The first median sternotomy has been performed in 1897, to remove lymph nodes.
Since then, only six decades after, the median sternotomy has become the standard
approach to the mediastinum, following Julian’s report [1] where the superiority over
thoracotomy was described underlining its less time-consuming, and higher tolerability
by the patients.
Briefly, to access the mediastinum through the median sternotomy, a skin incision
is made, approximately 2 cm under the sternal notch and extended below the xiphoid.
The exact midline over the sternum is marked with electrocautery to avoid faulty
sternotomy. Before the sternum incision is made, a pathway is create above the
suprasternal ligament and then continued beneath the manubrium and finally, per-
formed as well under the xiphoid to guarantee the separation of the mediastinum
structures from the posterior sternum bone.
Although, in comparison to extensive thoracotomy, midline sternotomy is less
traumatic, persistent postoperative pain remain the Achilles’ heel, affecting negatively
early postoperative respiratory function, delayed hospital discharge and increasing
costs [2–4]. The process of pain is difficult to assess [5], anyway the pain is considered
chronic when localized in the surgical site and persist over three months. In a
prospective study, a number of independent predictors for persistent thoracic pain
following sternotomy were identified, including urgent surgery, and re-sternotomy [6].
In this study, at one year, 42 (35%) patients reported chronic thoracic pain. Similarly,
another work reported the prevalence of post-operative pain as high as 39.3% at the
mean time of 28 months after surgery [7]. In 2001, Mazzeffi and Khelemsky [8]
estimated a 28% overall incidence of non-cardiac pain one year after surgery. Several
studies assessed that women are substantially more likely to suffer early and chronic
postoperative pain than men [6, 9], and that the prevalence of post-sternotomy chronic
pain decreases with age [6, 8]. These studies highlight the negative impact on daily life
of the population who experienced sternotomy and suffer postoperative pain. In fact
with the introduction of less invasive surgical procedures to treat cardiac pathologies,
although with limited surgical field, surgeons who are in favor claim that it is not only
cosmetic, but guaranteeing better and faster post-operative recovery.
Chronic post-sternotomy pain can be related to patient’s age, gender and degen-
erative process like osteoporosis. It can be also related directly to the surgical proce-
dure, including secondary sternal osteomyelitis and/or sternocostal chondritis,
incomplete bone healing. If the internal thoracic arteries are harvested in myocardial
revascularization, the different retractors, used to facilitated vessel exposure, add
additional trauma. In this contest the way of harvesting and the use of ellectrocautery
might affect wound healing and persistence pain.
To access to the mediastinum, retractors are being used to allow adequate surgical
field. Hemi-sternums separation extent of the force impressed during sternum opening
might lead to rib fracture, eventually associated to brachial plexus injury (BPI) [10–14].
Thus, there is an actual clinical need to provide to the surgeons suitable instrumented
retractors able to monitor in real time the forces exerted on the two halves during the
sternum opening procedure. In this way, by monitoring forces applied on the sternum,
it will be possible to find the balance between adequate surgical field and excessive
trauma to the chest. By evaluating major risk factors, width chest opening can be
tailored to the patients. For example if female gender is prone to chronic pain, chest
separation should be reduced to minimum.
Furthermore, with the increasing interest of shifting the cardiac surgery procedures
from full to partial sternotomy, including the “J” and “T” incisions, the proposed study
might be useful to evaluate and compare the forces applied on the sternum in the
An Electronic-Engineered Sensory Sternal Retractor Aimed 5
various surgical approaches to determine the best access, allowing at the same time the
optimal surgical view and successively good quality of life.
Only few data are available for the actual value of the forces exerted by a retractor
on the skeletal cage, dummy reproduced [15], or of corpses or animal models [16].
Aigner et al. [17] pointed out that data obtained from human patients are not presently
available in the literature probably due to the lack of an instrumented sternal retractor
readily suitable for the translation to surgery.
For this purpose, we designed and realized a sterilizable system based on a com-
monly adopted straight sternal retractor (Finocchietto) equipped with ultrathin force
sensors and conditioning electronic circuitry. The forces experienced during the
retraction were monitored in real-time by means of a home-made dummy.
The idea is to acquire data on the intensity and distribution of exerted retraction
forces during hemi-sternums separation in view of future challenging clinical studies
aimed to reduce the risk of chronic post-sternotomy pain.
2 Materials
A commonly adopted straight sternal retractor, Finochietto type (Fig. 1a), was equipped
with both ultra-thin force sensors (Fig. 1b and c) and electronic circuitry. The resulting
electronic-engineered “sensory retractor” was tested by means of a home-made dummy.
back
Fig. 1. (a) The Finochietto retractor equipped with the four sensors placed in positions designed
from 1 to 4 according to the figure. (b) The ultra-thin force sensors, HT201 (top) and A201
(bottom) types. (c) Aluminum sensors’ housings: front, back and cover [15].
6 G. Saggio et al.
Fig. 2. (a) LRX (Lloyd Instrument) testing machine used for sensor characterization under
compression, (b) and (c) Load cells used for sensor characterization.
An Electronic-Engineered Sensory Sternal Retractor Aimed 7
In such a way, the electric voltage value is referred to the force applied. The voltage
signal is amplified, calibrated and compensated with temperature, then processed by an
algorithm to correct the device nonlinearity. Consequently, the value of the applied
force was determined, taking into account deformation and material characteristics.
Two load cells of 50 N and 500 N, respectively (Fig. 2a and b) were used for sensor
characterization.
RP
VOUT ¼ VBIAS ð1Þ
RSENS þ RP
The values of each shunt resistors RP was determined taking into account some
needs. In particular, RP has to protect each sensor against excessive currents, and has to
guarantee the largest-as-possible output voltage swing so to allow adequate resolution
for the following digital conversion. With a force value ranging from 1 to 400 N, the
sensor resistance span of 1 MX roughly. Since the maximum allowable current for the
sensor is 2.5 mA, when the sensor is in short circuit, the RP value must be higher than
5V
RP [ ¼ 2 kX ð2Þ
2:5 mA
but correctly determined as the geometric mean of the extreme sensor resistance values,
in accordance with the findings in [19].
8 G. Saggio et al.
Fig. 3. (a) Adopted front-end circuits for the four FlexiForce A201 and HT201 sensors,
(b) front-end circuit for the FlexiForce A201 and HT201 sensors recommended by Tekscan.
5V
VLSB ¼ ’ 4:9 mV ð3Þ
210 1
The minimum voltage from the front-end circuit, in case of zero applied force,
calculated by Eq. (1) with RSENS = 1 MX, should be greater than VLSB
RP
VOUTmin ¼ 5 V [ 4:9 mV ð4Þ
1 MX þ RP
Table 1. Serial packet transmission for data acquisition from a single sensor device.
Table 2. Serial packet transmission for simultaneous data acquisition from four sensor devices.
Fig. 4. (a) The home made dummy built up using four gas pistons fixed to a wooden skeleton,
the compressible parts positioned outward in a face-to-face configuration. In vitro tests: (b) the
instrumented Finochietto retractor positioned into the dummy equipped with the four force
sensors (S1, S2, S3, S4).
3 Methods
3.1 Sensor Characterization
In this section the selected methods for static and dynamic characterization of the force
sensors will be deeply described. Both of them were accomplished with the LRX
testing machine with two load cells for compression up to 5 N and 500 N, respectively.
Due to the sensor thinness, the sensor sample was placed on a stainless steel platform,
and a 10 mm diameter steel disk was placed on the active area of the sensor, to achieve
the required pressure on device during the jack fall down. References on sensor
location help to replace the sample measurement in the same conditions.
Static Characterization. In the static characterization the descent rate of the jack was
set to 5 mm/min and the static measurements were acquired @5 N, 10 N, 20 N, 30 N
and 40 N with the 50 N load cell, and from 40 N to 400 N, step 40 N, with the 500 N
load cell. Two minutes break were set before each step to acquire measurements,
through the LabView-Arduino interface.
An Electronic-Engineered Sensory Sternal Retractor Aimed 11
Considering a baud rate of 19200 bps for 2 min or 120 s, 50 bit for each sensor
sample, the acquired resistance samples are (19200/50)∙120 = 46080. Waiting 52 s,
when the sensor response was considered enough stable, 3000 resistance samples were
considered for a duration of almost 7 s, of which the average and the standard deviation
were calculated. In the same time interval, 30 samples of the force magnitude were
considered among the only 384 samples stored by the test machine acquisition system
in 120 s, because the sampling frequency is much smaller than that of Luigino/
LabView interface.
Figure 5a and b shows the characterization of the HT201 and A201 devices,
respectively. The sensor resistance is represented as the average and standard deviation
among eight sensor samples as a function of the average compression force at each
step. Ultrathin flexible force sensors HT201 and A201 both showed exponential
resistance decay with the impressed force. Plots demonstrate the same behavior for
A201 and HT201 sensors, but in the force range 30–440 N standard deviations are very
small (7–42 kX, 8–51 kX respectively), whereas in the range 5–30 N standard devi-
ations become high (96–482 kX, 47–1000 kX respectively).
In order to investigate if HT201 sensors can effectively support autoclaving con-
ditions, these sensors were also characterized following the same procedure after five
cycles of autoclave treatment (VaporMatic 770, Asal Srl, Milan, Italy). HT201 sensors
did not show a significantly different behavior after five cycles of autoclave condi-
tioning, which is a reasonable result since these sensors have been specifically designed
for high temperature applications (up to 400 °F, approximately 200 °C). In any case, in
the occurrence of degradation in performances, those sensors can be easily and con-
veniently replaced.
Fig. 5. (a) Measured resistance versus force (R vs. F) for different sensor (a) HT201 type,
(b) A201 type.
and back for each speed rate, each time increasing and decreasing at a constant rate the
traction force from 0 N to 400 N and down again to 0 N. The repetition period depends
on the descent rate of the jack.
Results for the average sensor resistance against increasing force with different
speed, superimposed for comparison with two equal and symmetrical static charac-
teristics in Fig. 6a and b show A201 sensors behave with lower repeatability than
HT201 counterparts, changing the variation rate of the applied force.
Dynamic characterization was also repeated after five cycles of autoclave treatment
for HT201 sensors, and characteristics do not show a significantly different behavior
after treatment, as in the static case (so results are not presented for sake of brevity).
Fig. 6. Average sensor resistance against increasing force with different speed, superimposed
for comparison with two equal and symmetrical static characteristics for A201 (a) and HT201
(b) devices.
Aigner et al. assessed that the force distribution did not change significantly for the
other 3 retractions, for the different investigated spread widths (i.e. 5, 7.5, and 10 cm)
and was not gender dependent. The maximum force for full retraction was 493.6 N,
whereas the smallest maximum force was 159.0 N.
Fig. 7. (a) and (b) Response of the four sensors (housed as shown in Fig. 1a) in terms of force
[N] versus time [s] during the 5 cm opening procedure. (c) Response of the four sensors (housed
as showed in Fig. 1a) in terms of force [N] versus time [s] during the 10 cm opening procedure.
For example, the total average force for the mobile blade ranged between
60.1 N ± 6.0 N for 5 cm spread and 98.0 N ± 36.5 N for 10 cm spread, as expected
for a dummy built-up with 80 N gas pistons. The deviation with respect to this value is
also expected and is to be attributed to the uneven pressure distribution onto the circular
sensors due to the rough surface finishing of the contact area i.e. wood in the dummy.
An Electronic-Engineered Sensory Sternal Retractor Aimed 15
Table 3. Values of the mean, maximum and plateau forces (expressed in N) measured by
HT201 sensors positioned according to Fig. 1a (i.e. S1, S2, S3, S4). The related standard
deviation values are reported in parentheses [15].
Spread Force [N] S1 S2 S3 S4 S1+S2 fixed S3+S4
blade mobile blade
5 cm Mean 60.8 (5.7) 39.2 (4.4) 18.3 (2.1) 41.8 (7.7) 100.1 (8.5) 60.1 (6.1)
Max 97.3 (10.6) 115.0 (19.5) 62.7 (5.4) 156.4 (12.5) 212.2 (14.8) 219.1 (9.7)
Plateau 63.9 (5.8) 39.4 (4.9) 17.9 (1.8) 38.3 (8.3) 103.3 (9.0) 56.3 (7.2)
Max-mean 36.5 (8.7) 75.7 (21.1) 44.4 (4.0) 114.6 (12.9) 112.2 (12.5) 159.0 (12.6)
10 cm Mean 37.6 (8.8) 82.5 (5.60) 15.2 (10.4) 82.8 (26.8) 120.1 (12.8) 98.0 (36.5)
Max 79.6 (17.5) 126.3 (6.62) 66.9 (4.3) 199.7 (21.3) 205.9 (20.8) 266.6 (25.4)
Plateau 41.2 (6.7) 89.1 (7.02) 13.1 (12.9) 84.7 (32.0) 130.2 (11.7) 97.8 (44.4)
Max-mean 42.0 (14.1) 43.8 (6.98) 54.7 (9.4) 116.9 (16.9) 85.8 (13.6) 168.6 (26.3)
5 Conclusions
Median sternotomy is a surgical incision through the sternum, then after to allow access
to the mediastinum a retractor is positioned. Wide opening of the hemi-sternum, by
means of the retractor, guarantee better view and facilitate the surgical procedure.
However, its increase the stress on the sternum halves and ribs, leading to partial or
complete fractures and/or micro-fractures resulting in post-operative and chronic pain
in a non-negligible number of patients.
By measuring the forces during different opening procedures, we demonstrated
how it can be possible to understand and find the compromise between adequate
surgical field and the risk for sternum and ribs fractures, aiming at improving patients
postoperative coarse.
Within such a frame, this paper reports a new electronic-engineered sensory sternal
retractor aimed at measuring the forces impressed by the plates when opening a dummy
ribcage, so in an in-vitro median sternotomy condition.
We demonstrated that the impressed forces present “spikes”, i.e. sudden changes,
with peak force values meaningfully higher of the plateau values which, from a
mechanical point of view, can be the reason of the cracks/micro-cracks of the ribcage,
and so of the persistent postoperative pain suffered by a number of patients after the
surgical procedure.
References
1. Julian, O.C., Lopez-Belio, M., Dye, W.S., Javid, H., Grove, W.J.: The median sternal
incision in intracardiac surgery with extracorporeal circulation: a general evolution of its use
in heart surgery. Surgery 42, 753–761 (1957)
2. Defalque, R.J., Bromley, J.J.: Poststernotomy neuralgia: a new pain syndrome. Anesth.
Analg. 69(1), 81–82 (1989)
3. Bruce, J., Drury, N., Poobalan, A.S., Jeffrey, R.R., Smith, W.C., Chambers, W.A.: The
prevalence of chronic chest and leg pain following cardiac surgery: a historical cohort study.
Pain 112(3), 413 (2004)
An Electronic-Engineered Sensory Sternal Retractor Aimed 17
20. Orengo, G., Lagati, A., Saggio, G.: Modeling wearable bend sensor behavior for human
motion capture. IEEE Sens. J. 14(7), 2307–2316 (2014)
21. Bolotin, G., Buckner, G.D., Jardine, N.J., Kiefer, A.J., Campbell, N.B., Kocherginsky, M.,
Raman, J., Jeevanandam, V.: A novel instrumented retractor to monitor tissue-disruptive
forces during lateral thoracotomy. J. Thorac. Cardiovasc. Surg. 133(4), 949–954 (2007)
22. Ward, A.F., Grossi, E.A., Galloway, A.C.: Minimally invasive mitral surgery through right
mini-thoracotomy under direct vision. J Thorac Dis. 5(6), 673–679 (2013)
23. Reser, D., Holubec, T., Caliskan, E., Guidotti, A., Maisano, F.: Left anterior small
thoracotomy for minimally invasive coronary artery bypass grafting, Multimed. Man.
Cardiothorac. Surg. 1–5 (2015). https://www.ncbi.nlm.nih.gov/labs/articles/26420246/
24. Atluri, P., Stetson, R.L., Hung, G., Gaffey, A.C., Szeto, W.Y., Acker, M.A., Hargrove, W.C.:
Minimally invasive mitral valve surgery is associated with equivalent cost and shorter
hospital stay when compared with traditional sternotomy. J. Thorac. Cardiovasc. Surg.
151(2), 385–388 (2016)
µSmartScope: Towards a Fully
Automated 3D-Printed Smartphone
Microscope with Motorized Stage
1 Introduction
Neglected tropical diseases (NTDs) are a group of parasitic infectious diseases
that affect over 1.5 billion of the world’s poorest population, including 875 mil-
lion children [1]. The gold standard for detection of several NTDs is microscopic
examination, particularly via the visualization of different types of human bio-
logical products, like blood smears (e.g Malaria, Lymphatic filariasis, African
Trypanosomiasis), stool smears (e.g. intestinal helminths), and urine smears (e.g.
Schistosomiasis) [2]. Unfortunately, reliable identification of these parasitic infec-
tions requires not only proper microscopic equipment, but also high-standard
expertise for subsequent microscopic analysis. These requirements represents
the most common practical difficulties experienced in rural health facilities,
being closely related with the increasing interest of mobile health (mHealth)
and computer-aided diagnosis solutions for those particular scenarios.
The mobile phone is currently Africa’s most important digital technology. In
the year 2000 few Africans had a mobile phone, but today about three-quarters
do [3]. So it becomes natural that mHealth is starting to play an important role
when it comes to health in Africa, particularly through the usage of solutions
that allow skipping over centralized laboratories [4]. For instance, the usage
of advanced computer vision approaches coupled to the increasing processing
capabilities of mobile devices is already showing promising results in the area
of malaria diagnosis [5,6]. Moreover, considering the paramount importance of
microscopic examination for NTDs detection, the development of new portable
microscopic devices is an area that can greatly improve the chances of the suc-
cessful deployment of innovative solutions for NTDs diagnosis in underserved
areas [7]. To achieve that purpose, the constant advances and increasing pos-
sibilities coming from additive manufacturing should certainly be taken into
account, since 3D-printing currently allows faster and cheaper prototyping.
In this paper we report our most recent advances towards the development
of a fully automated 3D-printed smartphone microscope with a motorized stage,
termed µSmartScope [8]. The usage of this prototype can be resumed as follow:
the process starts by placing the smartphone in the µSmartScope along with
the smear, and have a set of magnified images acquired autonomously by the
smartphone camera sensor. This collection of images is then analyzed, either
automatically through computer vision approaches, or manually by a specialist
on a remote location. It is worth mentioning that we took into account several
particularities of the African reality during the design of this device, like the
high customs taxes and import duties currently in practice in many African
countries; this motivated us to favor solutions easily replicable in developing
countries. Several other additional requirements were equally considered, like
automating the device as much as possible, discarding the need of considerable
expertise and train of the technician in terms of maneuvering the microscope, or
supplying the energy needed for the illumination and/or any type of automation
through the mobile device battery, thus discarding the need of an additional
power source.
µSmartScope: Towards a Fully Automated 3D-Printed Smartphone 21
2 Related Work
Some research has been made in the last years to develop cell-phone based
systems that provide low-cost alternatives to conventional microscopy. The
microscopy designs of the proposed systems can be separated in three differ-
ent areas: lensless, on-lens and attachment-based approaches.
The lensless approaches are based on the principles of holographic
microscopy, i.e. the microscopic images are reconstructed from the holograms
captured by the cell-phone. This approach has the advantage of not requiring
any lenses or optical component as well as obtaining images with large field-of-
view (FOV). However, acceptable resolutions are only obtained for small mag-
nifications (∼40x magnification, NA = 0.65 objective) and processing power is
needed to reconstruct the image [9,10].
On-lens approaches usually employ a refractive element directly attached to
the smartphone camera at the focus, or a ball lens mounted in front of the
camera lens [11,12]. Despite being a low-cost alternative, the ball lens produces
a spherical focal plane, which creates aberrations and reduces drastically the
usable FOV. Moreover, magnification and radius of the ball lens are inversely
linked, so in order to achieve 1000x magnification we need a ball lens with radius
of 0.15 mm [12], which can turn the mounting and alignment process with the
camera lens really challenging.
The attachment-based approaches covers the majority of the solutions
already reported on the literature, which requires coupling additional hardware
to the cell-phone, such as commercial lenses or illumination modules [10,13,14].
This approach usually takes advantage of complex optical elements that allow
achieving suitable resolutions at high magnifications (e.g. ∼1000x), which
increases the overall cost of the system, but is currently a requirement for the
microscopic examination of several neglected diseases. With high magnifications
also emerges the limitation of having a small FOV, thus requiring the devel-
opment of mechanisms to move the smears in order to cover a large area of
the specimen. Moreover, it was verified that the majority of reported works
are designed for a unique cell-phone model, which can greatly compromise the
adoption of the proposed solution.
In this work, we present a 3D-printed microscope that can easily be attached
to a wide range of mobile devices models. To the best of our knowledge, this is
the first proposed smartphone-based alternative to conventional microscopy that
allows autonomous acquisition of a pre-defined number of images at 1000x mag-
nification with suitable resolution, by using a motorized automated stage, fully
powered and controlled by a smartphone, without the need of human interaction.
22 L. Rosado et al.
3 µSmartScope Overview
Considering that developing a cheap and easily replicable alternative to conven-
tional microscopes was one of the main goals of this work, the µSmartScope
was prototyped using Fused Filament Fabrication (FFF) technology. The FFF
is a 3D printing technology for rapid prototyping, which works by laying down
consecutive layers of material in very precise positions, in our case Polylactic
Acid (PLA) termoplastic. Layer by layer, a desired model is built from a digital
file. Despite the great advantages of this technology, 3D printing has limitations
like any other fabrication process, such as being sensitive to environmental vari-
ables like temperature, humidity or dust. Thus, we took those variables into
consideration during the design and printing process, for instance by accounting
Fig. 1. The µSmartScope prototype, with smartphone attached and microscopic slide
inserted.
Fig. 2. µSmartScope render models: (A) External view with microscope slide inserted
(at yellow); (B) Cut view with optical and electrical components highlighted; (C) Detail
of the cut view. (Color figure online)
µSmartScope: Towards a Fully Automated 3D-Printed Smartphone 23
the expected dimensional contraction during the cooling process or checking the
absence of any superficial humidity/dust on the filament, which showed to have a
significant impact in the quality and dimensional accuracy of the printed pieces.
As previously reported [8], the µSmartScope can be divided in 3 major mod-
ules (see Figs. 1 and 2): (i) Optics; (ii) Illumination; and (iii) Motorized Auto-
mated Stage (termed µStage). By analyzing the results previously reported, it
becomes clear that the µStage module was the bottleneck of the system. The
previous mechanical design had a significant unpredictable behavior due to small
irregularities of the 3D printed parts, and it was significantly slower than human
technicians in autonomously focusing microscopic fields. Thus, a major refactor
was made to the µStage module, which consequently had a pronounced impact
on the overall performance of the µSmartScope. Due to these pronounced modi-
fications, we opted to dedicate a separate section on this work to µStage module
(see Sect. 4). On the other hand, the Optics and Illumination modules suffered
only minor improvements when compared to the previously reported version [8],
so we following summarized these two modules under this section.
Fig. 3. Optics module: (A) Exploded view; (B) Assembled view; (C) Smartphone
holder; (D) Planachromat 100x oil-immersion objective lens; (E) Wide angle 10x eye-
piece lens.
24 L. Rosado et al.
Fig. 4. Illumination module: (A) Schematic of the developed light condenser gener-
ated with OpticalRayTracer® optics design software; (B) Exploded view of the light
condenser with lenses (at blue) and LED light (at yellow); (C) Assembled view of the
light condenser. (Color figure online)
µSmartScope: Towards a Fully Automated 3D-Printed Smartphone 25
4 µStage Module
The microscopic examination of blood smears usually requires the visual analysis
of different microscopic fields (i.e. positions) of the specimen. The minimum
number of required microscopic fields directly depends on the target disease and
used optical magnification, i.e. depends on the size of the target structures. For
instance, the World Health Organization (WHO) recomends the analysis of 100
different microscopic fields of a blood sample to perform a malaria microscopy
test [15]. It should be noted that this process is manually performed by trained
staff and can be extenuating, requiring that the operator takes regular breaks in
order to ensure maximum attention.
The approach proposed on this work aims to improve this process by per-
forming the microscopic slide movement autonomously and on-demand using a
smartphone. The µStage was developed for that specific purpose (see Fig. 5),
which consists on a motorized stage designed to be as cheap as possible, being
powered through the USB-OTG connection of the smartphone. In order ensure
its flexibility, most of the µStage structure is modular and can be adapted with-
out major refactors of the entire module. It is worth noting that during the
design of the µStage we tried to minimize the usage of mechanical components
and give priority to 3D-print as many parts as possible, in order to reduce costs
and ensure an easy replication of the prototype.
Regarding the µStage mechanical structure, it can be divided in two main
functional submodules (see Fig. 6): (i) the XY-plane Submodule, responsible
to move the stage on the XY plane (i.e. to obtain different microscopic fields);
and (ii) the Z-axis Submodule, responsible to move the stage in the Z axis
(i.e. find the focus point of the specimen in a fixed position on the XY plane).
Fig. 5. µStage module: External view (on the left) and cut view (on the right), with
detail of servo motors (at black), step motor (at yellow), electronics (at green) and
optics (at blue). (Color figure online)
Fig. 6. µStage functional submodules: (A) Exploded view of the XY-plane submodule,
with servo motors (at black); (B) Exploded view of the Z-axis submodule, with step
motor (at yellow) and electronics (at green). (Color figure online)
arrangement is not linear, since the servo movement is provided in a 90◦ arc.
Nevertheless, it allows the acquisition of different microscopic fields, particularly
by moving one of the servos while the other remains static. Taking into account
that repositioning to a specific microscopic field location with high precision
is usually not required in microscopic blood smear analysis, this module is not
required to perform high precision and resolution displacements. Covering a high
number of distinct microscopic fields from different XY positions is by far much
more important, in order to obtain a reliable overview of the specimen.
Composed by 3 tubular structures that slide against their negatives in the base
part (see Fig. 6(B)). This submodule allows the vertical movement of the µStage,
in order to find the Z-axis position that corresponds to the focus point of the
specimen. A simple stepper motor is used to provide the vertical movement,
being the rotary movement of the stepper engine translated to linear motion
through a threaded rod and a nut fixed in this part. Moreover, the base part of
this submodule is also used to fix the electronic board, the Z-axis actuator and
the µUSB connector.
The correct calibration and parameterization of the 3D-printer is evidently
important for all µSmartScope printed components, but particularly crucial for
this submodule. Obtaining the required gap between each tubular structure and
the corresponding negative is mandatory to achieve precise Z-axis displacements.
Being the µStage the bottleneck of the previously reported version, two major
mechanical changes were made in order to improve the performance of the Z-axis
submodule in terms of precision and stability:
4.3 Electronics
The power board was designed to power the 3 actuators (one stepper and two
servos) and the illumination LED, using exclusively the power from a USB con-
nection with 5 V and 500 mA. It should be noted that most of the currently
available smartphone models have USB-OTG interface in order to allow the
connection of USB peripherals. To be fully compatible with the µSmartScope,
the used smartphone model needs to support this power standard, otherwise it
will not be capable of powering the µStage. An image of the developed PCB can
be observed in Fig. 7, and the electronic system is composed by:
using the USB serial for Android library [17]. Moreover, an API for Android
was developed to allow interaction with the stage. This API was made to be
as simple as possible to integrate in any app, providing every function needed
to fully control the stage (i.e. stepping in X, Y and Z axes, as well as control
the LED light).
5 Automated Focus
Due to lens constrains, it is not possible to obtain maximal focus in the whole
area of the microscopic field. Therefore, a square region in the center of the pre-
view image was defined, which generally is the place where distortion interference
is lower (see Fig. 8). This square region will be used by the focus algorithm to
extract the metrics and make decisions accordingly.
30 L. Rosado et al.
Fig. 8. Focused image obtained using the µSmartScope, and respective central square
resulting from focus region selection: (A) Thin blood smear; (B) Thick blood smear.
Taking into consideration several works on the literature targeting the automatic
focus for microscopic devices using image processing, in this work we tested and
compared a wide range of focus metrics already proposed that were considered
highly relevant for automatic focusing. In detail, we tested: derivative-based as
the Brenner gradient and the Tenenbaum gradient; statistics based, like the
normalized variance; histogram-based, as entropy; and intuitive algorithms, like
thresholded content [20,21].
After this comparative analysis, the Tenenbaum gradient [22] standard devi-
ation (T EN GST D ) and sum (T EN GSU M ) were considered the most discrimina-
tive focus metrics, i.e. the metrics that allows better differentiation of the ideal
focus point. The Tenenbaum gradient is obtained by convolving the previously
selected central square of the image with Sobel operators, and by summing the
square of gradient vector components:
because the order of magnitude of the T EN GST D and T EN GSU M metrics pre-
sented a significant variation for different smartphone-specimen combinations.
We concluded that this pronounced variation was being cause by two main fac-
tor: (i) images of the same specimen captured with different smartphone models
may present different image characteristics (e.g. high variation in image con-
trast and/or color representation) due to the different camera sensor specifica-
tions; and (ii) microscopy specimen preparation greatly depends on good-quality
reagents and proper human skills, which can lead to huge intensity variations of
the stained components.
Despite this shifting behavior, we noticed however that the relative varia-
tion of both metrics was considerably more homogeneous when compared with
the respective absolute values. So instead of using directly the T EN GST D and
T EN GSU M metrics values, we used the following slope gradients:
ST D
(T EN GST
i
D
− T EN GST
j
D
)
SLOP Eij = , (3)
i−j
SU M
(T EN GSU
i
M
− T EN GSU
j
M
)
SLOP Eij = , (4)
i−j
where i and j are two different Z-axis positions, with i, j = {0, 1, . . . , Z M AX },
being Z M AX the highest Z-axis position reachable by the µStage.
1. The algorithm starts in the Rough phase, where the goal is to detect a pro-
ST D SU M
nounced increase of the focus metrics SLOP Eij and SLOP Eij , which
corresponds to the beginning of the visualization of the microscopic struc-
tures. At the beginning of the Rough phase, the µStage is on the bottom
Z-axis position (Z P os = 0), which from now on will be called the reset posi-
tion (Z Reset ).
2. While the µStage is on Z Reset , the user inserts the microscopic slide and
gives an order through the mobile application to start the image acquisition
process. The µStage starts autonomously ascending in the Z-axis to a specific
Z position (Z P os = 5000), which from now on will be called the initial position
(Z Initial ). The µStage is surely below the focus point at Z Initial , so no focus
metrics are computed until this position is reached.
32 L. Rosado et al.
3. When the Z Initial is reached, the µStage starts ascending on specific inter-
Interval Interval
vals (ZStep = 2). For each ZStep , the image preview obtained with
ST D
the smartphone camera sensor is used to compute the SLOP Eij and
SU M
SLOP Eij focus metrics. It should be noted that in the Rough phase we
use a frame rate of 20 frames per second (fps).
4. The transition from the Rough phase to the Precise phase occurs when the
ST D SU M
following condition is verified: SLOP Eij > 20 and SLOP Eij > 20.
5. When the Precise phase is reached, the µStage keeps ascending with the same
Interval
ZStep , but we decrease the frame rate to 4 fps. This adjustment is directly
related with the vibration caused by each mechanical step, which consequently
causes motion blur on the image preview. By guaranteeing more time between
the mechanical movement and the analysis of the preview image, we ensure
that the vibration dissipates and the computation of the focus metrics is not
affected by that artifact.
6. During the Precise phase, the T EN GST D is computed on the preview image
Interval
for each ZStep . If the T EN GST D of the current Z position has the high-
est value found on the Precise phase, this position is considered the best
candidate to focus point. When that happens, an image is captured with 5
Megapixels and saved on the smartphone memory. In the event of an higher
T EN GST D value for a posterior Z position, a new image is captured and the
best candidate to focus point is updated.
7. The focus point search logic ends with successful state when a candidate to
focus point was found and: (i) the T EN GST D presents always lower value in
Interval
the following 20 ZStep positions; or (ii) the T EN GST D has a value lower
than half of the best candidate T EN GST D in any of the following ZStep Interval
positions;
8. The focus point search logic ends with fail state if the maximum Z-axis posi-
tion (Z P os = 7000) is reached without a single candidate to focus point. In
the event of fail state, a reset procedure takes place: (i) the µStage descends
to the Z Reset position; (ii) the µStage moves to a different position in the XY
plane; and (iii) the focus point search logic described above is restarted. After
extensive testing, we noticed that the most common cause for the occurrence
of a fail state is usually the low number of microscopic structures in that
specific XY position.
The behavior of the proposed focus point search logic is illustrated on Fig. 9,
ST D SU M
where is depicted the variation of the SLOP Eij and SLOP Eij while the
µStage is ascending in the vertical axis.
We finish this section with a detailed analysis of the proposed automated focus
approach, covering the usage of a wide range of smartphone models and different
thin and thick blood smears.
6.1 Resolution
The magnified images of the smears obtained with the µSmartScope must have
an appropriate resolution over a sufficiently large area, so a conclusive decision
about the presence of a specific infectious agent can be made. The 1951 USAF
resolution test chart is a resolution test pattern conforming to MIL-STD-150A
standard, set by US Air Force in 1951. It is still widely accepted to test the
Fig. 9. Variation of the Tenenbaum focus metrics while the µStage is ascending in the
vertical axis. Illustrative examples of preview images obtained by the smartphone cam-
era at different Z positions are presented: (A) During the Rough phase; (B) Entering
the Precise phase; (C) Focus point; (D) Stopping condition after focus point detected.
34 L. Rosado et al.
Fig. 10. Images of READY OPTICS USAF 1951 microscope resolution target: (A)
Acquired using Bresser Microscope-5102000-Erudit DL; (B) Detail of Group 10 using
the Microscope; (C) Acquired using the µSmartScope; (D) Detail of Group 10 using
the µSmartScope [8].
Fig. 11. Minimum Michelson contrast for USAF Resolution Target Elements of Group
10 [8].
directions. The lower values of Michelson contrast for the µSmartScope might be
caused by the limitation in terms of the maximum power rate of the smartphone,
since both the illumination and motorized automated stage must be powered by
USB-OTG. Thus the power that feeds the LED is limited, which diminishes the
intensity of the LED and consequently the image contrast.
Resolution and field of view are inversely linked in standard laboratory micro-
scopes. In order to obtain microscopic fields with both high magnification and
resolution, the usage of objectives with higher numerical aperture is required,
which results in smaller Field of Views (FOV) [10]. To determine the FOV of the
µSmartScope, images of the READY OPTICS USAF 1951 Microscope Resolu-
tion Target were acquired and used to estimate the microns to pixels (µm/pixel)
relationship. The exact distances between bars of a specific Element is given by
the specifications of the resolution target, particularly Element 2 and 3 of Group
8 corresponds to 1.740 µm and 1.550 µm, respectively. By measuring the number
of pixels between bars of this 2 Elements on the acquired image, a relationship
of 0.085 µm/pixel was obtained for Element 2 and 0.084 µm/pixel for Element 3.
So we considered the average value, i.e. a relationship of 0.0845 µm/pixel. Fur-
thermore, the number of pixels for the vertical and horizontal axis that passes
through the center of the visible optical circle of the acquired image were deter-
mined. These values were then combined with the previously calculated µm/pixel
relationship, in order to estimate the FOV of the visible optical circle, which
resulted in a FOV of 214.38 µm and 206.87 µm for the vertical and horizontal
axis, respectively.
36 L. Rosado et al.
6.3 Illumination
In order to evaluate the uniformity of the illumination of the LED coupled to
the proposed condenser, an image was acquired with a blank microscope slide
(see Fig. 12(A)). A diagonal line scan starting on the top right corner of the
image was defined, in order to evaluate the variation of the pixels intensity
along this line. Particularly, a total of 200 pixel boxes were considered along the
scan line, being each box equally spaced and with a size of 10 × 10 pixels. The
mean and standard deviation of the pixels intensity was computed for each box,
as depicted in Fig. 12(B) and (C). This results demonstrate that the proposed
illumination set up ensures substantially uniform illumination with low noise,
being the small intensity variations probably caused by dust particles on the
lenses or components floating in the immersion oil.
Fig. 12. µSmartScope illumination uniformity analysis: (A) Original image; (B) Mean
pixel intensity of the 10 × 10 pixel boxes on the diagonal direction; (C) Standard
deviation of the 10 × 10 pixel boxes on the diagonal direction [8].
Fig. 13. Measured step size values after 100 repetitions: (A) XY plane; (B) Z axis.
For the displacement in the XY plane, a clear variation of the step size is
depict in Fig. 13(A). This behavior is probably caused by two main factors: (i)
the non-linear movement of the servo; and (ii) imperfections on the sliding parts
that are a consequence of the 3D printing process. Nevertheless, we obtained
an average displacement of 330 µm with a standard deviation of 81 µm, a result
that attests the suitability of the µStage XY-plane movement for microscopy
analysis purposes. As referred before, repositioning to a specific XY position
with high precision is usually not required in conventional microscopy analysis,
being much more important the coverage of a high number of different XY
positions. Please note that the obtained XY µStage displacement is marginally
higher than the FOV determined in Sect. 6.2, which means that we are obtaining
a new microscopic field every time we take a XY step.
Regarding the displacement in the Z axis, it should be taken into account
that our caliper is not able to measure such small steps. To estimate the average
step size, one full revolution of the motor was considered (which corresponds
to 512 steps), and the respective displacement was measured 100 times. This
distance was then divided by the number of steps, in order to estimate the
travel of a single step. As we can see in Fig. 13(B), the steps are less dispersed,
but some variability is still verified. Each step corresponds to an average of
0.763 ± 0.098 µm, which gives an indication that our steps in the Z axis are
within the theoretical values. By comparing these results with the previously
reported performance of 0.98 ± 0.18 µm [8], it becomes clear the major impact
of the mechanical improvements performed in this work.
38 L. Rosado et al.
Table 2. Average seconds per image for different smartphone models and thin blood
smears, calculated over 100 automated focus attempts for each combination.
Thin blood Seconds per image (average time over 100 automated focus attempts)
smears
Asus HTC One Motorola LG Nexus 5 Samsung
Zenfone 2 M8 Moto G5 Galaxy S6
Slide #1 11.30 8.50 8.39 9.04 7.46
Slide #2 10.76 7.39 8.11 7.37 6.58
Slide #3 10.45 7.25 8.12 7.77 6.99
Slide #4 11.85 7.15 7.95 8.14 7.16
Slide #5 9.21 7.74 7.99 8.11 6.86
Table 3. Average seconds per image for different smartphone models and thick blood
smears, calculated over 100 automated focus attempts for each combination.
Thick blood Seconds per image (average time over 100 automated focus attempts)
smears Asus HTC One Motorola LG Nexus 5 Samsung
Zenfone 2 M8 Moto G5 Galaxy S6
Slide #1 17.09 18.19 8.76 16.54 23.93
Slide #2 16.49 12.60 13.56 11.49 8.80
Slide #3 13.85 15.24 8.81 10.15 6.64
Slide #4 13.24 9.77 14.64 11.68 8.21
Fig. 14. Illustrative examples of focused images autonomously acquired with the
µSmartScope and different smartphone models for 5 different malaria-infected thin
blood smears: (A) Asus Zenfone 2; (B) HTC One M8; (C) Motorola Moto G5; (D) LG
Nexus 5; (E) Samsung Galaxy 6.
automatic focus of the smear. Considering the current battery capacity of smart-
phones, this obviously represents a huge burden in terms of power consumption,
and it is clear that the current autonomy of the system is low for a day of con-
tinuous use. As an alternative, we envision the usage of a power bank coupled
to a OTG splitter cable, which we aim to include in the next version of the
µSmartScope system.
40 L. Rosado et al.
Fig. 15. Illustrative examples of focused images autonomously acquired with the
µSmartScope and different smartphone models for 4 different malaria-infected thick
blood smears: (A) Asus Zenfone 2; (B) HTC One M8; (C) Motorola Moto G5; (D) LG
Nexus 5; (E) Samsung Galaxy 6.
Fig. 16. Images of different smears acquired with the µSmartScope: (A) Thick blood
smear infected with malaria parasites (P.falciparum species); (B) Thin blood smear
infected with malaria parasites (P.ovale and P.malariae species); (C) Thin blood smear
infected with Chagas parasites (Trypanosome cruzi species); (D) Liquid-based Pap
smear with high grade squamous lesions; (E) Thick blood smear infected with Lym-
phatic Filariasis parasites (Brugia malayi species); (F) Thick blood smear infected
with Lymphatic Filariasis parasites (Wuchereria bancrofti species). Images (A), (B)
and (C) were acquires with a LG Nexus 5, while images (D), (E) and (F) with a Sam-
sung Galaxy S5. All images were acquired with magnification of 1000x, except image
(D) which has magnification of 400x [8].
In this paper we report our most recent advances towards the development of
a fully automated 3D-printed smartphone microscope with a motorized stage
(µSmartScope). This prototype is the first proposed smartphone-based alterna-
tive to conventional microscopy that allows autonomous acquisition of a pre-
defined number of images at 1000x magnification with suitable resolution, by
using a motorized automated stage fully powered and controlled by a smart-
phone, without any human interaction.
All the components of the proposed device are described and properly eval-
uated. In terms of the optical module, a minimum resolution of 0.388 µm was
determined, with a FOV of 214.38 µm and 206.87 µm for the vertical and hor-
izontal axis that passes through the center of the visible optical circle, respec-
tively. Regarding the illumination module, the LED light coupled to the proposed
condenser demonstrated to achieve an uniform illumination suitable for bright-
field microscopy. The developed motorized automated stage (µStage) achieved
an average resolution of 330 ± 81 µm for the XY plane steps and an average reso-
lution of 0.763 ± 0.098 µm for the Z-axis steps, results that validates its effective
usage for microscopy analysis. Moreover, the proposed automated focus app-
roach was evaluated on 9 different blood smears and 5 different smartphone
models, achieving an average performance time of 8.31 ± 1.41 sec/image and
12.98 ± 4.20 sec/image, for thin and thick blood smears, respectively.
µSmartScope: Towards a Fully Automated 3D-Printed Smartphone 43
Several smears infected by different blood parasites responsible for the most
relevant neglected tropical diseases were used to test the device. The acquired
images showed that it was possible to detect those agents through images
acquired via the µSmartScope, which clearly illustrate the huge potential of
this device, specially in developing countries with limited access to healthcare
services. As future work, we want to test the performance and assess the practi-
cal usefulness of the µSmartScope on field trials. Particularly, the µSmartScope
will represent a component of a mobile-based framework for malaria parasites
detection currently being developed. Thus, we aim to integrate this prototype
into the referred framework, with the ultimate goal of creating a system that:
(i) provides an effective pre-diagnosis of malaria in medically-undeserved areas;
(ii) is low cost and mobile based; (iii) is easy to use, even for non-experts in
microscopy.
References
1. The END Fund: NTD Overview. http://www.end.org/whatwedo/ntdoverview.
Accessed 11 Sept 2017
2. Utzinger, J., Becker, S.L., Knopp, S., Blum, J., Neumayr, A.L., Keiser, J., Hatz,
C.F.: Neglected tropical diseases: diagnosis, clinical management, treatment and
control. Swiss Med. Weekly 142 (2012)
3. Zachary, G.: Technology alone won’t improve health in Africa [Spectral lines]. IEEE
Spectr. 52(1), 7 (2015)
4. Dolgin, E.: Portable pathology for Africa. IEEE Spectr. 52(1), 37–39 (2015)
5. Rosado, L., da Costa, J.M.C., Elias, D., Cardoso, J.S.: Automated detection of
malaria parasites on thick blood smears via mobile devices. Procedia Comput. Sci.
90, 138–144 (2016)
6. Rosado, L., da Costa, J.M.C., Elias, D., Cardoso, J.S.: Mobile-based analysis of
malaria-infected thin blood smears: automated species and life cycle stage deter-
mination. Sensors 17, 2167 (2017)
7. Rosado, L., da Costa, J.M.C., Elias, D., Cardoso, J.S.: A review of automatic
malaria parasites detection and segmentation in microscopic images. 14(1), 11–22
(2016). http://www.eurekaselect.com
8. Rosado, L., Oliveira, J., Vasconcelos, M.J.M., da Costa, J.M.C., Elias, D., Cardoso,
J.S.: µSmartScope: 3D-printed smartphone microscope with motorized automated
stage. In: Proceedings of the 10th International Joint Conference on Biomedical
Engineering Systems and Technologies - Volume 1: BIODEVICES, (BIOSTEC
2017), pp. 38–48. INSTICC, SciTePress (2017)
9. Tseng, D., Mudanyali, O., Oztoprak, C., Isikman, S.O., Sencan, I., Yaglidere, O.,
Ozcan, A.: Lensfree microscopy on a cellphone. Lab Chip 10(14), 1787–1792 (2010)
10. Pirnstill, C.W., Coté, G.L.: Malaria diagnosis using a mobile phone polarized
microscope. Sci. Rep. 5, 13368 (2015)
44 L. Rosado et al.
11. Arpa, A., Wetzstein, G., Lanman, D., Raskar, R.: Single lens off-chip cellphone
microscopy, pp. 23–28. IEEE, June 2012
12. Cybulski, J.S., Clements, J., Prakash, M.: Foldscope: origami-based paper micro-
scope. PLoS ONE 9(6), e98781 (2014)
13. Smith, Z.J., Chu, K., Espenson, A.R., Rahimzadeh, M., Gryshuk, A., Molinaro,
M., Dwyre, D.M., Lane, S., Matthews, D., Wachsmann-Hogiu, S.: Cell-phone-based
platform for biomedical device development and education applications. PLoS ONE
6(3), e17150 (2011)
14. Switz, N.A., D’Ambrosio, M.V., Fletcher, D.A.: Low-cost mobile phone microscopy
with a reversed mobile phone camera lens. PLoS ONE 9(5), e95330 (2014)
15. WHO: Basic malaria microscopy. Number Part 1. World Health Organization
(1991)
16. Budynas, R.G., Nisbett, K.J.: Shigley’s Mechanical Engineering Design, 10th edn.
McGraw-Hill Education, New York (2014)
17. Wakerly, M.: USB Serial for Android (2012)
18. Krotkov, E.: Focusing. Int. J. Comput. Vis. 1(3), 223–237 (1988)
19. Shih, L.: Autofocus survey: a comparison of algorithms. In: Electronic Imaging
2007, International Society for Optics and Photonics, p. 65020B (2007)
20. Sun, Y., Duthaler, S., Nelson, B.J.: Autofocusing algorithm selection in computer
microscopy. In: 2005 IEEE/RSJ International Conference on Intelligent Robots
and Systems, pp. 70–76. IEEE (2005)
21. Liu, X., Wang, W., Sun, Y.: Dynamic evaluation of autofocusing for automated
microscopic analysis of blood smear and pap smear. J. Microsc. 227(1), 15–23
(2007)
22. Tenenbaum, J.M.: Accommodation in computer vision. Ph.D. dissertation. Tech-
nical report, DTIC Document (1970)
A Portable Chemical Detection System
with Anti-body Biosensor for Impedance Based
Monitoring of T2-mycotoxin Bioterrorism
Agents
Tyndall National Institute, Lee Maltings, University College Cork, Cork, Ireland
vladimir.ogurtsov@tyndall.ie
1 Introduction
The development of miniaturized, portable biosensing systems for the rapid, reliable
and low cost determination of contaminating bio agents in the environment (atmo-
sphere, water, food etc.) has received considerable attention in recent years particularly
with the advancements in biotechnology. In the last decade with the growing threat of
bioterrorism, the demand on development of systems that can rapidly and accurately
detect bio agents, that can threaten human life or health, is significantly increased.
There exist a few options for the detection of bioterrorism agents [1]:
1. Generic detectors that provide an on/off alert
2. Biosensors for rapid screening and presumptive identification [2, 3] and
3. Identification systems for definitive confirmation.
Of the three technologies, biosensors can play an important role in detection of
Chemical, Biological, Radiological, Nuclear, and Explosive (CBRNE) materials
because they can provide a sensitive monitoring tool whilst also being amenable to
miniaturization and forming part of a portable system. Beside security, biosensing
systems are increasingly used in a range of other different applications including
environmental, clinical, food and agriculture [4–15]. The conventional approaches are
mainly lab-based and cannot be easily deployed at point-of-need. Modern microfab-
ricated biosensors, on the other hand, offer the advantages of a cost-effective and rapid
sample analysis, and specific and sensitive measurements over the traditional methods,
which tend to be a multi-step (e.g. ELISA), or to involve sophisticated and expensive
equipment (e.g. HPLC). The ability to apply semiconductor processing technologies,
more commonly used in the IC industry, in the sensor chip fabrication [8, 10] enables
large batch production and subsequent availability of cheap and disposable devices.
The International Union of Pure and Applied Chemistry (IUPAC) defines a
biosensor as follows “A biosensor is a self-contained integrated device which is cap-
able of providing specific quantitative or semi-quantitative analytical information using
a biological recognition element (biochemical receptor) that is in direct spatial contact
with a transducer element” [11, 12]. Thus, according to this classification a biosensor is
essentially composed of two elements:
1. A bioreceptor that is an immobilized sensitive biological element (e.g. enzyme,
DNA, antibody) recognizing the analyte (e.g. enzyme substrate, complimentary
DNA, antigen).
2. A transducer that converts (bio) chemical signal resulting from the interaction of the
analyte with the bioreceptor into its electrical equivalent. The intensity of the signal
generated is either directly or indirectly proportional to the analyte concentration.
Essentially, the selectivity or specificity of the biosensor depends on the
biorecognition element, which is capable of “sensing” the presence of an analyte [13,
14]. There are currently three generations of biosensors in existence. The term gen-
eration was originally created to describe the stages of integration in biosensors.
Biosensors with mediated response mainly generated by membrane-bound or mem-
brane entrapped biocomponents (first generation), were followed by virtually
A Portable Chemical Detection System with Anti-body Biosensor 47
reagentless biosensors where biocomponents are fixed directly to the sensor surface and
either the reaction of coimmobolised cosubstrates (bound to the electrode, polymer or
enzyme itself) or the direct heterogeneous electron transfer between the prosthetic
group and the electrode is exploited (second generation). While the immobilisation of
the receptors directly on an electronic circuit leads to systems with integrated signal
generation, transduction and processing (third generation) having the potential of
considerable miniaturisation [15]. The devices investigated in this study should be
considered as the third generation under this classification and this highlights the
progress sensing systems since their relatively recent development.
Antibody-based biosensors or immunosensor are found on an antigen-antibody
interaction where antibody (Ab), Y-shaped proteins, bind with a target, an invading
antigen (Ag), through a high affinity binding reaction. The specificity of the reaction is
due to a chemical constitution of antibody which is specific to the target antigen that
defines similarity of the antigen-antibody reaction to a lock (antibody) and key (anti-
gen) operation. This conditions such advantages of immunosensors as high versatility,
selectivity, stability and sensitivity. At the same time, the strong antigen-antibody
binding makes it very difficult to break these bonds, thus, a typical immunosensor is a
disposable one-shot sensor that cannot be used after the target detection [16–20].
Immunosensors utilizing antibody-based recognition elements, have been devel-
oped on a wide range of transduction platforms. The transducer element translates the
selective recognition of the analyte into a quantifiable signal and thus, has major
influence on sensitivity [21]. Since 1959 when an antibodies based biosensor for the
first time was used with radiolabeled assay for detection of insulin in the body fluids
[22] immunosensors have revolutionized impact on detection of a whole lot of different
target analytes such as disease markers, food and environmental contaminants, bio-
logical warfare agents and illicit drugs. Nowadays, transduction approaches used with
immunosensors include electrochemical, piezoelectric and optical systems [15, 23]
where the mostly used methods are based on optical and electrochemical methodology.
However, the classical optical transduction mechanism has suffered from poor sensi-
tivity when coupled with radioimmunoassay, the short half-life of radioactive agents,
concerns of health hazards, and disposal problems. As well the optical methods usually
require a complicated sample preparation that increases risk of sample contamination,
rises cost and time of analysis and makes them difficult for implementation in a
portable device capable of working in field conditions outside of specialized labs.
Electrochemical detection overcomes the most of these problems and provides means
for fast, simple, reliable and economical realization of portable autonomous
immunosensing systems capable of working at point-of-need without end-user inter-
vention [19, 24]. The advantage of electrochemical biosensors is also that they can
monitor bimolecular interactions in real time. In this process the biorecognition com-
ponents are immobolised on a solid surface (usually the sensor chip), and the com-
ponent to be detected is present in the solution phase [25].
A typical electrochemical biosensor usually represents an electrode or an array of
microdiscs/bands with corresponding bio functionalization acting as the working elec-
trode within a three-electrode electrochemical cell containing also a reference electrode
(i.e. Ag/AgCl) and an auxiliary/counter electrode (which varies) [26]. The working
electrode provides the sensing capability, the reference and the counter electrodes
48 V. I. Ogurtsov and K. Twomey
facilitate the measurements. A design option for the electrochemical biosensor is a bio
specific membrane that selectively interacts with the analyte. A physical transducer
coupled to the membrane detects this biointeraction and generates the resulting electrical
signal. The bio specific membrane can be made up of a material that undergoes electron
exchange reactions with the analyte so a direct electrical output signal consisting of a
potential or a current can be detected and interpreted [27].
Typical microfabricated biosensors incorporate a gold electrode, or other materials
e.g. platinum, Si3N4, active layer upon which different surface chemistries are applied
to form a complete device that is specific and sensitive to a target of interest [28, 29].
Within the branch of electrochemical sensors, there are amperometeric, potentiometric,
impedimetric and field-effect transistor (FET) biosensors [19, 23, 30, 31]. With the
amperometric type biosensors, the sensor current is monitored with time; for the
potentiometric – the voltage is monitored. Both of these types offer relatively straight
forward measurements and a rapid analysis. The impedimetric is more complex and
more sensitive technique. Its classical implementation is based on the measurement of
an AC current that forms in the response on the application to the sensor a sinusoidal
voltage over a set of frequencies. The advantage of impedemetric biosensor is that they
can directly sense the changes in interface between the bio functionalized electrode and
the sample solution due to the biorecognition reaction through monitoring of the sensor
impedance variations. For that they do not need any additional labelling that is very
often required in case of optical biosensor [22]. This circumstance along with relative
simplicity of electrochemical instrumentation defines the ease and cheapness of
impedemetric biosensors operation (no additional expensing labelling or sample
preparation) that with such their advantages as high sensitivity and the ability to
miniaturise the associated measurement instrumentation explains a growing interest in
label-free impedimetric biosensors for different application [4, 32]. A further advance
in electrochemical biosensors can be obtained by using of micro-sized electrodes,
which improve the mass transport (and hence the sensor sensitivity) owing to the
occurrence of radial diffusion over planar diffusion from macro-size electrodes
resulting in an increase in the signal-to-noise ratio and reduction in the iR drop [33].
The presented study describes the development of such portable biosensing plat-
form with immunosensor on the base of microfabricated microelectrode array capable
of biotoxin detection. The performance of the platform was validated by detection of
T2-toxin. It is a trichothecene mycotoxin, which is toxic to humans and animals. It is
naturally occurring mould byproduct of Fusarium spp. fungus that can be found in
grains such as barley, wheat and oats. It is well-known that some mycotoxins can be
used as chemical warfare agents [34, 35]. Inside of this group, trichothecenes can act
immediately upon contact, and exposure to a few milligrams of T-2 is potentially lethal.
The symptoms exhibited by purported victims included internal hemorrhaging, blis-
tering of the skin, and other clinical responses that are caused by exposure to
trichothecenes.
Detection methods for T-2 toxin can be classified into two types: screening
methods like ELISA and analytical methods like GC and HPLC [36, 37]. Since low
ppb concentrations are usually required to be detected, very sensitive analytical
methods are hence needed. Advantages of screening methods like ELISA are that these
methods generally do not require a clean-up other than dilution of the extract and/or
A Portable Chemical Detection System with Anti-body Biosensor 49
filtration before analysis. ELISA type screening methods rely on antibodies to detect
the T-2 toxin. The specific antibodies (monoclonal or polyclonal) distinguishes the 3D
structure of T-2 toxin. Usually an incubation period of about 1-2 h is required to
complete this bio recognition procedure. The T-2 toxin present in the sample being
tested competes with a labelled toxin for a limited number of antibody binding sites.
The more toxin in the sample solution, the lower the binding of the labelled toxin and
the lower the signal generated by the assay. The presence of the T-2 toxin is indicated
by a specific color change. However, there are disadvantages to this particular mech-
anism of detection. These two main disadvantages are high matrix dependence and
overestimation of the T-2 toxin concentration, which can occur if structurally related
mycotoxins or constituents present in the matrix interfere with the competitive
antibody-binding site mechanism required for the detection of T-2 toxin by ELISA
[38]. Hence, this could lead to false results. However, there are advantages of screening
methods like ELISA. They are high sample throughput and ease of operation. Thus,
detection limits of 0.1 ng/g obtained with this methodology have been reported [37].
Other recent screening methods include Molecularly Imprinted Polymers (MIPs) [38]
and Dipstick and Lateral Flow tests, where detection limits of 4 and 50 ng/kg were
stated [39].
The presented study describes the development of a portable autonomous
impedimetric biosensing system with a label-free electrochemical immunosensor for
T2-toxin detection for using at point of need. The immunosensor is based on a bio
functionalized microelectrode array [24, 40] which works as a part of a microfabricated
electrochemical cell realized as a single silicon packageless chip. The developed
biosensing system played role of a biohazard detection unit in a portable HANHOLD
platform developed for detection of CBRN threads for operating at point of need. The
platform was created during fulfilment an FP7 collaborative project called Handhold
(Handheld Olfactory Detector) [41] and included also an optical sensing module for
detection of chemical thread (explosive) and a radioactivity detector. All modules were
operating under integrated control from a single board PC that has possibility of
wirelessly transmitting the platform data to an outside decision center. The platform
was creating by the analogy of a sniffer dog that is used at airport custom control.
While the dog will remain a central part of the detection process at border crossing and
airports, sensor technology and low power computer based embedded system are
improving to the extent that the time is now right to develop substantially improved
mobile detection devices that can successfully complement the role played now mostly
by dogs. Moreover, these detection devices can be networked together to provide
enhanced alert facilities and also to enable easier management and field deployment of
the platforms themselves.
2 Experimental
operational characteristics of a sniffer dog. Unlike the situation with dog retraining, this
platform can be retargeted simply by removing one set of sensors and adding others,
the software load reacting dynamically to the reconfiguration. Figure 1 illustrates the
different technologies that play a role in the development of the Handhold biosensing
system.
These technologies include microfabrication and surface chemistry that will be used
for biosensor preparation, mix signal electronic design – for instrumentation hardware
development, signal processing algorithm and corresponding software – for imple-
mentation of the autonomous system operation without end-user interference and
integration – for completion of the system and providing its operability in the Hand-
hold CBRN threads detection platform.
2.2 Biosensor
As shown in Fig. 1 the biosensor consists of two main parts. They are a transducer
(working electrode in the three-electrode micro electrochemical system) and a bio
recognition layer that adds the bio specificity to the electrochemical sensor. For
improving biosensor sensitivity a microelectrode array structure was selected for
working electrode implementation. The development of the sensor transducer was
started from simulation of different electrode designs including single disc, disc array,
single band and band array. The simulation was performed in Multiphysics Engi-
neering Simulation Software Comsol. The aim of this part of the work was to analyse
the diffusion transport mechanism that occurs between the interface of the electrode and
the sample solution and to determine the optimum transducer layouts according to its
electrochemical performance. In this simulation, the model was created where each
microelectrode was positioned at a set distance below the sensor surface. This was to
A Portable Chemical Detection System with Anti-body Biosensor 51
represent the physical fabrication process where an etch layer opens up areas of
exposure in the top passivation layer to allow electrode contact with the environment.
This technological process allows very fine electrode patterning but supposes that each
electrode is located below the chip surface at the depth of few hundred of nm which is
the range of possible thicknesses of the silicon nitride layer deposited on the metal. The
hexagonal array configuration providing the most compact electrode placement [33]
was considered with this simulation. For the disc array model, the diffusion zone
approach was applied [42, 43], as schematically shown in Fig. 2.
The species concentration on the electrode surface was defined from the Nernst
equation Eq. (1)
CA nF
¼ exp E E= ð1Þ
CB RT
where CA and CB are concentrations of the electro active species involved into the
electrochemical reaction, n is a number of electrons participated in electron exchange,
R is the universal gas constant, F is the Faraday constant, E/ is the formal potential. The
electrode current was determined from integration of the diffusion flux over the elec-
trode area, Eq. (2):
Zra
@Cðr; x; tÞ
i ¼ 2pnFD rdr ð2Þ
@x x¼xs
0
where D is diffusion coefficient of electro active species. Some results attained for the
different electrode designs from this simulation are discussed below.
Examples of the calculated concentration profiles (occurring due to the diffusion
mechanism taking place) at the different disc electrode structures and dimensions are
shown in the diagrams presented in Fig. 3 in 10 s span after applying to the electrode a
voltage that changes, according to Nernst equation Eq. (1), the concentration of the
electro active species on the electrode surface to zero level.
Fig. 3. Concentration profiles for macro, micro and micro electrode array after 10 s. (Color
figure online)
As one can note, for a macro electrode, a planar type of diffusion occurs near the
electrode (Fig. 3a). Diffusion zones can be seen in the changing color from blue to
bright red. The dark red color situated at a distance from the electrode relates to the
bulk concentration, where no or minimal diffusion is occurring. A reduction in elec-
trode size from macro to micro changes the diffusion type from planar to radial. It can
be seen by an example of a microelectrode of radius 5 lm that spans out a semicircular
concentration wave from the electrode (Fig. 3b). This type of diffusion results in a high
current density that leads to increase of signal to noise ratio, hence it is desirable to
have. However, because of a small area of the microelectrode its total current is
reduced in absolute value that can be difficult to measure in practice due to existence of
electrochemical and electronic noises. To overcome this problem an array of micro-
electrodes can be used to increase the total signal. Care needs to be taken to ensure that
there is no diffusion zones overlap between the neighboring electrodes. A general
rule-of-thumb is to separate the electrodes in array by the length of more or equal to a
specific center-to-center distance dc= 20r [44]. An example of overlapping diffusion
zones is shown in Fig. 3c for a microdisc array of radius 5 lm and center-to-center
distance of 40 lm (8 disc radii). As can be seen here, at a short distance from the array
the concentration profiles transfer to the plain profile as in case of a macro electrode
(Fig. 3a). This circumstance reduces the transducer signal. In order to avoid this sit-
uation, the center-to-center spacing in the transducer in the biosensor chip prototype for
this disk radius was increased to 100 lm.
Dependence of the transducer current and current density from the time for the
micro electrode array of different designs can be seen in Figs. 4 and 5. The current
density of the whole electrode array is shown in the left plots and current of the whole
electrode array is presented in the right plots of these figures. The transient process for a
A Portable Chemical Detection System with Anti-body Biosensor 53
Fig. 4. Transient current density (a) and total current (b) for a microdisk electrodes array with
ra= 5 lm and different values of dc and Nel as indicated in Table 1. (Color figure online)
Fig. 5. Transient current density (a) and total current (b) for a microdisk electrodes array with
ra = 1 lm and different values of dc and Nel as indicated in in Table 1.
As was mentioned above, due to the technological peculiarities after fabrication the
microelectrode of radius re is located at the bottom of the pore in the isolation layer of
thickness dp deposited on the top of the metal. Thus, in the reality the microelectrode
54 V. I. Ogurtsov and K. Twomey
represents a recessed electrode with aspect ratio c = dp/re and performance of the
microelectrode depends on this ratio. The simulation results for a single inlaid electrode
and a recessed electrode with pore walls perpendicular to the electrode surface are shown
in Figs. 6 and 7. Figure 6a demonstrates transient currents for an inlaid disk electrode
(black curve) and for a recessed electrode with c = 0.2 (ra = 5 lm, dp = 1 lm)
(red curve) normalized to steady-state current of the inlaid electrode. Figure 6b shows
dependence of the normalized steady-state current of the recess electrode as a function of
the pore aspect ratio. Figure 7 presents the concentration profiles for the recess electrode
at different times after application of the stimulation voltage: at 0.01 s - Fig. 7a, at 0.1 s -
Fig. 7b, at 1 s - Fig. 7c and at 10 s - Fig. 7d. As can be seen from comparison of the inlaid
and recess electrodes behavior the existence of the electrode recess leads to decrease of
the steady-state current that can be quite significant for the large aspect ratio. From the
concentration profiles follows that at start time (high frequencies) until concentration
wave is inside of the pore the diffusion processes in a recess electrode are similar to
macroelectrode behavior with a planar type of diffusion (Fig. 3a).
Fig. 6. Normalised transient current for an inlaid disk electrode (black curve) and for a recessed
electrode with c = 0.2 (ra = 5 lm, dp = 1 lm) (red curve) (a); the normalized steady-state
current as a function of the pore aspect ratio (b). (Color figure online)
A Portable Chemical Detection System with Anti-body Biosensor 55
The results on the electrode topology optimization formed the basis of the design of
the biosensor chip prototype. The biosensor was realized as an on-chip packageless
three electrode microelectrochemical cell, which includes a counter electrode (CE), an
reference electrode (RE) and a working electrode (WE) patterned on a single silicon
chip as shown in Fig. 8. The WE was located at the end of the chip to provide its
separation from RE and CE. Such an arrangement was implemented in order to
facilitate the bio functionalization chemistry and prevent both RE and CE from
interference with this process. The chip size was 10.16 37.70 mm. To provide
connectivity to the instruments the chip is equipped with 4 rectangular contact gold
pads of 2.50 5.00 mm dimensions with a standard 2.54 mm pitch. The CE was
made from platinum of 5.25 5.00 mm dimensions. The RE was a circular Ag/AgCl
electrode of 1.00 mm diameter. The WE represents an array of microdisk or microband
electrodes of different parameters as shown in the Table 2. The best results for the
biosensor were obtained with the microdisc array with 40 lm diameter recessed gold
disks, arranged in a hexagonal configuration with 400 lm center-to-center spacing
which subsequently underwent surface modification as will be described below.
The on-chip micro electrochemical cell was fabricated at the Central Fabrication
Facility at Tyndall National Institute. For fabrication of the chip was used a standard
photolithography and lift-off techniques with different masks associated with the cell
electrodes as schematically shown in Fig. 9 and previously discussed in [10, 24, 45]. In
brief this process can be described as follows. Firstly, a silicon oxide layer of 1 lm
thickness was thermally grown on the silicon substrate of N-Type, <111>-orientation
and alignments marks were patterned on it. The selection of the substrate type is related
to its high resistivity. Then the 20 nm thick titanium and 150 nm thick gold layers were
patterned on the wafers to prepare for electrodes, connecting tracks and pads. The
titanium layer acts as an adhesion layer, which improved the quality of the gold layer
Fig. 7. Concentration distributions around the electrode for the recessed configuration electrode
at the indicated times in s.
56 V. I. Ogurtsov and K. Twomey
Fig. 8. Biosensor chip layout (a) and photos of fabricated chip, working electrode array and a
single recessed microdisk (b).
deposit. Another 20 nm thick titanium adhesion layer was used before the deposition of
the 150 nm thick Pt and of 250 nm thick Ag metals on the gold connection areas for
the counter and reference electrodes where to promote the adhesion of Ag the Ti layer
A Portable Chemical Detection System with Anti-body Biosensor 57
Fig. 9. Fabrication process for the three electrode on-chip micro electrochemical cell.
The required sensor specificity to the defined biotarget was introduced through
application of a bio functionalization procedure which attached an antibody to the
surface of a working electrode. Three surface modification techniques were
58 V. I. Ogurtsov and K. Twomey
2.3 Instrumentation
Instrumentation architecture implements impedance spectroscopy technique and rep-
resents low noise, highly sensitive measurement unit (analog part) under microcon-
troller control (digital part). The hardware provided stimulus signal, the front-end
control, conditioning of the biosensor signal, its amplification, filtration, frequency
transformation, data acquisition and communication with a control embedded
A Portable Chemical Detection System with Anti-body Biosensor 59
VSTIM jZF j
VOUT I ¼ cosðuS uF Þ ð3Þ
2 jZS j
VSTIM jZF j
VOUT Q ¼ sinðuS uF Þ ð4Þ
2 jZS j
where VOUT_I and VOUT_Q are the output voltages of the two signal paths, VSTIM is the
amplitude of the stimulation voltage; ZF is the feedback complex impedance with
module |ZF| and phase uF; ZS is the sensor complex impedance with module |ZS| and
phase uS. From the measured output voltages and the known feedback impedance the
module and phase of the sensor impedance ZS can be calculated as defined below
VSTIM jZF j
jZS j ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ð5Þ
2 V2 þ V2
OUT I OUT Q
VOUT Q
uS ¼ arctan uF ð6Þ
VOUT I
The method of signal extraction, which is used here, is called the lock-in technique.
The advantage of this method is the reduction of impact of transimpedance flicker (1/f)
noise on the measurement. This is because the noise in this technique can be reduced
60 V. I. Ogurtsov and K. Twomey
by a final low-pass filter. Thus, if this filter has a narrow bandwidth, only the noise
around the measurement frequency will influence the measurements.
The instrumentation architecture implemented in the system is shown in Fig. 12.
Voltammetric
Front-End Second stage ADC
channel
Amplifier amplifica on Mixer
Amplitude Amplitude
and offset and offset ADC
control control
Workin Low pass filter
Sensor
Mixer
ADC Data
Acquisi on
Sin(f0) Low pass filter & system
control To onboard PC
Cos(f0) USART
Mixer
Sin/Cos
Generator
ADC
Front-End Second stage
Amplifier amplifica on Low pass filter
Mixer
Amplitude Amplitude
μC
and offset and offset ADC
control control
JTAG
Low pass filter
Reference
Sensor
Stimulus
VDC1 VDC2
A1Sin(2π f0 +Δ φ )
described above but has some additional features to facilitate its practical implemen-
tation. A known AC voltage that is generated by the signal generator is applied across
the sensor and the current is converted to a voltage by the transimpedance amplifier.
The voltage is amplified and demodulated in the I/Q demodulator yielding real and
imaginary impedance components as described above. To increase the dynamic range
of the module, 2-bit electronically controlled instrumentation amplifiers are included in
the system structure after the transimpedance amplifier.
To Analog Board
From Sensor
Front-End
Amplifier
Potentiostat Temperature
Voltage stabilizer
sensor
a) b) c)
Fig. 14. Block diagram of potentiostat and front-end amplifiers module (a) and photographs of
the packaged unit with the sensor chip (b) and the unit PCB (c) (photographs are taken from
[24]).
Fig. 15. Photograph of signal processing and Micro-controller unit (the photograph is taken
from [24]).
the measurement mode at the selected frequency. Two command buttons in the upper
right corner allow for populating the table with default values (the upper large button)
or with parameters, which are set in the SpinEdit controls (the lower small button) for a
fast filling of table with the sweep parameter and hardware settings. Sweep parameters
include the start Fmin and finish Fmax frequencies and a number of frequencies per
decade, which are used to calculate frequency values distributed between Fmin and
Fmax. Hardware settings duplicate the tuning parameters from the Hardware Config-
uration panel that will be discussed below. Accurate adjustment of the sweep param-
eters at each table raw (frequency) is available with the help of the SpinEdit control that
can be activated in each cell. Frequency is tuned between the neighboring values in the
table; hardware settings are limited by the corresponding hardware specification. The
created Protocol table can be saved in the Excel file and conversely downloaded from
the file.
Frequency - ‘Zim vs F’, Real impedance part vs. Frequency - ‘Zre vs F’ and Phase vs.
Frequency - ‘Phase vs F’) plots as well as the processed Nyquist spectra obtained after
Kernel smoothing and subtraction of the reference background spectrum from the
measured spectra with analytical signal extraction (see the plot ‘Zim vs Zre processed’
where the analytical signals extracted are highlighted by red points on the corre-
sponding processed impedance spectra) and calculated analyte concentration in respect
to the biosensor calibration (plot ‘Anal. Signal vs. Concentration’) in both graphical
and table forms.
Fig. 17. Screen short of System configuration tab/window (the image is taken from [24]).
The measurements of imaginary Zim, real Zre, phase Ph values of the sensor
impedance is based on the Eqs. (3)–(6) and the corresponding module calibration. For
determination of the target concentration the special algorithm was developed. It is
based on analysis of equivalent circuit of the electrochemical biosensor followed by
experimental validation of the approach. The most used model for analysis of
impedimetric biosensors is the modified Randles equivalent circuit [52] shown in
Fig. 19. It includes resistor Rs that represents the resistance of the sample solution, in
which the biosensor is immersed, and the resistance of the working electrode itself and
the contacting track connecting the electrode with a corresponding pad presented an
external electrical contact of the biosensor; capacitor CD that arises from the series
combination of several capacitive elements, such as analyte binding (CAn) to a sensing
layer (CSe) on an background electrode (CBg) capacitances; charge transfer resistor Rct
that comprises the series connection of resistors incorporating the analyte binding (RAn)
to a sensing layer (RSe) on an electrode (RBg) and diffusion impedance ZD whose
frequency behaviour depends on the electrode system structure, shape and size of its
elements [53]. The CAn capacitance is most sensitive to binding of large species, such
A Portable Chemical Detection System with Anti-body Biosensor 65
Fig. 18. Screen short of EIS tab/window (the image is taken from [24]). (Color figure online)
as proteins or cells. One difficulty with capacitive based sensors is that their sensitivity
depends on obtaining the proper thickness of the sensing layer [54]. The charge transfer
resistor Rct can also be quite sensitive to analyte binding, particularly in case of large
species, such as proteins or cells, which significantly impede electron transfer.
It was found that for any (macroelectrode, microdisk or microband array) imple-
mentation of the biosensor transducer and any mechanism of manifestation of
biorecognition reaction through changes in charge transfer resistence or sensor
capacitance, the best results can be obtained through application of a differential
measurement scheme, realized by subtracting the reference spectrum from the mea-
sured impedance spectrum. In this case the measured spectrum is the biosensor
impedance spectrum after reaction with the target analyte and the reference spectrum is
the biosensor impedance spectrum before reacting with the target analyte or an
impedance spectrum obtained from the reference sensor if the last presents in the
system. After subtraction, the maximum value of the imaginary part of the differential
impedance spectrum can be taken as an analytical signal in the biosensor calibration
step and also determination of the target analyte concentration. Using this quantity as
an analytical signal instead of such equivalent circuit parameters as CD or Rct makes the
signal processing algorithm straightforward, thus it does not require high calculation
power from the control computer for its implementation. Additionally, this impedance
spectrum value is located in the middle of the frequency sweep. It reduces requirements
on the desired sweep range and therefore simplifies the hardware realization that is also
very important for the portable devices. Overall, the application of a differential
measurement scheme allows an increase in sensitivity and a reduction in the back-
ground signal. At the same time this approach suffers from noise. The injurious effects
66 V. I. Ogurtsov and K. Twomey
Fig. 19. The modified Randles equivalent circuit for biosensor analysis.
of noise can be reduced by both signals undergoing smoothing before their subtraction.
Different smoothing approaches including median, adaptive and Gaussian kernel
smoothing algorithms were tested to find the procedure that fits application specifi-
cation in the best way. It was found that Gaussian kernel smoothing method provides
required noise suppression and meets application requirement on simplicity and low
power of calculation. Since the window based smoothing methods start working
effectively only after the smoothing window is fully filled with samples the suggested
approach does not work in the beginning and at the end of the impedance frequency
sweep. This disadvantage is not critical in our case as the extracted parameter selected
as the analytical signal is located in the middle of the frequency sweep and this
limitation does not it affect.
Fig. 20. Experimental raw and smoothed (a) and processed differential impedance spectra with
extracted analytical signals (red points) (b) obtained with the developed biosensing system for
different T-2 toxin concentrations (0, 25 ppb, 50 ppb, 100 ppb, 250 ppb). (Color figure online)
can be seen in screenshot presented in Fig. 20. In order to provide better spectra
distinguishability the measurements were made without repetitions thus zero mea-
surement errors are shown in the corresponding table below the plots. Due to a rela-
tively high low boundary of the frequency sweep (10 Hz) the obtained impedance
spectra accounted only for high and middle frequency parts of a depressed semicircle
response of a microdisc electrode array associated with semi-infinite radial spherical
diffusion. When concentration of T2-toxin increases, real and imaginary impedance
parts both also increase that reflects a growth of a layer with targeted species captured
by antibodies due the antibody-antigen binding reaction. If toxin concentration exceeds
the certain level, this impedance increase is saturated due to a depletion of the free
antibodies capable of biorecognition reaction. The raw impedance spectra were sub-
jected to noise impact and application of Kernel’s smoothing algorithm allowed for
Fig. 21. Experimental smoothed Bode plots for real (a), imaginary (b) and phase (c) of the
immunosensor obtained with the developed biosensing system for different T-2 toxin
concentrations (0, 25 ppb, 50 ppb, 100 ppb, 250 ppb).
68 V. I. Ogurtsov and K. Twomey
Fig. 22. Calibration curve of the immunosensor obtained with the developed biosensing system
with five T-2 toxin concentrations (0, 25 ppb, 50 ppb, 100 ppb, 250 ppb). (Color figure online)
effective noise suppression. As was mentioned above the Kernel‘s smoothing cannot
not work in the beginning and the end of the frequency sweeping that explains the
signal oscillation in the plot with the smoothed differential spectra at high frequencies
(area of the small impedances in Fig. 20b). At the same time since for the analytical
signal are taken the maximal value of the imaginary part of the impedance located fare
away from this zone this attribute of the smoothing method does not affect its
measurement.
As follows from the smoothed Bode plots presented in Fig. 21 the most significant
difference between the spectra corresponding to the different toxin concentrations is
located in a low frequency range below 250 Hz. Here, the imaginary impedance values
are more than 2 times less than their real counterparts (80000 vs. 180000); they are also
more sensitive to variation of T2 toxin concentration. If frequency increases, both
impedance components decrease but the imaginary impedance stays relatively constant
at the frequencies below 250 Hz where the influence of the toxin concentration on the
biosensor impedance is most noticeable. This frequency behavior of the imaginary
impedance of the biosensor leads to the fact that frequency dependence of the impe-
dance phase represents unimodal smooth function with maximal values in the range of
56–62° around the frequency of 1 kHz. The dependence of imaginary impedance
against toxin concentration becomes more apparent from the differential impedance
spectra obtained by subtraction of the spectrum at zero concentration, which is used as
the reference background spectrum, from the impedance spectra corresponding to
nonzero toxin concentrations as demonstrated in Fig. 20b.
The maximum values of the imaginary impedances on the differential impedance
spectra marked by red points in this figure were used as the analytical signals to build a
calibration curve of the developed label-free impedimetric immunosensor. This curve
together with the analytical signals automatically extracted from the smoothed differ-
ential impedance spectra presented by the red squares is shown in Fig. 22.
A Portable Chemical Detection System with Anti-body Biosensor 69
The calibration curve represents a nonlinear function that reflects the competitive
nature of the antibody-antigen binding reaction. The model can be presented by Eq. (7)
y ¼ a lnðx þ bÞ þ c ð7Þ
where y is analytical signal, x is T-2 toxin concentration. Other parameters of the model
were determined by fitting the model with obtained experimental data. The following
values of these parameters were attained slope a = 2.667 ∙ 104, b = 22.449 and
c = −8.393 ∙ 104. The squared determination coefficient R2= 0.982 and the total
approximation error was 3.929 ∙ 103. As follows from the calibration curve, the
developed biosensing portable system can successfully detect biotoxin at the levels
below 25 ppm.
4 Conclusions
APTES – PDITC procedure. The instrumentation was realised as a two channel EIS
system, which used in-phase and quadrature demodulation of AC signal for extraction
of the real and imaginary impedance components. The biosensing system hardware
consisted of two modules: Potentiostat and Front-End Amplifiers (PFEA) and Signal
Processing and Microcontroller (SPM) units. The corresponding software capable of
autonomous operation contains two interrelated parts written for the host computer and
the embedded microcontroller. An automated procedure for analytical signal extraction
consists of: smoothing of the initial biosensor impedance spectrum by Kernel
smoothing procedure; subtraction of a reference background spectrum from the mea-
sured impedance spectrum; and extraction from differential impedance spectrum the
maximum of the imaginary component, which is used as an analytical signal for
biosensor calibration and calculation of the target analyte concentration. The PC
software was implemented with the option of user-friendly multi-tab GUI software
where each tab is associated with the program window. These are ‘System configu-
ration’, ‘Protocol’ and ‘EIS’ tabs/windows. They provide communication, instrumen-
tation and measurement setting control, and system operability testing; allow for design
of impedance measurement protocol for each frequency sweep point; and present
results of EIS and analyte concentration quantification in graphical and table forms.
The developed biosensing system was validated with the developed on-chip T2 toxin
biosensor. Calibration of the system was obtained for five toxin concentrations from
0 ppm to 250 ppm. It showed that the developed portable biosensing system can
provide successful detection of the biotoxin at the levels below 25 ppm.
References
1. Shriver-Lake, L.C., Ligler, F.S.: The array biosensor for counterterrorism. IEEE Sens. J. 5
(4), 751–756 (2005)
2. Ahmad, A., Moore, E.J.: Comparison of cell-based biosensors with traditional analytical
techniques for cytotoxicity monitoring and screening of polycyclic aromatic hydrocarbons in
the environment. Anal. Lett. 42(1), 1–28 (2009)
3. Brennan, D., Lambkin, P., Moore, E.J., Galvin, P.: An integrated optofluidic platform for
DNA hybridization and detection. IEEE Sens. J. 8(5), 536–542 (2008)
4. Bryan, T., Luo, X., Bueno, P.R., Davis, J.J.: An optimised electrochemical biosensor for the
label-free detection of C-reactive protein in blood. Biosens. Bioelectron. 39(1), 94–98 (2013)
5. Zhang, W., Du, Y., Wang, M.L.: Noninvasive glucose monitoring using saliva
nano-biosensor. Sens. Bio-Sens. Res. 4, 23–29 (2015)
6. Wang, X., Lu, X., Chen, J.: Development of biosensor technologies for analysis of
environmental contaminants. Trends Environ. Anal. Chem. 2, 25–32 (2014)
7. Yong, D., Liu, C., Zhu, C., Yu, D., Liu, L., Zhai, J., Dong, S.: Detecting total toxicity in
water using a mediated biosensor system with flow injection. Chemosphere 139, 109–116
(2015)
A Portable Chemical Detection System with Anti-body Biosensor 71
8. Herzog, G., Moujahid, W., Twomey, K., Lyons, C., Ogurtsov, V.I.: On-chip electrochemical
microsystems for measurements of copper and conductivity in artificial seawater. Talanta
116, 26–32 (2013)
9. Ogurtsov, V.I., Twomey, K., Herzog, G.: Development of an electrochemical sensing system
for environmental monitoring of port water quality to integrate on-board an autonomous
robotic fish. In: Hashmi, S. (ed.) Comprehensive Materials Processing. Sensor Materials,
Technologies and Applications, vol. 13, pp. 317–351. Elsevier Science, Oxford (2014)
10. Said, N.A.M., Twomey, K., Ogurtsov, V.I., Arrigan, D.W.M., Herzog, G.: Fabrication and
electrochemical characterization of micro- and nanoelectrode arrays for sensor applications.
J. Phys. Conf. Ser. 307, 012052 (2011)
11. Thevenot, D.R., Toth, K., Durst, R.A., Wilson, G.S.: Electrochemical biosensors:
recommended definitions and classification. Pure Appl. Chem. 71, 2333–2348 (1999)
12. Scheller, F.W., Wollenbergera, U., Warsinkea, A., Lisdata, F.: Research and development in
biosensors. Curr. Opin. Biotechnol. 12, 35–40 (2001)
13. Vo-Dinh, T., Cullum, B.: Biosensors and biochips: advances in biological and medical
diagnostics. Fresen. J. Anal. Chem. 366(6–7), 540–551 (2000)
14. Hock, B., Seifert, M., Kramer, K.: Engineering receptors and antibodies for biosensors.
Biosens. Bioelectron. 17(3), 239–249 (2002)
15. Wollenberger, U.: Third generation biosensors - integrating recognition and transduction in
electrochemical sensors. In: Gorton, L. (ed.) Biosensors and Modem Biospecific Analytical
Techniques, pp. 65–130. Elsevier, Amsterdam (2005)
16. Koyun, A., Ahlatcıoğlu, E., İpek, Y.K.: Biosensors and their principles. In: Kara, S. (ed.) A
Roadmap of Biomedical Engineers and Milestones. INTECH, Rijeka (2012)
17. Sharma, S., Byrne, H., O’Kennedy, R.J.: Antibodies and antibody-derived analytical
biosensors. Essays Biochem. 60(1), 9–18 (2016)
18. Byrne, B., Stack, E., Gilmartin, N., O’Kennedy, R.: Antibody-based sensors: principles,
problems and potential for detection of pathogens and associated toxins. Sensors 9(6), 4407–
4445 (2009)
19. Perumal, V., Hashim, U.: Advances in biosensors: principle, architecture and applications.
J. Appl. Biomed. 12(1), 1–15 (2014)
20. Dodeigne, C., Thunus, L., Lejeune, R.: Chemiluminescence as diagnostic tool. A review.
Talanta 51, 415–439 (2000)
21. Jiang, X., Li, D., Xu, X., Ying, Y., Li, Y., Ye, Z., et al.: Immunosensors for detection of
pesticide residues. Biosens. Bioelectron. 23(11), 1577–1587 (2008)
22. Yalow, R.S., Berson, S.A.: Assay of plasma insulin in human subjects by immunological
methods. Nature 184, 1648–1649 (1959)
23. Patel, Pd.: (Bio)sensors for measurement of analytes implicated in food safety. A review.
Trends Anal. Chem. 21(2), 96–115 (2002)
24. Ogurtsov, V.I., Twomey, K., Pulka, J.: A portable sensing system for impedance based
detection of biotoxin substances. In: 10th International Joint Conference on Biomedical
Engineering Systems and Technologies Proceedings, BIODEVICES, (BIOSTEC 2017),
Porto, Portugal, vol. 1, pp. 54–62 (2017)
25. Sadana, A.: Biosensors: Kinetics of Binding and Dissociation Using Fractals, 1st edn.
Elsevier, Amsterdam (2003)
26. Zhang, S., Zhao, H., John, R.: Development of a generic microelectrode array biosensing
system. Anal. Chimica Acta 421(2), 175–187 (2000)
27. Willner, I., Katz, E., Willner, B.: Layered functionalized electrodes for electrochemical
biosensor applications. In: Yang, V.C., Ngo, T.T. (eds.) Biosensors and Their Applications.
Elsevier, Amsterdam (2000)
72 V. I. Ogurtsov and K. Twomey
48. Vashist, S.K., Dixit, C.K., MacCraith, B.D., O’Kennedy, R.: Effect of antibody immobi-
lization strategies on the analytical performance of a surface plasmon resonance-based
immunoassay. Analyst 136(21), 4431–4436 (2011)
49. Raj, J., Herzog, G., Manning, M., Volcke, C., Maccraith, B., Ballantyne, S., Thompson, M.,
Arrigan, D.: Surface immobilisation of antibody on cyclic olefin copolymer for sandwich
immunoassay. Biosens. Bioelectron. 24, 2654–2658 (2009)
50. Barsoukov, E., Macdonald, J.R. (eds.): Impedance Spectroscopy Theory, Experiment, and
Applications, 2nd edn. Wiley, Hoboken (2005)
51. Lasia, A.: Electrochemical impedance spectroscopy and its applications. In: Conway, B.E.,
Bockris, J., White, R.E. (eds.) Modern Aspects of Electrochemistry, vol. 32, pp. 143–248.
Kluwer Academic/Plenum Publishers, New York (1999)
52. Suni, I.I.: Impedance methods for electrochemical sensors using nanomaterials. Trends Anal.
Chem. 27(7), 604–610 (2008)
53. Jacobsen, T., West, K.: Diffusion impedance in planar, cylindrical and spherical symmetry.
Elecrochimika Acta 40(2), 255–262 (1995)
54. Hu, S.Q., Wu, Z.Y., Zhou, Y.M., Cao, Z.X., Shen, G.L., Yu, R.Q.: Capacitive
immunosensor for transferrin based on an o-aminobenzenthiol oligomer layer. Anal. Chim.
Acta 458, 297–304 (2002)
Microfluidic Devices Integrating Clinical
Alternative Diagnostic Techniques Based
on Cell Mechanical Properties
1 Introduction
Innovative diagnostic tools are vital for accurate, early and prompt identification of
many diseases, which may save peoples’ lives. The fast development of microfluidics
opened a wide range of possibilities in developing point-of-care microfluidic devices
working as diagnostic tools, the so-called lab-on-chips.
Lab-on-chip systems swiftly evolved since their introduction by Manz [1] and are
now able to perform several programmed operations and biochemical analysis. One
major potential application of these microfluidic devices is in point-of-care diagnostics,
as they meet the various requirements identified by the World Health Organization for
the development of diagnostic tools for infectious diseases at resource-limited settings,
such as affordability, sensitivity, equipment free, robustness and portability [2]. Despite
claiming the ability to develop such systems at very low prices (approximately one
Euro), researchers working in this field recognize the need for further research, to
devise an effective system, prone to deliver the quick and accurate diagnostics required
in developing countries [3]. An important issue is related to the ambient conditions
which affect the fluid and therefore the samples transport. This issue, recently
approached by Moita et al. [4] is discussed in detail in this paper.
A paramount part of lab-on-chips development concerns the successful manipu-
lation and transport operations of the biosamples, which are mainly a matter of
momentum, energy and mass transport at liquid-solid interfaces. To cope with this, two
main streams were followed, namely the sample transport in microchannel-based
continuous microfluidics and the droplet-based digital microfluidics [5]. Although the
first has been widely explored in a large number of applications, there are still some
limitations associated to these systems, such as the fact that the functionality is not
generally reconfigurable after design and fabrication, limiting more flexible applica-
tions, the samples are difficult to access, the systems are difficult to clean and maintain,
clogging is more likely to occur and numerous auxiliaries such as tubes, valves, pumps,
among others, are usually required, thus increasing the complexity of these systems. In
addition, the efficiency of the valves and pumps at the micro-scale is still quite low,
turning these systems inefficient from the energetic point of view. Conversely, the
digital microfluidics based on droplets solves many of these issues, but the accurate
control of droplet dynamics is not yet well understood. Most of the droplet based
systems use a closed configuration with two plates, working as electrode and counter
electrode, respectively. Despite being stable, this configuration keeps some of the
problems pointed at the devices using microchannels, such as the clogging, the access
to the samples, among others. Alternatively, the open configuration systems have still
some problems to overcome. In the EWOD – Electrowettng on dielectric, the elec-
trodes are covered by a thin dielectric layer to avoid the occurrence of hydrolysis [6].
The dielectric material is often hydrophobic to promote the motion of the droplet. As
the droplet is deposited on the surface, the liquid-solid-vapour interfacial tensions are in
equilibrium, defining the angle formed between the tangent line at the droplet edge and
the surface, the so-called equilibrium contact angle [7]. Upon applying an electric
potential, the droplet acts as a capacitor and an electric force is generated at the
liquid-solid-vapour interface, unbalancing the interfacial tensions and decreasing the
contact angle. This mechanism is classically described by the Young-Lippmann
equation [7, 8] and despite the basic principles of EOWD are grounded for many
decades, as revised for instance by [9], various alternative models (e.g. [10, 11]) strive
to explain some unclear phenomena such as the contact angle saturation (i.e. the
limiting value of the angle bellow which it remains unchanged, independently of the
imposed electric potential), which are not predicted by Young-Lippmann theory.
The EWOD is therefore strongly dependent on the electrochemical properties of the
aqueous solutions of the droplets. However, information reported in the literature
concerning the transport of biofluids is actually quite sparse. Sirivasan et al. [12] report
the successful transport of physiological fluids. Also, several authors report the
effective electrowetting-induced transport of proteins and DNA [13], but it is not clear
which are the most suitable electrochemical properties of the fluids or the most
important parameters governing biofluids transport and manipulation. Adsorption of
the biocomponents on the dielectric substrate is also a problem that is not completely
described yet. For instance, [14, 15] report high adsorption of biocomponents and
76 A. S. Moita et al.
2 Experimental Procedure
The experimental procedure described in this section addresses the various steps fol-
lowed in the microfabrication of the chips, the protocol defined to characterize the chips
performance (mainly focusing on the transport and manipulation section) and diag-
nostic procedures to characterize the wettability and infer on the adsorption of the
biocomponents.
film by lithography and wet etch on a glass substrate with 102 45 mm2 and 700 lm
width. Finally, a thin film of a dielectric material was deposited on the chip assembly,
without covering the contacts. The different configurations tested only vary in the
electrodes width, w (between 120 and 1400 lm), being the numerous interdigitated
coplanar electrodes displaced with a fixed distance between them, 2a = 60 lm. The
length of the electrodes is 24 mm (Fig. 1a and b).
Fig. 1. (a) Definition of the main geometrical parameters for the chip dimensioning. (b) Chip
sample.
Since the coplanar electrodes have different polarities, the droplet must cover at
least two electrodes. These basic dimensions were taken following the results obtained
in the fundamental study previously completed [4]. Then, additional calculations on the
chip capacitances and other electrical parameters were performed, following the rec-
ommendations of Chen et al. [21]. The best performing configurations were then
selected empirically, based the analysis of the chips performance, as discussed in the
following sections.
Table 1. Physico-chemical properties of the solutions and suspensions used in the present
work [22].
Solution Density q Surface tension rlv Dynamic viscosity
[kg/m3] [mN/m] l [Ns/m2]
GFP (1.71 10−3 mM) 998 72.2 ± 0.7 1 10−3
GFP-expressing E. coli 998 73.8 ± 0.04 1 10−3
(1 109 cells/ml)
GFP-expressing E. coli 998 73.8 ± 0.04 1 10−3
(2 109 cells/ml)
all the solutions used here is very close to that of water, as shown in Table 1, which
summarizes the physico-chemical properties of the fluids used in the present work.
samples within the droplets (micro-to-nano scale) and the paramount role of wetta-
bility, an alternative method is explored to provide more accurate measurements as
proposed in Vieira et al. [25].
The performance of the chips is evaluated grounded on the dynamic response of the
droplets on the chips, which in turn is discussed based on the evaluation of the temporal
evolution of the droplet-surface contact diameter. The velocity of the contact line
motion is also evaluated. These quantities are determined from high-speed visualization
and post-processing. The high-speed images are taken at 2200 fps using a Phantom
v4.2 from Vision Research Inc., with 512 512 pixels@2100 fps resolution. For the
present optical configuration, the spatial resolution is 25 lm/pixel and the temporal
resolution is 0.45 ms. The post-processing is performed using a home-made routine
developed in Matlab. Temporal evolution of the contact (spreading and receding)
diameters is presented as the average curve of at least 3 events, obtained at similar
experimental conditions. Accuracy of the measurements is evaluated to be ±25 lm for
the contact diameter.
Regarding the actuation of the chips, although AC voltage is reported by some
authors to lead to better performances of the electrowetting systems, decreasing the
contact angle hysteresis and delaying the contact angle saturation [10, 26] these phe-
nomena seem to be more related to the wetting characteristics of the surfaces and with
the properties of the solutions. Also, to avoid the occurrence of limiting frequencies for
which the Lippmann equation is not satisfied [9], DC current was used on the chips
provided by a Sorensen DCR600-.75B power supply. The applied voltage was varied
from 0 to 245 V. However, since the basic principle for droplet motion in microchips
requires switching polarities between neighboring electrodes, within an imposed fre-
quency and following a similar approach explored by Fan et al. [27], a custom-made
frequency controller using an Arduino was applied on the chips using a square wave, to
assure the generation of an electrical force in the direction of the desired motion. This
wave defines the time during which the electrode is actuated by a certain applied
voltage. It has an internal clock with a frequency of 16 MHz. This device can vary the
imposed frequency between 50 Hz and 400 Hz, within 50 Hz increments. The negative
polarity of the chips was grounded, so, it will have a zero potential, i.e. the overall
potential difference applied to the chip corresponds to the voltage applied to the pos-
itive polarity of the power supply.
Fig. 2. Comparison between the equilibrium angles measured with the tensiometer and with the
LSCFM technique, for a smooth glass surface. Adapted from [22].
82 A. S. Moita et al.
Fig. 3. Dispersion analysis of the static contact angles measured with the tensiometer and with
the LSCFM technique, measured for smooth glass surfaces.
results, the safest way to alter wettability, avoiding the aforementioned issues, is
towards the chemical modification and/or selection of the appropriate dielectric
materials, as discussed in the following sub-section, while keeping the surface as
smooth as possible.
3 Device Configuration
The microfluidic device under development has three main working sections, namely,
the transport section, the diagnostic section and the selection section. The preferred
wetting properties for the materials selected in each of these sections can be quite
different. Hence, while in the transport section, droplet motion is governed by elec-
trostatic actuation aided by custom made wetting properties of the dielectric material,
so that hydrophobic/superhydrophobic regions are preferred to minimize the adhesive
forces, which are proportional to hysteresis and consequently minimize the energy
dissipated at the contact line between the droplet and the surface, in the diagnostic
section it is desirable to promote adhesion to constrain the sample in the sensor area.
Hence and following the arguments presented in the Introduction, the wettability
plays here a paramount role, so its characterization must be as accurate as possible. The
materials, namely the dielectric material, which is in contact with the biosolutions, must
therefore be chosen with care. For the transport and selection sections, the chosen
materials should maximize the hydrophobicity with the fluids in study, minimize the
hysteresis (to minimize the energy dissipation at the droplet contact line) and minimize
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 83
the adsorption, which locally affects the wettability, increasing the adhesion and,
consequently the energy dissipation [4]. Opposite trends are desired for the diagnostics
section. The various steps followed towards the selection of the materials is discussed
in the following sub-section.
Table 2. Equilibrium contact angles, obtained for each pair fluid-dielectric substrate considered
in the present work [22].
Dielectric coatingContact angle [°]
Water E-coli GFP
Teflon 112 ± 5 103 ± 6 121 ± 6
Teflon with Glaco 145 ± 1 141 ± 9 153 ± 3
PDMS 121 ± 1 112 ± 1 119.5 ± 0.4
PDMS with Glaco 153 ± 3 153 ± 2 155 ± 3
SU8 resist 67.1 ± 0.7 65 ± 2 71.8 ± 0.2
SU8 with Glaco 160 ± 7 162 ± 1 153 ± 4
Si3N4 64.1 ± 0.7 59 ± 4 65 ± 2
84 A. S. Moita et al.
Fig. 4. Contact angle hysteresis evaluated for GFP solution (1.71 10−3 mM), GFP-expressing
E-coli suspension (1 109 cells/mL) and distilled water on the tested dielectric substrates [22].
Concerning adsorption, Moita et al. [4] report that the GFP was adsorbed by Teflon
substrates, leading to a local increase of surface wettability and further contributing to
preclude the receding motion, as this wettability increase is irreversible, taking the
contact angles to values near saturation. In this context, the following analysis infers on
the possible adsorption of the GFP and E-coli cells by the PDMS and SU8 substrates,
as depicted in Fig. 5, which shows the TCDF value, as defined in Sect. 2, evaluated for
each substrate.
Figure 5a evidences minimum TCDF values for the adsorption of the GFP on
PDMS substrates. The alternative material with lower adsorption of the biocomponents
tested here was SU8, but the TCDF values obtained are about one order of magnitude
higher than those evaluated for PDMS. Even though the adsorption of cells is quite less
significant than the adsorption of the protein, Fig. 5b suggests a minor effect of the cell
concentration in the solutions in promoting passive adsorption mechanisms, i.e. leading
to slightly higher TCDF values. Nevertheless, adsorption of E-coli cells was likewise
observed to be minimum on the PDMS substrate. The application of the Glaco®
coating is observed to further reduce the adsorption of both proteins and cells,
decreasing the TCDF values in about one order of magnitude on the PDMS.
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 85
Fig. 5. Adsorption of the biocomponents (a) GFP and (b) E-coli cells by the dielectric substrates
quantified by the TCDF.
This analysis determined that the best dynamic response was found for the chips with
electrodes width w = 1200 lm and a distance between consecutive electrodes
2a = 60 lm. The detailed study performed is reported in [28].
The wetting properties have an important role in this dynamic response also,
together with other parameters such as, for instance, the thickness of the dielectric
layer, which is inversely proportional to the change in the contact angle during elec-
trostatic actuation, as predicted by Young-Lippmann equation.
Hence, the spreading diameter of GFP and E-Coli suspension droplets, was eval-
uated under electrostatic actuation for the test chips coated with PDMS, SU8 resist and
both materials further coated with Glaco®. As discussed in the previous sub-section,
this coating reduces the adsorption of the proteins and of the cells by the substrates, so
the irreversible contact angle decrease associated to the adsorption mechanisms is also
minimized. The coating also decreases substantially the contact angle hysteresis. This
modification of the wetting properties has a dramatic effect on the droplet transport on
the chips, as shown in Fig. 6 which depicts the temporal evolution of the spreading
diameter d(t) of a GFP droplet under electrostatic actuation, made non-dimensional
with the initial diameter of the deposited droplet, i.e. for 0 V (d(t)/d0V). t = 0 ms
corresponds to the beginning of the actuation.
Fig. 6. Electrowetting induced spreading diameter of GFP droplets for different dielectric
substrates, between coplanar electrodes for the configuration w = 1200 µm, at 230 V and
350 Hz. Initial droplet diameter is 2.8 mm.
The figure shows a significantly larger spreading diameter with the SU8 resist
coated with Glaco®, being followed by a pronounced recoiling, allowed by the
noteworthy hysteresis reduction. This recoiling will in turn allow the droplet to spread
again towards the next electrode, in a continuous motion. The worse response of the
droplet to the actuation obtained for PDMS coated with Glaco® is associated to the
large final thickness of the substrate, which, according to Young-Lippmann equation,
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 87
precludes the decrease of the contact angle under actuation. Despite this reduced
response, evident recoil is still observed.
However, as one looks at the overall maximum variation of the non-dimensional
droplet diameter and of the velocity of the contact line, as a function of different
imposed frequencies, the behavior of the droplet on the materials coated with Glaco® is
not particularly improved in the sense that the droplet does not depict the largest
diameters (Fig. 7a) or the highest velocities (Fig. 7b).
Fig. 7. Effect of the substrate wetting properties on the dynamic response of the droplets,
namely on (a) the maximum spreading dimensionless diameter and (b) the maximum spreading
velocities of GFP (1.71 10−3 mM) and GFP-expressing E-coli suspension (1 109 cells/mL),
for droplets moving between coplanar electrodes, for the configuration w = 1200 µm, for the
configuration w = 1200 µm, 2a = 60 lm. Initial droplet diameter is 2.8 mm.
88 A. S. Moita et al.
Once again, this result is attributed to the final thickness of the dielectric substrates,
which precludes the actuation. Hence, the final device must be fabricated under strict
control of the thickness of the dielectric layers used.
Droplet response is also affected by the nature and physico-chemical properties of
the solutions. Figure 8 depicts the dimensional diameter and velocity of droplets of
GFP and E-Coli solutions, as a function of the imposed frequency, for the same chip
configuration used in the previous analysis.
Fig. 8. Effect of the solution composition on the dynamic response of the droplets, namely on
(a) the maximum spreading dimensionless diameter [22] and (b) the maximum spreading
velocities of GFP (1.71 10−3 mM) and GFP-expressing E-coli suspension (1 109 cells/mL),
for droplets moving between coplanar electrodes, for the configuration w = 1200 µm,
2a = 60 lm. Initial droplet diameter is 2.8 mm.
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 89
The figure clearly shows the better response of the droplets of protein solutions,
although the surface tension and density only vary slightly between the protein solution
and the cell suspensions (Table 1). Hence, this particular behaviour is possibly
attributed to local variations in the wettability or in the reorganization of the electrical
density of the solution due to cell migration to the contact line region. However, this
particular behaviour must be further studied in the near future, by characterizing the
flow near the contact line.
As discussed in the previous sections, after actuation, the droplet suffers in many
cases an irreversible spreading process and the surface becomes locally hydrophilic. In
a qualitative approach, under these circumstances, the evaporation of the droplet
promotes the decreasing of the contact angle, while the wetting contact area remains
approximately constant due to the pinning of the contact line [31].
Looking at our experimental results, in the assay performed at HR = 99.90% two
stages can be identified (Fig. 9a): the first, where the liquid-solid contact area remains
constant, as the contact angle decreases, in agreement with [31], transiting at
t = 58.37 min for the second stage, where both the equilibrium contact angle and the
contact area liquid-solid decrease. The height of the droplet decreases continuously
over time (Fig. 9b).
Fig. 9. Evaluation of droplet evaporation by mass diffusion. Temporal evolution of: (a) equi-
librium contact angle and (b) droplet height and diameter, for the planes XZ and YZ of the
droplet profile. Din and hin stand for the initial diameter and height of the deposited droplet,
respectively.
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 91
The results further show that the contact angle decreases by about 53% of its
absolute initial value. In addition, there is a reduction of the droplet height, evaluated in
both planes XY and XZ, of approximately 80% in comparison to its initial height hin.
Also, the diameter of the deposited droplet decreases in 54% and 32% in planes XZ and
YZ, respectively, compared to the initial values Din.
For HR = 88.50%, the time resolution (2.25 min) is lower relatively to the HR =
99.90% assay (4.49 min) and does not allow identification of the two aforementioned
evaporation states. Regarding the contact angle, a linear decrease of 68% and 60% is
observed for the XZ and YZ planes, respectively – Fig. 9a.
Furthermore, the diameter of the contact area decreases by 36% (perspective XZ)
and 17% (perspective YZ) in comparison to its initial value. In turn, the height
decreases linearly throughout the whole period by roughly 67% of its initial height for
both planes – Fig. 9b. These trends are qualitatively in agreement with those reported
by Sobac and Brutin [32]. Quantitively, additional experiments must be performed to
take more sustained conclusions. For instance, the relative humidity measurements
require a more accurate evaluation within the entire chamber, since the evaporative
rates are quite high, even for large values of HR, which may be associated to an actual
global value of HR lower than that locally measured by the sensor. Despite this
limitation, this rough evaluation indicates that the evaporation of the droplets is likely
to occur even at controlled ambient conditions, so that a lab-on-chip system to be used
in harsh environments (e.g. countries with high temperatures and dry climate) should
address the possibility of a local increase of the HR inside the chip, for the time
required for the analysis.
4 Conclusions
References
1. Manz, A., Widmers, H.M., Graber, N.: Miniaturized total chemical analysis systems: a novel
concept for chemical sensing. Sens. Actuators B Chem. 1(1–6), 244–248 (1990)
2. Yager, P., Edwards, T., Fu, E., Helton, K., Nelson, K., Tam, M.R., Weigl, B.H.: Microfluidic
diagnostic technologies for global public health. Nature 442(7101), 412–418 (2006)
3. Dance, A.: The making of a medical microchip. Nature 545, 512–514 (2017)
4. Moita, A.S., Laurência, C., Ramos, J.A., Prazeres, D.M.F., Moreira, A.L.N.: Dynamics of
droplets of biological fluids on smooth superhydrophobic surfaces under electrostatic
actuation. J. Bionic Eng. 13, 220–234 (2016)
5. Geng, H., Feng, J., Stabryl, L.M., Cho, S.K.: Dielectroetting manipulation for digital
microfluidics: creating, transporting, splitting, and merging droplets. Lab Chip 17, 1060–
1068 (2017)
6. Berge, B.: Electrocapillarity and wetting of insulator films by water. Acad. Sci. Ser. II Mec.
317, 157–163 (1993)
7. Young, T.: An essay on the cohesion of fluids. Phil. Trans. R. Soc. Lond. 95, 65–87 (1805)
8. Lippmann, G.: Relation entre les phénomènes électriques et capillaires. Ann. Chim. Phys. 5,
494–549 (1875). (in French)
9. Mugele, F., Baret, J.C.: Electrowetting: from basics to applications. J. Phys. Condens. Matter
17, R705–R774 (2005)
10. Jones, T.B.: An electromechanical interpretation of electrowetting. J. Micromech. Microeng.
15, 1184–1187 (2005)
11. Bahadur, V., Garimella, S.V.: An energy-based model for electrowetting-induced droplet
actuation. J. Micromech. Microeng. 16, 1494–1503 (2006)
12. Srinivasan, V., Pamula, V.K., Fair, R.B.: An integrated digital microfluidic lab-on-a-chip for
clinical diagnostics on human physiological fluids. Lab Chip 4, 310–315 (2004)
13. Wheeler, A.R., Moon, H., Kim, C.J., Loo, J.A., Garrell, R.L.: Electrowetting-based
microfluidics for analysis of peptides and proteins by matrix-assisted laser
desorption/ionization mass spectrometry. Anal. Chem. 76, 4833–4838 (2004)
14. Rupp, F., Axmann, D., Ziegler, C., Geis-Gerstorfer, J.: Adsorption/desorption phenomena on
pure and Teflon® AF-coated titania surfaces studied by dynamic contact angle analysis.
J. Biomed. Mater. Res. 62, 567–578 (2002)
15. Yon, J.Y., Garrell, R.L.: Preventing biomolecular adsorption in electrowetting-based
biofluidic chips. Anal. Chem. 75, 5097–5102 (2003)
16. Suresh, S., Spatz, J., Mills, J.P., Micoulet, A., Dao, M., Lim, C.T., Beil, M., Seufferlein, T.:
Connections between single-cell biomechanics and human disease states: gastrointestinal
cancer and malaria. Acta Biomater. 1, 15–30 (2005)
17. Cross, S.E., Jin, Y.-S., Rao, J., Gimzewski, J.K.: Nanomechanical analysis of cells from
cancer patients. Nat. Nanotechnol. Lett. 2, 780–783 (2005)
Microfluidic Devices Integrating Clinical Alternative Diagnostic Techniques 93
18. Gosset, G.R., Tse, H.T.K., Lee, S.A., Ying, Y., Lidgren, A.G., Yang, O.O., Rao, J., Clark,
A.T., Di Carlo, D.: Hydrodynamic stretching of single cells for large population mechanical
phenotyping. PNAS 109(20), 7630–7635 (2009)
19. Remmerbach, T.W., Wottawah, F., Dietrich, J., Lincoln, B., Wittekind, C., Guck, J.: Oral
cancer diagnosis by mechanical phenotyping. Cancer Res. 69(5), 1728–1732 (2009)
20. Tse, H.T.K., Gosset, D.R., Moon, Y.S., Masaeli, M., Sohsman, M., Ying, Y., Mislick, K.,
Adams, R.P., Rao, J., Di Carlo, D.: Quantitative diagnosis of malignant pleural effusions by
single-cell mechanophenotyping. Sci. Transl. Med. 5(212), 1–9 (2013)
21. Chen, J.Z., Darhuber, A.A., Troian, S.M., Wagner, S.: Capacitive sensing of droplets for
microfluidic devices based on thermocapillary actuation. Lab Chip 4(5), 473–480 (2004)
22. Vieira, D., Mata, F., Moita, A.S., Moreira, A.L.N.: Microfluidic prototype of a lab-on-chip
device for lung cancer diagnostics. In: Proceedings of the 10th International Joint
Conference on Biomedical Engineering Systems and Technologies - Volume 1:
BIODEVICES, Porto, Portugal, 21–13 February 2017, pp. 63–68 (2017). https://doi.org/
10.5220/0006252700630068. ISBN: 978-989-758-216-5
23. Cheng, P., Li, D., Boruvka, L., Rotenberg, Y., Neumann, A.W.: Automation of
axisymmetric drop shape analysis for measurements of interfacial tensions and contact
angles. Colloids Surf. 43(2), 151–167 (1990)
24. Kietzig, A.M.: Comments on “an essay on contact angle measurements” – an illustration of
the respective influence of droplet deposition and measurement parameters. Plasma Proc.
Polym. 8, 1003–1009 (2008)
25. Vieira, D., Moita, A.S., Moreira, A.L.N.: Non-intrusive wettability characterization on
complex surfaces using 3D laser scanning confocal fluorescence microscopy. In: 18th
International Symposium on Applications of Laser and Imaging Techniques to Fluid
Mechanics, Lisbon (2016)
26. Chen, L., Bonaccurso, E.: Electrowetting - from statics to dynamics. Adv. Colloid Interface
Sci. 210, 2–12 (2014)
27. Fan, S.-K., Yang, H., Wang, T.-T., Hsu, W.: Asymmetric electrowetting–moving droplets by
a square wave. Lab Chip 7(10), 1330–1335 (2007)
28. Mata, F., Moita, A.S., Kumar, R., Cardoso, S., Prazeres, D.M.F., Moreira, A.L.N.: Effect of
surface wettability on the spreading and displacement of biofluid drops in electrowetting. In:
Proceedings of ILASS – Europe 2016, 27th Annual Conference on Liquid Atomization and
Spray Systems, September 2016, Brighton, UK, 4–7 September 2016 (2016). ISBN
978-1-910172-09-4
29. Adamo, A., Sharei, A., Adamo, L., Lee, B., Mao, S., Jensen, K.F.: Microfluidics-based
assessment of cell deformability. Anal. Chem. 84, 6438–6443 (2012)
30. Moita, A.S., Herrmann, D., Moreira, A.L.N.: Fluid dynamic and heat transfer processes
between solid surfaces and non-Newtonian liquid droplets. Appl. Therm. Eng. 88, 33–46
(2015)
31. Picknett, R., Bexon, R.: The evaporation of sessile or pendant drops in still air. J. Colloid
Interface Sci. 61(2), 336–350 (1977)
32. Sobac, B., Brutin, D.: Triple-line behavior and wettability controlled by nanocoated
substrates: influence on sessile drop evaporation. Langmuir 27(24), 14999–15007 (2011)
A Biochip Based Medical Device
for Point-of-Care ABO Compatibility:
Towards a Smart Transfusion Line
Abstract. ABO mismatch between donor and patient’s blood is still the cause
of accidents which are sometimes lethal. The main causes of mis-assignment are
human errors and wrong identification of patients or blood product. Only a final
compatibility test at the patient’s bedside can avoid these errors. In some
countries, this test is performed using manual procedures. This does not prevent
from human manipulation and interpretation errors. In this paper, we present a
prototype able to automatically perform a final ABO compatibility test. It relies
on the use of disposable antibodies grafted biochips inserted into a mobile
reader/actuator. Red blood cells are selectively captured on biochips grafted with
antibodies complementary of antigens present on the cells surface. Detection of
captured cells is based on optical absorption techniques. So far, our device
achieved blood compatibility test with 99.3% sensitivity and 97.9% specificity.
1 Introduction
In countries for which the hemo-vigilance is reliable and where a unique test is
performed, the ratio of adverse effects due to ABO incompatibility approaches 1/40000
red cell concentrates (RCC). This was the case in France before 2003 when only ABO
compatibility was tested. After this date, the use of both concordance and ABO tests at
the patient’s bedside reduced the adverse effects to about 1/600000. However, in most
countries, only the concordance test is considered.
There is a need for a second test at the patient’s bedside in order to reduce the
number of ABO errors. There exist ABO test cards, but they rely on delicate manual
operation and require a long and specific training. Furthermore, the compatibility card
requires a human interpretation of the agglutination test. Therefore, this method is a
source of various errors: manipulation, reading and interpretation difficulties. Also,
medical staff is exposed to blood when sampling patient’s blood. Concerning the
interpretation difficulty, iso-group compatibility is relatively straightforward but non
iso-group compatibility still leads to interpretation errors or stressful situations.
For these reasons, a point-of-care device able to automatically perform an ultimate
compatibility test with minimum manipulation and without human interpretation would
be profitable, in particular when considering the increase of blood product distributed
during the last decade. For example in France an increase of the RCC delivery of
almost 24% has been observed between 2000 and 2011 [1]. In 2016, more than 3
million of RCC where distributed [2].
Several methods have been proposed for blood typing. They are mainly based on
gel agglutination [3, 4]. SPR [5–7] and Surface Plasmon Resonance imaging (SPRi)
[8–11] techniques can also be used. However, these studies demonstrate the possibility
to detect captured cells with commercial laboratory apparatuses. Therefore, a direct
translation to the patient’s bedside may be difficult because the entire device used
should be re-thought for point-of-care use.
Recently, long-range surface plasmon-polaritons to detect red blood cells
(RBC) selectively captured by the surface chemistry was demonstrated [12]. However,
because packed RBC must be diluted in a buffer of controlled refractive index,
translation of the device to clinical use is still challenging. Techniques based on image
processing on plate test have been reported [13, 14]. In this case, image processing is
used to objectively observe and interpret red cell agglutination obtained manually.
Issues concerning blood and antibodies manipulation still exist. Spectroscopic methods
have also been reported [15, 16]. However, the use of an optical spectrometer to
measure absorption of diluted red cells may be difficult in clinical practice. In fact,
although these new devices are able to realize blood typing by objectively reading
agglutination, they still require hard translational research work before to be installed in
the patient’s room.
In this paper, we present a mobile device meant to address the above mentioned
issues. The main idea is to replace the four reaction zones of the current manual
compatibility card with four IgMs grafted biochips inspired from Surface Plasmon
Resonance (SPR) and SPRi biochips. Hemagglutination is therefore replaced by red
cell capture. The detection of captured red cells rely on a simple optical absorption
technique. Biochips are inserted in a mobile reader/actuator which drives the fluids
96 K. Charrière et al.
(blood, red cell concentrate (RCC) and physiological serum), performs the optical
reading and final interpretation. This concept is currently protected by two patents [17,
18] and was describing in a previous paper [19].
Research actions to set-up this device include four main steps. The first series of tests
consisted in studying the IgMs grafting and red cell capture using SPR and SPRi
methods with homemade biochips. This has been previously reported [20, 21]. The
second set of experiments consisted in translating the SPR biochip to biochips inserted
into cartridges and to detect the capture of red cells in these half-bulk conditions together
with the correlation between the number of captured cells and optical reading [19].
The third part of the experiments is the subject of the current publication. It consists
in using a large number of whole blood (WB) and RCC samples to test the automated
fluid flow control, optical reading and software interpretation of the ABO compatibility
result. The goal is to determine sensitivity and specificity of the device together with
the blood group concordance between what is expected and what the device reads. We
also studied the performance of the device according to the age of RCC. The last part of
the work consists in experimenting the use of the venous return to drive patient’s blood
into the device with reduces risk of exposure to blood for medical staff [18].
F.
E. E.
C.
AnƟ A AnƟ B
Fig. 1. Views of the device. A.B. There is 1 cartridge for the patient and another one for the red
cell concentrates. C. Each cartridge includes 2 biochips grafted with anti-A and anti-B antibodies.
D. Optical clamps are used to detect the capture of red cells at the chip’s surface. E. Fluids are
driven using syringes controlled by the internal micro-processor. F. The human-machine
interface allows setting the opto-fluic parameters and displaying compatibility test results.
In what follows, positive chips are defined as chips that have captured red cells,
regardless of the blood group. Conversely, negative chips correspond to chip with no
capture.
Fluids (blood, RCC and physiological serum) are driven by means of automated
syringes controlled via dedicated software. This software also drives the optical
measurement, human-machine interface and USB connection to a PC for data recording
and processing.
3 Results
3.1 Device’s Testing
The device was tested using 148 blood aliquots. This represents 296 biochips and
therefore 148 cartridges. Blood comes from both RCC and WB. Samples were pro-
vided by the Etablissement Français du Sang in accordance with the ethical rules and
with informed consent obtained from donors.
Among these 296 chips, 4 are not taken into account because errors occurred while
assembling the cartridges.
Therefore, only 292 biochips were tested. Remember that 2 biochips are required to
test 1 sample. For two samples, inversions of the anti-A and anti-B biochips were
made. Although the biochips behave correctly and are taken into account for biochip
testing, the corresponding samples were not taken into account for compatibility
testing. At the end, 142 samples were tested for compatibility.
The repartition of samples in terms of RCC, WB and blood group is given in
Table 1.
Fig. 2. Absorbance versus positive or negative biochips. Box plot: 5–95 percentiles.
Kruskal-Wallis test followed by Dunn’s multiple comparison tests. *** p < 0.001. Negative
values are due to a slight drift of the electronics. Red diamond shows cases of mis-assignment
when the threshold is set to 0.05. (Color figure online)
The best absorbance threshold to discriminate between positive and negative bio-
chip was set to 0.05 (minimization of mis-assignments). In this way, only 4 errors
occurred. One biochip represents a false negative. For it, not enough red cells were
captured although the biochip should have been positive. Indeed, red cell capture is not
homogenous on the surface, maybe due to an antibodies grafting problem. This means
that 1 patient of group AB was detected as A. Three other biochips were false positive.
For them a strong non-specific retention of red cells was recorded due to washing
problem. This means that 1 patient of group O was detected as A and 2 patients B
detected as AB.
From these results and differentiating between A and B biochips allows calculating
sensitivity and specificity of the device. This is summarized in Table 2. Almost all
sensitivities are 100%, except for anti-B biochips (97%) used with WB (false negative
described earlier). It is the same for specificities: all biochips are 100% specific, except
the anti-A biochips used with WB (3 false positives described previously).
Table 3 presents the same parameters regardless of the blood type and for the entire
device. At the end, specificity of the device is 99.3% and specificity is 97.9%.
Improving fabrication of the cartridges would probably resolve these mis-assignments
and improved sensitivity and specificity.
100 K. Charrière et al.
RCC WB
Anti-A Anti-B Anti-A Anti-B
Number of Biochips 82 78 68 64
Expected positives 36 39 39 33
Recorded positives 36 39 39 32
Sensitivity 100% 100% 100% 97%
Expected negatives 46 39 29 31
Recorded negatives 46 39 26 31
Specificity 100% 100% 89.7% 100%
Table 3. System performance (in terms of antibodies and for the entire device).
RCC + WB
Entire device
Anti-A Anti-B
Number of Biochips 150 146 292
Expected positives 75 72 147
Recorded positives 75 71 146
Sensitivity 100% 98.6% 99.3%
Expected negatives 75 70 145
Recorded negatives 72 70 142
Specificity 96% 100% 97.9%
Biochips performance was also tested as a function of the age of the blood dona-
tion. In this case, only RCC were considered because WB is meant to be fresh. For this
test we used 30 positive biochips with 6 to 8 days old donations, 31 negative biochips
with 6 to 8 days old, 29 positive with 43 to 44 days old and 40 negative with 43 to 44
days old. Figure 3 shows the absorbance obtained in terms of positivity/negativity. No
difference was observed between all kinds of positive biochips: 0.35 ± 0.011 for 6 to 8
days old RCC and 0.37 ± 0.01 for 43 to 4 days old RCC. The same is observed
between negative biochips: 0.001 ± 0.002 for 6 to 8 days old RCC and 0.005 ± 0.002
for 43 to 44 days old RCC. It is quite clear that the age of the blood donation does not
impact device performances (see Fig. 3).
For compatibility interpretation, 74 tests were performed. In all cases the software
delivered the right compatibility information, perfectly coherent with what happened at
the biochip surface. Of course, we mentioned cases where mis-assignments occurred.
However, this part of the test concerns the fact that the device delivers the right
information from the result of the optical measurement. For example, with sample
SO11 (WB of group O), a strong non-specific RCC retention has been observed on the
anti-A biochip, with an absorbance of 0.06 corresponding to a percentage of red cells of
A Biochip Based Medical Device for Point-of-Care ABO Compatibility 101
Fig. 3. Absorbance versus the age of the red cell concentrates in terms of positive and negative
biochips. Box plot: 5–95 percentiles.
21% on the biochip surface. This “patient” was considered A group. When testing the
compatibility with B group RCC, the device concluded that the transfusion should not
be allowed. Therefore the optical reading and the interpretation software work properly.
Experiments to test the ability of the device to correctly identify blood groups have
been conducted. Among 142 concordance tests performed 4 mis-assignments occurred.
The concordance performance is therefore 97% (see Table 4). Mis-assignments
reported here correspond to those already mentioned above.
Groups
A B AB O
N° of tests 19 25 14 20
RCC
Concordance (%) 100 100 100 100
N° of tests 17 13 20 14
WB
Concordance (%) 100 84.6 (2 m-a) 95 (1 m-a) 94.4 (1 m-a)
All groups
RCC + N° of test 142
WB Concordance (%) 97 (4 m-a)
102 K. Charrière et al.
The next experiment concerned the percentage of red blood cells on the biochips
surfaces as a function of the positive or negative nature of the biochips. While the
threshold in terms of absorbance was set to 0.05, the threshold in terms of percentage of
the surface covered with red blood cells is about 17.5% (see Fig. 4).
Fig. 4. Percentage of the biochip’s surface covered with red cells in terms of positive and
negative chips. RCC: red cell concentrate. WB: whole blood. Box plot: 5–95 percentiles.
Fig. 5. Frame captures of a video filmed while testing the efficiency of the blood venous return.
4 Discussion
5 Conclusion
To conclude, we believe that the concept described here may help enhancing blood
transfusion safety, not only in countries where a double ultimate test is already per-
formed, but especially in countries where only one test is considered. Furthermore,
such a device is meant to drastically reduce ABO compatibility accidents in countries
where the whole transfusion process (blood donation, conservation, delivery and
transfusion) is not yet fully satisfactory.
Acknowledgements. This work was partly supported by the EFS (grant DECO-13-0128), the
INSERM-CNRS (patent file CNRS/REF:02682-V), OSEO and the University of Franche-Comté
(grant A1105005I). This work is developed in the frame of the French RENATECH network and
the Biom’@x transversal axis at FEMTO-ST.
References
1. EFS: Rapport d’activité (2011). http://www.dondusang.net/rewrite/article/4900/efs/publicati
ons/feuilletez-en-ligne-le-rapport-d-activite-2011-de-l-efs.htm?idRubrique=790
2. ANSM: Bilans/Rapports d’activité - Bilans et rapports d’activité - ANSM : Agence nationale
de sécurité du médicament et des produits de santé. http://ansm.sante.fr/Mediatheque/
Publications/Bilans-Rapports-d-activite-Bilans-et-rapports-d-activite#folder_26762
3. Cid, J., Nogués, N., Montero, R., Hurtado, M., Briega, A., Parra, R.: Comparison of three
microtube column agglutination systems for antibody screening: DG Gel, DiaMed-ID and
ortho BioVue. Transfus. Med. 16, 131–136 (2006). https://doi.org/10.1111/j.1365-3148.
2006.00655.x
4. Langston, M.M., Procter, J.L., Cipolone, K.M., Stroncek, D.F.: Evaluation of the gel system
for ABO grouping and D typing. Transfusion 39, 300–305 (1999)
5. Malomgre, W., Neumeister, B.: Recent and future trends in blood group typing. Anal.
Bioanal. Chem. 393, 1443–1451 (2009)
6. Quinn, J.G., O’Neill, S., Doyle, A., McAtamney, C., Diamond, D., MacCraith, B.D.,
O’Kennedy, R.: Development and application of surface plasmon resonance-based
biosensors for the detection of cell-ligand interactions. Anal. Biochem. 281, 135–143
(2000). https://doi.org/10.1006/abio.2000.4564
7. Quinn, J.G., O’Kennedy, R., Smyth, M., Moulds, J., Frame, T.: Detection of blood group
antigens utilising immobilised antibodies and surface plasmon resonance. J. Immunol.
Methods 206, 87–96 (1997). https://doi.org/10.1016/S0022-1759(97)00092-6
8. Berthier, A., Elie-Caille, C., Lesniewska, E., Delage-Mourroux, R., Boireau, W.: Label-free
sensing and atomic force spectroscopy for the characterization of protein-DNA and
protein-protein interactions: application to estrogen receptors. J. Mol. Recognit. 24, 429–435
(2011). https://doi.org/10.1002/jmr.1106
9. Boozer, C., Kim, G., Cong, S., Guan, H., Londergan, T.: Looking towards label-free
biomolecular interaction analysis in a high-throughput format: a review of new surface
plasmon resonance technologies. Curr. Opin. Biotechnol. 17, 400–405 (2006). https://doi.
org/10.1016/j.copbio.2006.06.012
10. Campbell, C., Kim, G.: SPR microscopy and its applications to high-throughput analyses of
biomolecular binding events and their kinetics. Biomaterials 28, 2380–2392 (2007). https://
doi.org/10.1016/j.biomaterials.2007.01.047
A Biochip Based Medical Device for Point-of-Care ABO Compatibility 105
11. Mansuy-Schlick, V., Delage-Mourroux, R., Jouvenot, M., Boireau, W.: Strategy of
macromolecular grafting onto a gold substrate dedicated to protein–protein interaction
measurements. Biosens. Bioelectron. 21, 1830–1837 (2006). https://doi.org/10.1016/j.bios.
2005.11.021
12. Krupin, O., Wang, C., Berini, P.: Selective capture of human red blood cells based on blood
group using long-range surface plasmon waveguides. Biosens. Bioelectron. 53, 117–122
(2014). https://doi.org/10.1016/j.bios.2013.09.051
13. Ferraz, A., Carvalho, V., Soares, F.: Development of a human blood type detection
automatic system. In: Eurosensors Xxiv Conference, vol. 5, pp. 496–499 (2010)
14. Ferraz, A., Carvalho, V.: A prototype for blood typing based on image processing. In:
SENSORDEVICES 2013: The Fourth International Conference on Sensor Device
Technologies and Applications, pp. 139–144 (2013)
15. Ramasubramanian, M.K., Alexander, S.P.: An integrated fiberoptic–microfluidic device for
agglutination detection and blood typing. Biomed. Microdevices 11, 217–229 (2009).
https://doi.org/10.1007/s10544-008-9227-y
16. Ramasubramanian, M., Anthony, S., Lambert, J.: Simplified spectraphotometric method for
the detection of red blood cell agglutination. Appl. Opt. 47, 4094–4105 (2008)
17. Wacogne, B., Boireau, W., Morel, P., Pazart, L., Pieralli, C.: Device for taking a sample of a
body fluid and method for implementing same (2011)
18. Wacogne, B., Boireau, W., Morel, P., Pazart, L., Pieralli, C.: Secure perfusion system (2011)
19. Charrière, K., Rouleau, A., Gaiffe, O., Fertey, J., Morel, P., Bourcier, V., Pieralli, C.,
Boireau, W., Pazart, L., Wacogne, B.: Biochip technology applied to an automated ABO
compatibility test at the patient bedside. Sens. Actuators B Chem. 208, 67–74 (2015). https://
doi.org/10.1016/j.snb.2014.10.123
20. Charrière, K., Guerrini, J.-S., Wacogne, B., Rouleau, A., Elie-Caille, C., Pieralli, C., Pazart,
L., Morel, P., Boireau, W.: SmarTTransfuser - a biochip system for the final ABO
compatibility test. In: Présenté à International Conference on Biomedical Electronics and
Devices (BIODEVICES 2012), Vilamoura, Portugal (2012)
21. Charrière, K., Rouleau, A., Gaiffe, O., Morel, P., Bourcier, V., Pieralli, C., Boireau, W.,
Pazart, L., Wacogne, B.: An automated medical device for ultimate ABO compatibility test
at the patient’s bedside - towards the automation of point-of-care transfusion safety. In:
Présenté à International Conference on Biomedical Electronics and Devices (BIODEVICES
2015), Lisbon, Portugal (2015)
Novel Pattern Recognition Method
for Analysis the Radiation Exposure
in Cancer Treatment
1 Introduction
For the purposes of the dosimetry of the small areas in radiotherapy, typically
uses traditional micro ionisation chambers, semiconductor diode dosimeters and,
in the recent years, increasingly uses a very handy MOSFET transistors. The
MOSFET transistors provide good accuracy and repeatability of the results with
dimensions of a few millimetres, which is sufficient for practical applications in
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 106–118, 2018.
https://doi.org/10.1007/978-3-319-94806-5_6
Novel Pattern Recognition Method for Analysis the Radiation 107
medical diagnostics. In addition, they are joined harmoniously with the systems
of scanning and information processing [1,2]. For 3D doisimetry, Gel Dosimeters
proposed in the mid 80-s are widely used. However, this method is rather incon-
venient as it is associated with the use of special phantoms (models) made of
gel-like material changing its optical properties under the influence of ionising
irradiation. At the same time, it allows customising the parameters intensity and
the 3D geometry of irradiation. Polymer gel dosimetry remains one of the most
promising and widely used tools for 3D dose measuring [3]. There are also meth-
ods of extrapolating measurements at individual points or 2D measurements to
3D models.
The most modern methods [4] of measurement suggest usage of linear array
diodes with 98 measurement points for scanning of a water phantom. Then the
data are linked to the patient CT image and after Monte Carlo method are used
to extrapolate dose distributions inside the patient body and to control measure-
ments with point dosimeters (diodes) performed during the process treatment
in vivo. Impact of ionising irradiation at different MOS (metal oxide semicon-
ductor) structures have been studied for quite a long time, at least since the
mid 70-s due to the start of using of electronics based on MOS technology in
space systems [5]. The processes occurring in such structures under the influence
of radiation of various types and intensity is very well studied and described in
numerous articles addressing radiation hardness of MOS IS [6–9]. For medical
applications measuring of the accumulated radiation dose with such structures
is still rather new. As an indication of the accumulated dose the effect of degra-
dation of MOS structure is used, particularly the under-gate dielectric (SiO2).
Without going deep into the details of physical processes, we mention only the
main effect that is used for dosimetry.
Under irradiation, gate dielectric accumulates a positive charge which leads
in particular to a shift of the threshold voltage in a MOSFET transistor or to a
shift of the volt-farad characteristics of the MOS capacitor. Inducing of a positive
voltage on the Gate of a transistor (MOS capacitor) in the process of irradiation,
leads to the increase of the amount of accumulated charge. In the case of no
voltage applied, it makes possible to irradiate the passive MOS structure, and
then to measure the charge that is equivalent to the dose. Other effects occurring
in dielectrics during irradiation can be ignored in this case. In the range of doses
used in medicine, the charge accumulation is linearly proportional to the dose
and only at high doses about 6–8 Gy (depends on the technology) tends to
saturation and loses linearity. Moreover, dosimeters based on MOS structures
are small in size (around 1 sq. mm ) and very simple in production.
The particular focus of this paper is in the use of such sensors for development
and imaging technology of a net bandage dosimetry system (Fig. 1), with a MOS
capacitor sensor in every node of the grid. Such dosimetry net can be placed
(dressed) around any part of the body (or fantom) and will allow controlling the
dose of radiation for the incoming and outgoing flow of irradiation and from any
direction. This will allow building a 3D model of the absorbed dose inside the
108 D. Dubovitskiy and V. Kouznetsov
patient’s body, which is a new and highly demanded product in the market of
medical diagnostics tools.
The skin pattern recognition system is to deliver the location for net bandage
position relatively to body map. The skin patent is mathematically interpreted
by Fractal geometry. The self-similarity features of fractals are very suitable for
this application. The skin region is translated into the vector of Fractal features.
The Fractal features is then identifying the skin map. The first net bandage
placement is to select and record the skin regions. The future treatments requires
to position the sensor net at the very same skin area.
The automatic decision making system is implemented by Fuzzy logic tech-
nology. At the first net sensor bandage application the image recognition system
build membership function of Fuzzy logic and memorised skin map. The future
net bandage application will be automatically calculated offset to the first posi-
tion and it is up to future developer to use the offset or to move the sensors
net to the exact first position. We will cover the novel mathematical equation in
Sect. 3 and begin from dosimetry setup in next section.
this structure (for example thickness and type of gate-dielectric, the size of the
structure) to improve its operation as a dosimeter, due to the fact it’s simply a
capacitor it is not necessary to take into consideration the parameters required
for function as a transistor.
The scheme of the measurement of accumulated charge which corresponds to
the absorbed dose requires less number of contacts (only two, and one of them
is common for all sensors) that facilitates the creation of matrix or grid with a
large number of sensors.
Fig. 2. The measuring system for collecting data from sensors. 1- matrix of sensors,
2-multiplexer, 3-analog to digital converter.
The measurement process for each sensor will take approximately no more
than 5–10 ms. This means that the matrix of sensors with dimension of 256
sensors will be scanned in 1.5 to 2.5 s. These data is digitised by ADC (3) and sent
to a computer wirelessly, where data can be processed further. The advantages
of MOS as an absorbed dose dosimeters is also that they are able keep the
charge somewhat stable and readings can be taken for a long period of time
after exposure.
When MOS capacitors are exposed to ionizing radiation, the positive charge
is captured in the under-gate oxide [18–20]. It leads to a shift of volt-farad
characteristics, see Fig. 3. The magnitude of the shift depends on the absorbed
dose of ionising radiation and is approximately 200 mV in the absence of gate
voltage during irradiation. By applying a positive voltage on the gate, we can
increase the sensitivity of the sensors and the shift of the voltage can reach
400–500 mV per Gy.
To confirm the behaviour of MOS structures under radiation by standard
medical equipment, the most common samples taken, the gate oxide SiO2 were
grown in a dry environment at 1000C, thickness of oxide was 0.6 µm on Si wafer
4.5 Om/cm conductivity of n-type with F(fluorine) doping. Size of crystal was
1 × 1 mm. Irradiation was carried out at Photon clinical linear accelerator 6 MeV
(Varian 2100 EX), doses were 0 to 10 Gy, at room temperature. As a control
dosimeter, ionisation chamber ROOS was used. Results of the experiment are
shown on Fig. 4.
For our purposes we are going to use a successfully tested system of pattern
recognition [27] and adhere our system of sensors to the skin of the patient.
The image recognition system was used for skin cancer diagnostic and has been
published in [22–26,28] this redundant system was working well and we expect
it to be extremely effective.
Novel Pattern Recognition Method for Analysis the Radiation 111
The current medical practice includes several radiation exposure during the
course of treatment with a number of days in between. The position of the
net bandage on the patient’s skin is very important to allow consistency for the
next treatment of the same tumour. In order to address this, data from the CT
scan could be used to adjust position for the next treatment. Computation of
position is implemented by using Fractal Geometry theory [36] to get the pre-
cise pattern of the skin. The precise pattern of the skin is corresponded to the
calibration points on the net bandage. The real time computation system [33]
will allow a doctor to dynamically move the bandage and see the offset from the
last treatment position. When offset approaches zero, the exact same position
of the radiation sensors will be reached.
The skin has texture and a particular skin region can be characterised by Frac-
tal features called Fractal parameters [34]. An image of the skin sample is taken
by a specially designed dermatological image acquisition camera [38] on Fig. 5.
The correspondent points are calculated from Fractal parameters [35]. If we
consider the profile of a typical skin image [37], then the curve does not coincide
with a sine-wave signal. To obtain adequate accuracy, it is necessary to magnify
the resolution of the image [21], which in turn introduces distortion [29,31]. For
increased accuracy on low-resolution data, we consider a convolution function of
a form more consistent with the profile of a video signal [30–32]. For a signal I
we consider the representation
N
F (k) = I (n)
n=1
2π(k − 1)(n − 1) π π
arccos cos − −
N 2 2
2π(k − 1)(n − 1)
−i arcsin cos
N
and for an image I with resolution m × n,
M
N
F (p, q) = I (m, n) (1)
m=1 n=1
2π(p − 1)(m − 1) π π
arccos cos − −
M 2 2
2π(k − 1)(n − 1) π π
× arccos cos − −
N 2 2
2π(k − 1)(p − 1)
− i arcsin arccos
M
2π(k − 1)(n − 1)
× arcsin cos (2)
N
In this work, application of the power spectrum method used to compute the
fractal dimensions of a skin surface is based on the above representations for
F (k) and F (p, q) respectively. We then consider the power spectrum of an ideal
fractal signal given by P = c|k|−β , where c is a constant and β is the spectral
exponent. In two dimensions, the power spectrum is given by P (kx , ky ) = c|k|−β ,
where |k| = kx2 + ky2 . In both cases, application of the least squares method
or Orthogonal Linear Regression yields a solution for β and c, the relationship
between β and the Fractal Dimension DF being given by
3DT + 2 − β
DF =
2
for Topological Dimension DT . This approach allows us to drop the limits on
the recognition of small objects since application of the FFT (for computing the
Novel Pattern Recognition Method for Analysis the Radiation 113
power spectrum) works well (in terms of computational accuracy) only for large
data sets, i.e. array sizes larger than 256 and 256 × 256. Tests on the accuracy
associated with computing the fractal dimension using Eqs. (1) and (2) show an
improvement of 5% over computations based on conventional Discrete Fourier
Transform.
The setup calculates Fractal features dynamically from the centre of an
image. The testing GUI software is presented on Fig. 6:
The original skin image from the camera is presented on Fig. 7.
The current position of the net bandage and camera is given from optical
calibration marks Fig. 8.
The corresponding points of the current Fractal marks and optical position
gives us the offset number which guide the doctor to the original position of the
sensor net bandage.
114 D. Dubovitskiy and V. Kouznetsov
In order to map the skin surface, the ‘system’ has to know its mathematical
representation. Here, this representation is based on the features considered in
the previous section which are used to create an image of the object in the
‘electronic mind’. This includes the textural features for the skin object coupled
with the Euclidean and morphological measures defined. In the case of a general
application, all objects are represented by a list of parameters for implementation
of supervised learning for the first net bandage application in which a fuzzy logic
system automatically adjusts the weight coefficients for the input feature set.
The methods developed represent a contribution to pattern recognition based
on fractal geometry, fuzzy logic and the implementation of a fully automatic
recognition scheme as illustrated in Fig. 9 for the Fractal Dimension D (just
one element of the feature vector used in practice). The recognition procedure
Fig. 9. Basic architecture of the diagnostic system based on the Fractal Dimension D
(a single feature) and decision making criteria β.
Novel Pattern Recognition Method for Analysis the Radiation 115
uses the decision making rules from fuzzy logic theory [39–42] based on all, or a
selection, of the features defined above which are combined to produce a feature
vector x.
where wj is the weight coefficient matrix. This value is used to select the class
associated with
p(j) = min [(pj · wj − p) ≥ 0]
providing a result for a decision associated with the j th class. The weight coef-
ficient matrix is adjusted during the learning stage of the algorithm.
The decision criterion method considered here represents a weighing-density
minimax expression. The estimation of the decision accuracy is achieved by using
the density function
3 3
di = |xσmax − xi | + [σmax (xσmax ) − pj (xi )]
x = {xi }, which forms the membership function mj (x). The matrix of weight
factors wj,i is formed at this stage accordingly for the ith parameter and j th class
using the following expression:
N
wi,j = 1 − pi,j (xki,j ) − pi,j (xi,j ) pi,j (xki,j ) .
k=1
The result of the weight matching procedure is that all parameters which
have been computed but have not made any contribution to the characteristic
set of an object are removed from the decision making algorithm by setting wj,i
to null.
4 Conclusions
The focus of this paper is the development and validation of a simple, user-
friendly system which allows control of the spatial distribution of the accumu-
lated dose of radiation inside a body. The use of modern image recognition
techniques allows precise positioning of the sensors net bandage. The calcula-
tion of exact accumulation dose and its confirmation by correct measurements is
the key to the effective treatment. Simple and reliable monitoring of the 3D dose
distribution will allow us to provide treatment in the safest way. The safe way
means that the healthy cells will not be subject to unnecessary exposure, which
helps to maintain the healthy cells surrounding the tumour. This work represents
a new approach to accurate radiation exposure treatments. Implementation in
hospitals requires more experiments, calibration and technological input. Feasi-
bility studies, clinical validation and economical evaluation should be the next
steps. We hope that this research will contribute to the safer radiation exposure
treatment in cancer cure and prolonging the lives of many people.
References
1. Soubra, M., Cygler, J., Mackay, G.F.: Evaluation of a dual metal oxide-silicon
semiconductor field effect transistor detector as a radiation dosimeter. Med. Phys.
21(4), 567–572 (1984)
2. Thomson I., Reece M.H.: Semiconductor MOSFET dosimetry. In: Proceedings of
Health Physics Society Annual Meeting (1988)
3. De Deene, Y., Jirasek, A.: Uncertainty in 3D gel dosimetry. J. Phys. Conf. Series
573, 012008 (2015). 8th International Conference on 3D Radiation Dosimetry
(IC3DDose)
4. Nithiyanantham, K., Mani, G.K., Subramani, V., Mueller, L., Palaniappan, K.K.,
Kataria, T.: Analysis of direct clinical consequences of MLC positional errors in
volumetric-modulated arc therapy using 3D dosimetry system. J. Appl. Clin. Med.
Phys. 16(5), 296–305 (2015)
Novel Pattern Recognition Method for Analysis the Radiation 117
5. Ma, T.P., Dressendorfer, P.V.: Ionizing Radiation Effects in MOS Devices and
Circuits. Wiley Interscience, New York (1989)
6. Kohler, R.A., Kushner, R.A.: Total dose radiation hardness of MOS devices in
hermetic ceramic packages. IEEE Trans. Nucl. Sci. 35(6), 1492–1496 (1988)
7. Kaschieva, S.: Improving the radiation hardness of MOS structures. Int. J. Elec-
tron. 76(5), 883–886 (1994)
8. Claeys, C.: Simoen, E: Radiation Effects in Advanced Semiconductor Materials
and Devices. Springer Science and Business Media, Berlin (2002). https://doi.org/
10.1007/978-3-662-04974-7
9. Meurant, G.: New Insulators Devices and Radiation Effects, 1st edn. North Hol-
land, New York (1999). Print book ISBN 9780444818010
10. Kumar, A.S., Sharma, S.D., Paul Ravindran, B.: Characteristics of mobile MOS-
FET dosimetry system for megavoltage photon beams. J. Med. Phys. 39(3), 142–
149 (2014)
11. Gopidaj, A., Billimagga, R.S., Ramasubramanian, V.: Performance characteristics
and commissioning of MOSFET as an in-vivo dosimeter for high energy photon
external beam radiation therapy. Rep. Pract. Oncol. Radiother. 13(3), 114–125
(2008)
12. Choe, B.-Y.: Dosimetric characteristics of standard and micro MOSFET dosime-
ters as in-vivo dosimeter for clinical electron beam. J. Korean Phys. Soc. 55, 2566–
2570 (2013)
13. Briere, T.M., et al.: In vivo dosimetry using disposable MOSFET dosimeters for
total body irradiation. Med. Phys. 32, 1996 (2005)
14. Scalchi, P., Francescon, P., Rajaguru, P.: Characterisation of a new MOSFET
detector configuration for in vivo skin dosimetry. Med. Phys. 32(6), 1571–1578
(2005)
15. Sze, S.M.: Physics of Semiconductor Devices, 2nd edn. Willey, New York (1981)
16. Nicollian, E.H., Brews, J.R.: MOS (Metal Oxide Semiconductor) Physics and Tech-
nology. Wiley, New York (1982)
17. Zemel, J.N.: Nondestructive Evaluation of Semiconductor Materials and Devices.
Nato Science Series B. Springer US, New York (1979). ISSN 0258-1221
18. Hughes, H.L., Benedetto, J.M.: Radiation effects and hardening of MOS technology
devices and circuits. IEEE Trans. Nucl. Sci. 50, 500–521 (2003)
19. Oldham, T.R., McLean, F.B.: Total ionizing dose effects in MOS oxides and
devices. IEEE Trans. Nucl. Sci. 50, 483–499 (2003)
20. Adams, J.R., Daves, W.R., Sanders, T.J.: A radiation hardened field oxide. IEEE
Trans. Nucl. Sci. NS–24(6), 2099–2101 (1977)
21. Davies, E.R.: Machine Vision: Theory, Algorithms, Practicalities. Academic press,
London (1997)
22. Dubovitskiy, D.A., Blackledge, J.M.: Surface inspection using a computer vision
system that includes fractal analysis. ISAST Trans. Electron. Signal Process. 2(3),
76–89 (2008)
23. Dubovitskiy, D.A., Blackledge, J.M.: Texture classification using fractal geometry
for the diagnosis of skin cancers. In: EG UK Theory and Practice of Computer
Graphics 2009, pp. 41–48 (2009)
24. Dubovitskiy, D., Devyatkov, V., Richer, G.: The application of mobile devices for
the recognition of malignant melanoma. In: BIODEVICES 2014: Proceedings of the
International Conference on Biomedical Electronics and Devices, Angers, France,
p. 140, 03–06 March 2014. ISBN 978-989-758-013-0
118 D. Dubovitskiy and V. Kouznetsov
25. Dubovitskiy, D.A., Blackledge, J.M.: Moletest: a web-based skin cancer screening
system. In: The Third International Conference on Resource Intensive Applications
and Services, Venice, Italy, vol. 978-1-61208-006-2, pp. 22–29, 22–27 May 2011
26. Dubovitskiy, D.A., Blackledge, J.M.: Object Detection and classification with
applications to skin cancer screening. In: International Society for Advanced Sci-
ence and Technology (ISAST) Intelligent Systems, vol. 1, no. 1, pp. 34–45 (2008).
ISSN 1797–1802
27. Dubovitskiy, D.A., Blackledge, J.M.: Targeting cell nuclei for the automation of
raman spectroscopy in cytology. In: Targeting Cell Nuclei for the Automation of
Raman Spectroscopy in Cytology. British Patent No. GB1217633.5 (2012)
28. Dubovitskiy, D.A., McBride, J.: New ‘spider’ convex hull algorithm for an unknown
polygon in object recognition. In: BIODEVICES 2013: Proceedings of the Inter-
national Conference on Biomedical Electronics and Devices, p. 311 (2013)
29. Freeman, H.: Machine Vision: Algorithms, Architectures, and Systems. Academic
press, London (1988)
30. Grimson, W.E.L.: Object Recognition by Computers: The Role of Geometric Con-
straints. MIT Press, Cambridge (1990)
31. Louis, J., Galbiati, J.: Machine Vision and Digital Image Processing Fundamentals.
State University of New York, New-York (1990)
32. Nalwa, V.S., Binford, T.O.: On detecting edge. IEEE Trans. Pattern Anal. Mach.
Intell. 1(PAMI–8), 699–714 (1986)
33. Ripley, B.D.: Pattern Recognition and Neural Networks. Academic Press, Oxford
(1996)
34. Clarke, K., Schweizer, D.: Measuring the fractal dimension of natural surfaces using
a robust fractal estimator. Cartograph. Geograph. Inf. Syst. 18, 27–47 (1991)
35. Falconer, K.: Fractal Geometry. Wiley, Hoboken (1990)
36. DeCola, L.: Fractal analysis of a classified landsat scene. Photogram. Eng. Remote
Sens. 55(5), 601–610 (1989)
37. Snyder, W.E., Qi, H.: Machine Vision. Cambridge University Press, Cambridge
(2004)
38. Yagi, Y., Gilberson, J.R.: Digital imaging in pathology: the case for standardisa-
tion. J. Telemed. Telecare 11, 109–116 (2005)
39. Zadeh, L.A.: Fuzzy Sets and Their Applications to Cognitive and Decision Pro-
cesses. Academic Press, New York (1975)
40. Mamdani, E.H.: Advances in linguistic synthesis of fuzzy controllers. J. Man. Mach.
8, 669–678 (1976)
41. Sanchez, E.: Resolution of composite fuzzy relation equations. Inf. Control 30,
38–48 (1976)
42. Vadiee, N.: Fuzzy Rule Based Expert System-I. Prentice Hall, Englewood (1993)
Bioimaging
Automatic Segmentation of Neurons
from Fluorescent Microscopy Imaging
1 Introduction
Thanks to the great advances in microscopy technologies and cellular imaging,
we have many tools and techniques allowing to address fundamental questions in
neuron studies. We can capture high-resolution images of single cells or neuron
population that enable neurobiologists to investigate the neuronal structure and
morphological development associated to neurological disorders.
Recent studies [1] claim that the morphology (i.e., size, shape and dendritic
arborization) of a neuron is a key discriminant of its functional role. With differ-
ent morphological features and different functional tasks, distinct neurons have
diverse soma shapes and dendrite arborizations. Developmental abnormalities
might lead to neuron malfunction and can be early signals for variety of neu-
ropathies and neurological disorders.
To support neuroscientists in this study, fully automated tools for neuron
detection and segmentation are required. Different approaches have been pro-
posed [2,3], however still to date, the state of the art is not still satisfactory
given the complexity of the problem. For example, the manual interaction that
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 121–133, 2018.
https://doi.org/10.1007/978-3-319-94806-5_7
122 S. Baglietto et al.
some tools require [4] is time consuming, expensive and extremely subjective as
it depends on the user expertise and diligence.
Moreover the task is quite complex for different reasons. First of all, neu-
ronal samples are highly heterogeneous across different acquisitions. Images can
be characterized by high cell density and shape variety and usually there is a low
contrast at the neuron boundaries. Indeed, the fluorescence expressed is com-
monly non-uniform: it might present high intensity variability between soma, its
border and the processes, leading to bad morphological segmentation of cell and
significant fragmentation in dendrite appearance.
Traditional segmentation approaches that use basic method such as thresh-
olding and morphological operators are not precise enough for the task at hand
and lead to wrong segmentations. Learning approaches, nowadays broadly used
in object segmentation also thanks to Deep Learning, are not suitable for these
images because they require a huge amount of hand-labeled neuron samples for
training [5].
Among deformable models, active contour models have demonstrated good
performance in segmentation also in correspondence of challenging data [6,7].
Their main issue is the high sensitivity to the initialization, which often requires
user setting. To this aim, recent active contour models trying an hybrid approach
to automate the initial mask have been introduced [8,9].
Skeletonization is a global technique that extract the binary skeleton from
a given neuronal structure [10,11]. The key idea of these methods is an itera-
tive erasure of voxels from the volume of the segmented object preserving the
topology of the contained structure. Minimal path based tracing are other global
approaches which aim at linking seed points through an optimization problem
[12] or through Fast Marching algorithm [13]. Minimum Spanning Tree (MST)
tracing deals with the link between detected points into a tree representation [14].
In our work, we present a combined and fully automatic approach in order
to detect and segment the whole neuron. The first part [15] starts with soma
detection using a multiscale blob enhancement filtering. Then, neuron bodies are
segmented by an high performance active contour model. The resulting segmen-
tation is used to initialize the second part of the method, concerning dendrite
tree segmentation by hessian-phase based level set model.
The remainder of the paper is organized as follows. In Sect. 2 details on the
adopted dataset are provided. We present the combined method in Sect. 3: for
the detection and segmentation of cell bodies see Sect. 3.1 and for the dendritic
tracing see Sect. 3.2. In Sect. 4 results are presented and conclusions are provided
in Sect. 5.
2 Materials
In this work, we use two different datasets: Mouse Retina [15] and Larva
Drosophila [16]. Mouse Retina dataset shows populations of Retinal Ganglion
Cells (RGCs) which play a central role into the complex and stratified struc-
ture of the retina. Retina samples were imaged using Leica SP5 upright confocal
Automatic Segmentation of Neurons from Fluorescent Microscopy Imaging 123
Fig. 1. Some images containing Retinal Ganglion Cells (RGCs) used for testing the
proposed method [15]. The images show high variability across samples. In the bottom
line, there is some magnified crops of the upper images, showing the complexity of
images, where the analyzed structures are mixed with background and other structures.
Fig. 2. Some samples from the second dataset imaged single neurons. Images shows,
also for this dataset, the heterogeneity across samples and across microscopic acquisi-
tions. In the bottom line there are some details of the correspondent top line images.
3 Method
The pipeline for cells detection and segmentation is composed of three main
steps. As shown in Fig. 3 a Blob-based Filtering (second column) is followed by
an Active Contour (third column) and a Watershed Transform (last column).
The Multiscale Blob Enhancement Filtering is used to identify regions where
neuronal cells are likely located. After the blob filtering approach, blob-shapes
objects are binarized and used as initial ROIs for a Localizing Region-Based
Active Contour [17] that traces cell borders. Due to the fuzzy cell boundaries
and occlusions, multiple cells can be segmented all together as a unique entity.
To cope this issue, we apply a watershed transform.
Automatic Segmentation of Neurons from Fluorescent Microscopy Imaging 125
Fig. 3. Pipeline applied to two examples (from the top, Im1 (PV-EYFP) and Im2
(Thy1-EYFP)) with a crop in the central row, showing the difficulties caused by
contiguous cells [15]. In column, starting from the left side: Original Fluorescent
Microscopy Images; Results of the multiscale blob filter binarization; Results of the
active contour segmentation in blue transparency over the original image for getting
the suitable qualitative performance; Results of the watershed transform and of the
final threshold. (Color figure online)
where λx1 o ,s and λx2 o ,s are the eigenvalues of the Hessian matrix at point xo and
at scale s, β and γ are two thresholds which controls the sensitivity of the line
filter and S is defined as Λ = HF = j≤D λj , where D is the dimension of
2
the image. This measure is used for differing the structures from the background.
Starting from Frangi’s idea, we modify the filtering to filter-out line-like pat-
terns in favor of blob-like structures (as [19]).
126 S. Baglietto et al.
where smin and smax are the minimum and maximum scales where we expect
to find structures.
Watershed Transform and Size Filter. Active contour can fail to sepa-
rate groups of overlapping or contiguous cells. So, we exploit the simplicity and
computational speed of the watershed transform, introduced by Beucher and
Lantuéjoul [20], to split cells englomerates into groups of cells.
As a final step, we delete components which are too small or too large for
being cell bodies (a given example is in Fig. 3, first line) applying a size filter to
remove structures with size outside an acceptable range of soma dimensions. It
is possible to fix this range by a statistical analysis of the dimensions removing
the tails of the distribution.
Automatic Segmentation of Neurons from Fluorescent Microscopy Imaging 127
Crop of Im5
Fig. 4. Some cells are not easily visible to the human eye just visualizing the retina
images, but they are discovered and segmented by our algorithm (for example, in this
cropped figure, pink and blue cells were hardly detectable). Adding contrast to the
image makes these somata clearer but it increases noise and cell heterogeneity [15].
(Color figure online)
Fig. 5. Some example images from Larva Drosophila dataset. In column, from the left
side: 2D projection of the original volume; soma detection and segmentation applying
the first part of the proposed approach; whole neuron segmentation including dendrites.
segmented all the neurons (both soma and dendrites). For the gold dendritic
tracing, we adopted Simple Neurite Tracer [4].
Fig. 6. Variation of the % of detected cells in Mouse Retina dataset as a function of the
% threshold of overlap between detected cell and the corresponding annotated ground
truth [15].
130 S. Baglietto et al.
Table 1. Results for soma segmentation on Mouse Retina samples [15]. Dice Coeffi-
cient is computed for all steps in the pipeline (Blob Filter, Active Contour and Water-
shed Transform) and it shows improvements after each step. For the final stage of the
pipeline, there is also the percentage of detected cells computed assuming as detected
a cell with minimum overlap 50% with ground truth fixed at 50%.
Table 2. Soma segmentation results on Larva Drosophila dataset. Dice Coefficient has
been computed for each segmented soma.
Volume
DC #1 #2 #3 #4 #5 #6 #7 #8 #9 #10 #11 Average
Our method 0.82 0.71 0.78 0.71 0.80 0.91 0.86 0.86 0.88 0.83 0.85 0.82
Tuff 0.51 0.39 0.40 0.33 0.32 0.77 0.79 0.80 0.71 0.76 0.76 0.56
Automatic Segmentation of Neurons from Fluorescent Microscopy Imaging 131
Fig. 7. An example of the level set evolution starting from the soma segmentation as
seed point in the Sample #1. The level set is shown at different evolution steps.
132 S. Baglietto et al.
Larva Drosophila image results after the first part of the segmentation process
(middle column) and at the final segmentation (column on the right side).
To quantitatively evaluate our neuron segmentation, we compare our method
with a recent state-of-the-art automated approach proposed in [24], Tubularity
Flow Field (Tuff ). Tuff is a technique for automatic neuron segmentation that
performs directional regional growing guided by the tubularity direction of neu-
rites. We compute the DC on Larva Drosophila results for both methods and it
can be observed that our method significantly outperform Tuff (Table 3).
5 Conclusion
We have presented a new approach for whole neuron segmentation in challeng-
ing fluorescent microscopy images. Our method is comprehensive of two main
steps: soma detection and dendrite segmentation. In the first stage cell bod-
ies are detected by a new blob filtering approach and segmented by an active
contour model and a watershed transform. Then, a novel hessian-phase based
level set has been developed allowing to segment the whole neuron morphology.
Tests have been performed on large scale and single scale images. We obtained
high results for both detection and segmentation of the soma and for the whole
neuron reconstruction. Thanks to its generality and automation, this framework
could be applied to similar images and it is easily reproducible for the full net-
work reconstruction at the population level. Moreover, we could easily extend
the method to 3D dimensions since our theoretical model adopted a general
dimension formulation. Finally, it also opens new perspectives for the analysis
and the characterization of neuronal cells.
Acknowledgements. The research received financial support from the 7th Frame-
work Programme for Research of the European Commision, Grant agreement no.
600847: RENVISION project of the Future and Emerging Technologies (FET) pro-
gramme.
References
1. Baden, T., Berens, P., Franke, K., Rosón, M.R., Bethge, M., Euler, T.: The func-
tional diversity of retinal ganglion cells in the mouse. Nature 529(7586), 345–350
(2016)
2. Meijering, E.: Neuron tracing in perspective. Cytom. Part A 77(7), 693–704 (2010)
3. Basu, S., Aksel, A., Condron, B., Acton, S.T.: Tree2Tree: neuron segmentation for
generation of neuronal morphology. In: 2010 IEEE International Symposium on
Biomedical Imaging: From Nano to Macro, pp. 548–551. IEEE (2010)
4. Longair, M.H., Baker, D.A., Armstrong, J.D.: Simple neurite tracer: open source
software for reconstruction, visualization and analysis of neuronal processes. Bioin-
formatics 27(17), 2453–2454 (2011)
5. Zheng, Z., Hong, P.: Incorporate deep-transfer-learning into automatic 3D neuron
tracing. In: The First International Conference on Neuroscience and Cognitive
Brain Information, BRAININFO 2016 (2016)
6. Chan, T.F., Vese, L., et al.: Active contours without edges. IEEE Trans. Image
Process. 10, 266–277 (2001)
Automatic Segmentation of Neurons from Fluorescent Microscopy Imaging 133
7. Yezzi, A., Tsai, A., Willsky, A.: A fully global approach to image segmentation
via coupled curve evolution equations. J. Vis. Commun. Image Represent. 13(1),
195–216 (2002)
8. Ge, Q., Li, C., Shao, W., Li, H.: A hybrid active contour model with structured
feature for image segmentation. Signal Process. 108, 147–158 (2015)
9. Wu, P., Yi, J., Zhao, G., Huang, Z., Qiu, B., Gao, D.: Active contour-based cell
segmentation during freezing and its application in cryopreservation. IEEE Trans.
Biomed. Eng. 62(1), 284–295 (2015)
10. Lee, T.C., Kashyap, R.L., Chu, C.N.: Building skeleton models via 3-D medial
surface axis thinning algorithms. CVGIP: Graph. Models Image Process. 56(6),
462–478 (1994)
11. Palágyi, K., Kuba, A.: A 3D 6-subiteration thinning algorithm for extracting
medial lines. Pattern Recognit. Lett. 19(7), 613–627 (1998)
12. Meijering, E.H., Jacob, M., Sarria, J.C.F., Unser, M.: A novel approach to neurite
tracing in fluorescence microscopy images. In: SIP, pp. 491–495 (2003)
13. Benmansour, F., Cohen, L.D.: Tubular structure segmentation based on minimal
path method and anisotropic enhancement. Int. J. Comput. Vis. 92(2), 192–210
(2011)
14. Türetken, E., González, G., Blum, C., Fua, P.: Automated reconstruction of den-
dritic and axonal trees by global optimization with geometric priors. Neuroinfor-
matics 9(2–3), 279–302 (2011)
15. Baglietto, S., Kepiro, I.E., Hilgen, G., Sernagor, E., Murino, V., Sona, D.: Segmen-
tation of retinal ganglion cells from fluorescent microscopy imaging. In: BIOSTEC,
pp. 17–23 (2017)
16. Gulyanon, S., Sharifai, N., Kim, M.D., Chiba, A., Tsechpenakis, G.: CRF formu-
lation of active contour population for efficient three-dimensional neurite tracing.
In: 2016 IEEE 13th International Symposium on Biomedical Imaging, ISBI, pp.
593–597. IEEE (2016)
17. Lankton, S., Tannenbaum, A.: Localizing region-based active contours. IEEE
Trans. Image Process. 17(11), 2029–2039 (2008)
18. Frangi, A.F., Niessen, W.J., Vincken, K.L., Viergever, M.A.: Multiscale vessel
enhancement filtering. In: Wells, W.M., Colchester, A., Delp, S. (eds.) MICCAI
1998. LNCS, vol. 1496, pp. 130–137. Springer, Heidelberg (1998). https://doi.org/
10.1007/BFb0056195
19. Liu, J., White, J.M., Summers, R.M.: Automated detection of blob structures by
Hessian analysis and object scale. In: 2010 17th IEEE International Conference on
Image Processing, ICIP, pp. 841–844. IEEE (2010)
20. Beucher, S., Lantuéjoul, C.: Use of watersheds in contour detection. In: Inter-
national Workshop on Image Processing, Real-Time Edge and Motion Detection
(1979)
21. Lathen, G., Jonasson, J., Borga, M.: Phase based level set segmentation of blood
vessels. In: 19th International Conference on Pattern Recognition, ICPR 2008, pp.
1–4. IEEE (2008)
22. Läthén, G., Jonasson, J., Borga, M.: Blood vessel segmentation using multi-scale
quadrature filtering. Pattern Recognit. Lett. 31(8), 762–767 (2010)
23. Zijdenbos, A.P., Dawant, B.M., Margolin, R., Palmer, A.C., et al.: Morphometric
analysis of white matter lesions in MR images: method and validation. IEEE Trans.
Med. Imag. 13(4), 716–724 (1994)
24. Mukherjee, S., Condron, B., Acton, S.T.: Tubularity flow field—A technique
for automatic neuron segmentation. IEEE Trans. Image Process. 24(1), 374–389
(2015)
Evaluation of Dense Vessel Detection
in NCCT Scans
1 Introduction
Automatic detection of dense vessels in Non-Contrast Computed Tomography
(NCCT) scans of patient with suspected stroke may accelerate treatment triage.
In a thrombolysis-only service, detection of a dense vessel helps to inform the
decision to treat, since the length of the clot observed in thin slice NCCT is
related to the success rate of thrombolysis [1], a treatment aimed to recanalise
occluded arteries. If a thrombectomy service is available, then detection of proxi-
mal, large vessel occlusions helps identify a cohort of patients potentially eligible
for this treatment. Subsequent mandatory review of CTA would also benefit from
identification of which vasculature to review rst in the search for a denitive clot.
Detection of clots is complicated by the fact that increased intravascular
density might reect either stasis of blood ow at the clot blood interface, or a
high haematocrit (in which case it is usually symmetrical). In this paper we
aim to evaluate the performance of an automatic dense vessel detector versus
manual identfiication of dense vessel signs in NCCT. Nevertheless, whether a
dense vessel sign represents a true clot can only be determined by looking at the
CT angiography (CTA) scans, therefore we also compare automatic detection
and clinical identification in NCCT with CTA.
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 134–145, 2018.
https://doi.org/10.1007/978-3-319-94806-5_8
Evaluation of Dense Vessel Detection in NCCT Scans 135
2 Dataset
We use data from the South Glasgow Stroke Imaging Database provided by the
Institute of Neuroscience and Psychology, University of Glasgow, Southern Gen-
eral Hospital. It includes data from the following studies: ATTEST [2], POSH [3]
and WYETH [4]. The database provides information about presence or absence
of dense vessel signs in NCCT, and presence or absence of arterial occlusions in
the corresponding CTA scans.
For the purposes of this evaluation, we consider only the anterior circulation
to the brain which comprises proximal occlusions and more distal dot signs. In
order to train the detector, manual segmentations were required. These were
generated in 3D Slicer 4.5.0 by a clinical researcher under the supervision of an
experienced neuroradiologist. Annotations were collected on the acute NCCT
scan with sight of the radiology report which included laterality of symptoms,
but blind to additional information such as the CTA, CT perfusion and follow-up
scans.
3 Methods
3.2 Detector
Fig. 1. Steps in the dense vessel detection pipeline. From left to right: (1) Extraction
of the block of interest (anterior circulation region) from the novel volume which has
been aligned to the atlas. (2) Folding of the block along the anatomical midline. (3)
Corresponding bilateral intensity data is input to parallel CNN channels to allow direct
comparison of the left and right sides of the brain. The remaining three input channels
encode the x, y and z atlas coordinates. (4) CNN outputs probability volumes for the
target hemisphere, indicating the voxelwise probability of a dense vessel being present.
Table 1. Description of the connected components detected in the scan shown in Fig. 2.
The four detections are ranked from the most to the least confident.
For the evaluation, we used 58 scans from the POSH study (all patients for
which information was available about the presence or absence of dense vessels
in NCCT and arterial occlusion in CTA), 25 of which show dense vessel signs in
NCCT.
The output probability volumes are thresholded with t = 0.1. This threshold
is selected such that the geometric mean of precision and recall is maximised on
the voxel level predictions for the training data. After thresholding, connected
components are identified. If any of the connected components overlap with
the manual segmentation, then the the dense vessel sign is considered to be
detected i.e. a true positive. If the dense vessel sign is present and a detection
is output by the CNN, but the two do not overlap, then this is considered to
be a mislocated true positive. See Figs. 2 and 3 for examples of a true positive
and a mislocated true positive respectively. Table 1 gives details of the connected
components detected in one example of a true positive scan. To aid in assessing
clinical value, we further report the mean, minimum and maximum number of
connected components which require reviewing in a positive scan (either true
positive or false positive).
Fig. 2. True positive detection. NCCT is presented at the slice where the largest area
of dense vessel is visible. Expert segmentation and CNN outputs are presented as a
Maximum Intensity Projection (MIP). The brighter colours of the CNN output suggest
stronger confidence and darker weaker confidence. The thresholded output results in 4
connected components with peak confidences of 0.12, 0.19, 0.33 and 0.78 (see Table 1).
There is an overlap between at least one of the detected components and the expert
segmentation, therefore the detection is treated as a true positive. In this case the
number of detections that need reviewing is four; two detected components overlap
with manual segmentation and two are false positive detections. (Color figure online)
138 A. Lisowska et al.
Since not all arterial occlusions are visible in the NCCT scan (see Fig. 5),
we compare the CNN detections and manual segmentations with the thrombus
diagnoses according to the database in the CTA scans.
4 Results
Table 2 shows the agreement between the dense vessel identification by the clini-
cal researchers and by the automatic detection system. A good automatic detec-
tion system would achieve a level of agreement with a human expert equivalent
to the agreement between two human expert annotators. To obtain the inter-
rater agreement score we compare the clinical researcher’s manual segmentation
with the database dense vessel sign annotations. The CNN detection does not
reach the level of inter-rater agreement. Table 3 presents more detailed analysis
of the automatic detection in comparison to the manual segmentation in NCCT.
Table 3 shows that the CNN detector yielded 17 false positives. Comparing
the detection confidence range between the true positives and the false positives
suggests that the threshold selected based on the training data is not optimal
for the validation datasets (See Fig. 6B). The training data distribution has a
Fig. 3. Mislocated true positive. NCCT slice in which the largest area of dense vessel
is visible. The manual segmentation and CNN outputs are presented as a MIP. There
are only weakly confident detections. The thresholded output results in 8 connected
components with a maximum confidence level between 0.1 for the least confident seg-
ment and 0.3 for the most confident segment. There is no overlap between any of the
detected components and the manual segmentation, therefore the detection is treated
as a false negative detection of dense vessel. We also report that there were 8 detections
which needed reviewing.
Evaluation of Dense Vessel Detection in NCCT Scans 139
Fig. 4. False positive. NCCT slice in which the largest area of CNN detected dense
vessel is visible. The clinical researcher has not segmented any areas in this scan,
therefore this is a false positive detection. The number of components which require
review in this scan is 2.
Table 2. Agreement between observers for detection of dense vessel sign in NCCT.
The mislocated true positive are counted as missed detections.
Fig. 5. Large CTA occlusion but dense vessel is not visible in NCCT. Not all occlusions
present as dense vessels.
higher percentage of dense vessels (53%) than the validation data (43%), which
might affect the choice of threshold. Setting the detection threshold to 0.2 would
reduce the number of false positives without affecting the true positives and it
would increase the overall agreement with the expert segmentation to 0.7 and
with the database score to 0.72. It is worth noting that this would still be lower
than the intra-rater agreement.
140 A. Lisowska et al.
A B
Fig. 6. Box plots of scans with positive detections. (A) Number of connected compo-
nents idented in the scan for each type of positive (+ve) scan. (B) Confidence level of
the most confident connected component for each type of positive scan.
There are 6 datasets which are treated as false positive as dense vessels were
not identified in the NCCT scans, however arterial occlusions were identified in
the CTA scans. One of these detections turned out to be a true positive when
reviewed against CTA i.e. manual identification erroneously missed a dense vessel
sign.
There are 4 datasets in which the CNN failed to detect the dense vessel. All
of these are dense vessels at the distal segment of the vessel (see Fig. 7). These
are usually harder to detect, because they frequently present as dot signs, which
may be only a few voxels in size. The larger, easier to identify, distal occlusion
cases have been detected (See Fig. 8).
To assess the extent to which the detection of dense vessel signs would help in
the review of CTA we compute the agreement between detections in NCCT and
CTA (see Table 4). Most of the detected dense vessel cases have been identified
in CTA as the clots (arterial occlusion(AO)), however not all clots were visible in
NCCT, therefore all detection methods have higher specificity than sensitivity
when reviewed against CTA. There is a high proportion of cases that have a dense
vessel in our validation set, therefore the sensitivity of detection is higher than
reported in larger population studies [13], where the reported sensitivity is 52%
Evaluation of Dense Vessel Detection in NCCT Scans 141
Fig. 7. Missed distal detection. This is a mislocated true positive case i.e. the detections
were situated in the wrong part of the scan.
and specificity 95%. The automatic system did not reach detection specificity
of the clinical researchers, however the overall detection agreement with CTA
remained at the same level as the agreement with human annotations in NCCT.
5 Discussion
When we are considering the automatic detection for use in patient triage for
thrombectomy it is important that the CNN detects the dense vessels in patients
who have a large vessel occlusion in the internal carotid and in the M1 segment
of the middle cerebral artery [14] as these patients would be considered for
thrombectomy. In the validation dataset there were 19 patients who presented
occlusion in that region and the CNN detected all of them correctly (e.g. see
Fig. 9). In one case, a large vessel sign had been erroneously missed during man-
ual segmentation, and the CNN detector picked it up. This illustrates the fact
that identification of dense vessel in NCCT is not trivial and some signs can be
missed even by a human expert, therefore an automatic detector which acts as
a second reader may aid in reviewing the scan.
Any patient considered for thrombectomy would also have CTA to confirm
the presence and location of a clot, but this kind of detection could direct the
142 A. Lisowska et al.
clinician to the most likely vessel in CTA to speed up the review and treatment
decision. In order for an automatic system to be useful, the number of the can-
didate detections cannot be overwhelming. We report that the median number
of detection which required reviewing per positive scan is 2 (See Fig. 6A) and
average is between 3 and 4 (See Table 3). The highest number of detections per
scan was 16 and in this case it probably would not be clinically useful to run the
automatic detector. We do not use morphological operation e.g opening and clos-
ing to clean the scans from small detections, which are frequently false positive,
because we do not want to remove possible small distal dot signs. Nevertheless we
provide the detection confidence rank for each scan (see Table 1) and the clin-
ician may choose to review only the top confident detections. Encouragingly,
Evaluation of Dense Vessel Detection in NCCT Scans 143
Fig. 9. Example of large vessel occlusion with dense vessel, confirmed by CTA and
detected.
in all of the true positive scans, the most confident detection was the true
detection.
An interesting possibility would be to train a complementary system on CTA
scans, with or without the initial NCCT scan as additional input. In this task,
positive detection presents as the absence of a feature (the vessel) in the scan, and
the bilateral comparison inherent in our network might prove to be particularly
well suited. Such a network could act as a further second reader to aid the
radiologist.
6 Conclusions
Whilst there is currently an equal number of false positives, this number might
be reduced with better threshold selection.
This system will be beneficial to those patients with arterial occlusion who
exhibit the dense vessel sign in NCCT. There is a subset of patients who do not
exhibit this sign, and therefore we further propose a future objective of training
a system on both NCCT and CTA scans for complete detection of occluded
arteries.
References
1. Riedel, C.H., Zimmermann, P., Jensen-Kondering, U., Stingele, R., Deuschl, G.,
Jansen, O.: The importance of size successful recanalization by intravenous throm-
bolysis in acute anterior stroke depends on thrombus length. Stroke 42(6), 1775–
1777 (2011)
2. Huang, X., Cheripelli, B.K., Lloyd, S.M., Kalladka, D., Moreton, F.C., Siddiqui, A.,
Ford, I., Muir, K.W.: Alteplase versus tenecteplase for thrombolysis after ischaemic
stroke (ATTEST): a phase 2, randomised, open-label, blinded endpoint study.
Lancet Neurol. 14(4), 368–376 (2015)
3. MacDougall, N.J., McVerry, F., Huang, X., Welch, A., Fulton, R., Muir, K.W.:
Post-stroke hyperglycaemia is associated with adverse evolution of acute ischaemic
injury. In: Cerebrovascular Diseases, vol. 37, p. 267, Karger Allschwilerstrasse,
Basel (2014)
4. Wardlaw, J.M., Muir, K.W., Macleod, M.J., Weir, C., McVerry, F., Carpenter,
T., Shuler, K., Thomas, R., Acheampong, P., Dani, K., Murray, A.: Clinical rel-
evance and practical implications of trials of perfusion and angiographic imaging
in patients with acute ischaemic stroke: a multicentre cohort imaging study. J.
Neurol. Neurosurg. Psychiatry 84(9), 1001–1007 (2013)
5. Dabbah, M.A., et al.: Detection and location of 127 anatomical landmarks in
diverse ct datasets. In: SPIE Medical Imaging, p. 903415. International Society
for Optics and Photonics (2014)
6. Lisowska, A., Bereridge, E., Muir, K., Poole, I.: Thrombus detection in CT brain
scans using a convolutional neural network. In: Proceedings of the 10th Inter-
national Joint Conference on Biomedical Engineering Systems and Technologies
(BIOSTEC 2017), Bioimaging, vol. 2, pp. 24–33. Scitepress (2017)
7. Lisowska, A., O‘Neil, A., Dilys, V., Daykin, M., Beveridge, E., Muir, K., Mclaugh-
lin, S., Poole, I.: Context-aware convolutional neural networks for stroke sign detec-
tion in non-contrast CT scans. In: Valdés Hernández, M., González-Castro, V.
(eds.) MIUA 2017. CCIS, vol. 723, pp. 494–505. Springer, Cham (2017). https://
doi.org/10.1007/978-3-319-60964-5 43
8. Wolterink, J.M., Leiner, T., Viergever, M.A., Išgum, I.: Automatic coronary cal-
cium scoring in cardiac CT angiography using convolutional neural networks. In:
Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS,
vol. 9349, pp. 589–596. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-
24553-9 72
9. Chollet, F.: Keras (2015). https://github.com/fchollet/keras
10. Theano Development Team: Theano: a Python framework for fast computation of
mathematical expressions. arXiv e-prints abs/1605.02688, May 2016
11. Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint
arXiv:1412.6980 (2014)
Evaluation of Dense Vessel Detection in NCCT Scans 145
12. Lin, T.Y., Goyal, P., Girshick, R., He, K., Dollár, P.: Focal loss for dense object
detection. arXiv preprint arXiv:1708.02002 (2017)
13. Mair, G., Boyd, E.V., Chappell, F.M., von Kummer, R., Lindley, R.I., Sandercock,
P., Wardlaw, J.M.: Sensitivity and specificity of the hyperdense artery sign for
arterial obstruction in acute ischemic stroke. Stroke 46(1), 102–107 (2015)
14. Campbell, B.C., Donnan, G.A., Mitchell, P.J., Davis, S.M.: Endovascular
thrombectomy for stroke: current best practice and future goals. Stroke Vascu.
Neurol. 1(1), 16–22 (2016)
Tracking Anterior Mitral Leaflet
in Echocardiographic Videos Using
Morphological Operators and Active
Contours
1 Introduction
1.1 Motivation
Mitral valve diseases are widespread and are commonly affected by Rheumatic
Heart Disease (RHD) [1]. RHD is an autoimmune disease that usually begins in
childhood. It starts as a strep throat infection that is caused by the streptococci.
If the strep throat infection is untreated, it results into rheumatic fever. The
repeated episodes of rheumatic fever slowly damages the heart valves.
Following one of the most relevant published studies [2,3], about 15.6 million
people are affected globally from RHD, and require medical follow-up, being
responsible for 233,000 deaths per year. Heart valve diseases create a massive
economic burden on health authorities. The average surgery cost to treat mitral
regurgitation was 24.871 ± 13.940 dollars per patient in Europe [4–6]. The heart
valve treatments and operations are not only expensive, but also a highly risky
cardiac process [7].
The literature suggests the cost effective solution of using penicillin in the
early stage [8]. It reduces the probability of recurrence of the rheumatic fever,
resulting less risk of damage to the heart valves. Therefore, earlier detection is
considered vital to control disease progression and to estimate disease burden in
low-resource regions of the world [1].
RHD thickened the Anterior Mitral Leaflet (AML) that directly affects the
shape and mobility of the leaflet, resulting into pathologies like stenosis and
regurgitation. Quantifying the degree of change in morphological features helps
to identify early cases and to control disease progression.
The key benefits of using the Echocardiography modality are its non-ionizing,
non-invasive property and is able to analyze fast moving structures like AML
in real time. It is a low cost modality and is available as a portable device that
makes it the most appropriate choice to use it in low resource areas [9].
The Parasternal Long Axis view provides the most suitable window to mea-
sure and quantify the clinically relevant parameters of the mitral valve such
as, thickness, mobility and valvular anatomy (Fig. 1) [10]. To achieve this the
first step is to segment and track the structures throughout the cardiac cycle.
Manual segmentation is undesirable, given its impracticality, subjectivity and
expert knowledge required. Automatic and semi-automatics methods to identify
and track mitral valve structures can improve the diagnostic process, providing
quick and objective measurements of clinically relevant parameters, even without
any expert cardiology knowledge.
Active contour models were used to delineate the objects with deformable shapes
and were extensively used by the research community for segmentation and
tracking of the structures in medical images. The reason to adopt this kind
of approach is their robustness against image noise and shape fragmentation,
ability to track non-rigid motion and its capability to incorporate geometric
148 M. S. Sultan et al.
Fig. 1. Parasternal long axis view, (adapted from [24]) (A): showing Mitral valve (MV),
Anterior Mitral Leaflet (AML), Posterior Mitral Leaflet (PML) and other structures.
(B): Shows the MV in Diastole/Systole phase, the thickened and hockey shape leaflet.
constraints, such as the expected shape [11]. Optical flow was integrated in the
active contour framework to segment and track the AML in echocardiography
[12]. The limitations of the proposed method are, incapability to track large
frame to frame displacement of AML, require manual initialization in the first
frame and was computationally expensive to process a single cardiac cycle (20 m).
In another work, transformation fitting was used to obtain initial boundaries
that are further refined by the two connected active contours [13]. The proposed
method requires initialization, parameter adjustment, failed in high displacement
(≥10 pixels) and is computationally expensive with 1.8 s/frame for the ten itera-
tions. A fully automatic and unsupervised method was based on outlier detection
in a low rank matrix to track the region of the AML, in both 2D and 3D ultra-
sound images [14]. Despite the fact that it is fully automatic, it is very sensitive
rank and noise. Literature review demands a real time segmentation and track-
ing algorithm with less user interaction and the ability to efficiently track the
mitral valve when faced with a large frame to frame displacements [11–14].
Mathematical morphology is widely used in image processing for analysis of
shapes, geometrical and topological structures. They were previously used to
segment the left ventricle [15], myocardium, ischemic viable and non-viable in
echocardiography [16].
Tracking Anterior Mitral Leaflet in Echocardiographic Videos 149
2 Methodology
In the first step, echocardiography video is read, followed by contrast stretching
to normalize the illumination of the image. In this work, we assumed the per-
fect segmentation (manual) in the first frame and is provided in priori. In each
step, two successive frames are iteratively selected. Since AML is a thin region
that shows fast motion, the thin regions of the successive images are extracted
followed by the regions with large displacement, using basic mathematical
morphological techniques. These regions are subsequently merged with the seg-
mentation result of the preceding frame and filtered, in the candidate region
module. Obtained regions are classified by taking into account their shapes and
geometrical properties. The obtained boundaries of the AML with morphological
operator are not well localized. Therefore localized active contour model is used
to refine the obtained boundaries. After having the segmentation results, we pro-
ceed to the post processing step of AML analysis. A summary of the proposed
processing pipeline is depicted in Fig. 2, and each step will now be discussed in
detail.
In this stage, two consecutive frames were extracted iteratively until the whole
cardiac cycle was covered. The basic mathematical morphological operations
were used that requires, the input image and the structuring element of suitable
size and shape. These morphological operations can be used for both binary and
grayscale images. For the resolution of the videos used in this papers experiments
the maximum recorded thickness of the AML was 24 pixels. Following this, we
used the grayscale images with the disk shape structuring element of width
24 pixels, to extract the potential regions.
Finally, The thin AML region (Fig. 3C) is extracted by taking the difference
between the grayscale input image (Fig. 3A) and the grayscale opened image
(Fig. 3B) with the flat disk shape structuring element of 24 pixel diameter.
Fig. 3. (A) Grayscale image (B) Morphological opening (C) Top-hat transform
(adapted from [24]).
Based on the analysis of the AML in the PLAX view, the thin AML region
shows a very large displacement in successive frames, compare to other regions
in an image. Other regions such as, septum, inferior wall (Fig. 1) do not show
significant displacement in successive frames and thus the regions are overlapped.
This prior information is meaningful to overcome the problem of tracking in
frames with large AML displacement.
The focus of this module is to extract region that showed large displacement
from frame t − 1 to frame t. That can simply be achieved by taking the differ-
ence of successive frames followed by selecting only the positive intensity values
(Fig. 4). Hard threshold is then applied to get the binary image.
Fig. 4. Regions with high displacement at four different times (frames) (adapted
from [24]).
Fig. 5. Candidate image for final AML classification (adapted from [24]).
Localized Active Contours. Ultrasound images are very noisy and frequently
contain heterogeneous regions, and as such neither edge based contours, nor
region based contours are a suitable choice. In this situation, we need a model
that take the benefits from both edge and region based active contours. A local-
ized region-based active contour (LAC) framework [20] were proposed to address
this problem. This hybrid region-based curve evolution is robust to noise and
doesn’t rely on the global configuration of the image.
The rough boundaries of the AML obtained from the morphological opera-
tors were used to initialize the LAC framework, to refine the leaflet boundaries.
The algorithm is based on the analysis of the local circular regions with five
pixels radius, at each point on the curve. At each point the algorithm locally
identifies the background and foreground optimally by their mean intensities.
The formulation of the local energy function along the curve is defined as:
∂φ
2
∂t (x) = δφ(x) Ωy B (x, y) δφ (y) .((I (y) − ux )
(2)
2 ∇φ(x)
−(I (y) − vx ) )dy + λδφ(x)div |∇φ(x)|
Here, δ is the Dirac function, B (x, y) represents a region that locally defines
the interior and the exterior of the region at point x and the radius of the
local region is specified by the user. The uniform modelling energy is used as
an internal energy [19]. The localized version of the internal energy is defined
as the local interior regions and exterior regions at each point on the curve.
(υx , νx ) are the localized version of means at each point x. The second term
is the normalization term that keeps the curve smoother. It penalizes the arc
length based on the weights λ tuned by the user.
154 M. S. Sultan et al.
Skeletonization. Prior to perform the analysis, the shape of the AML is sim-
plified by using skeletonization. The morphological thinning is used to get a line
of one pixel width, while preserving the topological characteristics of the AML.
Skeletonization works in the same way as morphological operators, convolving
the structuring element (template) on the binary image. The Mark-and-Delete
based templates were found very reliable and effective for thinning algorithms
and thus used in this work [21]. Ultrasound images are usually affected by the
speckle noise resulting into irregular boundaries, producing superfluous minor
branches of the skeleton. These branches are filtered out to extract the fun-
damental part of the skeleton. This can be done by computing the Euclidian
distance between the branch and the end points. All those branches whose length
are less than the defined threshold (6 pixels) are discarded.
Motion Patterns. The focus of this module is to compute the motion pattern of
the AML and analyze to extract the meaningful information. The mean motion of
the X and Y coordinates of the skeleton was computed. The motion of the AML
in X-axis was small and doesn’t provide any meaningful information. However,
the motion of the AML in Y-axis has shown large motion with a unique pattern.
The mean of the y-coordinates of the AML skeleton for each frame is plotted
against time, showing the motion pattern of the AML (Fig. 7). The cardiac cycle
is divided into systole and diastole phase based on the maximum and minimum
of the peaks of the obtained motion pattern. The classification helps to label the
AML as open or close and will be useful for the analysis such as, computing the
thickness when the valve is open. Further work can help to classify frames in
early filling and late filling phase (Fig. 7). The late filing will be useful to extract
frames in which the AML is perpendicular to the ultrasound beam. This is the
best position to measure the thickness of the AML tip, which provides a strong
clue regarding the presence or not of diseases.
Shape. The hockey stick like appearance of the AML in PLAX view is an
indicator of stenosis. A condition in which the heart valve leaflets get restricted
(narrowed, blocked) resulting into interruption in the normal blood flow. In order
to identify this condition, we proposed the measurement of the local curvature
on the skeleton of the AML. A template based method is used to measure the
local bending of the AML [22]. We tested two template based methods, the
trigonometrical and crossover point method (Fig. 8).
The trigonometrical approach relates the crossover angle with the curvature
(Eq. 3). The crossover angle is the angle between the crossover point, where
the curve intersect the disk mask and the X-axis of the disk. This approach is
sensitive to noise to estimate the precise angle of the crossover points.
2sinθc
Ktr = (3)
1 − sin2 θc
Tracking Anterior Mitral Leaflet in Echocardiographic Videos 155
The obtained area is large for the small curvature. Thus the reciprocal of
squared area is close to zero that increase the confidence by avoiding infinity
and the reliability of the approach.
The experiments has shown that the area based method is less sensitive
to speckle noise and provide smoother results, and thus used for this analysis.
For each frame, we measured the local curvature at each point on the AML
skeleton, followed by computing the overall mean to obtain the global curvature.
In stenosis, the leaflet is restricted and can be identified by the curvature (shape)
change.
We observed that the motion pattern of the AML and the pattern of the
global curvature change are correlated. When the valve opens the curvature of
the AML tends to decrease showing the straightness of the leaflet and when the
valve opens the curvature start increasing, suggesting the bending of the leaflet.
This motion shape relation might help in future by providing a clue to identify
pathological condition.
For a better visual representation of the motion and curvature pattern, we
first smoothed the curved and then normalize to restrict it in the range (0–1)
(Fig. 9).
3 Results
3.1 Materials
An initiative from the Real Hospital Português, in Recife, Brazil lead to the
screening of 1203 childrens and pregnant women, looking for cardiac patholo-
gies. All patients were tested regarding the presence of streptococcal infection
and short mitral valve videos were recorded. The data were collected using dif-
ferent ultrasound devices (M-Turbo, Edge II model by SonoSite, Vivid my model
Fig. 9. Motion and curvature pattern of AML, red: motion pattern, green: curvature
pattern. (Color figure online)
Tracking Anterior Mitral Leaflet in Echocardiographic Videos 157
The Modified Hausdorff Distance [23] was proposed to obtain a distance measure
to match two objects. In this work, we extended this approach by categorizing the
segmented region as false positive, false negative and true positive. We assumed
that the nearest point between Automatic Segmentation (AS) and Ground Truth
(GT) with Euclidean distance smaller than 2 pixels are true positives. The part of
the AS that is falsely segmented as AML were considered false positives and the
parts of the GT that were missed by the automatic segmentation were considered
as false negatives, always using 2 pixels distance as reference T (Fig. 10).
In this section, we analyze and compare the tracking ability of the proposed
algorithm on 67 cases in 2D ultrasound videos, obtained from the PLAX view. To
validate the algorithm, we compared the results of the proposed algorithm with
the doctors annotation. Results were also compared with the reference state of
the art algorithm [25]. The reference algorithm used the modified internal energy
Fig. 10. Region classification (adapted from [24]). (Color figure online)
158 M. S. Sultan et al.
3.4 Quantification
The performance metric used are: the E-MHD error, the number of failures and
the processing time. The morphological operators are relatively fast to obtain the
boundaries of the AML, spending on average 0.4 s/frame, however the reference
algorithm consumes 1.2sec/frame. To refine the obtained boundaries, we used
the LAC that consumes about 0.7 s/frame. The boxplot is used for the statistical
analysis of the mean E-MHD and the number of failure in each video. The E-
MHD error was computed for each frame of the video and the mean E-MHD
error was saved in a vector for visual inspection.
The boxplot of E-MHD error (Fig. 11) shows that the median E-MHD error
of our method is smaller than the reference algorithm, 3.7 and pixels 5.2 respec-
tively. The most frequent error for our method lies between 3.14 to 4.6 pixels.
However, the reference algorithm covers the comparatively large range from 4.6
to 6.6 pixels. The overall range of our method is also improved, from 2.12 to
5.54, whereas the reference covers the higher values from 3.32 to 8.06 pixels. In
the Fig. 10, the red dots shows the outliers.
Proposed method has shown an improvement in tracking, with a median
number of failure in each video of 2 (Fig. 12). The reference method failed twice
as much than proposed method (median of 4). The most frequent range of failure
Fig. 13. Segmentation results, red: doctors’s annotation, green: proposed method.
(Color figure online)
160 M. S. Sultan et al.
The segmentation results were plotted for better problem understanding. The
Fig. 13A, B shows the reasonable results with fully segmented AML and with
the small region falsely segmented as the AML. However, in Fig. 13C we have
the discrete regions in an image and thus missed by the proposed algorithm
and is one of the reason, why the proposed method fails. In Fig. 13D, the pro-
posed method has segmented not only the AML, but also segmented the chordae
tendineae and the posterior aorta. These missing and over-segmented regions are
the responsible of having large E-MHD, with low sensitivity and recall.
In this work, a new tracking approach is proposed that uses morphological oper-
ators to predict the location and boundaries of the AML. To obtain more precise
boundaries, the localized active contour framework is adopted. The algorithm
is found robust to track in the difficult situations when the valve opens with
the mean AML displacement of about 35 pixels. In such situation, the active
contours often fails to segment the boundaries (edges) that are far from the con-
tour. The proposed algorithm outperform the reference algorithm with respect
to time, however it is still slow to get the real time performance of the algorithm.
The main limitation of the algorithm is its incapability to avoid segmenting
the neighbor regions (chordae tendineae, cardiac walls, septum etc.). It happens
because the intensity and texture of the neighbor regions are similar. The chor-
dae tendineae and the posterior aorta are directly connected with the AML,
containing the same features. So we dont have any reliable feature that identify
starting and ending of the leaflet. In this work, we define the region of interest
that minimize this problem, however we need an automatic system that robustly
define the region of interest and impose the shape constraints in the active con-
tour frame work to further improve the segmentation performance.
Another limitation of the proposed algorithm is its incapability to recover
from the failure. This is the situation that occur frequently due to low quality of
the image and due to the missing structures in several frame of the ultrasound
video.
In future, we will focus to overcome the limitations such as, the time con-
sumption, reduce failure, minimize segmenting irrelevant regions, and finally to
estimate and quantify morphological features to identify pathological cases.
References
1. Bisno, A., Butchart, E.G., Ganguly, N.K., Ghebrehiwet, T., et al.: WHO Expert
Consultation on Rheumatic Fever and Rheumatic Heart Disease, Geneva (2004)
2. WHO: World Heart Federation, World Stroke Organization. Global atlas on car-
diovascular disease prevention and control (2011). ISBN 9789241564373
3. WHF (2012). http://www.world-heart-federation.org/fileadmin/user upload/docu
ments/Fact sheets/2012/RHD.pdf
4. Trochu, J.N., Tourneau, T.L., Obadia, J.F., Caranhac, G., Beresniak, A.: Economic
burden of functional and organic mitral valve regurgitation. Arch. Cardiovasc. Dis.
108, 88–96 (2015)
5. Ribeiro, G.S., Tartof, S.Y., Oliveira, D.W.S., Guedes, A.C.S., Reis, M.G., Riley,
L.W., Ko, A.I.: Surgery for valvular heart disease: a population-based study in a
Brazilian urban center. PLoS One 7(5), e37855 (2012)
6. The economic cost of cardiovascular disease from 2014–2020 in six European
economies. Centre for Economics and Business Research, London (2014)
7. Mirabel, M., Iung, B., Baron, G., Messika-Zeitoun, D., Dtaint, D., Vanoverschelde,
J.-L., Butchart, E.G., Ravaud, P., Vahanian, A.: What are the characteristics of
patients with severe, symptomatic, mitral regurgitation who are denied surgery?
Eur. Heart J. 28, 1358–1365 (2007)
8. Colquhoun, S.M., Carapetis, J.R., Kado, J.H., Steer, A.C.: Rheumatic heart disease
and its control in the Pacific. Expert Rev. Cardiovasc. Ther. 7(12), 1517–1524
(2009)
9. Remnyi, B., Wilson, N., Steer, A., Ferreira, B., Kado, J., Kumar, K., Lawrenson,
J., Maguire, G., Marijon, E., et al.: World Heart Federation criteria for echocar-
diographic diagnosis of rheumatic heart diseasean evidence-based guideline. Nat.
Rev. Cardiol. 9(5), 297–309 (2012)
10. Omran, A.S., Arifi, A.A., Mohamed, A.A.: Echocardiography of the mitral valve.
J. Saudi Heart Assoc. 22, 165–170 (2010)
11. Sheng, C.: Segmentation in echocardiographic sequences using shape-based snake
model. Comput. Inf. 27, 423–435 (2008)
12. Mikic, I., Krucinski, S., Thomas, J.D.: Segmentation and tracking in echocardio-
graphic sequences: active contours guided by optical flow estimates. IEEE Trans.
Med. Imaging 17(2), 274–284 (1998)
13. Martin, S., Daanen, V., Chavanon, O., Troccaz, J.: Fast segmentation of the mitral
valve leaflet in echocardiography. In: Beichel, R.R., Sonka, M. (eds.) CVAMIA
2006. LNCS, vol. 4241, pp. 225–235. Springer, Heidelberg (2006). https://doi.org/
10.1007/11889762 20
14. Zhou, X., Yang, C., Yu, W.: Automatic mitral leaflet tracking in echocardiogra-
phy by outlier detection in the low-rank representation. In: IEEE Conference on
Computer Vision and Pattern Recognition, pp. 972–979. IEEE Computer Society,
Washington, DC (2012)
15. Yun-gang, L., Jacky, K.K., Shi, L., Guan, Y., Linong, L., Qin, J., PhengAnn, H.,
Winnie, C.C., Defeng, W.: Myocardial iron loading assessment by automatic left
ventricle segmentation with morphological operations and geodesic active contour
on T2* images, Scientific reports (2015)
16. Lascu, M., Lascu, D.: A new morphological image segmentation with application
in 3D echographic images. WSEAS Trans. Electron. 5(3), 72–82 (2008)
17. Noble, J.A., Boukerroui, D.: Ultrasound image segmentation: a survey. IEEE Trans.
Med. Imaging 25(8), 987–1010 (2006)
162 M. S. Sultan et al.
18. Kass, M., Witkin, A., Terzopoulouse, D.: Snakes : active contour model. Int. J.
Comput. Vis. 1(4), 321–331 (1988)
19. Chan, T., Vese, L.: An active contour model without edges. In: Nielsen, M.,
Johansen, P., Olsen, O.F., Weickert, J. (eds.) Scale-Space 1999. LNCS, vol. 1682,
pp. 141–151. Springer, Heidelberg (1999). https://doi.org/10.1007/3-540-48236-
9 13
20. Lankton, S., Tannenbaum, A.: Localizing region-based active contours. IEEE
Trans. Image Process. 17(11), 2029–2039 (2008)
21. Zhang, T.Y., Suen, C.Y.: A fast parallel algorithm for thinning digital patterns.
Commun. ACM 27(3), 236–239 (1984)
22. Masoud, A.S., Pourreza, H.R.: A novel curvature-based algorithm for automatic
grading of retinal blood vessel tortuosity. IEEE J-BHI 20(2), 586–595 (2016)
23. Dubuisson, M.-P., Jain, A.K.: A Modified hausdorff distance for object matching.
In: Proceedings of International Conference on Pattern Recognition, Jerusalem,
Israel, pp. 566–568 (1994)
24. Sultan, M.S., Martins, N., Costa, E., Veiga, D., Ferreira, M., Mattos, S., Coimbra,
M.: Real-time anterior mitral leaflet tracking using morphological operators and
active contours. In: Proceedings of International Joint Conference on Biomedical
Engineering Systems and Technologies, Porto, Portugal (2017)
25. Sultan, M.S., Martins, N., Veiga, D., Ferreira, M.J., Coimbra, M.T.: Tracking of
the anterior mitral leaflet in echocardiographic sequences using active contours. In:
EMBC, pp. 1074–1077 (2016)
Convolutional Neural Network Based
Segmentation of Demyelinating Plaques
in MRI
1 Introduction
2 Medical Background
Multiple sclerosis (MS) is a common, chronic disease involving the central
nervous system and leading progressively to different degrees of neurological
CNN Based Segmentation of Demyelinating Plaques in MRI 165
disability. In multiple sclerosis cells of the human immune system attack myelin
sheaths of the nerve fibres which represent white matter in the brain and spinal
cord. The consequence of the myelin damage is inflammation in the affected areas
and then creating scar tissue. This process is known as demyelination and the
afflicted areas within the nerves’ sheathes are called demyelinating plaques. To
diagnose MS the combination of clinical symptoms, typical history, cerebrospinal
fluid examination and magnetic resonance imaging (MRI) of the central nervous
system is required. MRI plays an important role in diagnosing MS as it enables
not only to confirm the diagnosis and defining its pattern but also to assess the
progress of the disease and the response to treatment. It is essential to know
which areas of the brain are affected because the process of demyelination as
well as some other lesions in the white matter could also be present in different
neurological disorders. White matter lesions in MS occur in some characteris-
tic locations. Thus most of the lesions appear typically in juxtacortical regions
(that is close to the brain cortex), periventricularly (that is around the ventri-
cles and these lesions tend to lay perpendicularly to the long axis of the lateral
ventricles), in corpus callosum, cerebellum (within hemispheres and cerebellar
peduncles) and peripherally in brainstem (that is in cerebral peduncles, pons
and medulla oblongata).
T2-weighted images (T2WI) are MR scans which are the most sensitive in
showing the white matter lesions that are presented as areas of a high signal
(that means they are hyperintense and appear white on the images) in regard
to normal white matter. More sensitive than conventional T2WI in detecting
juxtacortical and periventricular lesions are FLAIR (fluid-attenuated inversion
recovery) sequences because they suppress the signal of fluid, including cere-
brospinal fluid which fills the ventricles and subarachnoidal space. As a result
the cerebrospinal fluid has a low signal and appears black on the scans obtained in
FLAIR technique, as compared to the white matter abnormalities which remain
hyperintense. On conventional T2WI cerebrospinal fluid presents high signal like
demyelinating lesions in the white matter thus it may be difficult to recognize
plaques localized in the vicinity of the ventricles and juxtacortical areas. On the
FLAIR images the contrast between cerebrospinal fluid and the white matter
lesions disposed in its proximity is more clearer and makes the plaques can be
better detected.
The present study is based on indicating demyelinating lesions in the white
matter on head MR images. All MR scans chosen to the study were performed
in FLAIR sequences, in axial plane, with 3 to 5 mm slices using 1,5 Tesla scan-
ner. The study comprised a hundred patients with confirmed diagnosis of MS,
including fifty men and fifty women in the age range between 19 and 66 years old
and in various stages of the disease. All noticeable changes of the signal intensity
within the white matter were considered as demyelinating lesions.
3 Method
Convolutional neural networks are biologically inspired [7] machine learning tech-
niques, where the input has a form of a finite-dimensional linear space range.
166 B. Stasiak et al.
Fig. 1. Structure of the convolutional layer. Sample input matrices A1 , A2 , A3 are pro-
cessed with 2 groups of Fi,j filters (3 filters in each group). The results produced by
each filter group are summed up constituting separate output matrices: M1 , M2 . Image
originally published in [18].
work, instead of reducing the size of feature maps, keeps their dimensions con-
stant. The details and consequences of that approach are described in Sect. 3.2.
Thanks to the normalization of the CNN output (e.g. using unipolar sigmoid
function), a fuzzy mask is created where the value assigned to each point can
be interpreted as a probability of its belonging to the object. Further processing
(noise removal, thresholding) leads to a binary mask which is useful in some
applications. Our approach to thresholding is described in Sect. 3.3.
To describe a convolutional layer, the basic unit of CNN, let us denote the input
data as a tuple of matrices A1 . . . Ap of a fixed na × ma size (for the first layer
it could be for example a multi-channel digital image). The key element of the
layer are q filter groups where each group is a tuple of p matrices of nf × mf size
(Fi,j for i = 1 . . . p, j = 1 . . . q). The output is a tuple of feature maps M1 . . . Mq
where for each i = 1 . . . q
⎛ ⎞
p
Mi = Zi + ⎝ ⎠ Aj ∗ Fi,j .
j=1
The Zi used in the formula above is a bias matrix of the same size as Mi . Matrix
convolution Aj ∗ Fi,j is a matrix of elements (Aj ∗ Fi,j )r,c for r = 1 . . . (na )−(nf )
+ 1, c = 1 . . . (ma ) − (mf ) + 1 such that
⎛ ⎞⎛ ⎞
nf −1 mf −1
(Aj ∗ Fi,j )r,c = ⎝ ⎠⎝ ⎠ (Fi,j )(nf −dn ),(mf −dm ) · (Aj )(r+dn ),(c+dm ) .
dn =0 dm =0
rows and (mf − 1) columns) is independent from the input size – it is related
only to the filter size.
Each element of convolutional layer output is a result of processing some
nf × mf rectangle taken from each Aj . For the first feature map, nf × mf is a
size of visual field [7]. For further layers, the size of visual fields can be easily
calculated by tracking down the range of CNN input pixels affecting each output
element. Should the network consist of convolutional layers and element-wise
operations only, the visual field size would be nz × mz where nz = (nf1 + . . . +
nft ) − t + 1 and mz = (mf1 + . . . + mft ) − t + 1. In these formulas t denotes
a number of convolutional layers and nfw × mfw is w-th layer filter size for
w = 1 . . . t.
Fig. 2. For each additional row/column of input matrices, you get one more
row/column of the output: (a) - the original configuration, (b) - the extended input.
In (b) for a rectangle of the same size as Aj matrices results are similar to (a) configu-
ration (it would work in this way for the white pixels of Bj and Mi ). Image originally
published in [18].
from the detected objects. In order to avoid such problems with input regions
appearing in the bigger image but not present in smaller training samples, the
training set should include negative samples as it is described in Sect. 4.1.
3.3 Evaluation
As mentioned above, the size of the feature maps is kept constant from layer
to layer in the proposed neural network. We also do not use MLP layers at the
output and the goal of the training is regression instead of, as it would be typical
for CNNs, image classification. The raw MR scans are put on the CNN input,
and we expect the output to take a form of the same-sized image, clearly marking
the MS lesions as white regions, surrounded by black, neutral background. In
practice, however, the output image will not be truly black-and-white, and the
intensity of a given output pixel may be interpreted in terms of the probability
that it is a part of a lesion. Therefore, we have to apply thresholding in order
to make the final decision and to obtain a black-and-white result that may be
directly compared to the expert-generated ground-truth mask.
The value of the threshold T ∈ [0, 1], used for this purpose, determines the
standard evaluation measures of a binary classifier: precision and recall. Low
threshold means that many brighter regions of the output image will be inter-
preted as sufficiently bright to represent demyelinating lesions, thus increasing
the recall. For the lowest possible threshold value, T = 0, the whole image will be
regarded as a brain tissue lesion, and hence every actual ground-truth lesion will
be marked as properly detected (100% recall). The precision of this detection,
170 B. Stasiak et al.
however, will be very low. On the other hand, using high value of T will result
in the opposite: only the most outstanding regions will be detected as lesions,
yielding high precision, but many actual lesions will not have sufficient intensity
and they will remain undetected, yielding low recall. Therefore, the frequently
applied “balanced” measure of classification efficacy is to compute the harmonic
mean of precision and recall, known as the F-measure.
In our approach the F-measure is used to find the optimal value of the thresh-
old T . After the training is finished, we threshold the output images (obtained
for the input images from the training set) with several values of T , record-
ing the resulting F-measure values. The value of T maximizing the F-measure
then becomes the final threshold, which we subsequently use to compute the
classification results on a separate set of images (the testing set).
However, it should be noted that the exact method of computation of pre-
cision and recall may be defined in various ways. Below we will present 3
approaches that were used in the present study to obtain different evaluations
of the classification results.
where Ithres and Itarg denote the image obtained at the network output (sub-
jected to thresholding) and the target ground-truth image provided by the
human expert, respectively. Having computed the number of pixels in each set,
the precision and the recall are defined as:
|TP| |TP|
precision = = ,
|TP| + |FP| |P|
|TP| |TP|
recall = = ,
|TP| + |FN| |T|
where |X| denotes the cardinality of the set X. The precision is hence defined
as the proportion of the number of TP pixels (correctly reported within the
lesion areas) to all of the actually detected pixels (positive pixels, P). Similarly,
the recall is the proportion of TP to all the pixels that should be reported (true
pixels, T) [18].
CNN Based Segmentation of Demyelinating Plaques in MRI 171
In other words, the output region is matched if it contains at least one pixel
coincident to a lesion region in the ground truth image and similarly for the
matched regions in the target image. It should be noted that we need the dis-
tinction between MP and MT, because several different CCs in the target image
may be matched by a single connected component in the thresholded output
image and vice versa. The precision and recall are then defined as:
|MP|
precision = ,
|P|
|MT|
recall = .
|T|
“artificial” increase of the number of the positive regions (P) followed by the
drop of the precision value.
In order to overcome this and to make our evaluation more intuitive, we
decided to introduce the third evaluation, based on post-processing of the thresh-
olded output images. We aim at defining regions of interest (ROI) within them, so
that a single ROI may cover several nearby connected components. This is done
in several steps. First, we draw a bounding rectangle around every connected
component found in the thresholded output image. Every bounding box is then
padded (enlarged) by 10 pixels from all the four sides and this enlarged rectangle
is filled with foreground (white) pixels. After that, it is possible, that individ-
ual nearby CCs got merged, so we repeat the search of connected components
obtaining the final set of detected regions. On this “second-level” representation
we compute the standard evaluation measures (precision, recall and F-measure)
in the same way as in the CCE approach.
Additionally – for visualization purposes – we draw the bounding rectangle
around every of those enlarged and merged “second-level” regions. In this way
we obtain a very practical and useful outcome, that may be directly used by a
specialist to immediately spot the regions of interest, potentially containing the
demyelinating lesions in the MRI scan (Fig. 3).
4 Experiments
The initial data set consisted of 100 scans from different patients. In order to
guarantee the consistent image format, with fixed image resolutions and number
of scan levels for each patient, data from 4 scans was discarded. The processed
data was split into the set used for training and validation purposes (77 patients)
and the test set (19 patients). This means that the evaluation on the test set
is based only on the patients that were not known during the stages of weights
adaptation and model selection. MR scans taken from the patients were con-
verted to sets digital images, 448 × 512 pixels each. The scans that contained
demyelinating plaques were used in the further processing. This yielded 982
training-and-validation images and 242 test images.
The 982 images selected for training and validation purposes were further
processed in order to provide the training set where a significant part of surface
consists of the plaques of demyelination. Instead of the whole images, selected
50 × 50 pixel tiles were used. The objective of this step was to reduce the com-
putational complexity of training when compared to the full-resolution images,
since the tile surface is 90 times smaller that the surface of the whole image.
What is more, selecting tiles where the demyelinating plaques were overrep-
resented was intended to prevent the stochastic gradient-based training from
reaching the local minima of parameters that yielded “all-zero” results, erro-
neously indicating that every analyzed scan is completely free of MS symptoms.
The initial approach was to use only tiles with centered occurrences of demyeli-
nating plaques. The initial attempts to create the working solution revealed that
this method of building the training set does not cover all the phenomena visible
in the MRI scans. In the result, bright objects that were underrepresented in the
training set, such as skull bones, adipose body of the orbit and paranasal sinuses
resulted in falsely-positive labeling of the MS lesions. In order to provide the
model that recognizes such cases, additional tiles from the other regions of the
scans were included in the training set as well (Figs. 4 and 5).
The selected data sets can be summarized in the following way:
– Training set – 7856 tiles of 50 × 50 size picked from the 982 training-and-
validation images. Tiles were selected in a pseudo-random way, but areas with
high average brightness or contrast were preferred. Approximately one tile
out of three contained only healthy tissues, without demyelinating plaques.
In case of MS lesions that were positioned close to the image boundaries,
the image was extended appropriately. This data set is used for the weights
adaptation in the presented neural networks.
– “Quasi-validation set” – 982 full-sized (448 × 512) images. This set serves
similar purpose to the typical validation set, but due to limited amount of
labeled data, it is not separate from the training set. It must be emphasized,
however, that this set contains remarkably more data than the training set.
The quasi-validation set is used for monitoring the learning progress and
selection of additional parameters of the final solution, such as threshold
level.
– Test set – 242 full-sized (448 × 512) images, separate from all the other data
sets not only in terms of images extracted from MRI scans, but in terms of
the set of patients that were examined. This set is used only for the final
benchmarks of the selected models.
174 B. Stasiak et al.
Fig. 4. Example of the MRI scan used in the test set (left) and the corresponding
reference demyelinating plaques mask (right).
The structure of the network, i.e. the number of layers, the number of neurons,
the size of the receptive fields and the non-linearity types as well as different
training procedures were the subject of intensive experiments in the presented
study. Three selected solutions are described below.
All the experiments were done with Caffe deep learning framework on a
cluster node with Tesla K80M GPU accelerator. The training set of 7856 50 × 50
tiles was fed to the network in mini-batches of 199 tiles each. The proposed
solution is a CNN composed of convolutional layers only (no MLP layers), which
makes it behave like an image filter, which accepts input images of any size,
without any changes to the architecture or the weights. This mechanism was
explained in detail in Sect. 3.2. This property makes it possible to calculate mean
square error (Euclidean loss) between the network outputs and the ground-truth
masks achieved for the full-sized scans (“quasi-validation set”). This error value
was used as the indicator of the training progress. The “quasi-validation set” set
contained 982 images of 448 × 512 size from which training tiles were cut.
Fig. 5. Example of tiles cut from the training-and validation set (top) with the cor-
responding lesion masks (bottom). Both tiles and the masks were cut from the full-
resolution images of the same format as it was illustrated in Fig. 4. Note, that tiles (i),
(iv), (viii) do not contain the lesions. Tiles (i) and (iv), however, present some of the
great number of possible big, bright structures that are likely to cause false positives
when detecting the lesions.
Fig. 6. The sequence of layers used in the convolutional neural network used in the
basic architecture.
activation function on the output of the final convolutional layer. The final layer
yields a single matrix, which can be later compared to the expected mask with
marked demyelinating plaques.
The slow, but stable convergence in terms of mean squared error on quasi-
validation set is presented in the top plot of Fig. 7. For as long as 24 millions of
image propagations in the training process, the error on the quasi-validation set
clearly decreases. The network learns to detect lesions on the training tiles, and
the result is general enough to apply to the full-sized images. The minimum of
mean squared error is 199.0.
In order to provide a practical verification of the network effectiveness, the
network output was thresholded to obtain the binary image, which can be com-
pared directly to the target mask. The value of the threshold selected in order to
maximize the F-measure, as described in Sect. 3.3. It should be noted, however,
that the characteristics of the training set, which was composed of small tiles,
were so different from the testing set containing the full scans. The appropriate
way to select the most useful threshold was to maximize the F-measure on the
quasi-validation set.
176 B. Stasiak et al.
Fig. 7. Training of the basic architecture without dropout. Top: learning curve
(Euclidean loss) on quasi-validation set; bottom: F-measure on the quasi-validation
set and the testing set. The unit on the horizontal axis corresponds to 106 tiles, which
were grouped in batches of 199 images.
The maximum F-measure value on the test set is 0.551, but we have no formal
way of selecting that exact model. The best F-measure on the quasi-validation
set is 0.622, but the corresponding model is visibly overtrained and yields onle
0.496 on the test set. Following the lowest mean squared error would result in
selecting a model that yields F-measure values of 0.617 on the quasi-validation
set and 0.545 on the test set.
The per-pixel F-measure value observed on the quasi-validation set keeps
growing as well, as we can see int he bottom plot of Fig. 7. The second curve
presented in that plot, however, describes the dynamics of F-measure on the test
set. This result starts decreasing much earlier than the MSE from the top plot –
the generalization error becomes visible after processing 14 millions images. The
network is apparently getting overtrained, as the result keeps losing its general
properties. It must be emphasized, however, that this effect happens only after
the whole training set was iterated over for almost 1800 times, which corresponds
to ca. 26 h of training.
Fig. 8. The sequence of layers used in the convolutional neural network used in the
basic architecture with dropout mechanism.
(half of the input values for Conv6 layer were replaced with zeros), the number of
filters in Conv5 layer was increased twofold. The change involved only one level
of the network, so the resulting training speed decrease amounted for only 15%
when compared to the basic architecture.
As we can see in the plots from Fig. 9, the effects of network overtraining
in the architecture with dropout are remarkably less intense than in the basic
architecture. The minimum of minimal square error on the quasi-validation set
occurred after ca. 24 millions of image propagations, which is similar to the pre-
vious experiment. The rate of error increase in the overtraining stage, however,
is much lower than it was without dropout. Similar remark can be observed in
the bottom plots of Figs. 7 and 9. The F-measure on the quasi-validation has
similar dynamics in both cases. The F-measure on the test set, however, reaches
the minimum notably later in case of the model with dropout (after 20 million
images instead od 15 million), and does not start to drop as rapidly as it did in
the previous experiment. The minimum of mean squared error is 195.7, which
is slightly lower than it was without dropout.
In order to compare the F-measure values to the previous model, we use
the model that minimizes the mean square error again. This model, when used
with the optimal threshold, generates F-measure values of 0.620 on the quasi-
validation set and 0.539 on the test set, which is comparable to the previous
experiment.
Fig. 9. Training of the basic architecture with dropout. Top: learning curve (Euclidean
loss) on quasi-validation set; bottom: F-measure on the quasi-validation set and the
testing set. The unit on the horizontal axis corresponds to 106 tiles, which were grouped
in batches of 199 images.
Fig. 10. The sequence of layers used in the convolutional neural network used in the
new, improved architecture.
per processed image, which makes the training of this model almost 35% slower
than the basic architecture.
The plots presented in Fig. 11, indicate that the general properties are similar
to the basic architecture with dropout – the overtraining effects are not nearly
as intense as in the basic architecture without dropout, but occur nonetheless.
The number of processed images related to the mean square loss and F-measure
optima is similar to the basic architecture with dropout as well. The achieved
minimum of the mean squared error function, however, is the best amongst the
three models, assuming value of 189.5.
In order to compare the F-measure values to the previous model, we use
the model that minimizes the mean square error again. This model, when
used with the optimal threshold, generates F-measure values of 0.620 on the
CNN Based Segmentation of Demyelinating Plaques in MRI 179
Fig. 11. Training of the improved architecture. Top: learning curve (Euclidean loss)
on quasi-validation set; bottom: F-measure on the quasi-validation set and the testing
set. The unit on the horizontal axis corresponds to 106 tiles, which were grouped in
batches of 199 images.
quasi-validation set and 0.542 on the test set. This is apparently slightly bet-
ter than the basic architecture with dropout, but not necessarily than the basic
architecture. The three proposed solutions, despite of the differences, yielded
vastly similar F-measure values.
The results reported hereby have been obtained on the test set, with the three
presented network models (basic, basic with dropout, improved), and the three
evaluation methods (PPE, CCE and RIE). The threshold value was computed
individually for each of the nine experiment settings on the quasi-validation set,
CNN Based Segmentation of Demyelinating Plaques in MRI 181
as described above. Table 1 presents the precision, recall and F-measure values
for all the three network architectures and the PPE evaluation scheme. Similarly,
Tables 2 and 3 show the results for the connected-component-based approaches.
It should be noted the reported numbers of pixels and regions are counted across
the whole test dataset (292 images).
The results are generally quite similar, in terms of the obtained F-measure
values. Interestingly, the evaluation scheme involving the connected components
introduces only very small improvement over the pixel-based approach, although
it is based on completely different assumptions. Only the region-based method is
quite significantly better, exceeding 60%, probably due to the additional enlarge-
ment and merging of the connected components.
Example images where the quality of lesion detection was remarkably good
(Table 5), remarkably poor (Table 7) and acceptably successful (Table 4) were
182 B. Stasiak et al.
presented in the tables that consist of: the input image, the demyelinating
plaques mask (ground-truth annotations), and the results of all the three neu-
ral networks presented in this paper. Additionally, the successful image, which
can be considered easy in terms of lesion labeling, is used to demonstrate that
the presented methods need to be intelligent even in order to solve the simplest
tasks. Simple thresholding can be considered as an alternate method of marking
“bright, important points” in the image. However, when compared to the neural
network, such a simplistic attitude is likely to act remarkably poor, as it is shown
in Table 6.
5 Analysis
The sample results shown in the previous section are intended to demonstrate
the general possibilities of CNNs application for the image detection task. The
detection of typical demyelinating lesions was successful often enough to consider
the achieved results promising. It must be emphasized that the ability to detect
the demyelinating plaques is not only based on their intensity, but on their
shape and the characteristics of the surrounding tissue as well. This property
can be illustrated by the comparison of the results from Tables 5 and 6, where the
same sample is processed with the proposed neural network and with a simple
threshold operator.
It can be observed that the threshold of 50% is too low to detect only the
lesions, as it marks some other points between the cerebral hemispheres as well.
Higher threshold values, however, result in heavily reducing (55%) or totally
omitting (60%) one of the lesions. At the same time, even the threshold level of
60% is not sufficient to ignore the skull. Convolutional neural networks, however,
have no problem with labeling all of the demyelinating lesions contained in this
sample and with ignoring the skull. This specific sample is, apparently, easy
enough to process by all the proposed models.
184 B. Stasiak et al.
when the task of diagnosis assistance is considered. The expected masks used
in both training and evaluation of the neural networks consisted of polygons,
which were marked approximately by the medical specialists. Repeating the
same polygon shapes is virtually impossible – either for the neural network or for
another expert. Some of the significant sources of errors were related to the large
bright areas resulting in the false positives. This includes temporal bones and
optic nerves. Another common source of errors was related to the small regions
of noise that were erroneously detected as demyelinating plaques. Additional
consideration was devoted to the points that were detected in a general area
of the MS lesions – the proposed modifications to the measure of the object
detection quality provides us with some deeper insight into the results analysis.
The obtained result is promising, but the further room for improvement
remains apparent. The crucial room for improvement is related to the data set
size – greater number of training samples, covering better variety of cases, would
be likely to improve the result. The selected size of the training samples, which
is 50 × 50, is another parameter that might require further discussion. Larger
tile size would make it easier to include the whole temporal bones and optic
nerves in the training samples. Greater tile size, however, makes the training
additionally difficult, because generating plain black outputs becomes a remark-
able local minimum of the neural network cost function. This problem can be
addressed either by cost function modification or manual region growing on the
expected outputs that would increase the number of white points. Alternatively,
the irrelevant parts of the input images – namely, everything but the cerebral
tissue – can be removed manually by the separate tool.
The general field of the CNNs application for the medical image process-
ing is usually affected by the difficulty with collecting the sufficient data sets.
Unsurprisingly, this problem is visibly present in our work as well. Our analysis,
however, can be considered as an initial step towards even more efficient solu-
tions. Object localization based solely on convolutional layers, dynamic threshold
selection, and detailed description of the results involving F-measure and con-
nected components are some ideas, that – when used together – form an elegant
solution that can be applied to the great diversity of object localization problems.
Another way to improve the proposed method is to involve some well-known
pretrained CNN models instead of starting from the random weights. Neural
networks such as AlexNet [8] or VGG [22] consist of carefully trained weights that
are known to be useful in detection and classification of multiple normal, real life
objects. The mentioned networks are usually used as classifiers, but application
to the scale-preserving object localization solution is possible as well. Using
parts of the network with maximum-pooling layers is not necessarily impossible
in this task – the problem of restoring original resolution can be addressed with
techniques such as deconvolutional neural networks [12].
Acknowledgements. This project has been partly funded with support from
National Science Centre, Republic of Poland, decision number DEC-2012/05/D/ST6/
03091.
Authors would like to express their gratitude to the Department of Radiology of
Barlicki University Hospital in Lodz for making head MRI sequences available.
CNN Based Segmentation of Demyelinating Plaques in MRI 187
References
1. Tomczyk, A., Spurek, P., Podgórski, M., Misztal, K., Tabor, J.: Detection of elon-
gated structures with hierarchical active partitions and CEC-based image repre-
sentation. In: Burduk, R., Jackowski, K., Kurzyński, M., Woźniak, M., Żolnierek,
A. (eds.) Proceedings of the 9th International Conference on Computer Recogni-
tion Systems CORES 2015. AISC, vol. 403, pp. 159–168. Springer, Cham (2016).
https://doi.org/10.1007/978-3-319-26227-7 15
2. Tomczyk, A., Szczepaniak, P.S.: Adaptive potential active contours. Pattern Anal.
Appl. 14, 425–440 (2011)
3. de Brebisson, A., Montana, G.: Deep Neural Networks for Anatomical Brain Seg-
mentation. ArXiv e-prints arXiv:1502.02445 (2015)
4. Shelhamer, E., Long, J., Darrell, T.: Fully Convolutional Networks for Semantic
Segmentation. ArXiv e-prints arXiv:1605.06211 (2016)
5. Milletari, F., Navab, N., Ahmadi, S.A.: V-Net: Fully Convolutional Neu-
ral Networks for Volumetric Medical Image Segmentation. ArXiv e-prints
arXiv:1606.04797 (2016)
6. Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional Networks for Biomed-
ical Image Segmentation. ArXiv e-prints arXiv:1505.04597 (2015)
7. Hubel, D.H., Wiesel, T.N.: Receptive fields and functional architecture in two
nonstriate visual areas (18 and 19) of the cat. J. Neurophysiol. 28, 229–289 (1965)
8. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep con-
volutional neural networks. In: Pereira, F., Burges, C.J.C., Bottou, L., Weinberger,
K.Q. (eds.) Advances in Neural Information Processing Systems 25, pp. 1097–1105.
Curran Associates, Inc. (2012)
9. LeCun, Y., Bengio, Y.: Convolutional networks for images, speech, and time-series.
In: Arbib, M.A. (ed.) The Handbook of Brain Theory and Neural Networks. MIT
Press, Cambridge (1995)
10. Cireşan, D.C., Meier, U., Masci, J., Gambardella, L.M., Schmidhuber, J.: Flexible,
high performance convolutional neural networks for image classification. In: Pro-
ceedings of the Twenty-Second International Joint Conference on Artificial Intel-
ligence, IJCAI 2011, vol. 2, pp. 1237–1242. AAAI Press (2011)
11. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale
hierarchical image database. In: CVPR 2009 (2009)
12. Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks.
CoRR abs/1311.2901 (2013)
13. Nguyen, T.V., Lu, C., Sepulveda, J., Yan, S.: Adaptive nonparametric image pars-
ing. CoRR abs/1505.01560 (2015)
14. Cheng, G., Zhou, P., Han, J.: Learning rotation-invariant convolutional neural
networks for object detection in VHR optical remote sensing images. IEEE Trans.
Geosci. Remote Sens. 54, 7405–7415 (2016)
15. Mopuri, K.R., Babu, R.V.: Object level deep feature pooling for compact image
representation. CoRR abs/1504.06591 (2015)
16. Matsugu, M., Mori, K., Mitari, Y., Kaneda, Y.: Subject independent facial expres-
sion recognition with robust face detection using a convolutional neural network.
Neural Netw. 16, 555–559 (2003)
17. Dai, J., He, K., Sun, J.: Convolutional feature masking for joint object and stuff
segmentation. CoRR abs/1412.1283 (2014)
188 B. Stasiak et al.
18. Stasiak, B., Tarasiuk, P., Michalska, I., Tomczyk, A., Szczepaniak, P.: Localiza-
tion of demyelinating plaques in MRI using convolutional neural networks. In:
Proceedings of the 10th International Joint Conference on Biomedical Engineer-
ing Systems and Technologies (BIOSTEC 2017), BIOIMAGING, vol. 2, pp. 55–64.
SCITEPRESS (2017)
19. He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-
level performance on ImageNet classification. CoRR abs/1502.01852 (2015)
20. LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to
document recognition. In: Proceedings of the IEEE, pp. 2278–2324 (1998)
21. Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.:
Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn.
Res. 15, 1929–1958 (2014)
22. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale
image recognition. CoRR abs/1409.1556 (2014)
Bioinformatics Models, Methods and
Algorithms
Compositional Analysis of Homeostasis of
Gene Networks by Clustering Algorithms
1 Introduction
Homeostasis of biological systems is an important property of life. This property
emerges from the elaborate interaction among several components in cells. To
understand the mechanisms of homeostasis, therefore, we should investigate not
only individual components but also how they work in collaboration with other
components. The latter research discipline is called systems biology.
There are two paradigms in systems biology - qualitative and quantita-
tive - to model target biological systems. Quantitative approaches enable us
analysing real-valued properties such as response times or predicting a time-series
of concentration values of a certain molecular species. However, the quantitative
parameters to construct such precise model are not commonly available. In qual-
itative approaches, we cannot analyse such real-valued properties. Instead, it can
be applied to biological systems with incomplete information under parameter
uncertainty, which are common in most of organisms.
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 191–211, 2018.
https://doi.org/10.1007/978-3-319-94806-5_11
192 S. Ito et al.
Fig. 1. Simple circadian clock. Fig. 2. Time series of expression levels of gene
The transcription product of a, b and c in the network Fig. 1.
gene a inhibits gene b, gene
b similarly inhibits gene c and
finally gene c inhibits gene a.
This network produces periodic
expression pattern of each gene.
We can use these propositions to capture a state of the gene network, e.g.
between time t1 and t2 the propositions {on a , ab , bc } are true, which means that
gene a is ON, gene a is expressed beyond ab and gene b is expressed beyond
bc (but gene b is OFF). Using these propositions, we now have the symbolic
representation of the behaviour as depicted in Fig. 3 which consists of states
(represented as circle) and transitions (represented as arrows). The propositions
listed below the states represent the facts that are true in the states. We call
this symbolic representation of a behaviour as linear time structure. We easily
see that state 0 represents the state of the network at the beginning, state 1
represents the state between t0 and t1 , state 2 between t1 and t2 , and so on.
In general, there are many behaviours which can be produced by a single
network. Our purpose is to analyse a biological property for a given gene net-
work, e.g. whether the network can produce the behaviour in which gene a, b
and c oscillate. To solve this problem, we must characterise the set of possible
behaviours and check that a certain behaviour (e.g. genes oscillate) is contained
in the possible behaviours. To characterise (or model) the behaviours of a net-
work, we need some mathematical representation. In quantitative approach, ordi-
nary differential equations (ODEs) are widely used. In the current setting, we do
not handle a numerical time series but a symbolic time series of a behaviour (a
linear time structure). To characterise and reason about such structures, linear
temporal logic (LTL) is the suitable mathematical language. LTL can be seen
as propositional logic equipped with temporal operators such as G (Globally),
F (Future) and U (Until). Gφ means φ is always true, F φ means φ is eventually
true, and φU ψ means φ is true until ψ is true. The formal syntax and semantics
are shown in Appendix A.
We are to characterise the set of possible behaviours (linear time structures)
for a given network. This can be done by specifying an LTL formula φG for a given
network G such that the set of possible behaviours of a network is characterised
as {σ | σ |= φG }, i.e. all linear structures which satisfy the behaviour specification
φG . The problem of analysing network behaviours, e.g. checking whether there
is a behaviour which satisfies a certain property ψ (also written in LTL), can be
solved by finding σ such that σ |= φG ∧ψ, i.e. checking satisfiability of the formula
φG ∧ ψ. Thus analysing a gene network is reduced to satisfiability checking of
LTL. Once we have a formula φG ∧ ψ, the analysis can be automatically done
by LTL satisfiability checkers.
Now the question: how do we specify φG which characterises possible
behaviours of a given network G? The key idea is the following qualitative prin-
ciples of gene network behaviours:
Compositional Analysis of Homeostasis of Gene Networks 195
Formula (1) corresponds to the case that the current expression level of gene x
is basal. Then, gene x will exceed x1 in a future, otherwise gene x will become
OFF. Formula (2) corresponds to the case that the level is beyond x1 , which
means ‘if gene x is ON and the current expression level of gene x is above x1 ,
gene x at least keeps its level as long as gene x is ON’. Since this formula only
prohibits that the expression level of gene x falls down below x1 even though
gene x is expressed, it is possible for gene x to attain the next level x2 in some
future. Thus, by this formula, we ensure that the level of gene x will increase
(or keep its level) when gene x is expressed. Formulas (3) and (4) are similar to
Formula (2). Note that, in formula (4), since the maximum level of gene x is x3 ,
it only says that gene x will keep its level as long as gene x is ON.
If a gene is OFF, its expression level decreases over time. We have the sym-
metric formulae to the above:
For this network let us check the property ‘each gene oscillates’, which is written
in LTL as:
2
The definition is based on our previous publication [11].
Compositional Analysis of Homeostasis of Gene Networks 197
Fig. 4. A gene network consisting of two genes x and y. Gene x receives positive input
from environment. (Source: [17]).
∀x̃.behave RS (x̃) |= ϕ,
the level of the input is beyond the threshold ex above which gene x is activated.
As a result, we have the following propositions: {in x , on x , on y , ex , xy , yx }. The
division of external propositions E and internal propositions I is as follows:
E = {in x }, I = {on x , on y , ex , xy , yx }. Note that the environment only controls
in x , which means that the environment is only able to determine whether it
inputs to gene x or not. Whether the level of the input exceeds the level ex is
determined by the behaviour specification. Thus ex is internal propositions. Note,
however, that through the behaviour description, environment can indirectly
control the truth of ex . The specification for change of levels of inputs is the
same as that of genes:
G(ex ∧ ¬yx → on x ),
Compositional Analysis of Homeostasis of Gene Networks 199
which says that if the input level is beyond the threshold ex and gene y is not
expressed beyond the threshold yx (i.e. proposition yx is false; ¬yx is true), then
gene x is ON, and the switching condition on gene x is described as
G(xy ↔ on y ).
Gin x .
Among these steps, the first step is known to be the bottleneck, since known
algorithms to translate LTL formulae into equivalent automata run in expo-
nential time with respect to the sizes of formulae. For large LTL formulae, the
algorithm often fails in the first step.
The compositional algorithm is devised to overcome the difficulty in the first
step. Without loss of generality, reactive system specifications are written as
ϕ = ϕ1 ∧ ϕ2 ∧ · · · ∧ ϕn 3 . In compositional approach, each sub-formula ϕi is
translated into automaton Aϕi and thus translated into a local game G(Aϕi ),
then the winning strategy is computed for each G(Aϕi ). Thanks to the nice
property of safety games, the winning strategy for G(Aϕ ) – the strategy for
the original game – can be computed from the winning strategies for each local
game G(Aϕi ). Since each ϕi is much smaller than the original ϕ, the automata
construction for each ϕi is much less demanding. Note that, however, we have
to pay extra cost to merge the winning strategies for local games, which we
do not in the monolithic (i.e. non-compositional) algorithm. If we manage to
divide ϕ into a suitable set of sub-formulae, we will gain much benefit from the
compositional algorithm.
This compositional algorithm is implemented in the tool Acacia+ [5]. Accord-
ing to the experiments reported, the compositional algorithm outperforms the
monolithic algorithm. However, the formulae are divided heuristically by hand.
There is no guideline how to obtain a good division.
To facilitate applying compositional algorithm, we invent a new clustering
algorithm that suitably divide LTL specifications. Here the clauses are to be
distributed
into
several clusters c ,
1 2c , . . . , ck , i.e. ϕ = ( c1 )∧( c2 )∧· · ·∧( ck )
where ci = φ∈ci φ.
The basic idea how to divide LTL specification is based on the following
observation. Let us assume that we have 20 clauses and are distributing them
into 2 clusters. Consider two clusterings: (1) to put 1 clause into one cluster and
19 clauses into the other, (2) to put 10 clauses into each cluster. Intuitively, the
second clustering seems better, since in such uneven clustering as the first one,
the 19 clauses cluster will be a bottleneck and the efficiency may not improve
as a whole. Thus a good clustering distributes clauses into clusters as equally as
possible. However, the number of clauses may not be a good criterion, since the
sizes of formulae are not the same; in an extreme situation, the size of a certain
formula might be equal to the size of the rest. Thus we distribute clauses into
clusters whose sizes are as equal as possible.
To facilitate the winning strategy computation of each local game, it is desir-
able to obtain small automata. The size of automata Aϕ is determined by the
number of sub-formulae in ϕ. To compute the number of sub-formulae for each
clause is not realistic because it is exponential to the size of the formula. For
example, for a formula (a ∧ ¬b) → F b, we have the following variations of sub-
formulae: a, b, ¬b, F b, a∧¬b and (a∧¬b) → F b. The plausible criterion is the num-
ber of variations of propositions – the more variations of propositions we have,
the more will be the number of sub-formulae. For example, (a → b)∧(a → c) has
3
Our network specifications actually conform to this form.
Compositional Analysis of Homeostasis of Gene Networks 201
the winning strategies for each local game. This cost will grow as the number of
clusters increases. Although the number of clusters may affect the overall perfor-
mance, we have no idea how to estimate the best number of clusters in advance.
Hence, we demonstrate our clustering method with several number of clusters
so that we can see the impact of the number of clusters.
The networks that we used in this experiment are depicted in Figs. 5, 6, 7,
8, 9, 10 and 11. The network ‘bistable switch’ is the network that consists of
two genes x and y where x activates y, y activates x and x receives negative
input from environment (Fig. 5). The network ‘bistable switch2’ is the same as
‘bistable switch’ except both gene x and y receive negative inputs from envi-
ronment (Fig. 6). The networks ‘anti-stress response (a)(b)(c)’ are the stress
response networks studied in [23] (Fig. 7). The networks ‘3 genes’ to ‘6 genes’
are depicted in Figs. 8, 9, 10 and 11. They are artificially synthesised for the
experiment. The homeostatic properties we analysed are tendency to ease stress
(described as G(stress → F ¬stress)) for anti-stress response networks and sta-
bility of a certain gene expression (described as Gon) for the others. We impose
as environmental assumptions the following formulae that says every input is
oscillating: G((input → F ¬input) ∧ (¬input → F input)).
Table 1. Experimental results. Columns ‘E’ and ‘I’ respectively show the numbers
of external propositions and internal propositions. Column ‘S’ shows the size of for-
mula. (i.e. number of connectives and propositions). The columns ‘Time(s)’ show
the total time of the verification for various number of clusters (k). ‘k = 1’ means
non-compositional analysis. For ‘k ≥ 1’ the times include computation of clustering.
(Source: [17]).
Network E I S Time(s)
k=1 k=2 k=3 k=4 k=5 k=6
Bistable switch 1 6 232 0.40 0.07 0.08 0.10 0.10 0.07
Bistable switch2 2 8 353 206.81 0.81 0.41 0.42 0.51 0.53
Anti-stress response (a) 1 9 289 0.48 0.31 0.35 0.41 0.47 0.51
Anti-stress response (b) 1 7 237 0.22 0.12 0.13 0.15 0.16 0.18
Anti-stress response (c) 1 9 288 0.39 0.31 0.38 0.41 0.47 0.51
3 genes 3 10 418 1617.87 5.03 5.43 6.08 5.82 6.30
4 genes 2 12 444 >3600 4.07 4.48 5.16 5.69 6.21
5 genes 3 15 547 >3600 173.31 123.80 192.17 135.77 194.75
6 genes 3 18 646 >3600 2676.37 2792.61 2877.03 2942.62 3087.91
The figures shown in the table for k ≥ 2 include the computational times of
clustering. They are less than 0.01 regardless of the number of clusters for any
network except for ’6genes’ at k = 6 (0.011 s). We can see that the compositional
algorithm (k ≥ 2) outperforms the monolithic algorithm (k = 1). The improve-
ment by the compositional algorithm increases as the size of the network grows.
Next, we compare our method to random clustering. For each network spec-
ification, we generate 50 random clusterings. The way to generate random clus-
terings is that we first randomly choose the size of each cluster, in other words,
we first determine how many clauses each cluster should include. Then, we ran-
domly distribute the clauses of the original formula to each cluster according to
the determined size.
We show the results of random clustering in Table 2. The figures shown in
the table are the average of the verification times for 50 random clusterings.
During the experiment on random clusterings, the realisability checking for some
clusterings did not finish within 1 h. Such clusterings are considered to be verified
in 3600 s for the sake of convenience. We marked the figures with stars on such
figures in the table. Thus the real averages must be greater than them.
Compositional Analysis of Homeostasis of Gene Networks 205
Table 4. Relative standing of the verification time with our method within the run-
times of random clusterings. The value ‘best’ means that our method was the best.
(Source: [17].)
Network Time(s)
k=2 k=3 k=4 k=5 k=6
Bistable switch 0.05 0.05 0.06 0.05 0.02
Bistable switch2 0.56 0.12 0.12 0.13 0.12
Anti-stress response (a) 0.08 0.07 0.06 0.06 0.07
Anti-stress response (b) 0.06 0.06 0.07 0.06 0.07
Anti-stress response (c) 0.08 0.07 0.07 0.06 0.08
3 genes 1.14 0.67 0.59 0.54 0.57
4 genes 0.27 0.17 0.17 0.16 0.17
5 genes 4.92 1.75 1.26 1.09 1.05
6 genes 13.49 3.49 2.18 1.85 1.75
Network Time(s)
k=2 k=3 k=4 k=5 k=6
Bistable switch 0.09 0.07 0.08 0.07 0.07
Bistable switch2 11.95 4.13 1.95 1.3 0.88
Anti-stress response (a) 0.13 0.11 0.12 0.09 0.10
Anti-stress response (b) 0.08 0.07 0.07 0.07 0.07
Anti-stress response (c) 0.12 0.10 0.09 0.09 0.09
3 genes 78.18 25.36 46.75 6.23 6.48
4 genes 684.28 437.44 162.09 40.93 125.57
5 genes 1666.97 731.27 643.03 713.69 716.28
6 genes 2478.64 1515.71 1205.51 1400.31 894.75
Table 7. The winning strategy computation times for our clustering method.
Network Time(s)
k=2 k=3 k=4 k=5 k=6 -
Bistable switch 0.02 0.03 0.04 0.05 0.05
Bistable switch2 0.25 0.29 0.3 0.38 0.41
Anti-stress response (a) 0.23 0.28 0.35 0.41 0.44
Anti-stress response (b) 0.06 0.07 0.08 0.10 0.11
Anti-stress response (c) 0.23 0.31 0.34 0.41 0.43
3 genes 3.89 4.76 5.49 5.28 5.72
4 genes 3.80 4.31 4.99 5.52 5.68
5 genes 168.39 122.04 190.90 134.67 193.69
6 genes 2662.88 2789.11 2874.84 2940.76 3086.15
Table 8. The averages of the winning strategy computation times for 50 random clus-
terings. The stars on the figures show that some trials failed in automaton construction
due to time out of 1 h. The averages for such clusters are calculated by ignoring them
from the samples.
Network Time(s)
k=2 k=3 k=4 k=5 k=6
Bistable switch 0.03 0.03 0.04 0.04 0.05
Bistable switch2 0.24 0.27 0.29 0.32 0.34
Anti-stress response (a) 0.23 0.28 0.31 0.35 0.38
Anti-stress response (b) 0.06 0.07 0.08 0.09 0.09
Anti-stress response (c) 0.23 0.27 0.31 0.34 0.37
3 genes 3.72 4.09 4.12 4.21 4.46
4 genes 3.55 4.17 4.42 5.06 5.09
5 genes 132.12 136.11 150.42 152.28 150.55
6 genes 2181.37 2343.29 2297.50 2236.08 2226.78
6 Related Work
As for the compositional approach to realisability problems, the tool Unbeast [6]
is proposed other than Acacia+ [10]. Unbeast divides a given LTL specification
into the part which satisfies safety and the other which does not. The advantage
of this approach is that the safety part can be easily translated into an automa-
ton. Intuitively, safety means that a system never falls into some undesirable
state, and often appears in LTL specifications as G¬error . Since our logical
specification mainly consists of non-safety clauses, Unbeast may not be suitable.
Compositional Analysis of Homeostasis of Gene Networks 209
7 Conclusion
Let A be a finite set. We write Aω for the set of all infinite sequences on A. We
write σ[i] for the i-th element of σ ∈ Aω . Let AP be a set of atomic propositions.
A linear time structure is a sequence σ ∈ (2AP )ω where 2AP is the powerset of
AP .
The formulae in LTL are defined inductively as follows: (i) p ∈ AP is a
formula, (ii) if φ is a formula, ¬φ is a formula, (iii) if φ and ψ are formulae,
φ ∧ ψ, φ ∨ ψ and φU ψ are also formulae. Now we define the semantics of LTL.
Let σ be a linear time structure and φ be a formula. We write σ |= φ for φ is true
in σ and we say σ satisfies φ. The satisfaction relation |= is defined as follows.
210 S. Ito et al.
σ |= p iff p ∈ σ[0] for p ∈ AP
σ |= ¬φ iff σ |= φ
σ |= φ ∧ ψ iff σ |= φ and σ |= ψ
σ |= φ ∨ ψ iff σ |= φ or σ |= ψ
σ |= φU ψ iff (∃i ≥ 0)(σ i |= ψ and ∀j(0 ≤ j < i)σ j |= φ)
References
1. Abadi, M., Lamport, L., Wolper, P.: Realizable and unrealizable specifications of
reactive systems. In: Ausiello, G., Dezani-Ciancaglini, M., Della Rocca, S.R. (eds.)
ICALP 1989. LNCS, vol. 372, pp. 1–17. Springer, Heidelberg (1989). https://doi.
org/10.1007/BFb0035748
2. Aoshima, T.: On a verification system for reactive system specifications. Ph.D.
thesis, Tokyo Institute of Technology (2003)
3. Arthur, D., Vassilvitskii, S.: K-means++: the advantages of careful seeding. In:
Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algo-
rithms, SODA 2007, pp. 1027–1035. Society for Industrial and Applied Mathemat-
ics, Philadelphia (2007)
4. Bloem, R., Cimatti, A., Greimel, K., Hofferek, G., Könighofer, R., Roveri, M.,
Schuppan, V., Seeber, R.: RATSY – a new requirements analysis tool with syn-
thesis. In: Touili, T., Cook, B., Jackson, P. (eds.) CAV 2010. LNCS, vol. 6174, pp.
425–429. Springer, Heidelberg (2010). https://doi.org/10.1007/978-3-642-14295-
6 37
5. Bohy, A., Bruyère, V., Filiot, E., Jin, N., Raskin, J.-F.: Acacia+, a tool for LTL
synthesis. In: Madhusudan, P., Seshia, S.A. (eds.) CAV 2012. LNCS, vol. 7358, pp.
652–657. Springer, Heidelberg (2012). https://doi.org/10.1007/978-3-642-31424-
7 45
6. Ehlers, R.: Symbolic bounded synthesis. Formal Methods Syst. Des. 40(2), 232–262
(2012). https://doi.org/10.1007/s10703-011-0137-x
7. Elowitz, M.B., Leibler, S.: A synthetic oscillatory network of transcriptional regu-
lators. Nature 403, 335–338 (2000)
8. Emerson, E.A.: Temporal and modal logic. In: van Leeuwen, J. (ed.) Handbook of
Theoretical Computer Science, Volume B: Formal Models and Semantics (B), pp.
995–1072. MIT Press, Cambridge (1990)
Compositional Analysis of Homeostasis of Gene Networks 211
9. Filiot, E., Jin, N., Raskin, J.-F.: An antichain algorithm for LTL realizability. In:
Bouajjani, A., Maler, O. (eds.) CAV 2009. LNCS, vol. 5643, pp. 263–277. Springer,
Heidelberg (2009). https://doi.org/10.1007/978-3-642-02658-4 22
10. Filiot, E., Jin, N., Raskin, J.F.: Antichains and compositional algorithms for LTL
synthesis. Formal Methods Syst. Des. 39(3), 261–296 (2011)
11. Ito, S., Hagihara, S., Yonezaki, N.: A qualitative framework for analysing home-
ostasis in gene networks. In: Proceedings of BIOINFORMATICS 2014, pp. 5–16
(2014)
12. Ito, S., Hagihara, S., Yonezaki, N.: Approximate analysis of homeostasis of gene
networks by linear temporal logic using network motifs. In: BIOINFORMATICS
2015 - Proceedings of the International Conference on Bioinformatics Models,
Methods and Algorithms, Lisbon, Portugal, 12–15 January 2015, pp. 93–101 (2015)
13. Ito, S., Hagihara, S., Yonezaki, N.: Formulation of homeostasis by realisability on
linear temporal logic. In: Plantier, G., Schultz, T., Fred, A., Gamboa, H. (eds.)
BIOSTEC 2014. CCIS, vol. 511, pp. 149–164. Springer, Cham (2015). https://doi.
org/10.1007/978-3-319-26129-4 10
14. Ito, S., Ichinose, T., Shimakawa, M., Izumi, N., Hagihara, S., Yonezaki, N.: Modular
analysis of gene networks by linear temporal logic. J. Integr. Bioinf. 10(2), 12–23
(2013). https://doi.org/10.2390/biecoll-jib-2013-216
15. Ito, S., Ichinose, T., Shimakawa, M., Izumi, N., Hagihara, S., Yonezaki, N.:
Qualitative analysis of gene regulatory networks by temporal logic. Theor.
Comput. Sci. 594(23), 151–179 (2015). https://doi.org/10.1016/j.tcs.2015.06.017.
http://www.sciencedirect.com/science/article/pii/S0304397515005277
16. Ito, S., Izumi, N., Hagihara, S., Yonezaki, N.: Qualitative analysis of gene regula-
tory networks by satisfiability checking of linear temporal logic. In: Proceedings of
BIBE 2010, pp. 232–237 (2010). https://doi.org/10.1109/BIBE.2010.45
17. Ito, S., Osari, K., Hagihara, S., Yonezaki, N.: Efficient analysis of homeostasis of
gene networks with compositional approach. In: Proceedings of the 10th Inter-
national Joint Conference on Biomedical Engineering Systems and Technologies
(BIOSTEC 2017) - Volume 3: BIOINFORMATICS, Porto, Portugal, 21–23 Febru-
ary 2017, pp. 17–28 (2017). https://doi.org/10.5220/0006093600170028
18. Jobstmann, B., Bloem, R.: Optimizations for LTL synthesis. In: Proceedings of the
Formal Methods in Computer Aided Design, FMCAD 2006, pp. 117–124. IEEE
Computer Society, Washington (2006)
19. Jobstmann, B., Galler, S., Weiglhofer, M., Bloem, R.: Anzu: a tool for property
synthesis. In: Damm, W., Hermanns, H. (eds.) CAV 2007. LNCS, vol. 4590, pp.
258–262. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-73368-
3 29
20. Mori, R., Yonezaki, N.: Several realizability concepts in reactive objects. In: Infor-
mation Modeling and Knowledge Bases IV, pp. 407–424 (1993)
21. Pnueli, A., Rosner, R.: On the synthesis of a reactive module. In: POPL 1989:
Proceedings of the 16th ACM SIGPLAN-SIGACT Symposium on Principles of
Programming Languages, pp. 179–190. ACM, New York (1989)
22. Ward, J.H.: Hierarchical grouping to optimize an objective function. J. Am. Stat.
Assoc. 58(301), 236–244 (1963)
23. Zhang, Q., Andersen, M.E.: Dose response relationship in anti-stress gene regula-
tory networks. PLoS Comput. Biol. 3(3), e24 (2007)
Fast and Sensitive Classification of Short
Metagenomic Reads with SKraken
1 Introduction
Metagenomics is a study of the heterogeneous microbes samples (e.g. soil, water,
human microbiome) directly extract from the natural environment with the pri-
mary goal of determining the taxonomical identity of the microorganisms resid-
ing in the samples. It is an evolutionary revise, shifting focuses from the indi-
vidual microbe study to a complex microbial community. As already mentioned
in [1,2], the classical genomic-based approaches require the prior clone and cul-
turing for the further investigation. However, not all bacteria can be cultured.
The advent of metagenomics succeeded to bypass this difficulty.
The study of metagenomics is far more than just about labeling the species
of the microbes. The analysis can reveal the presence of unexpected bacteria and
viruses in a microbial sample, also allows the identification and characterization
of bacterial and viral genomes at a level of detail not previously possible. For
example, in the case of the human body, imbalances in the microbiome are related
with many diseases, e.g. inflammatory bowel disease (IBD) [3] and colorectal
cancer [4]. Furthermore it may aid the researchers to systematically understand
and characterize the microbial communities: how genes influence each others
activities in serving collective functions; grasp the collections of communities
that composites the biosphere where the humans be a part, etc. It has already
been applied in many areas like ecology, medicine, microbiology and some others
mentioned in [3–6].
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 212–226, 2018.
https://doi.org/10.1007/978-3-319-94806-5_12
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 213
of representative k-mers over the taxonomic tree; (ii) SKraken improves the
precision over Kraken on both simulated and real metagenomic datasets with-
out compromising the recall; (iii) benefit from the downsized database. As a
consequence, SKraken requires less memory RAM and the classification speed
increases w.r.t. Kraken. In the next section we will give an overview of Kraken
and analyze how to improve the classification. SKraken is presented in Sect. 2.2.
Both tools are tested on simulated and real metagenomic datasets in Sect. 3 and
the conclusions are drawn in Sect. 4.
2 Method
2.1 An Overview of Kraken
In order to better understand our contribution let’s briefly present Kraken in
the first place. Instead of utilizing the complete genome as reference, Kraken
considers only its k-mers, as well as many other tools [14,15], thus a genome
sequence is alternatively represented by a set of k-mers, which plays a role of effi-
ciently indexing a large volume of target-genomes database, e.g., all the genomes
in RefSeq. This idea is stemmed from alignment-free methods [18] and some
researchers have verified its availability in different applications. For instance,
the construction of phylogenetic trees, traditionally is performed based on a
multiple-sequence alignment technique, the process is be carried out on the whole
genomes [19,20], in practice it’s difficult to realize due to the improper length
of whole genomes. The alignment-free technique with the usage of k-mers is the
solver. Recently some variations of k-mers-based methods have been devised for
the detection of enhancers in ChIP-Seq data [21–25] and entropic profiles [26,27].
Recently, the assembly-free comparison of genomes and metagenomes based on
NGS reads and k-mers counts has been investigated in [28–31]. For a compre-
hensive review of alignment-free measures and applications we refer the reader
to [18].
Considering the taxonomic tree, taken from the complete NCBI taxonomic
information, this data structure is extended by annotating each node with k-
mers, including leaves and internal nodes. Every node is associated with a list
of k-mers that are critical for the future classification. More precisely, given a
dataset of target genomes, the construction of this annotated taxonomic tree
is carried out by scanning the k-mers of each genome in the dataset. If the
k-mer appears only in a given genome, it is associated only to one leaf node J
(representing that genome) and the k-mers list of node J is updated. If the k-mer
appears in more than one species, the k-mer is erased from those corresponding
nodes and moved to the lowest common ancestor of these nodes, see Fig. 1 for
an example. At the end of this step each k-mer belongs to only one node in the
taxonomic tree.
Once this database of annotated k-mers has been constructed, Kraken can
classify reads in a very efficient manner. Figure 2 reports an overview of the
classification process. Later when we launch the classification procedure, with a
given read, Kraken firstly decomposes the read into a list of its k-mers. Then each
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 215
Fig. 1. In this example the k-mer AGCCT , that is contained in the species 9 and 13,
is moved to the lowest common ancestor, the family node 2 (figure taken from [17]).
The most important step of Kraken is the construction of the augmented taxo-
nomic tree, annotating by a list of k-mers per node in order to implement the
classification. In this paper we propose SKraken that has an exclusive procedure,
216 J. Qian et al.
distinguishable from Kraken, we select and filter k-mers instead. Therefore, those
uninformative k-mers will be pruned away from the augmented taxonomic tree.
It may occur that one k-mers appears in more than one species, like demon-
strated in Fig. 1, sequence “AGCCT” belongs to two species. Kraken adds this
k-mers into the ancestor node 2 (representing a taxonomic family) out of node 9
and node 13. Since this k-mers will be used in the classification step, we would
like to be informative for the family node 2. However, the majority of species in
this family, nodes 10, 11 and 12, do not contain this k-mers.
To address this issue, for each k-mer, we define a scoring function that cap-
tures its representativeness within the taxonomic node. We recall that a k-mer
is associated with only one node in the tree. Let’s define T axID(m) that indi-
cates the taxonomic node associated with the k-mer m. However, before the
upward shift of k-mer, it is possible to occur in more than one species. We
define N umSpecies(m) as the number of species that contains m. By construc-
tion T axID(m) is the lowest common ancestor of all these species. Thus the
species in which m appears, they are all leaf nodes of the subtree T axID(m).
We define T otSpecies(n) as the total number of species in the subtree routed
in the node n. With these values we define q(m) the quality of a k-mer m as
(equation from [17]):
N umSpecies(m)
q(m) =
T otSpecies(T axID(m))
Figure 3 shows an example of the quality q(GAACT ). The quality of GAACT
can also be interpreted as the percentage of species nodes that contains
GAACT , i.e. N umSpecies(GAACT ), with respect to the family node 2, i.e.,
T axID(GAACT ), in this case 60%. Similarly, if we consider the example in
N umSpecies(AGCCT )
Fig. 1, the quality of q(AGCCT ) = T otSpecies(T 2
axID(AGCCT )) = 5 = 0.4,
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 217
that is 40%. We can conclude as, a k-mer has an high quality can be consid-
ered representative for a given taxonomic node and the relative sub-tree, thus
more likely will be informative for the classification. Based on these observations
SKraken tries to screen out the uninformative k-mers by means of their quality
value, and prunes the augmented taxonomic tree by removing the k-mers with
a quality below a given threshold Q.
In order to compute the quality scores q(m) for all k-mers we have to evaluate
N umSpecies(m) and T otSpecies(n) efficiently. Here is our setting: the construc-
tion of the augmented taxonomic tree of SKraken is divided into two steps. In
the first step, given a set of target genomes, we scan the k-mers of each genome
and build the augmented taxonomic tree, similarly to Kraken. Meanwhile, we
set a variable to keep tracking N umSpecies(m), as soon as m is found in a
new species we increment this variable. However, there can be genomes that are
further classified as sub-species of a given species node. In order to compute the
correct value of N umSpecies(m), we need to make sure that all genomes of a
given species are processed before moving to next species. This can be tackled
by scanning the input genomes in a particular order so that all genomes of a
species and the sub-species, are processed at once. Another problem is the fact
that a k-mer can appear in some sub-species of a given species node. When
computing N umSpecies(m) we must ensure not to overcount these occurrences,
and thus the corresponding variable is incremented only when m is found for the
first time in a given species. All other occurrences of m within the same species
will be discarded. At the end of the first step we have computed the augmented
taxonomic tree, with all k-mers, and the corresponding values N umSpecies(m).
In the second step SKraken computes the quality values q(m) and filters
uninformative k-mers. The number of leaf nodes that are the descendants of n,
indicated as T otSpecies(n), can be obtained for all nodes in the tree with a post-
order traversal of the taxonomic tree. Consequently all k-mers are processed and
the corresponding qualities q(m) are computed. If q(m) is below a given input
parameter Q, m is removed from the database.
Note that the size of the taxonomic tree is constant and much smaller with
respect to the number of k-mers. The overall process depends only on the total
number of k-mers and its size is linear to the input reference genomes. Once
the augmented taxonomic tree is build, reads can be classified with the same
procedure of Kraken.
3 Results
3.1 Datasets
Before the demonstration of result, we need to build a reference dataset. We
conduct the experiments both on real and simulated datasets with different
bacterial and archaeal genomes from NCBI RefSeq in order to capture a com-
prehensive understanding of the performance. The simulated and real datasets
are acquired from the original paper of Kraken [13] as well as from other related
studies [14,32–34]. The simulated datasets represent five mock communities that
218 J. Qian et al.
are constructed from real sequencing data: MiSeq, HiSeq, Mix1, Mix2, simBA5.
MiSeq and HiSeq metagenomes were built using 10 sets of bacterial whole-
genome shotgun reads. Mix1 and Mix2 are based on the same species of HiSeq,
but with two different abundance profiles.
The MiSeq dataset is particularly difficult to analyze because it contains five
genomes from the Enterobacteriaceae family (Citrobacter, Enterobacter, Kleb-
siella, Proteus and Salmonella). The high sequence similarity of this family can
elevate the difficulty of classification. The metagenome simBA5 was created by
simulating reads from the complete set of bacterial and archaeal genomes in Ref-
Seq, for a total of 1216 species. It contains reads with sequencing errors caused
in the reading process and it was created with the exact purpose of measuring
the stability with existing errors in a complex communities.
We also evaluated the performance of SKraken on a real stool metagenomic
sample (SRR1804065) from the Human Microbiome Project. Because there is
no ground truth for this dataset, we use BLAST with a sequence identity of
95% to find the reads that uniquely mapped to a genome and filter out all
other reads. If two paired-end reads do not map to the same genome, we discard
them. As a result, the real metagenomic sample contains 775 distinct species
and 1053741 reads. A summary of the main characteristics of all simulated and
real metagenomics datasets can be found in Table 1.
In order to compare the results we used the standard metrics of precision and
recall. Given N the number of reads, Y the number of reads classified and X the
number of reads correctly classified, we define precision as the fraction of correct
assignments over the total number of assignments (X/Y), and recall as the ratio
between the number of correct assignments and the number of reads (X/N).
Note that our classification task will be unfolded from both genus and species
level, where genus level is the parent of species level in the taxonomic tree. If
the classification step selected a node at genus level, while we are evaluating the
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 219
species level, even if the assigned genus-level node is the parent of the correct
leaf node, we considered it as mislabeled. On the other hand, when we evaluate
the genus-level, if the classification step selected a taxonomic node that is a
descendant (leaf node, species level) of the correct (genus-level) node, we consider
it to be correct.
When analyzing a metagenomic sample one needs to verify whether the esti-
mated abundance ratios of species is similar to the known profile, here we adopt
Pearson correlation that is a technique widely used. If the correlation value is
close to 1, which means the estimated abundance ratio perfectly matches the
known one. In the next sections, we will test the behavior of SKraken on these
three aspects: precision, recall and Pearson correlation.
3.3 Comparison
For a better perception of the performance of Skraken, we comparison it with
Kraken, since it is one of the cutting-edge tools as mentioned in [16].
For Kraken we use the default parameter k = 31, as suggest by the authors
[13], it is the best balance between precision and recall. For SKraken we use
the same value k = 31 and we test the performance by varying the tuning
parameter Q.
Fig. 4. Results on dataset Mix1 varying the filtering parameter Q, the figure taken
from [17].
We devised a series of tests with the variable parameter Q and the taxonomic
level at which the classification is evaluated. In the first set we want to test how
the filtering parameter Q impact the performance metrics. We run Kraken and
SKraken on the dataset Mix1 and evaluate the classification accuracy at the
species-level. The results are reported in Fig. 4. If the parameter Q = 0 (no
filtering) all k-mers are kept as we expected, thus the performance of Kraken
and SKraken are identical. As Q grows to 100%, we can see that the precision
improves from 63% to 75%, whereas the recall remains constant, which implies
the number of mistakenly-classified reads decreases. Another important obser-
vation is that the Pearson correlation with the known abundance ratios also
increases.
220 J. Qian et al.
Thus, we use the most stringent filtering (Q = 100%) to classify all dataset
at the species-level. Figure 5 shows a summary of precision and recall for all
simulated and real metagenomic datasets. This test confirms that SKraken is
able to improve the precision on all datasets without compromising the recall.
On simulated metagenomes the average precision increases from 73% of Kraken
to 81% of SKraken. On the real metagenome, the accuracy of Kraken turns out
to be 91%, SKraken achieves 96%.
Species Genus
Kraken SKraken Kraken SKraken
HiSeq Correct assignments 6049 6051 7800 7699
Reads classified 7890 7252 7890 7773
MiSeq Correct assignments 6260 6256 7258 7254
Reads classified 7994 7342 7994 7884
SimBA5 Correct assignments 7688 7687 8700 8561
Reads classified 9327 8513 9327 8929
Mix1 Correct assignments 535809 535798 848221 838720
Reads classified 854256 713935 854255 844077
Mix2 Correct assignments 529717 529734 756196 754441
Reads classified 766829 729856 766829 763114
SRR1804065 Correct assignments 902171 902270 966004 965859
Reads classified 994416 944000 994416 982962
222 J. Qian et al.
In the last experiment we test the ability to detect the correct abundance
ratios in a metagenomic sample, we compared it at both genus level and species
level of SKraken and Kraken, as displayed in Fig. 8. Both approaches have an
extraordinary result in genus level, in particular, they virtually peaked in dataset
HISeq, MIX1, MIX2 and SRR1804065. In dataset simBA5, SKraken increases
the score from 0.92 to 0.97, since it is one of the most complex and realistic
dataset, with 1216 species. If we compare these Pearson correlations with those
of species level classification in general the values decrease confirming that it
is more difficult to detect the correct species, rather than the genus. This is
the case where the classification accuracy can benefit from a careful selection
of discriminative k-mers. In fact in all dataset the correlation of SKraken is
better than Kraken. Again, in one of the most difficult metagenome(simBA5)
that contains 1216 species, the improvement is substantial from 0.61 of Kraken
to 0.77 of SKraken.
Fig. 8. The Pearson correlation of the estimated abundances with the correct ratios
for various level of classification and parameters.
All the results above have shown that SKraken enables to obtain a higher
classifying precision without any expense of recall. In other words, more reads
are correctly classified, and the estimated abundance ratios have a better match
with the known profile. An important property of SKraken is that the impact
on these metrics improves as the taxonomic level evaluated in the classifica-
tion becomes lower and thus more difficult. Moreover, as the number of newly
sequenced species grows, the probability that two non-related species share a
given k-mer will grow. For this reason we conjecture that SKraken will be able
to remove more uninformative k-mers as the number of sequenced genomes
increases.
Besides the enhancement of precision, the screening processing has another wel-
comed “side effect”, it reduced the database size since the number of annotated
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 223
k-mers decreases. The size of the database produced by Kraken, using all bac-
terial and archaeal complete genomes in NCBI RefSeq, is about 65 GB and it
contains 5.8 billion k-mers. In Fig. 9 we evaluate the percentage of k-mers fil-
tered by SKraken and the impact in memory for different values of threshold
Q. As expected, the percentage of k-mers filtered grows with the threshold Q
increases and it reaches the maximum of 8.1% with Q = 100. By construction,
the impact in memory depends linearly on the number of k-mers to be indexed.
When Q = 100, SKraken requires to index 5.3 billion k-mers in space of 60 GB.
Fig. 9. Percentage of k-mers filtered and database size as a function of the quality
threshold Q (taken from [17]).
Kraken SKraken
HiSeq 126 134
MiSeq 210 231
SimBA5 223 241
Mix1 473 498
Mix2 481 522
SRR1804065 694 754
4 Conclusion
We have presented SKraken, an efficient and effective method for the genus-level
and species-level classifications. It is based from Kraken, with the additional
224 J. Qian et al.
procedure where we detect the representative k-mers and filter the unquali-
fied ones, which produces an improved dictionary. The experimental result has
demonstrated that SKraken obtains a higher precision without compromising
the recall, moreover it boosts the processing speed. As future direction of inves-
tigation it would be interesting to explore alternative way to define the k-mer
scores incorporating other topological information of the tree of life.
Acknowledgement. The authors would like to thank the anonymous reviewers for
their valuable comments and suggestions. This work was supported by the Italian
MIUR project PRIN20122F87B2.
References
1. Felczykowska, A., Bloch, S.K., Nejman-Faleczyk, B., Baraska, S.: Metagenomic
approach in the investigation of new bioactive compounds in the marine environ-
ment. Acta Biochim. Pol. 59(4), 501–505 (2012)
2. Mande, S.S., Mohammed, M.H., Ghosh, T.S.: Classification of metagenomic
sequences: methods and challenges. Briefings Bioinform. 13(6), 669–681 (2012)
3. Qin, J., Li, R., Raes, J., et al.: A human gut microbial gene catalogue established
by metagenomic sequencing. Nature 464, 59–65 (2010)
4. Zeller, G., Tap, J., Voigt, A.Y., Sunagawa, S., Kultima, J.R., Costea, P.I., Amiot,
A., Böhm, J., Brunetti, F., Habermann, N., Hercog, R., Koch, M., Luciani, A.,
Mende, D.R., Schneider, M.A., Schrotz-King, P., Tournigand, C., Tran Van Nhieu,
J., Yamada, T., Zimmermann, J., Benes, V., Kloor, M., Ulrich, C.M., von Knebel
Doeberitz, M., Sobhani, I., Bork, P.: Potential of fecal microbiota for early-stage
detection of colorectal cancer. Mol. Syst. Biol. 10(11), 766 (2014)
5. Human Microbiome Project Consortium: Structure, function and diversity of the
healthy human microbiome. Nature 486(7402), 207–214 (2012)
6. Said, H.S., Suda, W., Nakagome, S., Chinen, H., Oshima, K., Kim, S., Kimura,
R., Iraha, A., Ishida, H., Fujita, J., Mano, S., Morita, H., Dohi, T., Oota, H.,
Hattori, M.: Dysbiosis of salivary microbiota in inflammatory bowel disease and
its association with oral immunological biomarkers. DNA Res.: Int. J. Rapid Publ.
Rep. Genes Genomes 21(1), 15–25 (2014)
7. Brown, C., Hug, L., Thomas, B., Sharon, I., Castelle, C., Singh, A., et al.: Unusual
biology across a group comprising more than 15% of domain Bacteria. Nature
523(7559), 208–211 (2015)
8. Zhang, Z., Schwartz, S., Wagner, L., Miller, W.: A greedy algorithm for aligning
DNA sequences. J. Comput. Biol. 7(1–2), 203–214 (2004)
9. Huson, D.H., Auch, A.F., Qi, J., Schuster, S.C.: Megan analysis of metagenomic
data. Genome Res. 17, 377–386 (2007)
10. Caporaso, J.G., Kuczynski, J., Stombaugh, J., Bittinger, K., Bushman, F.D.,
Costello, E.K., Fierer, N., Pea, A.G., Goodrich, J.K., Gordon, J.I., Huttley, G.A.,
Kelley, S.T., Knights, D., Koenig, J.E., Ley, R.E., Lozupone, C.A., McDonald, D.,
Muegge, B.D., Pirrung, M., Reeder, J., Sevinsky, J.R., Turnbaugh, P.J., Walters,
W.A., Widmann, J., Yatsunenko, T., Zaneveld, J., Knight, R.: Qiime allows anal-
ysis of high-throughput community sequencing data. Nat. Methods 7(5), 335–336
(2010)
Fast and Sensitive Classification of Short Metagenomic Reads with SKraken 225
11. Liu, B., Gibbons, T., Ghodsi, M., Treangen, T., Pop, M.: Accurate and fast estima-
tion of taxonomic profiles from metagenomic shotgun sequences. BMC Genomics
12, P11 (2011)
12. Segata, N., Waldron, L., Ballarini, A., Narasimhan, V., Jousson, O., Huttenhower,
C.: Metagenomic microbial community profiling using unique clade-specific marker
genes. Nat. Methods 9, 811 (2012)
13. Wood, D., Salzberg, S.: Kraken: ultrafast metagenomic sequence classification
using exact alignments. Genome Biol. 15, R46 (2014)
14. Ounit, R., Wanamaker, S., Close, T.J., Lonardi, S.: Clark: fast and accurate classi-
fication of metagenomic and genomic sequences using discriminative k-mers. BMC
Genomics 16(1), 1–13 (2015)
15. Ames, S.K., Hysom, D.A., Gardner, S.N., Lloyd, G.S., Gokhale, M.B., Allen, J.E.:
Scalable metagenomic taxonomy classification using a reference genome database.
Bioinformatics 29, 2253–2260 (2013)
16. Lindgreen, S., Adair, K.L., Gardner, P.: An evaluation of the accuracy and speed
of metagenome analysis tools. Sci. Rep. 6, 19233 (2016)
17. Marchiori, D., Comin, M.: Skraken: fast and sensitive classification of short metage-
nomic reads based on filtering uninformative k-mers. In: Proceedings of the 10th
International Joint Conference on Biomedical Engineering Systems and Technolo-
gies (BIOSTEC 2017), pp. 59–67 (2017)
18. Vinga, S., Almeida, J.: Alignment-free sequence comparison-a review. Bioinformat-
ics 19, 513–523 (2003)
19. Comin, M., Verzotto, D.: Whole-genome phylogeny by virtue of unic subwords. In:
2012 23rd International Workshop on Database and Expert Systems Applications
(DEXA), pp. 190–194, September 2012
20. Sims, G.E., Jun, S.R., Wu, G.A., Kim, S.H.: Alignment-free genome comparison
with feature frequency profiles (FFP) and optimal resolutions. Proc. Nat. Acad.
Sci. 106, 2677–2682 (2009)
21. Antonello, M., Comin, M.: Fast alignment-free comparison for regulatory sequences
using multiple resolution entropic profiles. In: Proceedings of the International
Conference on Bioinformatics Models, Methods and Algorithms (BIOSTEC 2015),
pp. 171–177 (2015)
22. Comin, M., Antonello, M.: On the comparison of regulatory sequences with mul-
tiple resolution entropic profiles. BMC Bioinf. 17(1), 130 (2016)
23. Comin, M., Verzotto, D.: Beyond fixed-resolution alignment-free measures for
mammalian enhancers sequence comparison. IEEE/ACM Trans. Comput. Biol.
Bioinf. 11(4), 628–637 (2014)
24. Goke, J., Schulz, M.H., Lasserre, J., Vingron, M.: Estimation of pairwise sequence
similarity of mammalian enhancers with word neighbourhood counts. Bioinformat-
ics 28(5), 656–663 (2012)
25. Kantorovitz, M.R., Robinson, G.E., Sinha, S.: A statistical method for alignment-
free comparison of regulatory sequences. Bioinformatics 23, i249–i255 (2007)
26. Comin, M., Antonello, M.: Fast computation of entropic profiles for the detection
of conservation in genomes. In: Ngom, A., Formenti, E., Hao, J.-K., Zhao, X.-M.,
van Laarhoven, T. (eds.) PRIB 2013. LNCS, vol. 7986, pp. 277–288. Springer,
Heidelberg (2013). https://doi.org/10.1007/978-3-642-39159-0 25
27. Antonello, M., Comin, M.: Fast entropic profiler: an information theoretic approach
for the discovery of patterns in genomes. IEEE/ACM Trans. Comput. Biol. Bioinf.
11(3), 500–509 (2014)
226 J. Qian et al.
28. Schimd, M., Comin, M.: Fast comparison of genomic and meta-genomic reads with
alignment-free measures based on quality values. BMC Med. Genomics 9(1), 41–50
(2016)
29. Comin, M., Leoni, A., Schimd, M.: Clustering of reads with alignment-free mea-
sures and quality values. Algorithms Mol. Biol. 10(1), 1–10 (2015)
30. Comin, M., Schimd, M.: Assembly-free genome comparison based on next-
generation sequencing reads and variable length patterns. BMC Bioinf. 15(9), 1–10
(2014)
31. Ondov, B.D., Treangen, T.J., Melsted, P., Mallonee, A.B., Bergman, N.H., Koren,
S., Phillippy, A.M.: Mash: fast genome and metagenome distance estimation using
MinHash. bioRxiv (2016)
32. Girotto, S., Pizzi, C., Comin, M.: Metaprob: accurate metagenomic reads binning
based on probabilistic sequence signatures. Bioinformatics 32(17), i567–i575 (2016)
33. Girotto, S., Comin, M., Pizzi, C.: Metagenomic reads binning with spaced seeds.
Theor. Comput. Sci. 698, 88–99 (2017)
34. Girotto, S., Comin, M., Pizzi, C.: Higher recall in metagenomic sequence classifi-
cation exploiting overlapping reads. BMC Genomics 18, 917 (2017)
Computational Identification of Essential
Genes in Prokaryotes and Eukaryotes
1 Introduction
If a disruption of a gene is lethal, the gene is defined to be an essential gene
(EG) [1,2]. Identification of these genes is not only important for understanding
gene function and the relationship between genotype and phenotype, but also
has numerous advantages in the fields of synthetic biology and biomedicine.
The EGs subset is considered as the minimal requirement to sustain the life of
an organism and researchers are re-engineering microorganisms with a minimal
genome [3]. EGs of pathogens can also be used as potential drug targets [4,5].
Different knock-out and gene-disruption experimental procedures were used
to identify EGs. The methods include single gene knock-out [6,7], transposon
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 227–247, 2018.
https://doi.org/10.1007/978-3-319-94806-5_13
228 D. Nigatu and W. Henkel
mutagenesis [8], and RNA interference [9]. Recently, the CRISPR (clustered reg-
ularly interspaced short palindromic repeats) gene-editing technology has also
been used [10–12]. The experimental methods are very accurate. However, they
are expensive and time-consuming. Hence, computational methods of EG pre-
dictions provide a worthwhile alternative. The earliest computational methods
were based on homology mappings [13]. Afterwards, as the EG annotations of
many model organisms became available in public databases, machine-learning
based methods using features that characterize EGs became popular.
Gene essentiality has been associated with a lot of features. Broadly, the fea-
ture types can be classified as features related to sequence composition [14–16],
network topology [17–20], subcellular localization [20–22], and gene expression
[20,23]. The sequence-based features can be directly obtained from the genetic
sequences. These include protein length, GC content, codon bias, amino acid
composition, and domain information. In addition, homology information is also
based on the genetic sequences. Nevertheless, homology mapping requires a com-
putationally expensive sequence alignment and data base search. Network topol-
ogy features are obtained from the properties of protein-protein interaction, gene
regulatory, or metabolic networks. The most common network based property
which is often exploited is degree centrality (EGs are often hubs). Features based
on gene-expression data make use of mRNA expression and fluctuation levels.
With the exception of the sequence-based features, computation of the other
features rely on the availability of additional experimental data. For the well
studied model organisms, gene-expression and interaction networks are avail-
able. However, for newly sequenced and understudied species, one needs to per-
form gene-expression analysis and construct protein/gene networks beforehand.
Hence, EG predictions which are based only on sequence information are very
important.
Many studies have proposed EG predictors which are sequence-based [14–
16,21,22,24–26]. Ning et al. [14] used nucleotide, di-nucleotide, codon, and amino
acid frequencies along with codon bias information provided by the analysis soft-
ware called CodonW (http://codonw.sourceforge.net). The most popular among
the CodonW features, which showed a good predictive power and, hence, used by
a lot of sequence-based predictors, is Codon Adaptation Index (CAI). CAI mea-
sures the relative adaptability of the codon usage in a gene to the codon usage of
highly expressed genes [27]. Ning et al. performed cross-validation experiments
considering 16 bacteria species. The other very effective EG predictor which
uses sequence and sequence-derived properties is Song et al.’s ZUPLS [15]. The
features used by ZUPLS are mainly from the so-called Z-curve features and
other sequence-based features such as size, CAI, strand, homology, and domain
enrichment scores. Cross-organism results were shown using models trained on
E. coli and B. subtilis. Among the sequence-based methods, ZUPLUS seems
to be the best method. Palaniappan and Mukherjee [21], in 2011, proposed a
machine learning based EG predictor using sequence and pysio-chemical prop-
erties, plus cellular localization information. In addition to predictions of EGs
between organisms, they showed results at higher taxonomic levels. In 2017,
Liu et al. [22], using a feature which measures long-rage correlation (the Hurst
Computational Identification of Essential Genes 229
Following our study in [35], Shannon and Gibbs entropies were used to character-
ize the degree of randomness and thermodynamic stability in the gene sequences.
MI and CMI has been widely used in the fields of bioinformatics and computa-
tional biology for a number of applications, including as a genomic signature [36],
metric for phylogeny [37], and for identifying single nucleotide polymorphisms
(SNPs) [38]. We used MI and CMI to measure the sequence organizational differ-
ences between EGs and NEGs. The Markov features were selected for measuring
statistical dependencies in the gene sequences.
In the present work, in addition to the 15 bacterial species, essential gene
predictions in Archea and Eukaryotic species were performed. Identification of
EGs in four eukaryotes, including humans, and 1 archaeon was performed using a
Random Forest classifier. More information-theoretic features, Kullback-Leibler
divergence (KLD) between the distribution of k-mers (k = 1, 2, 3) in the genes
and the background k-mer distributions in the training organism, the total MI,
total CMI, and two additional entropy features, were added. Moreover, with the
hope of increasing the prediction performance, other non-information-theoretic
features, which can be easily obtained from the genetic sequences and known
to have a correlation with gene essentiality, were included. The added features
are related to optimized stop codon usage, gene length, and GC content. The
predictive power of the five feature sets, both individually and collectively, was
assessed. In addition to intra-organism predictions where both the training and
testing examples are taken from a single species, cross-organismic predictions
were also performed based on a pairwise and leave-one-species-out schemes. In
the pairwise scheme, models trained on E. coli and B. subtilis were used to
predict the EGs of the other bacteria, while in the leave-one-species-out scheme,
EGs of the left out species are predicted using a model trained on the remaining
bacteria. The obtained results are with multiple existing and state-of-the-art EG
prediction methods.
The data for EGs and non essential genes (NEGs) for the 15 bacteria and 1
archaeon were downloaded from the database of essential genes (DEG 13.5).
DEG collects a comprehensive list of essential and non-essential genes identified
by various researchers through gene knock-out experimental methods [39]. In
DEG, although the EGs dataset for eukaryotes are available, the list of NEGs are
not included. One way to deal with this is, as done in most of the gene essentiality
prediction studies, to regard all other genes as NEGs. Since some studies consider
and test a small number of genes (small-scale screenings), taking the untested
genes as non-essential could be misleading. Hence, for eukaryotic EG predictions,
we used the dataset presented by the database of Online GEne Essentiality
(OGEE) [40]. The species used in this study along with the number of EGs
and NEGs are listed in Table 1 [34]. The genome sequences were obtained from
Computational Identification of Essential Genes 231
Table 1. Names and abbreviations of the species used in this study. The accession
numbers along with the number of essential and non-essential genes are listed.
where Ω is the set of nucleotides {A, T, C, G}, P (x) and P (y) are the marginal
probabilities and P (x, y) is the joint probability. In the corresponding gene
sequences, the probabilities are estimated from the relative frequencies. In addi-
tion to the total mutual information computed according to Eq. (1), for each
base pair (x, y), the quantity P (x, y) log2 PP(x)P
(x,y)
(y) is calculated and used as a
feature. Therefore, a total of 17 MI-related features were calculated.
where P (x, y, z), P (x, z), and P (y, z) are the joint pmfs of the random variables
shown in brackets. The three positions in a DNA triplet are regarded as the
random variables X, Z, and Y, respectively. In addition to the CMI between
X and Y conditioned on Z computed according to Eq. (2), for each possible
triplet XZY , the quantity P (x, y, z) log2 P (z)P (x,y,z)
P (x,z)P (y,z) was calculated and used as
a feature. Hence, in total, 65 CMI-based features were obtained.
Entropy (E). The Shannon entropy [42] quantifies the average information
content of the gene sequence from the distribution of symbols. The Shannon
entropy is defined as
HN = − Ps(N ) (i) log2 Ps(N ) (i), (3)
i
where Ps (i) is the probability of the ith word of block size N . Shannon
(N )
entropies of the DNA sequences of the genes were determined for block sizes
of 2 (base pairs) and 3 (triplets).
Similarly, we used Gibbs’ entropy to measure the thermodynamic stability
of the genes. Gibbs’ entropy is defined as
SG = −kB PGN (i) ln PGN (i), (4)
i
Computational Identification of Essential Genes 233
where PGN (i) is the probability to be in the ith state and kB is the Boltzmann
constant (1.38 × 10−23 J/K). Gibbs’ entropy is similar to Shannon’s entropy
except for the Boltzmann constant. Nevertheless, the probability distribution is
associated with the thermodynamic stability, which is quantified by the nearest-
neighbor free energy parameters. The probability distribution, PG (i), is modeled
by the Boltzmann distribution given by
E(i)
−
ni e kB T
PGN (i) = E(j)
. (5)
−
nj e kB T
j
T is the temperature in Kelvin and ni is the frequency of the ith word of block
size N . E(i) is the free-energy according to [43]. In [43], the free-energies for
base pairs at 37 ◦ C are provided. For larger block sizes, the energies can be
determined by adding the energies of neighboring di-nucleotides. Shannon and
Gibbs entropies for block size of 2 and 3 were calculated.
In addition to Shannon and Gibbs entropies, the relative entropy is computed
and used as a feature. The relative entropy, also known as the Kullback-Leibler
divergence (KLD) [44] measures the divergence between a probability distribu-
tion P (x) and a model distribution Q(x), and it is calculated as
P (x)
KLD = P (x) log2 . (6)
i
Q(x)
Markov (M). First, assuming that both EGs and NEGs were generated by two
separate Markov chains, the correct Markov chain orders, using EGs and NEGs
in the training data set, were estimated. Then, two Markov chains (M C+ (mE)
and M C− (mN )) of the estimated orders (mE and mN ) were constructed. After
that, the features are obtained by scoring every gene using the constructed
Markov chains.
Numerous Markov chain order estimators have been put forth in the litera-
ture. We have assessed the performances of selected estimators [45–49] on DNA
sequence data and the estimator proposed by Papapetrou and Kugiumtzis [50]
was chosen. The order estimation is based on CMI given in Eq. (2). A Markov
chain of order L has the following property.
For any m ≤ L, since the nth and (n − m)th nucleotides are dependent, the
CMI between the two conditioned on the m−1 bases in the middle will be greater
than zero. Conversely, for any m > L, two nucleotides are independent and
234 D. Nigatu and W. Henkel
the CMI will be zero. Using this observation, Papapetrou and Kugiumtzis have
proposed both parametric and non-parametric significance testing procedures
[50,51]. Compared to other approximations, the results of the gamma distribu-
tion was better [51]. Hence, we used the gamma distribution based parametric
ˆ
approach for estimating the orders. In a symbol sequence of length N, I(X; Y |Z),
the estimate of the CMI, is approximated by the gamma distribution as
ˆ |Z| 1
I(X; Y |Z) ≈ Γ ( (|X| − 1)(|Y | − 1), ). (8)
2 N ln 2
The gamma distribution is used as the distribution of the null hypothesis, H0 :
CM I(m) = 0. Since CM I ≥ 0 always holds, one-sided parameter testing is
performed. Thus, the p-value is computed from the complementary cumulative
distribution of the gamma distribution in Eq. 8. H0 is rejected if the p-value is
less than the nominal significance level (α = 0.05). Starting from order zero, the
null hypothesis is checked and if it is rejected, the next order is checked and
the process continues until the null hypothesis is accepted. Using the correct
estimated Markov orders, the Markov models are obtained by estimating the
transition probabilities from the training sequences.
After the two Markov chains are constructed, the Markov features are com-
puted by scoring the every gene sequences with the two Markov chains. If we
represent the sequence as b1 , b2 , b3 , . . . , bL , the score for an order m is calculated
as
L−m
P (bi+m |bi bi+1 . . . bi+m−1 )
Score = P (bi bi+1 . . . bi+m ) log2 ( ). (9)
i=1
P (bi+m )
The score measures how likely the sequence is generated by the given mE-th
and mN -th order Markov chain. The scores of the gene sequence on the Markov
chains M C+ (mE) and M C− (mN ) were used as features. For intra-organism
predictions, the Markov orders were estimated from the training sets whereas
for cross-organismic gene essentiality predictions, order estimation increased the
computational complexity without improving the result. Hence, we decided to
use a fixed order Markov chain. After experimenting with orders 1 up to 6, order
1 (i.e., mE = mN = 1) was selected.
alone. The results of MI and Markov features were also satisfactory while the
entropy and the newly incorporated non-IT features yielded a relatively small
prediction accuracy.
Fig. 1. EG predictions in E. coli. The estimated Markov orders were 5 for both EGs
and NEGs.
Similarly, the prediction results of our proposed method applied to the bacte-
ria MP are presented in Fig. 2. Using the complete feature set, an AUC score of
0.82 was obtained. Taken separately, all feature groups provided a score greater
than 0.72. The result shows that the added non-IT features (Stop + Len + GC)
also have the ability to distinguish between essential and non-essential genes with
a decent accuracy (0.72). A much higher AUC score of 0.90 was achieved for the
bacteria BS (Fig. 3). Both MI and CMI features achieved a score of 0.87, while
the entropy and Stop + Len + GC features yielded 0.79 and 0.76, respectively.
Compared to our previous results using the SVM algorithm [34], on average
a 4% improvement was obtained, which is due to the additional features. Ning
et al. [14] performed a five-fold cross-validation on EC and MP using sequence
composition features [14]. Our method improved the results slightly from 0.82
to 0.86 in EC and from 0.74 to 0.82 in MP.
Fig. 2. EG predictions in M. pulmonis. The estimated Markov orders were 6 for both
EGs and NEGs.
Fig. 3. The average ROC curves of B. subtilis EG prediction. The estimated Markov
orders were 5 for EGs and 4 for NEGs.
models trained with BS and EC. The average AUC score of both models tested
on 9 bacteria was 0.86, which is slightly better (4%) compared to our method.
However, due to the high computational effort for obtaining the homology- and
domain-related features, our method could provide an alternative with a little
penalty on the accuracy. We performed an extensive pairwise prediction produc-
ing a 15×15 AUC matrix. Similar large-scale pairwise predictions were presented
by two research groups, Cheng et al. [28] on 21 species and Liu et al. [22] on
31 species. To compare the performances of the three methods, we plotted the
distribution of the pairwise AUC scores of the 15 common species in Fig. 5. Our
predictor is significantly better than Liu et al.’s, while Cheng et al.’s predictor
is slightly better than ours.
0.95
0.90
0.85
0.80
0.75
AUC
0.70
0.65
0.60 BS model EC model
0.55
0.50
SA EC SE SA2 AB ST FN MT HI PA HP VC MP MG BS
Fig. 5. A comparison between pairwise prediction results of our method and two exist-
ing methods, proposed by Cheng et al. [28] and Liu et al. [22]. The diamond markers
show the mean values.
Fig. 6. Leave-one-species-out prediction results using RF and SVM [34]. For compari-
son, results of three other methods are presented [21, 22, 25]
Methanococcus maripaludis (MM) is the only archaeon whose EGs and NEGs
are available in DEG, generated experimentally by Sarimento et al. [53]. We
trained the RF classifier using 80% of the genes, which are randomly selected,
and predicted the remaining 20%. The ROC curves of the five feature groups
along with their combination are shown in Fig. 7. Using all features, an average
AUC score of 0.73 was obtained, which is good but smaller than the values
obtained for most of the bacteria. This can be due to the reduced quality of
240 D. Nigatu and W. Henkel
the data. Sarimento et al. could not confidently specify weather 419 genes are
essential or non-essential [53]. Hence, most of these genes are regarded as NEGs.
We further predicted the EGs of MM using the known essential and non-
essential genes of the bacteria EC and BS. The achieved AUC scores were not
satisfactory, 0.59 using EC and 0.64 using BS. The decline in performance is
expected because of the inherent differences in the genetic makeup between
bacteria and archaea.
essential in one data set and non-essential in the other, OGEE adopts a third
category named conditionally essential genes (CEG). A specific gene was covered
by up to 11 data sets and if all the studies do not agree on the essentiality, it is
labeled as CEG.
Out of 21,529 genes, 182 are EGs, 6,985 are CEGs, and 14,362 are NEGs. To
categorize the CEGs as essential or non-essential, we adopted a simple majority
voting scheme. That is, a CEG is regarded as EG if it is essential in a majority
of cell lines. This resulted in 1,632 EGs. We trained RF classifiers for each of the
five feature sets and their combinations. The data was split into 80% for training
and 20% for testing and 100 trials with different sets were performed. The ROC
curves are presented in Fig. 8. Using all the available features, a decent AUC of
0.76 was obtained. Similar to the prokaryotic EG prediction, the mutual informa-
tion features provided the largest contribution. However, the newly added non-IT
features were better than the entropy and Markov features. Guo et al. [26] pre-
dicted human EGs utilizing only intrinsic sequence information and obtained
an AUC score of 0.89. However, the approach used by Guo et al. is a little bit
different. They prepared the positive data set based on a majority decision on
essentiality in 11 cell lines. A gene is considered essential if it essential in more
than 6 cell lines. Otherwise, the gene is totally discarded rather than taking it as
a negative sample. Afterwards, they obtained 1,516 EGs and 10,499 NEGs. Con-
sidering that even the CRISPR-based experimental method proposed by Wang
et al. [12] yielded a 0.78 AUC score when validated using known EGs of a yeast
genome, our results are good.
cell) lines. Among the 13781 tested genes, only 267 were found to be EGs. That
is the smallest reported percentage of EGs among eukaryotic species. The pre-
diction results are presented in Fig. 9. The combined features yielded a very high
AUC score (0.87) and the performances of the individual feature groups are also
satisfactory.
were performed by training classifiers using DRO, CEL, and HSA data sets.
Prediction of CEL using models trained on DRO and HSA yielded an average
AUC score of 0.72 and 0.73, respectively. However, our method failed to predict
human EGs using both DRO and CEL models. Prediction of DRO EGs using
a classifier trained on HSA was also not possible (AUC = 0.47), while an AUC
score of 0.68 was obtained using CEL. All in all, cross-organismic EG predictions
in eukaryotes were not as successful as it was in bacteria.
4 Conclusion
We performed prediction of EGs in the three domains of life. The proposed
machine learning based predictor applies information-theoretic measures to
the DNA sequences of the genes and use them as features along with other
244 D. Nigatu and W. Henkel
References
1. Koonin, E.V.: How many genes can make a cell: the minimal-gene-set concept 1.
Annu. Rev. Genomics Hum. Genet. 1(1), 99–116 (2000)
2. Itaya, M.: An estimation of minimal genome size required for life. FEBS Lett.
362(3), 257–260 (1995)
3. Hutchison, C.A., Chuang, R.Y., Noskov, V.N., Assad-Garcia, N., Deerinck, T.J.,
Ellisman, M.H., Gill, J., Kannan, K., Karas, B.J., Ma, L., et al.: Design and syn-
thesis of a minimal bacterial genome. Science 351(6280), aad6253 (2016)
Computational Identification of Essential Genes 245
22. Liu, X., Wang, B.J., Xu, L., Tang, H.L., Xu, G.Q.: Selection of key sequence-based
features for prediction of essential genes in 31 diverse bacterial species. PLoS ONE
12(3), e0174638 (2017)
23. Deng, J., Deng, L., Su, S., Zhang, M., Lin, X., Wei, L., Minai, A.A., Hassett, D.J.,
Lu, L.J.: Investigating the predictability of essential genes across distantly related
organisms using an integrative approach. Nucleic Acids Res. 39(3), 795–807 (2011)
24. Li, Y., Lv, Y., Li, X., Xiao, W., Li, C.: Sequence comparison and essential gene
identification with new inter-nucleotide distance sequences. J. Theor. Biol. 418,
84–93 (2017)
25. Wei, W., Ning, L.W., Ye, Y.N., Guo, F.B.: Geptop: a gene essentiality prediction
tool for sequenced bacterial genomes based on orthology and phylogeny. PLoS
ONE 8(8), e72343 (2013)
26. Guo, F.B., Dong, C., Hua, H.L., Liu, S., Luo, H., Zhang, H.W., Jin, Y.T., Zhang,
K.Y.: Accurate prediction of human essential genes using only nucleotide compo-
sition and association information. Bioinformatics 33(12), 1758–1764 (2017)
27. Sharp, P.M., Li, W.H.: The Codon Adaptation Index-a measure of directional
synonymous codon usage bias, and its potential applications. Nucleic Acids Res.
15(3), 1281–1295 (1987)
28. Cheng, J., Wu, W., Zhang, Y., Li, X., Jiang, X., Wei, G., Tao, S.: A new compu-
tational strategy for predicting essential genes. BMC Genom. 14(1), 910 (2013)
29. Chen, Y., Xu, D.: Understanding protein dispensability through machine-learning
analysis of high-throughput data. Bioinformatics 21(5), 575–581 (2005)
30. Seringhaus, M., Paccanaro, A., Borneman, A., Snyder, M., Gerstein, M.: Predicting
essential genes in fungal genomes. Genome Res. 16(9), 1126–1135 (2006)
31. Yuan, Y., Xu, Y., Xu, J., Ball, R.L., Liang, H.: Predicting the lethal phenotype
of the knockout mouse by integrating comprehensive genomic data. Bioinformatics
28(9), 1246–1252 (2012)
32. Lloyd, J.P., Seddon, A.E., Moghe, G.D., Simenc, M.C., Shiu, S.H.: Characteristics
of plant essential genes allow for within-and between-species prediction of lethal
mutant phenotypes. Plant Cell 27(8), 2133–2147 (2015)
33. Guo, F.B., Ou, H.Y., Zhang, C.T.: ZCURVE: a new system for recognizing protein-
coding genes in bacterial and archaeal genomes. Nucleic Acids Res. 31(6), 1780–
1789 (2003)
34. Nigatu, D., Henkel, W.: Prediction of essential genes based on machine learning
and information theoretic features. In: Proceedings of BIOSTEC 2017 - BIOIN-
FORMATICS, pp. 81–92 (2017)
35. Nigatu, D., Henkel, W., Sobetzko, P., Muskhelishvili, G.: Relationship between
digital information and thermodynamic stability in bacterial genomes. EURASIP
J. Bioinf. Syst. Biol. 2016(1), 1 (2016)
36. Bauer, M., Schuster, S.M., Sayood, K.: The average mutual information profile as
a genomic signature. BMC Bioinf. 9(1), 1 (2008)
37. Date, S.V., Marcotte, E.M.: Discovery of uncharacterized cellular systems by
genome-wide analysis of functional linkages. Nat. Biotechnol. 21(9), 1055–1062
(2003)
38. Hagenauer, J., Dawy, Z., Göbel, B., Hanus, P., Mueller, J.: Genomic analysis using
methods from information theory. In: Information Theory Workshop, pp. 55–59.
IEEE (2004)
39. Luo, H., Lin, Y., Gao, F., Zhang, C.T., Zhang, R.: DEG 10, an update of the
database of essential genes that includes both protein-coding genes and noncoding
genomic elements. Nucleic Acids Res. 42(D1), D574–D580 (2014)
Computational Identification of Essential Genes 247
40. Chen, W.H., Minguez, P., Lercher, M.J., Bork, P.: OGEE: an online gene essen-
tiality database. Nucleic Acids Res. 40(D1), D901–D906 (2011)
41. Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley, Hoboken
(2012)
42. Shannon, C.: A mathematical theory of communication. Bell System Technical
Journal 27, 379–423, 623–656 (1948). Mathematical Reviews (MathSciNet): MR10,
133e
43. SantaLucia, J.: A unified view of polymer, dumbbell, and oligonucleotide DNA
nearest-neighbor thermodynamics. Proc. Natl. Acad. Sci. 95(4), 1460–1465 (1998)
44. Kullback, S., Leibler, R.A.: On information and sufficiency. Ann. Math. Stat. 22(1),
79–86 (1951)
45. Tong, H.: Determination of the order of a Markov chain by Akaike’s information
criterion. J. Appl. Probab. 12, 488–497 (1975)
46. Katz, R.W.: On some criteria for estimating the order of a Markov chain. Techno-
metrics 23(3), 243–249 (1981)
47. Peres, Y., Shields, P.: Two new Markov order estimators. ArXiv Mathematics e-
prints, June 2005
48. Dalevi, D., Dubhashi, D.: The Peres-Shields order estimator for fixed and variable
length Markov models with applications to DNA sequence similarity. In: Casa-
dio, R., Myers, G. (eds.) WABI 2005. LNCS, vol. 3692, pp. 291–302. Springer,
Heidelberg (2005). https://doi.org/10.1007/11557067 24
49. Menéndez, M., Pardo, L., Pardo, M., Zografos, K.: Testing the order of Markov
dependence in DNA sequences. Methodol. Comput. Appl. Probab. 13(1), 59–74
(2011)
50. Papapetrou, M., Kugiumtzis, D.: Markov chain order estimation with conditional
mutual information. Physica A: Stat. Mech. Appl. 392(7), 1593–1601 (2013)
51. Papapetrou, M., Kugiumtzis, D.: Markov chain order estimation with parametric
significance tests of conditional mutual information. Simul. Model. Pract. Theory
61, 1–13 (2016)
52. Berthold, M.R., et al.: KNIME: the Konstanz information miner. In: Preisach,
C., Burkhardt, H., Schmidt-Thieme, L., Decker, R. (eds.) GfKL 2007. Studies in
Classification, Data Analysis, and Knowledge Organization. Springer, Heidelberg
(2007). https://doi.org/10.1007/978-3-540-78246-9 38
53. Sarmiento, F., Mrázek, J., Whitman, W.B.: Genome-scale analysis of gene function
in the hydrogenotrophic methanogenic archaeon Methanococcus maripaludis. Proc.
Natl. Acad. Sci. 110(12), 4726–4731 (2013)
54. Fraser, A.: Essential human genes. Cell Syst. 1(6), 381–382 (2015)
55. Boone, C., Andrews, B.J.: The indispensable genome. Science 350(6264), 1028–
1029 (2015)
56. Dickinson, M.E., Flenniken, A.M., Ji, X., Teboul, L., Wong, M.D., White, J.K.,
Meehan, T.F., Weninger, W.J., Westerberg, H., Adissu, H., et al.: High-throughput
discovery of novel developmental phenotypes. Nature 537(7621), 508 (2016)
A Heuristic for the Live Parsimony
Problem
1 Introduction
Two main approaches for solving the problem of reconstructing the evolutionary
history of taxonomic objects and their relations with common ancestors are
present in the literature: distance-based and character-based.
In both cases a tree whose leaves represent the taxonomic objects and whose
internal nodes represent hypothetical ancestors must be constructed. In distance-
based methods the problem is to build a tree whose distances are equal to the
distances in an input distance matrix. In character-based methods the problem
is to build a tree that reflects the changes of character states during evolution
course, provided that the state of each character of the leaves are given as input.
Character-based phylogeny is the focus of this work. We investigate phy-
logeny reconstruction based on parsimony, where one tries to minimize the total
number of character state changes along the edges of the tree [5,15]. There are
two major problems related to parsimony: the Large Parsimony Problem (LPP),
where a tree must be build and labels must be assigned to the nodes, and the
Small Parsimony Problem (SPP), where the tree is given as input and the task
is assigning labels to the nodes. The later is easier to solve.
An extended theory, called Live Phylogeny, was defined in [16]. Live phy-
logeny generalizes traditional phylogeny by admitting the presence of living
ancestors among the input objects, called live internal nodes. Live phylogeny
suits well for sets of fast-evolving objects, like viruses [2,8], and may be used
to model the relationship of non-biological ones, such as documents, images or
relational database entries [3,13]. Figure 1 shows a phylogenetic tree with two
live internal nodes.
Fig. 1. Live phylogeny with two live internal nodes. Internal nodes are depicted with
double lines.
In a previous work [9] we introduced new versions of the large and of the
small parsimony problems called them Large Live Parsimony Problem (LLPP)
and Small Live Parsimony Problem (SLPP), and showed that LLPP is NP-
complete. These new problems generalize their traditional counterparts allowing
the existence of live internal nodes in the trees. We started our exploration of live
parsimony problems focusing on SLPP first, because it is easier and because it
may useful to solve LLPP in the same fashion that SPP is used by many heuris-
tics to solve LPP [17]. Also, while envisioning a branch-and-bound algorithm for
the LLPP, the need for a solver of SLPP became clear.
This article extends our prior work [9] as follows:
2 Parsimony Problems
The Large Parsimony Problem (LPP) takes n objects as input, each one labeled
by a string of m symbols s1 . . . sm where sj represents the state of character j,
and a symmetric score function δ(a, b) that expresses the cost of changing any
character from state a to state b. Each state of a character is a symbol from
a set S, with |S| = k. The output is a binary rooted tree T where the leaves
are the input objects, and each internal node v is labeled with a string of m
symbols v1 . . . vm
, mhaving symbol vj representing
the state of character j, such
that d(v, w) = j=1 δ(vj , wj ) and S(T ) = (v,w)∈T d(v, w) is minimum. The
distance d(v, w) between adjacent nodes v, w expresses the cost of changes that
occurred between them. The minimum score of a phylogeny is called minimum
parsimony score. This problem has been proved NP-hard [7].
If the tree topology is given, the problem is easier. This version is called Small
Parsimony Problem (SPP) and consists in labeling the internal nodes minimizing
the total score [11].
st (v) = min{si (u) + δ(i, t)} + min{sj (w) + δ(j, t)}, (1)
i∈S j∈S
The initialization for the algorithm consists in setting, for each leaf v and
each state t, st (v) = 0 if v is labeled with t, or st (v) = ∞ otherwise. At each
step, after computing st (u) and st (w) for every state t, st (v) may be calculated
also for every state t. At the end of the algorithm, the minimum parsimony score
is given by mint∈S st (root).
As in any dynamic programming algorithm, an additional backtracking step
is necessary in order to reconstruct an optimal assignment of labels. This is done
by tracking the choices made at nodes. The whole algorithm takes O(nk 2 ) time.
Even before Sankoff [14], Fitch [6] solved SPP for a binary score function using
a similar approach. Fitch’s Algorithm uses an auxiliary structure Sv , the set of
possible values for label v. The algorithm has three steps and takes O(nk) time.
Fitch’s Algorithm firstly initializes Sv for all leaves. If v is a leaf labeled with
t, Sv = {t}. Secondly, it does a post-order traversal on the tree to calculate Sv
for the internal nodes. If v is a internal node with u and w as children, Sv can
be calculated as follows [9]:
Su ∪ Sw if Su ∩ Sw = ∅,
Sv = (2)
Su ∩ Sw if Su ∩ Sw = ∅.
Finally, it does a pre-order tree traversal labeling the nodes. The root is
labeled with any element in Sroot . Then, every node w with parent v is labeled
as follows [9]:
label (v) if label (v) ∈ Sw ,
label (w) = (3)
any element of Sw if label(v) ∈ / Sw .
It is important to notice that Eq. 3 works even if w is a leaf, since Sw contains
only one element.
In this section we describe solutions for SLPP for binary and general functions,
based on Sankoff’s and Fitch’s solutions. These solutions were originally defined
in [9].
For both types of score function, we basically need to change how internal
nodes representing objects, named live internal nodes, are dealt with. Like a leaf,
a live internal node has its label already defined by the input and it cannot be
changed.
assigning, for each leaf v and each live internal node, zero to st (v) if v is labeled
with t, or ∞ otherwise. Figure 2 shows an example of the initialization, taking
the score function δ showed in Table 1. Gray rectangles show the values of sA (v),
sC (v), sG (v), sT (v) for each leaf or live internal node v.
Table 1. Function δ used in the examples, with characters from S ={A, C, G, T} [9].
δ A C G T
A 0 2 1 2
C 2 0 2 1
G 1 2 0 2
T 2 1 2 0
Fig. 2. First step of the algorithm for SLPP: evaluation of st (v) for leaves and live
internal nodes in a tree with one live internal node [9].
Although the label of v does not change, st̄ (v) needs to be calculated in order
to make sure that the minimum parsimony score for the subtree rooted by v is
correctly computed, provided that t̄ is the label of v. We can see this calculation
in Fig. 3. The arrow shows the direction of calculation (from leaves to root).
Recovering the best choices and to label the nodes may be done exactly as
before, since if the root is a live internal node with label t̄ then st (root) = ∞
for each t = t̄, and st̄ (root) = ∞. Thus, as in Sankoff’s algorithm, the minimum
parsimony score is given by mint∈S st (root).
A Heuristic for the Live Parsimony Problem 253
Fig. 3. Second step of the algorithm for SLPP: evaluation of st (v) for internal nodes
in a tree with one live internal node [9].
We can see an example of this step in Fig. 4. In the figure, the large arrow
shows the direction of labeling (from root to leaves) and small arrows show, at
each internal node, the characters of children that minimizes operations in Eq. 1.
The algorithm correctness is as follows. Let T be a binary rooted tree with
a labeling of all live internal nodes and leaves, and let δ(a, b) be the symmetric
function defining the cost of changing any character from state a to state b. By
induction on the height h of T , if h = 0 then T has only one labeled live internal
node r. After initialization, st̄ (r) = 0, and st (r) = ∞ for each t = t̄, where t̄ is
the label of r. The minimum possible value that can be reached by st̄ (r) is zero.
Fig. 4. Final labeling of nodes after the algorithm for SLPP on a tree with a single
node [9].
As an induction hypothesis, assume that for each tree rooted by node r with
height less than h, each state t is such that st (r) = ∞ or is the minimum
parsimony, and there exists at least one t such that st (r) = ∞. Now, let T be a
tree rooted by r with height h > 0. Let u, w be the children of r. For each state
254 R. Güths et al.
In order to solve SLPP efficiently with a binary score function, two modifications
were made to Fitch’s algorithm. In the initialization step the algorithm sets
Sv = {t} for each leaf and for each live internal node v labeled with t. Figure 5
shows the tree after this first step of the algorithm.
Fig. 5. First step of the algorithm for SLPP with binary score function: evaluating sets
for leaves and live internal nodes on a tree with one live internal node [9].
During the post-order traversal of the tree we cannot apply Eq. 2 for live
internal nodes because in SLPP they already have their corresponding sets with
a single letter representing their labels. Figure 6 illustrates this step.
Fig. 6. Second step of algorithm for SLPP with binary score function: final evaluation
of sets on a tree with one live internal node [9].
A Heuristic for the Live Parsimony Problem 255
In the final step, the pre-order traversal is the same as in Fitch’s Algorithm,
since Eq. 3 works well for live internal nodes. Figure 7 shows the tree after the
final step.
Fig. 7. Final step of algorithm for SLPP with binary score function: labeling the nodes
of a tree with one live internal node.
The correctness of this algorithm follows easily from the traversals in the tree.
In the top-down phase, if the label of a node was computed from the intersection
of its children sets, then both children are labeled with the same state assigned
to their father, resulting in cost zero for both edges. Otherwise, the label was
computed from their union, and since the intersection of the children sets was
empty, there is an edge of cost zero and the other edge has cost one, which is
the minimum number of changes.
Fig. 9. A tree that cannot be reached when the order of species inclusion is 0, 1, 2, 3, 4 [9].
Fig. 10. A central tree after including nodes 0, 1, 2, and the possible trees after includ-
ing node 3.
In order to guarantee all that every possible tree is generated, the proposed
solution shown in Algorithm 1 is generating all possible orders for including
nodes and executing the previous strategy for each one. Obviously we need to
pay an extra computational time, as we will see later.
A Heuristic for the Live Parsimony Problem 257
Algorithm 1 works with three nested loops: the outer loop generates all pos-
sible orders of inclusion of the species and stores them in a n-position array OS.
The second loop generates all possible orders of construction of trees, controlling
at which edge or node each species will be included (this order is stored in an
n-position array OC). The inner loop repeats the following sequence of steps,
controlled by i, to complete the tree or abandon the current construction:
– traverse the partial tree setting numbers for the edges and internal nodes.
– insert species OS[i] breaking the edge OC[i] or at the internal node indicated
by OC[i] (making it live).
– test whether the partial tree can generate an optimal solution. If it can’t,
interrupt the loop, indicating i as the position where the current search for
the optimal tree failed and requesting the next OC position. If it can, we
invoke the polynomial-time solution of SLPP to calculate the parsimony score
of the partial tree and then test against the best score so far.
258 R. Güths et al.
Except for the outer loop and of course for the possibility of creating live
internal nodes, the algorithm works on the same way as the branch-and-bound
proposed in [10].
Algorithm 1 uses an array OC to control tree construction. Given a partial
tree T with i − 1 species already inserted in it, the algorithm traverses T setting
numbers for edges and also for internal nodes. OC[i] indicates the position where
the i-th species will be inserted. To deal with the inclusion of species as live
internal nodes, we use values larger than the number of edges to flag during
the construction order of the partial tree. For i ≥ 2, OC has values larger
than the number of edges in the partial tree, meaning that OC[i] ≥ 2(i − 1).
So, if OC[i] < 2(i − 1), then the i-th species will be inserted by breaking the
edge numbered OC[i], otherwise the species will be placed in the internal node
indicated by OC[i] − 2(i − 1).
To illustrate the operation of arrays OC and OS consider, for example,
species 0, 1, 2, 3, 4 included in this order. Figure 11 presents the resulting trees
when applying the construction orders 00105 (Fig. 11a) and 00123 (Fig. 11b). In
these examples no live internal node was created because OC[i] < 2(i − 1), ∀i.
Fig. 11. Resulting trees when applying construction orders (a) 00105 and (b) 00123.
Figure 12 illustrates the construction of trees with live nodes for species
0, 1, 2, 3, 4 inserted in this order and applying the construction orders 00152 and
00126. In the first order, species 3 is positioned in the internal node numbered 1
(= 5 − 4 = OC[3] − 2(3 − 1)). In the second order, the last species is positioned
at the root (numbered 0).
To illustrate the branch, we will use the input matrix M shown in Fig. 13
with six species and two characters. By using the inclusion order 0, 1, . . . , 5 and
construction order 000000, we obtain the tree shown in Fig. 14.
As another example, the central tree shown in Fig. 15 is obtained using the
inclusion order of species 0, 1, 2, 3 and construction order 001. The figure shows
all possible inclusions of species 3.
A Heuristic for the Live Parsimony Problem 259
Fig. 12. Resulting trees when applying construction orders (a) 00152 and (b) 00126.
1 2
0 A A
1 T T
2 C G
3 A C
4 G A
5 C T
Fig. 13. Input matrix M with species 0, 1, . . . , 5 and two characters 1 and 2 with states
A, C, G and T [9].
Fig. 14. Initial tree with score 7, obtained by inclusion order 0, 1, . . . , 5 and construc-
tion order 000000 [9].
Figure 16 shows a most parsimonious tree with minimal score for M obtained
by the proposed branch-and-bound. Note that a most parsimonious tree obtained
by branch-and-bound may be equal to the tree obtained by traditional branch-
and-bound, or have the same score.
To see that all trees with l ≥ 0 live internal nodes can be generated, consider
T a tree with l live internal nodes. Because the traditional branch-and-bound
is embedded in our Algorithm, all trees with l = 0 live internal nodes can be
generated. Otherwise, if l > 0, it is sufficient to consider an order of species
such that the l species that are live in T are the latest in the inclusion order
of species. All construction orders are analyzed, in particular those orders that
generate a partial tree T equal to T , except for the labeling of the l live nodes.
260 R. Güths et al.
Among those trees, the algorithm will process the one that promotes the l last
species in the order to live internal nodes, exactly the same way that they are
placed in T .
As pointed out by [10], the running time of traditional branch-and-bound for
n species and m characters is O(mnn ), although this time is not reached in most
cases with biological sequence data. Because we included another external loop
that generates all permutations of species, our branch-and-bound running time
is now O(n!mnn ).
Fig. 15. Partial trees to be built and tested by the inclusion of species 3 [9]: (a) as a
leaf and (b) as a live internal node.
Fig. 16. Most parsimonious tree for M with score 6, obtained by branch-and-bound.
construction orders with l > 0 live internal nodes. This is done by changing
N extOC() in a such way that only construction orders with at least one live
internal node are returned. Figure 17 shows a most parsimonious tree obtained
by branch-and-bound with this restriction for matrix M (Fig. 13). Of course,
there is no guarantee that the most parsimonious live phylogeny found using
this approach will have the overall best parsimony score.
Fig. 17. Tree with live internal nodes and parsimony score 6.
Fig. 18. Partial trees to be built and tested by the inclusion of species 2 by the Live
Sequential Addition heuristic.
In fact, the execution of the heuristic allows the generation of trees with live
internal nodes, but does not guarantee that the tree generated at the end has
at least one live internal node. Then we will make one more addition to the
heuristic. In each step, the best partial tree with live internal nodes obtained
after the insertion of the i-th species will be stored in T Livei .
Let’s see how this will work. At each step T1 , the best tree with live inter-
nal node generated in the current step, will be stored temporarily. It will also
temporarily store T2 , the best tree generated from T Livei−1 by inserting Ei the
i-th species. Note that since T Livei−1 already has a live internal node, T2 will
have a live internal node not matter how inserting Ei . To finalize the step, we
compare the parsimony score of T1 and T2 saving the lowest in T Livei .
Fig. 19. Tree obtained for the 20 strains of Zika virus of Table 2 [12]. Node numbering
follows the last column of the table.
264 R. Güths et al.
Dataset
We used the 20 Zika virus strain genomic sequences, aligned using MUSCLE [4],
resulting in 10,109 columns. After aligning, we used GBLOCKS [1] to remove
non-informative characters, resulting in a shorter alignment with 2,169 columns.
7.1 Results
We run our heuristic taking as input the 20 × 2169 matrix (see Test 3 in this
section). Three tests have been executed. In Test 1 we compared the branch-
and-bound to the Live Sequential Addition heuristic. In Test 2 we compared the
branch-and-bound in its plain and only with live phylogenies versions. Finally,
in Test 3 we run our Live Sequential Addition heuristic using the whole input
matrix.
The input matrix is large enough to make our branch-and-bound method
computationally prohibitive. So, we built a set of 100 submatrices to be used in
tests 1 and 2. These submatrices have been obtained by selecting 10 strains (with
at least one representative strain of each clade shown in Table 2) and randomly
picking 500 columns.
Test 1
This test evaluates the performance of the heuristic in searching a most parsimo-
nious tree. In 50 cases out of the 100 10 × 500 submatrices, the heuristic worked
very well, finding a tree with the same score as that one obtained by branch-
and-bound, that is, the heuristic returned a most parsimonious tree. Overall, the
heuristic obtained scores close to those of branch-and-bound ones, with average
error of 0.19% and a maximum error of 2.07%.
It is interesting to note that in 24 cases the heuristic produced a tree con-
taining at least one live internal node. In 9 of those, the tree returned by the
heuristic was also most parsimonious.
Test 2
Fig. 20. Most parsimonious trees for a case. Figure 20(a) was built by branch-and-
bound and Fig. 20(b) was built by branch-and-bound with only live phylogenies. Both
have the same parsimony score 609.
Test 3
In Test 3 we used all 20 strains of Zika virus and all 2, 169 columns on our
heuristic. Figure 21 shows the resulting live phylogeny. The score of the tree
obtained by Lanciotti and colleagues [12] is 3, 233. Our heuristic obtained a score
of 3, 235, but with a live internal node. Note that all three clades suggested in
[12] were preserved in our tree.
Fig. 21. Tree obtained by Sequential Addition heuristic for the 20 species of Zika virus.
266 R. Güths et al.
8 Conclusion
In this article we extended the work presented in BIOSTEC’2017 [9], where we
introduced the Large Live Parsimony and the Small Live Parsimony problems.
These problems are an important step in the characterization of the problems
in the family of Live Phylogeny.
As expected, the live small problem is easy, while the live large problem is
hard. In this article we refined the branch-and-bound that was introduced earlier
adding the possibility of constraining the trees to those that contain at least a
live internal, broadening the possibilities of applying branch-and-bound to some
instances of the problem, for instance to construct benchmarks to test future
heuristics for LLPP.
In this article we also introduced a greedy heuristic for the large live problem,
which was applied to datasets of Zika virus genomes. The results, when compared
to those obtained by branch-and-bound, were good for these datasets.
The live parsimony problems still need further investigations. As mentioned
in the introduction, they have a broad range of applications where there are
typically large datasets to process. Better pruning and heuristics are probably
welcome future research results.
References
1. Cartresana, J.: Selection of conserved blocks from multiple alignments for their use
in phylogenetic analysis. Mol. Biol. Evol. 17, 540–552 (2000)
2. Castro-Nallar, E., Perez-Losada, M., Burton, G., Crandall, K.: The evolution of
HIV: inferences using phylogenetics. Mol. Phylog. Evol. 62, 777–792 (2012)
3. Cuadros, A.M., Paulovich, F.V., Minghim, R., Telles, G.P.: Point placement by
phylogenetic trees and its application to visual analysis of document collections.
In: Proceedings of the 2007 IEEE Symposium on Visual Analytics Science and
Technology, pp. 99–106 (2007)
4. Edgar, R.C.: Muscle: multiple sequence alignment with high accuracy and high
throughput. Nucleic Acids Res. 32(5), 1792–1797 (2004)
5. Felsenstein, J.: Inferring Phylogenies. Sinauer Associates, Sunderland (2004)
6. Fitch, W.M.: Toward defining the course of evolution: minimum change for a spe-
cific tree topology. Syst. Zool. 20, 406–416 (1971)
7. Goëffon, A., Richer, J., Hao, J.: Heuristic methods for phylogenetic reconstruction
with maximum parsimony, pp. 579–597. Wiley (2011)
8. Gojobori, T., Moriyama, E.N., Kimura, M.: Molecular clock of viral evolution, and
the neutral theory. P. Natl. Acad. Sci. 87(24), 10015–10018 (1990)
9. Güths, R., Telles, G.P., Walter, M.E.M.T., Almeida, N.F.: A branch and bound
for the large live parsimony problem. In: Proceedings of the 10th International
Joint Conference on Biomedical Engineering Systems and Technologies (BIOSTEC
2017), pp. 184–189. SCITEPRESS - Science and Technology Publications, Lda
(2017)
A Heuristic for the Live Parsimony Problem 267
10. Hendy, M.D., Penny, D.: Branch and bound algorithms to determine minimal evo-
lutionary trees. Math. Biosci. 59(2), 277–290 (1982)
11. Jones, N.C., Pevzner, P.A., Pevzner, P.: An Introduction to Bioinformatics Algo-
rithms, vol. 2004. MIT Press, Cambridge (2004)
12. Lanciotti, R., Lambert, A., Holodniy, M., Saavedra, S., Signor, L.: Phylogeny of
Zika virus in western hemisphere, 2015. Emerg. Infect. Dis. 22(5), 933–935 (2016)
13. Paiva, J., Florian, L., Pedrini, H., Telles, G., Minghim, R.: Improved similarity
trees and their application to visual data classification. IEEE Trans. Vis. Comp.
Graph. 17(12), 2459–2468 (2011)
14. Sankoff, D.: Minimal mutation trees of sequences. SIAM J. Appl. Math. 28(1),
35–42 (1975)
15. Setubal, J.C., Meidanis, J.: Introduction to Molecular Computational Biology, vol.
1997. PWS, Boston (1997)
16. Telles, G.P., Almeida, N.F., Minghim, R., Walter, M.E.M.T.: Live phylogeny. J.
Comput. Biol. 20(1), 30–37 (2013)
17. Yan, M., Bader, D.A.: Fast character optimization in parsimony phylogeny recon-
struction. Technical report TR-CS-2003-53. University of New Mexico (2003)
Evaluating Runs of Homozygosity in Exome
Sequencing Data - Utility in Disease Inheritance
Model Selection and Variant Filtering
1 Introduction
which is usually a prerequisite for gene discovery. Population history and geographical
events, such as genetic bottlenecks (i.e. a sharp reduction in the size of a population),
and cultural factors such as marriage between close biological relatives (related as
second cousins or closer), increases the probability of the offspring inheriting two
deleterious copies of a recessive gene. These at-risk kindreds represent a particular
context where the incidence of autosomal recessive disorders and its transmission to
offspring is considerably higher [9]. Consanguinity is not evenly distributed worldwide.
The global prevalence of consanguineous marriage is about 6,067 million. The com-
munities of North Africa, Western and South Asia, particularly the highly religious
transcontinental region of the middle-east, collectively account for 20–50% of all
consanguineous marriages (http://consang.net/index.php/Summary). In 2013, 44 mar-
riages between related individuals were reported in Portugal, which represents only 0.
1% of total marriages [10]. The affected offspring will have not only two identical
copies of the ancestral allele, but the surrounding DNA segments will also be
homozygous. Thus, the affected individual carries long stretches of DNA segments that
are identical by descent (IBD), i.e. homozygous segments inherited from each parent.
These homozygous DNA stretches are called ‘runs of homozygosity’ (ROH) (Fig. 1).
The ROH length is dependent on the degree of shared parental ancestry and on the
number of generations that share the DNA region. Therefore, longer ROH (measuring
tens of Mb) can be identified in individuals from geographically isolated populations or
belonging to ethnical groups with a higher consanguinity rate, due to cultural or reli-
gious reasons. In contrast, in larger and older populations with reduced consanguinity
Evaluating Runs of Homozygosity in Exome Sequencing Data 271
rates, ROH length is generally much shorter, as homozygous stretches have been
broken down over successive generations by repeated recombination events during
meiosis [11]. Nevertheless, there are small regions of our genome in which the
recombination rates are considerably lower and thus IBD regions are always observed.
2 Homozygosity Mapping
3 Next-Generation Sequencing
During the last decade we have witnessed the development of new alternative strategies
aiming to improve and eventually replace Sanger sequencing, those are collectively
referred to next generation sequencing (NGS). Although initially developed to study
the genome, they also allow the study of the transcriptome (RNA analysis, or RNA-
seq) [19], the study of protein–DNA interactions (known as ChIP-Seq) [20], and the
study of DNA patterns of methylation (or bisulfite-seq) [21]. NGS, enables sequencing
of an entire human genome within a single day, being faster and more cost-effective
than Sanger sequencing. Currently, at least four platforms are available for NGS:
Illumina and Ion Torrent (2nd generation), PacBio and Oxford Nanopore Technologies
(3rd generation). This number is expected to increase as technology improves and new
272 J. Oliveira et al.
equipment is being developed. Although these 2nd generation platforms are completely
different, all follow a quite similar generic workflow. All start with library preparation,
in which nucleic acids (study sample) are the template. It includes the random frag-
mentation of DNA, followed by in vitro ligation of platform-specific adaptors that
allow the clonal amplification of the sequences by PCR. PCR products are spatially
clustered, by different approaches depending on which can be recovered and arrayed.
Finally, the sequencing process itself consists of alternating cycles of enzyme-driven
biochemistry and signal-based data acquisition [8, 22].
As previously mentioned, NGS has enabled great advances in the study of genetic
diseases and gene discovery [6, 23]. This holds true especially in genetically hetero-
geneous conditions where any one of several distinct genes, when mutated, may give
rise to the same phenotype. As hereditary myopathies and primary ciliary dyskinesia
(PCD), in which neither gene-by-gene Sanger sequencing nor linkage analysis are cost-
effective or efficient approaches [24, 25]. Since nearly 85% of the known disease-
causing variants are located in exonic regions or neighboring intronic sequences (donor
or acceptor splice-sites) [26], whole-exome sequencing (WES) became a common
application of NGS. WES includes all exonic regions of *20,000 genes, performing a
comprehensive analysis of the genome while still being manageable in terms of costs
and amount of data generated. This approach is extremely useful for highly heteroge-
neous conditions and even for the identification of novel disease genes. Nevertheless,
unveiling the genetic mechanism of a disease is not straightforward. Genetic and phe-
notypic heterogeneity, in which a disease can be caused or modulated by pathogenic
variants in different genes, is not only confined to complex multigenic diseases but has
also been recognized to occur in rare diseases. This, together with the enormous amount
of novel variants of unknown clinical significance makes WES difficult to interpret and
could lead to inconclusive results. But exome analysis has been shown to be a powerful
tool, enabling the association of unlikely candidate genes with specific diseases. For
instance, Gripp and collaborators identified truncating mutations in the last exon of the
NOTCH3 gene in patients with lateral meningocele syndrome (OMIM 130720, also
known as Lehman syndrome) [27]. This gene was previously known to cause cerebral
arteriopathy with subcortical infarcts and leukoencephalopathy 1 (CADASIL, OMIM
612954), a progressive disease with onset during adulthood, affecting the brain’s small
arterial vessels, and manifested by migraine, strokes, and white matter lesions, with
subsequent dementia in some patients [28]. In spite of this usefulness, WES has still
some limitations to that need addressing [29]. Firstly, there are intrinsic limitations,
since some deleterious variations are located in non-coding regions and therefore not
covered by WES. Sequencing errors are related to poor capture efficiency in difficult
regions (such as GC-rich regions), mechanical and analytical errors, as well as
misalignment of reads in repetitive regions, leading to erroneous results. These limita-
tions make data analysis more difficult and impose the need to validate candidate
causative variants by Sanger sequencing. Moreover, as a considerable number of
variants are obtained (*35,000 to *50,000 depending on the strategy), WES analysis
requires the application of several filtering strategies, depending on the frequency of the
disease, the predicted model of disease inheritance and the impact of the variants. This
filtering step is critical as it will influence data analysis and final results outcome. For
example, variants may be excluded for being synonymous; however, even though these
Evaluating Runs of Homozygosity in Exome Sequencing Data 273
Table 1. Main algorithms and bioinformatics tools used for ROH detection.
Algorithm Software [Ref.] UI OS Input data files Range ROH
size (Mb)
Sliding blocks Homozygosity-Mapper GUI Unix/Linux VCF files + SNP >1.5
[51, 52] web servera genotypes
PLINK CLI Unix/Linux, BED, PED, and FAM [0.5–1.5]; > 1.5
[53] & Mac OS, files
GUI Windows
Sliding-window GERMLINE CLI Unix/Linux PED, MAP and >1.5
[54] Hapmap files
HomSI GUI Unix/Linux, VCF files >1.5
[55] Mac OS,
Windows
Heterogeneous H3M2 CLI Unix/Linux BAF profiles <0.5;
hidden Markov [56] [0.5–1.5]; > 1.5
model
Frequentistic Agile- Genotyper and - GUI Windows SAM + tab-delimited >1.5
genotype VariantMapper text files
assigment [57]
a
http://www.homozygositymapper.org; BAM- Binary Alignment Map; BED- Browser Extensible Data; CLI-
command-line interface; GUI- graphical user interface; OS- operative system; Ref.- references; SAM-
Sequence Alignment Map; SNP- Single Nucleotide Polymorphism; UI- user interface; VCF- Variant Call
Format.
Evaluating Runs of Homozygosity in Exome Sequencing Data 277
Through the use of this function, the algorithm determines the end of each
homozygous block, ignoring single heterozygous genotypes with seven or more
homozygous/unknown genotypes on either side. HomozygosityMapper calculates a
“homozygosity score”, and it is very robust against genotyping errors due to the per-
missivity of the DetectBlockEnd function as mentioned above. Homozygos-
ityMapper is not intended to replace other linkage tools, such a GENEHUNTER, but, as it
is much faster it may be used in combination therewith. The software can rapidly identify
the possible disease regions and then subsequently generate LOD scores and haplotypes
with conventional software. HomozygosityMapper is linked with GeneDistiller [58],
which is a database that includes information from various data sources such as gene-
phenotype associations, gene expression pattern, and protein-protein interactions,
allowing researchers to easily search information for the genes within a candidate
interval, for instance with a high homozygosity score [51, 52]. Therefore, the
HomozygosityMapper can be used as an indicator of the type of inheritance pattern and/or
be used as a filter for the analysis of NGS data.
PLINK [53], GERMLINE [54] and EXome-HOMozygosity [59] are examples in
which a sliding-window algorithm is applied for WES-based ROH detection. In a
sliding window analysis, the statistics are calculated for a small frame of the data. The
window incrementally advances across the region of interest and, at each new position,
the reported statistics are calculated. In this way, chromosomes are scanned by moving
a window of a fixed size along their entire length and variation in genetic markers
across the region of interest can be measured. In practical terms, a sliding window is a
sub-list that runs over an underlying collection of data (Fig. 2). This type of analysis
allows investigation into the way patterns of variation change across a surveyed
genomic segment [60].
PLINK is a user-friendly software tool, designed to facilitate the analysis of whole-
genome data, with an efficient computationally routine analysis and new analyses that
take advantage of whole-genome coverage [53]. It maps disease loci that contain
multiple rare variants in a population-based linkage analysis, computes allele and
genotypes frequencies, provides missing genotype rates, inbreeding, IBS and IBD
statistics for individuals and pairs of individuals, detection of non-Mendelian trans-
mission in family data, does sex checks based on X chromosome SNPs and tests non-
random genotyping failure. The software’s focus is the analysis of genotype/phenotype
data, so there is no support for steps prior to this (e.g. study design and planning,
generating genotype or copy number variation (CNV) calls from raw data). Very large
whole-genome association studies (WGAS) data sets can be analyzed using standard
hardware, and there are no fixed limits on the number of samples or SNPs [53]. The
software uses a sliding-window algorithm to search for ROH along the genome. In this
algorithm, the window moves forward from the 5′ to the 3′ extremity of a DNA
sequence on a SNP-per-SNP basis. Each SNP is weighted as to whether looks ‘ho-
mozygous’ enough (yes/no) (i.e. allowing for a number of heterozygous or missed
calls). Then, for each SNP, the proportion of ‘homozygous’ windows that overlap that
position is calculated. If this proportion is higher than a defined threshold, the SNP is
designated as being in a homozygous state. By default, PLINK consider a ROH only
when containing at least 100 SNPs in a homozygous segment, and of total length
1000 kb, a ROH must have at least one SNP per 50 kb on average and the SNP-to-
SNP distance is never greater than a user-specified threshold (default value = 1,000 kb)
[62]. PLINK is the main WGAS analytic software that can run either as a stand-alone
tool (from the command line or via shell scripting) or in conjunction with gPLINK, a
Java-based graphical user interface. gPLINK also offers a simple project management
framework to track PLINK analyses and facilitates integration with Haploview [61].
The combination of PLINK with gPLINK and Haploview offers users support for the
subsequent visualization, annotation and storage of results.
GERMLINE is a tool designed for genome-wide discovery of IBD segments shared
within large populations (by SNP arrays). It takes as input genotype or haplotype
marker data for individuals (as well as an optional known pedigree) and generates a list
of all pairwise segmental sharing [54]. Compared with similar tools, GERMLINE uses
a novel hashing and extension algorithm, which is a linear-time algorithm (i.e. an
algorithm in which plotting the runtime against the size of its input, a line is obtained)
for identifying short identical genomic “slices” between pairs of individuals, and then
extending the boundaries of these slices to discover long shared segments represen-
tative of IBD. For ROH detection, like the previous tool, it adopts a sliding-window
algorithm that is flexible with respect to several parameters such as window size,
minimum length of the ROH and tolerance for heterozygous mismatches. In contrast
with PLINK, GERMLINE breaks up SNP stretches into non-overlapping windows of a
user-specified length in terms of SNP, and only if several consecutive windows tagged
as homozygous exceed a threshold in terms of physical or genetic distance is the region
then labeled as homozygous [63]. GERMLINE is still significantly faster than similar
IBD algorithms and compared to with PLINK has both a higher sensitivity and a
Evaluating Runs of Homozygosity in Exome Sequencing Data 279
overall lower false-positive rate. GERMLINE can identify shared segments of any
specified length, as well as enable any number of mismatching markers.
EX-HOM (EXome HOMozygosity) demonstrated the possibility of exome
sequencing to combine, in a single step, the identification of all the coding variants of a
genome with the ability to perform homozygosity mapping as a way to limit candidate
gene search to specific chromosomal regions [59]. EX-HOM authors used a sliding-
window algorithm for WES based ROH detection by applying PLINK. This approach
was developed specifically for when a single, small consanguineous family is available
for mapping and identifying the genetic defect underlying a disease phenotype. As it
uses WES data only, it was susceptible to the limitations which affect an exome
sequencing approach, namely specific variants might have been missed due to low
coverage or the lack of identification of large insertions or deletions, which would
therefore have been missed. The efficiency of EX-HOM largely depends on the
capacity to filter out all the variants which are not related to the disorder. It showed that
regions >1 Mb in length overlap substantially with those identified as LOD score peaks
by linkage analysis, leading to the conclusion that the EX-HOM approach can correctly
identify disease-related long homozygous regions [59].
The three bioinformatic tools presented apply the sliding-window algorithm to
identify ROH. This approach is efficient when working with longer ROH, as these may
have few and unevenly spaced SNPs, whereas smaller and isolated ROH need a higher
marker density. Thus, the sliding window algorithm is not as effective when used with
short/medium ROH sizes. Magi et al. proposed a new tool, H3M2, that can detect small
to longer-sized ROH [56]. The algorithm is based on a heterogeneous HMM, that
incorporates distances between consecutive SNPs into transition probabilities matrix, to
distinguish between the homozygosity and the heterozygosity states. Through this
approach, the authors state that it can detect with high sensitivity and specificity ROH
of every genomic sizes and being the main feature of the algorithm the heterogeneity,
making it well-suited for WES data. To measure the homozygous/heterozygous
genotype state of each SNP, this model uses B-allele frequency (BAF), which is
defined as the ratio between B-allele counts (NB, the number of reads that match with
the 1000 Genomes Project alternate allele at specific position) and the total number of
reads mapped to that position (N, the depth of coverage) [56]. H3M2 showed better
performances than GERMLINE and PLINK with the same dataset, it was less sensitive
to parameter specification (which ensures that analysis results are not harshly affected
by the user’s chosen parameter configuration) and was considerably faster, namely in
preparation of input genotype calls, as it only requires BAF profiles instead of NGS
genotype calls [56].
Pippucci and co-workers tested H3M2, PLINK and GERMLINE and reported that
all tools successfully identified the same 18 Mb homozygous region harboring the
CACNA2D2 gene, in which mutations were associated with epileptic encephalopathy.
Although, higher accuracy was shown with H3M2 and PLINK [63]. Furthermore,
H3M2 exhibited the highest accuracy in the detection of short and medium ROH, which
highlights the applicability of WES-based HM in outbred individuals (i.e. non-
consanguineous) [63].
AgileGenotyper/AgileVariantMapper [57] and HomSI [55] are two other examples
of algorithms that can identify homozygous regions from an individual’s exome
280 J. Oliveira et al.
sequence data. AgileGenotyper uses massive parallel sequencing reads to determine the
genotypes at over 0.53 million known SNP identified by the 1000 Genomes Project, all
located within exonic regions (also the flanking introns). Once the files imported into
the software (in SAM format), it collates the sequence reads aligned to each exon and
determines the read depth for each sequence variant at each known polymorphic
position. It will then deduce the position’s genotype, e.g. is called as heterozygous if
25% or more of the reads identify the minor variant base, and compares this to the
possible genotypes as noted by the 1000 Genomes Project. If the genotype cannot be
deduced or does not match the known genotypes, this are scored as “no-calls”. The
genotype data are exported as a single tab-delimited text file, which can then be
analyzed by other programs such as AgileVariantMapper. AgileVariantMapper uses
the total read depth and the read depth of the minor allele to determine the genotype at
each position. These values can be interactively adjusted to determine the optimum
values for defining the homozygous region [57].
In summary, all the presented methods have proven to be extremely valuable for the
identification of ROH from WES data, with the possibility of performing HM mapping
without resorting to SNP arrays. Still, although powerful, there is an error rate asso-
ciated with this strategy that is difficult to estimate, since it is highly dependent on the
exome metrics, sequencing platform and the bioinformatic tool used to infer ROH.
Moreover, they have to take on the technical limitations of WES, namely the inade-
quate coverage of some exonic regions or sequencing errors which may be a source of
false positive and/or false negative calls. Another issue is related with the incomplete
annotation of the human genome, which can affect the accuracy of the mapping and
annotation of variants; for instance, a deep intronic (pathogenic) variant, could indeed
be the causative mutation in a non-annotated exon. With the improvement of NGS
technologies, these limitations will be surpassed and this method will be even more
powerful.
ROH regions are above a predetermined threshold value this would indicate that the
homozygous AR model could contribute to identify the disease-causing variant.
A lower ROH overall score, on the other hand, would indicate that this model is
unsuitable for variant filtering (Fig. 3).
As a preliminary study, we analyzed ROH in WES data of twelve unrelated
patients. In six cases the data was compatible with an autosomal recessive (AR) disease
model and further six cases whose disease model (and disease-causing variants) are still
unknown (UKN), used as test group. Among the AR disease model, we further sub-
divide the patients in two groups: the first where the disease-causing variants were in
homozygous state, thus we called this group “homozygous-AR” (n = 3, example in
Fig. 4A); and the other group where the disease-causing variant was found in com-
pound heterozygosity, known as “heterozygous-AR” (n = 3). To systemize this anal-
ysis, ROH were determined using HomozygosityMapper. In this software, we varied
the block length (40, 60, 80, 100, 150, 200 and 250) and collected several parameters
Fig. 3. Proposed workflow to identify disease-causing variants from WES data, with and
without family history.
282 J. Oliveira et al.
(highest homozygosity score, number of ROH, total size of ROH, average size of ROH,
and the size of the largest ROH). With the generated data we could then evaluate how
each parameter changed according to block length. A clear cut-off between the two
groups (hom-AR and het-AR) for homozygosity score was identified (Fig. 5). In
addition, for cases with longer ROH the block length of 80 was not sufficient and
should be optimized (increased) to determine the highest score accurately (up to a value
of 250 in the samples tested). It was interesting to note that scores above 60 were only
detected in homozygous-AR cases (Fig. 5). We found statistically significant
Evaluating Runs of Homozygosity in Exome Sequencing Data 283
Fig. 5. Correlation between highest homozygosity score and block length and comparison
between heterozygous and homozygous autosomal recessive (AR) groups. Statistical analysis
carried out with IBM SPSS statistics v.24; graphic in GraphPad Prism v.5. There is a clear cut-off
between the two groups.
differences between the two groups using Kruskal–Wallis test (p = 0.0146). Further,
we applied a Mann-Whitney U test, between homozygous-AR and heterozygous-AR
groups which was found to be statistically significant, for the following parameters:
highest score (Z = −3.302, p = 0.001), ROH average size (Z = −5.211, p = 0.000) and
ROH largest size (Z = −3.429, p = 0.001). To further verify if the differences empir-
ically observed in our cases could be statistically supported, we correlated the highest
homozygosity score obtained in HomozygosityMapper with the total size in Mb of the
respective ROH for each case. Next, we applied k-means clustering algorithm, a
method of cluster analysis that allows the partition of n observations into k clusters, in
which each observation is part of the cluster with the nearest mean.
Two different clusters were identified (Fig. 6). The first cluster corresponds to the
heterozygous AR samples (k1) whose values seem to be more homogeneous compared
with second cluster (k2, homozygous-AR) which has more dispersed data points.
Kruskal-Wallis test demonstrated statistically significant differences between the two
study groups (v2(2) = 6.2 and p = 0.044). No statistically significant differences were
found between the UKN group and the other two (homozygous- and heterozygous AR),
showing that it does not form a specific cluster. In the test group (unknown in Fig. 6),
four samples are located in the vicinity of cluster k1 whereas the other two are nearer k2.
It should be noted that the number of cases analyzed are still very low, and these
figures should be increased at least at 50 to 100 fold so that we can draw further
conclusions. Nonetheless, we were able to show that, by computing the homozygosity
highest score and ROH total size, it is possible to infer whether a homozygous or a
heterozygous variant would be the cause of a particular disease.
284 J. Oliveira et al.
The existence of longer stretches of homozygous SNP (ROH) may indeed provide a
dual role: (i) indicate the pathogenic variant zygosity status (homozygous or
heterozygous), and (ii) point towards the location of the disease-causing gene, since is
likely to be located within these longer homozygous stretches. For instance, the case
presented in Fig. 3A, represents a patient diagnosed with PCD; an AR genetic disorder
that leads to anomalies in cilia/flagella structure, which results in chronic respiratory
tract infections and in some patients, abnormalities in the position of internal organs
(situs inversus), and/or infertility [65]. This case exemplifies how two longer ROH
were identified, one in chromosome 1 and other on the 17, with the causative variant
ultimately being found in the gene CCDC103, located in chromosome 17 [64].
Fig. 6. Scatter plot presenting data analysis resorting tok-means clustering algorithm. Two
clusters were identified for two groups: heterozygous AR (k1, green) and homozygous AR (k2,
blue). Data analysis and plot performed with IBM SPSS statistics v.24. (Color figure online)
6 Conclusion
This work reviewed HM as an approach for gene discovery and the identification of
disease-causing variants. The application of HM should be considered, not only in
research but also in a routine clinical genetics setting, in cases where an autosomal
recessive disease is suspected. The currently available algorithms and bioinformatic
tools designed for ROH detection from array-SNP and WES data were also reviewed,
highlighting their main features, advantages and limitations. Selection of appropriate
algorithms should mainly consider the technology used to generate the data, but also
Evaluating Runs of Homozygosity in Exome Sequencing Data 285
specific features of the patient/family under study, such as genetic context, ROH
average size and the number of relatives affected by the same condition. As depicted in
Fig. 3, ROH analysis should be an option to include in the variant filtering strategies, in
cases with familial AR transmission pattern or in sporadic cases.
As for identifying new associations between a gene and a particular phenotype, the
presence of consanguinity in a specific kindred has its particularities; not only is the
disease inheritance model self-evident (even in sporadic cases), but also concentrating
the analysis of variants located in longer ROH regions reduces considerably the ana-
lytical burden. The term ‘consanguineous’ is somewhat ambiguous, as there are dif-
ferent levels, from the first to the 4th degree (e.g., first cousins). So the average
percentage of DNA shared ranges from 50 to 12.5%. It should be noted that these
percentages may influences the number and size of ROH. However, ROH analysis
should not be confined to consanguineous cases as it may be found to be useful for
cases with unknown parental consanguinity, especially if parents are from the same
remote/confined village, where mutation founder effects or inbreeding events might be
an issue.
This paper also shares some preliminary data that show how the analysis of ROH
regions may also be extremely useful for variant filtering and for the selection of a
disease inheritance model. More specifically, by correlating the highest homozygosity
score and total size of ROH obtained with HomozygosityMapper two distinct clusters
were obtained. If our assumption is correct, in the presence of a particular subject
whose WES data clusters in k1 heterozygous variants, this should be verified first as a
possible cause for the phenotype, corresponding to autosomal dominant (de novo) or
compound heterozygous AR disease models. If the patient is a male, we cannot exclude
an X-linked recessive inheritance, thus hemizygous variants should also be inspected.
However, it should be noted that the number of cases analyzed is still limited, and
further data needs to be collect and processed to verify these preliminary results.
In conclusion, the combination of WES and HM allow researchers to identify
candidate loci and underlying genetic defect itself (at the nucleotide level) in a single
step. Nonetheless, there are still several limitations and further bioinformatic devel-
opments are required. Considering the data presented, there are sensitivity issues that
require addressing, especially if the genetic defect is located in a small ROH or if the
pathogenic variant is de novo in an individual born to consanguineous parents. Finally,
we consider that it would be useful to develop a tool that combines variant filtering and
homozygosity mapping, which currently can only be performed individually.
Acknowledgements. The authors acknowledge support from: (i) Fundação para a Ciência e
Tecnologia (FCT) [Grant ref.: PD/BD/105767/2014] (R.P.); (ii) Research grant attributed by
“Fundo para a Investigação e Desenvolvimento do Centro Hospitalar do Porto” [Grant ref.: 336-
13(196-DEFI/285-CES)] (J.O.). The work was also supported by the Institutions of the authors
and in part by UMIB, which is funded by through FCT under the Pest-OE/SAU/UI0215/2014.
The authors would like to thank the clinicians for patient referral.
286 J. Oliveira et al.
References
1. Miko, I.: Gregor mendel and the principles of inheritance. Nat. Educ. 1(1), 134 (2008)
2. Christianson, A., Howson, C.P., Modell, B.: Global report on birth defects: the hidden toll of
dying and disabled children, New York (2006)
3. Lobo, I., Shaw, K.: Discovery and types of genetic linkage. Nat. Educ. 1(1), 139 (2008)
4. Slatkin, M.: Linkage disequilibrium—understanding the evolutionary past and mapping the
medical future. Nat. Rev. Genet. 9(6), 477–485 (2008)
5. Sanger, F., Nicklen, S., Coulson, A.R.: DNA sequencing with chain-terminating inhibitors.
PNAS 74(12), 5463–5467 (1977)
6. Boycott, K.M., et al.: Rare-disease genetics in the era of next-generation sequencing:
discovery to translation. Nat. Rev. Genet. 14(10), 681–691 (2013)
7. Xia, J., et al.: NGS catalog: a database of next generation sequencing studies in humans.
Hum. Mutat. 33(6), E2341–E2355 (2012)
8. Koboldt, D.C., et al.: The next-generation sequencing revolution and its impact on genomics.
Cell 155(1), 27–38 (2013)
9. Bittles, A.H.: Consanguinity and its relevance to clinical genetics. Clin. Genet. 60(2), 89–98
(2001)
10. Instituto Nacional de Estatística: Marriages (Between persons of the opposite sex - No.) by
Place of registration (NUTS - 2002), Sex, Relationship or affinity between the spouses and
Spouse previous marital status; Annual
11. McQuillan, R., et al.: Runs of homozygosity in European populations. Am. J. Hum. Genet.
83(3), 359–372 (2008)
12. Lander, E.S., Botstein, D.: Homozygosity mapping: a way to map human recessive traits
with the DNA of inbred children. Science 236(4808), 1568–1570 (1987)
13. Alkuraya, F.S.: Autozygome decoded. Genet. Med. 12(12), 765–771 (2010)
14. Goodship, J., et al.: Report autozygosity mapping of a Seckel syndrome locus to
chromosome 3q22.1-q24. Am. J. Hum. Genet. 67, 498–503 (2000)
15. Alkuraya, F.S.: Homozygosity mapping: one more tool in the clinical geneticist’s toolbox.
Genet. Med. 12(4), 236–239 (2010)
16. Syvänen, A.-C.: Toward genome-wide SNP genotyping. Nat. Genet. 37(6s), S5 (2005)
17. Gibbs, J.R., Singleton, A.: Application of genome-wide single nucleotide polymorphism
typing: simple association and beyond. PLoS Genet. 2(10), e150 (2006)
18. Evans, D.M., Cardon, L.R.: Guidelines for genotyping in genomewide linkage studies:
single-nucleotide–polymorphism maps versus microsatellite maps. Am. J. Hum. Genet. 75
(4), 687–692 (2004)
19. Wang, Z., Gerstein, M., Snyder, M.: RNA-seq: a revolutionary tool for transcriptomics. Nat.
Rev. Genet. 10(1), 57–63 (2009)
20. Park, P.J.: ChIP-seq: advantages and challenges of a maturing technology. Nat. Rev. Genet.
10(10), 669–680 (2009)
21. Li, Y., Tollefsbol, T.O.: DNA methylation detection: bisulfite genomic sequencing analysis.
In: Tollefsbol, T. (ed.) Methods in Molecular Biology, vol. 791, pp. 11–21. Springer,
Heidelberg (2011). https://doi.org/10.1007/978-1-61779-316-5_2
22. Mardis, E.R.: The impact of next-generation sequencing technology on genetics. Trends
Genet. 24(3), 133–141 (2008)
23. Ng, S.B., et al.: Exome sequencing identifies the cause of a mendelian disorder. Nat. Genet.
42(1), 30–35 (2010)
24. Oliveira, J., et al.: New splicing mutation in the choline kinase beta (CHKB) gene causing a
muscular dystrophy detected by whole-exome sequencing. J. Hum. Genet. 60(6), 305 (2015)
Evaluating Runs of Homozygosity in Exome Sequencing Data 287
25. Pereira, R., et al.: Mutation analysis in patients with total sperm immotility. J. Assist.
Reprod. Genet. 32(6), 893–902 (2015)
26. Antonarakis, S.E., Krawczak, M., Cooper, D.N.: Disease-causing mutations in the human
genome. Eur. J. Pediatr. 159(Suppl), S173–S178 (2000)
27. Gripp, K.W., et al.: Truncating mutations in the last exon of NOTCH3 cause lateral
meningocele syndrome. Am. J. Med. Genet. Part A 167(2), 271–281 (2015)
28. Norton, N., et al.: Genome-wide studies of copy number variation and exome sequencing
identify rare variants in BAG3 as a cause of dilated cardiomyopathy. Am. J. Hum. Genet. 88
(3), 273–282 (2011)
29. Sirmaci, A., et al.: Challenges in whole exome sequencing: an example from hereditary
deafness. PLoS ONE 7(2), e32000 (2012)
30. Sauna, Z.E., Kimchi-Sarfaty, C.: Understanding the contribution of synonymous mutations
to human disease. Nat. Rev. Genet. 12(10), 683–691 (2011)
31. Meienberg, J., et al.: Clinical sequencing: is WGS the better WES? Hum. Genet. 135(3),
359–362 (2016)
32. Belkadi, A., et al.: Whole-genome sequencing is more powerful than whole-exome
sequencing for detecting exome variants. Hum. Genet. 135, 359–362 (2016)
33. Xu, W., et al.: Model-free linkage analysis of a binary trait. Stat. Hum. Genet.: Methods
Protoc. 850, 317–345 (2012)
34. Bailey-Wilson, J.E.: Parametric and nonparametric linkage analysis. In: Encyclopedia of
Life Sciences. Wiley, Chichester (2006)
35. Pulst, S.M., et al.: Genetic linkage analysis. Arch. Neurol. 56(6), 667 (1999)
36. Ott, J.: Estimation of the recombination fraction in human pedigrees: efficient computation of
the likelihood for human linkage studies. Am. J. Hum. Genet. 26(5), 588 (1974)
37. Elston, R.C., Stewart, J.: A general model for the genetic analysis of pedigree data. Hum.
Hereditary 21, 523–542 (1971)
38. Kruglyak, L., et al.: Parametric and nonparametric linkage analysis: a unified multipoint
approach. Am. J. Hum. Genet. 58, 1347–1363 (1996)
39. Ghahramani, Z.: An introduction to hidden Markov models and Bayesian networks. Int.
J. Pattern Recogn. Artif. Intell. 15(1), 9–42 (2001)
40. Goedken, R., et al.: Drawbacks of GENEHUNTER for larger pedigrees: application to panic
disorder. Am. J. Med. Genet. 96(6), 781–783 (2000)
41. Sobel, E., Lange, K.: Descent graphs in pedigree analysis: applications to haplotyping,
location scores, and marker-sharing statistics. Am. J. Hum. Genet. 58(6), 1323–1337 (1996)
42. Geyer, C.: Introduction to Markov chain Monte Carlo. In: Brooks, S., et al. (eds.) Handbook
of Markov Chain Monte Carlo, pp. 3–48. CRC Press, Boca Raton (2011)
43. Romero-Hidalgo, S., et al.: GENEHUNTER versus SimWalk2 in the context of an extended
kindred and a qualitative trait locus. Genetica 123(3), 235–244 (2005)
44. Abecasis, G.R., et al.: Merlin—rapid analysis of dense genetic maps using sparse gene flow
trees. Nat. Genet. 30(1), 97–101 (2002)
45. Dudbridge, F.: A survey of current software for linkage analysis. Hum. Genomics 1(1), 63
(2003)
46. MacCluer, J.W., et al.: Pedigree analysis by computer simulation. Zoo Biol. 5(2), 147–160
(1986)
47. Gudbjartsson, D.F., et al.: Allegro, a new computer program for multipoint linkage analysis.
Nat. Genet. 25(1), 12–13 (2000)
48. Alkuraya, F.S.: The application of next-generation sequencing in the autozygosity mapping
of human recessive diseases. Hum. Genet. 132(11), 1197–1211 (2013)
288 J. Oliveira et al.
49. Krawitz, P.M., et al.: Identity-by-descent filtering of exome sequence data identifies PIGV
mutations in hyperphosphatasia mental retardation syndrome. Nat. Genet. 42(10), 827–829
(2010)
50. Becker, J., et al.: Exome sequencing identifies truncating mutations in human SERPINF1 in
autosomal-recessive osteogenesis imperfecta. Am. J. Hum. Genet. 88(3), 362–371 (2011)
51. Seelow, D., Schuelke, M.: HomozygosityMapper2012—bridging the gap between homozy-
gosity mapping and deep sequencing. Nucleic Acids Res. 40(W1), W516–W520 (2012)
52. Seelow, D., et al.: HomozygosityMapper—an interactive approach to homozygosity
mapping. Nucleic Acids Res. 37(Web Server issue), W593–W599 (2009)
53. Purcell, S., et al.: PLINK: a tool set for whole-genome association and population-based
linkage analyses. Am. J. Hum. Genet. 81(3), 559–575 (2007)
54. Gusev, A., et al.: Whole population, genome-wide mapping of hidden relatedness. Genome
Res. 19(2), 318–326 (2009)
55. Görmez, Z., et al.: HomSI: a homozygous stretch identifier from next-generation sequencing
data. Bioinformatics 30(3), 445–447 (2013)
56. Magi, A., et al.: H3M2: detection of runs of homozygosity from whole-exome sequencing
data. Bioinformatics 30(20), 2852–2859 (2014)
57. Carr, I.M., et al.: Autozygosity mapping with exome sequence data. Hum. Mutat. 34(1), 50–
56 (2013)
58. Seelow, D., et al.: GeneDistiller—distilling candidate genes from linkage intervals.
PLoS ONE 3(12), e3874 (2008)
59. Pippucci, T., et al.: EX-HOM (EXome HOMozygosity): a proof of principle. Hum. Hered.
72(1), 45–53 (2011)
60. Tang, R., et al.: A variable-sized sliding-window approach for genetic association studies via
principal component analysis. Ann. Hum. Genet. 73(Pt 6), 631–637 (2009)
61. Barrett, J.C., et al.: Haploview: analysis and visualization of LD and haplotype maps.
Bioinformatics 21(2), 263–265 (2004)
62. Chang, C.: PLINK: whole genome data analysis toolset-identity by descent. https://www.
cog-genomics.org/plink/1.9/ibd#homozyg
63. Pippucci, T., et al.: Detection of runs of homozygosity from whole exome sequencing data:
state of the art and perspectives for clinical, population and epidemiological studies. Hum.
Hered. 77(1–4), 63–72 (2014)
64. Oliveira, J., et al.: Homozygosity mapping using whole-exome sequencing: a valuable
approach for pathogenic variant identification in genetic diseases. In: Proceedings of the 10th
International Joint Conference on Biomedical Engineering Systems and Technologies,
BIOINFORMATICS, (BIOSTEC 2017), vol. 3, pp. 210–216 (2017)
65. Leigh, M.W., et al.: Clinical and genetic aspects of primary ciliary dyskinesia/Kartagener
syndrome. Genet. Med.: Off. J. Am. Coll. Med. Genet. 11(7), 473–487 (2009)
Virus Disassembly Pathways Predicted
from Geometry and Configuration Energy
Abstract. Virus are supramolecular structures that are responsible for some of
the most significant epidemics around the world. The disassembly of virus
particles, a key event during viral infection is triggered by numerous intracel-
lular factors. The investigation of the mechanisms of protein subunit loss during
viral disassembly has generally been overlooked, in sharp contrast with the
research on the assembly process of virus particles, which has been the focus of
both experimental and theoretical studies. In this work, we address the problem
of predicting the sequence of protein subunit removal from a viral capsid, by
assuming that the order of subunit loss is mainly determined by each capsid’s
structural geometry and configuration energy. We modelled the early stages of
virus disassembly in a sample of 51 icosahedral viruses of class T = 1, pre-
dicting the sequence of removal of up to five subunits. Due to the high sym-
metry of viral structures, we established the geometrical equivalence of subunit
configurations of capsid fragments, decreasing the size of the search space. The
energy of a given configuration was estimated by using heuristic functions of the
number and types of inter-subunit contacts. We found a disassembly pathway
common to a large group of viruses, consisting of the removal of a triangular
trimer. Exceptions to this general pattern include the loss of a pentagon-shaped
pentamer. These results point at specific subunit interactions as putative targets
for novel antiviral drugs developed to interfere with the disassembly process.
1 Introduction
Viruses are intracellular parasites that replicate inside living cells by using its genetic
and protein synthesis machinery to create new copies [1, 2]. Although they can infect
several different organisms, from bacteria and fungi to algae, plants, insects and ver-
tebrates, human viruses have deserved a special attention by researchers all over the
world [3]. Since ancient times, viruses have coexisted and have evolved with humans
(reviewed in [4]). Examples include several strains of herpes, measles and dengue
viruses. More recent examples include the HIV and Ebola [4]. Viruses caused severe
epidemic outbreaks during human history, like the smallpox virus (also known as
variola virus), causing the death of more than 400 thousand people in Europe each year
in the 18th century [5], or the four influenza pandemics, responsible for over 4 million
deaths all over the world [6].
Viral particles are composed by a nucleic acid (DNA or RNA, single or double
stranded) and, in most cases, a protein-based supramolecular structure, the capsid, that
protects the genetic material between the infection phase of the virus life cycle [1, 7, 8].
Round-shaped viruses have a fixed number of proteins surrounding the genetic material
in an icosahedral symmetry [7, 8]. This geometrical feature was first described in 1962
by Caspar and Klug, who developed a scheme to classify the different levels of
icosahedral symmetry by triangulating the icosahedron facets [7]. The Triangulation
number (or T number) represents the number of equilateral triangles that compose a
triangular face of the icosahedron [8].
The assembly of a complete viral capsid from its protein subunits has been addressed
in many studies, both experimentally and theoretically (reviewed in [2, 9]). Viruses’
capsid subunits are held together by non-covalent interactions such as electrostatic salt
bridges, hydrophobic contacts and hydrogen bonds [10–12]. The combined effect of
many of these interactions accounts for the stability of the virus capsid [13].
Virus assembly is spontaneous in vitro under close to physiological conditions [2]
and the resulting capsids are stable for a long period of time. Many of the theoretical
models of capsid assembly agree on the prediction that three bound protomers, pen-
tamers of trimers and capsids that have just a triangular icosahedron face missing are
stable intermediate forms found during the formation of a complete capsid [10, 12, 14–
17]). If capsid disassembly is just the reverse process of capsid assembly, these findings
suggest that the same structures may be found during capsid disassembly.
In sharp contrast to the number of studies addressing the assembly process, there
are very few experimental studies on the pathways of capsid disassembly. Furthermore,
since it is expected that intermediates of this process are transient and very difficult to
detect, theoretical models of capsid disassembly are also difficult to validate experi-
mentally. Moreover, the assembly and disassembly pathways are, most likely, not
entirely symmetric since some viruses undergo maturation steps after assembly com-
pletion, such as proteolysis, cross-linking or conformational changes [13], which are
not expected to occur during disassembly. In an in vitro study of capsid disassembly
using atomic force microscopy (AFM), Castellanos et al. observed the removal of a
triangle of subunits (a face of the icosahedron) when force was applied on capsids of
Minute Virus of Mice, followed sometimes by the loss of an adjacent triangle [18]. In
other cases, a removal of a pentamer of triangles (15 proteins) was observed. These
disassembly pathways follow the predictions of the assembly models [14–17]. Horton
and Lewis stated that, when possible intermediates are surveyed, energy minimums
appear every multiple of three subunits [19]. Ortega-Esteban et al. have shown, also by
AFM, that the first step of disassembly of Human Adenoviruses, starting from both
mature and immature capsids, is the loss of a pentagon of proteins [20].
In this work, we address the problem of predicting the sequence of subunit loss
during the early steps of viral capsid disassembly, the “disassembly pathway”, focusing
on T = 1 icosahedral capsid structures. We assume that the main factors that determine
these pathways are geometry and subunit interaction energy. Due to symmetry, many
Virus Disassembly Pathways Predicted 291
of the fragments obtained from removing protein subunits from a virus capsid are
geometrically equivalent, if the remaining subunits remain in the same place with the
same inter subunit interactions. To establish such equivalence, we applied rigorous
geometrical and combinatorial considerations.
The work is meant as a contribution to overcome the lack of theoretical studies
targeting the disassembly process, with the long-term goal of providing insights which
may guide the development of antiviral drugs designed to interfere with the formation
of intermediate fragments during this process.
2 Methods
Table 1. Groups and PDB IDs of viral capsids analyzed in this work. Viruses marked with † are
Parvoviruses (Family Parvoviridae).
Group PDB IDs
Adeno-Associated Virus† 1LP3, 2G8G, 2QA0, 3J1Q, 3J4P, 3NTT, 3RA2, 3RA4, 3RA8,
3RA9, 3RAA, 3UX1, 4IOV, 4RSO, 5EGC
Avian Birnavirus 1WCD
Bombyx mori Densovirus† 3P0S
Bovine Parvovirus† 4QC8
Canine and Feline 1C8D, 1C8E, 1C8G, 1C8H, 1FPV, 1IJS, 1P5W, 1P5Y, 2CAS,
Panleukopenia Virus† 4DPV, 1C8F
Galleria mellonella 1DNV
Densovirus†
Hepatitis E Virus 2ZTN, 2ZZQ, 3HAG
Human Adenovirus 1X9T, 4AQQ, 4AR2
(HAPD)
Human Parvovirus† 1S58
Penaeus stylirostris 3N7X
Densovirus†
Porcine Circovirus 3JCI, 3R0R
Porcine Parvovirus† 1K3V
Rodent Protoparvovirus† 1MVM, 1Z14, 1Z1C, 2XGK, 4G0R, 4GBT
Satellite Tobacco Mosaic 1A34, 4OQ8, 2BUK, 4BCU
Virus
292 C. A. Piedade et al.
Fig. 1. Graph representation of the Deltoidal Hexecontahedron. Each vertex represents a subunit
of the T = 1 virus capsid. Edges represent geometrical edges [25].
Table 3. Donor and acceptor’s atoms used to calculate the number of hydrogen bonds.
Donor Arg (Ne; Nη1; Nη2), Asn (Nd2), Cys (Sc), Gln (Ne2), His (Nd1; Ne2), Lys (Nf), Ser
(Oc), Thr (Oc1), Trp (Ne1), Tyr (Oη)
Acceptor Asn (Od1), Asp (Od1; Od2), Gln (Oe1), Glu (Oe1; Oe2), His (Nd1; Ne2), Ser (Oc),
Thr (Oc1), Tyr (Oη)
294 C. A. Piedade et al.
vectors representing the configuration of a capsid fragment with a list of numbers that
enumerate all the possible fragments (with n − 1 subunits removed) from which the
fragment may have originated, also following the numbering scheme shown in Fig. 1.
Edges in this graph were assigned a transition energy weight, Ei!j, based on the
variation of the average heuristic energy per protein, as indicated by Eq. 4. In this
equation, Ei represents the energy of fragment i and mi = 60 − ni is the number of
proteins in that fragment.
The Bellman-Ford algorithm [30] was then used on this graph to calculate the
shortest path from the complete capsid to every possible configuration, recording the
five shortest paths. The energy of a path is simply the sum of the transition energies of
the edges in the path.
The final fragments resulting from the minimal energy paths of disassembly for the 51
capsid structures and n from 2 to 5 are shown in Table 4 assuming Heuristic I for
energy calculations. It can be readily observed that most Parvoviruses (marked with a
†) follow the same sequence of disassembly: {1} ! {1, 10} ! {1, 10, 23} ! {1, 2,
10, 23} ! {1, 2, 10, 22, 23}. Fragment {1, 10, 23} represents the loss of proteins
forming a triangle on the capsid structure (Fig. 2A). This triangle is just one face of the
Icosahedron from which the Deltoidal Hexecontahedron is derived (three adjacent
deltoids).
In the mechanical removal of proteins by AFM applied on the Minute Mice Virus
(on our work represented by the group of the Rodent Protoparvovirus) it was observed
that the disassembly process tends to start by the loss of one of those triangular blocks,
followed by the removal of another adjacent triangle [18]. Furthermore, the theoretical
studies of Rapaport et al. revealed the existence of long-lived transient structures with
just one last triangle of proteins missing to form the complete capsid [16, 17]. The same
pattern for models of capsid assembly was showed by Reddy et al. using both indi-
vidual proteins and trimers [14], suggesting that the path for disassembly also follows
the triangle removal.
Going two more steps further in the disassembly of Parvoviruses, most of them end
up with fragment {1, 2, 10, 22, 23} for n = 5 (Table 4). This fragment results from the
removal of a trapezium-like shape group of subunits, centered on the triangle {1, 10,
23}, (Fig. 2C). Supposing this trend continues, it is not hard to see that there is a
296 C. A. Piedade et al.
Table 4. Capsid groups and final configurations for minimal energy paths after the removal of
n subunits. Energy was computed by Heuristic I (Eq. 1). Viruses marked with † are Parvoviruses
(Family Parvoviridae).
Group PDB IDs n
2 3 4 5
Adeno-Associated 1LP, 2G8G, 2QA0, {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Virus† 3J1Q, 3J4P, 3NTT,
3RA2, 3RA4, 3RA8,
3RA9, 3RAA,
3UX1, 4IOV, 4RSO,
5EGC
Avian Birnavirus 1WCD {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Bombyx mori 3P0S {1,2} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Densovirus†
Bovine Parvovirus† 4QC8 {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Canine and Feline 1C8D, 1C8E, 1C8G, {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Panleukopenia Virus† 1C8H, 1FPV, 1IJS,
1P5W, 1P5Y, 2CAS,
4DPV,
1C8F {1,10,23,34}
Galleria mellonella 1DNV {1,2} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Densovirus†
Hepatitis E Virus 2ZTN, 2ZZQ {1,6} {1,2,23} {1,2,6,23} {1,2,6,10,23}
3HAG {1,2,10,23}
Human Adenovirus 1X9T, 4AQQ, 4AR2 {1,2} {1,2,3} {1,2,3,4} {1,2,3,4,5}
(HAPD)
Human Parvovirus† 1S58 {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Penaeus stylirostris 3N7X {1,2} {1,2,23} {1,2,22,23} {1,2,6,10,23}
Densovirus†
Porcine Circovirus 3JCI {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
{1,2,10,23,42}
3R0R {1,2,10,23,42}
Porcine Parvovirus† 1K3V {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Rodent 1MVM, 1Z14, {1,10} {1,10,23} {1,2,10,23} {1,2,10,22,23}
Protoparvovirus† 1Z1C, 2XGK, 4G0R
4GBT
Satellite Tobacco 1A34, 4OQ8 {1,6} {1,2,23} {1,2,6,23} {1,2,6,10,23}
Mosaic Virus 2BUK, 4BCU {1,2,6} {1,2,6,24} {1,2,6,10,24}
chance of removing the proteins around the five-fold axis. These could be the follow-up
steps of the disassembly of these capsids, potentially resulting in the loss of 15 proteins
such as {1, 2, 3, 4, 5, 6, 10, 22, 23, 33, 34, 41, 42, 49, 50}.
Castellanos et al.’s results [18] confirm the predictions of Reddy et al. [14], since
the lowest energy configuration, just before the complete capsid, is missing a pentamer
of triangles (corresponding to such 15-protein), inclusively for Parvoviruses [15].
Virus Disassembly Pathways Predicted 297
Fig. 2. Disassembly stages of Human Parvovirus with PDB ID 1S58. A: complete capsid, B:
fragment {1, 10, 23}, C: fragment {1, 2, 10, 22, 23}.
The removal of four proteins can occur with more than one possibilities. Most
Parvoviruses lose subunits {1, 2, 10, 23}. However, the structure 1C8F of Canine and
Feline Panleukopenia virus is stabilized by the removal of the configuration {1, 10, 23,
34}, maintaining the triangular breach already formed (Table 4). Densoviruses, which
belong to the Parvoviridae family, follow the main trend, except for the loss of two
proteins, where configuration {1, 2} is favored, and in the case of Penaeus stylirostris
Densovirus which follows the pathway {1} ! {1, 2} ! {1, 2, 23} ! {1, 2, 22,
23} ! {1, 2, 6, 10, 23}, (Table 4). Avian Birnavirus follow the same path as the
majority of Parvovirus, as do Porcine Circoviruses, except when it comes to losing five
subunits, in which case the latter family settles on configurations {1, 2, 10, 23, 42} or
{1, 2, 10, 22, 23}, (Table 4).
Hepatitis E Virus capsids loses proteins in the order {1} ! {1, 6} ! {1, 2,
23} ! {1, 2, 6, 23} ! {1, 2, 6, 10, 23} or, alternatively, {1} ! {1, 6} ! {1, 2,
23} ! {1, 2, 10, 23} ! {1, 2, 6, 10, 23}, (Table 4). Although distinguishable from
the Parvovirus, Hepatitis E Virus loses a pentamer that forms a triangular hole on the
capsid structure.
HAPD follows the disassembly path {1} ! {1, 2} ! {1, 2, 3} ! {1, 2, 3,
4} ! {1, 2, 3, 4, 5}, an exception to all the others (Table 4). Human Adenovirus is the
only group undergoing the loss of the subunits around the five-fold axis, the pentamer
{1, 2, 3, 4, 5}, (Fig. 3B).
On Fig. 3 we can observe the structure of an HAPD, which is formed by very
condensed clusters of pentagons, having very few contacts with the 2-fold and 3-fold
proteins. The distance between the pentagonal clusters on the full capsid structure
might make it easier for this set of subunits to be removed, in opposition to creating a
bigger gap by removing, for example, proteins {1, 2, 10, 22, 23}. Our results are
supported by those of Ortega-Esteban et al. which showed, through AFM, a loss of a
pentagon-shaped pentamer of proteins for Human Adenoviruses [20].
Satellite Mosaic Tobacco Virus disassembly can be separated into two distinct
groups, in which the structures 1A34 and 4OQ8 follow the pathway {1} ! {1,
6} ! {1, 2, 23} ! {1, 2, 6, 23} ! {1, 2, 6, 10, 23}, whereas structures 2BUK and
4BCU follow the pathway {1} ! {1, 6} ! {1, 2, 6} ! {1, 2, 6, 24} ! {1, 2, 6, 10,
24}. The first two structures were inferred from crystallized viruses with RNA still
298 C. A. Piedade et al.
Fig. 3. Disassembly stages of Human Adenovirus (HAPD) with PDB ID 4AQQ. A: complete
capsid (the highlighted proteins are the subunits around the five-fold axis), B: fragment {1, 2, 3,
4, 5}.
attached to the virus capsid while the latter structures were not. Since nucleic acids can
assist the assembly of viral capsids [2], we can speculate that, although not accounted
for in our methods, the presence of these extra macromolecules will contribute to
significantly different full-capsid or fragment configuration energies, which will
influence the pathway of disassembly.
We also investigated the effect of considering the other two heuristic functions for
the calculation of configuration energies (Eqs. 2 and 3). Results were very similar to
Heuristic I, with a slightly higher degree of branching. In particular, the same general
trend of the loss of {1, 10, 23} and five proteins was observed in all three heuristics.
Detailed results can be found as supplementary tables in the code repository associated
with this work (see Sect. 2.6).
4 Conclusions
The aims of this study were to investigate whether there was a prevalent disassembly
pathway among all virus capsids and, if not, if there were any pathways specific to
some virus families. It seems clear that the results are in line with the idea that, for
many T = 1 viruses, the disassembly of the capsid proceeds through the loss of three
adjacent proteins. Geometrically, these subunits form one of the triangular faces of the
icosahedron from which the deltoidal hexecontahedron is derived. The capsid fragment
thus obtained still preserves a very high degree of symmetry. Families that follow this
pattern include some Parvoviruses (Adeno-Associated Virus, Bovine Parvovirus,
Human Parvovirus, Porcine Parvovirus, Rodent Protoparvovirus, Canine and Feline
Panleukopenia Virus), Avian Birnavirus, Hepatitis E Virus and Porcine Circovirus. If
our predictions are correct, then we expect to observe in many more viral types the
experimental results obtained by Castellanos et al. with Rodent Protoparvovirus [18].
We can speculate that once this type of structure is formed, a capsid is sufficiently
disassembled to export the genetic material out of the virus.
Virus Disassembly Pathways Predicted 299
Exceptions to the general pattern can be found among the Human Adenoviruses
(HAPD) which have a very particular pathway of disassembly, allowing for the
removal of a pentagons of subunits.
Our 60-subunit model of viral capsids (Deltoidal Hexecontahedron) and a combi-
natorial method of searching for geometrical equivalence based on symmetry and
geometry is generally superior to models based on 20 subunits (Icosahedron) or 12
subunits (Dodecahedron). Although we reached results comparable to those reported in
previous literature, we were also able to investigate cases such as the Human Aden-
oviruses (HAPD type), which do not lose triangular faces of the Icosahedron, but a
pentagon of faces of the Deltoidal Hexecontahedron. Our method is quite sensitive,
being able to detect different structures with or without the presence of a nucleic acid,
such as in the Satellite Mosaic Tobacco Viruses.
Two improvements can be envisaged for this work: one is the increase of the set of
viral structures studied, which could provide more significance to the conclusions about
the prevalence of the patterns found, and the other is the extension of the study above
the removal of five subunits, a follow-up step which would be of considerable more
computational cost.
As a major insight of this work, the prevalence of the removal of the triangular
trimer from a complete capsid suggests that it can be a viable target for interfering with
virus disassembly process and the subsequent infection mechanism. Antiviral drugs can
thus be designed to disrupt the interactions between these specific capsid proteins.
References
1. Poranen, M.M., Daugelavicius, R., Bamford, D.H.: Common principles in viral entry. Ann.
Rev. Microbiol. 56, 521–538 (2002)
2. Mateu, M.G.: Assembly, stability and dynamics of virus capsids. Arch. Biochem. Biophys.
531(1–2), 65–79 (2013)
3. Strauss, J.H., Strauss, E.G.: Overview of viruses and virus infection, Chap. 1. In: Viruses and
Human Disease, 2nd edn, pp. 1–33. Academic Press, London (2008)
4. Strauss, J.H., Strauss, E.G.: Emerging and reemerging viral diseases, Chap. 8. In: Viruses
and Human Disease, 2nd edn, pp. 325–343. Academic Press, London (2008)
5. Babkin, I.V., Babkina, I.N.: The origin of the variola virus. Viruses 7(3), 1100–1112 (2015)
6. Saunders-Hastings, P.R., Krewski, D.: Reviewing the history of pandemic influenza:
understanding patterns of emergence and transmission. Pathogens 5(4), 66 (2016)
7. Caspar, D.L., Klug, A.: Physical principles in the construction of regular viruses. In: Cold
Spring Harbour Symposia on Quantitative Biology, vol. 27, pp. 1–24 (1962)
300 C. A. Piedade et al.
8. Prasad, B.V.V., Schmid, M.F.: Principles of virus structural organization. In: Rossmann, M.,
Rao, V. (eds.) Viral Molecular Machines. AEMB, vol. 726, pp. 17–47. Springer, Boston
(2012)
9. Perlmutter, J.D., Hagan, M.F.: Mechanisms of virus assembly. Ann. Rev. Phys. Chem. 66,
217–239 (2015)
10. Zlotnick, A.: To build a virus capsid. An equilibrium model of the self assembly of
polyhedral protein complexes. J. Mol. Biol. 241(1), 59–67 (1994)
11. Zlotnick, A.: Are weak protein-protein interactions the general rule in capsid assembly?
Virology 315(2), 269–274 (2003)
12. Zlotnick, A., Johnson, J.M., Wingfield, P.W., Stahl, S.J., Endres, D.: A theoretical model
successfully identifies features of hepatitis B virus capsid assembly. Biochemistry 38(44),
14644–14652 (1999)
13. Zlotnick, A., Stray, S.J.: How does your virus grow? Understanding and interfering with
virus assembly. Trends Biotechnol. 21(12), 536–542 (2003)
14. Reddy, V.S., Giesing, H.A., Morton, R.T., Kumar, A., Post, C.B., Brooks, C.L., Johnson, J.
E.: Energetics of quasiequivalence: computational analysis of protein-protein interactions in
icosahedral viruses. Biophys. J. 74(1), 546–558 (1998)
15. Reddy, V.S., Johnson, J.E.: Structure-derived insights into virus assembly. Adv. Virus Res.
64, 45–68 (2005)
16. Rapaport, D.C.: Role of reversibility in viral capsid growth: a paradigm for self-assembly.
Phys. Rev. Lett. 101(18), 186101 (2008)
17. Rapaport, D.C.: Modeling capsid self-assembly: design and analysis. Phys. Biol. 7(4),
045001 (2010)
18. Castellanos, M., Perez, R., Carrillo, P.J., de Pablo, P.J., Mateu, M.G.: Mechanical
disassembly of single virus particles reveals kinetic intermediates predicted by theory.
Biophys. J. 102(11), 2615–2624 (2012)
19. Horton, N., Lewis, M.: Calculation of the free energy of association for protein complexes.
Protein Sci. 1(1), 169–181 (1992)
20. Ortega-Esteban, A., Perez-Berna, A.J., Menendez-Conejero, R., Flint, S.J., San Martin, C.,
de Pablo, P.J.: Monitoring dynamics of human adenovirus disassembly induced by
mechanical fatigue. Sci. Rep. 3, 1434 (2013)
21. Carrillo-Tripp, M., Shepherd, C.M., Borelli, I.A., Venkataraman, S., Lander, G., Natarajan,
P., Johnson, J.E., Brooks, C.L., Reddy, V.S.: VIPERdb2: an enhanced and web API enabled
relational database for structural virology. Nucleic Acids Res. 37(Database issue), D436–
D442 (2009)
22. Berman, H.M., Westbrook, J., Feng, Z., Gilliland, G., Bhat, T.N., Weissig, H., Shindyalov, I.
N., Bourne, P.E.: The protein data bank. Nucleic Acids Res. 28(1), 235–242 (2000)
23. Burnside, W.: On the theory of groups of finite order. Proc. Lond. Math. Soc. s2–7, 1–7
(1909)
24. Vincent, A.: Molecular Symmetry and Group Theory: A Programmed Introduction to
Chemical Applications, 2nd edn. Wiley, Chichester, New York (2001)
25. Piedade, C., Ferreira, A., Cordeiro, C.: How to disassemble a virus capsid - a computational
approach. In: Proceedings of the 10th International Joint Conference on Biomedical
Engineering Systems and Technologies, BIOSTEC 2017, pp. 217–222. SCITEPRESS, Porto
(2017)
26. Jeffrey, G.A.: An Introduction to Hydrogen Bonding. Oxford University Press, Oxford
(1997)
27. Kumar, S., Nussinov, R.: Relationship between ion pair geometries and electrostatic
strengths in proteins. Biophys. J. 83(3), 1595–1612 (2002)
Virus Disassembly Pathways Predicted 301
28. Onofrio, A., Parisi, G., Punzi, G., Todisco, S., Di Noia, M.A., Bossis, F., Turi, A., De Grassi,
A., Pierri, C.L.: Distance-dependent hydrophobic-hydrophobic contacts in protein folding
simulations. Phys. Chem. Chem. Phys. 16(35), 18907–18917 (2014)
29. Atkins, P.W., De Paula, J.: Physical Chemistry for the Life Sciences. Oxford University
Press, Oxford (2006)
30. Cormen, T.H., Leiserson, C.E., Rivest, R.L., Stein, C.: Introduction to Algorithms, 2nd edn.
The MIT Press, Cambridge (2002)
31. Csardi, G., Nepusz, T.: The igraph software package for complex network research.
InterJ. Complex Syst. 1695, 1–9 (2006)
Bio-inspired Systems and Signal
Processing
Towards Swarm Intelligence of Alcoholics
Andrew Schumann(B)
Abstract. I distinguish the swarm behaviour from the social one. The
swarm behaviour is carried out without symbolic interactions, but it
is complex, as well. In this paper, I show that an addictive behaviour
of humans can be considered a kind of swarm behaviour, also. The
risk of predation is a main reason of reducing symbolic interactions in
human group behaviours, but there are possible other reasons like addic-
tion. An addiction increases roles of addictive stimuli (e.g. alcohol, mor-
phine, cocaine, sexual intercourse, gambling, etc.) by their reinforcing
and intrinsically rewarding and we start to deal with a swarm. I show
that the lateral inhibition and lateral activation are two fundamental
patterns in sensing and motoring of swarms. The point is that both pat-
terns allow swarms to occupy several attractants and to avoid several
repellents at once. The swarm behaviour of alcoholics follows the lateral
inhibition and lateral activation, too. In order to formalize this intelli-
gence, I appeal to modal logics K and its modification K’. The logic K is
used to formalize preference relation in the case of lateral inhibition in
distributing people to drink jointly and the logic K’ is used to formalize
preference relation in the case of lateral activation in distributing people
to drink jointly.
1 Introduction
sensing and motoring of swarms. In Sect. 4, some statistical data on 107 peo-
ple addicted to alcohol and questioned by us are collected. In Sect. 5, a modal
logic for simulating lateral activation and lateral inhibition effects in the swarm
behaviour of alcoholics is introduced.
There are many optical illusions which show that there are different modalities in
perceiving signals which change our picture of reality. Let us consider the Müller-
Lyer illusion, see Fig. 1. Traditionally, this illusion was explained as a combina-
tion of two opposing factors: (i) lateral inhibition increasing contrast when two
points are seen closer than the objective display would justify (Fig. 1(a)), and
(ii) lateral activation decreasing contrast when two points are seen too far apart
to be considered currently (Fig. 1(b)). The matter is that simple cells in primary
visual cortex have small receptive fields and respond preferentially to oriented
bars. Then neurons increase or decrease a receptive field of visual cortex as a
whole according to concentrations of stimuli, see [20]. In case of increasing we
see Fig. 1(a), in case of decreasing we see Fig. 1(b).
Fig. 1. The Müller-Lyer illusion. The lines of (a), (b), (c) are of the same length.
Nevertheless, it seems to us that the line with the inward wings (i.e. (a)) is shorter
than line (b), and the line with the outward wings (i.e. (b)) is longer than line (a).
Hence, perceiving the same line lengths is strongly biased by the neural-
computation process in accordance with the stimulus distributions causing lat-
eral inhibition or lateral activation. It is possible to show that the same pat-
terns of perceiving signals can be detected at different levels of behaviours:
from behaviours of unicellular organisms even to swarming, such as to group
behaviours of ant nests. So, in [21] it was shown that the Müller-Lyer illusion
Towards Swarm Intelligence of Alcoholics 309
Fig. 2. An example from experiments performed in [21] to show that the Müller-Lyer
illusion is detected in the swarm behaviour of ants. The authors of [21] located an
attractant (honeydew solution of 50% w/w) on cardboard in the shape of the Müller-
Lyer figure. The figure consisted of a 7.5-cm central shaft with two 3-cm wings pointing
either inward or outward. As a result of this experiment proved statistically, the distri-
bution of ants repeats the Müller-Lyer illusion. It means that the group sensing of ants
repeats two patterns of perceiving performed by our neurons: (i) the first pattern is
under condition of lateral inhibition (a); and (ii) the second pattern is under conditions
of lateral activation (b).
holds for foraging ants, as well, see Fig. 2. It means that their swarm behaviour
embodies lateral activation and lateral inhibition in the group perceptions of
signals. The authors of [21] explain this phenomenon by that each swarm of ants
has the following two main logistic tasks: (i) to build a global route system con-
necting the nest with food sources to monopolize all reachable food sources (it
corresponds to the lateral activation, i.e. to the colony’s ability to discover new
food sources through exploration); (ii) to exploit effectively and efficiently each
310 A. Schumann
Fig. 3. A possible experiment analogous to the experiment carried out in [21] to show
that the Müller-Lyer illusion is detected in the swarm behaviour of Physarum poly-
cephalum. We can locate an attractant (high nutrient concentration) in the shape of
the Müller-Lyer figure. As a result, we can expect that the concentration of plasmodium
repeats the Müller-Lyer illusion.
found food source (it corresponds to the lateral inhibition, i.e. to the colony’s
ability to concentrate on some food sources). And there is an economic balance
which is analogous to the neurophysiological balance that generates the Müller-
Lyer illusion [21]. In this analogy, each ant in a colony corresponds to a neuron
or retinal cell and the behaviour of a swarm of ants corresponds to the behaviour
of a neurological field.
The patterns of lateral activation and lateral inhibition in sensing and motor-
ing can be detected even at the level of unicellular organisms, such as the slime
mould, also called Physarum polycephalum. So, it is shown in [31] that Kanizsa
illusory contours appear in the plasmodium pattern of Physarum polycephalum.
We can show that the Müller-Lyer illusion holds for the plasmodium too, see
Fig. 3. Indeed, we can observe (i) a lateral inhibition when there are two points
of higher concentration of plasmodium with a distance that is shorter than the
distance between two extremal points of the straight line (Fig. 3(a)), and (ii)
Towards Swarm Intelligence of Alcoholics 311
a lateral activation when there are two points of higher concentration of plas-
modium with a distance that is longer than the distance between two extremal
points of the straight line (Fig. 3(b)).
Thus, some optical illusions, such as the Müller-Lyer illusion, are connected
to combining lateral inhibition and lateral activation effects. Therefore, these
illusions are repeated in the swarm behaviour of Physarum polycephalum and
foraging ants.
Some bacteria can be grouped into swarms. For example, Myxococcus xanthus
is “a predatory surface-associated bacterium that moves in large multicellular
groups and secretes digestive enzymes to destroy and consume other bacteria
in the environment” [14]. Sometimes these multicellular groups form biofilms in
which sessile bacteria secrete an extracellular matrix. Now, it is on open ques-
tion how to define all the details for attracting and repelling bacterial swarms
[14]. Nevertheless, there are some unicellular organisms, such as Physarum poly-
cephalum, which have all the basic swarm patterns in their behaviour, but for
them the mechanism of attracting and repelling is studied well.
Each bacterium can be attracted or repelled directly. But there are no bac-
teria which can occupy several attractants simultaneously or which can avoid
several repellents at once. Only swarms can do it. And Physarum polycephalum
312 A. Schumann
Fig. 4. The Müller-Lyer illusion can be detected in the following behaviour of plas-
modium of Physarum polycephalum. Let an inoculum of plasmodium be located at the
centre of a short line. Then the Müller-Lyer illusion holds if (i) at first we locate two
repellents at two ends of the straight line respectively, (ii) then second we locate two
attractants at both ends respectively. In the first case we observe a lateral inhibition.
In the second case we observe a lateral activation. Hence, the distance between active
zones is different for (a) and for (b), although the straight line is of the same length.
can do it, too. Let us consider an example of Fig. 4. On the one hand, the slime
mould can react to both repellents (Fig. 4(a)) to run away from them. On the
other hand, the slime mould can react to both attractants (Fig. 4(b)) to occupy
them. As a consequence, we see a behavioural pattern of the Müller-Lyer illusion:
(i) the distance between two extremal points of plasmodium distribution became
shorter under condition of lateral inhibition in Fig. 4(a), and (ii) the distance
between two extremal points of plasmodium distribution became longer under
condition of lateral activation in Fig. 4(b). To compare, in the case of Fig. 3(a)
the plasmodium behaves under conditions of informational noise, when there
were too much attractants at a closer distance, to perform a lateral inhibition.
Hence, we can assume that lateral inhibition and lateral activation are impor-
tant fundamental mechanisms which are combined for sensing and motoring in
reactions of different swarms. So, swarms can have different biochemistries (dif-
ferent attractants and repellents), but algorithmically they realize the same pat-
terns such as the Müller-Lyer illusion. In the next section, the alcoholic behaviour
will be regarded as a new example of swarming with the same lateral inhibition
and lateral activation in sensing and motoring of groups.
Towards Swarm Intelligence of Alcoholics 313
Most interviewees (68.2%) claim that their need for alcohol is not satisfied
properly, see Table 2. It means that logistics optimizing the drink behaviour is
an actual trouble for them still.
Furthermore, most respondents (77.6%) buy alcohol at different places, see
Table 3. It is correlated to their group behaviour. Joining different small groups
for drinking or meeting the same small group at different places, they buy alcohol
not at the same place. Some respondents live in the country, where there is only
one store. This fact simplifies logistics for them.
Some respondents (19.6%) prefer to drink alone, but sometimes they drink
jointly. Others drink only in small groups (19.6%). So, the majority (60.8%)
prefer to drink jointly, but sometimes they drink alone. See Table 4.
Hence, we have detected that drinking is a form of group behaviour. The gen-
der characteristics of these groups are collected in Table 5. Some women (42.1%
off all women) drink in small groups consisting only of women and some men
(44.3% of all men) drink in small groups consisting only of men. So, 56% of
the respondents drink in mixed-gender groups. In the meanwhile, a sex/gender
behaviour is mainly reduced in these groups.
Table 5. The question whether the group for drinking jointly consists of men or women.
Joining small groups for drinking jointly supposes a kind of solidarity from
participants. So, the companions can buy drinks if the interviewee has not enough
money, see Table 6. For 68.4% of women and for 86.4% of men it is a common
practice.
Table 6. The question whether the companions buy drinks for you sometimes.
On the other hand, joining small groups for drinking jointly has a requirement
to help others; and 73.7% of women and 92% of men buy alcohol for others
sometimes if their companions have not enough money currently, see Table 7.
Table 7. The question whether you buy drinks for your companions sometimes.
According to our survey, all the respondents have affirmed that sometimes or
always they drink in small groups from 3 to 7 people, but the same respondents
can join different small groups in due course. The number of stable friends to
drink commonly is from 2 to 5. The alcohol-addicted people distinguish their
groups from relatives or colleagues and 63% of the respondents think that their
family and job hinder them to drink safely.
Thus, these small groups from 3 to 7 people can be regarded as human swarms
which help their members to drink safely and to logistically optimize the task to
drink. 83.1% of all the respondents have responded that members of the group
can pay for drinks if the respondent does not have money (Table 6); and 88.8%
of all the respondents have claimed that they can buy alcohol for somebody
from the group who does not have money (Table 7). So, we deal with a form of
solidarity in helping to drink.
In the case of involving new members into groups the main reasons are as
follows: they are neighbours or colleagues and they can help (it means, pay).
Entering new groups is possible if a friend/acquaintance has invited to join them
because it is more safe and interesting for the respondent to join the new group
where there is his or her friend. Without an invitation it is impossible to enter
the group.
Groups are very friendly and the only reason to expel somebody from the
group is that (s)he quarrels (in particular, (s)he does not want to pay). 32% of
respondents have noticed that it would be better to expel one member in their
groups.
Only 28% of respondents have stated that in their groups there are leaders.
They are men or women more than 40 years old. The leadership consists in a
support of the group to drink together.
We have discovered that alcoholics form a network consisting of several small
groups. And the task of optimizing common drinks is solved not by a small
group, but by the whole network, i.e. by several groups whose members are
interconnected. The point is that each small group of alcoholics appears and dis-
appears under different conditions, but the network, these alcoholics belong to,
is almost the same. We have studied that small groups of alcohol-addicted people
are not stable and, by exchanging their members, they can fuse or split in the
316 A. Schumann
Table 8. The question why you choose this place to drink (jointly).
◦A := •A,
where • are any performative verbs for expressing a preference relation with a
strong modality: ‘like’, ‘want’, ‘desire’, etc.
Now let us add countable many new one-place sentential connectives •ki to
the language of K:
These •ki A are read as follows: “the k-th utterance of preference relation
uttered by agent i to fulfil an illocutionary act”. The weaker modality ◦ki is
defined thus:
◦ki A := ¬ •ki ¬A.
We assume that •ki and ◦ki satisfy the necessity rule and distribution axiom
as well.
Let us denote the new extension by Ki .
Now let us define in Ki the four basic preference relations as atomic syllogistic
propositions:
ki (S good
LI P ), ki (S bad bad good
LI P ), ki (S LI P ), ki (S LI ). They are defined
as follows.
ki (S good
LI P ) := •ki (S ⇒ P ) (1)
The atomic proposition ki (S good LI P ) means: “for agent i, alternative P is
at least as good as alternative S by the k-th utterance” and it is defined under
conditions of lateral inhibition. In the model of alcohol-addicted swarms it means:
“for the grouping of alcohol-dependent people i, alternative P is at least as good
as alternative S at the k-th utterance under conditions of lateral inhibition”.
Let us define a model M.
Semantic meaning of ki (S goodLI P ):
M |= ki (S good
LI P ) := at the utterance k uttered by i, there exists a data
point A ∈ M such that AS ∈ S and for any A ∈ M, if AS ∈ S, then AP ∈ S
and it is defined under conditions of lateral inhibition.
Semantic meaning of ki (S good LI P ) in alcohol-addicted swarms: there is a
group of alcoholics i at a place A such that places A and S are connected by
exchanging of some members of i and for any place A, if A and S are connected
by exchanging of some members of i, then A and P are connected by exchanging
of some members of i.
ki (S bad
LI P ) := ◦ki (S ∧ P ) (2)
An example of the weak agent: (s)he prefers not to like not-A instead of that
to like A. An example of the strong agent: (s)he prefers to desire A instead of
that to accept A.
Hence, in logic Ki we have the four kinds of atomic syllogistic propositions:
ki (S good
LI P ), ki (S bad bad good
LI P ), ki (S LI P ), ki (S LI P ) for different k, i, S,
and P . All other propositions of Ki are derivable by Boolean combinations of
atomic propositions. Models for these combinations are defined conventionally:
M |= ¬A iff A is false in M;
M |= A ∨ B iff M |= A or M |= B;
M |= A ∧ B iff M |= A and M |= B;
M |= A ⇒ B iff if M |= A, then M |= B.
Proposition 1. Logic Ki is a conservative extension of K.
Proposition 2. In Ki , the conventional square of opposition holds, i.e. there
are the following tautologies:
ki (S good bad
LI P ) ⇒ ki (S LI P );
good
ki (S bad
LI P ) ⇒ ki (S LI P );
ki (S good good
LI P ) ∨ ki (S LI P );
¬(ki (S good good
LI P ) ∧ ki (S LI P ));
ki (S bad bad
LI P ) ∨ ki (S LI P );
¬(ki (S bad bad
LI P ) ∧ ki (S LI P )).
M |= ki (S good
LA P ) := there exists a data point A ∈ M such that AS ∈ S
and for any A ∈ M , AS ∈ S and AP ∈ S and it is defined under conditions of
lateral activation.
Semantic meaning of ki (S good
LA P ) in alcohol-addicted swarms: there is a
string AS and for any place A which is reachable for S and P by exchanging
of members of i, there are strings AS and AP . This means that we have an
occupation of the whole region where the places S and P are located.
ki (S bad
LA P ) := •ki (¬S ∧ ¬P ). (6)
M |= ki (S bad
LA P ) := for any data point A ∈ M , both AS is false and AP
is false and it is defined under conditions of lateral activation.
Semantic meaning of ki (S badLA P ) in alcohol-addicted swarms: for any place
A which is reachable for S and P by exchanging of members of i, there are no
strings AS and AP . This means that the group of alcoholics cannot reach S
from P or P from S immediately.
ki (S bad
LA P ) := ◦ki (S ∨ P ). (7)
M |= A ∧ B iff M |= A and M |= B;
M |= A ⇒ B iff if M |= A, then M |= B.
Proposition 3. Logic Ki is a conservative extension of K .
Proposition 4. In Ki ,
the unconventional square of opposition holds, i.e. there
are the following tautologies:
ki (S good bad
LA P ) ⇒ ki (S LA P );
good
ki (S bad
LA P ) ⇒ ki (S LA P );
ki (S good good
LA P ) ∨ ki (S LA P );
¬(ki (S good good
LA P ) ∧ ki (S LA P );
ki (S bad bad
LA P ) ∨ ki (S LA P );
¬(ki (S bad bad
LA P ) ∧ ki (S LA P )).
324 A. Schumann
Now, let us consider pairs •ki A and •mi A, where different performative verbs
•ki and •mi occur and these verbs belong to different groups of illocutions in
expressing a preference relation, i.e. both cannot be simultaneously representa-
tives, directives, declaratives, expressive, or comissives. For instance, ‘believing’
and ‘knowing’ are both representatives and ‘ordering’ and ‘insisting’ are both
directives. Assume, ‘believing’ be denoted by •ki and ‘advising’ by •mi . Notice
that ‘assuming’ is modally weaker than ‘believing’, and ‘advising’ is modally
weaker than ‘insisting’. So, ‘assuming’ can be denoted by ◦ki , and ‘advising’ can
be denoted by ◦mi , such that •ki A ⇒ ◦ki A and •mi A ⇒ ◦mi A. The construction
•ki A ⇒ •mi ¬A (respectively, ◦mi A ⇒ ◦ki ¬A) fits the situation that a belief
that A is ever stronger than some other illocutions (belonging to other illocution
groups) related to not-A.
Let us distinguish different swarms according to the acceptance of stronger
or weaker modality:
Meditative Agent. (i) agent i prefers •mi ¬A instead of •ki A iff •ki A ⇒ •mi ¬A;
and (ii) agent i prefers ◦ki ¬A instead of ◦mi A iff ◦mi A ⇒ ◦ki ¬A.
Active Agent. (i) agent i prefers •ki A instead of •mi ¬A iff •ki A ⇒ •mi ¬A; (ii)
and agent i prefers ◦mi A instead of ◦ki ¬A iff ◦mi A ⇒ ◦ki ¬A.
(k ∪ m)i∪j (S good
LA (P1 ∧ P2 ))
.
ki (S good
LA P1 ); mj (S good
LA P2 )
The illocutionary logic Ki can express the preference relations of alcoholics
in respect to attractants under the conditions of lateral activation.
6 Conclusions
I have shown that the lateral inhibition and lateral activation are two fundamen-
tal patterns in sensing and motoring of swarms. The point is that both patterns
allow swarms to occupy several attractants and to avoid several repellents at
once. Then I have shown that a habit of joint drinking of alcohol-addicted peo-
ple in small groups can be considered a swarm behaviour controlled by outer
Towards Swarm Intelligence of Alcoholics 325
stimuli (places to drink jointly). And this behaviour follows the lateral inhibi-
tion and lateral activation, also. As a result, the swarms of alcoholics can be
managed by localization of safe places for meeting to drink jointly. Generally,
the logic of lateral inhibition and lateral activation of groups of alcoholics has
the same axioms as the logic of parasite propagation for Schistosomatidae sp.
[26] as well as the same axioms as the logic of slime mould expansion [27]. The
difference is that instead of syllogistics for Schistosomatidae sp. and for slime
mould, where preference relations are simple and express only attractions by
food, we say about performative actions (or verbs), which express a desire to
drink together at a place. For simulating the swarm behaviour of alcoholics I
propose two logics: the logic Ki to formalize a lateral inhibition in distribut-
ing people to drink jointly and the logic Ki to formalize a lateral activation in
distributing people to drink jointly.
Hence, we can state that some forms of human group behaviour are not
social in fact. This unsocial behaviour can proceed as a human version of swarm
behaviour. Many forms of human swarming have recently been studied – from
crowds of people in escape panic [9] to aircraft boarding [22]. However, some
stable patterns of interconnected people have never been analyzed as a swarm. I
have proposed to consider a network of coordinated alcoholics as human swarm-
ing. The reasons are as follows: (i) their behaviour is controlled by replacing
stimuli: attractants (places where they can drink jointly and safely) and repel-
lents (some interruptions which can appear for drinking); this control is executed
by the same algorithms as for standard swarms from social bacteria to eusocial
mammals; (ii) the behaviour of alcoholics is collective and even cooperative,
but it is subordinated to the only one uncontrolled intention, namely, how to
drink; so, this motivation bears no symbolic meanings in the terms of symbolic
interactionism [4] and, then, it cannot be evaluated as social.
References
1. Abrahams, M., Colgan, P.: Risk of predation, hydrodynamic efficiency, and their
influence on school structure. Environ. Biol. Fishes 13(3), 195–202 (1985)
2. Adamatzky, A., Erokhin, V., Grube, M., Schubert, T., Schumann, A.: Physarum
chip project: growing computers from slime mould. Int. J. Unconv. Comput. 8(4),
319–323 (2012)
3. Beni, G., Wang, J.: Swarm intelligence in cellular robotic systems. In: Dario, P.,
Sandini, G., Aebischer, P. (eds.) Robots and Biological Systems: Towards a New
Bionics. NATO ASI Series, pp. 703–712. Springer, Heidelberg (1993). https://doi.
org/10.1007/978-3-642-58069-7 38
4. Blumer, H.: Symbolic Interactionism; Perspective and Method. Prentice-Hall,
Englewood Cliffs (1969)
5. Bull, R.A., Segerberg, K.: Basic modal logic. In: The Handbook of Philosophical
Logic, vol. 2, pp. 1–88. Kluwer (1984)
326 A. Schumann
6. Costerton, J.W., Lewandowski, Z., Caldwell, D.E., Korber, D.R., et al.: Microbial
biofilms. Annu. Rev. Microbiol. 49, 711–745 (1995)
7. Duffy, J.E.: The ecology and evolution of eusociality in sponge-dwelling shrimp.
In: Kikuchi, T. (ed.) Genes, Behavior, and Evolution in Social Insects, pp. 1–38.
University of Hokkaido Press, Sapporo (2002)
8. Helbing, D., Keltsch, J., Molnar, P.: Modelling the evolution of human trail sys-
tems. Nature 388, 47–50 (1997)
9. Helbing, D., Farkas, I., Vicsek, T.: Simulating dynamical features of escape panic.
Nature 407(6803), 487–490 (2000)
10. Jacobs, D.S., et al.: The colony structure and dominance hierarchy of the Dama-
raland mole-rat, Cryptomys damarensis (Rodentia: Bathyergidae) from Namibia.
J. Zool. 224(4), 553–576 (1991)
11. Jarvis, J.: Eusociality in a mammal: cooperative breeding in naked mole-rat
colonies. Science 212(4494), 571–573 (1981)
12. Jarvis, J.U.M., Bennett, N.C.: Eusociality has evolved independently in two genera
of bathyergid mole-rats but occurs in no other subterranean mammal. Behav. Ecol.
Sociobiol. 33(4), 253–360 (1993)
13. Jones, J.D.: Towards lateral inhibition and collective perception in unorganised
non-neural systems. In: Pancerz, K., Zaitseva, E. (eds.) Computational Intelligence,
Medicine and Biology. SCI, vol. 600, pp. 103–122. Springer, Cham (2015). https://
doi.org/10.1007/978-3-319-16844-9 6
14. Kearns, D.B.: A field guide to bacterial swarming motility. Nat. Rev. Microbiol.
8(9), 634–644 (2010)
15. Krützen, M., Mann, J., Heithaus, M.R., Connor, R.C., Bejder, L., Sherwin, W.B.:
Cultural transmission of tool use in bottlenose dolphins. PNAS 102(25), 8939–8943
(2005)
16. Michener, C.D.: Comparative social behavior of bees. Annu. Rev. Entomol. 14,
299–342 (1969)
17. Nalyanya, G., Moore, C.B., Schal, C.: Integration of repellents, attractants, and
insecticides in a “push-pull” strategy for managing German cockroach (Dicty-
optera: Blattellidae) populations. J. Med. Entomol. 37(3), 427–434 (2000)
18. Olson, R.S., Hintze, A., Dyer, F.C., Knoester, D.B., Adami, C.: Predator confusion
is sufficient to evolve swarming behaviour. J. R. Soc. Interface 10(85), 20130305
(2013)
19. Parsons, T.: Social Systems and The Evolution of Action Theory. The Free Press,
New York (1975)
20. Riesenhuber, M., Poggio, T.: Neural mechanisms of object recognition. Curr. Opin.
Neurobiol. 12(2), 162–168 (2002)
21. Sakiyama, T., Gunji, Y.-P.: The Müller-Lyer illusion in ant foraging. In: Hemmi,
J.M. (ed.) PLoS ONE, vol. 8, no. 12, p. e81714 (2013)
22. Schadschneider, A., Klingsch, W., Klpfel, H., Kretz, T., Rogsch, C., Seyfried,
A.: Evacuation dynamics: empirical results, modeling and applications. In: Mey-
ers, R.A. (ed.) Encyclopedia of Complexity and Systems Science, pp. 3142–3176.
Springer, Berlin (2009). https://doi.org/10.1007/978-0-387-30440-3 187
23. Schumann, A.: On two squares of opposition: the Leniewski’s style formalization
of synthetic propositions. Acta Analytica 28, 71–93 (2013)
24. Schumann, A.: From swarm simulations to swarm intelligence. In: 9th EAI Interna-
tional Conference on Bio-inspired Information and Communications Technologies
(formerly BIONETICS). ACM (2015)
25. Schumann, A., Pancerz, K., Szelc, A.: The swarm computing approach to business
intelligence. Studia Humana 4(3), 41–50 (2015)
Towards Swarm Intelligence of Alcoholics 327
26. Schumann, A., Akimova, L.: Syllogistic system for the propagation of parasites.
The Case of Schistosomatidae (Trematoda: Digenea). Stud. Log. Gramm. Rhetor.
40(1), 303–319 (2015)
27. Schumann, A.: Syllogistic versions of go games on physarum. In: Adamatzky, A.
(ed.) Advances in Physarum Machines. ECC, vol. 21, pp. 651–685. Springer, Cham
(2016). https://doi.org/10.1007/978-3-319-26662-6 30
28. Schumann, A., Woleński, J.: Two squares of oppositions and their applications in
pairwise comparisons analysis. Fundamenta Informaticae 144(3–4), 241–254 (2016)
29. Schumann, A., Fris, V.: Swarm intelligence among humans – the case of alco-
holics. In: Proceedings of the 10th International Joint Conference on Biomedical
Engineering Systems and Technologies, BIOSIGNALS, (BIOSTEC 2017), vol. 4.
SCITEPRESS (2017)
30. Skinner, B.F.: About Behaviorism. Random House, Inc., New York (1976)
31. Tani, I., Yamachiyo, M., Shirakawa, T., Gunji, Y.-P.: Kanizsa illusory contours
appearing in the plasmodium pattern of Physarum polycephalum. Front. Cellular
Infect. Microbiol. 4(10), 1–11 (2014)
32. Tsang, N., Macnab, R., Koshland, D.E.: Common mechanism for repellents and
attractants in bacterial chemotaxis. Science 181(4094), 60–63 (1973)
33. Viscido, S., Parrish, J., Grunbaum, D.: Individual behavior and emergent proper-
ties of fish schools: a comparison of observation and theory. Mar. Ecol. Prog. Ser.
273, 239–249 (2004)
34. Whiten, A., Goodall, J., McGrew, W.C., Nishida, T., Reynolds, V., Sugiyama,
Y., Tutin, C.E., Wrangham, R.W., Boesch, C.: Cultures in chimpanzees. Nature
399(6737), 682–685 (1999)
Health Informatics
Technologies for Ageing in Place: A Systematic
Review of Reviews and Meta-analyses
1 Introduction
The normal ageing process involves different types of alterations with repercussions on
motor (e.g. less mobility), cognitive (e.g. attention capacity or short-term memory) or
social (e.g. smaller social networks) capacities. These alterations influence perceived
abilities, such as increasing difficulty of focusing on nearby objects or distinguishing
them in poorly lit environments or even less ability to perform multitasking [1].
Consequently, diverse barriers tend to hamper the daily lives of older people, in par-
ticular about decision-making, problem-solving or day-to-day planning. On the other
hand, ageing is associated with many chronic diseases that influence the quality of
older adults. All these circumstances have implications for the autonomy and inde-
pendence of older adults and can lead to their institutionalization.
The concept of ageing in place [2] intents to satisfy the people’s desire to stay in
their own residential environments as they get older, despite possible declines in their
physical and cognitive capacities. Ageing in significant places where many of the daily
routines have taken place throughout life - such as neighbourhoods, churches, grocery
stores, markets, cafes, libraries, recreational associations or gardens - allows a greater
mediation of the alterations related to the ageing processes, and also the continuity of a
(more or less) active lifestyle and, consequently, of contributions to the community.
Technological solutions emerge as potentially cost-effective to meet the needs of
citizens and to promote the services’ reorganization [3], which are the aims of concepts
such as Medicine 2.0 [4], connected health [5], or holistic health [6, 7]. Moreover,
many of the current health and social policies focus on providing products and services
to compensate the declines of older adults, thus favouring the permanence in their
homes. In this sense, ageing in place [2] reflects the idea that familiar environments
provide a sense of belonging not only to the family, but also to the community, strongly
related to the notion of identity, security, familiarity and autonomy, which are essential
variables for a healthy ageing [8]. This perspective also enforces the role of information
technologies, which may help older adults in their homes (i.e. technologies for ageing
in place [9]) to overcome multiple impairments, and to allow them to live safely,
independently, autonomously and comfortably, with the necessary support services to
their changing needs [2, 10, 11].
There are several systematic reviews of reviews and meta-analyses related to
technologies for ageing in place [12–15]. However, their studies focus on specific
technologies (e.g. short message services [12]), or specific pathologies (e.g. congestive
heart failure [13]). Therefore, the broad analysis of the study reported in the present
article, which was partially presented in [16] (i.e. home monitoring of patients with
chronic diseases), is useful to inform older adults, practitioners and researchers about
the state of the art of technologies for ageing in place.
2 Methods
The systematic review of reviews and meta-analyses reported in this article aimed to
systematize current evidence of technologies for ageing in place and was informed by
the following research questions:
• RQ1: What are the most relevant applications of technologies for ageing in place?
• RQ2: What are the types of technologies for ageing in place?
• RQ3: What are the outcomes of technologies for ageing in place?
Since a large number of articles have been published, the study reported in this
article was planned to include systematic reviews or meta-analyses only. Moreover, it
followed the guidelines of the Preferred Reporting Items for Systematic Reviews and
Meta-Analyses (PRISMA) [17].
Technologies for Ageing in Place: A Systematic Review 333
The literature search was conducted using two general databases (i.e. Web of
Science and Scopus) and two specific databases (i.e. PubMed, a medical sciences
database, and IEEE Explorer, a technological database).
The database queries were prepared to include: (i) all the systematic reviews and
meta-analyses where any of the keywords ‘telecare’, ‘telehealth’, ‘telemedicine’,
‘homecare’, ‘telemonitoring’, ‘home monitoring’, ‘remote monitoring’, ehealth’,
‘telerehabilitation’, ‘mobile health’, ‘mhealth’ or ‘assisted living’ were presented in the
title or abstract; and (ii) all the systematic reviews and meta-analyses where any the
keywords ‘technology-based’, ‘information technology’, ‘information and communi-
cation’, ‘internet-based’, ‘web-based’, ‘on-line’, ‘smartphones’, ‘mobile apps’, ‘mobile
phone’, ‘monitoring devices’ or ‘consumer health information’ were presented in the
title or abstract together with any of the keywords ‘healthcare’, ‘health care’, ‘patient’,
‘chronic disease’, ‘older’ or ‘elderly’.
The search was performed on 30 of April 2016 and intend to include systematic
reviews and meta-analyses published during the preceding 10 years.
(x) total number of participants; (xi) technologies being reported; (xii) target chronic
disease, when applicable; (xiii) primary outcomes; (xiv) secondary outcomes;
(xv) evidence of the impact of the reported applications; (xvi) authors’ interpretations;
(xvii) conclusions; and (xviii) recommendations for the future.
Afterwards, the collected data was analysed to answer the research questions of the
present study.
3 Results
Figure 1 presents the PRISMA flowchart of the systematic review of reviews and
meta-analyses reported in the present paper.
A total of 2681 articles were retrieved from the initial searches on Web of Science
(1263 articles), Scopus (550 articles), PubMed (822 articles), and IEEE Explorer (46
articles). The initial screening yielded 1429 articles by removing the duplicates (1210
Technologies for Ageing in Place: A Systematic Review 335
articles) or the articles without abstracts (42 articles). After exclusions based on title
alone, 563 articles were retrieved. Additionally, 315 articles were eliminated based
upon review of their abstracts.
The full texts of the 248 remaining articles were assessed and 175 articles were
eliminated. Consequently, 73 articles [18–90] were included in this systematic review.
significant percentage of the primary studies are exploratory or descriptive, the review
concludes about several benefits of the social media, including: (i) increased interaction
with others; (ii) more available, shared and tailored information; (ii) increasing
accessibility and widening access to health information; (iv) peer support; and (v) po-
tential influence of healthy policies. However, reliability, confidentiality and privacy
are general concerns when using social media to communicate health issues and the
quality and reliability of the exchanged information is quite diverse [51].
Information technologies also offer a medium to assist healthcare providers to meet
educational-related responsibilities. One review [26] reports 25 RCT that evaluate the
ability of interactive computer based education programs that can be viewed at home or
during periodic clinic visits. Other eight systematic reviews reported the use of different
technologies (e.g. social media or personal health records) for citizens’ education [26,
41, 51, 69, 78, 90], covering a diverse population in terms of age, sex, education level,
health status and healthcare intervention or disease state. The programs present sig-
nificant variations in terms of features, implementations and integration strategies.
Physical Activity. Physical activity is one aspect of lifestyle changing that might be
effective in reducing rates of hospital admission and reducing risk of mortality.
Increasing physical activity to low or moderate levels can result in lower risk of
mortality from all causes.
Eleven of the included systematic reviews and meta-analyses consider the pro-
motion of physical activity among the behaviour change interventions analysed [27, 41,
44, 45, 48, 49, 56, 69, 80, 86, 89]. There are many applications that suggest a range of
physical exercises, complemented with demo videos and the measurement of several
outcomes.
When reported as an outcome, physical activity is being measured through vali-
dated questionnaires (e.g. the Short Questionnaire Assessing Health-enhancing -
SQUASH, the Baecke Physical Activity Questionnaire, or the comprehensive evalu-
ation of the Minnesota Leisure Time Physical Activity Questionnaire), and quantitative
measures [65, 92]. Examples of quantitative measures are steps per day or minutes of
weekly exercise [96] that can be performed by different types of devices [44, 97].
Technological applications appear to have positive effects on physical activity
behaviours [27, 44, 48, 56, 80, 89], but the results are not consistent [27]. For instance,
a review [44] reports that within the 14 RCT being analysed, six studies found that the
intervention group had significant differences compared with the control group, four
studies had mixed results, and another four had no significant differences between
groups [92].
However, the retrieved studies are characterized by small sample sizes, heteroge-
neous effect sizes, and a lack of analysis of the cost-effectiveness of the interventions,
especially in conjunction with clinical practice.
Nutrition and Weight Management. Nine of the included systematic reviews and
meta-analyses [27, 34, 42, 62, 65, 68, 69, 79, 80] deal with nutrition and weight
management. In terms of nutrition several studies reported nutrition applications able to
capture dietary intake [42, 62] that can be divided into applications that allow con-
trolling calories and keep food diary (e.g. applications that allow users to select food
Technologies for Ageing in Place: A Systematic Review 337
and portion size or applications that process food photograph taken by the users), and
specific applications for food safety, namely considering people with allergies.
Overall, positive feedback was reported: the use of nutrition applications resulted in
better self-monitoring adherence and changes in dietary intake when compared to
conventional techniques (e.g. paper records).
Concerning weight management, remote interventions are being delivered using a
wide range of technologies, including web-based applications and smartphone appli-
cations, or even more traditional methods, such as telephone calls [92]. The inter-
ventions might include [92]: (i) counselling or advice; (ii) self-directed or prescribed
exercise; (iii) home based or facility based exercise; and (iv) written education or
motivational support material, namely the promotion of knowledge using serious
games [41].
In terms of outcomes, questionnaires such as the Food Frequency Questionnaire
(FFQ) or the MEDFICTS score have been used [65, 92]. When compared to control
groups, the majority of primary RCT studies of the included systematic reviews and
meta-analyses pointed a significant net difference in weight between the intervention
and control groups. This is valid for different types of technologies (e.g. web-based
applications [42]), but is particularly evident when mobile applications are being used
(e.g. periodic prompts of a smartphone applications [27, 80]).
Long-term, sustainable behaviour change and health benefits are not shown by the
included systematic reviews and meta-analyses because of the lack of consistent
follow-up data collection and reporting.
studies on the effect the interventions and the identification of how the training and
implementation of such interventions should be tailored to maximize their effects.
Diabetes. Of the retrieved articles, 19 dealt with home monitoring of patients with
diabetes [18, 23, 24, 35–37, 40, 47, 52–54, 63, 64, 73, 75, 76, 82, 84, 85].
Furthermore, since self-management of diabetes requires patient adherence to best
practice recommendations (e.g. glycaemia control, dietary management or physical
activity) there has been an interest in increasing compliance of self-care applications:
(i) self-management and care knowledge [35, 40, 47, 52, 54, 76, 88]; (ii) prescribed
medication adherence [37, 72, 76]; and (iii) behaviour outcomes (e.g. diet and healthy
eating or physical activity) [24, 35, 37, 40, 46, 47, 54, 72, 73, 76, 82, 85, 88].
A significant number of articles focuses both type 1 and type 2 diabetes [23, 24, 36,
47, 54, 63, 73, 75, 82, 84]. Other articles focus type 2 diabetes [18, 35, 37, 40, 75, 85].
Only one of the retrieved articles focuses exclusively on type 1 diabetes [64].
The articles of the diabetes domain include primary studies with high quality
scientific evidence. All the retrieved articles targeting diabetes considered RCT primary
Technologies for Ageing in Place: A Systematic Review 339
studies and 11 of them considered RCT as one of the inclusion criteria [23, 24, 36, 37,
52–54, 63, 75, 84, 85]. On the other hand, aggregating all the primary studies it is
evident that the number of the involved participants is relatively significant (e.g. one
article reports the involvement of 3578 patients [52] and other reports the involvement
of 3798 patients [75]).
In technological terms, several articles [35, 40, 47, 52–54, 75, 82, 84] refer
web-based applications. In general, these applications allow synchronous (e.g. instant
messaging or chat) and asynchronous (e.g. electronic mail or bulletin board) com-
munications together with web pages to register clinical parameters (e.g. weight or
blood pressure) and medication.
Besides web-based applications, there are other technological solutions reported in
different articles:
• Computer-assisted applications integrating the management of clinical data with
electronic practice guidelines, reminder systems, and feedback to the patients
[18, 47].
• Smartphones (e.g. standalone smartphones or smartphones integrating specific
devices such as glucometers for automatic glucose level upload) [36, 37, 40, 47, 52,
64, 73, 76, 84].
• Automatic patient data transmission by means of monitoring devices (e.g. devices
to monitor vital signals or devices to monitor behaviour outcomes such as
pedometers or accelerometers connected by wireless communications to monitor
physical activity) [23].
• Video-conference [24, 47].
• Telephone calls [82].
The main outcome of most of the articles included in the diabetes domain is the
control of glycaemia by using glycosylated haemoglobin (Hb1c) as a proxy. However,
in all the studies, this aim is complemented with other health related outcomes (e.g.
health related quality of life [24, 35, 53, 54], weight [35, 52, 73, 75], depression [52],
blood pressure [24, 63, 73, 82], cholesterol level [35, 63], triglycemius level [63], or
fluctuation index [35]), behaviour outcomes (e.g. physical activity) [18, 24, 35, 37, 54,
73, 82, 85], patient self-motivation [84], patient-clinician communication [84], medi-
cation adherence [42], and structural outcomes related to care coordination [23, 24].
Most of the articles of the diabetes domain report moderate to large significant
reduction of Hb1c when compared with usual care [18, 23, 37, 40, 53, 63, 64, 73, 75,
76, 82, 84, 85]. However, several studies are not conclusive about the reduction of
Hb1c [24, 35, 36, 52]. In particular, computer-based diabetes self-management inter-
ventions [52] and consultations supported by video-conference [24] appear to have a
small beneficial effect on glycaemia control.
One article [47] reporting research gaps of the technological approaches identifies
the need to improve the usability of the applications as well the need for more com-
prehensive solutions, including real-time feedback to the patients and integration with
electronic health record systems.
Congestive Heart Failure. The number of RCT and non-RCT primary studies
included in the nine articles dealing with congestive heart failure varies from nine to 42
340 L. Pereira et al.
[19–21, 25, 30, 39, 57, 58, 61]. Most of these systematic reviews (i.e. six systematic
reviews [20, 25, 30, 39, 57, 61]) considered RCT as one of the inclusion criteria.
Concerning the supporting technologies, automatic patient data transmission by
means of monitoring devices [19, 25, 30, 39, 61] is being used together with
video-conference and standard telephone calls [20, 21, 57, 58] to allow the assessment
of symptoms and vital signs, as well as the transmission of automatic alarms.
In terms of clinical outcomes, the main concerns are the impacts of home moni-
toring in heart failure-related hospitalizations and all-cause mortality [57] when com-
pared with usual care. However, several secondary outcomes are also considered such
as self-care behaviour (e.g. adherence to prescribed medication, daily weighing or
adherence to exercise recommendations [39]).
According the reviewed articles, home monitoring has a positive effect on clinical
outcomes of patients with congestive heart failure. Home monitoring reduces mortality
when compared with usual care and it also helps to lower both the number of hospi-
talizations and the use of other healthcare services [25, 30, 57, 61].
However, there is a need for high-quality trials [20]. Additionally, Grustam and
colleagues [58] state that evidence from the scientific literature related to home mon-
itoring to support congestive heart failure patients is still insufficient. Also, more
comprehensive economic analyses are needed to reach a sound conclusion. This means
that further research is required in terms of comparisons of home monitoring with usual
care of patients with congestive heart failure.
Chronic Obstructive Pulmonary Disease. All the four retrieved articles dealing with
chronic obstructive pulmonary disease analyse RCT primary studies [29, 33, 79, 81]. In
particular, three of them considered RCT as one of the inclusion criteria [29, 79, 81].
Home monitoring is supported by commercially available devices to measure and
transmit different types of information (e.g. weight, temperature, blood pressure,
oxygen saturation, spirometry parameters, symptoms, medication usage or steps in
6-min walking distance). In some cases, the automatic data acquisition is comple-
mented by telephone interviews of clinical staff using questionnaires [29, 81], or by
telephone or video-conference calls to provide feedback to the patients [79]. In this
respect, one article also reports that telephone and videoconference calls when used to
provide feedback are associated to improvements in the quality of life of patients with
chronic obstructive pulmonary disease [79].
In what concerns the primary and secondary outcomes, three studies [29, 33, 81]
compare home monitoring with usual care of patients with chronic obstructive pul-
monary disease, considering mortality, admissions to hospital or other healthcare uti-
lization as primary outcomes. Secondary outcomes include, among others, health
related quality of life, patient satisfaction, physical capacity and dyspnea.
Home monitoring was found to reduce rates of hospitalization and emergency
department visits, while the findings related to hospital bed days of care varied between
studies [29, 81]. However, one study reports a greater mortality in a telephone-support
group compared with usual care [29]. Additionally, there is evidence that home mon-
itoring has a positive effect on physical capacity and dyspnea [79] and it is similar or
better than usual care in terms of quality of life and patient satisfaction outcomes [85].
Technologies for Ageing in Place: A Systematic Review 341
Cancer. To mitigate the undesirable side effects that can negatively affect the quality
of life of patients with cancer, their empowerment is important [93]. The patients’
empowerment can contribute to them being autonomous and respected, having
knowledge, or psychosocial and behavioural skills, receiving support from community,
family, and friends [74]. In this respect, technologies for ageing in place can support
knowledge transmission, including electronic survivorship care plans, patient-to-patient
and patient-to-caregiver communication, and electronic patient-reported outcomes [74].
The retrieved studies report that technologies for ageing in place (e.g.
computer-based and web-based applications) were found to be effective in different
health outcomes such as: (i) pain, depression and quality of life [66]; (ii) fatigue,
depression, anxiety, and overall quality of life [77]; (iii) knowledge; (iv) satisfaction;
and (v) other outcomes which were both directly and indirectly related to the healthcare
interventions [22].
However, some of the primary studies included in the retrieved articles present only
preliminary evaluations of the technological solutions being used [22] and the findings
suggest that the application of technologies for ageing in place in cancer care is still at a
very early stage [66].
Other Chronic Conditions. Finally, in terms of other chronic conditions, four articles
were identified [93]: (i) one article considers that remote interventions, including the
use of mobile applications designed for smartphones and tablets, appear to be
well-accepted for the care of patients with asthma and increase self-care behaviours
[50]; (ii) one article related to telerehabilitation for people with multiple sclerosis [67],
including web-based applications and videoconference, highlights the lack of
methodologically robust trials; (iii) one article shows the potential clinical benefits of
dietary interventions based in mobile applications for chronic renal patients, although
there is a need for additional robust trials [70]; and (iv) one article which aims was to
analyse the impact of mobile applications targeting overweight or obese people con-
cluded that when considering exclusively the use of mobile technologies (21 RCT) a
significant reduction in weight (5–10%) was observed [43].
4 Discussion
The following subsections discuss the findings according to the different perspectives
associated to the research questions formulated to inform the study reported in the
present article.
Since physical inactivity and obesity, together with smoking and alcohol con-
sumption, are well-established risk factors with impact in health conditions and mor-
tality, the promotion of healthy lifestyles assumes great importance [95]. In particular,
several articles report the impact of health education and behaviour changing in terms
of physical activity, nutrition and weight management.
One of the systematic reviews and meta-analyses target loneliness and social iso-
lation and other target home safety. Loneliness and social exclusion are generally
understood as dramatic consequences of the population ageing and the response to
emergencies in terms of home safety is generally accepted as an important issue for
older adults.
Concerning healthcare, according to the findings of the systematic review reported
in this article, diabetes, congestive heart failure, chronic obstructive pulmonary disease
and hypertension are the most relevant chronic diseases in terms of the use of tech-
nologies for ageing in place to support home monitoring. Type 1 and type 2 diabetes
stand out from other chronic conditions with a total of 19 articles. In order of relevance,
the second chronic condition was congestive heart failure (nine articles), which was
followed by chronic obstructive pulmonary disease (four articles). Furthermore, one
article reporting a systematic review related to home monitoring of patients with
hypertension was also included in the present systematic review.
In terms of the empowerment of older adults and their informal caregivers, diabetes
also stands out from other chronic conditions with a total of 15 articles. In order of
relevance, seven articles are related to mental health and four articles are related to
patients with cancer. Furthermore, five articles are related to one of the following
chronic conditions: (i) hypertension; (ii) asthma; (iii) multiple sclerosis; (iv) renal
chronic disease; and (v) obesity.
Diabetes requires constant monitoring of the glucose levels, congestive heart failure
has a high rate of hospital readmission [102], and key aspects of the natural history of
the chronic obstructive pulmonary disease are episodes of acute exacerbations [103].
On the other hand, self-management of diabetes requires patient adherence to best
practice recommendations, engagement and participation [63], as well as efficient
contacts between patients and clinical staff are essential when dealing with mental
health conditions, and patients undergoing cancer treatment may experience many
undesirable side effects that can negatively affect their quality of life. Therefore, the
results of the systematic review reported in this article are in line with the current strong
motivation for using technological solutions as a way to monitor older adults with
chronic diseases at home and to promote an increasing compliance of self-care.
Some technologies have gotten more attention than others. Particularly, mobile
applications such as applications to deliver periodic prompts and reminders [27] were
reported by a considerable number of articles. In this respect, the portability of
smartphones enables users to have access 24 h a day, making possible the long-term
management and reinforcement of health behaviours. This seems to promote the focus
of the participants and their adherence to the programs.
Concerning home monitoring of patients with chronic diseases, the results show that
the technological solutions being used include web-based applications, computer-based
applications, smartphones applications, automatic patient data transmission by means of
monitoring devices, video-conference and standard telephone calls. On the other hand,
in terms of the empowerment of older adults and their informal caregivers, the tech-
nological solutions being used include computer-based applications, web-based appli-
cations, smartphone applications, videoconference, telephone calls and serious games.
Despite a high level of technological innovation and implementation, one of the
findings is that telephone calls are still an important channel for the communication
between patients and care providers.
Furthermore, it seems that important aspects are neglected during the technological
developments, since there are reports of usability drawbacks as well as reports of the need
for more comprehensive solutions, including provision of real-time feedback and the
integration of the electronic health records systems being used by the care providers [47].
Therefore, the results show that not only disruptive technological solutions have a
key role, since practical and robust solutions are required, which means that the inte-
gration and the interoperability of existing technologies assume a great importance [104].
• Secondary outcomes are quite diverse and include health related quality of life,
weight, depression, blood pressure, behaviour change, self-management, care
knowledge, medication adherence, patient-clinician communication, or structural
outcomes related to care coordination.
Furthermore, in terms of technologies for ageing in place to promote the empow-
erment of patients with chronic conditions, the systematic reviews and meta-analyses
compare the use of technological solutions with usual care and the outcomes depend on
the type of the patients being considered, namely:
• Glycaemia control, health related quality of life or patient adherence to best practice
recommendations, including care knowledge, behaviour outcomes (e.g. diet and
healthy eating or physical activity) and adherence to prescribed medication, for
patients with diabetes.
• Improvement of the symptoms, engagement and participation for patients suffering
from mental health conditions.
• Knowledge, pain, depression, fatigue, anxiety, satisfaction and overall quality of
life for patients with cancer.
Independently of the outcomes being measured, the retrieved articles show that the
usage of technologies for ageing in place has positive effects with a moderate to large
improvements in different outcomes when compared with conventional practices.
5 Conclusion
Considering the large number of articles reporting studies related to technologies for
ageing in place, the authors decided to perform a systematic review of reviews and
meta-analyses.
Although the authors tried to be as elaborate as possible in methodological terms to
guarantee that the review selection and the data extraction were rigorous, it should be
acknowledged that the study reported in this article has, however, some limitations that
must be considered when interpreting the results. First, the search was not exhaustive
because of the language restriction. Then, only the articles published in the scientific
literature were included with a potential underestimation of the real number of works
and researches related to this topic. Finally, it should be noted as the continuous
evolution of the technology makes very difficult to provide an updated reporting of the
evidence available.
However, it should also be pointed that the study reported in this article presents
some strengths: (i) the methodological design was as elaborate as possible to guarantee
that the review selection and the data extraction were rigorous; and (ii) the systemat-
ically collected evidence contributes to the understanding of the technologies for
ageing in place to support older adults in their home environments.
The systematic reviews and meta-analyses being retrieved pointed out that tech-
nologies for ageing in place might facilitate immediate, intermediate and long-term
health related outcomes in older adults.
346 L. Pereira et al.
One of the problems that emerge from the study reported in the present article is
related to the outcomes being measured. Besides their diversity, it is important to refer
that different measurement methods are being applied. This is an important difficulty
when aggregating and analysing data from different trials, which is essential to achieve
the statistical and clinical significance that is required to promote the adoption of new
services.
On the other hand, it becomes increasingly important that the development tech-
nologies for ageing in place musts consider a variety of possible dangers to older
adults.
One aspect to be considered is the risk of the reduction of the human relations,
particularly between patients and healthcare providers. Therefore, technologies for
ageing in place should not interfere on the quality of the human relations, which might
require the implementation of specific training interventions for all the stakeholders.
These training programs are indispensable especially when it becomes evident that
even highly educated people may experience difficulties in understanding instructions or
other healthcare-related information. Therefore, there are significant problems in terms
of health literacy that can be exacerbated when dealing with older adults. Moreover,
health literacy is part of a complex issue known as ‘digital divide’, which is related to the
existence of a gap between people who can effectively deal with information and
communication tools and those who cannot [51] (e.g. difficulties in accessing specific
services due to non-existent adequate communication infrastructures).
Finally, since technologies for ageing in place manage different types of personal
information, including clinical data, there could be serious problems related to privacy
of older adults. These problems might be exacerbated when using mobile technologies,
such as mobile phones (e.g. when a smartphone is lost and the control or accesses is not
performed).
Despite the potentially negative connotations, given the findings, technologies for
ageing in place present an enormous potential and deserves further research. In par-
ticular, there is the need of additional scientific evidence from RCT and longitudinal
studies with large sample sizes and robust methods.
References
1. Cota, T., Ishitani, L., Vieira, N.: Mobile game design for the elderly: a study with focus on
the motivation to play. Comput. Hum. Behav. 51, 96–105 (2015). https://doi.org/10.1016/j.
chb.2015.04.026
2. Plaweck, H.: Aging in place: the role of housing and social supports. J. Gerontological
Nurs. 18, 53–54 (1990). https://doi.org/10.3928/0098-9134-19921001-14
Technologies for Ageing in Place: A Systematic Review 347
3. Genet, N., Boerma, W.G., Kringos, D.S., Bouman, A., Francke, A.L., Fagerström, C.,
Melchiorre, M.G., Greco, C., Devillé, W.: Home care in Europe: a systematic literature
review. BMC Health Serv. Res. 11, 207 (2011). https://doi.org/10.1186/1472-6963-11-207
4. Eysenbach, G.: Medicine 2.0: social networking, collaboration, participation, apomedia-
tion, and openness. J. Med. Internet Res. 10, e22 (2008). https://doi.org/10.2196/jmir.1030
5. Kvedar, J., Coye, M.J., Everett, W.: Connected health: a review of technologies and
strategies to improve patient care with telemedicine and telehealth. Health Aff. 33, 194–199
(2014). https://doi.org/10.1377/hlthaff.2013.0992
6. Koch, S.: Achieving holistic health for the individual through person-centered collaborative
care supported by informatics. Healthc. Inform. Res. 19, 3–8 (2013). https://doi.org/10.
4258/hir.2013.19.1.3
7. Mori, A.R., Mazzeo, M., Mercurio, G., Verbicaro, R.: Holistic health: predicting our data
future (from inter-operability among systems to co-operability among people). Int. J. Med.
Inform. 82, e14–28 (2013). https://doi.org/10.1016/j.ijmedinf.2012.09.003
8. Wiles, J.L., Leibing, A., Guberman, N., Reeve, J., Allen, R.E.S.: The meaning of “aging in
place” to older people. Gerontologist 52, 357–366 (2011). https://doi.org/10.1093/geront/
gnr098
9. Connelly, K., Laghari, K.U.R., Mokhtari, M., Falk, T.H.: Approaches to understanding the
impact of technologies for aging in place: a mini-review. Gerontology 60, 282–288 (2014).
https://doi.org/10.1159/000355644
10. Nutbeam, D.: Health promotion glossary. Health Promot. Int. 13, 349–364 (1998). https://
doi.org/10.1093/heapro/13.4.349
11. Walker, P.R.: Report from the World NGO Forum on Ageing and the United Nations
Second World Assembly on Ageing. PsycEXTRA Dataset. https://doi.org/10.1037/
e523502012-005
12. Househ, M.: The role of short messaging service in supporting the delivery of healthcare:
an umbrella systematic review. Health Inform. J. 22, 140–150 (2014). https://doi.org/10.
1177/1460458214540908
13. Kitsiou, S., Paré, G., Jaana, M.: Effects of home telemonitoring interventions on patients
with chronic heart failure: an overview of systematic reviews. J. Med. Internet Res. 17, e63
(2015). https://doi.org/10.2196/jmir.4174
14. Mcbain, H., Shipley, M., Newman, S.: The impact of self-monitoring in chronic illness on
healthcare utilisation: a systematic review of reviews. BMC Health Serv. Res. 15, 565
(2015). https://doi.org/10.1186/s12913-015-1221-5
15. Slev, V.N., Mistiaen, P., Pasman, H.R.W., Leeuw, I.M.V.-D., Uden-Kraan, C.F.V.,
Francke, A.L.: Effects of eHealth for patients and informal caregivers confronted with
cancer: a meta-review. Int. J. Med. Inform. 87, 54–67 (2016). https://doi.org/10.1016/j.
ijmedinf.2015.12.013
16. Queirós, A., Pereira, L., Dias, A., Rocha, N.P.: Technologies for ageing in place to support
home monitoring of patients with chronic diseases. In: van den Broek, E.L., Fred, A.,
Gamboa, H., Vaz, M. (eds.) Proceedings of the 10th International Joint Conference on
Biomedical Engineering Systems and Technologies - Volume 5: HEALTHINF, pp. 66–76.
INSTICC, Porto (2017). https://doi.org/10.5220/0006140000660076
17. Moher, D., Liberati, A., Tetzlaff, J., Altman, D.G.: Preferred reporting items for systematic
reviews and meta-analyses: the PRISMA statement. PLoS Med. 6, e0000097 (2009).
https://doi.org/10.1371/journal.pmed.1000097
18. Jackson, C.L., Bolen, S., Brancati, F.L., Batts-Turner, M.L., Gary, T.L.: A systematic
review of interactive computer-assisted technology in diabetes care. J. Gen. Intern. Med.
21, 105–110 (2006). https://doi.org/10.1007/s11606-006-0242-5
348 L. Pereira et al.
19. Martínez, A., Everss, E., Rojo-Álvarez, J.L., Figal, D.P., García-Alberola, A.: A systematic
review of the literature on home monitoring for patients with heart failure. J. Telemedicine
Telecare 12, 234–241 (2006). https://doi.org/10.1258/135763306777889109
20. Chaudhry, S.I., Phillips, C.O., Stewart, S.S., Riegel, B., Mattera, J.A., Jerant, A.F.,
Krumholz, H.M.: Telemonitoring for patients with chronic heart failure: a systematic
review. J. Card. Fail. 13, 56–62 (2007). https://doi.org/10.1016/j.cardfail.2006.09.001
21. Clark, R.A., Inglis, S.C., Mcalister, F.A., Cleland, J.G.F., Stewart, S.: Telemonitoring or
structured telephone support programmes for patients with chronic heart failure: systematic
review and meta-analysis. BMJ 334, 942 (2007). https://doi.org/10.1136/bmj.39156.
536968.55
22. Gysels, M., Higginson, I.J.: Interactive technologies and videotapes for patient education in
cancer care: systematic review and meta-analysis of randomised trials. Support. Care
Cancer 15, 7–20 (2006). https://doi.org/10.1007/s00520-006-0112-z
23. Jaana, M., Paré, G.: Home telemonitoring of patients with diabetes: a systematic assessment
of observed effects. J. Eval. Clin. Pract. 13, 242–253 (2007). https://doi.org/10.1111/j.
1365-2753.2006.00686.x
24. Verhoeven, F., Gemert-Pijnen, L.V., Dijkstra, K., Nijland, N., Seydel, E., Steehouder, M.:
The contribution of teleconsultation and videoconferencing to diabetes care: a systematic
literature review. J. Med. Internet Res. 9, e37 (2007). https://doi.org/10.2196/jmir.9.5.e37
25. Dang, S., Dimmick, S., Kelkar, G.: Evaluating the evidence base for the use of home
telehealth remote monitoring in elderly with heart failure. Telemedicine e-Health 15, 783–
796 (2009). https://doi.org/10.1089/tmj.2009.0028
26. Fox, M.P.: A systematic review of the literature reporting on studies that examined the
impact of interactive, computer-based patient education programs. Patient Educ. Couns. 77,
6–13 (2009). https://doi.org/10.1016/j.pec.2009.02.011
27. Fry, J.P., Neff, R.A.: Periodic prompts and reminders in health promotion and health
behavior interventions: systematic review. J. Med. Internet Res. 11, e16 (2009). https://doi.
org/10.2196/jmir.1138
28. García-Lizana, F., Muñoz-Mayorga, I.: Telemedicine for depression: a systematic review.
Perspect. Psychiatr. Care 46, 119–126 (2010). https://doi.org/10.1111/j.1744-6163.2010.
00247.x
29. Polisena, J., Tran, K., Cimon, K., Hutton, B., Mcgill, S., Palmer, K., Scott, R.E.: Home
telehealth for chronic obstructive pulmonary disease: a systematic review and meta-analysis.
J. Telemedicine Telecare 16, 120–127 (2010). https://doi.org/10.1258/jtt.2009.090812
30. Polisena, J., Tran, K., Cimon, K., Hutton, B., Mcgill, S., Palmer, K., Scott, R.E.: Home
telemonitoring for congestive heart failure: a systematic review and meta-analysis.
J. Telemedicine Telecare 16, 68–76 (2009). https://doi.org/10.1258/jtt.2009.090406
31. Sharp, I.R., Kobak, K.A., Osman, D.A.: The use of videoconferencing with patients with
psychosis: a review of the literature. Ann. Gen. Psychiatry 10, 14 (2011). https://doi.org/10.
1186/1744-859x-10-14
32. Saksena, A.: Computer-based education for patients with hypertension: a systematic
review. Health Educ. J. 69, 236–245 (2010). https://doi.org/10.1177/0017896910364889
33. Bolton, C.E., Waters, C.S., Peirce, S., Elwyn, G.: Insufficient evidence of benefit: a
systematic review of home telemonitoring for COPD. J. Eval. Clin. Pract. 17, 1216–1222
(2010). https://doi.org/10.1111/j.1365-2753.2010.01536.x
34. Kelders, S.M., Kok, R.N., Van Gemert-Pijnen, J.E.W.C.: Technology and adherence in
web-based interventions for weight control. In: Haugtvedt, C., Stibe, A. (eds.) Proceedings
of the 6th International Conference on Persuasive Technology: Persuasive Technology and
Design: Enhancing Sustainability and Health - PERSUASIVE 2011, p. 3. ACM, Columbus
(2011). https://doi.org/10.1145/2467803.2467806
Technologies for Ageing in Place: A Systematic Review 349
35. Ramadas, A., Quek, K., Chan, C., Oldenburg, B.: Web-based interventions for the
management of type 2 diabetes mellitus: a systematic review of recent evidence. Int.
J. Med. Inform. 80, 389–405 (2011). https://doi.org/10.1016/j.ijmedinf.2011.02.002
36. Baron, J., Mcbain, H., Newman, S.: The impact of mobile monitoring technologies on
glycosylated hemoglobin in diabetes: a systematic review. J. Diab. Sci. Technol. 6, 1185–
1196 (2012). https://doi.org/10.1177/193229681200600524
37. Cassimatis, M., Kavanagh, D.J.: Effects of type 2 diabetes behavioural telehealth
interventions on glycaemic control and adherence: a systematic review. J. Telemedicine
Telecare 18, 447–450 (2012). https://doi.org/10.1258/jtt.2012.gth105
38. Cheng, S.K., Dizon, J.: Computerised cognitive behavioural therapy for insomnia: a
systematic review and meta-analysis. Psychother. Psychosom. 81, 206–216 (2012). https://
doi.org/10.1159/000335379
39. Ciere, Y., Cartwright, M., Newman, S.P.: A systematic review of the mediating role of
knowledge, self-efficacy and self-care behaviour in telehealth patients with heart failure.
J. Telemedicine Telecare 18, 384–391 (2012). https://doi.org/10.1258/jtt.2012.111009
40. Frazetta, D., Willet, K., Fairchild, R.: A systematic review of smartphone application use
for type 2 diabetic patients. Online J. Nurs. Inform. (OJNI) 16(3) (2012)
41. Kennedy, C.M., Powell, J., Payne, T.H., Ainsworth, J., Boyd, A., Buchan, I.: Active
assistance technology for health-related behavior change: an interdisciplinary review.
J. Med. Internet Res. 14, e80 (2012). https://doi.org/10.2196/jmir.1893
42. Lieffers, J.R.L., Hanning, R.M.: Dietary assessment and self-monitoring: with nutrition
applications for mobile devices. Can. J. Diet. Pract. Res. 73, e253–e260 (2012). https://doi.
org/10.3148/73.3.2012.e253
43. Bacigalupo, R., Cudd, P., Littlewood, C., Bissell, P., Hawley, M.S., Woods, H.B.:
Interventions employing mobile technology for overweight and obesity: an early systematic
review of randomized controlled trials. Obes. Rev. 14, 279–291 (2012). https://doi.org/10.
1111/obr.12006
44. Blackman, K.C., Zoellner, J., Berrey, L.M., Alexander, R., Fanning, J., Hill, J.L.,
Estabrooks, P.A.: Assessing the internal and external validity of mobile health physical
activity promotion interventions: a systematic literature review using the RE-AIM
framework. J. Med. Internet Res. 15, e224 (2013). https://doi.org/10.2196/jmir.2745
45. Buchholz, S.W., Wilbur, J., Ingram, D., Fogg, L.: Physical activity text messaging
interventions in adults: a systematic review. Worldviews Evid.-Based Nurs. 10, 163–173
(2013). https://doi.org/10.1111/wvn.12002
46. Connelly, J., Kirk, A., Masthoff, J., Macrury, S.: The use of technology to promote physical
activity in type 2 diabetes management: a systematic review. Diabet. Med. 30, 1420–1432
(2013). https://doi.org/10.1111/dme.12289
47. El-Gayar, O., Timsina, P., Nawar, N., Eid, W.: A systematic review of IT for diabetes
self-management: are we there yet? Int. J. Med. Inform. 82, 637–652 (2013). https://doi.
org/10.1016/j.ijmedinf.2013.05.006
48. Foster, C., Richards, J., Thorogood, M., Hillsdon, M.: Remote and web 2.0 interventions
for promoting physical activity. Cochrane Database Syst. Rev. (2013). https://doi.org/10.
1002/14651858.cd010395.pub2
49. Geraedts, H., Zijlstra, A., Bulstra, S.K., Stevens, M., Zijlstra, W.: Effects of remote
feedback in home-based physical activity interventions for older adults: a systematic
review. Patient Educ. Couns. 91, 14–24 (2013). https://doi.org/10.1016/j.pec.2012.10.018
50. Belisario, J.S.M., Huckvale, K., Greenfield, G., Car, J., Gunn, L.H.: Smartphone and tablet
self management apps for asthma. Cochrane Database Syst. Rev. (2013). https://doi.org/10.
1002/14651858.cd010013.pub2
350 L. Pereira et al.
51. Moorhead, S.A., Hazlett, D.E., Harrison, L., Carroll, J.K., Irwin, A., Hoving, C.: A new
dimension of health care: systematic review of the uses, benefits, and limitations of social
media for health communication. J. Med. Internet Res. 15, e85 (2013). https://doi.org/10.
2196/jmir.1933
52. Pal, K., Eastwood, S.V., Michie, S., Farmer, A.J., Barnard, M.L., Peacock, R., Murray, E.:
Computer-based diabetes self-management interventions for adults with type 2 diabetes
mellitus. Cochrane Database Syst. Rev. (2010). https://doi.org/10.1002/14651858.
cd008776
53. Tao, D., Or, C.K.: Effects of self-management health information technology on glycaemic
control for patients with diabetes: a meta-analysis of randomized controlled trials.
J. Telemedicine Telecare 19, 133–143 (2013). https://doi.org/10.1177/1357633x13479701
54. Vugt, M.V., Wit, M.D., Cleijne, W.H., Snoek, F.J.: Use of behavioral change techniques in
web-based self-management programs for type 2 diabetes patients: systematic review.
J. Med. Internet Res. 15, e279 (2013). https://doi.org/10.2196/jmir.2800
55. Alvarez-Jimenez, M., Alcazar-Corcoles, M., González-Blanch, C., Bendall, S., Mcgorry,
P., Gleeson, J.: Online, social media and mobile technologies for psychosis treatment: a
systematic review on novel user-led interventions. Schizophr. Res. 156, 96–106 (2014).
https://doi.org/10.1016/j.schres.2014.03.021
56. Bort-Roig, J., Gilson, N.D., Puig-Ribera, A., Contreras, R.S., Trost, S.G.: Measuring and
influencing physical activity with smartphone technology: a systematic review. Sports Med.
44, 671–686 (2014). https://doi.org/10.1007/s40279-014-0142-5
57. Conway, A., Inglis, S.C., Clark, R.A.: Effective technologies for noninvasive remote
monitoring in heart failure. Telemedicine e-Health 20, 531–538 (2014). https://doi.org/10.
1089/tmj.2013.0267
58. Grustam, A.S., Severens, J.L., Nijnatten, J.V., Koymans, R., Vrijhoef, H.J.M.:
Cost-effectiveness of telehealth interventions for chronic heart failure patients: a literature
review. Int. J. Technol. Assess. Health Care 30, 59–68 (2014). https://doi.org/10.1017/
s0266462313000779
59. Hawley-Hague, H., Boulton, E., Hall, A., Pfeiffer, K., Todd, C.: Older adults’ perceptions
of technologies aimed at falls prevention, detection or monitoring: a systematic review. Int.
J. Med. Inform. 83, 416–426 (2014). https://doi.org/10.1016/j.ijmedinf.2014.03.002
60. Lazar, A., Thompson, H., Demiris, G.: A systematic review of the use of technology for
reminiscence therapy. Health Educ. Behav. 41, 51S–61S (2014). https://doi.org/10.1177/
1090198114537067
61. Nakamura, N., Koga, T., Iseki, H.: A meta-analysis of remote patient monitoring for
chronic heart failure patients. J. Telemedicine Telecare 20, 11–17 (2013). https://doi.org/10.
1177/1357633x13517352
62. Ngo, J., Engelen, A., Molag, M., Roesle, J., García-Segovia, P., Serra-Majem, L.: A review
of the use of information and communication technologies for dietary assessment. Br.
J. Nutr. 101, S102–S120 (2009). https://doi.org/10.1017/s0007114509990638
63. Or, C.K., Tao, D.: Does the use of consumer health information technology improve
outcomes in the patient self-management of diabetes? A meta-analysis and narrative review
of randomized controlled trials. Int. J. Med. Inform. 83, 320–329 (2014). https://doi.org/10.
1016/j.ijmedinf.2014.01.009
64. Peterson, A.: Improving type 1 diabetes management with mobile tools. J. Diab. Sci.
Technol. 8, 859–864 (2014). https://doi.org/10.1177/1932296814529885
65. Vegting, I., Schrijver, E., Otten, R., Nanayakkara, P.: Internet programs targeting multiple
lifestyle interventions in primary and secondary care are not superior to usual care alone in
improving cardiovascular risk profile: a systematic review. Eur. J. Intern. Med. 25, 73–81
(2014). https://doi.org/10.1016/j.ejim.2013.08.008
Technologies for Ageing in Place: A Systematic Review 351
66. Agboola, S.O., Ju, W., Elfiky, A., Kvedar, J.C., Jethwani, K.: The effect of technology-
based interventions on pain, depression, and quality of life in patients with cancer: a
systematic review of randomized controlled trials. J. Med. Internet Res. 17, e65 (2015).
https://doi.org/10.2196/jmir.4009
67. Amatya, B., Galea, M., Kesselring, J., Khan, F.: Effectiveness of telerehabilitation
interventions in persons with multiple sclerosis: a systematic review. Multiple Sclerosis
Relat. Disord. 4, 358–369 (2015). https://doi.org/10.1016/j.msard.2015.06.011
68. Bardus, M., Smith, J.R., Samaha, L., Abraham, C.: Mobile phone and web 2.0 technologies
for weight management: a systematic scoping review. J. Med. Internet Res. 17, e259
(2015). https://doi.org/10.2196/jmir.5129
69. Bert, F., Giacometti, M., Gualano, M.R., Siliquini, R.: Smartphones and health promotion:
a review of the evidence. J. Med. Syst. 38, 9995 (2013). https://doi.org/10.1007/s10916-
013-9995-7
70. Campbell, J., Porter, J.: Dietary mobile apps and their effect on nutritional indicators in
chronic renal disease: a systematic review. Nephrology 20, 744–751 (2015). https://doi.org/
10.1111/nep.12500
71. Chandak, A., Joshi, A.: Self-management of hypertension using technology enabled
interventions in primary care settings. Technol. Health Care 23, 119–128 (2015). https://
doi.org/10.3233/THC-140886
72. Farmer, A.J., Mcsharry, J., Rowbotham, S., Mcgowan, L., Ricci-Cabello, I., French, D.P.:
Effects of interventions promoting monitoring of medication use and brief messaging on
medication adherence for people with type 2 diabetes: a systematic review of randomized
trials. Diab. Med. 33, 565–579 (2015). https://doi.org/10.1111/dme.12987
73. Garabedian, L.F., Ross-Degnan, D., Wharam, J.F.: Mobile phone and smartphone
technologies for diabetes care and self-management. Curr. Diab. Rep. 15, 109 (2015).
https://doi.org/10.1007/s11892-015-0680-8
74. Groen, W.G., Kuijpers, W., Oldenburg, H.S., Wouters, M.W., Aaronson, N.K., Harten, W.
H.V.: Empowerment of cancer survivors through information technology: an integrative
review. J. Med. Internet Res. 17, e270 (2015). https://doi.org/10.2196/jmir.4818
75. Huang, Z., Tao, H., Meng, Q., Jing, L.: MANAGEMENT OF ENDOCRINE DISEASE:
effects of telecare intervention on glycemic control in type 2 diabetes: a systematic review
and meta-analysis of randomized controlled trials. Eur. J. Endocrinol. 172, R93–R101
(2014). https://doi.org/10.1530/eje-14-0441
76. Hunt, C.W.: Technology and diabetes self-management: an integrative review.
World J. Diab. 6, 225 (2015). https://doi.org/10.4239/wjd.v6.i2.225
77. Kim, A.R., Park, H.A.: Web-based self-management support interventions for cancer
survivors: a systematic review and meta-analyses. MedInfo (2015). https://doi.org/10.3233/
978-1-61499-564-7-142
78. Laranjo, L., Arguel, A., Neves, A.L., Gallagher, A.M., Kaplan, R., Mortimer, N., Mendes,
G.A., Lau, A.Y.S.: The influence of social networking sites on health behavior change: a
systematic review and meta-analysis. J. Am. Med. Inform. Assoc. 22, 243–256 (2014).
https://doi.org/10.1136/amiajnl-2014-002841
79. Lundell, S., Holmner, Å., Rehn, B., Nyberg, A., Wadell, K.: Telehealthcare in COPD: a
systematic review and meta-analysis on physical outcomes and dyspnea. Respir. Med. 109,
11–26 (2015). https://doi.org/10.1016/j.rmed.2014.10.008
80. Mateo, G.F., Granado-Font, E., Ferré-Grau, C., Montaña-Carreras, X.: Mobile phone apps
to promote weight loss and increase physical activity: a systematic review and
meta-analysis. J. Med. Internet Res. 17, e253 (2015). https://doi.org/10.2196/jmir.4836
81. Pedone, C., Lelli, D.: Systematic review of telemonitoring in COPD: an update.
Pneumonol. Alergol. Pol. 83, 476–484 (2015). https://doi.org/10.5603/piap.2015.0077
352 L. Pereira et al.
82. Riazi, H., Larijani, B., Langarizadeh, M., Shahmoradi, L.: Managing diabetes mellitus
using information technology: a systematic review. J. Diab. Metab. Disord. 14, 49 (2015).
https://doi.org/10.1186/s40200-015-0174-x
83. Richards, D., Richardson, T.: Computer-based psychological treatments for depression: a
systematic review and meta-analysis. Clin. Psychol. Rev. 32, 329–342 (2012). https://doi.
org/10.1016/j.cpr.2012.02.004
84. Tildesley, H.D., Po, M.D., Ross, S.A.: Internet blood glucose monitoring systems provide
lasting glycemic benefit in type 1 and 2 diabetes. Med. Clin. North Am. 99, 17–33 (2015).
https://doi.org/10.1016/j.mcna.2014.08.019
85. Arambepola, C., Ricci-Cabello, I., Manikavasagam, P., Roberts, N., French, D.P., Farmer,
A.: The impact of automated brief messages promoting lifestyle changes delivered via
mobile devices to people with type 2 diabetes: a systematic literature review and
meta-analysis of controlled trials. J. Med. Internet Res. 18, e86 (2016). https://doi.org/10.
2196/jmir.5425
86. Beishuizen, C.R., Stephan, B.C., Gool, W.A.V., Brayne, C., Peters, R.J., Andrieu, S.,
Kivipelto, M., Soininen, H., Busschers, W.B., Charante, E.P.M.V., Richard, E.: Web-based
interventions targeting cardiovascular risk factors in middle-aged and older people: a
systematic review and meta-analysis. J. Med. Internet Res. 18, e55 (2016). https://doi.org/
10.2196/jmir.5218
87. Chen, Y.-R.R., Schulz, P.J.: The effect of information communication technology
interventions on reducing social isolation in the elderly: a systematic review. J. Med.
Internet Res. 18, 18 (2016). https://doi.org/10.2196/jmir.4596
88. Christensen, J., Valentiner, L.S., Petersen, R.J., Langberg, H.: The effect of game-based
interventions in rehabilitation of diabetics: a systematic review and meta-analysis. Telemed.
e-Health 22, 789–797 (2016). https://doi.org/10.1089/tmj.2015.0165
89. Matthews, J., Win, K.T., Oinas-Kukkonen, H., Freeman, M.: Persuasive technology in
mobile applications promoting physical activity: a systematic review. J. Med. Syst. 40, 72
(2016). https://doi.org/10.1007/s10916-015-0425-x
90. Shingleton, R.M., Palfai, T.P.: Technology-delivered adaptations of motivational inter-
viewing for health-related behaviors: a systematic review of the current research. Patient
Educ. Couns. 99, 17–35 (2016). https://doi.org/10.1016/j.pec.2015.08.005
91. Ghapanchi, A.H., Aurum, A.: Antecedents to IT personnels intentions to leave: a systematic
literature review. J. Syst. Softw. 84, 238–249 (2011). https://doi.org/10.1016/j.jss.2010.09.022
92. Queirós, A., Santos, M., Rocha, N.P., Cerqueira, M.: Technologies for ageing in place to
support community-dwelling older adults. In: 2017 12th Iberian Conference on Information
Systems and Technologies (CISTI) (2017). https://doi.org/10.23919/cisti.2017.7975990
93. Queirós, A., Pereira, L., Santos, M., Rocha, N.P.: Technologies for ageing in place to
support the empowerment of patients with chronic diseases. In: Rocha, Á., Correia, A.M.,
Adeli, H., Reis, L.P., Costanzo, S. (eds.) WorldCIST 2017. AISC, vol. 570, pp. 795–804.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-56538-5_80
94. Kaplan, A.M., Haenlein, M.: Users of the world, unite! The challenges and opportunities of
social media. Bus. Horiz. 53, 59–68 (2010). https://doi.org/10.1016/j.bushor.2009.09.003
95. Boggatz, T., Meinhart, C.M.: Health promotion among older adults in Austria: a qualitative
study. J. Clin. Nurs. 26, 1106–1118 (2017). https://doi.org/10.1111/jocn.13603
96. Southard, B.H., Southard, D.R., Nuckolls, J.: Clinical trial of an internet-based case
management system for secondary prevention of heart disease. J. Cardiopulm. Rehabil. 23,
341–348 (2003). https://doi.org/10.1097/00008483-200309000-00003
Technologies for Ageing in Place: A Systematic Review 353
97. Consolvo, S., Klasnja, P., Mcdonald, D.W., Avrahami, D., Froehlich, J., Legrand, L.,
Libby, R., Mosher, K., Landay, J.A.: Flowers or a robot army? In: McCarthy, J., Scott, J.,
Woo, W. (eds) Proceedings of the 10th International Conference on Ubiquitous Computing
- UbiComp 2008, pp. 54–63. ACM, Seoul (2008). https://doi.org/10.1145/1409635.
1409644
98. Dickens, A.P., Richards, S.H., Greaves, C.J., Campbell, J.L.: Interventions targeting social
isolation in older people: a systematic review. BMC Public Health 11, 647 (2011). https://
doi.org/10.1186/1471-2458-11-647
99. Holt-Lunstad, J., Smith, T., Layton, J.: Social relationships and mortality risk: a
meta-analytic review. SciVee, e1000316 (2010). https://doi.org/10.4016/19865.01
100. Chronic Diseases and Associated Risk Factors in Australia. Australian Institute of Health
and Welfare, Canberra (2006). ISBN 978-1-74024-619-4
101. Thrall, J.H.: Prevalence and costs of chronic disease in a health care system structured for
treatment of acute illness. Radiology 235, 9–12 (2005). https://doi.org/10.1148/radiol.
2351041768
102. Bonow, R.O.: ACC/AHA clinical performance measures for adults with chronic heart
failure: a report of the American College of Cardiology/American Heart Association task
force on performance measures (writing committee to develop heart failure clinical
performance measures): endorsed by the heart failure society of America. Circulation 112,
1853–1887 (2005). https://doi.org/10.1161/circulationaha.105.170072
103. Calvo, G.S., Gómez-Suárez, C., Soriano, J., Zamora, E., Gónzalez-Gamarra, A.,
González-Béjar, M., Jordán, A., Tadeo, E., Sebastián, A., Fernández, G., Ancochea, J.:
A home telehealth program for patients with severe COPD: the PROMETE study. Respir.
Med. 108, 453–462 (2014). https://doi.org/10.1016/j.rmed.2013.12.003
104. Queirós, A., Carvalho, S., Pavão, J., Da Rocha, N.P.: AAL information based services and
care integration. In: Stacey, D., Solé-Casals, J., Fred, A., Gambo, H. (eds.) Proceedings of
the 6th International Conference on Health Informatics - HEALTHINF 2013, pp. 403–406.
INSTICC, Barcelona (2013). https://doi.org/10.5220/0004326004030406
A Model-Based Approach for Jump
Analyses Regarding Strength
and Balance
The Human as an Oscillating System
1 Introduction
Geriatric assessments gain an increasing relevance with the ongoing age-related
demographic shift. These assessments are well-established to identify early
changes associated with functional and cognitive decline [1–3] and usually consist
of several tests such as the Short Physical Performance Battery (SPPB) [4], the
de Morton Mobility Index (DEMMI) [5] or the Frailty Criteria [6]. It is impor-
tant to detect early changes in functional decline and to start interventions at
c Springer International Publishing AG, part of Springer Nature 2018
N. Peixoto et al. (Eds.): BIOSTEC 2017, CCIS 881, pp. 354–375, 2018.
https://doi.org/10.1007/978-3-319-94806-5_19
A Model-Based Approach for Jump Analyses 355
Table 1. Selection of geriatric assessment tests, their classification regarding the com-
ponents of physical fitness (− none, + significant, ++ highly significant), and test
duration. The test durations are based on literature and estimated on own experiences
(*) in a study with 250 participants [16].
Fig. 1. Major phases of a counter movement jump: standing (a), preparation (b),
take-off (c), flight (d), landing (e), and recovery (f). The orange marks indicate the
participant’s center of mass (COM) in each jump phase. Arrows mark the direction of
the movement. This Figure is an extended version of Fig. 1 in [17]. (Color figure online)
A Model-Based Approach for Jump Analyses 357
Force Platform: Force platforms measure the ground reaction force, which
acts on the plate. Figure 2 shows the coordinate orientation and dimensions of
the force platform used in our study. The main force, which acts perpendicular
to the plate in the vertical direction is Fz. The force in the mediolateral direction
is Fy and in the anterior-posterior direction Fx. The force platform in our study
measures with a sample rate of 200 Hz. Jump parameters such as power and jump
height are calculated automatically by the equations mentioned in Sect. 3.1.
Fig. 2. Coordinate orientation of the force plate and the IMU integrated into a hip-
worn belt. The force plate measures the force in the vertical (Fz ), mediolateral (Fy ),
and anterior-posterior direction (Fx ). The coordinate orientation of the triaxial sensors
of the IMU is illustrated as well.
3 Related Work
3.1 Strength: Calculation of Jump Power and Jump Height
The Counter Movement Jump is an established test for strength evaluation of
the lower extremities and can show an age-related decline in jump power [20].
According to [14,15] the parameters jump power and jump height seems to be
significant indicators for muscle performance and strength and thus, for the
functional ability. The jump power P can be calculated by the force acting on
the plate during the take-off phase. According to Newton’s second law, the force
F is equal to the mass m of an object times its acceleration a.
F =m·a (1)
The power P is defined by the force F times the velocity v:
P = F · v = ma · v (2)
The velocity v can be calculated by the integration of the acceleration a:
v = a · dt (3)
and the distance s of the center of mass (COM) from the neutral position by
s = v · dt = a · dt (4)
360 S. Hellmers et al.
The parameter jump height h can be estimated by the following equation, which
is based on projectile motion equations:
1
h = (vt · t) − · g · t2 , (5)
2
where vt is the vertical COM velocity at take-off, t is the time to peak flight
and g the gravity. The height represents the maximum vertical coordinate of the
COM trajectory following take-off of the jump.
10 Hz), 28–282% for trial lengths (20, 14, 10, 7, 5 and 3 s), as well as calculation
methods. Thereby they clarified the difficulties to compare TTS results recorded
among different systems based on the power measure.
While these analyses are performed based on single jump measurements for 25
healthy younger adults (20–53 years), its insights will apply due to the indicated
computational differences and the drastic effect sizes. Consequently, alternative
measures are desired, which are more robust regarding measurement-variations
such as sample rates.
Ideally, these measures should be equally applicable to rather mobile mea-
surement devices such as inertial measurement units (IMU), which will be
increasingly applied due to their lower price and the higher grade of mobility
[25–27].
Fig. 4. Relation between input function F(s), the transfer function of a system H(s)
and the output function Y(s).
5 Model-Based Approach
Current CMJ-based balance measures, such as COP and TTS, have been shown
to be significantly prone to algorithmic and technical variations (see Sect. 3.2).
Therefore, we propose the use of the oscillation behavior as an alternative app-
roach to drawing conclusions about muscle strength, balance ability, postural
stability, and neuromuscular control instead of using the DPSI, TTS, COP or
ROM. The advantage of the model-based approach of the oscillation behavior
(during the landing and recovery phase after a jump) over existing amplification-
based methods might be its potentially lower dependability on sample rates, and
trial lengths.
Our approach shall enable balance and strength analyses, which are more
robust than present methods, independent of measurement setting, and applica-
ble for different measuring devices.
In detail, we aim to model (as schematically illustrated in Fig. 6) the human’s
lower extremities as a spring that oscillates during the landing and recovery
phase. This means that the spring is slack during free fall in the flight phase and
will be compressed at the impact on the floor. After maximum compression, the
spring depresses during the recovery and stabilization phase to the steady state
in one or more oscillations.
A Model-Based Approach for Jump Analyses 363
Fig. 5. The phase of landing after a jump: the natural frequency describes the oscilla-
tions per time and the damping the decay of oscillations, which can be mathematically
described by an exponential function.
From a physical point of view, the spring (assumed as human’s lower extremities)
can be described by Hooke’s law:
F = −kx, (12)
with the force F , the displacement x and the spring constant k. The frequency
can be estimated with
ω0 = k/m. (13)
This equation shows that the frequency correlates to the spring constant k.
In our model, the spring is characterized by the spring constant. Conse-
quently, if comparing the spring with the muscles of the humans’ lower extrem-
ities, the spring constant characterizes the stiffness of the muscles in a first
approximation and is related to the damping. The natural frequency ω0 of a
system describes the dynamic quality and the ability how fast the system can
react on a stimulation. Higher frequencies indicate a system of a higher quality
since it reacts faster on a stimulation to compensate the force.
The damping ratio D characterizes the influence within or upon an oscillatory
system that has the effect of reducing its oscillations and might also be a relevant
parameter for the characterization of the balance ability, the postural stability,
and strength. The damping ratio is given by the following equation:
c
D= √ , (14)
2 mk
where c is the damping coefficient, k the spring constant and m the mass.
364 S. Hellmers et al.
Fig. 6. Comparing of the human’s lower extremities with a spring during the landing
and stabilization phase of a jump: the spring is slack during the flight phase and will
be compressed during the landing. After maximal compression, the spring depresses
to steady state in one or more oscillations. This figure is an extended version of Fig. 5
in [17].
6 Study Design
We analyzed 30 participants aged 70–87 years (50% female, 50% male). Each
participant performed three counter movement jumps in a sequence with a rest
of one minute between the jumps to avoid signs of fatigue. Further characteristics
of the study population are listed in Table 2. The group covers a typical range
of age, body weight, and body height for the group of pre-frail elderlies.
The jumps were performed on the AMTI AccuPower force platform and
with a sensor system integrated into a hip-worn sensor belt. The coordinate
orientations of both systems are illustrated in Fig. 2. The sampling rate of the
force platform is 200 Hz. The IMU consists of a triaxial accelerometer, gyroscope,
and magnetometer measuring with a frequency of 100 Hz.
For standardization, we analyzed the second jump of each participant regard-
ing the jump power, jump height, frequency in anterior-posterior, mediolateral
and vertical direction, as well as the spring constant, which is defined by an
acting force on the system and the resulting displacement due to this force (see
Table 2. Characteristics of our study population (n = 30) with the minimum (min),
maximum (max), and mean-value (mean) as well as the standard deviation (SD) of
age in years, body weight in kg and body height in cm.
Eq. 12). In contrast to [17] in which we made automatic analyzes of the param-
eters with the System Identification Toolbox of MATLAB, we decided to make
a manual analyze in this article since this seems to lead to better results.
The test procedures were approved by the local ethics committee (ethical
vote: Hannover Medical School No. 6948) and conducted in accordance with the
Declaration of Helsinki.
7 Results
7.1 Analyses of Jump Phases
Understanding the specific jump phases and the related signals of our measure-
ment systems is essential for the jump evaluation and parameter determination.
Figure 7 shows exemplarily the force plate measurement of the counter move-
ment jump of one participant. The blue line shows the acceleration, which was
calculated from the measured force on the basis of Eq. 1. The velocity (red line)
was calculated by the integration of the acceleration according to Eq. 3 and the
center of mass (COM) displacement from neutral position (green line) by Eq. 4.
We can clearly separate the phases mentioned in Sect. 2.
In detail: The jump starts from standing (a) with an almost constant accel-
eration a, velocity v, and COM-displacement s of about zero (a = v = s = 0).
The transition from standing to the preparation phase (b) can be recognized
by changes in the amplitudes of acceleration a and velocity v. In this phase,
a downward movement can be observed (negative value of COM-displacement
from neutral position) until the velocity amounts zero (v = 0). At this point (c)
the movement direction changes in an upwards movement (positive values of
velocity v). The Take-off with a maximum acceleration is marked by (d). In
the flight phase (e) the acceleration amounts a = −9.81 m/s2 , which equals the
gravity because the plate doesn’t measure anything when the participant is in
the air and standing in a rest position is defined as a = 0. At (f) in the middle
of the flight phase the velocity amounts zero again (v = 0) and the movement
direction changes in a downwards movements after a standstill. The jumper falls
downwards until he lands on the plate at point (g). The downward movement
continues until (h). In phase (i) the jumper tries to compensate the forces in
order to enter the phase of standing or resting (j) again.
The acceleration data of the IMUs can be analyzed in the same way. In
comparison to the force plate data, further evaluation of the flight phase are
possible.
Figure 8 shows the force measurement of the force plate during a jump. Each
phase described in Fig. 7 can also be identified in this Figure. Regarding Eq. 1 the
force is defined by F = m·a and thus, has a similar form than the acceleration. It
is clear that the mean force acts perpendicular to the force plate in the vertical (z)
direction. But there are also reaction forces in the mediolateral (y) and especially
in the anterior-posterior (x) direction. In the anterior-posterior direction is a peak
during the take-off phase pushing the feet off the ground, and during the landing
while the toes and heels strike the ground and compensate the movements.
366 S. Hellmers et al.
Fig. 7. Force plate measurement of a counter movement jump. The acceleration (blue),
velocity (red) and COM-displacement (green) are shown over the time. The jump
phases are standing (a), preparation (b) + (c), take-off (d), flight (e) + (f), landing and
stabilization (g)–(i) and resting (j). (Color figure online)
Therefore, the relationship between strength, jump power, and jump height is
given by:
Strength ∼ Jump P ower ∼ Jump Height
Table 3 lists the jump characteristics of our study population. It can be seen
that we have a broad variety regarding the jump performance and therefore,
observe strong as well as weak jumpers.
A Model-Based Approach for Jump Analyses 367
Fig. 8. Force per axes (x, y, z) over time during a counter movement jump. The major
phases of the counter movement jump are marked in the graph: standing (a), prepara-
tion (b), take-off (c), flight (d), landing (e), and recovery (f). This Figure was already
presented in [17].
Table 3. Jump characteristics of our study population with minimum (min), maximum
(max), and mean-value (mean), as well as the standard deviation (SD).
Fig. 10. Frequency in anterior-posterior, and vertical direction in relation to the DPSI.
In order to investigate the relation between frequency and balance ability, the
participants are divided into subgroups with a good and a low balance ability.
This was done based on the results of the DEMMI test item “tandem stand with
eyes closed”. If the participants were able to hold this position for 10 s, a good
balance ability was assumed. Figure 11 shows the frequencies in relation to the
DPSI for both subgroups.
The subgroup with good balance ability shows the similar slopes than in
Fig. 10. The group with the lower balance ability shows inverse relationships
between frequency and DPSI in both directions. Since a higher frequency indi-
cates a faster reaction on a stimulation and a better dynamic quality an inverse
relationship between DPSI and frequency was expected. The reason for the lin-
ear relation in vertical direction for the group with good balance ability might be
due to the need of a greater compensation movement when jumping higher. This
A Model-Based Approach for Jump Analyses 369
(a) Subgroup with good balance ability. (b) Subgroup with low balance ability.
Fig. 11. Frequency in anterior-posterior and vertical direction in relation to the DPSI
for a subgroup divided by their balance ability.
results in a higher DPSI, although the system can react with high frequencies
and a good dynamic quality.
Linear regression analyses show that there is no significant relationship
between DPSI and the frequencies for the whole study population and the sub-
group with low balance ability. A significant correlation between DPSI and the
frequency in vertical direction could only be found for the group with good
balance ability. Considering the DPSI as observable element, we performed a
stepwise model selection by the Akaike information criterion (AIC). The results
of the linear regression are shown in Table 4.
Table 4. Influence of the frequencies on the DPSI for the subgroup with good balance
ability. Significant results are marked by an asterisk.
Fig. 12. Spring constant in relation to the jump power. A linear regression shows a
significant relationship between these parameters.
The results for the spring constants of the participants regarding the jump
power are shown in Fig. 12. A significant linear relationship can be found. There-
fore, following correlation is given:
Table 5. Influence of the spring constant, the jump height, and the frequencies per
axis on the jump power. Significant results are marked by an asterisk.
8 Discussion
The main purpose of the present study was to develop a more robust analysis
strategy than present methods, which is independent of measurement settings
and applicable for different measuring devices. Therefore, we proposed to model
the human lower extremities as an oscillating system and tried to describe this
system by the parameters natural frequency and spring constant, which are the
characterization variables (besides the damping ratio) in a transfer function. In
order to investigate the suitability and sensitivity of our model, we conducted
a prospective study with 30 probands aged 70 and above and analyzed the
biomechanics of a jump as well as the resulting signals of our measurement
372 S. Hellmers et al.
systems. Besides the identification of the single jump phases, we focused on the
oscillating behavior in the phase of landing and recovery after a vertical jump.
The expected oscillating behavior could be observed in various degrees in all
participants. Thus, the suitability and the general validity of our model could
be confirmed.
In order to examine the applicability of our proposed model and the corre-
sponding novel parameters for the description of the strength and balance ability,
we have analyzed the spring constant and the natural frequency of each jump
and investigated their correlation with the common parameters jump height and
jump power as representing the strength and the DPSI as representing the bal-
ance. The found linear correlation between jump power and jump height are
consistent with previous results in literature since they are sensitive indicators
of the muscle performance and the muscle strength. A significant linear rela-
tion between jump power and spring constant was as well observed. Therefore,
the spring constant can be assumed as an alternative and reliable parameter
for strength and confirms our model’s fundamental assumption that the spring
constant characterizes the stiffness of the muscles in a first approximation and
is related to the damping.
Between the natural frequency and the DPSI no significant correlation was
found, except the relationship between frequency in the vertical direction and
DPSI for the subgroup with good balance ability. Since the natural frequency of
a system describes the dynamic quality and the ability how fast the system can
react on stimulations, higher frequencies indicate a system of a higher quality,
because it reacts faster on a stimulation to compensate the force. Therefore, an
inverse relationship between DPSI and frequency was expected. An explanation
for the linear relation might be the need of a greater compensation movement
when jumping higher. This results in a higher DPSI, although the system can
react with high frequencies and a good dynamic quality. However, since the oscil-
lating behavior was observed in all jumps, the general validity of our approach
could be confirmed.
Our findings confirmed also the independence of the natural frequency from
influences of demographic parameters such as age, sex, body height, and body
weight. The spring constant was only influenced by the body weight, which was
expectable due to the relation between spring constant and mass.
In summary, our model is applicable for different measuring devices and
independent of the system’s sampling rate since the observed frequencies lie in a
range below 20 Hz. Also, the trail length and the filter settings shouldn’t influence
the frequency. However, our investigations have been carried out on a limited
study population (n = 30) and further examinations are necessary to analyze the
relationship between frequency and balance ability and to make a systematic and
more in-depth analysis regarding the influences of the measurement settings.
9 Conclusion
Counter movement jumps are well-suited to measure the essential parameters
of functional ability (mobility, strength, and balance) within a single test. But
A Model-Based Approach for Jump Analyses 373
References
1. Clegg, A., Young, J., Iliffe, S., Rikkert, M.O., Rockwood, K.: Frailty in elderly
people. Lancet 381(9868), 752–762 (2013)
2. Cruz-Jentoft, A.J., Baeyens, J.P., Bauer, J.M., Boirie, Y., Cederholm, T., Landi,
F., Martin, F.C., Michel, J.P., Rolland, Y., Schneider, S.M., et al.: Sarcopenia:
European consensus on definition and diagnosis report of the European working
group on sarcopenia in older people. Age Ageing 39, 412–423 (2010)
3. Elsawy, B., Higgins, K.E.: The geriatric assessment. Am. Fam. Phys. 83(1), 48–56
(2011)
4. Guralnik, J.M., Simonsick, E.M., Ferrucci, L., Glynn, R.J., Berkman, L.F., Blazer,
D.G., Scherr, P.A., Wallace, R.B.: A short physical performance battery assessing
lower extremity function: association with self-reported disability and prediction
of mortality and nursing home admission. J. Gerontol. 49(2), M85–M94 (1994)
5. de Morton, N.A., Lane, K.: Validity and reliability of the de morton mobility
index in the subacute hospital setting in a geriatric evaluation and management
population. J. Rehabil. Med. 42(10), 956–961 (2010)
6. Fried, L.P., Tangen, C.M., Walston, J., Newman, A.B., Hirsch, C., Gottdiener, J.,
Seeman, T., Tracy, R., Kop, W.J., Burke, G., et al.: Frailty in older adults: evidence
for a phenotype. J. Gerontol. Ser. A: Biol. Sci. Med. Sci. 56(3), M146–M157 (2001)
7. Siglinsky, E., Krueger, D., Ward, R.E., Caserotti, P., Strotmeyer, E.S., Harris, T.B.,
Binkley, N., Buehring, B.: Effect of age and sex on jumping mechanography and
other measures of muscle mass and function. J. Musculoskelet. Neuronal Interact.
15(4), 301 (2015)
8. Cooper, R., Kuh, D., Cooper, C., Gale, C.R., Lawlor, D.A., Matthews, F., Hardy,
R., FALCon and HALCyon study teams: Objective measures of physical capability
and subsequent health: a systematic review. Age Ageing 40(1), 14–23 (2010)
374 S. Hellmers et al.
9. Hellmers, S., Steen, E.E., Dasenbrock, L., Heinks, A., Bauer, J.M., Fudickar, S.,
Hein, A.: Towards a minimized unsupervised technical assessment of physical per-
formance in domestic environments. In: Oliver, N., Czerwinski, M., Matic, A. (eds.)
Proceedings of the 11th EAI Conference on Pervasive Computing Technologies for
Healthcare, pp. 207–216 (2017)
10. Hellmers, S., Kromke, T., Dasenbrock, L., Heinks, A., Bauer, J.M., Hein, A.,
Fudickar, S.: Stair climb power measurements via inertial measurement units -
towards an unsupervised assessment of strength in domestic environments. In:
Fred, A., Gamboa, H., Zwiggelaar, R., Bermdez i Badia, S. (eds.) Proceedings of
the 11th International Joint Conference on Biomedical Engineering Systems and
Technologies, pp. 39–47 (2018)
11. Granacher, U., Gollhofer, A., Hortobágyi, T., Kressig, R.W., Muehlbauer, T.: The
importance of trunk muscle strength for balance, functional performance, and fall
prevention in seniors: a systematic review. Sports Med. 43(7), 627–641 (2013)
12. Buehring, B., Krueger, D., Fidler, E., Gangnon, R., Heiderscheit, B., Binkley, N.:
Reproducibility of jumping mechanography and traditional measures of physical
and muscle function in older adults. Osteoporos. Int. 26(2), 819–825 (2015)
13. Rittweger, J., Schiessl, H., Felsenberg, D., Runge, M.: Reproducibility of the jump-
ing mechanography as a test of mechanical power output in physically competent
adult and elderly subjects. J. Am. Geriatr. Soc. 52(1), 128–131 (2004)
14. Dietzel, R., Felsenberg, D., Armbrecht, G.: Mechanography performance tests and
their association with sarcopenia, falls and impairment in the activities of daily
living-a pilot cross-sectional study in 293 older adults. J. Musculoskelet. Neuronal
Interact. 15(3), 249–256 (2015)
15. Kalyani, R.R., Corriere, M., Ferrucci, L.: Age-related and disease-related muscle
loss: the effect of diabetes, obesity, and other diseases. Lancet Diab. Endocrinol.
2(10), 819–829 (2014)
16. Hellmers, S., Fudickar, S., Büse, C., Dasenbrock, L., Heinks, A., Bauer, J.M., Hein,
A.: Technology supported geriatric assessment. In: Wichert, R., Mand, B. (eds.)
Ambient Assisted Living. ATSC, pp. 85–100. Springer, Cham (2017). https://doi.
org/10.1007/978-3-319-52322-4 6
17. Hellmers, S., Fudickar, S.J., Dasenbrock, L., Heinks, A., Bauer, J.M., Hein, A.:
Understanding jump landing as an oscillating system: a model-based approach of
balance and strength analyses. In: van den Broek, E., Fred, A., Gamboa, H., Vaz,
M. (eds.) HEALTHINF, pp. 159–168 (2017)
18. Palma, S., Silva, H., Gamboa, H., Mil-Homens, P.: Standing jump loft time
measurement-an acceleration based method. In: Encarnao, P., Veloso, A. (eds.)
Biosignals, no. (2), pp. 393–396 (2008)
19. Bui, H.T., Farinas, M.I., Fortin, A.M., Comtois, A.S., Leone, M.: Comparison and
analysis of three different methods to evaluate vertical jump height. Clin. Physiol.
Funct. Imaging 35(3), 203–209 (2015)
20. Michaelis, I., Kwiet, A., Gast, U., Boshof, A., Antvorskov, T., Jung, T., Rittweger,
J., Felsenberg, D.: Decline of specific peak jumping power with age in master
runners. J. Musculoskelet. Neuronal Interact. 8(1), 64–70 (2008)
21. Wikstrom, E., Tillman, M., Smith, A., Borsa, P.: A new force-plate technology
measure of dynamic postural stability: the dynamic postural stability index. J.
Athletic Training 40(4), 305–309 (2005)
22. Meardon, S., Klusendorf, A., Kernozek, T.: Influence of injury on dynamic postural
control in runners. Int. J. Sports Phys. Ther. 11(3), 366 (2016)
A Model-Based Approach for Jump Analyses 375
23. Fransz, D.P., Huurnink, A., de Boode, V.A., Kingma, I., van Dieën, J.H.: Time
to stabilization in single leg drop jump landings: an examination of calculation
methods and assessment of differences in sample rate, filter settings and trial length
on outcome values. Gait Posture 41(1), 63–69 (2015)
24. Wikstrom, E., Tillman, M., Borsa, P.: Detection of dynamic stability deficits in
subjects with functional ankle instability. Med. Sci. Sports Exerc. 37(2), 169–175
(2005)
25. Choukou, M.A., Laffaye, G., Taiar, R.: Reliability and validity of an accelerometric
system for assessing vertical jumping performance. Biol. Sport 31(1), 55 (2014)
26. Elvin, N.G., Elvin, A.A., Arnoczky, S.P.: Correlation between ground reaction force
and tibial acceleration in vertical jumping. J. Appl. Biomech. 23(3), 180 (2007)
27. Milosevic, B., Farella, E.: Wearable inertial sensor for jump performance analysis.
In: Proceedings of the 2015 Workshop on Wearable Systems and Applications, pp.
15–20 (2015)
28. Lawson, M., Webster, A., Stenson, M.: Effects of ankle bracing on the performance
of the dynamic postural stability index. In: Conference Proceedings on Interna-
tional Journal of Exercise Science, vol. 12, p. 42 (2015)
Regression, Classification and Ensemble
Machine Learning Approaches to Forecasting
Clinical Outcomes in Ischemic Stroke
1 Introduction
Stroke is defined as the rapid loss of brain function caused by disturbances in the blood
supply to the brain. It is one of the leading causes of death worldwide [1]. There are
broadly two types of stroke – ischemic and hemorrhagic. The former occurs due to lack
of blood flow to the brain, and the latter is caused by internal bleeding. Ischemic stroke,
which accounts for almost 87% of all strokes [2], is the focus of this study. We consider
different machine learning approaches that can build predictive models on the clinical
data of ischemic stroke patients, with the aim of forecasting clinical outcomes 90 days
after stroke. The data were collected retrospectively from the University of Mas-
sachusetts Medical School, Worcester, Massachusetts, USA and comprise demographic
information, medical history and treatment records of 439 patients.
There are several reasons for wishing to predict how well a stroke patient will
recover from stroke. If any such prediction can be made with a reasonable degree of
confidence, it may assist medical experts manage stroke more effectively and allocate
resources more efficiently. Furthermore, when the predictive models are built from the
stroke data, they can reveal information on the factors that affect stroke outcome.
In this paper, the outcome of stroke patients is measured in terms of the modified
Rankin Scale (mRS) score, an integer value between 0 and 6 measuring the degree of
disability or dependence in daily activities of people who have suffered a stroke [3].
There are three approaches one may use to solve the problem of predicting this value.
The first is to treat the target as a numeric attribute and apply some form of regression.
The second approach would be to think of the several different mRS scores as different
categories, in which case the problem becomes that of multinomial classification. The
third approach is to note the ordered nature of the target attribute, and use ordinal
regression techniques for prediction. We have addressed the prediction task from all
three perspectives.
The outcome prediction of the patients is performed on the data we have about the
patient at the time of discharge. The independent attributes in our study consist of
demographic information, medical history and treatment records. The response or target
attribute is mRS-90, the mRS score at 90 days following stroke onset (described in
Table 1). Treating mRS scores as integer values, we perform regression analysis, in
which we find that M5 model trees used in tandem with bootstrap aggregating (bagging)
significantly outperforms other common regression methods such as linear regression.
We then treat the target as a multinomial categorical attribute and apply several clas-
sification techniques such as logistic regression and C4.5 decision trees. Our studies
show that classification using the aforementioned regression technique followed by
translation of the target to a discrete value performs better than the standard classifi-
cation methods mentioned. Finally, we treat mRS-90 as an ordered categorical variable,
and use methods of ordinal regression using different link functions, of which the logit
link function yield the best results. We visualize and analyze the models obtained
through the different algorithms. The work presented in this paper is an extended and
enhanced version of the work presented in a conference paper by the same authors [4].
In Sect. 2 of this paper, we present some background information on the different
techniques used, and mention some related work from the literature. The methodology
of our research is described in Sect. 3. That section deals with the steps that are taken to
prepare and preprocess the data, and describes in full details the parameter choices we
made in our prediction techniques. Section 4 presents a comparison of different pre-
diction methods, and analyzes the results to gain more insights about the models dis-
covered. Section 5 concludes with a summary of findings and directions for future work.
current form by a group of researchers during the late 1980s [5]. The score is an integer
between 0 and 6 signifying the various degrees of impairment caused by stroke, with 0
being the least amount of impairment and 6 being death. Table 1 presents a summary
description of the different mRS scores. This study, for reasons described later,
excludes patients with mRS scores = 6.
Usually the score is assessed by a medical expert with the help of a questionnaire at
various stages of stroke. Figure 1 depicts a simplified questionnaire developed by [7]
that is used to assess mRS scores. In this study, the mRS scores are recorded in three
different stages. The first is before admission, in which the degree of disability the
patient had before the onset of stroke is retrospectively assessed. The next is at the time
of discharge from the hospital after initial treatment of stroke. The last one is at 90 days
after stroke onset, the score this study attempts to predict.
of the predictor generates a separate prediction output. These individual outputs are
then aggregated according to the type of task in question. For classification, a majority
vote is taken. For the task of regression, the average of the predictor versions is taken
[13]. Figure 3 summarizes the bagging process. Bagging improves generalization error
by reducing the variance of the individual predictive models. If a base predictor is
unstable - if it is not robust to fluctuations - the bagging process helps to stabilize it [8].
Fig. 2. (a) An example model tree built with the M5 algorithm with input attributes X and Y.
Linear models LM 1 to LM 4 are built in the leaves. (b) The corresponding problem space
showing separate subspaces defined by the tree and how each linear model fits points in the
subspace. With permissions of both the authors and the publisher, this figure is adopted from [4].
The size and the number of the bootstraps are parameter that can be changed. In the
most common case, the size of each bootstrap Bi is n, the same as that of the entire
training set. In this case, on an average Bi contains approximately 63% of the original
training data since each sample has a probability of 1 − (1 − 1/n)n of being picked,
which converges to about 0.63 for sufficiently large n [14]. This is because sampling is
done with replacement, resulting in duplicate instances in each bootstrap.
Some examples of link functions are: logit, probit and complementary log-log (cloglog).
Ordinal regression with a logit link function is similar to logistic regression, so a com-
parison of the two may serve as a good explanation of how ordinal regression works.
Fig. 3. Summary of the process of bagging. From the training set, k bootstraps are created. Each
bootstrap B1, …, Bk is used to build predictor versions V1, …, Vk which make separate
predictions P1, …, Pk. The final prediction Pf is a combination (average for regression, majority
voting for classification) of all the predictions. With permissions of both the authors and the
publisher, this figure is adopted from [4].
Let us first start with the binary logistic regression model. For consistency with the
literature, we use in this subsection the term variable instead of attribute. The predictive
attributes are referred to as independent variables and the target attribute is referred to
as response variable. Let X = {X1, …, Xm} be the vector of independent variable and
Y be the response variable with possible outcomes 0 and 1. Let p be the probability that
Y = 1. That is, p = P(Y = 1). In logistic regression, it is assumed that there is a linear
relationship between X and the logit transform of p:
p
logit ¼ ln ¼ a þ b1 X1 þ . . . þ bm Xm ð1Þ
1p
pi
logitðY ¼ iÞ ¼ ln ¼ ai þ bi1 X1 þ . . . þ bim Xm ; i ¼ 1; . . .; k 1 ð2Þ
pk
Now let us examine the case in which the response variable Y is ordinal. The
multinomial logistic regression model of Eq. 2 will not make use of the information
about the ordering. One way to take advantage of the ordering is to use cumulative
probabilities and cumulative logits. Considering k ordered categories the cumulative
probabilities are defined by P(Y i) = p1 +, …, + pi and the cumulative logits are
defined by
PðY iÞ p1 þ . . . þ pi
logitðY iÞ ¼ ln ¼ ln
1 PðY iÞ pi þ 1 þ . . . þ pk ð3Þ
¼ ai þ bi1 X1 þ . . . þ bim Xm ; i ¼ 1; . . .; k 1
However, for ordered categories under certain conditions, the logistic coefficients
do not depend on i, resulting in only one common bij for each covariate. In this case the
cumulative odds are given by eai eb1 X1 þ ... þ bm Xm , which means that the k odds for each
category i differ only with regards to the intercepts ai. This means that instead of the
k − 1 different linear models of Eqs. 2 and 3 can be expressed by one linear model and
the k − 1 intercepts a1…k−1. The a1…k−1 can be thought of as the cut-off points in the
underlying latent continuous variable. The equations and explanations in this subsec-
tion are adapted from [19].
Apart from the logit function shown in our example, other link functions following
the same basic principles of ordinal regression may also be used. Table 2 shows a
summary of the different link functions we experimented with.
Table 2. Different link functions used in the ordinal regression experiments in this paper. The
formulae are given in terms of the probability p. For the probit function, u is the cumulative
distribution function for the normal distribution [18].
Link function Formula
Logit p
ln 1p
Probit u1 ð pÞ
Complementary log log (cloglog) lnð lnð1 pÞÞ
Cauchit tanðpðp 1=2ÞÞ
the effect of which on stroke outcome needs to be examined. The coefficients and odds
ratios computed from the logistic regression model may lead to useful conclusions. We
present here a selected few among such studies. [20] reported that using statins for
treatment of ischemic stroke improved stroke outcome since the statins obtained an
odds ratio of 1.57 in a logistic regression model predicting mRS-90 2. This means
that the patients who are administered statins have 1.57 times the probability of
attaining mRS-90 2 than those who are not treated with statins. A similar study with
a different goal was done in [21] – the authors studied the effects of atrial fibrillation in
stroke outcome. [22] studied hyperglycemia as a factor of stroke and concluded that it
is associated with poor outcome. It was found in [23] that successful revascularization
is associated with good outcome. [24] reported that leukoaraiosis is a factor influencing
poor 90-day outcome of stroke. All the above studies dichotomized the mRS score to
two levels – one consisting of mRS-90 2 and the other of mRS > 2.
Some studies, however, have duly observed the ordinal nature of the measure of
stroke outcome, and taken advantage of this by using ordinal regression. [25] used
ordinal regression for their study on how comorbid conditions affect stroke outcome. In
[26] it was observed that choosing an ordinal rather than a binary method of analysis
allows trials of stroke outcome to have much smaller sample size for a given statistical
power. [27] used ordinal logistic regression to study the effectiveness of angiotensin-
receptor blocker in improving stroke outcome based on the modified Rankin score.
A point to note, however, is that the studies mentioned are concerned more with the
details of the regression models – binary or ordinal, and less with the evaluation of the
models’ performance from a prediction perspective. There have not been many studies
that focused solely on predicting the stroke outcome and employing machine learning
models to assist in the prediction task. [28] aimed to predict stroke outcome using
linear regression, but used the functional independence measure (FIM) which is a scale
that measures stroke recovery in terms of activities of daily living [29]. A similar effort
of predicting FIM was made by in [30]. Both these papers only used linear regression
for the prediction task. To the best of our knowledge, there is no study that has
methodically explored machine learning methods to predict the mRS-90 score as a
measure of stroke outcome.
3 Methodology
homogeneous values were also removed. The final set of attributes included demo-
graphic information (such as age and gender), medical history (such as diabetes and
hypertension), habits history (such as smoking and drinking), subtype of stroke (such
as large vessel and cardioembolic) [31], prescribed medication (such as anticoagulants),
and mRS scores at different stages (before admission, at discharge and at 90 days).
A measure of stroke severity determined by the National Institutes of Health Stroke
Scale (NIHSS) score [32] were also included. Table 3 presents summary statistics of all
the attributes of the stroke dataset used in this study. For the multivalued attribute
stroke subtype, five binary attributes for the five possible values were created, with
each attribute value specifying whether (1) or not (0) the patient had that particular
subtype of stroke. This is done since there is no ordinal relationship among the different
stroke types; so giving them numeric scores would make the model incorrect.
One important point to note is that patients who died within 90 days of stroke,
therefore having a mRS score of 6, were excluded from this analysis. The reason is that
patient death can occur for a combination of several reasons apart from stroke, such as
advanced age or additional comorbid conditions. Therefore, for stroke outcome pre-
diction, a decision was made to conduct experiments only with the patients who
survived the stroke after 90 days. Prominent works on this area such as the Copenhagen
Stroke Study [33] have also excluded dead patients in some of their models.
3.2 Regression
Algorithms and Parameters. We considered several regression techniques to predict
stroke outcome. We used the machine learning tool Weka (version 3.8) [34] to run
experiments with different regression methods. From our experiments, the most suc-
cessful regression method was bagged M5 model trees. For bagging, the bootstrap size
was kept equal to the size of the training set. 10 bootstraps were created from each
training set. The model trees were pruned to reduce overfitting. This was done by
requiring a minimum of 10 instances per leaf in the model trees. For every other
parameter, the default values in Weka (version 3.8) were applied. For comparison with
the bagged M5 model trees, we ran experiments with the most commonly used
regression algorithm – linear regression. We used linear regression with and without
bagging for performance comparison.
Evaluation Criteria. The performance of the regression models can be evaluated by
measuring the degree of similarity between the actual values of the target attribute, and
the predicted values returned by the models. 10-fold cross validation [35] was used to
assess how well the models will generalize to an independent dataset. The training data
set was pseudo-randomly partitioned into 10 subsets. Ten training and test rounds were
then performed; on the i-th round, the union of all subsets except the i-th was used to
train the regression model, which was then tested on the i-th subset. The reported
performance value is the mean of the ten performance values obtained in this way. We
measured three criteria for evaluation: the Pearson correlation coefficient, mean
absolute error and root mean squared error.
Regression, Classification and Ensemble Machine Learning Approaches 385
Table 3. Summary statistics of the attributes of the stroke dataset. The total number of patients
is 439. For continuous attributes, the mean and standard deviation are shown in a Mean ± Std.
Dev. format. For categorical attributes the percentages of different values are given. For binary
attributes, only the percentages of TRUE values are shown. For mRS scores at different stages,
we summarize the overall mean and standard deviation along with the distribution of individual
scores. For some of the attributes, shorter names are devised and shown in the rightmost column.
These shorter versions of attribute names are used when describing the models in Sect. 4.
Attribute Distribution of values Short name
(if applicable)
Stroke subtype Small vessel: 12.3%, Large vessel: 23.7%, Type = {SmVes,
Cardioembolic: 31.4%, Cryptogenic: 23.7%, LgVes, Cardio,
Others: 8.9% Crypto, Others}
Gender Male: 57.4%, Female: 42.6%
Age 67.2 ± 14.6, Range: 19–97
NIHSS score at 7.2 ± 7.1, Range: 0–32 NIHSS-adm
admission
Hypertension 74.7% HTN
Hyperlipidemia 58.8% HLD
Diabetes 29.8%
Smoking 29.4%
Alcohol problem 14.6% Alcohol
Previous history 19.4% Hist-stroke
of stroke
Atrial Fibrillation 27.7% AFib
Carotid Artery 21.0% CAD
Disease
Congestive Heart 8.7% CHF
Failure
Peripheral Artery 6.4% PAD
Disease
Hemorrhagic 11.2% Hem-Conv
conversion
tPA 20.5%
Statins 47.4%
Antihypertensives 62.9% Anti-HTN
Antidiabetics 20.5%
Antiplatelets 45.3%
Anticoagulants 10.3% Anti-Coag
Perfusion 8.7%
Neurointervention 18.7% NeuroInt
mRS before 0.41 ± 0.86 (0: 74.0%, 1: 15.0%, 2: 5.9%, mRS-adm
admission 3: 2.1%, 4: 1.4%, 5: 0.5%)
mRS at discharge 1.60 ± 1.63 (0: 35.3%, 1: 13.7%, 2: 15.3%, mRS-disch
3: 9.8%, 4: 11.6%, 5: 5.0%)
mRS at 90 days 1.28 ± 1.46 (0: 46.9%, 1: 17.5%, 2: 14.4%, mRS-90
3: 11.6%, 4: 6.2%, 5: 3.4%)
386 A. Kabir et al.
1X n
MAE ¼ jzi ^zi j ð5Þ
n i¼1
sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
1X n
RMSE ¼ ðjzi ^zi jÞ2 ð6Þ
n i¼1
Where n is the number of predictions, z1 ; . . .; zn are the actual and ^z1 ; . . .; ^zn are the
predicted values respectively [37]. Since MAE and RMSE measure differences
between the actual and predicted outputs, lower values of these measures indicate better
prediction performance.
3.3 Classification
Algorithms and Parameters. The best classification method found in this study is a
technique where a classification is done based on the regression results obtained from
bagged M5 model trees (as described in Sect. 3.2). When the numeric prediction is
obtained from this regression method, we round it to the nearest integer and assign the
instance to the class corresponding to that integer. For example, a regression output of
0.35 is assigned to class “0” and an output of 3.83 is assigned to class “4”. We denote
this approach here as classification via regression. For comparison, we consider two
more widely used classification algorithms: logistic regression [15] and C4.5 decision
tree [38]. The choice of logistic regression is motivated by the fact that it is the standard
classification method used in clinical trial studies. As for decision tree, it gives a good
diagrammatic representation of the prediction process as well as proving to be
empirically successful in classification tasks. The machine learning tool Weka (version
3.8) [34] was used for our classification experiments as well. To reduce overfitting, the
C4.5 Decision trees were pruned by requiring at least 10 instances per leaf. For every
other parameter, the default values in Weka (version 3.8) were applied.
Regression, Classification and Ensemble Machine Learning Approaches 387
Evaluation Criteria. The main evaluation criterion for the classification algorithms
used in this study is accuracy – the percentage of cases where the actual and the
predicted classes are the same. But since there are six different classes with subtle
variations between two adjacent mRS-90 scores, we may consider predictions that are
close enough to the actual score to be fairly accurate as well. We therefore define
“near-accuracy” to refer to the percentage of cases in which the classifier either makes
an accurate prediction, or makes a wrong prediction which is either one more or one
less than the correct mRS score. For example, if the correct class is 3, only a prediction
of 3 will be correct in terms of accuracy, but a prediction of 2, 3 or 4 will be acceptable
in terms of near-accuracy. Once again, 10-fold cross validation [35] (using the process
described in Sect. 3.2) was used to assess how well the models will generalize to an
independent dataset.
Results show that bagging used in tandem with M5 model trees performs much
better than all the other techniques. Even without bagging, the M5 model trees perform
better than linear regression. The improvement in performance is most impressive
when the mean absolute error is considered, but not so much when we consider the root
mean squared error. This leads us to an important point.
Large errors have a relatively greater influence when the errors are squared. So, if
the variance associated with the frequency distribution of the error magnitude increa-
ses, the difference between MAE and RMSE also increases [41]. From our observation
of the relative MAE and RMSE values of the M5 model tree, we can conclude that
there is a high variance in the prediction error of the M5 model tree. It therefore makes
sense that a variance-reducing procedure like bagging should improve the performance
of the model tree, as observed in Table 4. Note however that bagging does not have the
same kind of effect in improving the performance of linear regression.
To see if any statistically significant improvement in performance is achieved, we
performed paired t-tests in terms of correlation coefficient on each pair of the five
methods considered. The difference between means for each pair are examined at a
p-value of 0.05. The results of the tests are presented in Table 5. It shows that the
performances of linear regression and M5 model trees (without bagging) are not sta-
tistically different from each other. When bagging is used with linear regression, it is
unable to improve the performance significantly. However, when bagging is used with
M5 model trees, the resulting regression model performs significantly better than the
models of all the other methods.
Table 4. Comparison of different regression methods on stroke data in terms of R, MAE and
RMSE. For R, higher values indicate better model fit, whereas for the MAE and RMSE metrics
lower values are better. Table reused from [4].
Method R MAE RMSE
Average prediction −0.136 1.235 1.461
Linear regression 0.779 0.654 0.916
M5 model tree 0.785 0.577 0.905
Bagging with linear regression 0.783 0.649 0.908
Bagging with M5 model trees 0.822 0.537 0.832
Analysis of the Linear Regression Model. The linear regression model returns a
linear equation delineating the amount of influence each predictive attribute has on the
final outcome. Figure 4 shows the model obtained through linear regression. The
attributes with a positive coefficient contribute to an increase in the mRS-90 score
(worse outcome). The magnitude of the coefficient points to the degree of contribution.
Note that the values of the continuous attributes age and NIHSS were scaled to a range
between 0 and 1 before running linear regression. This allows comparison of their
coefficients with those of all the other attributes.
From the linear regression model, we observe that older age, higher NIHSS at
admission (more severe initial stroke) and higher mRS at discharge all have large
positive coefficients. These implies that they are all predictive of a poor outcome.
Regression, Classification and Ensemble Machine Learning Approaches 389
Alcohol also has noticeable contribution towards a poorer outcome. All the stroke
subtypes have a negative coefficient, so it is difficult to draw conclusions from these
coefficients. Among the other attributes that have a negative coefficient, Perfusion is
found to have fairly high contribution towards a better outcome.
Fig. 4. Linear regression model to predict 90-days outcome of stroke from patients’ data.
Analysis of the M5 Model Tree. We investigate the model returned by the M5 model
tree algorithm to find insights about stroke outcome. Figure 5 shows the model tree
where each leaf is a linear equation. The linear equations are shown alongside the tree.
The sign and magnitude of coefficients of each predictive attribute in the equations give
an indication of how the output attribute responds to changes in the given input
attribute. The continuous variables age and NIHSS at admission are scaled to the range
between 0 and 1, so that the magnitudes of these attributes are within the [0,1] range.
390 A. Kabir et al.
From the model tree of Fig. 5, it is clear that the tree divides the input space based
on the mRS score at discharge, and builds linear models for different ranges of that
score. By following the decision branches of the tree, we can see that the linear models
LM 1 and LM 2 correspond to mRS discharge scores of 0 and 1 respectively. LM 3 is
associated with mRS discharge scores of 2 and 3, and LM 4 with scores of 4 and 5.
Let us now take a closer look at each of the linear models. In LM 1, the y-intercept
is a very small value and there is no other attribute that has a large enough coefficient to
change the prediction substantially. This means that the mRS-90 prediction for almost
all patients reaching this point of the tree will be close to 0. At LM 2, the mRS-disch
value is 1 with a coefficient of 0.1265. Since the y-intercept is 0.3596, if all the other
attributes are absent, the output is 0.4861. Let us call this the baseline prediction for
this leaf. Other attributes will either increase or decrease this baseline prediction based
on their coefficients. Older age, higher NIHSS at admission and antihypertensives
contribute towards increasing the mRS-90 score. On the other hand, cardioembolic and
cryptogenic strokes contribute significantly towards lowering the mRS-90 score.
At LM 3, the mRS score can be either 2 or 3. In the former case, the baseline prediction
is 2*0.0951 − 0.3414 = −0.1512 and in the latter case, it is 3*0.0951 − 0.3414 =
−0.0561. However, there are several attributes in this model that may have a major
impact on the final prediction, notably age, NIHSS at admission, diabetes, large vessel
stroke subtype and mRS before admission. Higher values for some or all of the above
attributes will result in increased mRS-90 score. For LM 4, the baseline prediction is
either 2.6762 (for mRS discharge = 4) or 4.1181 (for mRS discharge = 5). If a patient
reaches this leaf, the output mRS-90 prediction is likely to be quite high. Only neu-
rointervention has a major effect of lowering the mRS-90 score.
Fig. 5. The M5 model tree built on the stroke dataset with minimum 10 instances in each leaf.
Each leaf is a linear model (LM1–LM4) predicting the target attribute mRS-90. The numbers
under the leaves indicate how many instances are covered under that particular linear model.
Each of the linear models are shown in detail alongside the tree.
Regression, Classification and Ensemble Machine Learning Approaches 391
1 2
3 4
5 6
Fig. 6. Model trees obtained from bootstraps 1–6 when using bagging with M5 algorithm on the
stroke data. Bootstrap numbers are shown alongside the tree for reference.
392 A. Kabir et al.
7 8
9 10
Fig. 7. Model trees obtained from bootstraps 7–10 when using bagging with M5 algorithm on
the stroke data. Bootstrap numbers are shown alongside the tree for reference.
Table 6. Comparison of logistic regression, C4.5 and classification via regression (bagging with
M5 model trees) on the stroke dataset in terms of accuracy and near-accuracy. Table reused
from [4].
Method Accuracy Near-accuracy
Majority class 46.9% 64.4%
Logistic Regression 54.2% 83.6%
C4.5 (with pruning) 56.7% 86.8%
Classification via regression 59.7% 90.0%
Analysis of the C4.5 Decision Tree. For classification of mRS-90 values, we created
a decision tree using the C4.5 algorithm. In order to avoid overfitting, we pruned the
trees by imposing restrictions of having a minimum of 10 instances per leaf. The
decision tree we obtained is shown in Fig. 8. The structure of the classification tree is
similar to that of the regression trees we discussed before. The difference is that the leaf
nodes of the classification tree are mRS-90 scores (0,…,5) rather than linear models to
calculate the scores. Like the trees in our regression models, this tree also has mRS at
discharge as the primary factor. Patients with mRS at discharge = 0 are predicted to
have mRS-90 = 0. The same is true for patients with mRS at discharge = 1 unless they
have a non-cryptogenic stroke and use antihypertensives, in which case the mRS-90
prediction is 1. Patients having mRS at discharge of 2 or 3 usually are predicted to end
Table 8. Confusion matrix for the method of supervised classification via regression using
bagging with M5 model trees. The rows show the actual mRS scores while the columns show the
ones predicted by the model. The diagonals (in darker gray) are the correct predictions. The cells
adjacent to the diagonals (in lighter gray) are near-correct predictions missing the actual score
by 1. Table adapted from [4].
Actual Predicted
0 1 2 3 4 5
0 159 36 11 0 0 0
1 10 40 19 8 0 0
2 2 15 31 14 1 0
3 0 8 19 21 3 0
4 0 3 5 8 10 1
5 0 3 1 2 8 1
Regression, Classification and Ensemble Machine Learning Approaches 395
up with mRS-90 of 2. There are, however, exceptions in two cases: If the patients have
low (0 or 1) mRS before admission, has a history of alcohol consumption and is
younger than 87, they are predicted to do better and have mRS-90 of 1; On the other
hand, if mRS before admission is higher than 1 and NIHSS at admission is higher than
5, they are predicted to do worse and have mRS-90 of 3. Patients having mRS at
discharge of 4 have predicted mRS-90 of 2 or 3. Patients having mRS at discharge of 5
are predicted to have mRS-90 of 4 or 5.
Fig. 8. C4.5 Decision tree constructed from the stroke data by treating the mRS-90 scores as
multinomial categories.
Analysis of the Logistic Regression Model. Among the six categories of mRS-90, we
chose the last category (mRS-90 = 5) as the reference category for multinomial logistic
regression. Table 9 shows the coefficients that were obtained for each attribute/category
combination. When analyzing this model, we need to consider the fact that all the
coefficients are with respect to the reference category. So positive coefficients imply
higher probability of being classified in that category than the reference category.
A negative coefficient would imply the opposite. As a concrete example, diabetes has
negative coefficients for all the categories, so a diabetic patient is less likely to have
mRS scores 0–4 than mRS score of 5. We inspected the sign and magnitude of each
attribute, and compared their coefficients across categories. The following attributes
were found to have very high negative coefficients in all categories: age, NIHSS score
at admission, and hypertension. In addition, several other attributes such as gender,
diabetes and alcohol have negative coefficient across all categories. This means that all
these attributes have a lower probability of being in the lower mRS scores (lower
probability of better outcome). As for the different treatment methods, only perfusion
has positive coefficients in all categories. Antidiabetics and antiplatelets also have
positive coefficients for mRS scores of 0–3. This indicates that these methods of
treatment improve the probability of going to a lower mRS score (better outcome).
396 A. Kabir et al.
Table 9. Logistic regression coefficients for each category of the mRS-90 attribute for the stroke
data. Here mRS-90 = 5 is used as the reference category.
Attribute Coefficients for each category
0 1 2 3 4
Subtype = Large vessel −0.886 1.652 −1.080 −13.28 −2.426
Subtype = Small vessel −0.858 0.162 −2.960 −14.45 −4.619
Subtype = Cardioembolic 2.725 4.685 0.987 −12.24 −1.222
Subtype = Cryptogenic −1.310 0.798 −1.758 −14.46 −4.119
Gender (female) −3.200 −3.233 −3.076 −4.001 −3.098
Age −19.60 −18.64 −16.43 −18.81 −13.25
NIHSS score at admission −9.696 −9.376 −8.096 −6.668 −8.224
Hypertension −29.73 −30.99 −31.28 −31.05 −30.68
Hyperlipidemia 2.869 2.598 2.882 2.735 2.670
Diabetes −4.155 −4.010 −2.869 −3.472 −3.672
Smoking 4.538 4.680 4.029 4.353 2.530
Alcohol problem −4.463 −5.363 −3.866 −4.179 −1.675
Previous history of stroke −2.316 −2.018 −1.045 −1.331 −0.469
Atrial fibrillation 0.810 1.133 2.010 3.047 2.038
Carotid artery disease −0.038 0.889 0.487 1.468 2.168
Congestive heart failure −0.162 −0.007 −0.656 −1.405 −1.294
Peripheral artery disease −1.909 −1.444 −0.274 −1.458 −2.282
Hemorrhagic conversion −0.327 −1.130 −1.952 −1.029 −1.737
tPA −0.641 −0.860 −0.565 −1.374 −1.289
Statins −2.597 −2.323 −2.715 −2.057 −2.641
Antihypertensives −1.595 −0.875 −1.052 −0.222 −1.340
Antidiabetics 0.030 0.658 0.478 1.194 −0.246
Antiplatelets 1.397 0.515 0.182 0.129 −0.244
Anticoagulants −1.725 −1.295 −1.897 −1.797 −2.735
Perfusion 4.584 3.587 2.132 3.163 2.693
Neurointervention −5.795 −6.002 −5.304 −6.513 −5.680
mRS before admission −1.994 −1.222 −1.274 −0.636 −1.109
mRS at discharge −3.754 −1.954 −1.133 −0.561 0.042
Regression, Classification and Ensemble Machine Learning Approaches 397
Table 10. Performance of ordinal regression on stroke data for different link functions. For AIC,
smaller values indicate better performance whereas for log-likelihood, larger values (lower
magnitude of negative values) indicate better performance.
Link function AIC Log-likelihood
Logit 651.96 −292.98
Probit 654.17 −294.09
Cloglog 663.71 −298.85
Cauchit 685.53 −309.77
Table 11. Ordinal regression (using the logit link function) coefficients for each attribute of the
stroke data. The p-values are shown in the rightmost column, and the attributes with p < 0.05 are
marked with asterisks.
Attribute Coefficient p-value
Subtype = Large vessel 0.438 0.461
Subtype = Small vessel −0.259 0.738
Subtype = Cardioembolic −0.062 0.918
Subtype = Cryptogenic 0.356 0.550
Gender 0.343 0.235
Age 4.164 0.001*
NIHSS score at admission 0.280 0.723
Hypertension −0.545 0.201
Hyperlipidemia 0.118 0.730
Diabetes 0.426 0.289
Smoking −0.396 0.199
Alcohol problem 0.407 0.297
Previous history of stroke 0.610 0.073
Atrial fibrillation 0.242 0.526
Carotid artery disease 0.378 0.319
Congestive heart failure −0.037 0.939
Peripheral artery disease 0.553 0.253
Hemorrhagic conversion −0.123 0.739
tPA 0.101 0.747
Statins −0.096 0.776
Antihypertensives 0.388 0.289
Antidiabetics −0.096 0.832
Antiplatelets −0.302 0.164
Anticoagulants 0.347 0.406
Perfusion −0.767 0.096
Neurointervention −0.552 0.153
mRS before admission 0.269 0.089
mRS at discharge 2.002 <0.001*
398 A. Kabir et al.
(p < 0.05) impact. Other attributes having relatively strong impact according to the
model are mRS before admission, history of previous stroke, and perfusion. The first
two contribute to higher mRS scores whereas the last one contributes to lower mRS
scores.
This paper has presented the results of predicting the 90-day outcome of ischemic
stroke patients based on the data consisting of their demographics, medical history and
treatment records. The target attribute for prediction is the modified Rankin Scale score
90 days after stroke. The problem of prediction was approached from several angles: as
a regression problem, as a multinomial unordered classification problem, and as an
ordinal regression problem. For regression, a meta-learning approach of bootstrap
aggregating (bagging) using M5 model trees as the base learner proved to be very
effective, significantly outperforming the standard linear regression method. The same
method, after translation of the target output from numeric to nominal, works suc-
cessfully as a multinomial classification scheme, and significantly outperforms logistic
regression, the most commonly used classification method in clinical trials. We also
experimented with ordinal regression using different link functions, and obtained a
useful model using the logit link function.
A model tree divides the input space into a number of subspaces defined around the
attributes represented by internal nodes of the tree, and builds a model for each sub-
space. In our dataset, mRS at discharge proved to be an important attribute dominating
the final outcome. Therefore, predictive algorithms that construct specialized models
for different ranges of mRS score at discharge are likely to do well. This is exactly what
we observed from our M5 model trees obtained by a single iteration of the M5 algo-
rithm, as well as the many trees obtained by multiple iterations of bagging. By
examining the model tree prediction errors for the stroke dataset considered, it is found
that the variability of errors is much higher for model trees than for other regression
methods such as linear regression. Since bagging is empirically known to reduce the
instability and error variance of its base learners, it shows good performance for this
particular dataset.
Through our experiments we obtained several predictive models that can take
inputs in the form of patient data, calculate some function, and produce output in the
form of the patient’s 90-day stroke outcome. The regression models’ output is a
continuous real number estimating the numeric mRS score, and the classification
models’ output is a nominal value corresponding to the predicted category of the mRS
score. For regression we analyzed the tree induced by the M5 model tree algorithm. We
found that the internal nodes of the tree simply “direct” the decision towards linear
models (at the leaves) based on the mRS score at discharge. By examining the indi-
vidual linear models, we were able to identify different attributes that were associated
with either improved or worsened outcome of stroke patients. For example, older age,
more severe initial stroke, and presence of hypertension were associated with a poorer
90-day outcome even when their condition at the time of discharge was relatively good.
Regression, Classification and Ensemble Machine Learning Approaches 399
On the other hand, Neurointervention was found to be a procedure that was associated
with improved outcome of patients who were in relatively poor condition at the time of
discharge. We also analyzed the ten different model trees constructed when the bagging
technique was applied to M5 model trees. Some of these trees were identical to the
single tree discussed above, but there was a considerable amount of differences among
the other trees. However, mRS at discharge was found to be the most important
attribute in all of the trees. The other regression model that we observed was that of
linear regression, which comes in the form of a linear equation. We analyzed the
coefficients of the equation and found that, according to the model, older age, more
severe initial stroke, higher mRS at discharge, and alcohol were correlated with poorer
outcome, while the Perfusion procedure was correlated with better outcome.
As for the classification part of our study, we obtained models induced by the C4.5
Decision Tree and Logistic Regression algorithms. We also performed ordinal logistic
regression and obtained a linear equation of the attributes that can be used to predict
stroke outcome. The contributions of attributes towards a better or a worse outcome in
these models were found to be similar to the ones observed in the regression models.
To summarize some of the important observations made from both the regression and
the classification models, we found that higher mRS score at the time of discharge,
older age and more severe initial stroke are the most influencing factors that are
associated with worsened stroke outcome. The presence of diabetes and hypertension
also have some correlation with worse outcome. On the other hand, treatment proce-
dures such as Perfusion and Neurointervention were shown to be associated with
improved outcome of the patients.
There is room for future work to take this study forward. One limitation of the study
is the exclusion of the patients who died within 90 days of stroke. As mentioned before,
this is in line with other work in the literature (e.g., the Copenhagen Stroke Study [33]),
but it would be interesting in future work to extend our approach to include these
patients. We are also limited by a large amount of missing values in attributes that are
part of the medical records but are not included in this study. By collecting data about
more patients, and more data about each patient, we may be able to uncover some more
useful patterns. Another future goal is to improve the process of classification via
regression by discovering better ways to translate the numeric predictions to discrete
classes. Future work includes also an in-depth analysis of the ordinal regression results
and models that we obtained.
Acknowledgements. The authors thank Prof. Dr. Klaus Brinker for suggesting using ordinal
regression as an additional technique in this research.
References
1. Raffeld, M., Debette, S., Woo, D.: International stroke genetics consortium update. Stroke 47
(4), 1144–1145 (2016). https://doi.org/10.1161/STROKEAHA.116.012682
2. Mozaffarian, D., Benjamin, E.J., Go, A.S., et al.: Heart disease and stroke statistics—2016
update. Circulation 133(4), e38–e360 (2016). https://doi.org/10.1161/CIR.00000000000
00350
400 A. Kabir et al.
3. Rankin, J.: Cerebral vascular accidents in patients over the age of 60: II. Prognosis. Scott.
Med. J. 2, 200–215 (1957)
4. Kabir, A., Ruiz, C., Alvarez, S.A., Moonis, M.: Predicting outcome of ischemic stroke
patients using bootstrap aggregating with M5 Model trees. In: van den Broek, E.L., Fred, A.,
Gamboa, H., Vaz, M. (eds.) 10th International Joint Conference on Biomedical Engineering
Systems and Technologies, BIOSTEC 2017, Proceedings Volume 5: HealthInf, 21–23
February, Porto, Protugal, pp. 178–187 (2017)
5. van Swieten, J.C., Koudstaal, P.J., Visser, M.C., Schouten, H.J., van Gijn, J.: Interobserver
agreement for the assessment of handicap in stroke patients. Stroke 19(5), 604–607 (1988)
6. Banks, J.L., Marotta, C.A.: Outcomes validity and reliability of the modified rankin scale:
implications for stroke clinical trials: a literature review and synthesis. Stroke 38(3), 1091–
1096 (2007)
7. Bruno, A., Shah, N., Lin, C., Close, B., Hess, D.C., Davis, K., Baute, V., Switzer, J.A.,
Waller, J., Nichols, F.T.: Improving modified rankin scale assessment with a simplified
questionnaire. Stroke 41(5), 1048–1050 (2010). https://doi.org/10.1161/STROKEAHA.109.
571562
8. Tan, P.-N., Steinbach, M., Kumar, V.: Introduction to Data Mining. Addison-Wesley,
Boston (2005)
9. Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and Regression Trees.
Wadsworth, Belmont (1984)
10. Quinlan, J.R.: Learning with continuous classes. In: Adams, A., Sterling, L. (eds.) 5th
Australian Joint Conference on Artificial Intelligence, Hobart, Australia, pp. 343–348 (1992)
11. Wang, Y., Witten, I.H.: Induction of model trees for predicting continuous classes. In: van
Someren, M., Widmer, G. (eds.) 9th European Conference on Machine Learning. University
of Economics, Prague (1996)
12. Etemad-Shahidi, A., Mahjoobi, J.: Comparison between M5′ model tree and neural networks
for prediction of significant wave height in Lake Superior. Ocean Eng. 36(15–16), 1175–
1181 (2009). https://doi.org/10.1016/j.oceaneng.2009.08.008
13. Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)
14. Aslam, J., Popa, R., Rivest, R.: On estimating the size and confidence of a statistical audit.
USENIX/ACCURATE Electron. Voting Technol. Workshop 7, 8 (2007)
15. McCullagh, P.: Regression models for ordinal data. J. R. Stat. Soc.: Ser. B (Methodol.) 42
(2), 109–142 (1980)
16. Herbrich, R., Graepel, T., Obermayer, K.: Regression models for ordinal data: a machine
learning approach. Technische Universität Berlin, Fachbereich 13, Informatik, Berlin (1999)
17. Anderson, J.: Regression and ordered categorical variables. J. R. Stat. Soc.: Ser.
B (Methodol.) 46(1), 1–30 (1984)
18. Agresti, A.: An Introduction to Categorical Data Analysis. Wiley-Interscience, Hoboken
(2007)
19. Bender, R., Grouven, U.: Ordinal logistic regression in medical research. J. R. Coll. Phys.
Lond. 31(5), 546–551 (1997)
20. Moonis, M., Kane, K., Schwiderski, U., Sandage, B.W., Fisher, M.: HMG-CoA reductase
inhibitors improve acute ischemic stroke outcome. Stroke 36(6), 1298–1300 (2005)
21. Marini, C., De Santis, F., Sacco, S., Russo, T., Olivieri, L., Totaro, R., Carolei, A.:
Contribution of atrial fibrillation to incidence and outcome of ischemic stroke. Stroke 36(6),
1115–1119 (2005)
22. Yong, M., Kaste, M.: Dynamic of hyperglycemia as a predictor of stroke outcome in the
ECASS-II trial. Stroke 39(10), 2749–2755 (2008). https://doi.org/10.1161/STROKEAHA.
108.514307
Regression, Classification and Ensemble Machine Learning Approaches 401
23. Nogueira, R.G., Liebeskind, D.S., Sung, G., Duckwiler, G., Smith, W.S.: Multi MERCI
writing committee: predictors of good clinical outcomes, mortality, and successful
revascularization in patients with acute ischemic stroke undergoing thrombectomy: pooled
analysis of the mechanical embolus removal in cerebral ischemia (MERCI) and multi
MERCI trials. Stroke 40(12), 3777–3783 (2009). https://doi.org/10.1161/STROKEAHA.
109.561431
24. Henninger, N., Lin, E., Baker, S.P., Wakhloo, A.K., Takhtani, D., Moonis, M.:
Leukoaraiosis predicts poor 90-Day outcome after acute large cerebral artery occlusion.
Cerebrovascular Diseases. 33(6), 525–531 (2012). https://doi.org/10.1159/000337335
25. Kissela, B., Lindsell, C.J., Kleindorfer, D., Alwell, K., Moomaw, C.J., Woo, D., Flaherty,
M.L., Air, E., Broderick, J., Tsevat, J.: Clinical prediction of functional outcome after
ischemic stroke: the surprising importance of periventricular white matter disease and race.
Stroke 40(2), 530–536 (2008). https://doi.org/10.1161/STROKEAHA.108.521906
26. Optimising Analysis of Stroke Trials Collaboration: Calculation of sample size for stroke
trials assessing functional outcome: comparison of binary and ordinal approaches. Int.
J. Stroke 3(2), 78–84 (2008). https://doi.org/10.1111/j.1747-4949.2008.00184.x
27. Sandset, E.C., Bath, P.M., Boysen, G., Jatuzis, D., Kõrv, J., Lüders, S., Murray, G.D.,
Richter, P.S., Roine, R.O., Terént, A., Thijs, V., Berge, E.: SCAST study group: the
angiotensin-receptor blocker candesartan for treatment of acute stroke (SCAST): a
randomised, placebo-controlled, double-blind trial. Lancet 377(9767), 741–750 (2011).
https://doi.org/10.1016/S0140-6736(11)60104-9
28. Gialanella, B., Santoro, R., Ferlucci, C.: Predicting outcome after stroke: the role of basic
activities of daily living. Eur. J. Phys. Rehabil. Med. 49(5), 629–637 (2013)
29. Granger, C.V., Hamilton, B.B., Keith, R.A., Zielezny, M., Sherwin, F.S.: Advances in
functional assessment for medical rehabilitation. Top. Geriatr. Rehabil. 1(3), 59–74 (1986)
30. Brown, A.W., Therneau, T.M., Schultz, B.A., Niewczyk, P.M., Granger, C.V.: Measure of
functional independence dominates discharge outcome prediction after inpatient rehabili-
tation for stroke. Stroke 46(4), 1038–1044 (2015). https://doi.org/10.1161/STROKEAHA.
114.007392
31. Adams Jr., H.P., Bendixen, B.H., Kappelle, L.J., Biller, J., Love, B.B., Gordon, D.L., Marsh
3rd, E.E.: Classification of subtype of acute ischemic stroke. Definitions for use in a
multicenter clinical trial. TOAST. Trial of Org 10172 in Acute Stroke Treatment. Stroke 24
(1), 35–41 (1993)
32. Brott, T., Adams Jr., H.P., Olinger, C.P., Marler, J.R., Barsan, W.G., Biller, J., Spilker, J.,
Holleran, R., Eberle, R., Hertzberg, V., et al.: Measurements of acute cerebral infarction: a
clinical examination scale. Stroke 20(7), 864–870 (1989)
33. Nakayama, H., Jorgensen, H.S., Raaschou, H.O., Olsen, T.S.: The influence of age on stroke
outcome. The Copenhagen Stroke Study. Stroke 25(4), 808–813 (1994)
34. Frank, E., Hall, M.A., Witten, I.H.: The WEKA Workbench. Online Appendix for “Data
Mining: Practical Machine Learning Tools and Techniques”. Morgan Kaufmann, Burlington
(2016)
35. Kohavi, R.: A study of cross-validation and bootstrap for accuracy estimation and model
selection. In: Proceedings of the International Joint Conference on Artificial Intelligence
(IJCAI), vol. 2, pp. 1137–1145 (1995)
36. Lee Rodgers, J., Nicewander, W.A.: Thirteen ways to look at the correlation coefficient. Am.
Stat. 42(1), 59–66 (1988)
37. Moore, D.S., Notz, W.I., Fligner, M.A.: The Basic Practice of Statistics. W.H. Freeman and
Co., New York (2015)
38. Quinlan, J.R.: C4.5 - Programs for Machine Learning. Morgan Kaufmann, San Francisco
(1992)
402 A. Kabir et al.
39. Christensen, R.H.B.: Regression Models for Ordinal Data [R package ordinal version
2015.6-28]. https://cran.r-project.org/web/packages/ordinal/index.html
40. Bozdogan, H.: Model selection and Akaike’s Information Criterion (AIC): the general theory
and its analytical extensions. Psychometrika 52(3), 345–370 (1987)
41. Willmott, C.J., Matsuura, K.: Advantages of the mean absolute error (MAE) over the root
mean square error (RMSE) in assessing average model performance. Clim. Res. 30(1), 79–
82 (2005). https://doi.org/10.3354/cr030079
Author Index