S R M I T: Ecurity ISK Anagement For The Nternet of Hings
S R M I T: Ecurity ISK Anagement For The Nternet of Hings
JOHN SOLDATOS
(Editor)
Published, sold and distributed by:
now Publishers Inc.
PO Box 1024
Hanover, MA 02339
United States
Tel. +1-781-985-4510
www.nowpublishers.com
sales@nowpublishers.com
ISBN: 978-1-68083-682-0
E-ISBN: 978-1-68083-683-7
DOI: 10.1561/9781680836837
Suggested citation: John Soldatos (ed.). (2020). Security Risk Management for the Internet of Things.
Boston–Delft: Now Publishers
The work will be available online open access and governed by the Creative Commons
“Attribution-Non Commercial” License (CC BY-NC), according to https://creativecommons.org/
licenses/by-nc/4.0/
Table of Contents
Foreword xi
Preface xv
Glossary xxi
Chapter 1 Introduction 1
By John Soldatos
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Overview and Limitations of Security Risk Assessment
Frameworks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.1 Overview of Security Risk Assessment . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.2 Limitations of Security Risk Assessment Frameworks for IoT . . . . . 7
1.3 New Technology Enablers and Novel Security Concepts . . . . . . . . . . 9
1.3.1 IoT Security Knowledge Bases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3.2 IoT Reference Architectures and Security Frameworks . . . . . . . . . . . 10
1.3.3 Blockchain Technology for Decentralized Secure Data Sharing
for Security in IoT Value Chains . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.3.4 Technologies Facilitating GDPR Compliance . . . . . . . . . . . . . . . . . . 12
1.3.5 Machine Learning and Artificial Intelligence Technologies for
Data-driven Security . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
1.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
iii
iv Table of Contents
Epilogue . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 247
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
About the Editor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 252
Contributing Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253
Foreword
Over twenty years following the introduction of the term Internet of Things (IoT)
by Kevin Ashton at MIT, IoT technologies are growing at a rapid pace. In-line
with early visions about IoT applications, IoT technologies are currently enabling
advanced services that interconnect physical and virtual things based on interop-
erable communication technologies. Most importantly, IoT applications are no
longer limited to small scale laboratory environments, as they are part of robust
enterprise deployments. Consumers are already seeing IoT applications improving
their lives and making the world a better place to live. For instance, thousands
of internet-connected devices are currently optimizing resource usage and improv-
ing environmental performance in modern smart cities. As another example, IoT
is a key enabler of advanced automation in industrial applications in sectors like
manufacturing, energy, and supply chain management. During the past weeks we
are also witnessing the importance of IoT technologies in fighting the COVID19
outbreak, through facilitating processes like diagnostic testing, contact tracing and
disease spreading estimation.
These developments are however the start of the IoT journey rather than its
destination. In several cases, state of the art IoT deployments have just scratched
the surface of potential use cases and there is still a long way to go to realize the
full potential of the IoT paradigm. The latter will be shaped by recent advances
in technologies like Artificial Intelligence (AI), big data, robotics, blockchain and
5G. These technologies are expected to enable an even broader and progressing
revolution in the future, across a wide variety of verticals, including healthcare,
smart cities, energy, agriculture, and industry. Specifically, they will enable next
generation IoT applications that will employ advanced and distributed computing
to bring intelligence and automation at the point of action. Emerging IoT systems
like connected and autonomous vehicles will be therefore able to take faster, more
xi
xii Foreword
intelligent, and more efficient decisions at the right place and at the right point
in time.
In this landscape, the number of IoT devices is rising constantly with an expected
40 billion IoT devices to be in used worldwide by 20251 , including not only passive
devices, but also semi-autonomous smart objects. Nevertheless, the rising sophisti-
cation of IoT systems, technologies and applications is becoming associated with a
more complex IoT ecosystem. This ecosystem is characterised by a combination of
very diverse products, systems, and services, which rely on data and connectivity to
deliver their value.
Delivering real value in the scope of the complex IoT ecosystem is primarily
about addressing specific and pragmatic challenges, as needed for delivering appli-
cations that can be integrated into society. These applications must be part of an
open, predictable, and competitive IoT market where individual rights and free-
doms are respected. When it comes to addressing pragmatic challenges, the cyber-
security aspects of IoT come into the foreground. Specifically, security, privacy and
data protection aspects cannot be negotiated when it comes to deploying real-life
IoT applications that are ethical, trustworthy and protect the citizens’ rights. In
this context, the rising complexity of IoT technologies and ecosystems puts into
question conventional methodologies for risk assessment and traditional regula-
tory paradigms. In fact, IoT applications are characterised by a very large attack
surface and complex ecosystem consisting of diverse actors, heterogeneous physical
and virtual spaces, as well as devices of different size, nature, and complexity. In this
multi-actor, multi-stakeholder environment, cyber risks can mean different things
to different actors, and hence, it is not always possible to have a single approach for
confronting them. Furthermore, the complexity and fragmentation of applicable
legal frameworks for safety and liability, at both EU and national levels, proves to
be another significant challenge. Advanced digital technologies can become part of
the solutions to these challenges, as they provide powerful testing and validation
capabilities for managing cyber-security in complex environments.
The development of secure, trustworthy, and human-centric IoT systems and
applications is a top priority for the European Commission. A human-centric and
trustworthy approach is necessary to facilitate the uptake of digital solutions among
European citizens. This approach reflects the EU’s specific way and vision of human
progress and evolution. Ensuring a cybersecure IoT is an essential foundation of
this approach and vision. In Europe, any IoT solution must be compliant with the
1. https://www.statista.com/statistics/802690/worldwide-connected-devices-by-access-technology/
Foreword xiii
“Human Centred” approach. It is also very positive that this book is offered as an
Open Access publication, which could help it reach a wider readership and boost its
overall impact. Finally, it is our hope and expectation that this book will be proven
an effective resource and a good reading experience for the IoT community.
June 2020
Salvatore Scalzo
EC Policy Officer, “Internet of Things” Unit of DG CONNECT
IoT Security and Privacy Cluster Coordinator
Franck Boissiere
EC Programme Officer, “Internet of Things” Unit of DG CONNECT
IoT Security and Privacy Cluster Coordinator
Preface
In recent years, Internet of Things (IoT) systems have rapidly evolved in terms of
functional and technological sophistication. Early small-scale sensing systems and
wireless sensor networks have given their place to massively scalable, cloud-based
systems that comprise many thousands of internet-connected objects. At the same
time, the advent of edge computing has provided opportunities for advanced inter-
net of things deployments that process and analyze data in real time, while closing
the loop to the field and influencing the status of the physical world. Moreover, we
are gradually witnessing the rise of smart objects with semi-autonomous behavior
(e.g., drones, robots, automated guided vehicles), which intensify the decentraliza-
tion and the intelligence of edge/cloud architectures.
This rising complexity of IoT systems provides opportunities for more auto-
mated and intelligent business applications. However, it also introduces new cyber-
security challenges such as new ways for conducting large-scale attacks and a wealth
of vulnerabilities and risks at different parts of an IoT system such as smart objects,
IoT networks, edge gateways, and cloud elements. These vulnerabilities are evident
in the scope of recent large-scale cybersecurity incidents, such as the Mirai malware
that turned IoT devices into remotely controlled bots able to launch Distributed
Denial of Service (DDoS) attacks. The Mirai-based based DDoS back in 2016 took
advantage of vulnerabilities (e.g., hard-coded passwords, poorly patched software)
of internet-connected CCTV (Closed Circuit Television) cameras and DVR (Dig-
ital Video Recorders). As another example, the “Lizard Stressor” attacks few years
ago compromised many commercial home routers at a large scale.
In addition to confronting such attacks, developers, and operators of IoT sys-
tems must comply with stringent regulatory requirements, such as requirements
xv
xvi Preface
stemming from the General Data Protection Regulation (GDPR) and the Net-
work Information Systems (NIS) directive. To this end, there is a need for deploy-
ing effective security and data protection methods that boost compliance to the
security, privacy, and data protection mandates of these regulations.
In this context, security risk management methods can be a powerful tool for
developers, solution integrators, and operators of IoT systems. International stan-
dards and frameworks for risk management can be used to support the identifi-
cation of risks or threats, and to assess their respective probabilities. Nevertheless,
state-of-the-art technologies for security risk assessment have prominent limitations
when it comes to confronting risks associated with large-scale, cyber-physical, and
interconnected IoT systems. For example, risk assessments for modern IoT sys-
tems must be more frequent and must take into account knowledge about both
cyber and physical assets. Furthermore, they should be more proactive and auto-
mated, and able to leverage information that is shared across IoT supply chains.
In this direction, organizations can take advantage of emerging technologies (e.g.,
edge/fog computing architectures, machine learning, blockchain), to implement
novel risk assessment approaches that are characterized by automation, intelligence,
and transparency.
During the last couple of years, a group of European Commission (EC) funded
projects have been researching and implementing novel security and privacy mech-
anisms for the Internet of Things, including risk assessment and mitigation. The
projects have formed a Cluster of IoT Security and Privacy projects, as a means of
boosting their collaboration. The purpose of the present book is to present detailed
information about the novel risk assessment techniques that have been developed
by these projects and their role in the IoT security risk management process. Specif-
ically, the book presents several architectures and platforms for end-to-end security,
including their implementation based on the edge/fog computing paradigm. It also
highlights machine learning techniques that boost the automation and proactive-
ness of IoT security risk assessments. Furthermore, blockchain solutions for open
and transparent sharing of IoT security information across the supply chain are
introduced. Moreover, several chapters of the book present frameworks for privacy
awareness, along with technical measures that enable privacy risk assessment and
boost GDPR compliance. Likewise, techniques for security certification of IoT sys-
tems along with frameworks for IoT security interoperability towards end-to-end
security are discussed.
Overall, the book is structured into fifteen chapters that present different IoT
security technologies, including novel techniques for IoT security risk assessment.
Specifically:
Preface xvii
Overall, the book introduces novel IoT security technologies with emphasis on
solutions for IoT risk assessment. It also presents future trends in IoT Security,
including how novel digital technologies (e.g., Machine Learning, AI, blockchain)
can be used to secure emerging IoT systems. Furthermore, it provides insights on
privacy and data protection solutions, notably solutions that can support organi-
zations in their regulatory compliance efforts. The book is made available as an
Open Access publication, which could make it broadly and freely available to the
IoT community. I would like to thank Now Publishers for the opportunity and
their collaboration in making this happen.
Most importantly, I take the chance to thank the contributing projects for their
valuable inputs and contributions in developing the presented IoT security tech-
nologies, as well as in documenting them as part of the book. I would also like
to acknowledge funding and support from the European Commission as part of
the H2020 SecureIoT, Brain-IoT, CHARIOT, ENACT, IoT Crawler, NGIOT,
PDP4E, SEMIoTICS, SerIoT, and SOFIE projects.
March 2020
John Soldatos
Glossary
Symbols
API - Application Programming Interfaces. 109, 110, 116, 117, 128, 135, 197, 203,
224–226
xxi
xxii Glossary
ARD - Automated Responsible Disclosure. 229, 232, 234, 235, 237, 239, 243
ARM - Advanced RISC (Reduced Instruction Set Computing) Machine. 226
ASCII - American Standard Code for Information Interchange. 222
AVD - Automated Vulnerability Detector. 144, 148, 149
B
BIN - Binary. 204
BISMON - . 199–201
BMS - . 111, 112
BRAIN-IoT - model-Based fRamework for dependable sensing and Actuation in
INtelligent decentralized IoT systems. 208, 215
C
CA - Certificate Authorities. 221, 224
CAM - Compliance Auditing Manager. 27
CAN - Controller Area Network. 54, 55
CAP - Cybersecurity Assurance Program. 182
CAPEC - Common Attack Pattern Enumeration and Classification. 9, 25, 33
CAS - Compliance Auditing Service. 26, 27
CC - Common Criteria. 180, 182
CCRA - Common Criteria Recognition Arrangement. 182
CCTV - Closed Circuit Television. xv, 2
CDPA - Consumer Data Protection Act. 18
CERT - Computer Emergency Response Team. 9, 37, 175
CERT/CC - Computer Emergency Response Team/Coordination Center. 236
CFG - Control Flow Graph. 204
CHARIOT - . 196–202, 204, 206, 207, 220–227
CLI - Command Line Interface. 223, 225
Glossary xxiii
D
DARPA - Defense Advanced Research Projects Agency. 53
DBN - Deep Belief Network. 52
xxiv Glossary
E
EAL - Evaluation Assurance Levels. 182, 191
EAL4+ - . 183
EAP - Extensible Authentication Protocol. 94
EC - European Commission. xvi
ECDHE - Elliptic Curve Diffie-Hellman Ephemeral. 214
ECSO - European Cyber Security Organisation. 180, 191
ECSO WG1 - . 181
ELF - Executable and Linkable Format. 204
EMALCSA - . 82
ENISA - . 93, 179, 180
ETAP - . 81
Glossary xxv
GDPR - General Data Protection Regulation. xvi–xviii, 5, 12–14, 18, 26, 69–79,
81, 84, 85, 144, 146, 153, 157, 162–166, 168–173, 175, 177, 183, 235,
247, 248
GHOST - . 163
IT - Information Technology. 2, 3, 6, 8, 9, 12, 18, 20, 49, 51, 183, 208, 209, 211,
212, 247
LL - . 117
LPWAN - Low Power Wide Area Network. 209–211, 213, 214, 218
NGIoT - . 177
NIST - National Institute of Standards and Technology. 7, 25, 26, 33, 37, 90, 91,
93, 179
OMA - . 94
OS - . 116, 203
PIA - Privacy Impact Assessment. xvii, 69, 71–75, 79, 81, 84, 85
PKI - Public Key Infrastructure. 111, 113, 119, 213, 217, 220, 221, 223–225
QoS - Quality of Service. 89, 90, 96, 123, 130, 133, 134
QT - . 54, 61–63
SerIoT - Secure and Safe Internet of Things. 88, 89, 95, 99, 100
T
TITAN - . 189, 190
TLS - Transport Layer Security. 94, 213, 214
TMUC - Trustworthiness Metrics and Utility Calculation. 25
TOE - Target of Evaluation. 182, 185, 186, 188–192
TPM - Trusted Platform Modules. 213
TSIS - . 143, 144, 146–148, 153, 156–158
TTCN3 - Testing and Test Control Notation Version 3. 189, 190
TTP - Trusted Third Party. 10, 11
TXT - Text File. 176
U
UAV - Unmanned Aerial Vehicle. 50
UID - . 205
UK - United Kingdom. 182
UL - Underwriters Laboratories. 182
ULD - . 183
UML - Unified Modeling Language. 73, 74, 189
UPnP - Universal Plug and Play. 127
URL - Uniform Resource Locator. 94, 201
xxxii Glossary
V
V2X - Vehicle-to-everything. 54
VAE - Variational Autoencoder. xvii, 52, 53, 55, 65, 248
VIM - Virtual Infrastructure Manager. 136
VNF - Virtual Network Function. 135, 136
W
W3C - World Wide Web Consortium. 127, 235
WiFi - Wireless Fidelity. 72
WP29 - Working Party article 29. 164, 165, 171
X
XACML - eXtensible Access Control Markup Language. 25
XLS - Excel Spreadsheet. 176
XLSX - Excel Microsoft Office Open XML Format Spreadsheet file. 176
XML - eXtensible Markup Language. 9, 19, 20, 28, 47, 176, 189
XMPP - eXtensible Messaging and Presence Protocol. 127
DOI: 10.1561/9781680836837.ch1
Chapter 1
Introduction
By John Soldatos
This chapter is the introduction to the book. It presents the main security and
privacy challenges for modern IoT systems, along with state-of-the-art frameworks
for security risk assessment and mitigation. Accordingly, it discusses the limitations
of these frameworks when it comes to implementing risk management systems for
modern IoT systems. Moreover, it outlines some of the enabling technologies and
security concepts that could help organizations alleviate these limitations. Some
of these concepts are driving the security and technologies that are introduced in
subsequent chapters.
1.1 Introduction
In recent years, we have witnessed the rise of the Internet of Things (IoT) paradigm,
which is empowered and propelled by the proliferating number of internet-
connected devices. The latter enable a wide range of innovative IoT applications in
sectors like trade, transport, healthcare, and industry. A main characteristic of these
applications is their ability to take intelligent decisions based on the collection and
processing of large amounts of data from the physical world. In several cases, these
decisions involve actuation and control activities that influence the status of the
physical world and the surrounding environment of the IoT systems.
1
2 Introduction
For over a decade, IoT systems have been evolving in functional and technolog-
ical sophistication. Early IoT applications were based on the collection and pro-
cessing of physical world data based on sensors, wireless sensor networks, and other
types of internet-connected devices. This early paradigm has gradually evolved in
terms of scalability, thanks to the integration of IoT systems and devices with cloud
computing infrastructures. The latter infrastructures have also enabled IoT appli-
cations to benefit from the elasticity, flexibility, and quality of service of the cloud,
which has empowered the development of larger scale applications. Nevertheless,
this scalability was based on a heavily centralized model, where very large num-
bers of internet-connected devices integrate their data and services in the cloud.
In recent years, this model has been decentralized based on the emergence and
expanded deployment of edge computing architectures and Cyber-Physical Sys-
tems (CPS), which include various smart objects like drones, smart wearables,
and autonomous guided vehicles. On the one hand, edge computing deployments
decentralize IoT functions by placing them closer to the field, while on the other
smart objects exhibit (semi) autonomous behaviors that can be implemented inde-
pendently from the cloud. Hence, emerging IoT applications feature decentralized
intelligence, which can be split across the cloud, edge nodes, and smart objects.
Edge computing and CPS systems are two of the main pillars of the fourth indus-
trial revolution (Industry 4.0), which closes the loop to the field and enables the
digital control of physical processes in a way that blurs the boundaries between
the physical and the cyber worlds. Likewise, Industry 4.0 is enabling the conver-
gence of Information Technology (IT) with the Operational Technology (OT) [e.g.,
Supervisory Control And Data Acquisition (SCADA) systems, DCS (Distributed
Control Systems), Programmable Logic Controllers (PLC)] that is widely used in
industrial settings.
However, this increased sophistication of IoT systems has introduced a host of
security challenges as well, including:
• New ways for large-scale security attacks: The proliferation of IoT devices
and their deployment has opened new cybersecurity vulnerabilities and
enabled new ways of conducting large-scale attacks. As a prominent exam-
ple, in October 2016, we witnessed the first large-scale distributed denial
of service (DDoS) attack based on IoT devices, which took advantage
of vulnerabilities (e.g., hard-coded passwords, poorly patched software) of
internet-connected CCTV (Closed Circuit Television) cameras and DVR
(Digital Video Recorders), in order to deploy the notorious Mirai malware
on them [1]. Likewise, back in January 2015, the “Lizard Stressor” attacks
compromised many commercial home routers at a large scale.
Introduction 3
• Identifying the assets that are most valuable to the company. In the case of
IoT security risk assessment, the assets include IoT devices and IoT services.
6 Introduction
• Determining the potential threats to each asset. This step leverages infor-
mation about the IoT asset (e.g., information collected from a device via an
appropriate probe).
• Estimating the likelihood that events associated with the threat will
occur. Usually, a probability estimate is produced based on some determin-
istic or stochastic model about the asset and the attacker’s behavior.
• Determining the impact criticality of each loss event. Events that could
have a catastrophic impact on the assets and the organization should be
avoided as a matter of priority.
• Calculate risk scores based on combinations of information about events’
probabilities and their impact criticality.
In-line with the above-listed steps, risk management methods rely on the estima-
tion of a risk situation for the various assets. This estimation is based on information
about the processes, the assets, and the infrastructure of an organization. The main
goal of the process is the identification of potential risks and the development
of appropriate protective measures. The tangible outcome of a risk analysis is in
most cases a list of risks or threats to a system, together with the corresponding
probabilities.
International standards in the field of risk management are used to support
the identification of risks or threats, and to assess their respective probabilities.
These standards range from general considerations and guidelines for risk man-
agement processes (e.g., [2, 3]) to specific guidelines for the IT sector such as the
ISO20000 [4], ISO27000 [5], ISO/IEC 27002 [6], ISO27005 [7], and others
(e.g., [8]), all the way to highly specific frameworks. Most of these standards spec-
ify framework conditions for the risk management process, but rarely go into detail
on specific methods for the risk analysis or risk assessment. This is one reason why
often differences in the risk assessment arise within the specific areas of an applica-
tion, making a direct comparison of the results of the various frameworks difficult.
In this context, choosing the right method and the right tool for risk analysis
and risk evaluation proves to be complicated. In recent years, various concepts,
algorithms, and tools have evolved from research, specially designed to protect IT
infrastructures and systems. Since their historical background is settled in a business
context, in these methods a quantitative risk assessment is usually performed based
on monetary costs [9, 10]. This is the case with the French EBIOS1 and the UK’s
1. EBIOS stands for Expression des Besoins et Identification des Objectifs de Sécurité – Expression of Needs
and Identification of Security Objectives.
Overview and Limitations of Security Risk Assessment Frameworks 7
CRAMM2 methods, but also with the ISO/IEC 27005:2013 standard [7]. Most
of the risk assessment methods and tools (e.g., most of the ones listed in [11]) use
the commonly known rule of thumb “risk = probability × potential damage” [12].
Depending on the applied method, the terms and scales for the assessment of the
probabilities as well as the potential damage are predefined (such as in the NIST
(National Institute of Standards and Technology) policy [13] or in the Mehari
method [14]). In practice, the selection of a specific risk-assessment tool is based
on practical considerations and depends on how well the present terminology of
the application can be mapped onto the predefined specific terminology of the risk
assessment methodology. In order to structure the process of risk assessment, there
are various attempts to develop models for general risk assessments. For example,
the AURUM system [15] provides a graphical tool for the risk modeling based on
ontologies. It uses a Bayesian approach for determining threat probabilities (which
is also done by the method proposed in [16]). The OCTAVE method [17] is based
on subjectively estimated probabilities and thus can be understood as an a priori
distribution with regards to the Bayesian approach. Also, the CORAS method [18]
allows the integration of several different risk assessment processes, whereas the
identification of the probability of an attack is not done automatically but a priori
to any risk assessment.
Finally, the assessment per se is performed in order to optimize the existing con-
figuration of an operation environment. A variety of mathematical models that
embed different reasoning techniques have been proposed, such as Integer Linear
Programming, Constraint Programming, Stochastic Optimization, Multi-criteria
Optimization, Robust Optimization, Metaheuristics, etc. In the literature, a variety
of optimization and decision support approaches have been reported. They inte-
grate techniques and models from Artificial Intelligence, Operations Research, and
Computational Logics (e.g., [19–21]).
2. CRAMM (CCTA Risk Analysis and Management Method) is a qualitative risk analysis and management
tool developed by UK government’s Central Computer and Telecommunications Agency (OGC since April
2001).
8 Introduction
cyber only) nature. Specifically, IoT systems present unique challenges that cannot
be adequately tackled based on conventional approaches. These include:
SKBs can also collect and store external CTI data sources through available
documents in various formats like JSON (JavaScript Object Notation) and XML
(eXtensible Markup Language). There are several SKBs, including commercial
SKBs from major security vendors and SBKs from standards development bodies
(e.g., the OWASP (Open Web Application Security Project) Security Knowledge
Framework). Nevertheless, these SKBs are general purpose and do not comprise
10 Introduction
suitable security knowledge databases for the IoT assets, which is a setback to
supporting security applications like risk analysis and security management activi-
ties for IoT environments.
In order to enable effective and automated risk assessments for IoT systems,
SKBs should also collect and consolidate information from additional sources, as
means of covering IoT assets. For example, the OVAL (Open Vulnerability and
Assessment Language) specifications enable the description of IoT assets.
• The need for a Trusted Third Party (TTP) that owns and operates the
centralized database infrastructure. However, in IoT supply chains, there is
not always a prominent and accepted TTP.
New Technology Enablers and Novel Security Concepts 11
latency and handling of more transactions per second). The latter benefits stem
from the fact that permissioned blockchains can be decoupled from the compu-
tationally expensive Proof of Work (PoW) mechanisms. Moreover, permissioned
blockchains are often hosted on cloud platforms that have robust cybersecurity
controls across different layers of the technology stack.
The use of permissioned blockchain networks for sharing security information is
not widespread, and security implementations are in early stages. For example, the
authors in [27] describe an early implementation of blockchain-based solution for
sharing security information across financial institutions. There are also blockchain
solutions for information sharing in general, as well as for protecting specific IT
systems and IoT devices (e.g., [26]) and enabling the trusted exchange of data
between them (e.g., [28]).
Later chapters of the book present practical uses of blockchain technology for
privacy and data protection in IoT systems.
technologies that facilitate the tracking and mapping of data workflows in order to
identify incidents like data loss, data theft, and security attacks.
Several chapters of the book illustrate GDPR compliant solutions, along with
methodologies (e.g., privacy by design) for implementing and deploying them in
practical IoT security contexts.
1.4 Conclusion
This chapter has presented some of the most prominent cybersecurity challenges
for modern IoT environments. It has also illustrated how some of the main charac-
teristics of IoT systems are driving these challenges. Furthermore, it has presented
an overview of security risk assessment functionalities, including the standards
and frameworks that support their implementation. State-of-the-art techniques for
security risk assessment have prominent limitations when it comes to confronting
risks associated with large-scale, cyber-physical, and interconnected IoT systems.
Recent attacks against IoT systems (e.g., the Mirai-based DDoS attack) have mani-
fested these limitations and opened the discussion for new approaches. In this direc-
tion, security systems integrators can benefit from a range of novel technologies
14 Introduction
that can boost the timeliness and efficiency of the risk assessment processes, while
dealing with other concerns like GDPR compliance. For example, ML/DL tech-
nologies can detect complex attack patterns in order to accelerate the preparedness
of security teams. Likewise, blockchain techniques can boost the transparency of
data sharing across connected IoT systems. Furthermore, IoT security knowledge
bases can increase the automation of the IoT risk assessment process, by helping
organizations to identify IoT-specific threats and to provide advanced cyber threat
intelligence for IoT systems. Following chapters of the book illustrate a host of novel
techniques for IoT security, which leverage advanced technology enablers such the
ones outlined above.
Acknowledgments
Part of this work has been carried out in the scope of the H2020 SecureIoT project,
which is funded by the European Commission in the scope of its H2020 program
(contract number 779899).
References
[1] John Soldatos, “Why the DDoS attack happened and what can be done to
prevent more episodes”, The Internet of All Things. October, 2016 available
at: https://www.theinternetofallthings.com/why-the-ddos-attack-happened-
10262016/ [Accessed: February 2020].
[2] International Standardization Organization, “ISO 31000: Risk Manage-
ment – Principles and Guidelines”, Geneva, Switzerland, 2009.
[3] International Standardization Organization, “ISO 31010: Risk management –
Risk assessment techniques”, Geneva, Switzerland, 2009.
[4] International Standardization Organization, “ISO 20000: Information Tech-
nology Service Management”, Geneva, Switzerland, 2005.
[5] International Standardization Organization, “ISO 27001: Information Secu-
rity Management System Requirements”, Geneva, Switzerland, 2013.
[6] International Organization for Standardization (ISO)/International Elec-
trotechnical Commission (IEC), “ISO/IEC 27002 Information technology –
Security techniques – Code of practice for information security management”,
2005.
[7] International Standardization Organization, “ISO 27005: Information secu-
rity risk management”, Geneva, Switzerland, 2011.
References 15
Chapter 2
17
18 Security Data Modelling for Configurable Risk Assessment
2.1 Introduction
• The emergence of new types of large-scale security attacks against IoT sys-
tems, such as the large-scale Distributed Denial of Service (DDoS) attacks
that exploit the vulnerabilities of specific IoT devices.
• Vulnerabilities in cyber-physical systems (CPS), which are associated with the
cyber resilience challenges of CPS systems.
• Security threats associated with the interplay between Information Technol-
ogy (IT) and Operational Technology (OT) infrastructures, given the dif-
ferent nature of IT and OT assets and the conflicting IT and OT security
requirements.
• Complex regulatory compliance requirements in a demanding and volatile
landscape, such as the need for compliance to the General Data Protec-
tion Regulation (GDPR) in Europe and the Consumer Data Protection Act
(CDPA) in California, in the U.S.
In order to cope with these challenges, there is a need for end-to-end solutions
and tools that can protect IoT assets from the many different types of attacks. Secu-
rity monitoring and security data analytics is one of the primary functionalities
in this direction. It is also one of the security building blocks of standards-based
IoT architectures, such as the Industrial Internet Reference Architecture (IIRA)
of the Industrial Internet Consortium (IIC). The IIRA has been specified along
with an accompanying security framework, namely the Industrial Internet Security
Framework (IISF) [2]. The IISF specifies monitoring, analysis, and action work-
flows as a core security functionality for IoT systems. Based on such IoT monitoring
Introduction 19
workflows, it is possible to monitor the various devices, systems, and services that
comprise an IoT system in effective and integrated ways. Accordingly, monitoring
can then be a foundation for the identification, assessment, and mitigation of risks,
i.e., for the implementation of risk management functionalities that are an integral
and important element of every non-trivial security policy.
The implementation of data-driven security mechanisms is based on the employ-
ment of advanced analytics of large amounts of security data for the various
assets that comprise the IoT system [3]. These analytics include the application
of machine learning (ML) and artificial intelligence (AI) techniques, which can
enable the extraction of insights about security incidents, vulnerabilities, and risks,
as a means of executing risk assessment functions and boosting the early and proper
preparedness of security teams [4]. Likewise, ML and AI algorithms can be used
for extracting IoT security knowledge, including insights not already known. As a
prominent example, the use of predictive analytics can enable the extraction of pre-
dictive security insights about IoT assets, in order to facilitate timely preparedness.
The collection and consolidation of security data from multiple heterogenous
sources is a key prerequisite for the design and implementation of data-driven secu-
rity techniques, including ML/AI algorithms for security analysis [5]. Indeed, secu-
rity data stem from a variety of different Probes that are deployed in the diverse
assets and devices of an IoT system. These Probes provide data in different seman-
tics and formats, which makes it difficult to consolidate and analyze them in a com-
mon and unified way. This is a setback to the implementation of integrated, intelli-
gent, data-driven security techniques, which are capable of correlating information
and events associated with different assets, as a means of identifying vulnerabili-
ties and attack patterns beyond existing security knowledge [e.g., knowledge avail-
able in popular vulnerability databases like Common Vulnerability and Exposures
(CVE)] [6].
In order to remedy this situation, there is a need for collecting and mapping
security data in a common model for IoT security knowledge modeling and cyber
threat intelligence. Such a common data model can be a foundation for security
interoperability of IoT security data monitoring systems that collect and analyze
information from different platforms and devices [5, 7]. In this direction, stan-
dards organizations have developed knowledge models for cyber threat intelligence
like STIX (Structured Threat Information Expression) [8], which serves both as
a language and as a serialization format that enables the exchange of cyber threat
intelligence (CTI). As another example, the Industry Consortium for Advance-
ment of Security on the Internet (ICASI) has specified the Common Vulnerability
Reporting Framework (CVRF) [9]. CVRF is an XML-based data exchange format
that automates data reporting and consumption of information about software
vulnerabilities, while providing support for describing the vulnerability handling
20 Security Data Modelling for Configurable Risk Assessment
security model that is presented in the chapter has been designed, developed, and
validated in the scope of H2020 SecureIoT project [5], which provides end-to-end
predictive IoT security solutions for various use cases, along with security services
like risk assessment and compliance auditing. As already outlined, the presented
data model is used for modeling security data for various use cases in connected
transport, healthcare, and smart manufacturing. At the same time, it is used for
configuring the SecureIoT data collection platform, as a means of facilitating differ-
ent deployment configurations at both design and run time. These configurations
boost the implementation of actionable security intelligence, as well as allow the
security services of the platform to reconfigure the operation of the various Probes.
This can also enable scenarios that close the loop to the field upon the identification
or the grading of certain security risks.
The rest of the chapter is structured as follows: Section 2.2 introduces the archi-
tecture of the SecureIoT platform [5], which has driven the specification of some
of the entities of the data model. Section 2.3 presents the data model, including
its main entities and the relationships between them. The presentation focuses on
the structural characteristics of the model and its constructs that enable data col-
lection and security data analysis. Note, however, that an exhaustive presentation
of the data model, including details for each one of the entities and their attributes,
is beyond the scope of this chapter. Section 2.4 is devoted to the description of
the SecureIoT Risk Assessment infrastructure, with emphasis on how it leverages
the introduced data model. This infrastructure is currently used to support mul-
tiple IoT security use cases based on different instantiations of the data model.
Section 2.5 is the final and concluding section of the chapter.
2.2.1 Overview
The SecureIoT project developed a BigData architecture for the security monitor-
ing and protection of IoT systems at scale. The architecture serves as a blueprint for
the development and deployment of IoT security systems. It enables the offering
of security services (e.g., risk assessment, compliance auditing) based on a managed
security, cloud-based model, i.e., a SECaaS (Security-as-a-Service) model.
The SecureIoT architecture specifies various components, which are grouped
together based on their functional relevance. As shown in Figure 2.1, the compo-
nents’ groups of the SecureIoT architecture are:
Figure 2.1. SecureIoT architecture for data-driven security services for IoT systems [5].
• Analytics Group: This group clusters components that process and analyze
security data in order to derive insights that are accordingly used for security
risk assessments, compliance auditing and other security functionalities. Data
analytics are supported by a powerful mechanism of reusable “security tem-
plates,” which addresses different types of threats, vulnerabilities, and attacks.
• Global Repository: This group comprises databases and repositories where
information generated from the various components is persisted. The global
repository is at the center of the architecture and interfaces to all the differ-
ent/other groups of components.
• SLA Group: This group comprises components that keep track and audit
Service Level Agreements (SLAs) between operators of IoT systems and the
SecureIoT services providers in the scope of the SECaaS paradigm.
• Security and Privacy Group: This group consists of components that model
and implement security policies.
• Risk Assessment Group: This group comprises the set of components that
deliver data-driven risk assessments. These components enable the services
that are described in Section 2.4 of this chapter.
• Compliance Auditing Group: This group comprises the set of components
that enable and implement the compliance auditing services of SecureIoT-
compliant platforms.
• Programming Support Group: This group comprises the set of components
that support programmers and developers in securing their programs, based
Data-driven Security Architecture 23
• Deployed Probes: Probes collect data from the target IoT systems or appli-
cations and stream them to a cloud platform that follows the architecture.
The routing is performed by the data routing components.
• Probe Management and Configuration: This component is responsible for
managing and configuring the deployed Probes. It interacts with the SPEP
(Security Policy Enforcement Point), which provides automatic Probe con-
figuration commands that are used to configure the managed Probes. Manual
Probe configuration commands may also be issued by security operators.
• Probe Registry: This is a registry that maintains a record of the deployed
Probes. Probe deployment data, as well as state and configuration data, are
maintained by the registry. The registry provides Probe creation, reconfigu-
ration, and search capabilities.
• SPEP (Security Policy Enforcement Point): It executes actuating functions
towards the IoT system or application under protection. Examples of sup-
ported actuating functions include: (i) The automatic (re)configuration of
Probes, which can be triggered by security data analysis performed in the Ana-
lytics Engine; and (ii) Manual (re)configuration of Probes and other compo-
nents triggered by the Risk Assessment service, i.e., as a risk mitigation action.
training data and templates. The Analytics Engine trains itself using historical
data that are stored in the Observations part of the Global Repository. The
outcome of the training is a set of templates that are stored in the Security
Template Database. The templates are executed by the Engine and provide
the data-driven logic that enables real-time identification of security issues
based on the analysis of real-time data that are streamed from the deployed
Probes through the data bus. The Analytics Engine raises alerts for security
issues that may be detected while monitoring the IoT systems. The gener-
ated alerts can be intercepted either by the SPEP component or by the Risk
Assessment Engine. Depending on the defined reconfiguration policies, they
may carry reconfiguration commands that are to be applied to the deployed
Probes. The alerts that are intercepted by the SPEP result in the reconfig-
uration of Probes, which can be put into effect through the Management
and Configuration component. Alternatively, Probe reconfigurations can be
executed manually through the Management and Configuration dashboard.
• Security Template Extraction: This component trains models based on his-
torical data and extracts reusable templates. In particular, the training part of
the analytics algorithm results in the generation of the security templates that
are stored into the Security Templates Database and are subsequently used by
the monitoring part of the analytics algorithm (e.g., in order to generate alerts
or trigger risk assessments).
• Security Templates Database: Stores the Security Templates, i.e., pre-
trained and ready to use machine learning and other data mining models. The
templates mechanism boosts reusability and modularity in handling many
different security risks.
• Risk Assessment Engine (RAE): This engine implements the logic of the
service itself. It makes use of a Risk Assessment database that contains sets of
known vulnerabilities and interacts with its user through the Risk Assessment
Dashboard. The RAE comprises its own database (RAE DB) and operates
based on its own set of rules (RA rules) that implement the risk assessment
logic.
• Risk Assessment Dashboard: This component provides a user interface to
the Risk Assessment Engine. Mitigation actions that are proposed by the Risk
Assessment Engine are presented to the Dashboard allowing the user to decide
for their enforcement. These actions are eventually carried out through the
SPEP component of the Data Management Group.
The operation of the RAE and of the dashboard based on the SecureIoT data
model is illustrated in Section 2.4 of this chapter.
• SLA Probes: These are specialized Probes that collect SLA-related informa-
tion. They are placed in different parts of the SecureIoT architecture, includ-
ing Probes in the Data Management, Analytics, and Risk Assessment Groups.
The Probes are marked as “SLA Probes” in the architecture figure. They
enable the acquisition of (raw) data that are sent to the data routing com-
ponent. The latter transforms raw data to properly structured Observations
and stores them to the Observations Repository.
• SLA Processing Engine: This engine processes raw data and generates SLA
reports, i.e., reports that provide information about the execution of the SLA
(e.g., data collected, data processed, risk assessment reports generated, etc.).
The latter are stored in a dedicated part of the Global Repository, i.e., the
SLA Repository, which is destined to support SLA Management. The SLA
Repository hosts SLA reports along with configuration data about each SLA.
28 Security Data Modelling for Configurable Risk Assessment
• SLA Admin Console: This console provides the means for configuring and
managing SLAs, based on the processing of configuration data of the SLA
Repository.
A more detailed presentation of the SecureIoT and its use in supporting data-
driven security scenarios is available in [5] and [12]. The previous paragraph aimed
at outlining the main components and their relationships between them, as the
latter have driven the design and implementation of the data model for structuring
security data about IoT assets, but also for enabling the configuration of the various
Probes that comprise SecureIoT-compliant platforms.
2.3.1 Overview
To support the security data flow and exchange across the different components of
the SecureIoT architecture, a common data model has been specified. The model is
XML-based and comprises the semantics needed for structuring and storing secu-
rity data, along with metadata regarding the configuration of a SecureIoT compli-
ant platform. It is meant to be a lightweight model that is tailored to the needs of
data-driven IoT applications. Nevertheless, based on the Security Knowledge Base
(SKB) Repository, the data model can be linked to conventional sources of security
knowledge, as a means of automating risk assessment and cyber threat intelligence
processes.
In practice, the model is structured as a collection of interconnected and inter-
related schemas, which can be considered as stand-alone data models. In particular,
the SecureIoT Data Models form six main logical groups, which are driven by cor-
responding groups of components in the SecureIoT architecture. These include the
following groups:
• Data Management, which focuses on the modeling of the data collection and
data routing functionalities. These are integral functionalities of any security
data collection platform that collects information from multiple sources and
Probes, and routes them to different consuming applications (e.g., such as
low-latency edge analytics applications).
• Analytics, which focuses on the Analytics configuration and management at
the Security Intelligence layer.
• Knowledge Base, which focuses on modeling of Security Knowledge from
third-party standardized sources.
Data Modelling for Security Systems Interoperability 29
Some of the Logical Groups are focusing on different aspects of the SecureIoT
infrastructure and some are shared along the different layers. All together they form
an integrated solution, which is used in the SecureIoT Infrastructure. Figure 2.2
shows the SecureIoT Root Entity of the data model with the different Logical
Groups.
entities are the Probe, which is responsible for collecting data and streaming it to
the processing components of the platform and the Observation, which represents
the collected datum. Each Observation concerns a metric that is measured, and this
is represented as the QuantityKind in the model. A given Probe collects data of a
given kind, and if different metrics are to be measured, then different Probes must
be deployed. Moreover, each Probe is associated to a Data Source from which data
originate and a Data Destination to which data are fed. The structure of a Data
Source sub-model is depicted in Figure 2.4. Overall, the data management group
comprises the following core entities:
These entities are linked based on standardized schemata provided from each
(knowledge provider) organization. These schemata are imported to the SecureIoT
Data Models and become available for use by other logical groups.
• Risk Model, which provides the means for representing a model for a
risk assessment service. It provides the means for modeling different risk
34 Security Data Modelling for Configurable Risk Assessment
Figure 2.8. Focus of the main building blocks of the risk assessment group in SecureIoT
architecture.
Risk Assessment Services 37
• Analytics Engine: which is using the trained templates to analyze the real-
time data coming from the IoT Systems and are stored to the SecureIoT
Global Repository (Elasticsearch)
• Data Bus: which is used as a messaging channel implementing publish/sub-
scribe for the Analytics reports
• CMDB: where specific use case data describing the involved Assets and the
potential vulnerabilities/threats are stored
• Risk Assessment Engine: which is analyzing the published reports from the
Analytics Engine to the Data Bus
• Risk Assessment Dashboard: which is responsible to visualize the risk assess-
ment reports with possible mitigation actions.
1. The possibility to represent and protect the needs of the use cases and the
validation using SecureIoT by transferring the data from the use cases to the
RAE using communication channels of the project.
2. Its different processing components to transform the data from the Probes.
Regarding the first phase, we list all the inputs required in order to start work-
ing in the RAE: a tool for modeling the risk scenarios and transform the model to
computer data where in collaboration with the use case owner the security require-
ments are collected for the threat scenario. With this procedure, we collect infor-
mation about users, assets, threats, attacks, impact in the system, etc., which we
provide as sample below. Once the information is compiled, the modeling of the
system is conducted with the help of the modeling tool.
For the Risk Assessment data flow, the direct related components are the CMDB,
Global Repository (Observations), Analytics Engine, Data Bus, Risk Assessment
Engine and Risk Assessment Dashboard. Figure 2.9 below depicts a high-level
dataflow of the validation runtime. More specifically, we can see that first we instan-
tiate the Risk Assessment Engine, which retrieves all the configuration data from the
CMDB to be initialized (not depicted in the sequence diagram), and start listening
for the Analytics Engine reports by subscribing to a Data Bus topic named after
the ID of the Analytics Engine. Then, we set up the Analytics Engine to consume
incoming data from the Elasticsearch (Global Repository) where the captured Raw
Data are stored. The data are then analyzed for specific Attacks. When an attack is
identified, it is pushed to the Data Bus by using as topic name the ID of the instan-
tiated Analytics Algorithm. The report is then picked up from the Risk Assessment
Engine which applies preconfigured rules for the specific scenario and produces a
40 Security Data Modelling for Configurable Risk Assessment
report with possible mitigation actions. This report is then picked up from the Risk
Assessment Dashboard in order to provide it to the end user.
• Attack scenario description: In the first phase, we discuss about the scenarios
of the use case in order to understand what the normal way of working is and
what the attack scenarios are. These scenarios disrupt the normal activity and
could make the system not to work as intended, data to be stolen, etc.
• Definition of assets and impact from the technical and business point of
view: After having collected the scenarios, we defined what are their assets
and the impact the attacks could have on them. This was studied from a use-
case user and cybersecurity expert perspective so we could give ideas of how
attacks could impact their services and business in a way they didn’t think of.
Risk Assessment Services 41
• Modeling of the scenarios: Using the CORAS tool [13], we modeled dif-
ferent scenarios together with risks, assets, attacks, indicators, cybersecurity
properties, etc. Following we will examine the models created for each use
case in their own sub-sections.
• Creation of different models and data that are used as input for the Risk
Analysis Engine: Once the models are created, we extract information from
them to generate different data files that are necessary for the Risk Assessment
Engine. The files we need to generate are the indicators of the cybersecurity
threats of each scenario, mitigation actions, risks, and the DEXi model [14].
This last one is a mathematical model that is used for calculating the proba-
bility for a specific risk to happen.
• Analysis: Finally, we perform the analysis of the model, and the result is
shown in the Risk Assessment Engine on the one hand and in a console on
the other hand. We followed this approach so we can check the information
in both platforms and in the next iteration create our own interface that is
more IoT-oriented and provide more useful information of the one coming
from the Probes and assurance system.
"1": "Datum1",
"2": "Datum3"
}
}
Asset
Asset entity is an abstraction of an Asset belonging to the previous IoT System in
the real world. The user needs to provide, among other information, a unique name
for the Asset, the IP address and port through which it can be remotely accessed,
and the CMDB identifier of the System entity it belongs to. An Asset sample can
be seen in the Table 2.2.
Abuse Case
Abuse Case entity is an abstraction of an attack scenario that may affect an Asset.
The user needs to provide, among other information, a unique name for the Abuse
Case, the CMDB identifier of the vulnerable Asset entity, information on the Con-
trols and Mitigation Measure that will be used as countermeasures against the
attack, as well as a numerical estimation of the attack severity (impact). An Abuse
Case sample can be seen in Table 2.3.
Table 2.3. Connected vehicle system abuse case.
{
"name": "Man in the middle attack",
"assetIDs": ["6e348dd4−e067−4456−8361−48ef8751b868"],
"controlIDs": [
"e14d4ce7−344e−475d−a847−d20c1de168ee",
"0d8557aa−623f−4cd9−b12e−ef05d782be99",
"11b0220a−84cf−4eea−8e5d−c6cb6db08efe"
],
"mitigationMeasures": [{
"extendedDescription": "Use PKI for encrypting the
data before sending to the server. PKI is recommended
as it is more secure and allows for better interaction
with other entities of the system",
"longDescription": "Use encryption mechanisms using a
PKI strategy",
"shortDescription": "IN−C1−M1"
}],
"processorOrchestratorID":
"cb53a12b−3a42−4f9e−871d−b23f2a5481e7",
"attackID": "8dd92273−573b−49c0−a61d−cde79d92ada8",
"securityData": {
"capecID": [
"154 Resource Location Spoofing",
"212 Functionality Misuse",
"441 Malicious Logic Insertion"
],
"cpeID":
["cpe:2.3:a:1024cms:1024_cms:1.4.1:∗:∗:∗:∗:∗:∗:∗"],
"cveID": [
"CVE−2017−14084", "CVE−2017−12697",
"CVE−2018−17187"
],
"cweID": ["Man in the middle (200)"]
},
"impact": {
"availability": 3,
"confidentiality": 8,
"integrity": 8
}
}
44 Security Data Modelling for Configurable Risk Assessment
Indicators
The indicators are introduced in the system via JSON and using a specific format.
An example is shown in Table 2.4.
All the fields are mandatory except for the types of sensors used. This is done for
helping in the testing of the information in the system. Regarding the other fields,
they are:
These data are processed by the Risk Analysis Engine and linked to other data of
the system such as the risks, mitigations, etc. The input validated for the indicators
comes from the Probes of SecureIoT. That way the “Rule” needs to identify a valid
Probe for linking these data to the correct scenario of risk.
Mitigations
Together with risk analysis, the Risk Assessment Engine is able to provide mitiga-
tions for identified risks. Although it is not shown in the models of each use case, we
define the information in a computer-oriented file so it can be processed and linked
to risks in the tool. In Table 2.5, we show one of the mitigations of the connected
cars use case.
Table 2.5. Connected vehicle RAE mitigation action.
{
"short_description": "IN−C1−M1",
"extended_description": "Use PKI for encrypting the
data before sending to the server. PKI is recommended as
it is more secure and allows for better interaction with
other entities of the system.",
"detailed_description": "Use encryption mechanisms using
a PKI strategy.",
"id": 58
}
Risks
The document of risks describes all the risks shown in the models in a computer-
oriented way. The objective of this is to allow the Risk Assessment Engine to link the
risks to the indicators and mitigations that we introduced previously. In Table 2.6,
we show one of the risks we used in the document with a description of the
fields.
Table 2.6. Connected vehicle RAE risk.
{
"id": 19,
"short_description": "WRP11−R1",
46 Security Data Modelling for Configurable Risk Assessment
The descriptions of each of the fields used for the risks are:
• Mode_of_operation: indicates if the risk is activated or not for the risk assess-
ment
• Mitigation_measures: the link to the mitigations for this risk
• Detailed_description: a description of the risk
• Short_description: identifier used for linking in the internal database
• Organization: the identifier of the organization having this risk
• Sections: the number of possible linked items
• Id: the unique identifier of the risk.
2.5 Conclusions
This chapter introduced a lightweight model for exchanging and reporting security
data about IoT devices, systems, and other assets. The model makes provisions for
representing security data for various IoT assets and devices. It also provides the
means for modeling the ways these data can be analyzed in order to implement
security services for IoT systems such as risk assessment services. In the scope of
the chapter, we have presented a concrete example of how this model can be used
to support a data-driven IoT risk assessment service. Furthermore, the model can
represent metadata about the IoT data collection system, as means of facilitating
its (re)configuration. The latter can also serve as basis for implementing action-
able intelligence by closing the loop to the field in response to security incidents
or even in response to specific outcomes of the risk assessment process. Likewise,
the integration of configuration metadata facilitates the configurability and pro-
grammability of the IoT security system in dynamic environments.
The use of the data model for the implementation of various security services
in different use cases demonstrates that there is merit in the overall data model-
ing approach. In particular, benefits associated with the speed of the configuration
and implementation of the data-driven security services have been observed and
References 47
validated. Moreover, the versatility of the model has been proven, given its ability to
support diverse data collection scenarios for different use cases. However, the adop-
tion of a lightweight approach comes with some limitations in the expressivity of
cyber threat intelligence applications, which can be a setback to higher automation
and intelligence. This is a trade-off that should be taken into account by designers
and deployers of IoT security systems based on the introduced data model. Fur-
thermore, the extent at which the proposed model is tied to the SecureIoT platform
should be investigated as well. There are certain parts of the model that are com-
pletely agnostic to the architecture of the SecureIoT platform. This is, for example,
the case for the modeling of data sources, Observations, and security analytics func-
tionalities. The modeling of these entities is general purpose and can be used in the
scope of virtually any IoT system for security data collection. Nevertheless, other
parts of the model are more closely affiliated to the modules and to the structure
of the SecureIoT architecture, which renders their customization and use in other
platforms more time consuming.
Overall, the chapter introduced a novel approach to modeling security data
and IoT security systems configurations. The approach emphasized security data
collection and analytics, towards delivering intelligent functionalities for security
teams, including data-driven risk assessment and actionable intelligence tied to their
outcomes. The approach can be certainly extended and customized to different
requirements, thanks to the use of XML-based, extensible and rather general-
purpose entities for modeling security data and their analysis.
Acknowledgements
This work has been carried out in the scope of the H2020 SecureIoT project, which
is funded by the European Commission in the scope of its H2020 program (con-
tract number 779899). The authors acknowledge valuable help and contributions
from all partners of the project.
References
[1] Q. Jing, A. V. Vasilakos, J. Wan, J. Lu, and D. Qiu, “Security of the Internet of
Things: perspectives and challenges,” in Wireless Networks, vol. 20, issue 8,
November 2014, pp. 2481–2501.
[2] R. Martin, S. Schrecker, H. Soroush, J. Molina, J. P. LeBlanc, F. Hirsch, M.
Buchheit, A. Ginter, H. Banavara, S. Eswarahally, K. Raman, A. King, Q.
Zhang, P. MacKay, and B. Witten. (2016). Industrial Internet Security Frame-
work Technical Report. 10.13140/RG.2.2.28143.23201.
48 Security Data Modelling for Configurable Risk Assessment
[3] A. Bolster and A. Marshall, “Analytical metric weight generation for multi-
domain trust in autonomous underwater MANETs” Proc. IEEE 3rd Under-
water Commun. Netw. Conf. pp. 1–5, 2016.
[4] U. Jayasinghe, G. M. Lee, T. Um and Q. Shi, “Machine Learning based Trust
Computational Model for IoT Services”, in IEEE Transactions on Sustainable
Computing, May 2018.
[5] A. Roukounaki, S. Efremidis, J. Soldatos, J. Neises, T. Walloschke, N.
Kefalakis, Scalable and Configurable End-to-End Collection and Analysis of
IoT Security Data: Towards End-to-End Security in IoT Systems. GIoTS
2019: 1–6.
[6] K. Umezawa, Y. Mishina, S. Wohlgemuth, and K. Takaragi, (2018). Threat
Analysis using Vulnerability Databases – Matching Attack Cases to Vulnera-
bility Database by Topic Model Analysis.
[7] H. C. Chen, M. A. Al Faruque, and P. H. Chou, “Security and privacy chal-
lenges in IoT-based machine-to-machine collaborative scenarios,” 2016 Inter-
national Conference on Hardware/Software Codesign and System Synthesis
(CODES+ISSS), Pittsburgh, PA, 2016, pp. 1–2.
[8] F. Sadique, and S. Cheung, I. Vakilinia, S. Badsha, and S. Sengupta, (2018).
Automated Structured Threat Information Expression (STIX) Document
Generation with Privacy Preservation. 10.1109/UEMCON.2018.8796822.
[9] The Common Vulnerability Reporting Framework (CVRF), Version 1.1,
available at: https://www.icasi.org/the-common-vulnerability-reporting-fra
mework-cvrf-v1-1/
[10] R. Danyliw, et al. “The Incident Object Description Exchange Format.” RFC
5070 (2007): 1–92.
[11] T. Takahashi, et al. “An Incident Object Description Exchange Format
(IODEF) Extension for Structured Cybersecurity Information.” RFC 7203
(2014): 1–28.
[12] S. Astaras, S. Efremidis, A.-M. Despotopoulou, J. Soldatos, and N. Kefalakis,
“Deep Learning Analytics for IoT Security over a Configurable BigData Plat-
form”, In the Proc. of the 22nd International Symposium on Wireless Per-
sonal Multimedia Communications, IEEE WPMC, 2019, Lisbon, Portugal,
November 2019.
[13] F. Vraalsen, and F. Braber, M. Lund, and K. Stølen, (2005). The CORAS Tool
for security risk analysis. 402–405. 10.1007/11429760_30.
[14] M. Bohanec, V. Rajkovič, I. Bratko, B. Zupan, and M. Žnidaršič, DEX
methodology: Three decades of qualitative multi-attribute modelling. Infor-
matica 37, 49–54, 2013.
DOI: 10.1561/9781680836837.ch3
Chapter 3
For nearly two decades, machine learning techniques have been extensively
researched in terms of their ability to indicate cybersecurity attacks in the network-
ing and cloud assets of IT systems. To a lesser extent, they have been deployed
in cybersecurity for IoT systems. Nevertheless, their potential for detecting attacks
in all layers and components of an IoT system has not been adequately investi-
gated. This chapter introduces deep learning techniques for the detection of abnor-
mal behaviors and anomalies in the operation of state-of-the-art IoT systems that
comprise smart objects such as connected vehicles and socially assisted robots. The
presented approach uses variational autoencoders and has been validated on real-
life datasets derived from smart objects with very promising results. Moreover, the
introduced algorithms have been integrated within a data-driven IoT security plat-
form, namely the SecureIoT platform that has been presented in a previous chapter.
Overall, the present chapter discusses a novel deep learning approach for IoT secu-
rity, while validating it in realistic datasets other than the legacy openly available
datasets that are commonly used by security researchers.
49
50 Data-driven IoT Security Using Deep Learning Techniques
3.1 Introduction
In-line with these workflows, the SecureIoT project has implemented a data-
driven security platform, which provides the means for collecting security data from
various IoT components in scalable and highly efficient (e.g., with low latency)
ways. A brief description of the platform is provided in the previous chapter of the
Introduction 51
book, while more detailed descriptions are also available in other publications of
the authors [2, 4]. Data-driven security platforms like SecureIoT enable the inte-
gration of different types of algorithms for analyzing security-related datasets, for
the purpose of identifying patterns and events that are interesting from a security
viewpoint. In fact, the algorithms used are among the more critical components of
the data-driven security systems, as their efficiency determines the overall intelli-
gence of the security mechanism.
For nearly two decades, a great deal of data analytics and machine learning
techniques have been researched in terms of their ability to identify security risks,
attack patterns, and potentially abnormal behaviors. For example, the authors in
[5] present a range of network intrusion detection techniques, including statistical,
knowledge-based, and machine-learning approaches. For another example, compu-
tational intelligence methods have been also employed [6], including deep learning
techniques such as Artificial Neural Network (ANNs). Techniques based in ANNs
have also been employed for misuse detection in [7]. The system developed in
[7] took advantage of a knowledge base that contained attack signatures. Machine
learning techniques have also been put to service for more complex security tasks,
such as the identification of botnets [8] and intrusion detection in sophisticated sce-
narios [9]. Intrusion detection use cases have been also addressed by taking advan-
tage of genetic algorithms [10]. In general, most of the research efforts for applying
machine learning in cybersecurity have been focused on techniques for identify-
ing misuse and/or anomaly detection. Their main goal has been to identify known
intrusions with high confidence, while at the same time decreasing false-positive
rates for the identification of unknown attacks.
Most of the above-listed systems have been focused on network attacks. The
emergence of new distributed computing models, like cloud computing, gave rise
to the application of similar techniques for different types of IT assets and infras-
tructures [11]. For example, in [12] supervised machine learning techniques were
used for identifying security incidents compromising cloud infrastructures. The
application of machine learning for cloud security revealed promising results along
with limitations. The use of larger amounts of historical data for training was among
the techniques that later works employed in order to remedy those limitations [13].
As already outlined, the emergence of IoT systems and applications introduced
new cyber assets, along with a set of new challenges. Machine learning techniques
show significant promise for identifying anomalies and attacks associated with IoT
assets, yet they have not systematically been studied yet. Deep learning models
can be also applied, as IoT devices produce large amounts of data that could be
utilized for security analysis. Specifically, deep learning is a class of machine learn-
ing graph algorithms where each base input feature vector is classified, clustered, or
otherwise modeled in a multi-stage process that performs both higher-order feature
52 Data-driven IoT Security Using Deep Learning Techniques
extraction and class fitting in a unified way. The defining characteristic is that the
entire process is simultaneously fitted on the data, and hence, there is no need for
a pre-designed feature extraction or pre-processing step. Deep learning models are
further divided into numerous sub-classes based on their topology, such as Recur-
rent Neural Networks (RNNs), Deep Belief Networks (DBNs), and Convolutional
Neural Networks (CNNs) [14].
This chapter presents a deep learning approach to anomaly detection in IoT
systems, notably systems that comprise smart objects like connected vehicles or
robots. Specifically, the chapter focuses on a deep learning method that has been
proved capable of detecting systematic faults and intrusions in IoT systems, namely
the Variational Autoencoder (VAE) [15]. This deep learning structure has multi-
ple hidden layers that first reduce the dimensionality of the input and then scale
up to reproduce it (Figure 3.1). It is conceptually separated into the encoder (i.e.,
the initial layers that perform dimensionality reduction) and the decoder (i.e., the
output layers that recreate the output). In the variational approach, the code layer
includes an adaptive probability distribution for the latent variable. The presented
approach has been developed and validated based on real-life datasets that have been
made available by OEM vendors and IoT integrators in the scope of the H2020
SecureIoT project [2]. Furthermore, the deep learning algorithms introduced on
this chapter have been integrated in the SecureIoT platform for data-driven secu-
rity [4], which has been presented in the previous chapter.
One of the principle innovations of the technique lies in the use of real-
life datasets from smart objects for the training and evaluation of the deep
Methodology and Datasets 53
learning models. Most of the above-presented methods have been developed and
evaluated based on reference datasets, such as the one used in the scope of the
Third International Knowledge Discovery and Data Mining Tools Competition
back in 1999 and several datasets with network-level information provided by
DARPA [16–18]. The VAE models of this chapter have leveraged true IoT datasets
derived from novel smart objects/devices like a socially assisted robot and a con-
nected car. These datasets are therefore described in subsequent sections of the
chapter.
The rest of the chapter is structured as follows: Section 3.2 following this intro-
ductory section describes the methodology followed and the datasets used. Sec-
tion 3.3 presents the VAE models developed and used. Section 3.4 presents valida-
tion results on IoT real-life datasets. Section 3.5 is the final and concluding section
of the chapter.
• Business Understanding, which is the starting point of the process and refers
to the need for understanding the anomaly detection problem at hand.
• Data Understanding, which aims at examining the data in order to gain valu-
able insights on the applicability of certain data mining schemes. Data under-
standing enables the identification of data patterns in the datasets, which will
serve as a basis for synthesizing candidate machine learning schemes.
• Data Preparation, which will transform the maintenance datasets into a for-
mat appropriate for employing data mining and machine learning models.
This step involves Extract Transform Load (ETL) processes.
• Modeling, which discerns proper machine learning or statistics schemes for
the anomaly detection problem at hand. It interacts closely with the data
preparation phase in order to ensure that the available training datasets are
appropriate for fitting the target/identified models.
• Evaluation, which entails the assessment of the produced model in terms of
efficiency as the latter is reflected in the speed of training, the speed of model
execution, its noise tolerance, as well as its expressiveness and explanatory
ability. Metrics such as classification accuracy, errors in numeric prediction,
lift and conviction measures and more are frequently used.
54 Data-driven IoT Security Using Deep Learning Techniques
1. https://www.applusidiada.com/global/en/
Variational Autoencoders for Anomaly Detection 55
datasets can be found, each associated to a set of sensors from the same simula-
tion. They are described as number 1 to 5 in the section about scenarios of socially
assistive robots. The first dataset contains environmental sensing data such as tem-
perature or humidity, and the second one provides data from wearable sensing data
such as the physical activity of the patient. The third and fourth datasets repre-
sent, respectively, visual sensing data, which give information about the patient
such as their position or age, and resting furniture data to know if the patient is on
the sofa. The last dataset contains measurements of a patient’s vitals such as their
heartbeat.
• Two fully connected layers in parallel. The first produces the mean of the
distribution, and the second the variance.
56 Data-driven IoT Security Using Deep Learning Techniques
• A sampling layer that follows the previous two and samples a Gaussian dis-
tribution with the parameters it received.
where:
The first term of the loss function is the expected log-likelihood that the decoder
reconstructs the input accurately. It is zero for a perfect reconstruction and grows
58 Data-driven IoT Security Using Deep Learning Techniques
as the model becomes less accurate. We need to use a practical substitute for per-
formance and numerical stability, a common choice being the root of the mean
absolute error.
The second term acts as a regularizer. Normally, the encoder can learn very local-
ized, essentially meaningless representations, leading to overfitting. However, the
Kullback–Leibler divergence will grow as the encoder’s distribution becomes more
complex and drifts away from our base distribution N (z).
As such, minimizing this loss function leads to an accurate model via the first
term, while it mitigates overfitting and keeps latent representations meaningful via
the second term.
1. The model can reconstruct missing features. We can artificially replace some
features from input vectors with “missing value” placeholders in the training
phase, but still include the real intended values in the loss calculation. The
model will learn to infer missing values.
2. The model can be fitted to new abnormal cases without retraining. We can
resume training with the new data and provide the new label as input. This is
especially helpful when transitioning from preliminary training and scaling
the dataset.
3. The model can detect abnormalities that weren’t encountered before. It can-
not specifically identify them, but if an input vector fails to be accurately
reproduced in the output with any label, it is evidence of a new use case that
was not encountered during training.
4. The model is very efficient on data of low dimensionality and isolated attacks.
It can learn the correlations between the features very well and isolated attacks
(i.e., attacks that are limited on the number of features they affect) are easily
recognizable and do not make the input similar to a normal situation or
another use case.
Application and Validation Results 59
• At least one sensor report is being overwritten with a static, usually extreme
value (e.g., speed or RPM appear as their maximum values despite the current
throttle position).
• At least one sensor report is being overwritten with a seemingly regular value,
but the way the value evolves is abnormal (e.g., the vehicle does not appear
to properly consume fuel as expected while it is being driven).
conditions and subsequently detect many kinds of attacks without needing all pos-
sible malicious contexts from the beginning.
However, there are some challenges. The normal driving operation is not sim-
ple, as the state of the vehicle can vary according to the traffic and road condi-
tions, although we expect the relations between the speed, throttle, RPM, gear, and
steering angle to stay the same. In addition, one kind of attack takes place over
a duration and cannot be detected from a single time instance. We handle these
challenges by introducing data augmentation. Specifically, we resample the dataset
so that the various normal driving conditions are equally frequent during training,
and we aggregate sensor readings and inject the calculated sensor slopes along their
current values.
To validate the performance of the algorithm, we randomly split the post-
processed dataset into training (70%) and testing (30%) sets. Because there is no
strict confidence measure, we used the maximum squared reconstruction error of
the normal case as the classification threshold (i.e., inputs with a reconstruction
error on the normal label over a threshold are classified as abnormal). This allows
taking advantage of this model’s flexibility on abnormal cases and leads to the per-
formance curves depicted in Figures 3.4 and 3.5.
Expectedly, very low and very high thresholds lead to performance equivalent
to a random choice, as all test inputs are either rejected as abnormal or accepted as
Figure 3.5. Performance variations on the connected car dataset as a function of the
reconstruction error threshold.
normal, respectively. The optimal error threshold for this model is t = 0.76, where
the accuracy score is 92% on this test set.
• Motor readings. The head, shoulders, and elbows are monitored for their
angular position and velocity in terms of yaw, pitch, and roll (depending on
the degrees of freedom of each motor).
• Gesture reports. The type of gesture to be performed by the QT robot, its
duration, and whether it was performed normally.
• Where applicable, event timestamps. Specifically, the start and end times
of the various runs of the QT Memory Game are recorded.
The task is to detect abnormalities in the robot motors with respect to the
intended gesture. Although it is hard to discern the gestures from the motor
62 Data-driven IoT Security Using Deep Learning Techniques
readings alone, the latter follow a regular pattern dependent on the gesture.
The duration also does not vary in the normal cases.
A variational autoencoder is ideal, as it can accurately model the normal behavior
because there is little variation among normal values, despite the complexity of
the relation. We can also detect all kinds of abnormalities without knowing them
beforehand.
Similar to the other use cases, we validate the performance of the algorithm
by randomly splitting the post-processed dataset into training (70%) and testing
(30%) sets. For this test, the motor position readings were used for the input, and
the label ground truth was derived from the gesture reports. The model we tested
provided the performance curves of Figures 3.6 and 3.7.
In this case, the ideal error threshold is t = 0.02, where the accuracy score is
73%. In general, the optimal threshold is dataset dependent as well as model depen-
dent. Here we can observe that the normal use case is more easily representable,
owing to more tightly coupled data, while abnormal data are more sparsely spread,
as attacks or faults can transpire in multiple ways. As such, the false-positive rate
tends to drop more abruptly than the false-negative rate rises.
Figure 3.7. Performance variations on the QT robot dataset as a function of the recon-
struction error threshold.
However, due to the nature of the intended sensors, the various measurements
tend to be correlated with the current timestamp, especially on the normal case.
For example, the step counter consistently records two days of low step counts fol-
lowing five days of average step counts, simulating a workday/weekend separation
in behavior. The state of the user (sleeping, walking, grooming, etc.) is correlated
with the time of day, as do the various home sensors (lightning, illuminance, etc.).
The abnormal data, even when the measured values are not out of their specified
range, do violate the above assumptions.
Aside from absolute temporal correlations, there is more coupling between the
sensors. The user’s body state is dictated by their current activity. The temperature,
humidity, and gas sensor measurements tend to increase together, and this happens
when the user is performing an activity in that room. Following an entertaining
activity, the user’s reported boredom drops.
A variational autoencoder learns this series of correlations and properly models
the normal operation. This means that we do not need to know all the possible
abnormalities beforehand and do not need to be restricted with a pure discrimi-
nator. Rather, we can expect to detect many different deviations from the normal
conditions. Moreover, we get the advantage of flexible training. For example, in a
real setting, seasonal changes will change the usual values of the sensors. We do
64 Data-driven IoT Security Using Deep Learning Techniques
not need all the data beforehand, the model will operate correctly with the current
conditions, and we can adapt it by extending the training as the expected changes
are observed.
For this dataset, we followed the same procedure of randomly splitting the post-
processed dataset into training (70%) and testing (30%) sets. Not all features were
considered for classification, as many of them were mostly static or otherwise iso-
lated in range. To demonstrate the early fitness of the algorithm for this use case,
we used a subset of environmental sensors. The trained model produces the perfor-
mance curves of Figures 3.8 and 3.9.
We can see that, in this case, the ROC curve is much simpler than the other use
cases. The performance rates also show an inversion of the common behavior of
the previous datasets: the false-positive rate is gradually dropping here, while the
false-negative rate is abrupt. This is still a characteristic of the provided dataset, as
abnormal readings have one or more features with error values that are atypical of
that sensor’s normal range, directly leading to a higher reconstruction error. Normal
readings do not have a simple defining characteristic and their representations have
more natural variance.
Here, a threshold value of t = 0.3 appears to correctly capture all the abnormal
inputs of this test while keeping false positives relatively low, leading to an accuracy
measure of 90%.
This chapter has introduced a deep learning approach based on Variational Autoen-
coders for detecting anomalies in IoT applications that comprise smart objects like
assistive robots and connected cars. The developed deep learning models managed
to detect the abnormalities in the behavior of the IoT devices and smart objects, in
three different application scenarios. Furthermore, the VAE components have been
integrated in the SecureIoT platform for data-driven IoT security.
66 Data-driven IoT Security Using Deep Learning Techniques
The developed model exhibits acceptable prediction time. Most of the time is
spent on initializing the algorithm, i.e., loading the data and the model and com-
piling the model for the target platform. On an Intel Core-i7 4771 CPU and an
nVidia Geforce GTX 770 GPU, initialization takes several seconds, while evalua-
tion of the data currently takes 80 ns for the proposed model. In terms of scalabil-
ity, the number of parameters of a model of a given depth scales linearly with the
dimensionality of the input. As such, so do the training and evaluation times.
There is necessarily a stochastic element in training a deep learning model, which
is further increased by using a variational approach. Nevertheless, a fully trained
encoder will always output the same probabilistic parameters for the same input.
Hence, the consistency of the model is deemed acceptable. Moreover, the current
implementation requires a script that will invoke the model with the correct param-
eters and input files. However, the parameters do not change between invocations
and the input file may or may not change, depending on the measurement storage
process. In this way, some degree of automation can be achieved.
Overall, our work is one of the first efforts to exploit deep learning for detecting
anomalies in smart objects. It is also an effort to integrate the deep learning model
in an IoT security platform, i.e., to integrate deep learning with a real-life data
collection system. However, future work is required towards identifying anoma-
lies and assessing risks in more complex scenarios involving multiple smart objects
and demanding correlation of diverse anomaly indicators. In this direction, rele-
vant datasets need to be made available, beyond the publicly available network and
IoT data.
Acknowledgments
This work has been carried out in the scope of the H2020 SecureIoT project, which
is funded by the European Commission in the scope of its H2020 program (con-
tract number 779899). The authors acknowledge valuable help and contributions
from all partners of the project.
References
[1] Q. Jing, A. V. Vasilakos, J. Wan, J. Lu, and D. Qiu, “Security of the Internet of
Things: perspectives and challenges,” in Wireless Networks, vol. 20, issue 8,
November 2014, pp. 2481–2501.
References 67
Chapter 4
69
70 Privacy Awareness, Risk Assessment, and Control Measures
4.1 Introduction
In the recent years, increasing number of IoT products and services are being widely
deployed in all professional and mass-market usage scenarios (Manyika et al., 2015
and Gartner, 2015). This exponential growth of diverse IoT platforms leads to mag-
nitude of market opportunities associated with the IoT products, but also in the rise
of some associated challenges and concerns. Furthermore, the IoT technology and
market landscape will become increasingly complex in the longer term, i.e., ten or
more years from now, as indicated in Holland (2012), especially after IoT technolo-
gies will prove their full potential in business-critical and privacy-sensitive scenarios.
Privacy can be conceptualized as “the right to be left alone” (Panagiotou and
Zaharakis, 2018). It refers to the process of disclosing and mobilizing personal data
under certain conditions and safeguarding measures. From technical perspectives,
IoT systems require context-based technological advancement regarding privacy
awareness, keeping consumers convenience as the primary concern, as indicated
by Sabo et al. (2013). Solutions suitable to tackle such challenges are still missing.
For instance, in smart city scenarios, many initiatives fall into the temptation of
developing new IoT platforms, protocols, models, or tools aiming to deliver the
ultimate solution that will solve all the IoT challenges and become the reference
IoT platform or standard. Instead, usually, they result in the creation of yet-another
IoT solution or standard.
In this paper, the authors explore the challenges of privacy awareness in futuristic
IoT environments, considering actuation in dependable fashion, by introducing
privacy awareness and control approach in the BRAIN-IoT platform, which is a
novel solution for building decentralized IoT platforms, based on inter-networking
across heterogeneous existing IoT systems.
BRAIN-IoT has been designed to support dynamicity and heterogeneity in
terms of new systems, protocols, and technologies, by allowing dynamic edge/-
cloud reconfiguration. Another advanced approach proposed by BRAIN-IoT is
security and privacy protection approach, which supports a set of end-to-end
security, privacy, and trust enablers suitable for fully decentralized environments.
In BRAIN-IoT platform, the authors address the challenges of privacy awareness
for IoT platform using GDPR core principles. The authors divide this research
goal into three research questions: (i) RQ1: Which features can be used to assess
privacy awareness in the IoT platforms? (ii) RQ2: Which privacy impact assess-
ment approach can be defined on top of the GDPR-based privacy principles? (iii)
RQ3: How to validate the security and privacy impact of a given IoT platform?
In response to RQ1, the authors explore key privacy principles intro-
duced in GDPR regulation and International Organization for Standardization
Literature Review 71
This section provides an overview of the State of The Art (SoTA) in the context of
(i) GDPR implication in IoT domain and (ii) current standards and tools for PIA.
weak cyber security standards, often owing to the limited computational power of
identifying technologies such as Wireless Fidelity (WiFi) or Radio Frequency IDen-
tification (RFID), can exacerbate privacy risks (Wachter, 2018). Together, these
risks make free and well-informed consent challenging in the IoT. Considering the
risk of personal data processing and linkage of user records, privacy policies often
fail to communicate clearly the privacy risks, as indicated by Basso et al. (2015a).
In Libertés (2017), the authors explain that GDPR regulation creates govern-
ing principles of personal data processing (Articles 5) and guideline for individual
rights (Article 25 and 33) that can help to perform PIA relevant for IoT devices.
In particular, GDPR (Voigt and Bussche, 2017) helps for transparency (Article 5),
data storage, access, rectification, and deletion (Articles 5, 15–17), informed con-
sent (Article 7), notification duties (Articles 13–14 and 33), automated decision-
making and profiling (Articles 21–22), privacy by design and privacy by default
(Article 25), cyber security (Articles 33–34), and data protection impact assessment
(Article 35–36). Tools created based on GDPR for PIA, like the one described in
Libertés (2017), can help to undermine the tendency of IoT devices and services to
collect, share, and store large and varied types of personal data, to operate seamlessly
and covertly, and to personalize functions based on prior behavior. For example, EU
H2020 project Synchronicity (Synchronicity, 2018) explores the benefit of GDPR
privacy principles in the domain of smart city. In BRIAN-IoT context, the term
privacy denotes “the right of an entity to be secure from unauthorized disclosure of
personal information that are contained in BRIAN-IoT platform.” In this paper,
the authors explore the key privacy principles from GDPR to perform privacy com-
pliance evaluation for any IoT system.
Considering tools designed for privacy awareness and control, the nationale de
l’informatique et des libertés (CNIL) PIA tool described in Libertés (2017) has been
designed around three principles: (i) a didactic interface to carry out PIAs: the tool
relies on a user-friendly interface to allow for a simple management of your PIAs. It
clearly unfolds the PIA methodology step by step. Several visualization tools offer
ways to quickly understand the risks. (ii) A legal and technical knowledge base:
the tool includes the legal points ensuring the lawfulness of processing and the
rights of the data subjects. (iii) A contextual knowledge base, available along all
the steps of the PIA, adapting the contents displayed. The data are extracted from
several sources, e.g., the GDPR, the PIA guides and the Security Guide from the
CNIL, to the aspect of the processing studied. BRAIN-IoT will extend the IoT-
Modeling Language (IoT-ML) adding the some specific modeling profile that can
be used to describe privacy awareness polices. IoT-ML is a UML-based modeling
language built upon the Internet of Things Architecture (IoT-A) architecture frame-
work, described in Atzori (2017). It adds to the key concepts provided by IoT-A to
describe IoT systems, the ability to integrate new IoT domain models, as indicated
in Gérard et al. (2017).
A PIA approach can assist the controller of an IoT system to identify any specific
risks of privacy breaches involved in an envisaged operation. More specifically, a
PIA is an essential approach to identify privacy risks, and perform risk analysis, risk
evaluation, personal data flow, and storage. In the BRAIN-IoT scope, the PIA can
be performed in the context of two components, IoT devices and IoT systems for
various use cases.
The proposed approach is based on the following four components: (i) Context:
Define and describe the context of the IoT system under consideration. (ii) Privacy
principles: Analyze the controls guaranteeing compliance with the GDPR funda-
mental principles considering the proportionality and necessity of processing, and
the protection of data subjects rights. (iii) Privacy risks: Identify the threats that are
associated with data privacy and ensure they are properly treated. (iv) Privacy com-
pliance evaluation: Analyze the overall privacy safeguarding requirements related to
the processing of information assets in a particular setting.
4.3.1 Context
At the beginning of a PIA, it is important to outline the use case under consid-
eration, its nature, scope, context, and purposes. Also, identify the data controller
and any processor of information assets in a particular setting. In the context of
A Conceptual Privacy Awareness Framework 75
BRAIN-IoT, for PIA an information asset could be: (a) software (e.g., an operating
system), (b) hardware (e.g., a sensor, Central Processing Unit – CPU – memory,
etc.), (c) processed data (e.g., personal data stored in BRAIN-IoT platform, sen-
sor status transmitted over a network, robot location in memory, etc.). In certain
instances, identifiability of the processed data might vary. To determine whether
or not information about a person considered as processed data in IoT platform,
several factors need to be taken into account. Personal data can be considered to be
information assets in at least the following instances (Standardization, 2011): (i) if
it contains or is associated with an identifier, which refers to a natural person; (ii)
if it contains or is associated with an identifier, which can be related to a natural
person; (iii) if it contains or is associated with an identifier, which can be used to
establish a communication with an identified natural person, and (iv) if it contains
references, which link the data to any of the identifiers above. Information does
not need to have an identifier to be considered information asset. Information will
also be considered personal data, if it contains or is associated with a characteris-
tic which distinguishes specific purpose. Any attribute which takes on a value that
uniquely identifies a information assets has to be considered as a distinguishing
characteristic, as indicated in Standardization (2011). Further, the authors explore
the processed data based on three categories: (a) information about the user (first
name, date of birth, email, etc.), (b) recorded data (sounds, images, movements,
temperature, etc.), and (c) calculated data (data from robot operating system, etc.).
These concepts have been used to define a template to provide a brief description
of the use cases for PIA.
Basso et al. (2015b). This section outlines privacy control criteria based on the six
key principles and four additional principles.
(P1) Lawfulness, fairness, and transparency: The lawfulness, fairness, and trans-
parency for information assets are set out in Article 6 of the GDPR. Six privacy con-
trol criteria are defined for this principle (P1-01...P1-06, corresponding to GDPR
Article 6 points a...f ). At least one of these criteria must apply whenever processing
personal data.
(P2) Purpose limitation: In the BRAIN-IoT context, purpose limitation principle
helps to ensure controller plan to use or disclose processed data (GDPR article 5).
The controller needs to specify purpose that is additional or different from the
original specified purpose. Also, it is important that the new specification must be
fair, lawful, and transparent (Standardization, 2011). In this case, the criteria are:
P2-01: clearly identified purposes for data processing; P2-02: purposes must be
documented; P2-03: regularly review the processing and, where necessary, update
the documentation and privacy information for individuals.
(P3) Data minimization: Data minimization is closely linked to the principle of
“collection limitation” but goes further than that. Whereas “collection limitation”
refers to limited data being collected in relation to the specified purpose, “data min-
imization” strictly minimizes the processing of information assets (GDPR article
5). The criteria for this principle are: P3-01: ensure adoption of a “need-to-know”
principle, i.e., one should be given access only to the information assets which is
necessary for the conduct of his/her official duties; P3-02: delete and dispose of
information asset whenever the purpose for personal data processing has expired;
there are no legal requirements to keep the information asset or whenever it is prac-
tical to do so.
(P4) Data quality: The data accuracy controls criteria for BRAIN-IoT system
requirements based on the Article 5(1)(d) of the GDPR (Parliament, 2016) are:
P4-01: ensure the processed data present in the IoT platform is not incorrect or
misleading; P4-02: ensure the reliability of processed data collected from a source;
P4-03: keep a note of any challenges to the accuracy of the IoT platform processed
data.
(P5) Storage limitation: Article 5(1)(e) of GDPR explicitly pointed out that a stor-
age duration must be defined for each type of process data, and it must be justified
by the legal requirements and processing needs (Parliament, 2016). Thus, storage
of personal data in a IoT platform must be justified using privacy control crite-
ria. The evaluation criteria for the storage limitation in the context of BRAIN-IoT
system are: P5-01: ensure mechanism must be implemented to archive common
data or purge archived data at the end of their storage duration; P5-02: functional
A Conceptual Privacy Awareness Framework 77
traces will also have to be purged, as will technical logs which may not be stored
indefinitely.
(P6) Integrity and confidentiality: Article 5(1)(f ) of the GDPR concerns the
“integrity and confidentiality” of personal data. It can be referred to as the GDPR’s
“security principle.” This principle explores the broad concept of information secu-
rity. Poor information security leaves any systems and services at risk and may
cause real harm and distress to individuals—lives may even be endangered in some
extreme cases. The common control criteria for integrity and confidentiality are:
P6-01: presentation, when the device is activated, of the terms and conditions
for use/confidentiality; P6-02: possibility of accessing the terms & conditions for
use/confidentiality after activation; P6-03: legible and easy-to-understand terms;
P6-04: existence of clauses specific to the device; P6-05: detailed presentation of
the data processing purposes (specified objectives, data matching where applica-
ble, etc.); P6-06: detailed presentation of the personal data collected; P6-07: pre-
sentation of any access to the identifiers of the device, the smartphone/tablet or
computer, specifying whether these identifiers are communicated to third par-
ties; P6-08: information for the user if the app is likely to run in the back-
ground; P6-09: information on the secure data storage method, particularly in
the event of sourcing; P6-10: information on protection of access to the device;
P6-11: arrangements for contacting the company (identity and contact details)
about confidentiality issues; P6-12: where applicable, information for the user
on any change concerning the data collected, the purposes, and confidentiality
clauses.
(P7) The right to be informed: In the BRAIN-IoT platform, various components
of IoT systems are integrated for modeling tasks. An IoT platform should be able
to demonstrate that an individual has consented the processed data. The individ-
ual must be able to withdraw his/her consent easily at any time (Articles 7 and 8
of the GDPR). The following is a list of controls intended to ensure that: users
of BRAIN-IoT consent has been informed; there has been a reminder and con-
firmation of their consent; the settings associated with the latter have been main-
tained: P7-01: express consent during activation; P7-02: consent segmented per
data category or processing type; P7-03: express consent prior to sharing data with
other users; P7-04: consent presented in an intelligible and easily accessible form,
using clear and plain language adapted to the target user; P7-05: for each new user,
consent must once again be obtained; P7-06: where the user has consented to the
processing of special data (e.g., his/her location), the interface clearly indicates that
said processing takes place (icon, light).
(P8) The rights of access and to data portability: Where the data processing
benefits from an exemption from the right of access, as presented in Articles 15 of
78 Privacy Awareness, Risk Assessment, and Control Measures
(P10) The rights to restriction of processing and to object: The list of control cri-
teria considering the rights to restriction of processing and to object of personal data
(from GDPR article 20 and 25) in a IoT platform are: P10-01: existence of “Pri-
vacy” settings; P10-02: invitation to change the default settings; P10-03: “Privacy”
settings accessible when activating the device; P10-04: “Privacy” settings accessible
after activating the device; P10-05: existence of technical means for the data con-
troller to lock access to and use of the individuals to restriction; P10-06: possibility
of deactivating some of the device’s features (microphone, Web browser, etc.); P10-
07: existence of alternative apps for accessing the device; P10-08: compliance in
terms of tracking; P10-09: effective exclusion of processing the user’s data in the
event consent is withdrawn.
Maximum (Scale: 3): It seems extremely easy for the selected risk sources to mate-
rialize the threat by exploiting the properties of supporting assets (e.g., theft
of paper documents stored in the public lobby). Further, privacy stakeholders
may encounter significant, or even irreversible consequences, which they may not
overcome.
Considering the likelihood and severity scale, it is possible to evaluate a privacy
principle (P) using each privacy control criterion (Ci ). The risk level for a privacy
principle (P) can be defined as:
n
1X
P= Ci , i = 1 . . . n (4.1)
n
i=1
Here, n is the number of privacy control criteria. Based on the risk value of a
privacy principle, it can be assumed that if the value is greater than 2, then the
risk level for a use case is significant and need further attention from controllers.
The severity and likelihood evaluation can be performed by monitoring how many
information assets affected by a specific privacy threat.
This section describes the experiments performed on IoT scenario, namely Criti-
cal Water Infrastructure, provided by the BRAIN-IoT project. Firstly, an use-case
Experimental Analysis 81
description is presented, and then, the results of privacy compliance analysis are
reported.
Below in Table 4.1, the authors present an overview of Critical Water Infras-
tructure use case. Considering the processed data, there are no real personal data
present in this use case. However, the generalized PIA approach based on GDPR,
presented in Section 4.3, is applicable also to this type of use cases. Based on this
approach, the authors can get an overall view on impact of privacy threats consid-
ering GDPR regulation. The result of the privacy compliance evaluation of Critical
Water Infrastructure scenario is briefly presented in Table 4.2.
Use-case The services of water treatment and supply in large cities are
description fundamental for the existence of life and for its quality. These services
are associated to several infrastructures that are considered, in
accordance with European norms, as critical, and as such are therefore
bound to a series of conditions for their development. Within this
context, it’s about how best to apply the technological improvements to
our work environment preserving the established guarantees. It’s about
progressing in the application of technology in the processes related to
potable water and its urban distribution, in a way that within the
sustainability context, citizens receive a better service. In other words,
from a social point of view, but also from economic environmental one.
Processing An important part of this effort is orientated to systems’
purposes implementation, improving the general information of the companies
processes, thus those associated with IoT technologies and their
integration as company’s information models. The idea is that a greater
amount and better flow of information will allow taking better
decisions, to achieve resilient and sustainable systems to guarantee high
quality of life.
Processing Online monitoring system of the quality of surface waters, with two
stakes profilers and two quality control stations in the dam and the river that
supply the metropolitan area around the city of Corun̈a.
Treatment Plant (ETAP), with an automatized management system
that allows to control all process aspects to optimize the inputs demand,
especially reactive and energy as fundamental elements of sustainability.
Tele-control system of supply grids, with more than 80 control points,
i.e., deposits, pumps, and sectors, with access to water quality data and
the correct infrastructures’ functioning. This system serves also as a feed
of the hydraulic model of the supply grid, allowing a constant
improvement of water distribution efficiency amongst the population.
(Continued)
82 Privacy Awareness, Risk Assessment, and Control Measures
Risk level
Privacy control criteria Privacy threats Severity scale Likelihood scale
Risk level
Privacy control criteria Privacy threats Severity scale Likelihood scale
4.5 Discussion
This section reports a synthesis of the main findings of the experimental analysis
presented in Section 4.4.
The privacy compliance analysis is presented in a radar chart in Figure 4.2. This
figure illustrates the PIA results in a radar chart using 10 (P1 to P10) GDPR-based
privacy principles. The severity (expresses the magnitude of a risk) and likelihood
(expresses the possibility of a risk occurring) evaluation can be performed by mon-
itoring how many information assets (such as door controller, etc.) are affected by
a specific privacy threat (for instance, unauthorized access to services, illegal access,
etc.). The scale used is the one presented in Section 4.3, so it can be assumed that
if the risk value is greater than 2, then the risk for a use case is significant and needs
further attention from the use-case owner. From the PIA results, it is possible to see
that Integrity and confidentiality (Security) privacy principle (P6) has high sever-
ity and likelihood scale value (2) due to risks regarding illegal access, unauthorized
access to services, data from untrustworthy sources, vulnerable IoT devices, and
data from untrustworthy sources. Figure 4.3 illustrates the main findings of this
work which is labeled A to F.
Complex IoT application scenarios present a set of challenges that were not consid-
ered in more traditional IoT applications, such as heterogeneity and lack of interop-
erability, security and privacy concerns. Moreover, the challenges in implementing
privacy awareness in IoT platforms consider the needs of allowing autonomous
behaviors in isolation. Further, as more IoT systems get connected to the inter-
net, there is a need for a highly scalable and extensible PIA approaches. Authors’
belief is that adopting the GDPR key privacy principles to remodel PIA approach
will enhance the reliability of IoT platform. This paper has explored the lingering
problems in the context of privacy awareness in complex IoT applications by inte-
grating GDPR in IoT platforms. More specifically, this paper has presented the way
in which the BRAIN-IoT project intends to address challenges of privacy aware-
ness in IoT platforms by leveraging GDPR and ISO/IEC standards. This paper
has further proposed a PIA approach for IoT platforms, which will also be advan-
tageous from the developers’ point of view. The proposed framework is yet to be
fully implemented and tested but it presents a direction of investigation for the
IoT community on ways the problem of scalability, ease of development, ease of
deployment, and lightweight implementation can be resolved.
86 Privacy Awareness, Risk Assessment, and Control Measures
Acknowledgments
This work is supported by European Union’s Horizon 2020 research and innovation
programme under grant agreement No 780089, project BRAIN-IoT (model-Based
fRamework for dependable sensing and Actuation in INtelligent decentralized IoT
systems).
References
Atzori, L., A. Iera, and G. Morabito. 2017. “Understanding the Internet of Things:
definition, potentials, and societal role of a fast evolving paradigm.” Ad Hoc
Networks. 56: 122–140.
Basso, T., M. Leonardo, M. Regina, and J. M. e B. Andrea. 2015a. “Towards a
UML profile for privacy-aware applications.” In: Conference on Computer and
Information Technology; Ubiquitous Computing and Communications; Depend-
able, Autonomic and Secure Computing; Pervasive Intelligence. Ed. by I. Interna-
tional. and Computing.
Basso, T., R. Moraes, and M. J. e M. Vieira. 2015b. “Requirements, design and
evaluation of a privacy reference architecture for web applications and services.”
In: Proceedings of the 30th Annual ACM Symposium on Applied Computing.
BRAIN-IoT, C. 2018. “Initial Threat modelling and Security assessment.”
Tech. Rep.
Cavoukian, A. 2006. Creation of a Global Privacy Standard. [Online]. Available.
URL: http://www.ehcca.com/presentations/privacysymposium1/cavoukian%
5C_2b%5C_h5.pdf .
Etsi, C. Y. B. E. R. 2019. “Cyber Security for Consumer Internet of Things.” ETSI.
2019.
Gartner. 2015. “M2M Global Forecast & Analysis Report.” Tech. Rep. Machine
Research.
Gérard, S., C. Dumoulin, P. Tessierck, and B. Selic. 2017. “Papyrus: A UML2 Tool
for Domain-Specific Language Modeling.” pp. 361–368.
Group, O. M. 2011. “OMG Unified Modeling Language (OMG UML).” Super-
structure, Version. 2(4): 1.
Holland, J. H. 2012. Signals and Boundaries: Building Blocks for Complex Adaptive
Systems. MIT Press.
Libertés, C. N. de l’Informatique des. 2017. Privacy Impact Assessment (PIA).
[Online]. Available. URL: https://www.cnil.fr/en/privacy-impact-assessment-
pia.
References 87
Chapter 5
Cyberattacks on the Internet of Things (IoT) can be the source of major economic
damage. They can disrupt production lines, manufacturing processes, and supply
chains. They can adversely impact the physical safety of vehicles and transportation
systems, and damage the health of living beings both through supply chains for
food, medicines, and other vital items, as well as through direct attacks on sensors
and actuators that may be connected to vital functions. Thus, securing the IoT is
of primary importance to our societies. This paper describes the technical approach
that we adopt for IoT security in the SerIoT Research and Innovation Project that
is funded by the European Commission. We first discuss the risk scenario for the
IoT and briefly review approaches that have been developed to mitigate such risks.
Then, we discuss a policy-based lightweight approach that mitigates risks at the
level of the attachment of IoT devices to a network. We follow this with a detailed
proposal based on using a distributed Machine Learning approach to risk and attack
detection in real time, as well as suggestions for future work.
88
Introduction 89
5.1 Introduction
The IoT may extend to billions of objects and sensors connected to Clouds,
databases, decision systems, and actuators [1]. It has the potential to improve the
critical processes that are at the heart of our socio-economic systems [2, 3] compos-
ing a data-driven society [4]. However, the pervasive nature of the IoT raises risks
that go way beyond the individual technologies such as the internet or wireless
networks [5] and machine-to-machine systems [6]. In addition to risks related to
system malfunctions, Quality of Service (QoS) failures, and excessive energy con-
sumption, they also include the theft and tampering of data, conventional network
attacks, and other attacks that attempt to deplete the energy of autonomous sensors
and actuators [7–12].
In Information and Communication Technology (ICT), risk management
embraces different processes intended to deal with the identification, assessment,
and mitigation of risks, which are derived from vulnerable ICT systems or potential
cybersecurity attacks [13]. However, the scale and heterogeneity of IoT systems sets
out significant challenges for the implementation of a risk management approach.
On the one hand, IoT represents the current trend to hyperconnected systems;
therefore, risk management aspects must consider the relationships and dependen-
cies among different devices that could affect to the security level of a certain sys-
tem or deployment. On the other hand, IoT deployments are usually composed
by components with resource constraints, which are installed in uncontrolled envi-
ronments with default security configurations. These constraints also make IoT
devices and systems an attractive target for possible attacks. Furthermore, due to
the potential huge number of devices in IoT deployments, there is a real need to
consider risk management approaches with a high degree of automation to effi-
ciently identify and mitigate new risks affecting such systems. These approaches
should be able to represent the relationships among devices, vulnerabilities, and
attacks of the overall deployment.
To create a suitable risk management methodology for the IoT, holistic
approaches are required to understand the probability and impact of potential
risks affecting devices and systems. As a core process of risk management, risk
assessment has been strongly considered in recent years by IoT researchers [14].
However, assessing risk is in itself insufficient, and we must design IoT networks
that can detect and then mitigate security risks, but also mitigate other risks that
may arise regarding the overall performance of the system by preserving QoS
and offering energy efficient operation of the system [15]. Thus, a continuing
overall evaluation of the risks introduced by IoT systems and their networks
is needed, and this paper offers a view of risk identification and mitigation as
applied to IoT systems, based on our work in the SerIoT Project supported by the
90 IoT Network Risk Assessment and Mitigation
vehicle) can be composed of many different devices that, in turn, could be com-
prised of several components with a different security level. This means that the risk
assessment of a certain system will depend on the security level associated to each
part of the system. For this aspect, the main challenge is related to the way in which
each risk value could be aggregated to provide a reliable value for the whole system.
A number of risk assessment methodologies have been developed in the literature
and applied in some cases to the IoT domain. For example, the Common Weakness
Scoring System (CWSS) [30] is a methodology to prioritize software weaknesses.
The main motivation is to provide means to different stakeholders (e.g., software
testers, or manufacturers) for quantifying the risks associated to a specific weakness.
This way, the corresponding stakeholder can prioritize the weakness to be solved
based on the estimated risk. Various metrics (i.e., Base Finding, Attack Surface and
Environmental metric groups, which in turn contain different metric to quantify
the CWSS score associated to a certain weakness) are used to produce a final CWSS
score between 0 and 100. The main challenge is obviously the quantification pro-
cess as cybersecurity risk is more difficult to quantify especially in the IoT domain
[14]. A similar approach is also used in the Common Vulnerability Scoring System
collection (CVSS) [31], which is based on three group metrics: Base, Temporal,
and Environmental. The CVSS is widely used today as it is used in the National
Vulnerability Database (NVD) created by the NIST. In the IoT literature, CVSS
and CWSS schemes are used to assess the risks in Bluetooth technology, where the
authors extended the authentication metric to include new security factors inher-
ent to the Bluetooth technology [32]. CVSS and CWSS have also been used as an
input in the process of cybersecurity certification of IoT devices in [33, 34]. Finally,
another quantitative risk assessment scheme is DREAD, which is used to compute
a risk value associated to a certain threat or vulnerability based on the use of five
categories: Damage potential, Reproducibility, Exploitability, Affected users, and
Discoverability. It was used at Microsoft, and currently, it is used by OpenStack. In
[35], DREAD is applied to the mobile healthcare domain due to its simplicity.
As already mentioned, risk identification is a core process of a risk assess-
ment approach. In this direction, threat and vulnerability assessments are usually
employed to identify risks to organizational operations, and the evaluation of those
risks in terms of likelihood of occurrence and impact if they occur. In turn, for
the identification of threats and vulnerabilities, Intrusion Detection System (IDS)
have been widely used [36] by monitoring and analyzing the network and/or sys-
tem events. However, previous considerations make the application of well-known
IDSs more challenging into IoT systems. Indeed, one of the popular approaches
in literature is to analyze the data produced by IoT device to identify potential
anomalies. A potential weakness of this approach is the difficulty to anticipate the
92 IoT Network Risk Assessment and Mitigation
attack as the anomaly is created when the attack is already ongoing or it is com-
pleted. Moreover, machine learning capabilities have been increasingly become
more powerful in recent times, and a complex security attack may be composed by
a sequence of attacks steps. Then, the research objective is to the detect the initial
steps in the attack.
These aspects are already highlighted by [37], which provides a comprehensive
taxonomy of IDSs for IoT. Authors classify IDSs according to different parameters,
such as the placement strategy (distributed, centralized, and hybrid) and detection
method (signature-based, anomaly based, specification-based, and hybrid). In par-
ticular, anomaly based detection techniques are used to detect new attacks by com-
paring the normal behavior of a certain system with its actual activity. This approach
could leverage the inherent nature of IoT systems, which are usually composed of
special purpose devices. This aspect is considered by the recent Manufacturer Usage
Description (MUD) standard [38], which is aimed to restrict the communication
to/from a certain IoT device. To generate such intended behavior, anomaly detec-
tion techniques have widely used machine learning techniques, and in [39], a dis-
tributed anomaly detection approach in which each node monitors its neighbors
is proposed, where monitoring nodes inform other nodes about security problems.
In [40], deep autoencoders to detect anomalous network traffic of IoT devices are
discussed and tested for different commercial IoT devices in the presence of IoT
botnets such as Mirai. In our case, we describe a risk monitoring approach based
on a multi-agent system and DL techniques for automatic feature extraction and
anomaly detection that is described in Section 5.4.
As the next step of a risk management approach, risk mitigation focuses on the
goal to mitigate the risks through different means: (a) avoiding the risk (e.g., not
using a specific interface), (b) reducing the risk by implementing specific counter-
measures (e.g., intrusion detection), or (c) transfer the risk to some other entity
(e.g., an insurance company). In the IoT domain, the mitigation of risk is more
difficult to achieve because of the limited computing capabilities of IoT devices,
which makes more difficult the implementation of sophisticated countermeasures.
The transfer of the risks is also difficult as IoT devices are often standalone systems
(e.g., a sensor). A potential approach to mitigate IoT security risks is the integra-
tion of Software-Defined Network (SDNs), in order to restrict communications
from/to IoT devices. Indeed, the application of SDN techniques into IoT scenarios
has attracted a significant interest from academia in recent years. For example, [41]
describes an architecture for managing the obtaining and enforcement of MUD
restrictions, which are enforced by SDN switches. In this direction, our approach
is based on an architecture to identify and mitigate security risks based on the com-
munication profiles of IoT devices that is described in the next section.
Autopolicy System 93
1. Authentication/
Bootstrapping 3. Request Profile
IoT Device
Profile
Profile Builder
Publisher
Local database
The obtained traffic profile is then sent to the Policy Enforcer, which only allows
IP traffic under strict rules defined by the profile. The role of the Policy Enforcer
could be embedded in the controller (or switch) acting as Device Authenticator.
Listing 5.1 shows a simple example of traffic profile to define restrictions on the
communications to/from the device. In particular, rate specifies the maximum
bandwidth (in Mbps); allowed_dst and blocked_dst indicate the IP address, pro-
tocol, and port of allowed/denied communications. Furthermore, connections rep-
resents the maximum number of connections allowed with a certain device.
While this represents a simple example of traffic profile, we plan to extend this
initial data model and align it with the MUD standard. The described mechanism
was designed for the needs of the SerIoT project as one of the components to pre-
vent and mitigate potential risks in IoT systems. Its role is to secure the network
from the malicious endpoints and ensure that IoT devices behave according to spe-
cific traffic rules. It should be noted that the described mechanism is intended to
reduce the attack surface by enforcing the rules associated to the intended behav-
ior of IoT devices. However, it still requires a dynamic approach to continuously
monitor the risks that can be identified through traffic analysis. For this purpose,
next section provides an overview of the risk monitoring approach that is being
implemented in the scope of the SerIoT project.
The recent proliferation of IoT technologies has given rise to a dramatic increase
in the number of edge devices. More devices not only introduce more security vul-
nerabilities but also greatly increase the volume of transferred data that must be
analyzed in order to detect and mitigate network anomalies. When undetected,
96 IoT Network Risk Assessment and Mitigation
such anomalies can have a great impact on the robustness and Quality of Service
(QoS) of the attacked IoT infrastructure. The main challenge for anomaly detec-
tion methods in today’s IoT network is the analysis of the big amounts of data that
arise from the growing number of IoT devices. Such large volumes of information,
as well as the distributed, large-scale, and heterogeneous nature of the IoT network,
render the goal of (near) real-time anomaly detection difficult to accomplish using
traditional centralized monitoring approaches [51].
In this respect, this work proposes the use of Multi-agent Systems (MAS) for
processing the large amounts of data exchanged by the IoT devices in a distributed
manner. The MAS allows for monitoring of the network traffic using parallel com-
putation, resulting in faster detection times, and thus, improving the security and
QoS of the IoT network. Additionally, when the agents have redundant roles (i.e.,
detection of anomalies regarding the same IoT device by different agents), the MAS
system offers robustness, since it can tolerate failures of one or multiple agents [52].
Finally, MAS also offers scalability, since due to the modular nature of the approach,
adding new agents to the network is easy and straightforward.
The main challenge when implementing MAS is how the agents will exchange
information in order to solve a problem in a cooperative manner [53]. In this
respect, this work utilizes recent advances in AI and Deep Learning (DL) in order
to enable automatic feature extraction and anomaly detection by the agents of the
MAS. Specifically, due to the graph structure of the network communication events,
where nodes represent devices and edges communication, it is natural to consider
DL approaches that deal with graph structured datasets, i.e., Graph Neural Net-
work (GNN) [18]. This work utilizes as an inspiration the more generic GNN
formulation, proposed by DeepMind [17].
The IoT network is defined as a graph G(V, E, f v , f e ), where each node vi ∈ V
represents either an IoT device or a router, and each directed edge e j ∈ E represents
a directed communication between nodes. Each node and edge is augmented with a
feature vector through functions f v : V → R Nv and f e : E → R Ne , respectively,
where Nv and Ne , the size of node and edge feature vectors, respectively. These
features are calculated using network traffic over a specific time window 1T . The
list of utilized features is presented in Table 5.1.
The agents of the MAS are installed in a subset of the nodes V 0 ⊆ V and
exchange information using the same set of communication edges E. Each agent
comprised of two networks utilized for updating the feature representation of the
adjacent edges and the node they are installed on. These updated features are after-
wards utilized for multi-class anomaly detection through another pair of deep learn-
ing networks. An overview of this formulation for updating feature representa-
tion through information exchange between the different agents is illustrated in
Figure 5.2. Under this consideration, the role of the edge Deep Neural Network
Towards Distributed Attack Detection 97
Table 5.1. The list of network features used for anomaly detection. The number of features
for each node is Nv = 5 and for each edge is Ne = 3.
Figure 5.2. The procedure for updating the feature representations of the edges and
nodes visible by each agent. The edge Deep Neural Network (DNN) takes as input the
previous edge feature values and the adjacent agent features, and updates the fea-
ture representation of each edge. The node DNN aggregates the information from the
updated edge feature values, and updates the feature representation of the specific node.
Edge DNNs are the same for all agents (i.e., have shared weights), and thus, each
agent is able to learn from events happening to other agents.
Based on the previous definitions, there are in total four DNNs, two used for
updating node and edge feature representations, namely MLP n and MLP e , and
two used for classification of neighbor nodes and the node in which the agent is
installed, namely Classifier neighbor and Classifier node .
Given an agent installed in node v j , the procedure followed for updating the
features of this node and the adjacent edges, as well as detecting anomalies in the
entire neighborhood is formulated as follows:
0 j
i
e1T = MLPe (x1T , x1T
i
, e1T
i
)
N
!
0j 1 X 0i
x1T = MLPn concat(e1T , x1T
i
)
N (5.1)
i=1
0
pi = Classifierneighbor (e1T
i
, x1T
i
)
0
p j = Classifier node (e1T
i
, x1T
i
)
where x j is the feature vector of node v j where the agent is installed, x i is the
feature of the neighboring nodes in the graph, ei is the feature vector of the adjacent
edges, and N is the number of neighboring nodes. This computation is performed
with traffic data from a specific window 1T . The MLP n network is applied on the
average of the updated features of each adjacent edge and node. The concant (.)
function takes as input two vectors and returns a larger vector which represents
the concatenation of the two vectors. Equation (5.1) represents a single iteration of
information exchange between agents. Stacking multiple blocks of such operations,
the agents can exchange information multiple times within each time period 1T .
The architecture of this procedure is illustrated in Figure 5.3. The entire network
is trained in a supervised manner using the back-propagation algorithm with cross-
entropy as loss for multi-class classification.
In this formulation of MAS, each agent is responsible for performing anomaly
detection not only on itself but also on its neighboring nodes. This redundancy
of the role of the agents enables the system to be robust in cases where one or
multiple agents may fail, since other agents will take over the role of detecting
anomalies in neighboring nodes instead of them. The issue that arises, however,
is how to combine the overlapping decisions of the different agents into a single
decision for each node in the IoT network. This work utilizes a simple aggregation
method, where a node is considered anomalous if at least one agent reported it as
Conclusions 99
Figure 5.3. The architecture of each DNN agent in the multi-agent anomaly detection
system. The edge Multi-layer Perception (MLP) takes information from the N neighboring
agents and updated the values of the edge features, while the node MLP combines the
updated edge feature values and updates the feature values of the specific node (agent).
Edge features are used for predicting anomalies in the neighborhood, while node features
are used for detecting anomalies in the specific node, using the classifier networks. The
training is performed in a supervised manner using back-propagation with cross-entropy
loss.
5.5 Conclusions
Acknowledgments
References
[1] Bera, S., Misra, S., and Vasilakos, A.V.: Software-defined networking for inter-
net of things: A survey. IEEE Internet of Things Journal 4(6), 1994–2008
(2017)
[2] Elhammouti, H., Sabir, E., Benjillali, M., Echabbi, L., and Tembine, H.: Self-
organized connected objects: Rethinking QoS provisioning for IoT services.
IEEE Communications Magazine 55(9), 41–47 (2017)
[3] Melcherts, H.E.: The internet of everything and beyond. Human Bond Com-
munication: The Holy Grail of Holistic Communication and Immersive
Experience p. 173 (2017)
[4] Ramos, J.L.H., Geneiatakis, D., Kounelis, I., Steri, G., and Fovino, I.N.:
Toward a data-driven society: A technological perspective on the development
of cybersecurity and data protection policies. IEEE Security & Privacy (2019)
[5] Gelenbe, E., Gorbil, G., Tzovaras, D., Liebergeld, S., Garcia, D., Baltatu,
M., and Lyberopoulos, G.: Security for smart mobile networks: The nemesys
approach. In: Privacy and Security in Mobile Systems (PRISMS), 2013 Inter-
national Conference on. pp. 1–8. IEEE (2013)
[6] Ratasuk, R., Prasad, A., Li, Z., Ghosh, A., and Uusitalo, M.A.: Recent
advancements in M2M communications in 4G networks and evolution
towards 5G. In: Proc. 18th IEEE International Conference Intelligence in
Next Generation Networks (ICIN). pp. 52–57. Paris, France (Feb 2015).
https://doi.org/10.1109/ICIN.2015.7073806
[7] Collen, A. et al.: Ghost: Safeguarding home IoT environments with person-
alised real-time risk control. In: Recent Cybersecurity Research in Europe:
Proceedings of the 2018 ISCIS Security Workshop, Imperial College London.
vol. LNCCIS 821. Springer Verlag (2018)
[8] Gelenbe, E. and Kadioglu, Y. M.: Energy life-time of wireless nodes with
network attacks and mitigation. 2018 IEEE International Conference on
Communications Workshops (ICC Workshops), Kansas City, MO, pp. 1–6,
(2018). doi: 10.1109/ICCW.2018.8403561.
[9] He, D., Chan, S., Qiao, Y., and Guizani, N.: Imminent communication secu-
rity for smart communities. IEEE Communications Magazine 56(1), 99–103
(Jan 2018). https://doi.org/10.1109/MCOM.2018.1700587
References 101
[10] Kalkan, K. and Zeadally, S.: Securing internet of things (IoT) with software
defined networking (sdn). IEEE Communications Magazine (2017). https:
//doi.org/10.1109/MCOM.2017.1700714
[11] Lu, X., Spear, M., Levitt, K., Matloff, N.S., and Wu, S.F.: A synchronization
attack and defense in energy-efficient listen-sleep slotted MAC protocols. In:
Emerging Security Information, Systems and Technologies, 2008. SECUR-
WARE’08. Second International Conference on. pp. 403–411. IEEE (2008)
[12] Pirretti, M., Zhu, S., Vijaykrishnan, N., McDaniel, P., Kandemir, M., and
Brooks, R.: The sleep deprivation attack in sensor networks: Analysis and
methods of defense. International Journal of Distributed Sensor Networks
2(3), 267–287 (2006)
[13] Purdy, G.: Iso 31000: 2009 — setting a new standard for risk management.
Risk Analysis: An International Journal 30(6), 881–886 (2010)
[14] Nurse, J.R., Creese, S., and De Roure, D.: Security risk assessment in internet
of things systems. IT Professional 19(5), 20–26 (2017)
[15] Sen, S., Koo, J., and Bagchi, S.: Trifecta: Security, energy-efficiency, and com-
munication capacity comparison for wireless IoT devices. IEEE Internet Com-
puting 22(1), 74–81 (2018)
[16] Gelenbe, E., Domanska, J., Czachórski, T., Drosou, A., and Tzovaras, D.:
Security for internet of things: The SerIoT project. In: 2018 International
Symposium on Networks, Computers and Communications, ISNCC 2018,
Rome, Italy, June 19-21, 2018. pp. 1–5. IEEEXplore (2018), https://doi.org/
10.1109/ISNCC.2018.8531004
[17] Battaglia, P., Hamrick, J.B.C., Bapst, V., Sanchez, A., Zambaldi, V., Mali-
nowski, M., Tacchetti, A., Raposo, D., Santoro, A., Faulkner, R., Gulcehre,
C., Song, F., Ballard, A., Gilmer, J., Dahl, G.E., Vaswani, A., Allen, K., Nash,
C., Langston, V.J., Dyer, C., Heess, N., Wierstra, D., Kohli, P., Botvinick, M.,
Vinyals, O., Li, Y., and Pascanu, R.: Relational inductive biases, deep learning,
and graph networks. arXiv (2018), https://arxiv.org/pdf/1806.01261.pdf
[18] Zhang, Z., Cui, P., and Zhu, W.: Deep learning on graphs: A survey. arXiv
preprint arXiv:1812.04202 (2018)
[19] Brun, O., Yin, Y., and Gelenbe, E.: Deep learning with dense random neu-
ral network for detecting attacks against IoT-connected home environments.
Procedia Computer Science 134, 458–463 (2018)
[20] Gelenbe, E., Liu, P., and Lainé, J.: Genetic algorithms for route discovery.
IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics)
36(6), 1247–1254 (2006)
[21] Mehmood, Y., Ahmad, F., Yaqoob, I., Adnane, A., Imran, M., and Guizani, S.:
Internet-of-things-based smart cities: Recent advances and challenges. IEEE
Communications Magazine 55(9), 16–24 (2017)
102 IoT Network Risk Assessment and Mitigation
[22] Francois, F. and Gelenbe, E.: Towards a cognitive routing engine for software
defined networks. In: Communications (ICC), 2016 IEEE International Con-
ference on. pp. 1–6. IEEE (2016)
[23] Dobson, S. et al.: A survey of autonomic communications. ACM Transactions
on Autonomous and Adaptive Systems (TAAS) 1(2), 223–259 (2006)
[24] Galis, A., Denazis, S., Brou, C., and Klein, C.: Programmable Networks for
IP Service Deployment. Artech House Inc. (2004)
[25] Rubio-Loyola, J., Astorga, A., Serrat, J., Chai, W.K., Mamatas, L., Galis, A.,
Clayman, S., Cheniour, A., Lefevre, L., Mornard, O., Fischer, A., Paler, A.,
and Meer, H.D.: Platforms and software systems for an autonomic internet.
In: 2010 IEEE Global Telecommunications Conference GLOBECOM. IEEE
(2010)
[26] Tsarouchis, C., Denazis, S., Kitahara, C., Vivero, J., Salamanca, E., Magana,
E., Galis, A., Manas, J.L., Carlinet, L., Mathieu, B., and Koufopavlou, O.: A
policy-based management architecture for active and programmable networks.
IEEE Network 17(3), 22–28 (2003)
[27] Joint Task Force Transformation Initiative: Guide for applying the risk man-
agement framework to federal information systems: a security life cycle
approach. Tech. Rep. NIST SP 800-37r1, National Institute of Standards and
Technology (Jun 2014). https://doi.org/10.6028/NIST.SP.800-37r1, https:
//nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-37r1.pdf
[28] USC, S.: 3502. CNSSI-4009
[29] Radanliev, P., De Roure, D.C., Nicolescu, R., Huth, M., Montalvo, R.M.,
Cannady, S., and Burnap, P.: Future developments in cyber risk assessment
for the internet of things. Computers in Industry 102, 14–22 (2018)
[30] Martin, B. and Coley, S.: Common weakness scoring system (CWSS). Inter-
net, http://cwe.mitre.org/cwss (2014)
[31] Mell, P., Scarfone, K., and Romanosky, S.: Common vulnerability scoring sys-
tem. IEEE Security & Privacy 4(6), 85–89 (2006)
[32] Qu, Y. and Chan, P.: Assessing Vulnerabilities in Bluetooth Low Energy (BLE)
Wireless Network Based IoT Systems. In: 2016 IEEE 2nd International Con-
ference on Big Data Security on Cloud (BigDataSecurity), IEEE Interna-
tional Conference on High Performance and Smart Computing (HPSC),
and IEEE International Conference on Intelligent Data and Security (IDS).
pp. 42–48. IEEE, New York, NY, USA (Apr 2016). https://doi.org/10.1109/
BigDataSecurity-HPSC-IDS.2016.63, http://ieeexplore.ieee.org/document/
7502262/
References 103
[43] Miettinen, M., Marchal, S., Hafeez, I., Asokan, N., Sadeghi, A.R., and
Tarkoma, S.: IoT sentinel: Automated device-type identification for secu-
rity enforcement in IoT. In: 2017 IEEE 37th International Conference on
Distributed Computing Systems (ICDCS), 5–8 June 2017. pp. 2177–2184.
IEEE (2017). https://doi.org/10.1109/ICDCS.2017.284
[44] Barrera, D., Molloy, I., and Huang, H.: Standardizing IoT network secu-
rity policy enforcement. In: Workshop on Decentralized IoT Security and
Standards 2018 (01 2018). https://doi.org/10.14722/diss.2018.23007, http:
//wp.internetsociety.org/ndss/wp-content/uploads/sites/25/2018/07/
diss2018_7_Barrera_paper.pdf
[45] NIST: Securing Small-Business and Home Internet of Things Devices: NIST
SP 1800-15 (2019)
[46] Garcia-Morchon, O., Kumar, S.S., and Sethi, M.: Internet of Things Security:
State of the Art and Challenges (RFC 8576) (2019)
[47] Simon, D., Ph.D., D.B.D.A., and Eronen, P.: Extensible Authentication Pro-
tocol (EAP) Key Management Framework. RFC 5247 (Aug 2008). https:
//doi.org/10.17487/RFC5247, https://rfc-editor.org/rfc/rfc5247.txt
[48] Rao, S., Chendanda, D., Deshpande, C., and Lakkundi, V.: Implementing
lwm2m in constrained IoT devices. In: 2015 IEEE Conference on Wireless
Sensors (ICWiSe). pp. 52–57. IEEE (2015)
[49] Selander, G., Mattsson, J., Palombini, F., and Seitz, L.: Object security for
constrained restful environments (oscore) (July 2019), https://tools.ietf.org/
html/rfc8613
[50] Neisse, R., Hernández-Ramos, J.L., Matheu, S.N., Baldini, G., and Skarmeta,
A.: Toward a blockchain-based platform to manage cybersecurity certification
of IoT devices. In: 2019 IEEE Conference on Standards for Communications
and Networking (CSCN). pp. 1–6. IEEE (2019)
[51] Khodadadi, F., Dastjerdi, A.V., and Buyya, R.: Internet of things: an overview.
In: Internet of Things, pp. 3–27. Elsevier (2016)
[52] Qin, J., Ma, Q., Shi, and Y., Wang, L.: Recent advances in consensus of multi-
agent systems: A brief survey. IEEE Transactions on Industrial Electronics
64(6), 4972–4983 (2016)
[53] Gupta, J.K., Egorov, M., and Kochenderfer, M.: Cooperative multi-agent
control using deep reinforcement learning. In: International Conference on
Autonomous Agents and Multiagent Systems. pp. 66–83. Springer (2017)
DOI: 10.1561/9781680836837.ch6
Chapter 6
Industrial IoT applications in various domains including large office sites and trans-
port infrastructure are getting more complex with multiple computer-based man-
agement systems handling the workload. These systems can generate data that can
be used to introduce cognitive capability in many critical tasks. Handling the con-
cerns related to privacy and security is important to enable any complex function-
ality [1–3]. For this reason, the CHARIOT IPSE (IoT Privacy and Safety Engine)
shows a solution that links the privacy, security, and safety supervision functionality
in a novel way.
In summary, the safety supervision engine will propose a mechanism to pro-
cess and handle the signals generated by sensors for safety-related scenarios. Lately,
IoT deployments are moving away from an approach of offloading the cognitive
capabilities to one central approach towards distributed intelligence with mul-
tiple nodes handling the cognitive capabilities of the IoT deployment [1, 4].
The implementation of machine learning-based policies and propagating the
changes on the systems require edge cloud orchestration done by the system [5, 6].
To align with the CHARIOT objectives, CHARIOT develops a safety supervision
105
106 Chariot-integrated Approach to Safety, Privacy, and Security
engine that can run on multiple Fog nodes of an IoT setup. In addition, the safety
supervision engine provides collaborative optimization capability to dynamically
change priorities of sensor based on the operation mode of the system. By doing
dynamic prioritization of signal transfer to cloud, the scalability issue which is seen
as one of the major research challenges for industrial IoT [7] is tackled.
The CHARIOT Privacy engine uses blockchain to avoid privacy breaches
in the IoT network. The Privacy engine encrypts the data on a southbound
dispatcher (sensor side in the architecture) by using keys distributed on the
blockchain network. It keeps the policies and the access rights for particu-
larly sensitive data in the policy repository and enforces the access rights for
CHARIOT. The Privacy engine is particularly suitable for large industrial IoT
deployment environments where multiple contractors may access different data
streams [8, 9].
The CHARIOT Security Engine checks the firmware updates and related
changes in network and network transmission to make sure that the data shared
in the system is bona fide with no security breaches [10, 11]. CHARIOT uses open
source tools and heuristics where historical firmware binary data will be used to
identify updates with security vulnerabilities and inform the operator.
In this chapter, we will provide an overview of the core capabilities of CHARIOT
safety supervision engine, privacy engine, and security engine. An overview of the
Engine’s capabilities is also provided.
Figure 6.1. A sample industrial IoT topology for a work campus environment.
• Zone: Zone is any physical area that includes some networked components.
A zone might be a train (Trenitalia), building floor (IBM), or airport room
(Athens Airport). A zone may be hierarchically attached to a parent zone (e.g.,
a room in a building floor). A zone may have an associated CHARIOT Fog
Node. All the Fog nodes run the safety supervision engine separately to enable
fault tolerance.
• Gateway: Gateway is any computational node that has network connection
with one or more sensors. A gateway in CHARIOT can also be a fog node
with relatively more computational power. These fog nodes can store the par-
tial state of the IoT Setup.
• Sensor: A sensor is any networked component that can collect data. The
sensor can be simple or complicated. A sensor can be optionally part of a
gateway. This relation is optional since some sensors are complicated enough
to have direct connection.
108 Chariot-integrated Approach to Safety, Privacy, and Security
The three layers of elements can have hierarchical dependencies. For exam-
ple, a zone might have two gateways connected to multiple sensors. CHARIOT
uses an IoTL (Internet of Things Language – A domain-specific language devel-
oped for CHARIOT Project) to keep all these dependencies consistent across
various nodes in the industrial IoT setup. The dependencies are kept consis-
tent on any node that is keeping the state of the industrial IoT Setup. In addi-
tion, using IoTL, one can easily define custom types and hierarchies for custom
scenarios.
IoTL is a language for communicating the state of the system. Its grammar is
created using ANTLR, and a Python 3 parser is generated by ANTLR. IoTL is used
to share and track the topology of an IoT setup across different components. This
language enables communicating safety-related actions and state changes between
multiple nodes of a system. The core specs of the IoTL are listed as follows:
The CHARIOT Safety Supervision Engine can run on multiple nodes within
a distributed IoT deployment as seen in Figure 6.2. The nodes are associated with
zones in the safety supervision model. In a deployment, more than one node can
be deployed. Deployment of the same runtime on multiple instances ensures the
system availability even when the network is partitioned during a safety hazard. The
zone hierarchy is the same as the zone hierarchy in this setup. The safety supervision
engine is a web application that provides a REST-APIs and connectors for live
data streams. All the Safety Supervision Engine functionality is written on Python
Ecosystem and using open source Python libraries. Flask framework was chosen
due to its compactness, popularity, and easy integrability with machine learning
functionality written in Python Language.
The CHARIOT Safety Supervision Engine can run on multiple networked
nodes in complicated IoT deployments. In this case, multiple fog nodes can be
associated with different zones. Based on the hierarchy and the dependencies, each
fog node can observe only associated parts of the system, and they synchronize
periodically to keep the state of the system consistent across the nodes. Every change
Figure 6.3. A simple example of global and local state in the safety supervision engine.
is hierarchically synchronized across all the computation nodes across the Fog Net-
work using the commands API with the “!sync” command. The high-level repre-
sentation of the nodes can be seen in Figure 6.2. The communication across the
safety supervision platform instances is carried out. In the current living lab envi-
ronments, only one Fog node exists for purposes of the CHARIOT proof of concept
and industrial cases.
The dependency relation across zones is kept within the state of the IoT Deploy-
ment. In the case of synchronization across different fog nodes, only state relevant
to one or more dependent component is shared with a fog node. For example, if
there is no relevant change in the state for one fog node, the state is not updated to
conserve resources. In Figure 6.3, we provide a simple example for this state shar-
ing mechanism. In this simple scenario, the global view has a bunch of gateways
associated with a single zone in an IoT deployment. The state relevant to the gate-
way named pi1 is kept by the fog node associated with that gateway. In the case of a
state change, only the relevant portions to the element are shared. By definition, the
global state associated with the root zone “designstudio” is associated with all the
elements in the IoT deployment, so it receives all the updates. On the contrary, the
fog node associated with pi1 only receives if a portion of the deployment relevant
to it has been updated.
“Privacy” has been a fundamental issue since the early days of computing, especially
IoT. IoT is becoming part of everyday living and everything is being shared over
internet—photos, videos, health measurements, gaming records, etc—privacy IS
the area of matter of worry. In IoT, privacy has different scenario as compared to
others as mode of data collection, mining, and provisioning is different. Technically,
The CHARIOT Privacy Engine 111
data can be collected in such a manner that it may make difficult to interpret and
realize that personal data are ready to be compromised. Privacy of individuals in
IoT is a serious compromise factor. Within CHARIOT, privacy is given special
attention by ensuring that the system has control over:
• Data being collected by individual sensors should enter the system if only the
sensor is known and registered in the topology.
• If data are from a known sensor, data encryption must be applied using a
public key stored in a blockchain PKI.
1. Principal: Defines the person or entity that has listened on CHARIOT for
new messages. A principal example is an Industrial System Gateway.
2. Resource: Defines the things that the ACL rule applies to. Some examples
are the Sensors, Logs, or the alerts.
3. Operation: Identifies the action that the rule governs. CHARIOT supports
only READ.
4. Guard: Is the responsible service to apply ACL rules, for us the Privacy
Engine.
5. Action: Identifies the action of the rule. It must be one of: ALLOW, DENY.
Figure 6.5. Define and external service and its public key.
• Listener
Listener is a component waiting for any sensor measurement—data in ana-
logue or digital format, within the IoT system.
The CHARIOT Privacy Engine 113
• Topology Synchronization
This component is responsible to keep information describing new sen-
sors topology within the IoT system and any update in sensor description
(location, sensitivity, etc.). Also, all valid recipients should be registered in
this component.
• Inspector
Inspector is the component responsible to raise an Alert Flag every time a
message is generated by a “Sensitive” labeled sensor, in order all actors within
the IoT Network to be aware of that type of sensors. To achieve this, the
network administrator should label a sensor as “Sensitive” so as to prevent a
privacy information violation.
• Blockchain PKI client
This component assures verification of the Public Key used by the recipient.
If there is a breach, an alert will be raised.
• Alert Generator
Alert Generator, if triggered, will generate an alert and will also keep record
of all alerts.
• Filter
CHARIOT Admin should define the Access List Controls with IoTL in order
to guard user data. Filtering rule (Figures 6.9) will effectively avoid leakage of
114 Chariot-integrated Approach to Safety, Privacy, and Security
In today’s context, smart devices belonging to the Internet of Things ecosystem (IoT
ecosystem) are increasingly present. However, the technology companies producing
these evolved systems tend to underestimate the potential risks to which they are
exposed to, since the latter are connected almost continuously to the internet. The
greatest danger is caused by superficial behavior conducted by the manufacturers
themselves who are oriented to release the firmware updates with an excessively low
frequency and sometimes, fortunately only in extreme cases so far, the updates are
not released at all. This superficial behavior allows hackers to take advantage of a
vast amount of potentially vulnerable devices.
The CHARIOT Security Engine 115
The firmware, an expression comprised of the terms firm (stable) and ware (com-
ponent), is a program present within each electronic component that performs a
critical task. Its fundamental role is to start the function of the same component
allowing it to communicate with further integrals or cards present in the device
in which they are installed. This determines the importance for releasing firmware
updates; otherwise, an IoT device that has not been updated since its purchase can
become dangerous for the consumer when cybercriminals discover a vulnerability
of the system. This usually occurs when a device manufacturer decides to allocate
its internal resources to the production of a new product to be introduced to the
market, instead of the development and release of updated firmware versions for
existing products. Consequently, the most significant risk for the consumer is find-
ing obsolete and potentially dangerous security hardware.
Failure to update the software should not always be attributed to the manufac-
turer; the end user plays an active and harmful role. Quite often, due to personal
inexperience or lack of knowledge for the “smart” product, the consumer does not
check for updates and does not install the firmware versions released and recom-
mended by the manufacturer. To overcome this particular situation and reduce the
security risks related to the exposing the security of the software platforms, the most
advanced devices, including computers, are programmed to perform automatic
updates without the need of any customer intervention. Unfortunately, updating
and protecting the security features of the systems belonging to the Internet of
Things is not, and will not be, trivial to the implementation of the proposed solu-
tions of this project.
To achieve a high-level security for IoT devices, it is necessary to introduce a
shared standard that can guarantee a common methodology for the design and
development of the device, but above all, issues that are related to the updates.
On the other hand, however, updates can lead to vulnerability (vulnerable sur-
faces), potentially exploitable leaving a strong attack vector.
• Reverse-engineer the entire firmware, extract the file system, and understand
how the entire device works, including knowing the possible use of known-
to-be-vulnerable out-of-date API/libraries or unknown exploitable vulnera-
bilities in the firmware code itself;
• Insert a firmware backdoor, making the device covertly connected to a mali-
cious Command & Control server;
• Change the device behavior, altering its performance, without altering the
telemetry values reported to administrators;
• Find hard-coded private symmetric-cryptography keys/passwords/usernames
or private certificates used to encrypt communications between the device
and other systems; the attacker may be able to eavesdrop these private com-
munications;
• Roll back the firmware to a previous legitimate version with known vulner-
abilities he/she wants to exploit; this attack is particularly insidious because
the pushed firmware is authentic, so it can easily survive most of the in-place
controls, as usually they tend to check just the firmware source and/or the
firmware integrity.
In parallel with the above CHARIOT IPSE components, CHARIOT develops the
Cognitive System and related methodology as the accompanying supervision, ana-
lytics, and prediction models, enabling high security and integrity of Industrial
IPSE Dashboard and User Interfacing 117
IoT devices and systems. This includes the development of the intelligent behav-
ior of the CHARIOT platform to enable IoT critical systems’ secure interaction.
In this direction, a web-of-things environment is being developed interacting with
heterogeneous IoT devices and systems. This includes the development of the
suitable interfaces for the topological and functional behavior models for the IoT
system devices (components) as well as additional safety and privacy structures
that communicate through open APIs and include security services employing
Blockchain, IoT security profiles, and fog services.
This module will be mainly used to predict and avoid potentially endangering
behavior in the IoT system that is now handled in an optimized way in cooperation
with safety critical systems and run-time environments of the IoT system itself.
This is using the existing IoT platform (IBM’s Watson IoT) to demonstrate the
CHARIOT innovative concept and capability and support integration with other
safety, privacy, and machine-learning cloud services via relevant open APIs, thus
supporting third-party integration and innovation.
At the same time, a security, privacy, and safety awareness configurable dashboard
is being developed in order to remotely monitor in near real time the various IoT
gateways and sensors in the CHARIOT IoT network. This will be the actual end-
user interface to the CHARIOT solution and platform. The dashboard has been
designed as part of the entire IPSE implementation and will also be used for both
understanding of the IoT ecosystem topology as well as for the post data analytical
purposes to assist in the refinement and improvements of PSS policies.
This advanced intelligence web-based dashboard utilizes the latest state-of-the-
art web technologies in order to deliver rich content information to the LL users and
achieve cross-browser and multi-device compatibility. The web development has
been performed using the .NET Core framework and with the utilization of user
interface implementation technologies like HTML, CSS, and JavaScript. More-
over, support on publish-subscribe-based messaging protocols as well as REST-API
have also been utilized. Further to that, the dashboard follows rules regarding user
friendliness, responsiveness, and configurability as web solution, based on the LLs
needs. The Dashboard supports various components that are expected to support
and enhance the user experience as well as the actual system operation and configu-
ration such as interactive widgets, customizable entries, multiple components, data
filtering, gauss controls and fusion charts, map, and top view visualization controls,
etc. The following Figure presents a snapshot of this Dashboard for monitoring the
devices in a train wagon.
As presented in the following (Figure 6.11), via the IPSE dashboard, the status of
the IPSE Engines and fog node services can also be monitored. “Health visibility”
(Figure 6.12) is a service which is checking the “live” status of each CHARIOT
component in a graphical way.
118 Chariot-integrated Approach to Safety, Privacy, and Security
Moreover, via IPSE dashboard, the end user is able to perform administrative
and management activities related to:
Figure 6.13. Firmware update and process visibility via the dashboard.
Each engine of the CHARIOT IPSE provides several new innovations over the
state-of-the-art in industrial IoT. Distributed safety supervision introduced by
the safety supervision engine ensures fault-tolerant operation of the cognitive
safety assurance components such as anomaly detectors and other online machine
learning-based models. The relevant state is shared across multiple components in
the Fog network to enable edge cloud orchestration by using the domain-specific
language developed for IPSE. Having multiple active computation nodes in the
industrial IoT network enables fault tolerance during safety hazards against partial
network outages.
The Privacy engine ensures data privacy through encrypting data at the source,
namely at the southbound dispatcher through a PKI supported by CHARIOT
blockchain infrastructure. For industrial IoT setups with multiple contractors, our
approach provides detailed privacy policies for every sensor data stream. On top of
that, through the privacy-related constructs in the domain-specific language, pri-
vacy policies can be enforced even during partial network outage. Using CHARIOT
Blockchain solution for handling PKI provides secure encryption for the multiple
data streams handled by CHARIOT.
Finally, the security engine applies machine learning to identify security prob-
lems in possible firmware updates dynamically, and the signed firmware updates
120 Chariot-integrated Approach to Safety, Privacy, and Security
are distributed across the industrial IoT setup. Deployment of the integrated IPSE
provides CHARIOT living labs the capability of addressing their business-critical
problems related to safety supervision, security, and privacy.
Acknowledgments
This project has received funding from the European Union’s Horizon 2020
research and innovation program (No. 780075). The authors acknowledge the
research outcomes of this publication belonging to the CHARIOT consortium.
References
[1] Ranjan, Rajiv, et al. “The Next Grand Challenges: Integrating the Internet of
Things and Data Science.” IEEE Cloud Computing 5.3 (2018): 12–26.
[2] Sfar, Arbia Riahi, et al. “A roadmap for security challenges in the Internet of
Things.” Digital Communications and Networks (2018): 118–137.
[3] Sanchez-Iborra, Ramon, and Maria-Dolores Cano. “State of the art in LP-
WAN solutions for industrial IoT services.” Sensors 16.5 (2016): 708.
[4] Pradhan, Dhiraj K., and Sudhakar M. Reddy. “A fault-tolerant communica-
tion architecture for distributed systems.” IEEE transactions on Computers 9
(1982): 863–870.
[5] Ukil, Arijit, Soma Bandyopadhyay, and Arpan Pal. “Iot-privacy: To be private
or not to be private.” Computer Communications Workshops (INFOCOM
WKSHPS), 2014 IEEE Conference on. IEEE, 2014.
[6] Shi, Weisong, et al. “Edge computing: Vision and challenges.” IEEE Internet
of Things Journal 3.5 (2016): 637–646.
[7] Sun, Xiang, and Nirwan Ansari. “EdgeIoT: Mobile edge computing for the
Internet of Things.” IEEE Communications Magazine 54.12 (2016): 22–29.
[8] Da Xu, Li, Wu He, and Shancang Li. “Internet of things in industries: A sur-
vey.” IEEE Transactions on industrial informatics 10.4 (2014): 2233–2243.
[9] Porambage, Pawani, et al. “The quest for privacy in the internet of things.”
IEEE Cloud Computing 2 (2016): 36–45.
[10] Sadeghi, Ahmad-Reza, Christian Wachsmann, and Michael Waidner. “Secu-
rity and privacy challenges in industrial internet of things.” Design Automa-
tion Conference (DAC), 2015 52nd ACM/EDAC/IEEE. IEEE, 2015.
[11] Zhao, Yan Ling. “Research on data security technology in internet of things.”
Applied Mechanics and Materials. Vol. 433. Trans Tech Publications, 2013.
DOI: 10.1561/9781680836837.ch7
Chapter 7
7.1 Introduction
attempt to integrate this evolving technology and the exponential growth of IoT
by overcoming significant hurdles such as dynamicity, scalability, heterogeneity,
and end-to-end security and privacy. The introduction of digital technologies in
economic and societal processes is key to addressing economic and societal chal-
lenges such as aging of population, ensuring societal cohesion, and sustainable
development. IoT appears to be an important pillar of 5G. Global networks like
IoT create enormous potential for new generations of IoT applications, by leverag-
ing synergies arising through the convergence of consumer, business, and industrial
internet, and creating open, global networks connecting people, data, and things.
A series of innovations across the IoT landscape have converged to make IoT prod-
ucts, platforms, and devices technically and economically feasible. However, despite
these advances, significant business and technical hurdles must be overcome before
the IoT’s potential can be realized.
Some important challenges and complexities include:
1. SEMIoTICS: Smart End-to-end Massive IoT Interoperability, Connectivity and Security: http://www.
semiotics-project.eu
Background and Challenges 123
The background of the SPDI properties and the necessity for the usage of pat-
terns to drive the fulfillment of the end users’ goals in the directions of security,
privacy, dependability and interoperability for the IoT are provided in this section.
Furthermore, the IoT challenges stemming from each of the four domains are also
briefly mentioned in order to highlight that being able to address them system-wide
and in a by-design-fashion is of paramount importance in industrial or larger IoT
environments.
124 Pattern-driven Security and Privacy in IoT
Further to being a legal or business challenge, privacy could also be seen as a human
right in either way privacy matters in the IoT domain and as such we discuss some
privacy challenges which have to be taken care of at the architectural level:
The latter mentioned challenge of monitoring allows the data subject to exercise
some form of oversight, which adds to the transparency of the data processing.
different stakeholder domains and to reuse proven solutions for problems which
occur over and over again. The initial works date back to the 1977 book “A Pat-
tern Language: Towns, Buildings, Construction” by Alexander et al. [46], where
the concept of reusable design solutions for architectural problems was intro-
duced. The pattern approach has ever since been successfully applied in other
domains, e.g., software design [47], security [48], privacy [49]. Here, SEMI-
oTICS, especially, builds upon ideas from similar pattern-based approaches used
in service-oriented systems [50, 51], cyber-physical systems [52], and networks
[53, 54], covering more aspects in addition to Security, and especially privacy
[55, 56] and also providing guarantees and verification capabilities that span both
the service orchestration and deployment perspectives, as detailed in Section 7.3
above.
P1 ∧ P2 ∧ . . . ∧ Pn −
→ Pn + 1 (7.1)
The run-time adaptations that can be enabled by SPDI patterns may take three
forms: (a) to replace particular components of an orchestration; (b) to change the
structure of an orchestration; and (c) a combination of (a) and (b).
Figure 7.3. Pattern reasoning engines at all layers of the SEMIoTICS architecture.
Backend Pattern Engine may receive fact updates from the individual Pattern
Engines present at the lower layers (Network & Field), allowing it to have an
up-to-date view of the SPDI state of said layers and the corresponding com-
ponents
• Network Pattern Engine: Integrated in the SDN controller to enable the capa-
bility to insert, modify, execute, and retract network-level patterns at design
or at run time. It is supported by the integration of all required dependencies
within the network controller, as well as the interfaces allowing entities that
interact with the controller to be managed based on SPDI patterns at design
and at run time.
• Field Layer Pattern Engine: Typically deployed on the IoT/IIoT gateway, able
to host design patterns as provided by the Pattern Orchestrator. Since the
compute capabilities of the gateway can be limited, the module is able to host
patterns in an executable form compared to the pattern rules as provided in
the other layers.
Pattern Enforcement’s and Evaluation in SEMIoTICS Use Cases 133
however, in the event of an emergency, e.g., a patient fall, this information needs
to be available to a selected set of authorized users, e.g., in this case medical per-
sonnel. The process, also depicted in Figure 7.5, includes a doctor’s initial request
to a service where the request is checked for validity by a policy enforcement point
(PEP). Asking a SEMIoTICS component called Security Manager is only needed if
the service would be exposed by the field level and after the synchronization and
evaluation of the Security Manager. Based on this information, the PEP will then
decide if the request or the response will be granted or not.
With respect to the scope of privacy, the SEMIoTICS Security Manager as
depicted in Figures 7.3 and 7.5 acts to ensure that the privacy properties imposed by
the pattern are technically enforced. In detail, this happens by enabling the interac-
tion of a Security Manager located in the back-end with the Pattern Engine through
the associated privacy pattern rule as depicted in Figure 7.5. The goal is that the
Pattern Enforcement’s and Evaluation in SEMIoTICS Use Cases 135
Pattern Engine is able to check if the current policy used to make the decisions
inside the Backend Security Manager is conforming to the SPDI Patterns as speci-
fied by the application. The Pattern Engine thus periodically makes requests to the
Security Manager to obtain the necessary information to judge the quality of the
currently enforced policy. For example, to check that no one has access to a patient’s
location, it queries to obtain a list of entity that would be granted access. Based on
the response, the appropriate pattern rule expressing the SPDI pattern can reason
on the answers received from the Security Manager. For example, under normal
health conditions of all patients that response shall be an empty list. By reasoning
over the response, the Pattern Engine is able to identify if the policy being enforced
in SEMIoTICS is compliant with the privacy pattern; this is done without having
to check the actual technical enforcement (e.g., the access control policy as specified
in the PEP), but rather by checking its effectiveness; further, it is able to reflect this
to the outside via an API call that will allow other components to retrieve the SPDI
status.
7.5 Conclusion
Acknowledgments
This work has received funding from the European Union’s Horizon 2020 research
and innovation program under grant agreements No. 780315 (SEMIoTICS).
References
Chapter 8
8.1 Introduction
The next-generation IoT systems need to perform distributed processing and coor-
dinated behavior across IoT, edge, and cloud infrastructures; manage the closed
loop from sensing to actuation: and cope with vast heterogeneity, scalability, and
dynamicity of IoT systems and their environments (Ferry et al., 2018). To unleash
the full potential of IoT, it is essential to facilitate the creation and operation of
trustworthy Smart IoT Systems or, for short, TSIS. TSIS typically operate in chang-
ing and often unpredictable environments. Thus, the ability of these systems to
continuously evolve and adapt to their new environment is essential to ensure and
increase their trustworthiness, quality, and user experience. Besides, by 2021, the
number of “connected things” will grow to 25 billion, according to Gartner.1 Thus,
processes that were formerly run by humans will be automated, making it much
more difficult to control data ownership, privacy, and regulatory compliance.
Yan et al. (2014) described the different dimensions of trust for IoT sys-
tems, concluding that risk management is essential to guarantee trustworthiness.
1. https://www.gartner.com/en/newsroom/press-releases/2018-11-07-gartner-identifies-top-10-strategic-iot-
technologies-and-trends
143
144 Enabling Continuous Privacy Risk Management in IoT Systems
Poor risk management together with a reactive strategy usually forces companies
to continuously re-factor application architectures to improve software quality
and security, incurring high re-implementation costs (Boehm and Turner, 2003).
In general, there is a lack of solutions to support continuous control of risks through
evidence collection. Companies have little control on actual effectiveness of the
mitigation actions defined during risk management processes. Besides, many com-
panies fill this gap by using manual procedures based on storing all the information
in spreadsheets, by departments and locally (Ahmad et al., 2012). This approach
rapidly turns inefficient as projects or teams grow.
In parallel, GDPR discusses data protection by design and by default, remarking
that it is essential to consider privacy from the beginning to address related issues
successfully. This is especially true in the IoT arena, where technologies are not
consolidated yet and mixing legal requirements with a deep technical understand-
ing is challenging. In particular, understanding privacy-related vulnerabilities and
including privacy considerations in a continuous risk management process is dif-
ficult. On the one hand, knowledge related to vulnerabilities connected to privacy
issues is not so commonplace. On the other hand, continuous evidence collec-
tion to support risk management is usually key, but most monitoring approaches
focus on collecting evidences from the TSIS infrastructure or technical architec-
tural components. However, privacy-related risks are usually detected by analyzing
functional descriptions of the system (Wuyts, 2015) (e.g., data flows). Connecting
this functional level with the components of the architecture that are being moni-
tored is not trivial. Recognizing the overlap between privacy and security is key to
determining when existing security risk models may be applied to address privacy
concerns (Brooks et al., 2017).
In this chapter, we establish the basis to create an Automated Vulnerability
Detector (AVD) for early detection of privacy issues. We also improve continu-
ous risk management in TSIS development by embedding privacy-related risks
explicitly through the combined use of models for both the architecture and the
data flow implemented on the components of the architecture. We present a new
approach that, through linking GeneSIS (Ferry et al., 2019) models and Data Flow
Diagrams (DFDs) (DeMarco, 1979), improves risk assessment process for privacy-
related risks. We achieve this by enabling the use of the information that is typically
collected from the infrastructure to control security, thanks to the link that LIND-
DUN (Wuyts, 2015) establishes between privacy and security threads in STRIDE
(Shostack, 2014).
This chapter is organized as follows. Section 8.2 provides an analysis of the
state of the art. Section 8.3 presents our methodology. Then, Section 8.4 describes
the basic steps to create an AVD taking LINDDUN as the baseline reference.
In Section 8.5, we describe the mapping between the architecture model and DFDs
as well as our risk model. Finally, in Section 8.6, we draw some conclusions.
State of the Art 145
There exist quantitative risk methodologies and tools, like RiskWatch2 or ISRAM
(Karabacak and Sogukpinar, 2005) and qualitative risk methodologies such as
OCTAVE (Alberts and Dorofee, 2002), CORAS (Lund et al., 2010), or STRIDE
(Shostack, 2014). Traditional risk management methodologies focus on the assess-
ment of risks at a singular stage in time and do not address the challenge of man-
aging continuously evolving risk profiles.
Continuous change management in software development has been studied
from different perspectives. As an example, Fitzgerald and Stol (2017) proposed
a roadmap and agenda for continuous software engineering. In particular, there is
a growing interest on security and privacy, and this has motivated work on con-
tinuous security (Merkow and Raghavan, 2011), which was proposed to prioritize
security throughout the whole software development life cycle.
Previous work discussed risk management in complex and evolving environ-
ments. For instance, Omerovic (2016) and Gupta et al. (2015) propose manag-
ing risk related to accountability, assurance, agility, or financial aspects in multi-
cloud applications. Risks analysis is used to guide the selection of cloud service
providers. Shrivastava and Rathod (2017) present a risk management framework
for distributed agile development, studying risks related to software development
life cycle, project management, group awareness, etc. However, they focus on ana-
lyzing risk factors that represent a threat to the successful completion of a soft-
ware development project, rather than risk related to non-functional requirements
such as security or privacy. Aslam et al. (2017) performed a systematic literature
review on risks and control mechanisms for distributed software development.
Moran (2014) explicitly tackles issues related to risk management for agile software
development. Moran proposes a risk modified kanban board and user story map.
Finally, Muntés-Mulero et al. (2018) propose a framework to manage risks that
supports collaboration, agility, and continuous development. In the MUSA Risk
Assessment tool,3 in order to assess the risks in the different components of an appli-
cation, the tool asks users to choose among predefined threats that may potentially
affect each individual component. Once threats are selected, they are automatically
classified in the STRIDE security-oriented framework (Spoofing identity, Tamper-
ing, Repudiation, Information disclosure, Denial of service, and Elevation of privilege).
We take this tool as the baseline for the proposal presented in this chapter. Finally,
Khan et al. (2017) perform a DFD-based analysis of vulnerabilities and threads,
using STRIDE as a framework, in Cyber-physical Systems. Casola et al. (2019)
propose an almost completely automated process for threat modeling and risk
assessment.
None of these contributions tackle the automatic detection of privacy-related
vulnerabilities or monitor privacy-related risks for continuous risk manage-
ment. In general, even for security, risks are not analyzed and monitored con-
tinuously, and historical data are not taken into account (Ahmad et al., 2012,
Baskerville et al., 2014). Besides, security risks are assessed based on speculation
rather than evidence (Shameli-Sendi et al., 2016, Webb et al., 2014). The inter-
action between analytics capabilities and Information Security Risk Management
(ISRM) capabilities can help organizations to perform continuous risk assessments
and enable evidence-based decision-making (Naseer et al., 2017).
In parallel, Article 25 in GDPR4 discusses data protection by design and by
default, underlining that considering privacy from the beginning is essential to
address privacy successfully. GDPR establishes binding data protection principles,
individuals’ rights, and legal obligations to ensure the protection of personal data
of EU citizens. However, legal measures need to come along with technical mea-
sures to protect privacy and personal data in practice. PDP4E H2020 project
(Martin and Kung, 2018) focuses on the importance to involve engineers in the
loop, for Privacy-by-design (PbD) to be viable.
LINDDUN
LINDDUN (Wuyts, 2015) is a threat modeling methodology to support risk ana-
lysts when addressing privacy risks affecting end users in an application. This
methodology provides some guidance to identify and categorize threats under a set
of general risks (Linkability, Identifiability, Non-repudiation, Detectability, Disclosure
4. Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protec-
tion of natural persons with regard to the processing of personal data and on the free movement of such data,
and repealing Directive 95/46/EC (General Data Protection Regulation). Official Journal of the European
Union, L119:1–88, May 2016.
Model-based Continuous Risk Management Methodology 147
In this section, we propose a methodology for risk management for the creation of
TSIS, and we indicate the actors involved in each of those steps. Our methodology
(see Figure 8.1) can be summarized in 6 steps:
• S1: TSIS Assets Definition: First, we edit or load DFDs representing the
functional description of the system. These DFDs are useful to manage risks
Figure 8.1. Proposed model-based risk management methodology. Roles: (O) Risk Man-
agement Owner; (P) Product Owner; (A) Architect; (V) Developer and (R) Risk Analyst.
148 Enabling Continuous Privacy Risk Management in IoT Systems
Note that this methodology is continuous in the sense that new evidence is con-
tinuously collected to challenge previous knowledge upon which risk-related deci-
sions were made in the past. While immediate triggering of mitigation actions is
not our main focus, conditional options based on the status of the system could
be included in S4 to strengthen the continuous approach. Also, the methodology
is particularly customized for TSIS, specifically in S1, where GeneSIS models are
used as the baseline, and S6, where evidences are collected to consider privacy-
related issues. However, the methodology can be easily adapted to other types of
systems and used generically.
Automatic Vulnerability Detection 149
Figure 8.2. Analysis of vulnerability scope in the LINDDUN threat tree for detectability
in a data flow.
5. LINDDUN Detectability threat tree catalog: www.linddun.org/detectability. Last accessed in Jan 18th 2020.
150 Enabling Continuous Privacy Risk Management in IoT Systems
conditions that must hold for a subset of vulnerabilities extracted from LIND-
DUN threat trees, related to those areas. For instance, in area C, vulnerabilities
related to steganalysis become relevant if the channel is not encrypted. As a second
example, in area B, depending on the volume of traffic in the channel, “insufficient
dummy traffic” may or may not actually be a vulnerability. In the latter, note also
that this condition may change along time, increasing or decreasing the relevance
Table 8.1. Example of conditions for vulnerabilities related to detectability in a data flow
(as defined in LINDDUN) to be relevant.
Vulnerability/ies
(Wuyts, 2015) Conditions for relevance Rationale
Table 8.1. Example of conditions for vulnerabilities related to detectability in a data flow
(as defined in LINDDUN) to be relevant (continued).
Vulnerability/ies
(Wuyts, 2015) Conditions for relevance Rationale
of this vulnerability. Therefore, continuous risk management may also include the
continuous monitoring of metrics that allow measuring the level of relevance of
vulnerabilities or the likelihood of threats to occur.
Note though that Figure 8.2 shows an example for a simple case where the
LINDDUN threat tree is not related to any other threat tree. However, in most
cases vulnerabilities related to a particular LINDDUN threat trees are related to
vulnerabilities detected in other LINDDUN or STRIDE threat trees. Figure 8.3
describes the detail of this connection in the form of a graph, where every node is
one of the LINDDUN threat trees (blue nodes) or one of the STRIDE threat trees
related to LINDDUN trees (red nodes). As it can be observed, Detectability of a
data flow (D_d f ) is one of the rare cases where the vulnerabilities in the threat tree
are not related to those in other trees. In general, most of the trees are related to oth-
ers. In order to understand the vulnerabilities of a particular component in a DFD,
it is important to navigate through these connections. For instance, the analysis
of vulnerabilities related to the Identifiability of an entity (I _e) generates a cascade
analysis of vulnerabilities that may include I _ds, I _d f , I D_d f , I D_ds, S_e, and
T _ p, following directed edges in the graph. Note that edges are colored in gray if
threat trees refer to the same DFD element (e.g., I _ds → I D_ds), and they are
colored in red if they refer to different types of DFD elements (e.g., I _e → I _ds).
For those relationships represented in gray, we assume that we refer to the same
element in the same DFDs. For those relationships represented in red, we have
explored them one by one and established a rule to propagate the analysis from one
152 Enabling Continuous Privacy Risk Management in IoT Systems
Figure 8.3. Graph describing the relationship between LINDDUN (blue) and STRIDE (red)
threat trees. The first part of the ids stand for each of the categories of LINDDUN (Linka-
bility, Identifiability, Non-repudiation, Detectability, Unawareness, and Non-compliance –
Disclosure of information is not represented as it is understood as the Information dis-
closure already captured by STRIDE) and STRIDE (only for those related to LINDDUN:
Spoofing identity, Tampering, Information disclosure and Elevation of privilege). The sec-
ond part after the “_” represents the type of DFD component (entity, data flow, data store,
process). Labels in the edges represent indications extracted from LINDDUN descriptions
in the corresponding threat trees.
GDPR establishes a set of duties imposed on the data processors, controllers, and
third parties which are aimed at honoring the corresponding data subject’s rights.
In GDPR, risk is explicitly scoped (Rec. 76) with regard to the rights and free-
doms of the data subject. In order to understand whether these data subject rights
and principles (described in GDPR) are being effectively protected by mitigating
existing risks, current approaches tend to analyze the data flow in the system (e.g.,
LINDDUN and DFDs). However, elements in a DFD tend to be defined at a
functional level (e.g., process to collect profile data from a user or process to merge data
from two existing data sources). In this situation, collecting system data to monitor
a threat related to the identifiability of an entity as defined by LINDDUN, just to
take an example, is not obvious, if we do not understand the relationship of a data
flow in the application with the architecture of the infrastructure that we monitor.
Without a proper connection of these elements to architectural levels that can be
monitored, continuous risk management for privacy becomes much more difficult.
Most monitoring tools monitor the infrastructure level. These tools may monitor
system aspects in a data center or a particular server, tools for network monitoring,
etc, and collect metrics.
The main contribution of the approach proposed in this section is the map-
ping of TSIS architecture models, such as the one proposed by GeneSIS, with data
flows in the application under analysis (e.g., DFDs). In S6, we may assume that
the elements from both models have been mapped establishing a link between the
two types of models. Details on this mapping are provided below. For instance, a
NoSQL data store in our system architecture may be mapped to a data store com-
ponent in a DFD. As another example, a particular process element in a DFD may
be mapped to the server the process is running on. As an alternative, it is possi-
ble that the mapping is not pre-established before starting the risk analysis process.
In this situation, we consider that phase S1 can be extended to allow the user to
establish this mapping manually.
Model Mappings
In order to drive the mapping between the architecture model and DFDs, we take as
the baseline the connection between privacy-related threats studied in LINDDUN
and the vulnerabilities that are related to security aspects of the architecture, as
captured by STRIDE. In Table 8.2, we show a sample of vulnerabilities extracted
from (Shostack, 2014) related to the STRIDE threats in Figure 8.3. The table shows
a brief description and some usual mitigation actions. In the last two columns, we
show examples of metrics that could be used for continuous security monitoring
154
Table 8.2. Sample of the mapping between STRIDE threats (related to LINDDUN threats in Figure 8.2) and related metrics that can be detected
from monitoring architectural components. STRIDE tree node extracted from Shostack (2014).
Tampering at Bypassing ACLs, permissions, Ensure data is created Random data editor Database/Data
Data Store protection rules policies, etc allow people with appropriate reporting the number of store
(T _ds) because of no or to alter data without a permissions. Change successful attempts to
weak protection clear justification. permissions. change data with no or low
privileges.
Information Timing Code time execution can Design for Monitoring execution time Software
Disclosure of a reveal confidential cryptography to take and control variability. components to
Process (I D_ p) information. constant time. solve tasks requiring
secrecy.
155
156 Enabling Continuous Privacy Risk Management in IoT Systems
and what type of architectural components are being monitored. We would like to
remark that this table is not meant to be an exhaustive list of all the potential threats,
but some examples for illustrative purposes that serve the purpose of analyzing
which DFD components are to be mapped to which architectural components.
Note that these vulnerabilities are specially relevant in an IoT context. For
instance, let us take the example in Table 8.2 related to Information Disclosure of
Data Flow. There may be different vulnerabilities associated in particular to TSIS.
For instance, if we have a device IoT hub, eavesdropping or interfering the commu-
nication between the device and the gateway would be a potential threat. You may
also find similar threats in the communication between devices, where data may be
read in transit, tampering with the data or overloading the device with new con-
nections or even with the cloud gateways, where eavesdropping or communication
interference between devices and gateways may occur.
In general, mappings between architectural models and DFD link:
• Entities in DFDs with the associated network channels in which entity infor-
mation is transmitted or data stores in which the information is stored;
• Data flows in DFDs with the corresponding network channels;
• Data stores in DFDs with the corresponding databases or data stores used in
the system architecture; and
• Processes in DFDs with the corresponding software components executing
tasks related to those processes, especially when these tasks entail some level
of secrecy.
actions defined in S3 and S4. Once the risk model is developed for each DFD, we
propose two ways to exploit the established mappings to enable continuous risk
management:
8.6 Conclusions
Handling privacy-related risks is not well understood in the IoT context. There
is a lack of mechanism to effectively detect vulnerabilities related to privacy and
to collect meaningful evidences and continuously monitor risks. We need to find
the intersection between data protection challenges and the enabling of TSIS. We
have presented a first step towards the enactment of continuous privacy-related
risk control for TSIS, leveraging the link offered by LINDDUN between privacy
and security threat categories and combining functional and architectural models.
As future work, we need to establish a stronger connection between privacy threat
models like LINDDUN and the actual requirements imposed by GDPR, as current
threat models were devised before GDPR and they may not fully cover all derived
requirements. In particular, the relationship between data subject rights and GDPR
158 Enabling Continuous Privacy Risk Management in IoT Systems
Acknowledgments
This work is supported by the European Commission through the ENACT project
under Project ID: 780351 and the PDP4E project under Project ID: 787034.
David Sanchez-Charles was part of the staff of Trialog SA by the time this arti-
cle was written. We would like to thank A. Kung and the rest of Trialog’s team for
their continuous support in this research.
References
Ferry, N., P. H. Nguyen, H. Song, P.-E. Novac, S. Lavirotte, J.-Y. Tigli, and A. Sol-
berg. 2019. “GeneSIS: Continuous Orchestration and Deployment of Smart
IoT Systems.” In: the proceedings of the IEEE COMPSAC conference, Mil-
waukee, USA, July 15–19. Springer.
Ferry, N., A. Solberg, H. Song, S. Lavirotte, J.-Y. Tigli, T. Winter, V. Muntés-
Mulero, A. Metzger, E. R. Velasco, and A. C. Aguirre. 2018), ‘ENACT:
Development, Operation, and Quality Assurance of Trustworthy Smart IoT
Systems.” In: International Workshop on Software Engineering Aspects of Contin-
uous Development and New Paradigms of Software Production and Deployment.
pp. 112–127.
Fitzgerald, B. and K. Stol. 2017. “Continuous software engineering: A
roadmap and agenda.” Journal of Systems and Software. 123: 176–189.
DOI: 10.1016/j.jss.2015.06.063. URL: http://dx.doi.org/10.1016/j.jss.2015.
06.063.
Fleurey, F. and B. Morin. 2017. “ThingML: A generative approach to engineer het-
erogeneous and distributed systems.” In: 2017 IEEE International Conference
on Software Architecture Workshops (ICSAW). pp. 185–188.
Gupta, S., V. Muntés-Mulero, P. Matthews, J. Dominiak, A. Omerovic, J. Aranda,
and S. Seycek. 2015. “Risk-driven framework for decision support in cloud ser-
vice selection.” In: 2015 15th IEEE/ACM International Symposium on Cluster,
Cloud and Grid Computing. pp. 545–554.
Karabacak, B. and I. Sogukpinar. 2005. “ISRAM: Information Security Risk
Analysis Method.” Comput. Secur. 24(2), 147–159. issn: 0167-4048.
DOI: 10.1016/j.cose.2004.07.004. URL: http://dx.doi.org/10.1016/j.cose.
2004.07.004.
Khan, R., K. McLaughlin, D. Laverty, and S. Sezer. 2017. “STRIDE-based threat
modeling for cyber-physical systems.” In: 2017 IEEE PES Innovative Smart
Grid Technologies Conference Europe (ISGT-Europe). 1–6. DOI: 10.1109/ISG-
TEurope.2017.8260283.
Lund, M. S., B. Solhaug, and K. Stølen. 2010. Model-driven risk analysis: the
CORAS approach. Springer Science & Business Media.
Martin, Y.-S. and A. Kung. 2018. “Methods and tools for GDPR compliance
through privacy and data protection engineering.” In: 2018 IEEE European
Symposium on Security and Privacy Workshops (EuroS&PW). 108–111.
Merkow, M. and L. Raghavan. 2011. “An Ecosystem for Continuously Secure
Application Software.” RUGGED Software, CrossTalk March/April.
Moran, A.. 2014. Agile Risk Management. Springer International Publishing. isbn:
978-3-319-05007-2.
160 Enabling Continuous Privacy Risk Management in IoT Systems
Chapter 9
9.1 Introduction
With the progressive implementation of new technologies, more and more inter-
connected, we are surrounded by an increasingly synchronized, delocalized, and
correlated environment, animated by a continuous flow of data generated by our
choices and behaviors.
The disruptive advent of IoT technologies has profoundly changed the rela-
tionship between human beings and “things.” It is difficult to make a distinction
between online and offline dimensions, because most of the devices used in every-
day life are connected and interact with the surrounding reality in a continuous
exchange of information between material and virtual environments.1
It can be observed, therefore, that a new challenge has opened up for the pro-
tection of the rights and freedoms of Data Subjects and stakeholders in the new
technological environment that has changed since the broader adoption of IoT.2
161
162 Data Protection Compliance Requirements
Existing digital and electronic devices have implied the need to strongly focus
on personal data protection safeguards. The more the technologies have evolved,
the more it has been necessary to establish a legislative framework to protect a “per-
sonal” and “private” space. These needs have been at the heart of debates in the
Parliament and in the Commission of the European Union, bringing to the adop-
tion of a legal instrument to be valid in all the European Union Member States, the
Regulation n. 679/2016, the General Data Protection Regulation (GDPR). This
Regulation was introduced into an existing International and European normative
framework protecting fundamental rights and freedoms, which includes, among
others, the Universal Declaration of Human Rights and the Charter of Fundamen-
tal Rights of the European Union.
Even if the Regulation does not explicitly refer to IoT Devices or the IoT Sys-
tem, the principles emerging from the GDPR—first of all, the ones of purpose
limitation, storage limitation, lawfulness, transparency, fairness, integrity and con-
fidentiality, and data protection-by design/by default—are adaptable to any tech-
nological development and deployment, IoT technologies included.
The present document proposes some brief considerations, without claiming to
be exhaustive, on the issues of Data Subject awareness and accountability in the
IoT field, with a focus on technology developers in the light of the GDPR and
cybersecurity international best practices.
(or networks) of physical objects that are capable of sensing or acting on their environment, and able to
communicate with each other, other machines or computers”. See also ENISA, Baseline Security Recom-
mendations for IoT in the context of Critical Information Infrastructures, 2017. The document refers to the
concept introduced by Kevin Ashton: “a wide ecosystem where interconnected devices and services collect,
exchange and process data in order to adapt dynamically to a context.”
IoT and General Data Protection Regulation 163
between individuals and IoT Devices or the IoT System has been studied in order to
better understand practical and actual dynamics of relationship between individuals
and objects; as a result, it emerged the need to ensure a human-centric IoT envi-
ronment, as the EU project GHOST3 demonstrates.
A new IoT project may involve users and Data Subjects as actors or as factors.
Most of the IoT solutions are considering the possible factorization of people: people
as an entity which interacts with other sensors. Factorization could bring to limit
Data Subjects’ rights and freedom, with the risk of de facto downgrading their role to
the same level of “objects.” A way to empower individuals, without factorizing them,
is to educate and train them on the IoT ecosystem features, risks, and safeguards, so
as to promote the creation and perception of an environment that they can consider
“safe” and “manageable.”
A better education would also imply an increasing awareness of Data Subjects’
skills and rights, related to the adoption of IoT tools, as well as a major attention by
developers and producers to IT security and other technological features, in order
to protect, by default, individuals from possible violations and to enable them to
receive adequate information and to exercise their rights.
Moreover, the way in which personal information is collected and managed, in
an IoT scenario, may affect not only end users’ rights but also the rights of third
subjects which are observed and recorded by the devices.
The processing of personal data carried out through IoT systems is certainly
included in the material scope of application of the GDPR, which sets forth some
conditions for the lawfulness of the processing as listed in Article 6:
“[…] processing shall be lawful only if and to the extent that at least one of the following
applies:
(a) The Data Subject has given consent to the processing of his or her personal data
for one or more specific purposes;
(b) Processing is necessary for the performance of a contract to which the Data
Subject is party or in order to take steps at the request of the Data Subject prior
to entering into a contract;
(c) Processing is necessary for compliance with a legal obligation to which the Con-
troller is subject;
(d ) Processing is necessary in order to protect the vital interests of the Data Subject
or of another natural person;
(e) Processing is necessary for the performance of a task carried out in the public
interest or in the exercise of official authority vested in the Controller;
3. GHOST Safe-Guarding Home IoT Environments with Personalised Real-time Risk Control: “D3.9: Trials
use case specification and report” (1st release).
164 Data Protection Compliance Requirements
(f ) Processing is necessary for the purposes of the legitimate interests pursued by the
Controller or by a third party, except where such interests are overridden by the
interests or fundamental rights and freedoms of the Data Subject which require
protection of personal data, in particular where the Data Subject is a child.”
There are six available legal bases for the processing. No single basis is necessar-
ily more suitable than the others, due to the fact that the most appropriate basis
will depend on the purpose of the data processing and on the possible relationship
between Controllers and Data Subjects.
Among the various bases stated in the aforementioned Art. 6 of the GDPR, the
one that necessarily implies a good degree of awareness and information of the Data
Subjects is described in letter (a): consent.
In particular, according to Article 7 of the GDPR:
Article 4 paragraph 1 n. 11 GDPR rules that “consent of the Data Subject means
any freely given, specific, informed and unambiguous indication of the Data Subject.”
However, as provided by the WP29 in the opinion adopted in 2014 on the “Recent
Developments on the Internet of Things”:4 “In many cases, the user may not be aware
4. ARTICLE 29 DATA PROTECTION WORKING PARTY This Working Party was set up under Arti-
cle 29 of Directive 95/46/EC. It is an independent European advisory body on data protection and pri-
vacy. Its tasks are described in Article 30 of Directive 95/46/EC and Article 15 of Directive 2002/58/EC.
The secretariat is provided by Directorate C (Fundamental Rights and Union Citizenship) of the Euro-
pean Commission, Directorate General Justice, B-1049 Brussels, Belgium, Office No MO-59 02/013.
Website: http://ec.europa.eu/justice/data-protection/index_en.htm14/ENWP223 Opinion 8/2014 on the
Recent Developments on the Internet of Things Adopted on 16 September 2014.
IoT and General Data Protection Regulation 165
of the data processing carried out by specific objects. Such lack of information constitutes
a significant barrier to demonstrating valid consent under EU law, as the Data Subject
must be informed ” pursuant to Article 13 and Article 14 of the GDPR.
In all cases of different legal bases of personal data processing, other than Data
Subjects’ consent (think about the data processing in the legitimate interest of the
Controller or in the public interest), the challenge would be that of adequately
informing Data Subjects and to make them as aware as possible of the activities
carried out and of the related risks to their rights and freedoms.
Therefore, the ratio legis of this principle is to allow Data Subjects to have control
over their data and thus to make a choice about it.
In particular, paragraph 10 of the document states that: “the Data Subject should
be able to determine in advance what the scope and consequences of the processing entails
5. ARTICLE 29 DATA PROTECTION WORKING PARTY This Working Party was set up under Arti-
cle 29 of Directive 95/46/EC. It is an independent European advisory body on data protection and pri-
vacy. Its tasks are described in Article 30 of Directive 95/46/EC and Article 15 of Directive 2002/58/EC.
The secretariat is provided by Directorate C (Fundamental Rights and Union Citizenship) of the Euro-
pean Commission, Directorate General Justice, B-1049 Brussels, Belgium, Office No MO-59 02/013.
Website: http://ec.europa.eu/newsroom/Article29/news.cfm?itemtype=1358&tpaid=6936 17/EN WP260
rev.01 Article 29 Working Party Guidelines on transparency under Regulation 2016/679 Adopted on 29
November 2017. As last Revised and Adopted on 11 April 2018.
166 Data Protection Compliance Requirements
and that they should not be taken by surprise at a later point about the ways in which
their personal data has been used.”
The Data Controller would therefore be obliged to:
• Provide individuals with information including the purposes for the process-
ing of their personal data, the retention periods for personal data processed,
and with whom personal data will be shared;
• Provide privacy information to individuals at the time the IoT Data Con-
troller collects their personal data;
• Obtain personal data from other sources; the IoT Data Controller shall pro-
vide individuals with privacy information within a reasonable period and no
later than one month; there are few circumstances when the IoT Data Con-
troller does not need to provide people with privacy information, such as, if
an individual already has the information or if it would involve a dispropor-
tionate effort to provide information to them;
• Provide information to people in a concise, transparent, intelligible, easily
accessible manner, and it shall use clear and plain language;
• Regularly review, and where necessary, update its privacy information. The
IoT Data Controller shall provide the individuals with any information
related to the new uses of their personal data, before the IoT Data Controller
starts the processing.
could be equipped with well-visible interfaces, displaying such labels. For example,
an image sensor and an IoT object could be equipped with a screen, to increase
awareness on the presence of the sensor and increase user confidence in using
objects that gather data.
Another option, connected with the information just indicated above, could
result in the obligation to alert individuals about the presence of IoT sensors in a
given environment. Specifically, following the provision set forth by Article 12 para-
graph 8 of the GPDR, where it is expected that “The Commission shall be empowered
to adopt delegated acts in accordance with Article 92 for the purpose of determining the
information to be presented by the icons and the procedures for providing standardized
icons,” Solutions which provide for the use of icons or figurative representations of
other kind, easily understandable by the interested parties, may be considered.
This “alert signals” solution could be even more helpful – in all cases in which
Data Subjects are not “end users,” but simply “non-users,” since there are no inter-
faces and no features for interactivity between things and persons – in order to
prevent negative effects caused by IoT deployments.
Finally, a third option consists in using a dedicated smart phone application to
inform the data subjects about the presence of IoT in their vicinity. Such approach
has been researched and implemented in the European Large Scale Pilot on IoT
for smart cities, the H2020 Synchronicity European research project.6 A dedicated
application, titled Privacy App (www.privacyapp.info), has been developed.
The Privacy Application enables users to discover nearby IoT Devices or the
IoT System on an interactive map. For each IoT device, the app provides the avail-
able information related to the purpose of the data processing, the controllers and
processors involved, as well as information on retention period and cross-border
transfer. The user can also add new IoT Devices or the IoT System on the map and
request for more information. More importantly, the user can directly contact the
Data Protection Officer of the data controller of the IoT device.
6. The Privacy App was developed by Mandat International in the context of Synchronicity, the H2020 Euro-
pean Large Scale Pilot on the Internet of Things for Smart Cities: https://synchronicity-iot.eu/
168 Data Protection Compliance Requirements
7. The approach of the International Standard Organization already in 2013, with the ISO/IEC 27001 Stan-
dard, focuses on risk assessment as a source from which to derive the implementation of technical and
organizational security measures.
The Principles of Accountability, Data Protection by Design 171
The “data protection by default principle” finds substance in all those measures,
technical and organizational, functional to ensure that only personal data necessary
for the purposes pursued are processed. That obligation applies to the amount of
personal data collected, the extent of their processing, the period of their storage and
their accessibility. In particular, such measures shall ensure that by default personal
data are not made accessible without the individual’s intervention to an indefinite
number of natural persons.
In the light of the above considerations, in conclusion, it should be noted that
the GDPR, in its Articles 5-24-25 as analyzed above, does not provide for a merely
formal fulfillment of data protection obligations; quite the opposite, the Regulation
calls for a real assessment, concrete and substantial, by those who are responsible
for the processing of personal data. In this regard, the recent EDPB guidelines on
the topic have specified that (par. 8): “The term measures can be understood in
a broad sense as any method or means that a Controller may employ in the pro-
cessing. These measures must be appropriate, meaning that they must be suited to
achieve the intended purpose, i.e. they must be fit to implement the data protec-
tion principles effectively by reducing the risks of infringing the rights and freedoms
of Data Subjects. The requirement to appropriateness is thus closely related to the
requirement of effectiveness”.9
The GDPR introduces, of course, a cultural change that will inevitably impact
also IoT developers and manufacturers, even if it has to be taken into account that
the “accountability principle,” the “data protection by default principle” and the
“data protection by design principle” refer to actual Data Controllers, while, accord-
ing to Recital 78 of the GDPR, “producers of the products, services and applications
should be encouraged [but not obliged, authors’ note] to take into account the right
to data protection when developing and designing such products, services and applica-
tions and, with due regard to the state of the art, to make sure that Controllers and
processors are able to fulfil their data protection obligations.”
Notwithstanding this, the GDPR combines the mild “encouragement” of pro-
ducers with the principle of accountability to be respected by Data Controllers, thus
elevating data protection by design from an option to a strict parameter of com-
pliance to be assessed by Controllers in the selection of products and/or services
providers. And the mentioned Recital 78 of the GDPR provides that “the principles
of data protection by design and by default should also be taken into consideration in
the context of public tenders.”
9. EDPB, Guidelines 4/2019 on Article 25 Data Protection by Design and by Default. In particular, the concept
of effectiveness is extremely connected with the accountability principle of Article 5 paragraph 2 of the
GDPR. Indeed, the Data Controllers must be able to demonstrate that they have implemented dedicated
measures to protect these principles, and that they have integrated specific safeguards that are necessary to
secure the rights and freedoms of Data Subjects.
The Principles of Accountability, Data Protection by Design 173
Table 9.1. Cybersecurity measures for an IoT device and an IoT system (Source: ETSI).
The IoT Devices or the IoT System requires at least one administrative user, that is, a user
having the ability to operate with elevated privileges inside the IoT Devices or the IoT
System (e.g., definition of other users, reset of their passwords).
The IoT Devices or the IoT System requires the passing of an authentication procedure
(e.g., login) before being able to allow the processing of any personal data. This authenti-
cation procedure verifies the username and a password of at least of 8 characters in length
and containing alphanumeric, special, and uppercase characters.
(Continued)
10. ETSI, Cyber Security for Consumer Internet of Things, 2019. Document available at the following link:
https://www.etsi.org/deliver/etsi_ts/103600_103699/103645/01.01.01_60/ts_103645v010101p.pdf
174 Data Protection Compliance Requirements
Table 9.1. Cybersecurity measures for an IoT device and an IoT system (Source: ETSI)
(continued).
The IoT Devices or the IoT System requires strong authentication e.g., (multi-factor
authentication, possession or biometrics). For IoT Devices or the IoT System that have
stateless systems in general, the IoT Devices or the IoT System generates a token to asso-
ciate to the session. The token associated with the session of the web IoT Devices or the
IoT System or stateless systems is sufficiently long (64 or more alphanumeric characters)
and impossible to guess. The token associated with the session of the IoT Devices or the
IoT System or stateless systems has an expiration time.
The IoT Devices or the IoT System stores the password within its database in encrypted
form.
The IoT Devices or the IoT System uses a hashing algorithm suitable for password encryp-
tion.
The IoT Devices or the IoT System implements automated password selection restrictions
(e.g., a minimum number of characters is set, and it ignores common or user-referenced
passwords). When the user ID is associated to an email address, the IoT Devices or the IoT
System requires such email address to be verified. Email addresses associated with a user ID
are periodically verified to ensure that the email is still valid and in use.
The IoT Devices or the IoT System limits or throttles the availability of logins in the event
of an abnormal number of unsuccessful access attempts occurring within a short time frame.
The IoT Devices or the IoT System allows each of its administrative users to assign different
permission levels to different users.
The IoT Devices or the IoT System prevents any non-administrative user from changing
the permission levels assigned to other users.
The IoT Devices or the IoT System protects the data it allows to be processed through
pseudonymization techniques.
The IoT Devices or the IoT System protects the data that it allows to be processed through
transparent encryption techniques. Data processed through the IoT Devices or the IoT
System are appropriately classified (e.g., common, particular, judicial, subdivisions in per-
sonalized under systems).
The IoT Devices or the IoT System transmits network traffic in a protected from via state-
of-the-art security protocols (e.g., TLS1.2, valid certificates, HSTS). Data processed with
the help of the IoT Devices or the IoT System are backed up at least daily. Data processed
with the help of the IoT Devices or the IoT System can be restored quickly.
The IoT Devices or the IoT System is currently supported (e.g., through the release of
security updates and patches).
(Continued)
The Principles of Accountability, Data Protection by Design 175
Table 9.1. Cybersecurity measures for an IoT device and an IoT system (Source: ETSI)
(continued).
The IoT Devices or the IoT System is constantly kept up to date. The IoT Devices or the
IoT System is periodically subjected to sessions of vulnerability assessment and penetration
testing to assert its robustness to cyberattacks.
The IoT Devices or the IoT System generates access logs.
The IoT Devices or the IoT System generates logs of critical actions (e.g., creation or
removal of content or users).
The IoT Devices or the IoT System generates logs of the performed processes.
The logs are complete, unalterable, and stored for at least six months; the integrity of the
logs can be verified. If the IoT Devices or the IoT System is connected with smartphones
and requires permissions on the device, it provides policies that describe the purposes of the
processing enabled by each permission. If the IoT Devices or the IoT System is connected
with smartphones, it never uses the Device ID as a key to identify a record. If the IoT
Devices or the IoT System is for smartphones, it uses certified pinning techniques to avoid
MITM attacks.
The IoT Devices or the IoT System code does not contain confidential credential compo-
nents (e.g., passwords, tokens, keys …). The IoT code is developed in accordance with the
guidelines for secure code (e.g., CERT, OWASP …).
Table 9.2. Privacy measures for an IoT device and an IoT system.
The IoT Devices or the IoT System is accompanied by a specification of the type of data
of which it allows the processing.
The IoT Devices or the IoT System is accompanied by a specification of the data flows
from/to the outside.
The IoT Devices or the IoT System allows to define and modify the retention times for the
various types of data that it stores.
The IoT Devices or the IoT System makes it possible to record the source of the data it
stores (e.g., data supplied directly by the person concerned, data extracted from databases
…)
(Continued)
176 Data Protection Compliance Requirements
Table 9.2. Privacy measures for an IoT device and an IoT system (continued).
The IoT Devices or the IoT System stores only the data necessary for its operation (e.g., it
does not store unnecessary data).
If the process of verifying the accuracy of the data entered by the user identifies incorrect or
suspicious data, the IoT Devices or the IoT System sends an alert to the competent function
or reports it to an administrator (to allow the competent function to be informed).
The IoT Devices or the IoT System retains the date of the last update of each record.
The IoT Devices or the IoT System allows an administrator to “mark” data as restricted
(e.g., providing flags in the database that identify the associated field as restricted).
Where applicable, the IoT Devices or the IoT System prevents the processing of restricted
data fields (the restricted data field must not be read, modified, deleted, transmitted, dis-
played, etc. until it is unlocked by the platform administrator. Neither another user nor
IoT Devices or the IoT System should be able to do this).
The IoT Devices or the IoT System shall enable the Data Controllers to aggregate in a
comprehensible way all the data that it retains in relation to an interested party, allowing
the party the ability to modify and visualize it.
The IoT Devices or the IoT System shall enable the Data Controller to record aggregated
data in one or more common format files (.csv, .xlsx, .xls, .txt, etc.).
The IoT Devices or the IoT System shall enable the Data Controller to transfer aggregated
data relating to a Data Subject in an interoperable format (e.g., XML, CSV, JSON).
The IoT Devices or the IoT System allows to export the aggregated data related to a Data
Subject in an interoperable format (e.g., XML, CSV, JSON), flanking them with useful
metadata in order to identify them correctly.
If the IoT Devices or the IoT System collects data on minors, it requires the consent of the
parental guardians to be entered and given to the Data Controller.
Where applicable, if the IoT Devices or the IoT System collects data on minors, the consent
of the parental guardians shall require an express opt in consent.
If the IoT Devices or the IoT System generates scores relating to a Data Subject (e.g., third-
party data resulting from automatic processing) and the Data Subject did not give his or
her consent to automated processing, the IoT Devices or the IoT System makes it
possible that the decision having legal effects on the Data Subject comes from an operator
and not from an automated process.
(Continued)
The Principles of Accountability, Data Protection by Design 177
Table 9.2. Privacy measures for an IoT device and an IoT system (continued).
Where applicable, the IoT Devices or the IoT System works in accordance with the consent
given by the Data Subjects (e.g., it informs the operators about the consent given and does
not make certain types of data available for certain processing operations if their consent
was not been given).
Where applicable, the IoT Devices or the IoT System keeps a record of the consent lent or
denied, each with its own timestamp.
Where applicable, the IoT Devices or the IoT System shall keep a record of the requests by
the Data Subjects to exercise their rights.
The IoT Devices or the IoT System does not feed databases and/or does not transfer per-
sonal data to servers not allocated within the European Union in the absence of assessment
of adequacy of the country in which the data are transferred and explicit consent requested
and provided by the Data Subject/IoT Devices or the IoT System user.
If the IoT Devices or the IoT System is public, the Data Controller shall communicate and
ease access to privacy policies in order to allow the Data Subject to review them at any time.
Considering the pervasive nature of the Internet of Things, the risk for the rights
and freedoms of data subjects will be higher and higher. Adopting data protection
by design approach will be key for the adoption and compliance of IoT-related
services and applications.
The adoption of the GDPR has raised the level of awareness on data protection
regulations. However, ensuring compliance will require to bring together experts
in technology and in law in order to build a bridge between these two worlds.
The present chapter has been redacted in the context of the European Research
Project NGIoT in the context of the Horizon 2020 European Research Program.
DOI: 10.1561/9781680836837.ch10
Chapter 10
Cybersecurity Certification
in IoT Environments
This chapter will present and analyze current standardized IoT security certification
approaches and their lacks in terms of security certification challenges, such as the
security dynamicity or the life cycle management. Furthermore, the chapter will
present a methodology combining security risk assessment and testing to address
some of the challenges and to serve as a basis for future security certification schemes.
10.1 Introduction
The technology of the Internet of Things (IoT) is increasingly invading our daily
lives. The high scope involves devices of all kinds, such as the washing machine, traf-
fic lights, the car, the fridge, the surveillance camera, the smart band, the heart mon-
itor, or even the clothes. These smart devices connect to the internet and exchange
data in order to make our lives easier. This great interaction and exchange of infor-
mation is plagued in the definition of IoT given in [1], where it is defined as “an
open and comprehensive network of intelligent objects that have the capacity to
auto-organize, share information, data and resources, reacting and acting in face of
situations and changes in the environment.” However, the advantages of the inter-
connectivity of these devices pose a big challenge associated in terms of security
178
Introduction 179
and privacy. On the one hand, IoT devices gather a high amount of personal and
sensitive data that could be exposed to attackers. On the other hand, the attack
surface has been increased in such a way that a network of these devices can be
used to attack more juicy targets. This is clearly seen in attacks such as the Mirai
IoT Botnet [2], which causes several money loses to big companies such as Spo-
tify or Microsoft. This network was composed by different kinds of typical devices
(e.g., IP cameras, video cameras, routers, thermostats, and even computers), and
although the creators were arrested, new variants of the Botnet have been devel-
oped, leveraging to a worse menace, capable of performing Distributed Denial of
Service (DDoS) attacks at 1 terabyte per second. This gets still worse in situations in
which human life is involved, for example, in ehealth devices or automated vehicles,
as demonstrated in [3].
Being aware of this situation, the scientific community, companies, standardiza-
tion bodies, and, especially, the governments try to respond to this challenge, work-
ing together to build a next generation of more secure and standardized devices.
In this sense, it is needed a suitable security certification scheme to assess and com-
pare different security technologies, in order to provide a more harmonized IoT
security view to be leveraged by end consumers. The term “certification” is defined
by the National Institute of Standards and Technology (NIST) as “a comprehen-
sive assessment of the management, operational, and technical security controls in
an information system, made in support of security accreditation, to determine
the extent to which the controls are implemented correctly, operating as intended,
and producing the desired outcome with respect to meeting the security require-
ments for the system.” This way, the main aim of the security certification is to
validate and verify the security of the product. In Europe, this initiative is led by the
European Union Agency for Cybersecurity (ENISA), as established in the Proposal
for a REGULATION OF THE EUROPEAN PARLIAMENT AND OF THE
COUNCIL on ENISA, the “EU Cybersecurity Agency,” and repealing Regula-
tion (EU) 526/2013 and on Information and Communication Technology cyber-
security certification (“Cybersecurity Act”). As described in the Cybersecurity Act,
the main objective is to create a common cybersecurity certification framework
for digital products, including IoT, in a way the certificates could be valid inside
the whole European Union. However, while the current heterogeneity existent in
terms of security standards and certification schemes makes difficult the compari-
son between products certified by different standards and schemes, the dynamism
associated to security leverages on an out-of-date certificate that does not show the
security level in a realistic way. In this sense, agile and automated certification mech-
anisms are needed, implying a revolution of the current certification schemes that
are usually static, time costly, and expensive.
This chapter analyses in Section 10.2 the main challenges associated to cyber-
security certification and how current certification standards deal with them. As a
180 Cybersecurity Certification in IoT Environments
One of the main challenges associated with security certification is the harmoniza-
tion of the wide variety of security certification schemes that coexist together [5].
The current heterogeneity makes difficult the comparison of different solutions
and processes, especially when a product is evaluated under different certification
schemes at national levels. Currently, there is no unified solution that copes with
these issues; therefore, the process of comparing and assessing the cybersecurity level
of different IoT solutions is challenging. ENISA already remarked the need for har-
monization of security certification that could help to increase the trustworthiness
and competiveness of European products [6]. Following the recommendations of
ENISA, there are some elements that should be harmonized. This is the case of the
different assurance levels, the elements considered during the certification process,
and the roles of the involved stakeholders. Regulatory bodies have an important
role here, promoting the creation of cybersecurity framework through the consen-
sus of the main stakeholders and orchestrating its development and deployment.
In particular, the certification meta-scheme proposed by European Cyber Security
Organisation (ECSO) [7] represents an ambitious initiative that homogenizes and
aggregates different certification approaches under a common framework.
Another challenge is related with the standardization of the cybersecurity
scheme. Despite the limitations of the current approaches, a cybersecurity certi-
fication scheme should adopt the main concepts, terms, and operational aspects of
existing standard approaches [such as Common Criteria (CC) [8]]. The usage of
standardized concepts helps to a common understanding and to the harmonization
of the cybersecurity certification.
Existing approaches are usually expensive, slow, and complex, requiring many
documentation and processes [9, 10]. This could imply that a small company could
not afford the costs of the certification process or even a delay on the market release
of the product, with high money losses.
Cybersecurity is a very dynamic concept. Taking into account the frequency
of updates and patches of certain devices, a lightweight recertification process is
necessary to ensure an updated security certificate. Automated procedures are also
necessary to ensure the scalability of the (re)certification process. In this sense, the
Security Certification Challenges in Current Schemes 181
cybersecurity certification scheme should deal with the changes of the certificate.
On the one hand, the product should be monitored during its lifecycle in order
to detect new vulnerabilities and update its security level. On the other hand, the
security level should be modified when a cybersecurity recertification is required
due to an update/patching.
Another key point is that a system composed by IoT devices could be composed
by several components with different levels of security. Security composition could
be a desirable design feature in cybersecurity certification and deployment at least
for some security properties (e.g., confidentiality, authentication), but it is usually
hard to perform in a way that all the layers and threats are correctly weighted [11].
The security certification process should also take into account the protocol stack
to cover vulnerabilities at different layers. Indeed, RASEN project [12] proposes
a mechanism to aggregate risks from different layers. However, physical aspects
remains as a challenge, due to the context dynamicity of the IoT, which means that
security properties are difficult to separate from other layers of an IoT component.
As a result of the certification process, a label should be generated to provide
a simple, clear, and visual level of the security certified [13]. Companies such as
Bosch [14] add that customers need to compare the security of different prod-
ucts without feeling overwhelmed with technical details. In this sense, the label
has to address an important trade-off between the simplicity of the label and the
non-ambiguous and complete representation of the results of such process. This is
rather difficult, because in comparison to the energy label, which measures a phys-
ical quantity, the measurements of security are far more complex. In addition, the
label design should also take into consideration the dynamicity of the security. As
pointed out by ECSO WG1,1 a visual static cybersecurity label is not enough, since
it should also cope with the dynamism of the IoT paradigm to reflect changes on
the current security level. For this reason, the usage of a digital QR code, which can
be regenerated, can help to check the status of the cybersecurity label in a fast and
easy way, as it could also be easy to update.
The context in which the product will operate must be considered, in order to
make products comparable among each other and to specify the boundary condi-
tions of the context where the cybersecurity certification was applied. This aspect
is specially challenging, because it could not be known a priori.
To address the issue of label significance and the need to measure the security
properties, security metrics must be established. However, some of such metrics,
such as likelihood or impact, are difficult to be measured, due to its complexity,
which is reported in [15].
1. https://www.ecs-org.eu/working-groups/wg1-standardisation-certification-labelling-and-supply-chain-
management
182 Cybersecurity Certification in IoT Environments
While some of the previous challenges are being currently addressed by Euro-
pean organizations, the security dynamism of the IoT paradigm and the burden
of existent schemes make the adoption of an IoT cybersecurity certification frame-
work challenging. The next section gives a brief overview of the main current cer-
tification schemes and how they deal with these challenges. For a more detailed
description, [5] poses a complete source of information of the security certification
and standards.
that the product complies its security specifications) and efficiency analysis (measur-
ing the strength of the security functions and mechanisms). One of the key points
of CSPN is that the evaluation is performed in a short period of time through the
adaptation of the product development lifecycle, reconciling time needs of the man-
ufacturing with the security assessment. CSPN is complementary to CC and can
be used as a previous short and non-expensive assessment to CC certification [21].
The IoT Security Testing Framework [22], developed by International Com-
puter Security Association (ICSA) Labs, specifies security testing requirements for
different types of IoT systems. The approach is based on a periodic assessment and
update of the certification criteria, addressing the dynamicity inherent to IoT. To
strengthen the assessment process, ICSA Labs (or associated accreditation bodies)
also carry out random assessments of the current product, and a short period of time
(2 to 4 weeks) is given to fix any flaw encountered or the certificate may be revoked.
The certificate is renewed annually by repeating the whole process. However, there
is no MRA, and the criteria and processes are not standardized.
The European Privacy Seal2 certifies if an IT product is compliant with the Gen-
eral Data Protection Regulation (GDPR) [23]. Other legislative documents from
the Member States could be also taken into account. For the certification pro-
cess, the regulatory requirements are translated into questions to be answered or
addressed. However, the process is done manually by experts answering the related
questions, meaning that the assessment can be affected by personal judgments.
In addition, it is focused on the privacy aspect.
The Singaporean National IT Evaluation Scheme (NITES)3 is mandatory for
manufacturers that provide IT products to governmental agencies in Singapore.
Although the specifications are private, it is based on a customization of CC stan-
dard v3.1 [5] at the approximate level of EAL4+ with some additions, which are
mostly targeted to the vulnerability analysis. Therefore, they share similar flaws,
such as the heavy documentation and the management of the security changes.
The ULD Datenschutz-Gutesiegel4 is a German IT certification scheme used by
German public authorities in order to verify the compliance with the data protec-
tion and data security rules. Although for minor changes it considers a lightweight
certification process, for major changes, the overall assessment has to be repeated.
Finally, the Cellular Telecommunications Industry Association (CTIA) is a
recent IoT certification scheme for LTE and/or WiFi devices [24]. The certification
process takes into account security aspects such as authentication, encryption, patch
2. https://www.european-privacy-seal.eu
3. https://www.csa.gov.sg
4. https://www.datenschutzzentrum.de/guetesiegel
184 Cybersecurity Certification in IoT Environments
documenting the technical context of the TOE. In the parallel perspective, risk-
based testing, the Test Planning activity is intended to determine the test strategy
specifying the testing techniques and test phases. This perspective integrates the
test planning with the risk assessment process by using security risk assessment to
identify high areas of risk and optimize the testing efforts needed in consequence
(see 1 in Figure 10.1, right). Moreover, the ETSI proposal points out that a previ-
ous assessment process could be useful to establish the test strategies necessary to
deal with the critical risks.
The main part called Security Assessment, also common in both perspectives, rep-
resents the integration between the security risk assessment and the security testing
work streams. In particular, it combines typical activities considered in the ISO
31000 (Risk management—Guidelines) and ISO 29119 (Software and systems
engineering—Software testing) standards. The two perspectives differ in the way
these two streams interact between them.
On the one hand, in the test-based risk assessment perspective, security risk assess-
ment is composed by three activities. The first one, risk identification, is meant to
find, recognize, and analyze the associated risks of the TOE. Here, the expert opin-
ion, historical data, and stakeholder’s needs can be considered. In this perspective,
risk identification is improved by security testing, which is able to identify actual
vulnerabilities or areas particularly vulnerable (see 1 in Figure 10.1, left). This can
be done through the usage of network discovering techniques or vulnerabilities
scanners. The second one, risk estimation, determines the risk level, understanding
the origin of the risk and its consequences. However, risk estimation is usually a
hard activity, as the information for the estimation is often imprecise and subjective,
relying on the expert judge. This perspective tries to solve this issue by providing
additional input from the security testing process with the objective to improve the
confidence of the risk measurement (see 2 in Figure 10.1, left). Finally, risk evalu-
ation determines if the risk obtained in the risk estimation activity is acceptable or
not, depending on the context established at the beginning of the process.
On the other hand, and following the risk-based testing perspective, security
testing is also composed by three activities. Firstly, test design and implementation
is meant to design, implement, and derive the test cases that are going to be per-
formed over the TOE. Here, security risk assessment can provide information about
the potential vulnerabilities to systematically determine the test purposes. Also, the
risk estimation obtained from the risk assessment can be used to prioritize the test
cases (see 2 in Figure 10.1, right). Test environment setup and maintenance sets up
the environment in which the tests are going to be executed. Finally, test execu-
tion, analysis, and summary executes the tests and analyses their results. One of the
main challenges in this activity is the extension of the testing process taking into
account the budget, the time, and the likelihood of discovering new vulnerabilities.
186 Cybersecurity Certification in IoT Environments
example of the proposal instantiation. The different steps of the figure are going to
be described in the next subsections to clarify the insights of the process. For more
detailed examples, the reader can refer to [26–28].
As a starting point, the certification approach considers a set of five general secu-
rity properties obtained from the literature, in which more specific vulnerabilities
and threats are mapped (e.g., from existing vulnerabilities databases such as the
National Vulnerability Database (NVD)) [32]:
It should be noted that the proposed mapping is designed to measure the risk
associated with the lack of each property, so the resulting security label can be spec-
ified by using such properties to provide a multidimensional security description.
The first activity, Test design and implementation, designs a test suite to test the
risk of each vulnerability. With the objective of automating this process, we use a
Model Based Testing (MBT), which generates the tests from a high level model of
the TOE [29]. Indeed, MBT has shown its benefits and usefulness for systematic
compliance testing of systems [30]. On the one hand, our approach uses the Unified
Modeling Language (UML) class diagrams to model the architecture of the TOE,
that is, the entities involved and their relations. Figure 10.2 shows an example of
an architecture modeled in UML. Specifically, the figure shows two main entities
(smart object and server) that are connected with other two entities (request and
response), by the relations “receiveRequest”, “sendRequest”, “receiveResponse,” and
“sendResponse”. On the other hand, the TOE behavior is expressed in Object Con-
straint Language (OCL),5 using the CertifyIt tool [31]. It is worth noting that other
tools can be used to automate this process [32]. The high-level tests are defined in
CertifyIt by means of test purposes, referred to the high level model. Figure 10.2
shows an example of test purpose, in which any operation can be used any number
of times with the objective of reaching a specific tag called “smart_object_ends_ok.”
The power here resides on the tags we can specify in OCL and the ability of Cer-
tifyIt for generating all the steps of the test necessary to reach the tag. This way, it
is not necessary to specify step by step the test, implying less time to program and
more time to think about security.
The generated tests can be exported in several languages (XML, PDF, JUnit, and
TTCN3). In particular, we export the tests in Testing and Test Control Notation
(TTCN) v.3 language, as it is a standardized testing language whose aim is to sys-
tematically execute the test, improving efficiency and scalability. Figure 10.2 shows
an excerpt of a test exported in TTCN-3 language. To cope with the particulari-
ties of each IoT device, CertifyIt also produces some interfaces, called adapter, as a
middle interface between the TTCN-3 and the real device. The adapter has to be
implemented to link the operations and the types defined in the high-level model
with the real code.
In the Test environment set up and maintenance activity, the execution environ-
ment is set up (e.g., a local device or a large-scale platform). Our proposal con-
siders the usage of the FIT IoT-LAB platform, a large-scale infrastructure (about
2000 IoT nodes) for testing purposes without the need of cumbersome deployment
tasks [33]. In this case, the setup of the environment consists of the reservation of
the resources and the upload of the code to the remote devices.
Finally, the tests are executed in the Test execution, analysis, and summary activity.
To do so, we use the tool TITAN. TITAN is a TTCN-3 compilation and execu-
tion environment for different platforms that in combination with CertifyIt creates
5. http://www.omg.org/spec/OCL/2.4
190 Cybersecurity Certification in IoT Environments
executable tests. TITAN sends the test commands (TTCN3 language) to the real
device through the adapter in order to execute the tests. As a result of this process,
it generates a test report indicating the tests that have passed or failed as well as
additional information (e.g., time, number of devices, errors, sniffer information).
The automation of the testing process implies that if a new vulnerability is dis-
covered, the recertification process can be done in a non-expensive, fast, and easy
way, which is key to address the dynamic nature of cybersecurity in IoT.
It is worth noting that the proposed tools for the automation are independent of
the methodology used to automate the process. This chapter proposed a particular
combination but other one is also possible.
results are mapped into CVSS metrics. There, the percentage of non-ciphered data
is assigned to the attack vector, and the complexity of the attack is based on the
ciphersuite used (algorithm and key length). The metrics are combined using the
CVSS formula [34].
Finally, the risk evaluation activity compares the obtained risk in each general
vulnerability with the profiles available for a specific context. To do so, the numeri-
cal risk is mapped to a risk interval (low, medium, high and critical), and the profile
fulfilled is obtained by comparing the risk interval with the risk interval of the pro-
file. Figure 10.3 shows the process of mapping the risk intervals and comparing the
result with the profiles (A, B, C, and D). In this case, for the lack of confidentiality,
the TOE obtained a high risk, which fulfils profiles C and D. As the chosen profile
is always the highest one, the profile obtained for that general vulnerability is C.
This process is repeated for each general vulnerability.
10.5 Conclusion
This chapter has discussed the main challenges that have to be addressed towards
the creation of a cybersecurity certification framework, and how current certifica-
tion schemes fail to deal with some of them. As a response, we have proposed a
methodology based on the ETSI proposal that combines risk assessment and test-
ing to assess IoT devices in an objective and automated way, coping with one of the
main challenges: the security dynamism. The methodology proposed is intended
to serve as a basis for the development of a common security certification scheme,
with the aim of supporting the implementation of the Cybersecurity Act.
Acknowledgments
This work has been partially funded by the European Union’s Horizon 2020
research and innovation program under grant agreement no-779852 (IoTCrawler)
projects and grant agreement no-830929 (CyberSec4Europe) by the FPU-
16/03305 research contract of the Ministry of Education and Professional Training
of Spain.
References
[5] ECSO, “State of the Art Syllabus v2”. 2017. [Online] Available in: https://
ecs-org.eu/documents/publications/5a31129ea8e97.pdf
[6] H. Baars, R. Lassche, R. Massink, and H. Pille, “Smart grid security certifica-
tion in Europe. Challenges and recommendations”. ENISA, 2014.
[7] ECSO, “A Meta-Scheme Approach v1.0”. 2017.
[8] CCRA, “Common Criteria for Information Technology Security Evaluation.
Part 1: Introduction and general model.” 2017.
[9] S. Murdoch, M. Bond, and R. J. Anderson, “How Certification Systems Fail:
Lessons from the Ware Report”, IEEE Security & Privacy Magazine, vol. 10,
no. 6, pp. 1–1, 2012, doi: 10.1109/MSP.2012.89.
[10] S. P. Kaluvuri, M. Bezzi, and Y. Roudier, “A Quantitative Analysis of Common
Criteria Certification Practice”, en Trust, Privacy, and Security in Digital Busi-
ness, vol. 8647, Cham: Springer International Publishing, 2014, pp. 132–143.
[11] M. Bartoletti, P. Degano, and G. L. Ferrari, “Security Issues in Service Compo-
sition”, en Formal Methods for Open Object-Based Distributed Systems, Berlin,
Heidelberg, 2006, vol. 4037, pp. 1–16, doi: 10.1007/11768869_1.
[12] RASEN project, “D3.2.3. Techniques for Compositional Test-Based Security
Risk Assessment v.3”. 2015.
[13] AIOTI, “Report on Workshop on Security and Privacy in the Hyper-
Connected World”. 2016.
[14] J. Hubner and M. Lastovka, “BOSCH Political Viewpoint. Security in IoT.”
2017.
[15] J. Hearn, “Does the common criteria paradigm have a future?”, IEEE
Security & Privacy Magazine, vol. 2, no. 1, pp. 64–65, ene. 2004, doi:
10.1109/MSECP.2004.1264857.
[16] Common Criteria, “Arrangement on the Recognition of Common Criteria
Certificates In the field of Information Technology Security”. 2014.
[17] F. Keblawi and D. Sullivan, “Applying the common criteria in systems engi-
neering”, IEEE Security & Privacy Magazine, vol. 4, no. 2, pp. 50–55, Mar.
2006, doi: 10.1109/MSP.2006.35.
[18] CESG, “The Commercial Product Assurance (CPA) build standard”. 2014.
[19] Underwriters Laboratories, “UL 2900 Standards Process”, UL 2900 Stan-
dards Process. [Online]. Available in: https://industries.ul.com/cybersecurity/
ul-2900-standards-process.
[20] ANSSI, “Certification de sécurité de premier niveau (CSPN)”, ANSSI, 2008.
[Online]. Available in: https://www.ssi.gouv.fr/administration/produits-cer
tifies/cspn/.
[21] G. Baldini, G. Giannopoulos, and A. Lazari, “Annex 8: JRC Analysis and
recommendations for a European certification and labelling framework for
cybersecurity in Europe”, European Commission, 2017.
194 Cybersecurity Certification in IoT Environments
Chapter 11
The CHARIOT design method is oriented to bridge the systems engineering gap
that currently exists between (a) the formal safety engineering techniques applied in
the development and testing of safety critical systems and (b) the rapidly evolving
and ad hoc manner in IoT devices are developed and deployed. In this direction,
CHARIOT started with a classification and usage guidelines of relevant standards
and platforms to align the project developments with existing guidelines, standards,
and industrial challenges on Industrial IoT Security. A free software compilation
toolset is also being developed (based on existing open source technologies) targeted
towards IoT engineers designing IoT systems and developing source code running
on devices in Industrial IoT networks. This introduces new methods and tools for
more secure and safer IoT software development with static source code analyses
to help avoid safety and security defects and risks targeting the firmware developer
during the building steps and the software security engineer before the firmware
deployment. Source code analyses are executed offline and are strongly bound with
the CHARIOT Security Engine to online check results. For that purpose, the static
analyses add meta-data (cryptographic signature, static analysis results, and sup-
port for “proof-carrying code” approach [1]) into the binary, hence permitting that
binary executable to be suitably “filtered” or “authenticated” by the CHARIOT
196
Firmware Software Analysis at Source Code 197
Security Engine and the IoT devices and, in turn, shielding against cyberattacks at
source code (compilation and run-time) level.
At the same time, CHARIOT recognizes security vulnerabilities present at
devices producers’ level underestimating the potential risks to which they are
exposed to, since the latter are connected almost continuously to the internet. Great
risks are created by superficial behavior conducted by the manufacturers themselves
who are oriented to release the firmware updates with an excessively low frequency
and sometimes, fortunately only in extreme cases so far, the updates are not released
at all. Existing challenges related to the actual firmware running in IoT devices
include reverse-engineering of the entire firmware, extracting the file system and
understanding how the entire device works, including knowing the possible use of
known-to-be-vulnerable out-of-date API/libraries or unknown exploitable vulner-
abilities in the firmware code itself. They also include backdoor insertion, change
of the device behavior, altering its performance, without altering the telemetry val-
ues reported to administrators, finding hard-coded private symmetric-cryptography
keys/passwords/user-names or private certificates used to encrypt communications
between the device and other systems and rolling back the firmware to a previous
legitimate version with known vulnerabilities (to be exploited). This attack is par-
ticularly insidious because the pushed firmware is authentic, so it can easily survive
most of the in-place controls, as usually, they tend to check just the firmware source
and/or the firmware integrity.
CHARIOT introduces a firmware checking approach comparing the individual
functions that make up the firmware, looking for similarities between the functions
themselves (according to certain parameters to be extracted and analyzed) and the
behavior of the firmware according to statistical surveys. The firmware update pro-
cess is also affirming compilation level warnings (metadata), firmware hashing and
versioning through the blockchain system, as well as the firmware analysis results.
The CHARIOT firmware checking is multiple from the development of the
first firmware components to the deployment and the upload of the firmware at
the level of sensors, gateways, and other nodes. To prepare an industrial deploy-
ment, some verification steps are optional and depend on the Security Policy of
the Industrial. The mandatory firmware verification concerns the one performed
by the Security Engine in the Fog Node, just before the firmware deployment.
Hence, the Security Engine has the responsibility of the firmware deployment on
the sensors, gateways, and other nodes. Since the Security Engine has few times to
take a decision for the acceptance of the reject of an update, the firmware needs to
exhibit detailed information about its content. The different verification steps are
registered in the blockchain that gives insurance about the origin and the integrity
of the verification process.
The CHARIOT architecture is based on the deployment of Fog Nodes that
are the entry point of all the communications to the sensors, gateways, and other
198 Firmware Software Analysis at Source Code
Figure 11.1. Firmware development and deployment (in green) in the CHARIOT
architecture.
notes. The Figure 11.1 defines the firmware development and deployment path on
the architecture of CHARIOT.
the crucial analysis results will be checked by the Security Engine without any
assumptions on their origin (inspired from proof-carrying code approaches).
results with caller/callee information. All this work operates incrementally on the
analysis data provided by the developer.
We present now an example of BISMON usage. It concerns both safety and
security. To ensure some integrity for the analysis results, the developer needs to be
identified in the BISMON server as explained in Figure 11.2. Then, it launches
the compilation command, which produces the analysis results and send them to
BISMON. Let us see how it works on for an analysis that computes the maximal size
of the stack usage during the firmware execution. For Safety and Security reasons,
when the firmware runs, it should guarantee that the stack memory that it uses will
not overflow over the available memory. This can be ensured by a static analysis at
source code level. During the compilation process, the BISMON analysis computes
the maximal stack frame of every function. Such an analysis is only possible in the
compiler that produces the final binary firmware since the compiler knows these
data—it needs them to call and return from functions (update of the stack frame
registers). In the gcc plug-in = BISMON client, it is implemented as a gcc low-
level pass. The BISMON analysis also produces the couples caller/callee to be able
to produce the complete call graph. Then, BISMON sends these data: maximal
stack frame by function and the couples caller/callee to the BISMON server. At
the server side, BISMON detects the absence of recursive functions that could lead
to an infinite stack usage. It transforms the call graph into a call tree by “inlining”
the function calls at the caller site. It then incrementally update the call tree each
time the developer sends a new translation unit. When the developer operates the
link edition that generates the final executable, it sends this information to the
BISMON server. At this point, the server knows that it has the entire source code.
So it computes an overestimate of the stack height for every leaf of the call tree and
defines the maximal value as the result for the stack usage. It then converts this result
into CHARIOT meta-data and attaches them at the end of the binary firmware—
in the CHARIOT meta-data section. It informs the blockchain of its work. Later,
the firmware will be uploaded: the Security Engine can look if the meta-data are
present and if the stack usage is bounded. If it is the case, it authorizes the sending to
the CHARIOT Gateway. The gateway extracts the meta-data from the firmware:
it knows the sensors to update; so if the stack usage is greater than the available
memory it will not send it for the update; instead, it informs the CHARIOT Fog
node that the update has not occurred.
BISMON can automatically verify (safety and security) coding rules (some
buffer overflows, Insecure random numbers, etc.) and identify the lines of source
code that break a particular coding rule (no recursive function, no stack overflow)
as safety properties. As an example, BISMON security properties can check that
every call to a communication library should have URL addresses that are statically
well known with identified corporate prefixes. Such coding rules might prevent
some developer’s mistakes but it will not prevent an attacker from introducing
202 Firmware Software Analysis at Source Code
malicious code – this problem is more the scope of existing formal methods (e.g.,
axiomatic semantics, operational semantics, and abstract interpretation) developed
in the next section.
In the CHARIOT approach, additional specific static analyses embed semantic and
formal analyses (to check the absence of memory overflow, of Cross-site scripting).
If a formal analysis generates no alarm, then the firmware execution cannot expose
the vulnerability. Such type of analysis is conducting by a Security Engineer that
inspects the source code, the binary code to look for security vulnerability. Formal
analysis has also the advantage that their results are objective. Hence, the Security
Engineer does not need to sign that his analysis is correct, since the results can be
checked again by the Security Engine.
Consider the example of memory overflow: In some cases, they create security
breaches since an attacker can exploit a memory overflow to take the control of the
machine by executing own (external) code with an overflow of one or two bytes,
jumping between codes (code injection vulnerability). If a formal analysis generates
some alarms, the Security Engineer will perform different actions depending if the
alarm is a true one or a false one. For a true alarm, he will ask or do a code correction;
For a false alarm, he will write/generate a formal reasoning that complements the
analysis’ reasoning and so removes the false alarm for the security engine. At the end
of this process, the running system should not have any vulnerability targeted by the
analysis. The natural vulnerabilities targeted are Run-time Execution Errors (mem-
ory overflows, integer overflows, invalid pointers, undefined behaviors). Security
involves other risks and new software vulnerabilities will be investigated (tainting,
impact analysis) in accordance with the Security policies of industrial IoT players.
Some tools like Frama-C [3] can perform such formal analysis for IoT code [4].
Current development work in CHARIOT concerns a way to save such formal
results in the firmware meta-data to check these results again with the Security
Engine. It addresses formal source code verification and formal binary code verifi-
cation, depending on the presence on source code or not. It is an innovative chal-
lenge in both cases: the link between the source code and the binary code needs to
be established for source code verification; the formal analysis at binary code needs
to recover the data structures, the control, and data flows.
The Security Engine aims to keep the firmware (binary) of IoT devices secure
through the CHARIOT environment as an important attack vector and a highly
Technological Innovation and Security Alignment Per Outcome 203
• Reverse engineer the entire firmware, extract the file system, and understand
how it would be used for known-to-be-vulnerable out-of-date API/libraries
or unknown exploitable vulnerabilities in the firmware code itself;
• Insert a firmware backdoor, making the device controllable by remote;
• Change the device behavior, altering its performance and the telemetry val-
ues;
• Find hard-coded private symmetric-cryptography keys/passwords/user names
or private certificates used to encrypt communications between the device
and other systems; the attacker may be able to eavesdrop these private com-
munications.
During the software update process, the security engine extracts the static code
analysis results (meta-data) from the firmware binary as inserted earlier at the
binary compilation stage. The extracted static code analysis results are then cross-
referenced with a local security policy that dictates the tolerated static code analysis
limits, and if the firmware passes these limits, the binary analysis process is initiated.
During binary analysis, insight to the internal workings of the firmware is provided
by comparing jump addresses with historical firmware update data, when avail-
able, and databases of common vulnerability jump patterns. This binary analysis
204 Firmware Software Analysis at Source Code
is then consumed by the security engine to approve or reject the firmware update
and provide contextual messages to the initiator of the update, such as the type of
vulnerability pattern identified. The overall workflow achieves two things: firstly,
it prohibits anyone to issue a firmware update that possesses warnings that were
ignored from the static code analysis if they do not satisfy the criteria set forth in
the local security policy; and secondly, it prevents any firmware update of going
through if a vulnerability or erratic pattern in comparison to its previous versions
is identified in the jump registers of the binary file.
The heuristic approach first treats the system as different sub-systems so that the
sub-system’s solution must spread widely at the solution space. Similarly, following
this approach, the analysis was addressed by considering different aspects of the
characteristics of the firmware and its typical behavior.
So, the optimal sub-system solutions in the heuristic approach for CHARIOT
firmware devices are found:
Finally, the optimal solution will be determined by matching all the sub-system
solutions, calculating a final reference for the statement of security of the new
firmware, finding the global optimal solution more efficiently.
First phase is the Data Collection; the scope of this stage is to parse the ELF,
BIN, and HEX firmware images using RADARE2 opensource tool to load images,
collect data, parse an image with tool analyzer in order to get names and addresses
for functions, build a map of function names and extract the CFG (Control Flow
Graph), and perform initials automatic analysis.
Once each firmware function is extracted (together with basic information like
addresses, names and sizes), each function is disassembled to:
• Get all the instructions used inside the function, classify the commands based
on their behavior, get conditional/unconditional jumps, loops, etc.;
• Count the number of each type of instructions (standard and classified),
jumps, etc.;
• Note offsets of commands.
Technological Innovation and Security Alignment Per Outcome 205
In the second phase, the scope is to analyze the function structures taking care
marginally the order of instructions inside each function, as well as the offsets,
parsing the divided code in a function-by-function manner without any correlation.
Then, a fingerprint sha256 for each function (a sort of UID) has been calculated,
so that we are able to make a comparison between functions by matching their
fingerprint.
If they match (independently from their correlation), there would be a rea-
sonable confidence assurement for the possibility of a complete correspondence
between them. A function-by-function match allows the algorithm to map all
the instructions inside the two examined firmwares and perform an effective
comparison.
In addition, in order to give greater reliability to the algorithm, fingerprints that
take into consideration offsets have been created. This can be added as a separate
parameter in the implementation. Multiple options can be considered as contri-
butions to finger print: statistics, instructions’ order, and offsets. Obviously, these
options influence the final comparison results. Some functions may have the same
number of instructions but a different order. Some functions may be considered as
equal by these parameters, but have different offsets, etc.
Then, the generated fingerprints from two images are compared to define which
functions have been changed. It is noted that only fingerprints are compared. Func-
tions are considered “equal” if fingerprints match. If fingerprints do not match,
functions are considered as “removed” or “new,” depending on whether or not, it
is present in the first reference firmware.
Eventually, a similarity threshold is used to perform a more in-depth analysis;
two firmware images are compared function by function based on the number
of instructions. The number of differences for each instruction is calculated. If
this number is bigger than the similarity threshold, then the function is marked as
“removed” or “new.” If smaller, it is then marked as “similar.”
At the end of the process, the comparison algorithm calculates the confidence
rate which shows whether numerous changes were made. This rate can be used
like a useful but not final reference to establish if gross potential vulnerabilities can
affect the new release of the firmware and if it needs to be further analyzed to verify
other critical issues.
Based on the signatures, it is made by looking for all the printf functions in
the firmware code and analyzes the first parameter that is passed to it. If this is
a formatting string like “%s” or “%d”, then the code is safe. If we see a memory
address or a parameter of the function that has initially called the printf function,
then we have a potentially vulnerable place.
This feature provides functionality for looking through code and detecting all the
read and write memory accesses. After such places are detected, analysis is per-
formed to detect whether the address value was changed using some relative value,
i.e., whether it has increased or decreased comparing to the base address. Next the
obtained offset values are analyzed, and it is determined whether it is allowed to go
beyond the boundaries of the allocated memory. Due to the peculiarities of static
analysis, values of the parameters from which the functions are called cannot be
given and cannot return an exact answer whether this place is vulnerable or not.
This leads to false positives. If a metadata provided by the offline static analysis
provides a symbolic offset, then the security engine checks the formal reasoning
that ensures this metadata and it then uses this metadata to remove the alarm. This
can optionally be turned off via analysis options.
This feature uses signatures to look for the vulnerable functions (strcpy, strcat,
memcpy, etc.). After such functions are detected, all the places where these func-
tions are called are determined. Next, the parameters that are passed to them are
analyzed to determine if an overflow is possible in this place or not.
dashboard can alert the Security Engineer. Then, the Security Engineer can inspect
the code again and add in the meta-data formal explanations about why the code
is correct. The next firmware upload should be accepted because the CHARIOT
Security Engine will now check the formal reasoning of the Security Engineer in
relation with the incriminated portion of code.
Acknowledgments
This project has received funding from the European Union’s Horizon 2020
research and innovation program (No. 780075). The authors acknowledge the
research outcomes of this publication belonging to the CHARIOT consortium.
References
Chapter 12
According to market forecasts, Internet of Things (IoT) products and services are
spreading very quickly in all professional and mass-market usage scenarios in terms
of revenues and volumes of devices and services. In the meantime, lots of pilots and
commercial deployments are demonstrating the added value of IoT-based solutions
in real scale, and in all market segments, from the transportation industry to the
utility sector but also in smart building, factory of the future, or healthcare appli-
cations. The use of historical and live data in big data lakes will allow recognizing
anomalies, implementing alert functions, calculating solutions and optimizations
for a more efficient digitalized world. In order to build trustable services and solu-
tions, devices information needs to be trusted. This can only be ensured if IoT
systems are end-to-end protected, ensuring protection against corrupted third par-
ties. Indeed, while state-of-the-art IT communication protocols such as HTTPS
ensure end-to-end security, many IoT communication protocols have strong con-
straints in terms of bandwidth, power, and computation capabilities that do not fit
state-of-the-art security protocols. In this context, BRAIN-IoT project has tested an
innovative approach to ensure end-to-end security on constrained protocols with
limited impact on bandwidth and power consumption.
208
IoT Systems Architecture and Security Challenges 209
12.1 Introduction
While in the past we had hardware and software in charge of monitoring devices
and/or processes in a separate Operational Technology (OT) network, today’s IoT
systems are directly connected to IT networks that fuel data lakes with historical
and real-time data. This in return allows the recognition of anomalies, the imple-
mentation of alerts, the finding of solutions, and the dynamic optimization of the
overall resources allocation. As modern networks are now formed by hundreds or
thousands of connected nodes, the threat of cybersecurity breaches has increased
too. To reduce this threats coming from small sensors connected to the cloud and
then to the entire network, the IoT devices need to be protected.
Researchers have pointed out security vulnerabilities of Low Power Wide Area
Network (LPWAN). Applying traditional IT technologies on Cyber-Physical Sys-
tems (CPS) is not a solution as it creates many single points of failure that could
lead to data flow disruption and exposure. Another challenge is the use of corrupted
edge equipment, which could lead to secret leakage and device identity spoofing.
Moreover the huge number of devices and their specificities lead to complex
management activities with high operating costs, especially regarding device iden-
tity and security management. Last, LPWAN devices often make trade-offs between
cybersecurity and calculation and battery power to reach the promised battery life
time, which is typically around 10 years. Indeed, the technological advances in
network architectures, operational technologies, Cyber-Physical Systems, and mat-
uration of cloud services have opened new business opportunities for critical and
privacy sensitive application. However, the lack of software-based security embed-
ded in the sensors and the proprietary provisioning methods delivered by network
providers are preventing the commercial deployment from booming.
In this chapter, we will first have an overview of IoT systems architecture and
associated challenges, especially regarding LPWAN technologies introduction in
existing architectures.
After the architecture and security challenges overview, we will present relevant
associated countermeasures that are available at the state of the art.
Finally, we will describe an innovative approach, allowing to secure very low
power devices, which are currently the weakest point to infiltrate an IoT system.
LPWAN are a low cost and a reliable alternative to link low battery consump-
tion sensors with long-range (>2 km) as shown in Figure 12.1. At the moment,
LoRaWAN and SigFox are the well-known LPWA Networks. However, less known
LPWA Networks also exist, such as RPMA, Symphony Link and Weightless.
In 2017, LoRaWAN claimed to have 52 networks publically announced, 350+
ongoing trails and city deployments and presence in over 100 countries worldwide.
For its part, SigFox has presence in 36 countries and a coverage of more than 85%
of the population in 17 countries. The latest forecasts report that the accumulated
number of LPWAN devices shipment between 2018 and 2023 will be approxi-
mately 600 Millions as it is highlighted by Figure 12.2.
Those new architectures based on LPWAN communications, combined with
OT and Cyber-Physical Systems, open a wide range of new opportunities for busi-
ness critical and privacy sensitive applications. An exemple of a multi-networks IoT
system is presented in Figure 12.3 IoT Industrial Cloud systems are rapidly matur-
ing as packaged systems that can sense, collect, analyze, and action on edge devices.
However, the software-based device security and manual/proprietary provisioning
methods delivered by platform providers are preventing deployments from scaling.
IoT Systems Architecture and Security Challenges 211
LPWAN and mesh short-range radio networks today sacrifice security require-
ments in order (i) not to overload the limited memory and CPU power of very light
micro-controllers and (ii) to save energy so that they can reach the promised bat-
tery lifetime. Many researchers have pointed out security vulnerabilities of those
212 End-to-End Security for IoT
In order to address the security threats highlighted in the previous section, industri-
als set up mitigation measures to limit the impact of cyberattacks on those systems.
In this section, we will classify mitigation measures in 3 main categories:
• Device management
• Endpoint protection
• Communication security
State-of-the-Art Mitigation Measures 213
In the area of LPWAN, end devices have many constraints such as hardware cost,
lifetime duration, bandwidth, and computation power that are not compatible with
the use of a dedicated hardware component.
(and asymmetric cryptography) limits the need to share secrets to establish end-
to-end secure channel.
The arrival of LPWAN, however, reveals the limits of those technologies:
The Figure 12.6 above provides an overview of each IoT management solution.
We can see that the only solution offering end-to-end protection is highly power
consuming and therefore cannot be used to secure low power devices.
security footprint and to guarantee its robustness, the security layer relies on well-
proven authenticated encryption algorithms such as AES-CCM* and AES-GCM
and tends to integrate lightweight security protocols such as SPECK to go a step
further in energy consumption impact.
Regarding the device authentication, this technology offers the following state-
of-the-art advantages:
• For the IoT manufacturer, to add trust at the device layer at reasonable cost
by providing a “super-slim” security software
• For the IoT operator, to simplify IoT device deployment with self-registration
capabilities and thus to reduce the time and cost necessary to deploy a secure
IoT system.
12.5 Conclusion
including slim encryption layer to ensure trustable and optimized data transmis-
sion. This solution also allows self-registration of devices in a plug & play mode,
ensuring automatic key renewal over constrained protocols.
This solution combines perfectly with existing security measures such as use of
secure element and network security capabilities to ensure the trustability and the
confidentiality of data, which is key to build new digital services.
Acknoweledgments
The work presented in this Chapter has received funding from the European
Union Horizon 2020 research and innovation programme under grant agreement
No 780089, Brain-IoT project.
DOI: 10.1561/9781680836837.ch13
Chapter 13
easier to spoof a single entity rather than a distributed system and validate mali-
cious identities. As more trust and value, informational or otherwise, is placed in
IoT networks, these types of attacks will become much more widespread as the ben-
eficial yield would outweigh the resources required to take such a network down.
To address this, the abstraction of the PKI stack is proposed to a distributed
layout with the aid of blockchain technology [2]. In PKIs, Certificate Authorities
(CA) act as the validation oracles for any identities in the network by issuing identity
certificates and subsequently validating them cryptographically. These authorities
possess a single cryptographic keypair that, if compromised, can result in devastat-
ing consequences as the possessor of the keypair would be able to issue, adapt, and
revoke an unlimited number of identities, thus resulting in the compromisation of
the whole network.
CHARIOT decided to break down the components that make up the issuance
and verification process of a CA, simplifying them where possible and finally re-
implement them using blockchain-based smart contracts. As an initial step, the
bare necessities of a PKI network were defined in association with a cryptographic
identity, thus an identifier, with a set of metadata that could either reside in on-
chain data or simply act as a validator of off-chain data, such as a cryptographically
secure hash. To enhance the basic indications of a PKI identity, which are “valid,”
“expired,” or “revoked,” an additional field is added to signal the status of the iden-
tity. This field takes the form of an unsigned integer, which is utilized in CHAR-
IOT solution as a binary abelian group code. This code can be used to represent an
arbitrary amount of statuses and, being a simple representation of ones and zeroes,
depends on the software implementation for assigning a binary group to a status
code. This results in a modular system that can easily be expanded in scope.
By using smart contracts, we ensure that all on-chain computations will
be relayed across the network and ensure that any alterations to the underly-
ing database will be available to queriors of the blockchain instance from any
blockchain node on the network. This allows the network to remain operational
even if a percentage of the network is compromised. Although in the CHARIOT
implementation, the IoT gateways were utilized as nodes in the blockchain net-
work, as the forthcoming technological advancement in the IoT’s computational
capabilities will enable them to become blockchain nodes themselves, empowering
verification of identities in a bilateral format (both gateway and sensor identities
verified) rather than a unilateral format (only sensor identities verified).
With regard to the actual invocation process of a contract’s identity issuance,
revocation, and adaptation method, a multi-signature protocol has been imple-
mented rather than a single-signature one. As the acceptance policy is precon-
figured during the setup of the CHARIOT solution, it is possible to construct
an n-out-of-m scheme that would satisfy even the most rigorous security policies.
222 Blockchain Ledger Solution Affirming Physical, Operational
For instance, for the installation of a sensor, the four-eye principle could be applied
by requiring signatures by both the installer of the sensor and the manager of the
site in which the sensor is installed.
(d) The third and final member of the payload is the identifier of the asset we
wish to alter. This identifier usually takes the form of the encoded public key
of the identity, although the protocol we have devised supports an arbitrary
identifier.
While the blockchain itself is inherently scalable due to its segregated structure, the
overlaying services needed to be somehow adapted to exemplarily operate under a
scalability scenario. To this end, deployment scripts were coded to automate the
process of spawning an instance of the software and used these scripts to create
deployment pipelines that contain instructions on how to deploy the various com-
ponents of the CHARIOT blockchain in the Docker environment. By storing the
code of the components in virtualized containers, overlay services were set up to
manage the life cycle of these containers automatically and scale the network as
necessary, ensuring that the systems remain operational even under extreme net-
work load. Additionally, load-balancing services can easily be appended on top of
the architecture to properly delegate network load to the containers spawned under-
neath the load balancer.
Conclusions and Future Work 227
Acknowledgments
This project has received funding from the European Union’s Horizon 2020
research and innovation program (No. 780075). The authors acknowledge the
research outcomes of this publication belonging to the CHARIOT consortium.
228 Blockchain Ledger Solution Affirming Physical, Operational
References
Chapter 14
14.1 Introduction
1. Vendors are typically companies that aim to maximize their profits, which
can be expressed as: revenue-expenses.
2. It is not possible for a vendor to generate significant extra revenue by improv-
ing the security of its products, since proving or measuring security of the
product is practically impossible (it cannot be practically proved that a prod-
uct is secure, only that it is insecure).
1. https://www.hackerone.com/
Introduction 231
3. Similarly, insecure products will not make the vendor lose a significant
amount of revenue, since the current Responsible Disclosure scheme does not
reveal the number of vulnerabilities to the general public and allows the ven-
dor ample time to fix its products. Since most of vulnerabilities are released
to the general public only after a long period of time, most of customers do
not suffer too much or too often from using insecure products, and therefore,
they do not have a clear incentive to switch to another vendor.
4. Creating more secure products is expensive (it requires more development
time, more experienced developers, better processes, more time for testing,
etc.); the vendor will often rather minimise the costs by making less secure
products.
Per 2 and 3 above, vendor revenue will not be adversely affected by insecure
products, yet per 4, creating less secure products reduces expenses. Therefore, per
1, providing less secure products tends to maximise the vendor’s profits.
Now, if the vulnerabilities would be disclosed in more strict and transparent
manner, the situation would change. Vendors behind insecure products would lose
reputation and would have a harder time selling products. Customers could, for
instance, demand agreements from vendors with bad security history that in the
case a serious vulnerability is found in the product, the vendor would pay compen-
sation to the customer, etc. As a result, vendors would have strong incentives to
improve the security of their products, and overall, there would be financial incen-
tives to provide more secure products and fix vulnerabilities quickly, which in turn
would lead to more secure products in the market.
Transparency of vulnerabilities is also important because even if the vulnerability
has been found and responsibly disclosed by a one expert, it does not mean that
no one else in the world knows about it. In fact, it is very likely that the same vul-
nerability has already been found by other actors, such as criminal organisations,
which do not have any incentives to disclose it to the vendor and commonly trade
vulnerabilities on black markets [7]. From the customers’ point of view, it is cate-
gorically better to know that a vulnerability exists (even if the whole world knows
about it), compared to the situation where the vulnerability exists but the customers
do not know about it. At least in the former case the customers can take actions to
mitigate the risk (stop using the vulnerable product temporarily, install additional
safeguards, etc.), while in the latter case the customers cannot predict or mitigate
attacks in any way since they are not aware of the possibility of the attack. Therefore,
a long time between the vulnerability discovery and the related public disclosure
increases the possibility of security breaches and decreases the overall security.
To succeed, any practical implementation of Responsible Disclosure also has
to be able to deal with the scale of vulnerabilities, so because of the huge num-
ber of IoT devices with diverse capabilities and operations, a fully—or almost
232 Leveraging Interledger Technologies
14.2 Background
of failure and offer resilience against many attacks. It is relatively easy to determine
if any of the participating nodes in the DLT are misbehaving and even in an extreme
case, where an attacker manages to control the majority of the DLT’s resources, the
attacker still cannot modify the existing data, only control the addition of new data.
DLTs can be implemented with different levels of openness. They can be fully
open (permissionless), which means that anyone can join the DLT and propose
transactions; most well-known DLTs such as Bitcoin and Ethereum are based on
this principle. However, DLTs can also be permissioned, which makes them either
semi-open, in which case read access is open to everyone but write access is restricted,
or closed, in which case both read and write access are restricted.
The main practical innovation of DLTs is the enablement of distributed trust.
While there have been multiple proposals for distributed databases in the past,
they have mostly concentrated on the distributed implementation, while the trust
model has remained firmly centralised. In contrast, DLTs allow various entities that
may not fully trust each other, such as individuals, organisations, and companies, to
collaborate in a safe and transparent manner, with only a low risk of being cheated
by others.
to safety concerns). At the same time, the growing number of IoT devices intro-
duced in recent years and the new types of data coming with them have increased
the security risks both for the industry [17] and the in consumer IoT domain [18].
As an example, a case study on baby monitors [19] details how many baby mon-
itors ship with static credentials and other vulnerabilities allowing access to the
confidential information provided by the devices. Together, the above-mentioned
risks in software systems and the flaws in the IoT systems can be highly danger-
ous and cause significant economic losses [20], which draws considerable discus-
sion on how such risks can be controlled and how related vulnerabilities should be
disclosed [21].
Over time, many different approaches have been proposed for the mechanisms
of vulnerabilities disclosure. Traditionally, the so-called full vendor disclosure has
been used; there the vendor has full control over how and when a vulnerability
is disclosed, which leaves many vulnerabilities totally unfixed or fixed only after a
long delay [22, 23]. In order to solve the problem of full vendor disclosure with
public at risk, the immediate public disclosure has been proposed, which introduces
a direct and strong incentive for the related vendor to fix the released issue as fast
as possible [24]. Even though early disclosure of the vulnerabilities can drive faster
fixing and uptake of mitigating measures by related parties to reduce the risk, the
attackers can also exploit the released vulnerability to attack already before a patch
is delivered by the vendor [25]. To address the conflicts between vendor and users,
a hybrid disclosure approach, responsible disclosure, was introduced: the vendor is
allowed some time to develop a patch before the vulnerability is disclosed by the
finder [26]. This way, the vendor is given the opportunity to fix the vulnerabil-
ity in time without the dangers of immediate exposure, but the public can still
learn about the existence of the vulnerability in case the vendor fails to deliver
a patch.
To coordinate the process between vendors, finders, and users, often a Com-
puter Emergency Response Team/Coordination Center (CERT/CC) is used as
a trusted third party. The economics of the vulnerability disclosure process
has been studied [27] to shed light on the interaction between participat-
ing parties. There was a trend of using market strength to achieve the opti-
mization of social benefit in the vulnerability disclosure process [28], though
many [29, 30] criticize that the market solution performs no better than existing
solutions.
Different vulnerability disclosure approaches have been compared and evaluated
in their efficiency [31], and with the corresponding cost, benefits and risks provided
in [32]. The work in [33] offers another angle for the life cycles of software vul-
nerabilities, which is based on analysis in seven dimensions for more than forty
thousand vulnerabilities reported in the past few decades.
Automated Responsible Disclosure (ARD) 237
The parties to responsible disclosure, and the proposed Automated Responsible Dis-
closure (ARD) mechanism, can be divided into the following categories, as shown
in Figure 14.1.:
To improve responsible disclosure, the ARD solution has been designed with the
following goals and assumptions:
4a. Patch
General
Public
Public ledger
3. The Vendor is clearly notified about each vulnerability and has to agree to
fix it within the grace period; otherwise, the vulnerability is immediately dis-
closed.
4. Releasing a fix results in automatic disclosure of the vulnerability.
5. If a fix has not been released before the grace period has expired, the vulner-
ability is automatically disclosed.
6. Authorities are trusted to try to operate correctly, e.g., to prevent uninten-
tional leaks of vulnerability information, correctly screen reported vulnera-
bilities, etc., but disclosure cannot be prevented by the Authority.
The solution is designed to utilize two distributed ledgers, a private one main-
tained by the Authority and used for storing the details of the vulnerability during
the disclosure process, and a public one for first disclosing the existence of a vulner-
ability and later the details of the vulnerability. A key element is then the interledger
functionality, which facilitates the automatic disclosure of the information from the
private ledger to the public one once the conditions for disclosure have been met.
In the basic case, the interledger functionality is run by the Authority or some other
trusted party, but for additional security and resilience, the interledger functionality
can be implemented in a distributed manner over a consortium of parties.
Technically, the public and private ledgers can utilize any suitable ledger tech-
nology (permissioned for the latter, and permissionless or permissioned for the for-
mer). The public ledger is readable by all, while some of the write operations are
restricted. Anyone has a permission to store metadata of the vulnerability informa-
tion (including its hash), which can be achieved, e.g., by a smart contract owned
by the Authority. Additionally, write operations related to confirmation of vulnera-
bility, notification of patching the vulnerability, and full disclosure of vulnerability
information are limited to the Authority, Vendors, interledger, and the Security
Expert who reported the vulnerability.
The private ledger can be implemented, for example, using Hyperledger Fab-
ric, a permissioned DLTs, and a separate chaincode (which corresponds to a smart
contract in Fabric) per Vendor. In Fabric, the vulnerability information is stored as
a private data collection and the chaincode controls the access to the vulnerability
using a hash-lock (see Siris et al. [10]). The hash-lock’s secret is generated by the
Security Expert and also known by both the Authority and Vendor and is revealed
as necessary to the interledger, after which the interledger functionality can retrieve
the vulnerability information from the private ledger and write it to the public
ledger. The same secret is then used to lock the functionality Vendor is using for
notifying of fixing the vulnerability, so Vendor’s notification would reveal the secret,
which would then allow the interledger functionality to retrieve and publish the full
vulnerability information.
Automated Responsible Disclosure (ARD) 239
Vulnerability
discovered by Expert
1. Vulnerability reported to authories,
hash, secret, and meta-data
stored in public ledger
Vulnerability verified
by Vendor
3. b No acknowledgement
3.a Acknowledgement
from Vendor
from Vendor stored in public ledger
Vulnerability
fully disclosed
then commits a transaction to public ledger stating that the Authority con-
firms that the previously reported vulnerability (with hash H (V )) is valid.
If the report is a duplicate of a previous report, this is noted in the confirma-
tion and the disclosure will follow the timeline of the original report.
3. The Vendor is notified of the new vulnerability. They now have a fixed time
period (e.g., two weeks) to fetch the details of the vulnerability from the
private ledger and either:
a. confirm and agree to fix the vulnerability by storing their agreement to
the public ledger, in which case they are accorded the grace period of, e.g.,
six months,2 or
b. deny/ignore the vulnerability and do nothing, in which case the Authority
will reveal the secret s 3 to the public ledger and the interledger function-
ality will automatically disclose the vulnerability from the private ledger.
4. If the vendor has agreed to fix the vulnerability, and then
a. releases a fix and stores a corresponding notice on the public ledger to
indicate the vulnerability is now fixed, after which the interledger func-
tionality will publish the full vulnerability information, which in turn
will mark the vulnerability resolved. If the vendor wants its customers to
have some extra time to update their systems, it can first indicate the fix
is upcoming, and then after some time period marks the vulnerability as
fixed. If, however the grace period runs out with no fix,
b. the interledger functionality automatically discloses the vulnerability as
above.
5. After the vulnerability has been disclosed, the security expert can claim the
vulnerability report for credit and any bounty offered.
In Steps 1 and 2, the vulnerability report has to identify the Vendor (e.g., with
their DID) and the product (e.g., with a vendor-specific identifier provided by the
vendor). The same vendor and product ids are then included in the notice on the
public ledger, which allows general public to know the number of known but not
yet fixed vulnerabilities per vendor and per product.
In Step 3, if the vendor acknowledges the vulnerability, the vendor signs a proof
of delivery (with their private key, tied to the public DID) containing the times-
tamp, the hash of vulnerability disclosure H(V), and the relevant metadata (affected
product name and version, etc.). This acknowledgement is stored in the publicly
accessible ledger.
2. The Vendor can also mark the vulnerability as a duplicate of the previously reported one, in which case it
will be disclosed simultaneously with the original vulnerability, as mentioned above.
3. If the Authority does not reveal the secret s in a timely manner, the Security Expert can also do the same.
Analysis 241
The periods for acknowledging the vulnerability, after a fix has been released, and
the grace period can be customized by the authority based on country, necessary
security level, etc. For instance, for critical products, the period can be shorter.
Similarly, the bounty system can be customized: it can, e.g., contain a fixed com-
pensation by the authority and an additional compensation by the vendor or by
third parties for select products.
14.5 Analysis
The solution meets the five requirements listed in the beginning of Section 14.3:
To further evaluate the solution, the potential dishonesty of each party has to be
considered.
If the Security Expert reports a non-valid vulnerability. There is the potential risk
that the system is flooded with irrelevant information, e.g., to make the product
appear to contain more vulnerabilities than it actually has. To prevent that, the
Authority vets all reports, which can help prevent the obvious bogus reports, but
non-trivial fakes may pass depending on the level of scrutiny. However, in this
respect, the solution is not worse than existing solutions.
Duplicate reports of the same vulnerability. Valid but duplicate reports (either
intentional or unintentional) of the same vulnerability can make the product appear
of lower quality due to the increased number of reported vulnerabilities. To avoid
this, duplicate reports can be marked by either the Authority or Vendor as duplicates
of a previous report. Thus, they will not affect the number of individual vulnerabil-
ities discovered, but each report will be eventually disclosed using the same timeline
as the original report.
The Authority does not accept valid reports. This is always a potential issue as one
can either have a mechanism to vet the reports to prevent spamming or have guar-
anteed acceptance of reports with a related spam issue. To reduce the risk of valid
reports not being accepted, in addition to reporting the vulnerability to the Author-
ity and optionally to the Vendor, the Security Expert stores the hash of vulnerability
with associated metadata in the public ledger. If the Authority has not accepted a
valid report, the Security Expert can then publicly reveal the vulnerability,4 which
together with the previously stored hash will make it obvious that the Authority
did not handle the reported vulnerability properly.
The Authority does not disclose the vulnerability after the grace period has expired.
In this case, both the Security Expert and general public will know that the vul-
nerability information has not been disclosed properly by the Authority, and the
disclosure process can also be triggered by the Security Expert, who also possesses
the secret needed.
The Vendor denies the existence of a valid vulnerability. In that case, the vulnera-
bility is automatically disclosed thus providing a clear incentive to remain truthful
or risk lose customers’ trust.
The Vendor claims the fix is effective when it’s not. Releasing a fix triggers a disclo-
sure, which would reveal the fix is not effective thus providing a clear incentive to
remain truthful.
4. This can be performed either using the same smart contract used for reporting vulnerability metadata, or
any other channel.
Conclusions and Future Work 243
The Vendor tries to extend the grace period. All time periods are automatically
enforced preventing any extensions.
Furthermore, it is important to note that the actions performed by all par-
ties, including the actions of the entity providing the interledger functionality,
are transparently and immutably recorded on the public ledger. Hence, the trans-
parency is increased and any attempts for misbehavior can be identified in a
non-repudiable manner. The general public will also know the number of found,
confirmed, but unpatched vulnerabilities per Vendor and product. This provides
incentives for Vendors to patch vulnerabilities quickly and create more secure
products.
Acknowledgments
The research reported here has been undertaken in the context of SOFIE (Secure
Open Federation for Internet Everywhere) project, which has received funding
from the European Union’s Horizon 2020 research and innovation program under
grant agreement No. 779984.
References
The rise of the IoT paradigm provides enterprises with significant opportunities
to implement novel systems that improve the automation and intelligence of their
business processes. However, it also comes with a host of new cybersecurity chal-
lenges, which can compromise the ability of organizations to benefit from advanced
IoT capabilities. Most of these challenges stem from the unique characteristics of
IoT systems that differentiate them from other IT systems. Specifically, the cyber-
physical nature of IoT systems obliges organizations to deal with OT (Opera-
tional Technology) and IT (Information Technology) security at the same time.
Furthermore, IoT systems include new types of assets, such as IoT devices and
cyber-physical systems, which introduce novel vulnerabilities and threats that are
not covered by existing security knowledge bases. Likewise, industrial IoT systems
(e.g., digital automation platforms in industrial plants) are increasingly becoming
interconnected, which opens new opportunities for adversaries to attack industrial
platforms. Moreover, organizations that deploy IoT must make sure that they com-
ply with existing and emerging regulations, including general purpose regulations
(such as the GDPR) and sector-specific regulations. In the light of these challenges,
conventional cybersecurity methods must be enhanced with new tools for IoT sys-
tems and services.
In this book, we have introduced a range of novel IoT security systems and tech-
nologies that can cope with the above-listed challenges. The presented IoT secu-
rity systems take advantage of leading-edge digital technologies in order to provide
novel security tools and capabilities that are hardly possible based on conventional
cybersecurity techniques. Prominent examples of such digital technologies include
machine learning and blockchain technologies. Likewise, the novel techniques that
are presented in the book include techniques for IoT data modeling, cybersecu-
rity risk assessment, mapping of technical security measures to GDPR compliance
247
248 Epilogue
• Security data models for risk assessment in IoT systems, as well as for the
configuration of the associated IoT security platforms and mechanisms, and
ambient assisted living.
• Deep learning techniques based on Variational Autoencoders (VAE) for
anomaly detection in IoT applications that comprise intelligent and semi-
autonomous smart objects such as drones and robots.
• Privacy awareness techniques for IoT systems, along with methodologies and
approaches for data protection that boost privacy compliance.
• Policy-based approaches for risk mitigation of IoT networks and devices,
including approaches that leverage machine learning for detecting risks in
real time.
• Architectures and platforms for end-to-end security, privacy, and safety in
Industrial IoT systems, notably platforms that following popular edge/fog
computing architectures for securing IoT systems. The presented platforms
provide end-to-end security solutions, which address all types of IoT assets
from the device up to the services and applications levels.
• Pattern-driven approaches for ensuring Security, Privacy, Dependability and
Interoperability in IoT applications. The interoperability solutions include
semantic interoperability for security data and mechanisms.
• Use of the LINDDUN privacy threat modeling methodology and of the
STRIDE threat model towards evidence-based risk management and data
protection, as a means of facilitating GDPR compliance.
• Approaches for cybersecurity certification, which combine certification with
testing in order to alleviate the limitations of conventional certification tools
for IoT environments (e.g., their inability to cope with dynamic and volatile
environments).
• Device level security techniques, including techniques based on analysis of
the source code and the binaries of the firmware of IoT nodes (e.g., gateways,
sensing devices, smart objects).
• Decentralized (blockchain-based) solutions for boosting data legitimacy, pri-
vacy, and trust in industrial IoT infrastructures.
• Interledger technologies comprising multiple blockchain infrastructures
towards federating different IoT platforms and performing combined risk
assessments, i.e., risk assessments across different IoT platforms.
Epilogue 249
Most of the above-listed technologies and techniques are innovative and hold
the promise to increase the security and resilience of next generation large-scale
IoT systems, including systems that will comprise smart objects. The chapters of
the book present how these security techniques are used to protect IoT systems and
applications in a variety of sectors including, for example, transport, healthcare,
and industry. These applications are in several cases indicative of the future IoT
environments and their security challenges.
The solutions of the book are collectively presenting some of the main char-
acteristics of future IoT security environments. The latter are expected to play an
important role for the rapid adoption and wider use of IoT systems in the scope
of the fourth industrial revolution (Industry 4.0). The wave of Industry 4.0 sys-
tem will comprise large number of IoT devices, including different IoT assets (e.g.,
smart objects, IoT services, next generation networks (e.g., 5G)) in complex and
dynamic configurations. The solutions listed in the book could greatly boost IoT
adoption in the Industry 4.0 era, through lowering barriers associated with security,
privacy and data protection.
The solutions and technologies of the book will play a significant role in address-
ing security, privacy, and data protection in future IoT systems, notably systems that
are characterized by semi-autonomous behaviors, integration between physical and
cybersecurity, expanded use of AI as part of IoT functionalities, as well as richer and
more frequent security information sharing across IoT stakeholders. Even though it
is not easy to predict the future, we can safely assume that IoT security will be a chal-
lenging, yet very exciting journey for security experts, consultants, security research
organizations, IIoT solution providers and many other stakeholders. In this book,
we have provided knowledge and insights about where we stand in this journey. We
also attempted to develop a vision for the future. We really hope you will enjoy the
journey and will appreciate our ideas to help you start it on the right foot.
Index
anomaly detection, xvii, 51–53, 55, 59, 61, 162–166, 168–173, 175, 177, 183,
90, 92, 96–99, 248 235, 247, 248
autoencoders, 55, 57–59, 65, 92, 248 IIC, 10, 18, 26
big data, 21, 208 IISF, 10, 18, 50
CAPEC, 9, 25, 33 interledger, xix, 229, 232–234, 237–241,
CERTs, 9, 37, 175 243, 248
CMDB, 8, 25, 29, 33, 35, 39, 42, 43 Internet of Things (IoT), xv–xx, 1–3,
cognitive packet network, xix, 105, 116, 5–10, 12–14, 17–31, 34, 35, 37–39,
119 41, 42, 46, 47, 49–53, 58, 65, 66,
CPE, 9, 25, 33 69–81, 84, 85, 88, 108, 114, 115,
CTI, 9, 19, 25 121–131, 133, 137, 143, 144, 149,
CVE, 9, 19, 25, 33, 35 156, 157, 161, 164, 167, 163,
CWE, 9, 25, 33 177–184, 186, 189, 192, 196–198,
cyber physical systems (CPS), xvi, 2, 7, 13, 202, 203, 205, 208–218, 220, 221,
18, 35, 129, 145, 209, 210, 247 225, 226, 229, 230, 232, 234–237,
data sets, xvii, 13, 49, 51–55, 57–66, 96, 243, 247–249
131 interoperability, xvi, xviii, 19, 25, 28, 85,
deep learning, xvii, 13, 48, 49, 51, 52, 55, 121–123, 127, 128, 130, 137, 223,
57, 59, 65, 66, 90, 96, 248 232, 248
dependability, xviii, 121, 123, 125, 126, IoT security, xvi, xvii, xix, xx, 9, 124, 183,
128, 130, 135, 137, 248 196
design patterns, 125, 128, 132 JSON, 9, 44, 176, 224
digital twins for security, 247 machine learning, xvi, xviii, xx, 9, 13, 19,
distributed ledgers, xix, 238 24, 49, 51, 53, 88, 90, 92, 99, 105,
GDPR, xvi–xviii, 5, 12–14, 18, 26, 69–79, 109, 117, 119, 247, 248
81, 84, 85, 124, 144, 146, 153, 157, NFV, 123, 136
250
Index 251
252
Contributing Authors
253
254 Contributing Authors