Algorithms in Advanced AI - RNV Jagan Mohan
Algorithms in Advanced AI - RNV Jagan Mohan
Algorithms in Advanced Artificial Intelligence is a collection of papers on emerging issues, challenges, and new methods in
Artificial Intelligence, Machine Learning, Deep Learning, Cloud Computing, Federated Learning, Internet of Things, and
Blockchain technology. The book addresses the growing attention to advanced technologies due to their ability to provide
“paranormal solutions” to problems associated with classical Artificial Intelligence frameworks. AI is used in various subfields,
including learning, perception, and financial decisions. It uses four strategies: Thinking Humanly, Thinking Rationally,
Acting Humanly, and Acting Rationally. The authors address various issues in ICT, including Artificial Intelligence, Machine
Learning, Deep Learning, Data Science, Big Data Analytics, Vision, Internet of Things, Security and Privacy aspects in AI, and
Blockchain and Digital Twin Integrated Applications in AI.
ALGORITHMS IN
ADVANCED ARTIFICIAL INTELLIGENCE
Edtitors
Dr. R. N. V. Jagan Mohan
Dr. Vasamsetty Chandra Sekhar
Dr. V. M. N. S. S. V. K. R. Gupta
First edition published 2024
by CRC Press
4 Park Square, Milton Park, Abingdon, Oxon, OX14 4RN
© 2024 selection and editorial matter, Dr. R. N. V. Jagan Mohan, Dr. Vasamsetty Chandra Sekhar and Dr. V. M. N. S. S. V. K.
R. Gupta; individual chapters, the contributors
The right Dr. R. N. V. Jagan Mohan, Dr. Vasamsetty Chandra Sekhar and Dr. V. M. N. S. S. V. K. R. Gupta to be identified as the
authors of the editorial material, and of the authors for their individual chapters, has been asserted in accordance with sections
77 and 78 of the Copyright, Designs and Patents Act 1988.
All rights reserved. No part of this book may be reprinted or reproduced or utilised in any form or by any electronic, mechanical,
or other means, now known or hereafter invented, including photocopying and recording, or in any information storage or
retrieval system, without permission in writing from the publishers.
For permission to photocopy or use material electronically from this work, access www.copyright.com or contact the Copyright
Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA 01923, 978-750-8400. For works that are not available on
CCC please contact mpkbookspermissions@tandf.co.uk
Trademark notice: Product or corporate names may be trademarks or registered trademarks, and are used only for identification
and explanation without intent to infringe.
DOI: 10.1201/9781003529231
Contents
1. Convolutional Neural Networks Detect Alzheimer’s disease by Analyzing Facial Expressions and
Eye Movements 1
S. V. Swamy Kadali, R. N. V. Jagan Mohan and Lakshmi M.
2. Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 8
Gidla Sudheer Babu, A. V. S. S. Varma, B. V. Ramana and Srilali Siragam
3. Segment Anything: GPT-3 and Logistic Regression Approach for Skin Cancer Detection 20
V. N. V. Sri Harsha, S. Rao Chintalapudi and V. S. Manoj Kumar Chenna
4. Enhancing Metric Learning Reliability for Pose-Oriented Face Recognition by Visual
Assessment of Tendency 25
Pinisetty Rajasekhar and V. Ravindranath
5. Verifiable Secure Vehicle Connectivity Using Machine Learning Framework for Internet of Vehicles 30
Lanka Divya, Priyadarshini Voosala, R. Shiva Shankar, Ch. Ravi Swaroop
6. Disease Detection In Dental Patients Using Machine Learning Algorithms Through Image Analysis 36
Khadar Alisha Sheik and V. Kiran Kumar
7. Early Disease Diagnosis in Tomato Crops Using AI-Based Deep CNN 44
T. V. K. P. Prasad, V Dilip Kumar, T. Srinivasa Rao, Gude Sujatha and T. K. Priyanka
8. Improvement Over K-Means Algorithm Over Complex Data 49
D. D. D. Suribabu, T. Hitendra Sarma and B. Eswara Reddy
9. Visual Representation of Lung Cancer Image Classification Using Artificial Neural Network 54
B. Nandana Kumar, K. Surya Ram Prasad and G. V. Satya Sriram
10. Machine Learning Improve Predictive Analysis of Diabetes Disease 59
K. Durga Bhavani, CH. Vinod Varma and B. Mounika
11. Tackle Comorbid Obesity in T2DM by Applying New Strategies to Optimize Glycaemic Control and
Weight Management 64
Yugandhar Bokka, R. N. V. Jagan Mohan and M. Chandra Naik
12. A Literature Survey on Deep Learning Approach Used for Audio-to-Sign Conversion with
Gesture Recognition for the Deaf and Dumb 68
B. Veerendra and D. Ramakrishna
13. Federated Learning Approach Based on the MFCC for Speech Emotion Recognition 77
Banda SNV Ramana Murthy and Veluri Ravi Kishore
14. Automated Object Recognition with IoT for Visually Impaired Users 82
JMSV Ravi Kumar, M. Babu Reddy, M. Srikanth and D. Ratna Giri
15. Deep Learning Approach for Early Detection and Diagnosis of Teenager Interstitial Lung Disease 88
Ramesh Alladi, R. N. V. Jagan Mohan and K. V. Ramana
16. Robust Object Detection in Medical Imaging: Cross-Measure Refinement with Edge Detection and SSD 94
Bhanurangarao M. and Mahaveerakannan R.
17. AI-Based Breast Cancer X-Ray Image Detection Using Generative Adversarial Attacks 104
V. S. R. K. Raju Dandu, R. N. V. Jagan Mohan and M. Chandra Naik
vi Algorithms in Advanced Artificial Intelligence
57. Enhancing Network Security: Deep Ensemble-Based Attack Detection Framework 382
R. Tamilkodi, S. Ratalu, Gandham Santoshi, Vysyaraju Sarath Raju, Allampalli V M Mukesh Rao, and
Rampa Aditya Raghava Koundinya
58. Early-Stage Chronic Kidney Disease Detection using Machine Learning with Bigdata 388
Mamatha B and Sujatha P Terdal
59. An MDB-KMC and Firefly-Based Clustering Approach for Energy Optimization in
Wireless Sensor Networks 396
Veeraiah T., Sudhamsu Mouli and M. P. Singh
60. Software Requirements Based Software Effort Estimation using RSLU-GNL-GRU in
Software Project Management 401
K. Harish Kumar and K. Srinivas
61. The Evolution and Impact of Large Language Models in Artificial Intelligence 410
Chaitanya K. and Krishna Jayanth Rolla
62. Several Machine Learning Techniques Used to Forecast Parkinson Disease 418
O. Sri Nagesh, B. Rajarao and Voore Subrahmanyam
63. Fungal Disease Risk Assessment using Data-Driven Methods: Impacts on Food Security and
Crop Devastation 426
Kamidi Jeswanth Kumar
64. Redefining Glaucoma Identification using State-of-the- Art Machine Learning 431
D. Ratna Giri, P. Syamala Rao, J. V. Rama Kumar and JMSV Ravi Kumar
65. Probe Method: A Dependable Economy Data Methodology Feature Selection for Machine Learning 438
Chiranjeevi S. P. Rao Kandula and Srinivas Rao Parnadi
66. Estimating Human Life Expectancy through Sentiment Analysis, Population-based Optimisation,
and Machine Learning Models 444
Meduri Raghu Chandra, G. Jaya Raju, Lanka Atri Datta Ravi Tez and K.Lakshmaji
67. A Distributed-Back Propagation Procedure that uses Climate while Predicting the Spread of
Mosquitoes Using Least Squares Estimation 452
K. Gopala Varma, M. Chandra Naik and R. N. V. Jagan Mohan
68. Unveiling the Efficacy of Machine Learning in Addressing Imbalances in Credit Card
Fraud Detection Data 456
Ch Siva Subrahmanyam, N. Deshai, K. Samatha and J. Tulasi Rajesh
69. Blockchain-driven Security Paradigm: A Robust System Harnessing the Internet of
Medical Things (IoMT) Network for Enhanced E-Healthcare Monitoring 462
Tulasi Rajesh Jonnapalli, N. Deshai K Samatha and B. V. D. S. Shekar
70. Estimating Foreign Export Volume Using Machine Learning for Big Data Business Analytics 471
Yendrapati Geetha
71. Unmasking Deceit: Pioneering Deep Learning Hybrids to Expose Fabricated Reviews in the
Digital Realm 477
N. Deshai and B. Bhaskara Rao
72. YOLO CNN Approach for Object Detection 481
Aluri Dev Ananth, Abhiram Seemakurthi, Sasank Tumma and Prasanthi Boyapati
73. Multi-Crop Analysis Using Multi-Regression via AI-based Federated Learning 487
Mouna Penmetsa and R.N.V. Jagan Mohan
74. Empowering Inclusive Communication: Advancements in Wearable Technology with
GloSign—A Glove-Based Solution for Seamless Sign Language Interaction 492
L V Srinivas, R. Shiva Shankar, N. Deshai, K. Sravani and V. Maheswararao
Contents ix
75. AI-Based Voice Assistant Application for B5G and 6G Free Space Optic Technology is
Competent of Detecting Fake Words 502
R. N. V. Jagan Mohan and Vasamsetty Chandra Sekhar
76. GenerativeAI in Personal Dairy Information Retrieval for Criminal Investigation 507
KVSS Murthy, J. Rajanikanth, R. Shiva Shankar, CH. Ravi Swaroop and D. Ravibabu
77. PCACSO Feature Selection for Prediction of Breast Cancer NAC Response 514
Susmitha Uddaraju, G. P. Saradhi Varma and I.Hemalatha
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
List of Figures
1.1 Facial expressions of alzheimer detection using diffusion models in machine learning 2
1.2 Facial and eye expression of alzheimer patients 3
1.3 Face database of images 4
1.4 The average facial image across all photos, excluding the test photo 4
1.5 Shows the query face (test face) and matching face from the database, together with h the number of
CNN that were used to make the best classification. Convolutional Neural Network is referred to as CNN 4
1.6 The multislice view shows the activation of the hippocampus region in Alzheimer’s disease (AD) stages,
while MCI refers to mild cognitive impairment 5
1.7 The red area depicts the hippocampus in coronal, axial, and sagittal views, aiding in the classification of
Alzheimer’s disease stages 5
1.8 The Montreal Neurological Institute (MCI) coordinate system has identified voxels with the distinguished
stage of Alzheimer’s disease 5
1.9 The proposed CNN Algorithm recognizes red, green, and blue-colored voxels in a brain cutout, indicating
different stages of Alzheimer’s disease (AD), mild cognitive impairment (MCI), and AD 5
1.10 Displays nearly 10 stages of CNN used for the optimal classification of Alzheimer’s disease stages, where
CNN stands for Convolutional Neural Network 5
1.11 Graph for comparing algorithms RGB, YcbCr and CNN 6
2.1 Flow diagram of the self-caring autonomous medicinal and aromatic plants (MAP) nursery using arduino
microcontroller 9
2.2 Block diagram of the Arduino microcontroller 11
2.3 Five plants: (a) lemon grass, (b) basil, (c) aloe vera,
(d) rosemary, and (e) ashwagandha 11
2.4 Circuit diagram of the self-caring autonomous medicinal and aromatic plants nursery using
Arduino Microcontroller 15
2.5 Actuator connection 15
2.6 Flow chart of the methodology 15
2.7 The suggested system’s screen captures of the serial monitor of the Arduino Microcontroller 18
3.1 Disease segment anything: The GPT-3 of computer vision 22
3.2 Relating algorithms RGB, YcbCr and segment anything: GPT-3 23
4.1 Pose oriented cluster images 27
4.2 Face recognition process 28
4.3 Performance in DCT of pose oriented images i.e., clockwise and anti-clockwise 29
5.1 The route is divided into segments connected to SMUs, where vehicle characteristics are transmitted via
Wi-Fi, and data is sent to the cloud for further computation 31
5.2 Illustrates the creation of an SMU using BBN. 32
5.3 Results 34
xii Algorithms in Advanced Artificial Intelligence
List of Tables
Dr. R. N. V. Jagan Mohan working as Professor in Computer Science and Engineering Department from Sagi Rama Krishnam
Raju Engineering College, China Amiram, Bhimavaram. I have Ph.D completed from Acharya Nagarjuna University since 2015
under the esteemed guidance of Dr.Kurra Raja Sekhara Rao, M.Tech in CSE, University College of Engineering, Jawaharlal
Nehru Technological University, 2020. I have published papers around 43 in various international Journals and national journals.
I have published patents around 6 and 1 is Granted international. Published Books in various international publishers 2 and 6
National publishers. I have guidance in Ph.D from J.N.T.U, Kakinada as Supervisor since 2022 to till date. One Research Project
Completed Project on Dissecting Autism Trajectories in Longitudinal Electronic Health Records, collaboratively in India
and Israel, Govt of India, Ministry of Science and Technology, Dept of Science and Technology. DST-SERB Sponsored
International Conference on Algorithms in Advanced Artificial Intelligence, Organized dates at 22nd -24th December 2023, Dept
of CSE, SRKR Engineering College, Bhimavaram-534204. AICTE Sponsored National Conference on Productivity, Quality,
Reliability, Optimization and Computational Modelling, Organized dates at 18th – 20th December 2019, Dept of CSE & IT,
SRKR Engineering College, Bhimavaram-534204. Three Faculty programs organized Webinar on Blockchain Technology:
Insights and Applications,13th August, 2022 at Dept of CSE, SRKR Engineering College, Bhimavaram. Resource Person by
Dr. Hussein El Ghor, Professor in CSE, Lebanese University, Lebanon. Faculty Development Program on Data Science and Its
Application, Dept of CSE, Sponsored by SRKR Engineering College, June 10th – 15th, 2021. National Seminar Symposia DST
SERB Workshop on Machine Learning Evolve Predictive Data Analytics, Dept of IT, SRKR Engineering College, Sanction
Order No: SSY/2017/001121, Sanctioned Date: 13-12-2017, Organized Date:23rd to 28th, July, 2018. Attended many Faculty
Development Programs.
Dr. Vasamsetty Chandra Sekhar PhD is Professor and Head of the Department of Computer Science and Engineering
Department of Sagi Ramakrishnam Raju Engineering College, Andhra Pradesh, India. He has written and co written multiple
articles for IEEE and Elsevier, two peer-reviewed SCI journals for which he has also served as a reviewer. Additionally, he has
taken part in numerous international conferences. Software engineering and machine learning are two of his research interests.
His main area of study is investigating various IoT and software engineering techniques to address a number of difficult issues
in summarization, design, and analysis. Dr.V. Chandra Sekhar received his M.Tech (Computer Science and Technology) and
PhD degrees from Andhra University in Visakhapatnam. He has over 26 research papers, over book chapters, and one patent
published, one authored book published in peer-reviewed publications. Faculty Development Programs were arranged by him.
Software engineering, machine learning, and the Internet of Things are some of his research interests. Vice-Chair, Computer
Society, IEEE Vizag Bay Section.
Dr. V. M. N. S. S. V. K. R. Gupta PhD is Associate Professor of Computer Science and Engineering Department of Sagi
Ramakrishnam Raju Engineering College, Andhra Pradesh, India. He has written and co-written multiple articles for IEEE
and Elsevier, two peer-reviewed SCI journals for which he has also served as a reviewer. Additionally, he has taken part in
numerous international conferences. Data Mining and Healthcare are two of his research interests. His main area of study is
investigating various techniques to address a number of difficult issues in summarization, and analysis. Dr. Gupta received his
M.Tech (Computer Science and Technology) from Andhra University and PhD degrees from K. L. University in Guntur. He
has over 22 research papers, over three book chapters, and three patents published in peer-reviewed publications. He organized
faculty development programs. Among his areas of interest in research is machine learning.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
S. V. Swamy Kadali1
Research Scholar, School of Computing,
SRM Institute of Science and Technology, Kattankulathur, India
R. N. V. Jagan Mohan2
Associate Professor, Department of CSE,
Sagi Rama Krishnam Raju Engineering College(A),
Lakshmi M.3
Professor & HOD, School of Computing, Department of DSBS,
SRM Institute of Science and Technology, Kattankulathur, India
Abstract: The most common form of severe dementia, Alzheimer’s disease (AD), is a cumulative neurological disorder
because of the degradation and death of nerve cells in the brain tissue, intelligence steadily declines and most of its activities
are compromised in AD. Before diving into the level of AD diagnosis, it is essential to highlight the fundamental differences
between conventional machine learning (ML) and deep learning (DL). This work covers a number of photo-preprocessing
approaches that aid in learning because image processing is essential for the diagnosis of AD. The most crucial kind of neural
network for computer vision used in medical image processing is called a Convolutional Neural Network (CNN). The proposed
study will consider facial characteristics, including expressions and eye movements using the diffusion model, as part of CNN’s
meticulous approach to Alzheimer’s diagnosis. Convolutional neural networks were used in an effort to sense Alzheimer’s
disease in its early stages using a big collection of pictures of facial expressions.
Keywords: Alzheimer’s disease, Machine learning, Computer vision, Deep learning, Convolutional neural network etc.
DOI: 10.1201/9781003529231-1
2 Algorithms in Advanced Artificial Intelligence
of areas. Various facial expressions and eye postures are used 2. Proposed Work
to recognize diseases. Computer vision has undergone a
significant shift because of the application of deep learning The Diffusion model process involves adding noise to a face
for Alzheimer recognition. It is widely used to show image, learning to remove it, and then training a machine
computers how to perceive and make judgments similarly learning model to produce a denoise.CNN is one of the most
to humans. There are instances when “computer vision” popular DL classification techniques. The right dataset and
and “machine vision” are used interchangeably. To speed up the following strategies for categorizing AD can be used
image processing, the technique is frequently coupled with Apply transfer learning after using the feature selector to
AI, machine learning, and deep learning. In-depth discussion separate the features, do it fairly, do it with abstract CNN
of the face and eye expression object recognition request of models, examine the output of the two included models, and
deep learning in computer vision will be provided in this then use the Hyper parameters optimizer with one model.
paper. The limitations of standard cognitive tests’ screening Use the CNN method to recognize facial expressions of
procedures may be reduced using eye-tracking-based emotion in pictures. To a network is delivered a pixilated
paradigms which could help in the early diagnosis of AD by image. Filters in the first convolutional layer enable a feature
Alexandra Wolf, 2023[1]. In AD, nerve cells die and brain map process to be applied to each image pixel. This map is
tissue deteriorates, which will significantly reduce the size of subjected to a second layer of filters to create a third map, and
the brain over a period of time and impairs its major functions. so on until the final layer generates the prediction.
Before analyzing the level of AD diagnosis, it is important
to emphasize the significant differences between classical 2.1 Diffusion Model Using Machine Learning
deep learning (DL) rather than machine learning (ML). The The Diffusion model process involves adding noise to an
inclusion variety of photo preprocessing approaches in this face image and learning to remove it, then training a machine
work helps to improve learning because image preparation is learning model to produce a denoise of face image. The process
essential for the diagnosis of AD. of learning the mean matrix involves assuming a normal noise
The following is how the paper is set up: In part II, a succinct distribution and parametrizing the distribution mean and
explanation of CNN’s Alzheimer detection is provided. standard deviation matrix. This can be divided into a forward
Section III covers the findings of the experiment. Then, in and reverse process. Mathematicians often use the jargon
part IV, they came to conclusions, and in section V, references of physical processes to formalize mathematical concepts,
are made. such as diffusion equations for Fick diffusion, heat diffusion,
Fig. 1.1 Facial expressions of alzheimer detection using diffusion models in machine learning
Convolutional Neural Networks Detect Alzheimer’s disease by Analyzing Facial Expressions and Eye Movements 3
and Brownian motion. The Langevin equation, based on the layer, Albawi [3] selects the class (or label) with the highest
Wiener process, is a stochastic formulation with infinitely probabilities.
small steps and normal distribution of time increments. This To perform well on the classification assignment, CNN
intertwines diffusion with white noise generation, making must be able to handle large datasets by Han [7]. In terms
Machine Learning models called diffusion models. The of transfer learning by Dishashree Gupta, 2023 [4], CNN
diffusion models, which utilize a Gaussian prior to generate includes a variety of models were trained on the Image Net
data, form the core of text-to-image generative models. dataset by Sing [13]. The model’s designer can adapt a pre-
trained CNN model’s parameters (i.e., weights) to the new
2.2 Alzheimer Detection by Convolutional
task.
Neural Network Using Facial Features
The Face and Eye Expressions Algorithm Utilizing
The Classification approach is a crucial role, even though
Convolutional Neural Networks for Image Classification
the quantity of the disease object database has a substantial
as follows: Image input a collection of [height, width, and
impact on how well disease objects are identified. Machine
channel] pixel values.
learning includes deep learning. Since the properties are
automatically extracted, deep learning is more effective Feature Extraction
than traditional machine learning techniques. Additionally, 1. To obtain a feature map, use a convolution neural
“end-to-end learning” using deep learning involves giving network.
the network both tasks and unprocessed data. The majority (a) Convergence (RELU).
of the time, advances in Alzheimer’s disease is made in
(i) Choose a kernel whose size is 4x4 and whose
facial features including face and eye expressions using the
depth matches that of the input array.
convolutional neural network technique.
(ii) Convolutional processing is used to get
Features of the facial and eye expression.
(b) Pooling i.e. max pooling.
(i) After applying the dimensionality reduction
procedure to reduce the feature map’s spatial
size, extract the 2x2 image.
2. To extract low-level features from the image, carry out
the steps stated earlier until the fourth layer, altering the
channel size to a value of 16, 32, 64, or 128.
Classification
1. Smooth output is fed to a feed-forward neural network
Fig. 1.2 Facial and eye expression of alzheimer patients with back propagation in each training phase iteration.
2. Using the SoftMax Classification approach, a trained
A CNN is one of the best neural network techniques for
model is utilized to categorize photos like face and
categorization and recognition of images. Figure 1.1 shows
Eye expressions by identifying their dominating
the layers that make up CNN, including the classifier layer,
characteristics.
the pooling layer, the activation layer, the convolution layer,
and more. According to a study from 2020 [2], the activation
function is then used to decide whether or not to excite the 3. Experimental Result
neuron. In order to learn and handle tasks that get harder, it
The CNN has frequently been applied to issues with facial
adapts the information in a nonlinear way. A critical step in the
recognition. The simplicity and speed of the CNN algorithm
process of extracting feature maps by Ebrahimighahnavieh
make it superior to other face-recognition systems. There
is the convolution layer, which bypasses the learnt filter or
were 100 photos of 10 images (ten images of each person)
kernel with a certain size of the input picture 2019 [5]. By
used in the CNN working out. In Fig. 1.3, the face database
Sharma, 2017 [11], activation functions of the sigmoid, Tanh,
is displayed.
and ReLU kinds can be used to make feature maps. The size
is reduced but the most crucial components are retained by To each grayscale image is transformed obsessed by an NxN
pooling the layers. According to Ebrahim, 2020 [6], they pixel vector. As a result, the data matrix used to apply CNN
belong to the downscale group. All neuron from the layer has 100 columns, each of which contains an image. Estimate
above is linked to each neuron from the layer below in a the mean image of face expressions, as illustrated in Fig. 1.2,
completely connected layer. The final phase of the classifier excluding one image for testing.
4 Algorithms in Advanced Artificial Intelligence
Fig. 1.5 Shows the query face (test face) and matching face
from the database, together with h the number of
CNN that were used to make the best classification.
Convolutional Neural Network is referred to as CNN
Fig. 1.10 Displays nearly 10 stages of CNN used for the optimal
classification of Alzheimer’s disease stages, where
CNN stands for Convolutional Neural Network
0.8
YCbCr
0.6
0.4 CNN References
0.2
0 1. Alexandra Wolf, KornkanokTripanpitak, Satoshi Umeda, and
Mihoko Otake-Matsuura. 2023. “Eye-Tracking Paradigms for
the Assessment of Mild Cognitive Impairment: A Systematic
Review.” Frontiers Psychological 14 (June). https://doi.org/
doi.org/10.3389/fpsyg.2023.1197567.
Methods 2. Review of Activation Functions in Neural Networks.
n.d. Accessed https://www.geeksforgeeks.org/activation
Fig. 1.11 Graph for comparing algorithms RGB, YcbCr and functions-neural-networks/.
CNN 3. Albawi S, Mohammed TA, and Al-Zawi S. 2017.
“Understanding of a Convolutional Neural Network.” In
International Conference on Engineering and Technology
4. Discussion (ICET), 1–6. Institute of Electrical and Electronics Engineers.
An approach is made by using CNN algorithm for https://doi.org/10.1109/ICEngTechnol.2017.8308186.
classification of images as per the stages of Alzheimer’s 4. Gupta, Dishashree. 2017. “Transfer Learning and the Art of
Using Pre-Trained Models in Deep Learning,.” Analytics
disease as the algorithm is used for detecting problems
Vidhya. June 1, 2017. https://www.analyticsvidhya.com/
related to recognition of faces. Artificial Neural network is
blog/2017/06/transfer-learning-the-art-of-fine-tuning-a-pre
used in classification and identification of three dimensional trained-model/.
grid of brain. Vector features of image processing is being 5. Ebrahimighahnavieh MA, and Chiong DR. 2019. “Deep
used in this paper. Diagnosis of AD in early stage and Learning to Detect Alzheimer’s Disease from Neuroimaginga
providing proper treatment is challenging as it effects the Systematic Literature Review 105242.” Computer-Methods
brain cells. The hippocampus, a crucial brain region affected And-Programs-In-Biomedicine, November. https://doi.
by Alzheimer’s disease, was analyzed using fMRI data and org/10.1016/j.cmpb.2019.105242.
SPM, achieving a 95% classification rate. The method uses 6. Ebrahim D, Ali-Eldin AM, Moustafa HE, and Arafat H. 2020.
voxels as independent variables and extracts feature vectors “Alzheimer Disease Early Detection Using Convolutional
based on the largest size of discrete cosine transform matrix Neural Networks.” In 15th International Conference on
Computer Engineering and Systems (ICCES), 1–16. Institute
values for Alzheimer’s disease classification. Fig 7 explains
of Electrical and Electronics Engineers. https://doi.org/https://
the stages of AD and face recognition approach used in CNN
doi.org /10.1109/ICCES51560.2020.9334594.
algorithm. This algorithm helps in diagnosis of AD at an 7. Donngmei, Han , Liu Qigang, and Fan Weiguo. 2018. “A New
early stage and provides scope for doctors to understand the Image Classification Method Using CNN Transfer Learning
patient’s condition and provide better treatment. and Web Data Augmentation.” . Expert System Application 95
(April): 43–56. https://doi.org/10.1016/j.eswa.2017.11.028.
5. Conclusion 8. Jinglin Sun, Yu Liu, Hao Wu, Peiguang Jing, and Yong Ji.
2022. “A Novel Deep Learning Approach for Diagnosing
Deep learning, a machine learning technique using Alzheimer’s Disease Based on Eye-Tracking Data.” Front
Convolutional Neural Networks, is trained on large face and Hum Neuroscience 16 (September). https://doi.org/10.3389/
eye expression datasets for identifying Alzheimer disease fnhum.2022.972773.
through intricate, non-linear relationships between inputs 9. Jonathan S. Talahua , Jorge Buele , P. Calvopiña , and José
Varela-Aldás. 2021. Review of Facial Recognition System for
and outputs. It concludes that CNN achieves the most
People with and without Face Mask in Times of the COVID-19
positive results. CNN is proposed for human face recognition
Pandemic. Sustainability 13 (12). https://doi.org/10.3390/
using feature vectors from feature extraction. This method su13126900.
separates stages of Alzheimer’s disease (AD) using fMRI 10. Odusami, Modupe, Rytis Maskeliunas, Robertas
data, using activated hippocampus voxels as input vectors Damasevicius, and Tomas Krilavicius. 2021. “Analysis of
and feature values to cover maximum data variability. The Features of Alzheimer’s Disease: Detection of Early Stage
CNN-based algorithm, based on fMRI data from patients’ from Functional Brain Changes in Magnetic Resonance
hippocampus voxels, offers the best classification rate for Images Using a Finetuned ResNet18 Network”, Diagnostics
11 (6): 1071. https://doi.org/10.3390/diagnostics11061071.
Convolutional Neural Networks Detect Alzheimer’s disease by Analyzing Facial Expressions and Eye Movements 7
11. Gupta, P, N Saxena, M Sharma, and J Tripathi. 2018. Review 14. SP, Singh, Wang L, Gupta S, Goli H, Padmanabhan P, and
of Deep Neural Network for Human Face Recognition. Gulyas B. n.d. Review of 3D Deep Learning on Medical
International Journal Engineering Manufacturing 8 (1): 63– Images: A Review. Sensors 20 (18): 5097. https://doi.
71. 10.5815/ijem.2018. org/10.3390/s20185097,2020.
12. S, Sharma, and Sharma S. 2017. Review of Activation 15. Duan, Y, J Lu, and J Zhou. 2019. Review of Learning Deep
Functions in Neural Networks. Towards Data Science 6 (12): Equidistributed Representation for Face Recognition. IEEE
310–16. Conf. On Computer Vision and Pattern Recognition (CVPR),
13. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis 3415–24.
in Health Care Using Hybrid Model Algorithms”,Journal of
Note: All figures and table in this chapter were designed by the
Artificial Intelligence, Machine Learning and Neural Network
author.
(JAIMLNN), Volume 3, 2023.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
8 Algorithms in Advanced Artificial Intelligence
Abstract: Medicinal and aromatic plants (MAPs) are botanical raw materials, sometimes known as herbal pharmaceuticals,
that are generally utilized as ingredients in cosmetics, health, and medicinal products, as well as other natural health products
for therapeutic, aromatic, and/or culinary purposes. A nursery’s objective is to produce seedlings that are grown in optimum
conditions until they are ready for planting. MAPs make up a considerable component of natural vegetation, and all nurseries
fundamentally try to produce and offer enough high-quality seedlings to suit consumer demand. The primary goal of this
research is to use Arduino to maintain the MAP Nursery. MAPs require highly specific circumstances to flourish. Specific
accurate measurements of temperature, humidity, soil moisture, and sunshine must be kept in the MAP Nursery. Using the
Arduino, we can monitor all of these parameters within the plant to ensure that they are within the needed range for the
healthy growth of MAPs. Numerous MAP species are in high demand for both home use and commercial use in the herbal
sector. Resources for MAPs were abundant in forests, but anthropogenic pressure is rapidly destroying forests. In this article,
we looked at five distinct plant species that flourish under diverse atmospheric conditions: Lemon Grass, Basil, Aloe-Vera,
Rosemary, and Ashwagandha. An Arduino Microcontroller, combined with sensors and actuators, is utilized to keep the
appropriate circumstances for their growth.
Keywords: Medicinal and aromatic plants, Arduino, Nurseries, Microcontroller, Lemon grass
1. Introduction use and commercial use in the herbal sector. Resources for
MAPs were abundant in forests, but anthropogenic pressure is
A nursery’s objective is to produce seedlings that are grown rapidly destroying forests. As a result, growing healing plants
in optimum conditions until they are ready for planting. The in agricultural areas may be a straightforward technique of
basic purpose of all nurseries is to produce and provide enough obtaining progressively basic components. The relevance of
high-quality seedlings to meet client demand [1, 2]. MAPs cultivating medicinal and aromatic plants is fast increasing
constitute a significant percentage of the natural vegetation. due to the detrimental effects of chemical and artificial
Numerous MAP species are in high demand for both home medicines, which are rising awareness among people all
DOI: 10.1201/9781003529231-2
Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 9
of the soil should also be evaluated and, if necessary, soil moisture. They highlight the advantages of real-time
adjusted to ensure that it falls within the suitable range data collecting and remote control, which allow farmers
for the selected species. to make informed decisions and modify environmental
(e) Seed propagation: When the soil is ready, seeds conditions as needed [13]. Lee et al. report an automated
or propagation materials are placed in soil-filled greenhouse control system based on Arduino for temperature,
containers. To enhance germination and root humidity, and soil moisture adjustment. They concentrate on
development, the soil must be kept moist but not soggy. establishing feedback control algorithms in order to maintain
As seedlings grow, they must be checked and relocated optimal growing conditions and achieve energy efficiency.
to larger containers. They demanded that the testing findings illustrate the
(f) Fertilization: Specific nutrients may be required for system’s usefulness in enhancing plant growth and lowering
medicinal plants to develop properly and generate the energy consumption [14]. Gupta et al. present an Arduino
needed medicinal characteristics. Fertilizers can be based smart irrigation system that uses soil moisture sensors,
added to the soil to supply nutrients and keep plants weather data, and real-time feedback to optimize irrigation
healthy. Medical plants, like any other crop, are scheduling. They emphasize the significance of precision
susceptible to pests and illnesses. agriculture approaches in conserving water while increasing
crop yields [15].
(g) Pest and disease control: Pests and illnesses should be
recognized early and treated promptly to reduce the Finally, this study reveals the extensive usage of Arduino
risk of plant damage, such as employing organic pest based devices in plant cultivation to control temperature,
control methods. humidity, and soil moisture. These examples show the
(h) Harvesting: When the plants are ripe and ready to be potential and utility of utilizing Arduino as a platform for
harvested, they can be cut and prepared for use in a constructing smart, automated agricultural solutions. Real-
variety of medical applications [10]. time monitoring, data-driven decision-making, and enhanced
resource management are made possible by the integration
These are only a few of the general procedures for producing of sensors, actuators, and wireless communication. Further
medicinal plants in a nursery. The specific approaches and research and development in this subject have the potential
considerations will differ depending on the plant type and the to improve agricultural practices, increase crop productivity,
nursery’s goals. and contribute to more sustainable farming approaches.
The growth of technology has permitted the creation of Finally, this study reveals the extensive usage of Arduino
smart and automated systems for a variety of uses, including based devices in plant cultivation to control temperature,
agriculture. Maintaining ideal temperature, humidity, and humidity, and soil moisture. These examples show the
soil moisture levels is critical for plant health and growth potential and utility of utilizing Arduino as a platform for
in the field of plant cultivation. Arduino, an open-source constructing smart, automated agricultural solutions. Real-
microcontroller platform, has grown in popularity as a time monitoring, data-driven decision-making, and enhanced
low-cost and versatile alternative for building automated resource management are made possible by the integration
systems. This literature review intends to investigate of sensors, actuators, and wireless communication. Further
existing research and projects that use Arduino to control research and development in this subject have the potential
temperature, humidity, and soil moisture in plant cultivation. to improve agricultural practices, increase crop productivity,
Some of the authors reported numerous plant-related results and contribute to more sustainable farming approaches.
using the Arduino microcontroller. Smith et al. investigate
the use of Arduino-based sensors and actuators to monitor
and control temperature, humidity, and soil moisture in a 2. Proposed Work
greenhouse environment. They show how Arduino may The suggested system is depicted in the Fig. 2.1. An Arduino
be used to maintain ideal growing conditions, resulting board is linked to humidity, temperature, and soil moisture
in increased plant growth and productivity [11]. Johnson sensors, which continuously monitor the conditions of the
et al. demonstrate an Arduino-based automated plant watering MAP nursery. There will be 5 separate sensor sets for 5
system that includes sensors to assess soil moisture levels and different plant groups. Each group must be kept within a
actuators to control water flow. They illustrate the viability certain range of humidity, temperature, and soil moisture
of employing Arduino for efficient irrigation management, levels. When the humidity level drops, the humidifier will
ensuring plants receive appropriate moisture while avoiding use Arduino to supply enough humidity to the plant group.
overwatering [12]. Brown et al. investigate the application When Arduino detects a low humidity level, it activates the
of Arduino-based wireless sensor networks (WSNs) in humidifier. It will turn off the humidifier after the appropriate
agricultural situations to monitor temperature, humidity, and level has been attained. Similarly, if high humidity is detected
Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 11
(a) (b)
deeply watering the plant once or twice a week, cause root rot and other problems, while letting
depending on the weather circumstances. the soil to dry out too much can produce wilting
(iii) During the growing season: Lemongrass leaves and stress in the plant. As a general rule of
should be watered on a regular basis to keep thumb, water basil plants when the top inch of soil
the soil constantly moist. Watering frequency feels dry to the touch. Depending on temperature,
will vary depending on the weather, but it is humidity, and soil type, this could imply watering
normally recommended to water the plant every few days or once a week. To see if the soil is
deeply once a week, or more frequently if the moist enough, use a soil moisture meter or simply
weather is hot and dry. stick your finger up to the second knuckle. If the
(iv) Winter dormancy: Watering should be soil feels damp but not soggy, it is likely to be at the
reduced throughout the winter months when proper moisture level for basil. It is time to water if
the plant is dormant to avoid waterlogging it seems dry. It’s also worth noting that basil plants
the soil. The soil should be allowed to dry out can benefit from a layer of mulch around the base
slightly between waterings, but not fully dry of the plant, such as straw or leaves. This can assist
out. It is crucial to note that the particular soil to keep the soil moist and keep it from drying up
moisture requirements for lemongrass vary too rapidly.
based on climate, soil type, and other growing 3. Aloe Vera: Aloe vera is a succulent plant native to
factors. It’s always a good idea to keep an eye Africa’s hot, dry climates. It can thrive in a variety of
on the soil moisture levels and adjust watering situations and is well-adapted to high temperatures and
as needed to guarantee the plant’s best growth low humidity.
and health. (a) Temperature: Temperatures between 20°C and
2, Basil: Basil (Ocimum basilicum) is a culinary herb that 30°C (68°F and 86°F) are suitable for growing
comes in a variety of varietals with varying growing aloe vera. Temperatures less than 10°C (50°F) or
requirements. Basil, on the other hand, enjoys warm greater than 35°C (95°F) might stress or damage
temperatures and moderate humidity levels. The the plant. Although aloe vera may endure limited
following are some common temperature and humidity periods of cold temperatures, it is preferable to
ranges for basil cultivation: avoid exposing the plant to cold temperatures for
(a) Temperature: Basil thrives in warm climates, extended periods of time.
preferring temperatures ranging from 18°C to 27°C (b) Humidity: Aloe vera enjoys low humidity levels
(65°F to 80°F). Basil can withstand temperatures and can endure arid air. Aloe vera grows best
of up to 35°C (95°F), although it may begin to in humidity levels ranging from 30% to 50%.
exhibit indications of stress or limit its growth. However, if the humidity level is too high, it might
Temperatures below 10°C (50°F) can cause plant raise the danger of fungal illnesses and root rot.
damage or death. If the air is excessively dry, the plant may benefit
(b) Humidity: Basil enjoys humidity levels ranging from misting or placing a tray of water near it to
from 40% to 60%. Because high humidity can enhance humidity levels. It’s crucial to remember
promote fungal illnesses, it’s critical to allow that aloe vera is prone to frost damage, so if you’re
adequate air circulation and prevent crowding the growing it outside in a milder climate, bring it
plants. In dry or arid locations, we may need to inside or cover it during frost or freezing weather.
spritz the plants or use a humidifier to enhance (c) Soil Moisture: Soil Moisture: Aloe vera plants
humidity levels. It’s worth noting that different grow best in well-draining soil that is maintained
basil kinds may have slightly varied temperature slightly damp but not soggy. Allowing the soil to
and humidity requirements. Some kinds, for dry between waterings is critical to preventing
example, Thai basil, may prefer slightly higher root rot and other problems. Here are some broad
temperatures and humidity levels. It’s always a suggestions for aloe vera soil moisture ranges:
good idea to do some study on the variety of basil When watering aloe vera, keep the soil evenly
you’re growing to ensure you’re providing optimal moist but not wet. Allow the soil to dry slightly
growing conditions for maximum development between waterings, but not completely dry.
and output. Overwatering is a major concern with aloe vera, so
(c) Soil Moisture: Basil plants demand well-drained avoid watering it too regularly. Aloe vera prefers
soil that is regularly moist. Overwatering can soil moisture levels between 50 and 70% of field
capacity. This means the soil should be damp
Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 13
but not wet. Insert your finger up to the second moist but not waterlogged throughout this
knuckle into the soil to assess the moisture level. period.
If the soil at that depth feels dry, it’s time to water (ii) During the growing season: Once established,
the plant. Aloe vera plants, in general, require less rosemary prefers a dry soil. In general, the
water throughout the winter months when they soil should be allowed to partially dry before
are dormant and more water during the growing being watered again. However, it is critical
season in the spring and summer. It’s also critical not to allow the soil to dry out completely, as
to make sure the soil has sufficient drainage to this might stress the plant and restrict growth
keep water from accumulating around the roots. A and output. During the growing season, the
well-draining soil mix with sand or perlite can aid soil moisture range for growing rosemary is
in drainage. Remember that the appropriate soil approximately 25-50% of the soil’s maximum
moisture range for aloe vera might vary depending water-holding capacity.
on factors such as temperature, humidity, plant (iii) During the winter months: In locations
size, and container size. As a result, it is critical to with cold winters, rosemary may go into
routinely monitor the plant’s soil moisture levels hibernation, requiring less water. The soil
and modify watering as needed to ensure that should be allowed to dry up more than
the soil maintains within the acceptable moisture usual during this period, but not completely.
range. Growing rosemary in the winter requires a
4. Rosemary: Rosemary (Rosmarinus officinalis) is a soil moisture range of 10-25% of the soil’s
perennial herb native to the Mediterranean region. maximum water-holding capacity. It’s crucial
(a) Temperature: It thrives in warm, sunny, and to note that these are only guidelines; the ideal
dry conditions and tolerates a wide variety of soil moisture range for producing rosemary
temperatures and humidity levels. Growing will vary depending on elements including
rosemary requires a temperature range of 15°C to soil type, climate, and growing conditions. It
30°C (59°F to 86°F), with an optimal temperature is critical to routinely monitor soil moisture
of around 20°C to 24°C (68°F to 75°F). levels and adjust watering as needed to
Temperatures less than 10°C (50°F) or greater than promote optimal growth and yield.
35°C (95°F) might stress or damage the plant. 5. Ashwagandha: Ashwagandha (Withania somnifera)
(b) Humidity: In terms of humidity, rosemary favors is a tropical plant native to India that thrives in warm,
lower humidity levels ranging from 30% to 50%. humid climates.
High humidity levels can cause fungal illnesses and (a) Temperature: Ashwagandha grows well in
other problems, therefore it’s critical to promote temperatures ranging from 20°C to 30°C (68°F
proper air circulation and prevent overcrowding to 86°F), with an optimal temperature of around
the plants. If we’re growing rosemary inside or in a 25°C (77°F). Temperatures less than 15°C (59°F)
greenhouse, we may control humidity by opening or greater than 35°C (95°F) might stress or damage
vents or using a dehumidifier if necessary. It should the plant.
also be noted that rosemary is drought-tolerant and (b) Humidity: In terms of humidity, Ashwagandha
favors well-drained soils. Overwatering can cause favors moderate to high humidity levels ranging
root rot and other problems, so water the plant from 50% to 85%. It can, however, survive
thoroughly but seldom, allowing the soil to dry lower humidity levels, if necessary, as long as
somewhat between waterings. it receives enough water and the temperature is
(c) Soil Moisture: Rosemary is a drought-tolerant herb within the proper range. To keep humidity levels
that likes well-drained soil with a modest amount stable, spray the plants on a frequent basis or use
of rainfall. Here are some broad suggestions a humidifier in the growth area. It is important
for growing rosemary in different soil moisture to note that Ashwagandha can be cultivated both
ranges: outside in warm, humid conditions and indoors in
(i) During the establishment phase: It is critical a greenhouse or other controlled environment. If
to maintain the soil continually moist we’re growing Ashwagandha indoors, you might
when planting rosemary until the plant has need to enhance humidity by placing a tray of
established itself. This usually takes around water near the plants or using a humidifier to keep
2-3 months. The soil should be kept uniformly the correct levels.
14 Algorithms in Advanced Artificial Intelligence
liquids or gases. It works by employing an electromagnetic relay switch. By setting and resetting the digital pins, we may
coil to move a plunger or piston, which opens or closes a valve control all of the actuators (Fig. 2.5).
mechanism. A solenoid valve’s key components include:
• Solenoid Coil
• Plunger or Piston
• Valve Mechanism Fig. 2.5 Actuator connection
Solenoid valves come in various types The process begins with the sensors measuring the
• Normally Closed (NC) Solenoid Valve temperature, humidity, and soil moisture levels of the plants.
The sensed values are temporarily saved in separate variables
• Normally Open (NO) Solenoid Valve
in Arduino. If the values are less than the predefined values,
• Direct-acting Solenoid Valve the Relays will be turned off in the programme. If the values
• Pilot-operated Solenoid Valve surpass the restrictions, the relevant Relay attached to the
Solenoid valves are used to control the flow of water in actuator will go ON, allowing the actuator to activate. If the
irrigation systems, among other things. Temperature value reaches the value, for example, the relay
connected to the Cooling fan will turn on, allowing the Cooling
fan to reduce the temperature. If the temperature falls below
4. Methodology the set point, Arduino will activate the relay linked to the heat
The plant species mentioned above, namely Lemon Grass, light. When the temperature reaches the specified level, the
Basil, Aloe-Vera, Rosemary, and Ashwagandha, are grown relay turns off. Similar activities are taken for humidifiers
under various atmospheric circumstances. An Arduino and dehumidifiers to decrease and increase humidity, and
Microcontroller, combined with sensors and actuators, is for the water pump motor to decrease soil moisture value.
utilized to keep the appropriate circumstances for their The sensors update the parameter values every second. This
growth. The circuit diagram (Fig. 2.4) depicts the Arduino procedure is repeated continuously.
Microcontroller, as well as sensors and actuators.
Applications of Arduino for controlling Temperature: (iv) Use a Peltier module: Depending on the direction
Depending on your design and demands, there are a few of the electric current, a Peltier module can be
different ways to raise or lower the temperature in a plant utilized to cool one side and heat the other. To raise
nursery. Here are some general guidelines: the temperature, Arduino can be used to alter the
(a) Adjust the thermostat: If your nursery has one, we can direction of the electric current.
use it to raise or lower the temperature as needed. For (v) Control a heating system: Based on temperature
example, if the temperature is too low, we can activate data from sensors installed in various areas,
the heating system by turning up the thermostat. Arduino can be used to manage a heating system,
(b) Use heating or cooling systems to adjust the such as a central heating system or a space
temperature: Depending on your climate and the time heater. It is critical to remember that increasing
of year, we may need to use heating or cooling systems temperature with Arduino necessitates necessary
to regulate the temperature. For example, in colder safety precautions in order to avoid any harm or
months, we could use a heater or furnace to raise the risks. Furthermore, the specific components and
temperature, while in hotter months, we could use an programming required will vary depending on
air conditioning unit to lower the temperature. the application and requirements. In the proposed
system a colling fan has been incorporated for
(c) Provide shading: If our nursery is in a sunny location,
controlling the high temperature and a heat lamp
we may need to give shade to keep the plants cool. This
is incorporated as heating element.
may entail employing shade cloth or other materials
to decrease the amount of direct sunlight reaching the Applications of Arduino for controlling Humidity: There
plants. are several ways an Arduino could be used to raise humidity
(d) Use ventilation: By moving air and reducing heat levels in a certain location. Here’s a high-level overview of
buildup in specific regions, ventilation can assist one approach:
regulate temperature. To do this, we could employ fans, (a) Gather needed materials: will need an Arduino board,
vents, or other ventilation devices. a humidity sensor, a humidifier, and some wires.
(e) Insulate the nursery: If our nursery is not well-insulated, (b) Connect the humidity sensor: Use wires to connect the
heat can escape in the colder months or infiltrate in the humidity sensor to the Arduino board. To attach the
warm months, making temperature regulation difficult. sensor correctly, make sure to follow the instructions
Insulating the walls, ceiling, and/or floor can help to included with it.
keep the temperature steady. It is critical to continually (c) Write the code: Write a program in Arduino that reads
monitor the temperature and make adjustments as the humidity sensor and sends a signal to the humidifier
needed to ensure that the plants are developing in to turn on or off depending on the current humidity
optimal conditions. Arduino can be used to regulate level. If the humidity is too low, for example, the
a variety of electronic components that can aid in Arduino can activate the humidifier, releasing moisture
temperature growth in a controlled environment. Here into the air.
are some examples of how Arduino can be used to raise (d) Test and adjust: Run the setup and make any necessary
the temperature: changes to the code to verify that the humidifier turns
(i) Use a heating element: To raise the temperature, on and off according to the humidity levels.
Arduino can be programmed to operate a heating (e) Fine-tune: To fine-tune the system, you could add
element such as a resistor or a heating pad. The additional sensors or use more complex code to
heating element can be linked to a relay module, operate the humidifier based on other factors like as
which the Arduino can regulate. temperature or time of day.
(ii) Control a fan: Warm air can be circulated in a room It is crucial to note that there are numerous approaches
or enclosure using a fan. Based on temperature to controlling humidity levels using an Arduino, and the
readings from a sensor, such as a thermistor or a specific strategy will depend on our personal setup and
temperature sensor, Arduino can be used to control demands. Furthermore, if we are unfamiliar with electronics
the speed of a fan. or programming, it may be beneficial to seek advice from a
(iii) Use a heat lamp: A heat lamp can be used to professional or experienced enthusiast.
provide concentrated heat in one location. Based
(a) Humidity-Plant: We can use the following strategies to
on temperature readings from a sensor, Arduino
enhance or decrease humidity in a plant nursery:
can be programmed to control the on/off time of
the heat lamp. (i) Use a humidifier: A humidifier is a device that
raises the humidity in a room. Set your humidifier
Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 17
to the ideal humidity level in your plant nursery. (e) Provide adequate drainage: Proper drainage is
This is especially beneficial during the dry winter essential for preventing soggy soil and root rot. Make
months or in low-humidity environments. sure the pots or planters have drainage holes and use
(ii) Provide appropriate ventilation: good air well-draining soil.
circulation is essential for sustaining healthy plants (f) Reduce watering frequency: If the soil is overly wet,
and can also aid with humidity regulation. Make reduce the frequency of watering. Allow the soil to dry
sure your plant nursery has proper ventilation somewhat before watering again.
to avoid stagnant air and unnecessary moisture (g) Increased air circulation: Improved air circulation can
buildup. help to reduce fungal growth and promote healthy plant
(iii) Use a dehumidifier: If the humidity in your plant growth. Make sure the nursery has appropriate airflow.
nursery is too high, we can remove extra moisture It is critical to frequently evaluate soil moisture levels
from the air with a dehumidifier. This is especially and alter watering and other practices as needed to
vital in humid climates or during the summer maintain the ideal soil moisture level for the plants
months. being cultivated. In the proposed system a pump motor
(iv) Water plants correctly: Overwatering plants can controlled by Arduino is provided to water the plants.
cause surplus moisture in the air and promote A continuous monitoring is done by connecting a soil
fungal growth. Check that you are not overwatering moisture sensor with the Arduino which connects soil
your plants. moisture information every instant.
(v) Use Mulch: Mulch your plants to keep moisture We examined five separate plant species that thrive in a
in the soil and prevent excess moisture from variety of atmospheric circumstances in this article: Lemon
evaporating into the air. Grass, Basil, Aloe-Vera, Rosemary, and Ashwagandha. An
(vi) Use a humidity tray: A humidity tray is a water- Arduino Microcontroller, along with sensors and actuators, is
filled tray that is placed beneath your plants. used to maintain the proper conditions for their growth. The
Water evaporates, forming a humid environment aforementioned plants’ corresponding results were obtained
surrounding your plants. using an Arduino microcontroller, as illustrated in Figure 6
(vii) Group plants together: Plants naturally release (a-d).
moisture into the air through a process known as
transpiration. Planting them in groups can help to 5. Results and Discussion
raise the humidity level in the surrounding region.
We may change the humidity levels in our plant Results of the 5-plants using Arduino controller as shown in
nursery using these ways to provide the best the Fig. 2.7(a-d). The status of parameters can be observed on
growing conditions for our plants. In this proposed serial monitor. These results are the screen captures of serial
system a Humidifier and a Dehumidifier has been monitor of Arduino controller of the proposed system.
used. Figure 2.7(a) show the case for Aloe vera temperature
Applications of Arduino for controlling Soil Moisture: To monitoring when its temperature value raised over the limit
increase or decrease soil moisture in a plant nursery, you can (i.e., 35 o C), the cooling system is activated and the colling
do the following: fan is ON. Figure 2.6(b) shows that after activating cooling
system temperature become normal. It took 15 to 20 seconds
(a) Check the soil moisture level: To determine the present to get the things normal. we can observe it on the time stamp.
moisture level, use a soil moisture meter or insert our
finger into the soil. Figure 2.6(c) shows the monitoring of all the three parameters
Temperature, Humidity and soil Moisture values for all the
(b) Water the plants: If the soil is too dry, thoroughly water
above sited plants. From the Fig. 2.6(d), we can clearly
the plants to increase moisture levels. Water should be
observe the activation of cooling system on the raise of
applied gradually so that the soil can absorb it equally.
temperature for one plant Aloe vera.
Overwatering can result in soggy soil and root rot.
(c) Use mulch: A layer of organic mulch, such as wood
chips or straw, applied to the soil around the plants can 6. Conclusions
help to retain moisture and reduce evaporation. This technique will limit human intervention with the nursery
(d) Increase humidity: Raising the humidity level in the and offer the former more time to focus on successful plant
nursery can aid in the preservation of soil moisture. To growing procedures. Depending on the size of the nursery,
increase the humidity level, use a humidifier or mist the the scope of this system can be expanded. This technology
plants on a regular basis. will significantly reduce the nursery’s upkeep costs. This can
18 Algorithms in Advanced Artificial Intelligence
Fig. 2.7 The suggested system’s screen captures of the serial monitor of the Arduino Microcontroller
Self-caring Autonomous Medicinal and Aromatic Plants (MAP) Nursery Using Arduino Microcontroller 19
be improved by incorporating Internet of Things components 10. Anonymous, (1985). Wealth of India. Raw material, CSIR,
into the system to remotely monitor nursery conditions. We New Delhi. 11.
can use the Arduino to monitor all of these parameters within 11. R. Nandhini, S. Poovizhi, Priyanka jose, R. Ranjitha, S.
the plant to verify that they are within the required range for anila, (2017). Arduino based smart irrigation system using
IoT, 3rd National Conference on Intelligent Information and
MAP growth. Many MAP species are in high demand for
Computing Technologies.
both personal and commercial use in the herbal industry.
12. T. Saha, M. K. H. Jewel, M. N. Mostakim, N. H. Bhuiyan,
M. S. Ali and M. K. Rahman, (2017). Construction and
References Development of an Automated Greenhouse System Using
Arduino Uno, I.J. Information Engineering and Electronic
1. Deshpande D. J., (2005). Commercial cultivation of medicinal Business. 3, 1-8.
and aromatic plants. Himalayan publishing house, Mumbai. 13. Aishwarya Kagalkar, (2017). Smart Irrigation System,
2. Devi PR, Aparna D, Babu MR, Sekhar MR, Sunitha P et al., International Journal of Engineering Research & Technology.
(2017). Status, scope and promotion of medicinal and aromatic 6, 05.
plants in Andhra Pradesh. Journal of Pharmacognosy and 14. Vimal P V, K S Shivaprakasha, (2017). IOT Based Greenhouse
Phytochemical. 6(6):283-289. Environment Monitoring and Controlling System using
3. Farooqi AA, Sreeranu BS., (2001). Cultivation of Medicinal Arduino Platform, International Conference on Intelligent
and aromatic crops. University press (India limited), Computing, Instrumentation and Control Technologies
Hyderabad. (ICICICT).
4. Khare CP., (2007). Indian Medicinal Plants an Illustrated 15. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
Dictionary. Springer Publications, America. in Health Care Using Hybrid Model Algorithms”,Journal of
5. Kirtikar KR, Basu BD., Indian Medicinal Plants. Lalit Mohan Artificial Intelligence, Machine Learning and Neural Network
Basu, Allahabad, 1918;2 (I, VIII). (JAIMLNN), Volume 3, 2023.
6. Kurain A, Sankar AM., (2007). Medicinal Plants. New India 16. Gabriel Villarrubia, Juan F. De Paz, Daniel H. De La Iglesia
Publishing Agency, New Delhi-110088. and Javier Bajo, (2017). Combining Multi-Agent Systems and
7. Sharma R., (2013). Agro techniques of medicinal plants. Wireless Sensor Networks for Monitoring Crop Irrigation,
Daya Publishing House, New Delhi. Sensors. 17, 1775.
8. Anonymous, (2008). Trees of Gujarat. Gujarat Forest
Department, Gandhinagar. Note: All the figures and tables in this chapter were designed by
9. Trivedi PC., (2010). Medicinal Plants: Conservation and the author.
utilization. Edn 2, Aavishkar Publishers, Distributors, Jaipur.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
20 Algorithms in Advanced Artificial Intelligence
V. N. V. Sri Harsha1
Assistant Professor of CSE – AI&ML, CMR Technical Campus, Hyderabad, India
S. Rao Chintalapudi2
Professor of CSE - AI&ML, CMR Technical Campus, Hyderabad, India
V. S. Manoj Kumar Chenna3
Assistant Professor of CSE, CMR Technical Campus, Hyderabad, India
Abstract: The ability to identify and categorize computer vision analysis high dimensional data is crucial for machine learning.
Many Skin Cancer Diagnosis Disease Detection Systems based on artificial intelligence (AI) in Computer Vision Model use
the categorization of illnesses into benign and malignant groups, including the classification of skin cancer. Among the several
categorization methods, logistic regression stands out for its clarity, efficiency, and interpretability. This paper will cover the
theories underlying cancer diagnosis and illness recognition, the Hybrid model of Skin Cancer Detection operation of Segment
Anything: The GPT-3 of Computer Vision and logistic regression, as well as their potential Deep Learning application to the
classification complex associated with skin cancer.
Keywords: Computer vision, Logistic regression, Skin cancer, Segment anything, The GPT-3
DOI: 10.1201/9781003529231-3
Segment Anything: GPT-3 and Logistic Regression Approach for Skin Cancer Detection 21
and other forms of ultraviolet (UV) radiation are the leading drop precipitously after metastasis [9]. Artificial intelligence
causes of skin cancer. Sunlight and artificial UV rays in has the potential to improve skin cancer detection, lowering
tanning booths are the leading causes of skin cancer. New death rates from the condition [10]. Workload can be reduced
melanoma instances are expected to fall by 5.6% in 2023, and skin lesion diagnosis can be improved with the use of
while the mortality rate is expected to fall by 4.4% [6]. AI-based solutions [11, 12]. One form of deep learning AI
Dermatologists have a far better chance of entirely eradicating that can help with melanoma detection and sickness outcome
skin cancer if it is detected early and treated with minimal prediction is convolutional neural networks (CNN) [4, 13]. To
scarring. Before a tumor becomes malignant or grows deeper detect skin cancer, Hasan et al. [13] used a CNN model. After
into the skin, a doctor will often detect it in its precancerous the Dermoscopic pictures have been segmented, the features
stage. Artificial intelligence (AI) has several potential of the damaged skin cells are extracted using a feature-
applications in dermatology. To foretell the characteristics of extracting technique. To classify the collected features, they
future samples and carry out tasks, machine learning (ML), employed a CNN classifier and got an accuracy of 89.5%.
a subfield of AI, employs statistical techniques and models Ningrumet et al. [14] have developed a skin cancer diagnostic
[7]. Despite its importance in the detection of skin cancer, model that has an accuracy of 92.34% using CNN and ANN
dermatology lags behind radiology in the embrace of artificial techniques.
intelligence. As its use, expansion, and development continue
and as new technologies emerge, AI is becoming increasingly 3. Proposed Work
accessible to the general public. The early identification of
skin cancer is aided by AI. For the detection of skin cancer, The suggested initiative includes the detection and diagnosis
for instance, deep convolutional neural networks might be of skin cancer. The segmented hybrid model for skin cancer
employed to assess skin-imaging data [8]. diagnosis: Skin cancer classification problems are resolved
by the GPT-3 of computer vision and logistic regression, as
The essay is structured as follows: The first section serves well as its forthcoming deep learning applications.
as an introduction. In Part 2 of this essay, you’ll find some
relevant material. In Section 3, we discuss the proposed
methods for detecting skin cancer with the Segment Anything 4. Skin Cancer Detection using
Model and for comprehending cancer classification with the Segment Anything Model
Logistic Regression. Section 4 is when everything wraps up
Regular sun exposure is a major risk factor for the development
in the experiment. The end of Section 5 is stated briefly in
of skin cancer, also known as the abnormal growth of skin
Section 6.
cells. This common kind of cancer can, however, appear
even in sun-protected areas of the skin. The three most
2. Related Work common forms of skin cancer are melanomas, squamous
cell carcinomas, and basal cell carcinomas. Lessening one’s
Because of its visual complexity, dermatology is at the
exposure to ultraviolet light can help prevent skin cancer.
forefront of the current AI revolution. The field of study that
Changes in the skin can be used as an early warning system.
seeks to understand the workings of the human brain and how
When skin cancer is detected at an early stage, it is easier to
they might be replicated in machines is known as artificial
treat, and the patient has a greater chance of survival [4].
intelligence, or AI for short. As AI becomes increasingly
prevalent in the scientific community, this section educates The Segment Anything Model (SAM) is new NLP model
readers on the benefits and latest research findings of AI in that generalizes to zero and a few short learning tasks, a
the detection and treatment of skin cancer. Recent studies trend in computer vision and NLP that enhances their field
and significant ML results in a wide variety of tumor advancements. Segmentation models train with images and
areas relevant to dermatology have also been evaluated. masks, relying on data distribution at training time. SAM offers
The incidence of skin cancer, which includes malignant “prompts” for user input, allowing for interactive predictions,
melanoma and non-melanoma skin cancer (NMSC), is on point locations, boundary boxes, masks, and text prompts.
the rise [8]. While doctors with advanced training may spot The model captures point and box prompts, converts them
cancer, the difficulty of gaining access to them has increased to positional embeddings, and learns additional embeddings
the need for automated methods. This has the dual benefit of to distinguish between boxes and points. Mask prompts are
reducing healthcare costs and saving lives. Since melanoma converted to embeddings using a convolutional network.
covers such a broad spectrum of features, it can be difficult to The text prompt is trained using image encodings from the
differentiate benign skin lesions from skin cancers. Melanoma CLIP model, eliminating data labeling. The researchers
has the worst survival rate of all skin cancers. If caught early generated the largest segmentation dataset by adapting mask
enough, surgery is a certain cure. However, survival rates predictions based on prompts, collecting, and retraining the
22 Algorithms in Advanced Artificial Intelligence
model on newly annotated images. The following steps are each input cancer data point into one of two groups. A
shown in Fig. 3.1. supervised learning technique created specifically for these
situations is logistic regression. Logistic regression is built
5. Understanding Cancer on the underlying sigmoid function (or logistic function).
This S-shaped curve may convert any number between 0 and
Classification Using Logistic 1 into a value between those extremes. In mathematics, the
Regression logistic function is represented as follows:
Consider developing a system that can identify benign and 1
S( x ) = (1)
malignant growths based on characteristics such as color, 1 + e z
weight, and structure. The binarization approach divides
Segment Anything: GPT-3 and Logistic Regression Approach for Skin Cancer Detection 23
The input cancer traits and their corresponding weights are 7. Conclusion
combined linearly to form x, where e serves as the base for
the natural logarithm. The result of the logistic function may Skin cancer has been one of the most prevalent forms of cancer
be viewed as the likelihood that a certain data point belongs in the recent decade. Since the skin is the biggest organ in the
to a specific class. The model predicts class 0 if S (x) is close human body, it can be considered that skin cancer is the most
to 0, and class 1 if S (x) is close to 1. prevalent kind of cancer in people. As a result, this work
was started with the objective of detecting and classifying
6. Experimental Result whether a person is having skin cancer or not. Although it
looks simple, a machine learning model: logistic regression
Comparisons are made between the F1-score, accuracy, has a considerable impact on classification problems.
sensitivity, specificity, and F1-score of the GPT-3, YcbCr, Despite the complex nature of neural networks and ensemble
and RGB algorithms. Segment Anything GPT-3 accuracy techniques, logistic regression is still a reliable collaborator
measures are compared to the RGB and YcbCr accuracy in attempting to categorize cancer diagnostic data as you
metrics from the research, and the outcomes are as follows: delve deeper into the field of machine learning. Various deep-
Accuracy: Accuracy evaluates model performance across learning techniques also have been applied to computer-based
all classes, with equal weight for each. It is determined by skin cancer detection in prior studies. This paper applied an
dividing forecasts by the proportion of accurate estimations. NLP model named “Segment Anything”, which is used in
image segmentation tasks for detecting skin cancer. It works
Sensitivity: The sensitivity, also known as the genuine by indicating which areas of an image should be separated
positive rate or recall rate, measures an individual’s ability to and can be used for a range of segmentation tasks without
identify positive events. the requirement for additional training. It is concluded that,
Specificity: How effectively a model can forecast actual among different models, the Segment Anything: GPT-3
negatives in each of the categories that are accessible depends model performed well with a 97 percent accuracy rate. In
on its specificity. the future, the YOLO techniques, a deep learning algorithm
Precision: Precision is determined by dividing the total usedfor object detection can also be applied for skin cancer
number of positively recognized samples by the number of detection using image segmentation.
correctly identified positively identified samples.
F1-Score:The F1 score is a metric used to evaluate the References
performance of two classifiers by averaging their accuracy 1. Bishop, C. M. “Pattern Recognition and Machine Learning”,
and recall. Springer, 60, 1, 78–78, 2006.
2. Géron, A.: Hands-On Machine Learning with Scikit-Learn,
Table 3.1 Graph for Likening algorithms RGB, YcbCr and Keras, and Tensor Flow, O’Reilly Media, 2019.
Segment Anything: GPT-3 3. Kinnor Das, Clay J. Cockerell, Anant Patil, PawełPietkiewicz,
Algorithm Accuracy Sensitivity Specificity Precision Fl-score Mario Giulini, Stephan Grabbe, Mohamad Goldust.“Machine
Learning and Its Application in Skin Cancer”, International
RGB 0.75 0.79 0.79 0.80 0.80
Journal Environ Res Public Health, 18(24): 13409, 2021.
YCbCr 0.89 0.88 0.85 0.84 0.85 DOI: 10.3390/ijerph182413409.
GPT-3 0.97 0.96 0.97 0.97 0.98 4. Reza Ahmadi Mehr and Ali Ameri.„Skin Cancer Detection
Based on Deep Learning”, Journal Biomedical Physic
Engineering, 12(6): 559–568, 2022. DOI: 10.31661/jbpe.
v0i0.2207-1517.
5. Ferlay J., Colombet M., Soerjomataram I., Parkin D.M.,
Pineros M., Znaor A., Bray F. “Cancer statistics for the year
2020: An overview”. International journal of cancer. 2021.
doi: 10.1002/ijc.33588.
6. Skin Cancer Facts & Statistics [Internet]. The Skin Cancer
Foundation. [(accessed on 22 June 2023)]. Available online:
https://www.skincancer.org/skin-cancer-information/skin
cancer-facts.
7. Hastie, T., Tibshirani, R., & Friedman, J.“The Elements of
Statistical Learning: Data Mining, Inference, and Prediction”,
Fig. 3.2 Relating algorithms RGB, YcbCr and segment Springer, 2009.
anything: GPT-3
24 Algorithms in Advanced Artificial Intelligence
8. Apalla Z., Nashan D., Weller R.B., Castellsaque X. Skin 13. Haenssle, H.A.; Fink, C.; Schneiderbauer, R.; Toberer, F.;
cancer: Epidemiology, disease burden, pathophysiology, Buhl, T.; Blum, A.; Kalloo, A.; Hassen, A.B.H.; Thomas,
diagnosis, and therapeutic approaches. Dermatology and L.; Enk, A.; et al. “Man against machine: Diagnostic
therapy, 7((Suppl. 1)): 5–19, 2017. doi: 10.1007/s13555-016 performance of a deep learning convolutional neural network
0165-y. for Dermoscopic melanoma recognition in comparison to 58
9. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis dermatologists”,Annals of oncology, 29, 1836–1842, 2018.
in Health Care Using Hybrid Model Algorithms”,Journal of 14. Hasan, M., Barman, S. D., Islam, S., Reza, A. W. “Skin cancer
Artificial Intelligence, Machine Learning and Neural Network detection using convolutional neural network”, In Proceedings
(JAIMLNN), Volume 3, 2023. of the 2019 5th international conference on computing and
10. Davis L.E., Shalin S.C., Tackett A.J. Current state of melanoma artificial intelligence, 254–258, 2019.
diagnosis and treatment. Cancer Biol. Ther.; 20: 1366–1379, 15. Ningrum DNA, Yuan SP, Kung WM, Wu CC, Tzeng IS, Huang
2019. doi: 10.1080/15384047.2019.1640032. CY, Li JY, Wang YC. “Deep Learning Classifier with Patient’s
11. Jutzi T.B., Krieghoff-Henning E.I., Holland-Letz T., Utikal Metadata of Dermoscopic Images in Malignant Melanoma
J.S., Hauschild A., Schadendorf D., Sondermann W., Fröhling Detection”,Journal of Multidisciplinary Healthcare,14, 877
S., Hekler A., Schmitt M., et al. Artificial Intelligence in Skin 885, 2021. doi: 10.2147/JMDH.S306284. PMID: 33907414;
Cancer Diagnostics: The Patients’ Perspective. Frontiers in PMCID: PMC8071207.
medicine, 2020; 7:233. doi: 10.3389/fmed.2020.00233. 16. Malvehy J., Pellacani G. “Dermoscopic, confocal
12. Sengupta S., Mittal N., Modi M. Improved skin lesions microscopy and other non-invasive tools for the diagnosis
detection using color space and artificial intelligence of non-melanoma skin cancers and other skin conditions”,
techniques. Journal of Dermatological Treatment, 31, 511– ActaDermato-Venereologica, 97((Suppl. 218)): 22–30, 2017,
518, 2020. doi: 10.1080/09546634.2019.1708239. DOI: 10.2340/00015555-2720.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Pinisetty Rajasekhar*
Assistant Professor, Dept of Mathematics, J.N.T.University, Kakinada, A.P., India
V. Ravindranath
Professor, Dept of Mathematics, J.N.T.University, Kakinada, A. P., India
Abstract: A major challenge with face recognition is that it is dependent on pose-oriented considerations (Ziad Hafid, 2001)
[1]. In order to look at the problem with face recognition from many angles, as proposed by Abhishek Sharma in 2012 [2], This
study suggests fuzzy pose-based integral clusters for both clockwise and anticlockwise images. To improve facial recognition,
the authors propose utilizing a technique known as “cluster tendency” to identify groups of people based on their similarities.
Assessing the clustering tendency is a crucial first step in cluster analysis. The Visual Assessment of Tendency (VAT) method
is one useful tool for determining cluster tendency. Using DCT and Metric Learning, the tendency of clustering in relational
or object data may be visually examined by normalising and extracting features from the picture matrix generated by VAT for
person matching. Examining the likelihood of producing a particular set of angle picture data with the same allocation allows
analysts to ascertain the collection’s clustering tendency. When it comes to detecting criminal identities, it also determines how
much spatial unpredictability there is in the posture image data. Also, look at the investigation’s findings to see how trustworthy
face recognition technology is for police work.
Keywords: Discrete cosine transform, Fuzzy cluster analysis, Pose oriented face recognition, Visual assessment tendency,
Metric learning of neighborhood component analysis etc.
DOI: 10.1201/9781003529231-4
26 Algorithms in Advanced Artificial Intelligence
recognition systems. In order to pre-treat photos with faces lighting approaches described by R.N.V. Jagan Mohan et al.
hidden, face hallucination algorithms must first be trained (2012) [9]. The unique angle orientation algorithm proposed
on similar images, both with and without disguise. However, by R.N.V. Jagan Mohan et al. (2014) [10] should be applied
because of fleeting facial emotions, these algorithms may to all face recognition inputs to ensure they are in an upright
not be able to effectively map the whole state of the face and frontal posture before they are compared to the database
(Abhishek Sharma, 2012 [2]).Deep learning-based facial image. In order to compare an angularly oriented input with
recognition technology is the least intrusive and most quick a database image, it must first be rotated to the correct angle
biometric identification method (Xinyi Wang, 2022 [13]). in either a clockwise or anticlockwise direction, as stated by
It is helpful in security applications like criminal detection R.N.V. Jagan Mohan, 2016 [11]. In a similar vein, this will
since it compares acquired photographs to stored face prints put the face in an upright, frontal position. The spinning axis
using image processing and machine learning. According to makes it easy to identify the face in the shot. The face will
Lixiang Li (2020), face recognition is a technology that takes spin anticlockwise if the input image flips from horizontal
images from surveillance footage and classifies them. Face to vertical. The face will also rotate clockwise if the supplied
recognition technology finds applications in face tracking, image goes from vertical to horizontal. If the input picture is
forensic surveillance, criminal detection, and airport security. skewed in any way, we straighten it out using the rotational
Training and database comparisons of previously saved axis before comparing it.
images are required. Unsupervised learning uses unlabeled Rotate the supplied image clockwise or anticlockwise as it is
data to examine a data collection for patterns; clustering oriented at an angle. Use this in a pose-oriented recognition
or cluster analysis is required to find patterns among system. The fuzzy rule can be used to cluster images with an
unlabeled data pieces. Preprocessing images, extracting angle orientation, as shown below.
features, clustering images based on similarity, and selecting
q 90∞
the optimal number of clusters are all necessary. R.N.V.
Jagan Mohan, 2020 [12] states that feature extraction Ú x(q )dq = Ú cost dq (1)
0 0
from mathematical models or trained convolutional neural
networks improves clustering results for a range of photo- q 30∞ 60∞ 90∞
clustering problems. Abhishek Sharma’s study proposes Ú x(q )dq = Ú cost dq + Ú cosq dq + Ú cos q dq (2)
fuzzy pose-based integral clusters for photos taken in both 0 0 30∞ 60∞
0∞
clockwise and anticlockwise directions. Facial recognition is = (sinq )30 60∞ 90∞
0 + (sinq )30∞ + (sin q )60∞
(3)
a hurdle in pose-oriented topics. Cluster tendency is a method
used to determine likely clusters from individual person The outcome, if the ∫x(θ)dθ =∫cosθdθ = Sin θ process is
identifications. When evaluating clustering tendencies in valued, has a range of fuzzy variable values from 0 to 1,
relational or object data, the Visual Assessment of Tendency which is a range of values for the same function. Choose one
(VAT) methodology is employed. This aids in determining of the next three clusters, and then take it. We choose the first
the dependability of angle-oriented image data gathering for cluster.
law enforcement applications and assesses its likelihood to 30∞ 30∞
cluster.The format of the paper is as follows: Part 1 covers Ú x(q ) dq = Ú cos q dq (4)
the introduction. Part 2 of this work provides a brief overview 0 0
of fuzzy cluster analysis for pose-oriented facial photos. When analyzing the second cluster,
Section 3 covers the DCT transformation method for pose- 60∞ 60∞
oriented face images. The use of neighbourhood component Ú x(q ) dq = Ú cos q dq (5)
analysis for face recognition will be covered in Part 4, which 30∞ 30∞
is a machine learning component. Step-by-step procedure for The third cluster is what we take.
pose-oriented face detection in Section 5. Section 6 contains 90∞ 90∞
the outcomes of the experiment. The last one is the conclusion
of Section 7, which is mentioned in Section 8. Ú x(q ) dq = Ú cos q dq (6)
60∞ 60∞
Select the sequentially arranged clusters for identification
2. Fuzzy Pose Oriented Face Image after classifying the angle-oriented clusters.
Cluster Classification
The alignment of the faces in the images should be the 3. Discrete Cosine Transform
determining factor in sorting them. To fix images that don’t The discrete cosine transform (DCT) has been used as a feature
have an angle of 900 degrees, first rotate them to that angle extraction step in various face recognition investigations by
and then use normalisation methods like the geometric and Annadurai et, 2004[3]. Until now, discrete cosine transforms
Enhancing Metric Learning Reliability for Pose-Oriented Face Recognition by Visual Assessment of Tendency 27
have been applied either in a holistic appearance-based the precision of nearest-neighbour classification. The
sense or in a local appearance-based sense while largely approach employs a stochastic variation to directly optimise
discarding the spatial information. During the classification the leave-one-out k-nearest neighbours (KNN) score on the
phase, feed certain neural network types with local DCT training set. Also, it can learn a low-dimensional linear data
coefficients or simulate them statistically. Ahmed, Natarajan, transformation (Goldberger, 2005 [4]), which can be used
and Rao introduced the discrete cosine transform (DCT) by for fast data presentation and classification. They determine
R.N.V.Jagan Mohan, 2012[8]. The DCT have recommended the softmax probability of the Mahalanobis distance by
a number of modifications. decomposing it. M = LTL and define the probability pij that xi
N is the neighbor of yj:
p (2n 1)(k 1)
y(k, 1) = w(k )Â x(n) cos ,
n=1 2N exp( || Lxi Lx j ||22 )
pij = , pii = 0 (9)
where k = 1, º, N (7) Â exp( || Lxi Lxl ||22 )
l πi
Where
The likelihood that xi the stochastic closest neighbours rule
Ï 1 will correctly classify the data is then:
Ô , k =1
Ô N
w( k ) Ì (8) pi = Â pij (10)
Ô 2 j : j π i , y j = yi
ÔÓ , 2£k£N
N The optimization challenge is the search for matrix L that
The two matrices, x and y, have the same dimensions and maximizes the overall likelihood of being correctly classified.
length, N. The deterministic column transformation (DCT) L = argmax  pi (11)
transforms the x-matrix. Since vectors truly run from 1 to N i
and not 0 to N-1, it is not customary to index the series from
n = 0 and k = 0, but rather from 1 and N. We use a discrete 5. Face Normalization Images with an
cosine transform to extract the feature vectors from the input
sequence in line with the previously reported formulas by
Angle Orientation Cluster Tendency
Aman (2012) [2]. Method
If you have an object vector or numerical dissimilarity values
4. Neighborhood Components for every pair in your object set, you can visually assess the
Analysis (NCA) clustering tendency of the object set. Before importing angle-
based photos into any clustering technique’s face recognition
An approach to distance-metric learning, in contrast to the system, it is imperative to confirm that the angle-oriented data
conventional use of the geometrical mean, seeks to improve sets contain meaningful clusters (i.e., non-random structures).
If so, tell me how many clusters there are. This approach faces that are supposedly the same size, orientation, position,
aims to evaluate the viability of the clustering analysis or and illumination after it has been normalised. Acquiring the
clustering tendencies. The propensity for clustering Abhishek attributes used in this comparison involved a transformation
Sharma et al. (2012) [1] examined an angle-oriented face procedure. The Discrete Cosine Transform (DCT), a widely
photo dataset to determine whether any meaningful, non used transformation in this field, is used for feature extraction
random groupings are present in the dataset. The following in several face recognition investigations. We split the input
approaches are examined to see if angle-oriented face images photographs into N × N portions to determine the regions that
can cluster, as suggested by Hui-Fuong Ng et al., 2006 [4]: will be processed locally. The data is transformed into the
• Images using homogeneous n-angles as samples frequency domain using a two-dimensional N × N Discrete
(a1 … an) from Angle Oriented Dataset D. Cosine Transform (DCT). According to Annadurai et al.
(2004), researchers can use statistical operators to compute
• Find the distance using, xi, between each angle oriented
various spatial frequency functions within the blocks
image and its closest neighbor: For each pose ai ∈ D,
and derive a DCT coefficient at the block level [2]. Face
find its nearest neighbor aj; then compute the distance
Recognition: The final stage is face matching. This technique
between ai and aj and denote it as xi = dist(ai, aj).
can recognise a specific input image by matching its feature
Because all of the input images in a clockwise or anti- vector to the feature vectors stored in the database, according
clockwise cluster assessment tendency have the same face to Zhang and Jin (2007) [14]. Several neighbourhood
but various angles, pose-oriented cluster pictures demonstrate component analysis classifiers for metric learning and
that each image is effective. distance learning are used. It is frequently necessary to
compute the averages for each column of the matrix after
6. Process of Face Recognition establishing the distances for the N × N matrix. The input
image and the database image are the same when the overall
System average is negative or zero, as stated by Goldberger et al.
Facial recognition is a challenging pattern recognition issue (2005) [3].
in computer science that aims to identify human faces in three
dimensions from two-dimensional images. The procedure
consists of four steps: face detection, face segmentation,
face alignment, face feature extraction, and face matching
against a database. The first step is to isolate the face from its
background; the second is to correct the image’s alignment
based on aspects including size, stance, and photographic
characteristics. The procedures for facial recognition with
a focus on poses are as follows: Getting a Consistent Face:
The initial stage of any face recognition system is face
normalisation. The face area is located as a first stage in the
face recognition process. The face normalisation technique
measures the difference between the size of the input image (N
x N) and the database picture. In order to fit, the input image
Fig. 4.2 Face recognition process
must be reduced if its size differs from the database image’s.
If the orientation of the selected image needs to be changed
to match the database image, rotate the angled face from 00 7. Experimental Result
to 900 until it aligns with the database image. According to
R.N.V. Jagan Mohan et al. (2012) [8], the rotation might The DCT with Neighborhood Component Analysis
be either clockwise or anti-clockwise, depending on the approaches are evaluated against the MIT and FERET face
chosen posture of the image. A measurement, or collection databases as well as the student population of the University
of measurements, is used in feature extraction to characterise Database in a typical execution setting by making use of the
a feature. Every measurement reveals the measurable quality input photos’ orientations (clockwise). Table 4.1 displays
of an object. Its calculation makes use of several important the percentage level of recognition for the two experimental
object-specific properties. All aspects can be categorised procedures. Neighbour Component Analysis reveals startling
using either high-level or low-level qualities. Extracting developments in DCT dependability. In Fig. 4.3, we can see
low-level traits from the source photographs derives them the results of the two experimental techniques’ recordings at
from high-level qualities. The face can be compared to other various recognition levels. The graph makes it quite evident
Enhancing Metric Learning Reliability for Pose-Oriented Face Recognition by Visual Assessment of Tendency 29
Table 4.1 Performance in DCT of pose oriented images i.e., 6. Hui-Fuang Ng: Pose-Invariant Face Recognition Security
clockwise and anti-clockwise System, Asian Journal of Health and Information Sciences,
2006.
S.No Scale of Pose Performance Performance in
7. Jason Brownlee: How to Perform Face Recognition with
Oriented Images in DCT for DCT for Anti-
VGGFace2 in Keras, Machine Learning Mastery, 2019.
Clockwise Clockwise
8. KH Teoh, RC Ismail, SZM Naziri, R Hussin, MNM Isa and
1 00-900 96.67 93.34 MSSM Basir: Face Recognition and Identification using
Deep Learning Approach, Journal of Physics: Conference
Series 1755(2021) 012006 IOP Publishing doi:10.1088/1742
100 Pose Oriented Images 6596/1755/1/012006, 2021.
9. Lixiang Li, Xiaohui Mu, Siying Li, Haipeng Peng: A Review
95 of Face Recognition Technology, IEEE Access (Volume-8),
Page(s): 139110 –139120, Electronic ISSN: 2169-3536,
90 INSPEC Accession Number: 19974172, DOI: 10.1109/
Performance in DCT for Performance in DCT for ACCESS.2020. 3011028, 2020.
Clockwise Anti-Clockwise 10. R.N.V. Jagan Mohan, R. Subbarao, K.Raja Sekhara Rao:
Similarity of Inference Face Matching on Angle Oriented Face
Fig. 4.3 Performance in DCT of pose oriented images i.e.,
Recognition, Published in Journal of Computer Engineering
clockwise and anti-clockwise
and Intelligent Systems from www.iiste.org, ISSN 2222-1719
(Paper) ISSN 2222-2863 (Online), Vol 3, No.2, 2012.
that the reliability performance of DCT for neighbouring 11. R.N.V. Jagan Mohan and K. Raja Sekhara Rao: Target
component analysis steadily improves with increasing data Inference on Evaluation of Angle Oriented Cluster, Computer
amounts. Science and Information Technology 2(3): 121-125, DOI:
10.13189/csit.2014.020301, Copyright © 2014 Horizon
Research publishing all rights reserved. http://www.hrpub.org,
8. Conclusion and Future Perspective 2014.
Fuzzy technique applications to nested clusters based on 12. R.N.V. Jagan Mohan: Angle Oriented Based Image Analysis
angles, specifically 0–300, 310–600, and 610–900. With Using L-Axial Semi-Circular Model, Published in Asian
Journal of Mathematics and Computer Research, ISSN No:
the aid of the visual assistance tendency model, an angle-
2395-4205(Print), 2395-4213(Online), Vol-10, ISSUE-4,
oriented approach is presented; the testing results show its
Page No-320-331, 2016.
effective implementation, and it can rotate in both clockwise 13. R.N.V.Jagan Mohan: Cluster Optimization Using Fuzzy
and anticlockwise directions. Based on average classifier Rough Images, International Journal of Multimedia and
values, Neighbourhood Component Analysis of Metric Image Processing,ISSN:2042-4647,DOI:10.20533 /
Learning suggests using this approach. Studies have shown ijmip.2042.4647.2020.0062, Impact Factor (IF):5.48,
that the suggested pose-oriented discrete cosine transform Calculated by Infonomics Society’s Indexing Citation Board
(DCT) increases the accuracy of criminal face detection. (ICB), Pages: 505 -510, Volume-10, Issue-1, March-2020.
14. UP Police launch ‘Trinetra’, its AI-powered face recognition
app to catch criminals, The Financial Express. December 27,
References 2018, Retrieved February 14, 2022.
15. Xinyi Wang, Jianteng Peng, Sufang Zhang, Bihui Chen,Yi Wang,
1. Abhishek Sharma and Murad Al Haj and Jonghyun Choi and
Yandong Guo:A Survey of Face Recognition, Computer Vision
Larry S. Davis and David W. Jacobs: Robust pose invariant
and Pattern Recognition,arXiv:2212.13038,2022,https://doi.
face recognition using coupled latent space discriminate
org /10.48550/arXiv.2212.13038, 2022.
analysis, Computer Vision and Image Understanding 116,
16. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
1095–1110, 2012.
in Health Care Using Hybrid Model Algorithms”,Journal of
2. Aman R. Chadha, Pallavi P. Vaidya, M. Mani Roja: Face
Artificial Intelligence, Machine Learning and Neural Network
recognition using discrete cosine transform for global and
(JAIMLNN), Volume 3, 2023.
local features, IEEE Xplore, DOI: 10.1109/ICONRAEeCE.
17. Zhigang Yu, Yunyun Dong, Jihong Cheng, Miaomiao Sun,
2011.6129742, 16 January 2012.
Feng Su: Research on Face Recognition Classification
3. Annadurai, S., and Saradha, A.: Discrete Cosine Transform
based on Improved Google Net, Hindawi Publications,
Based Face Recognition Using Linear Discriminate Analysis,
Volume 2022, Article ID 7192306, https://doi.
Proceedings of International Conference on Intelligent
org/10.1155/2022/7192306,2022.
Knowledge Systems, 2004.
18. Zhang, Jin: Visualization for Information Retrieval, Springer,
4. Crime and Criminal Tracking Network & Systems (CCTNS):
ISBN 978-3-540-75148-9, 2007.
National Crime Records Bureau. Archived from the original
19.Ziad M. Hafed: Face Recognition Using DCT, International
on February 18, 2022, Retrieved February 18, 2022.
Journal of Computer Vision, pp. 167-188, 2001.
5. Goldberger et al.: Neighborhood Components Analysis, NIPS,
2005. Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
30 Algorithms in Advanced Artificial Intelligence
Lanka Divya*, Lanka Divya, Priyadarshini Voosala, R. Shiva Shankar, Ch. Ravi Swaroop
Assistant Professor, Dept. of Computer Science and Engineering,
Sagi Ramakrishnam Raju Engineering College
Abstract: Vehicles and other objects are linked via the Internet of Vehicles (IoV), a system of interconnected road networks
that uses data processing, automated control, communication, and sensors. According to the research, segment-monitoring
units (SMU) might be used to manage traffic and vehicle characteristics. Reduced congestion, lowered speed restrictions,
and accident prevention are all possible because to the SMUs’ ability to regulate traffic and vehicle attributes. Better traffic
management in the IoV may be achieved using real-time vehicular traffic data. Simple and easy route segmentation using the
unsupervised Learning K-Means Approach would be the main focus of the proposed study.
Keywords: Internet of vehicle, Segment monitoring units, Transportation based network, Unsupervised learning K-Means
approach etc.
*l.divya44@gmail.com
DOI: 10.1201/9781003529231-5
Verifiable Secure Vehicle Connectivity Using Machine Learning Framework for Internet of Vehicles 31
deep neural networks and Q-learning. The AI layer consists These SMUs [2] will be able to successfully control traffic
of intelligent systems, cloud computing, and large-scale in the near imminent by minimizing congestion on routes,
data analysis as it pertains to the architecture of the IoT. lowering the speed limits by traffic, and preventing accidents.
Onboard Diagnostic Units (OBU), Road Side Units (RSU) To improve traffic management in the IoV, it is assumed that
and edge servers are used in IoV multimedia communication the suggested systems would learn from real-time traffic data
to facilitate data sharing, inter-vehicle connections, and describing how cars manage time delays and speed limits in
quality-of-service monitoring via sensor nodes. The IoV uses crowded areas. The basic, uncomplicated route segmentation
machine learning technology to provide secure vehicle-to that designates each segment with an SMU could potentially
vehicle communication [10]. function as the primary foundation for the suggested study.
The five portions of this paper’s response can be understood Internet of Vehicle Architecture: The route is divided into
as follows: section 1’s broad introduction. Proposed work segments linked to a SMU. Vehicles approach SMUs via
such as Internet of Vehicle Architecture and Network Wi-Fi, sending their characteristics and data to the cloud for
Establishment, Route Segmentation Using Unsupervised additional controls.
Learning K-Means Approach, and working with SMU i = Forward Direction (East to West)
covered in Section 2. Section 3 deals with experimental
results. Conclusion included in section 4. References are j = Backward Direction (West to East)
included in Section 5. Sj1 = segment 2 in backward direction
Sj1 = segment1 in backward direction
2. Proposed Work Network Establishment: With each vehicle serving as a
Route segmenting is a colloquial phrase for managing random variable with either continuous or discrete features,
traffic on pathways with caution. In the suggested study, IoV Bayesian Belief Networks (BBN) are constructed with
segment-monitoring units (SMU) are implemented. SMU’s cars as nodes and linked to SMUs and cloud edges. Using
responsibility is to control traffic and a range of vehicle a conditional probability to indicate the influence of each
characteristics. vehicle, an edge in an IoV Bayesian network depicts the link
between vehicle characteristics and the SMU.
Vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I)
are the most common IoV communications. The absolute Let IoV includes N1, N2, N3, …. Nn, then the vehicle
minimum condition for attaining inventive ability is high- relationship is determined as P(N1, N2, N3, …. Nn) =
level route segmentation. In any early approach, exhaustive Pb(N1|N1, N2, N3, …, Nn) Pb(N1, N2, N3, …. Nn) [3]
segmentation is relatively infrequent. The present study that = P(N1|N2, N3, … Nn)P(N2|N3, …, Nn) … P(Nn–1|Nn) P(Nn)
demands primary vehicle routing to specified destinations SMU in IoV Bayesian belief network can be depicted as
mentions segmentation in the crowd sources.
given by Divya et al.
Fig. 5.1 The route is divided into segments connected to SMUs, where vehicle characteristics are transmitted via Wi-Fi, and data
is sent to the cloud for further computation
32 Algorithms in Advanced Artificial Intelligence
Where is the mean centre value of the Query ‘i’ is blocked in The Process of Expectation and Maximization Procedure
the cluster k. steps can be defined:
Input: Vehicle parameters
4. Working of SMU xj, j = 1, 2, …, n and i ε{1, 2, …, k} label set.
A Gaussian mixture model with expectation and maximization Make ready:
( )
map reduction is used by each SMUto regulate traffic in the
Internet of Vehicles effectively. q (0 ) = p1(0 ) , º, pk(0 ) , m1(0 ) , º, m k(0 ) , s 1(0 ) , º, s k(0 ) (6)
Regression Statistics
Multiple (R) 0.809161445
2
R 0.654742243
2
Adjusted R 0.482113365
Error 0.006742324
Samples 10
to describe a stage for dispersed measurements used in 4. Gerla, M.; Lee, E.; Pau, G.; Lee, U: Internet of vehicles: From
vehicle administration, with a focus on experiment findings intelligent grid to autonomous cars and vehicular clouds, 2014
throughout time. In the Internet of Vehicles, multiple IEEE World Forum on Internet of Things (WF-IoT) (PDF).
regression is known to show that the timely assign (Y) of pp. 241–246. doi: 10.1109/WF-IoT.2014.6803166, ISBN 978
1-4799-3459-1, S2CID 206866025, 2014.
intelligent automobiles depends not only on the amount of
5. Hamid, Umar Zakir Abdul; et al: Internet of Vehicle (IoV)
delay, such as Full or last (x1), ReTX Count (x2), Hop Count
Applications in Expediting the Implementation of Smart
(x3), etc. Y = f(x1, x2, x3…xk) is the formula for the dependent Highway of Autonomous Vehicle: A Survey, Performability
variable in multiple regression, which is a function of several in Internet of Things, EAI/Springer Innovations in
independent factors. ReTX Count and Hop Count are used Communication and Computing: 137–157, doi: 10.1007/978
to assess the delayed time in milliseconds when an IoV is 3-319-93557-7_9, ISBN 978-3-319-93556-0, S2CID
completed within a given time frame. 69362954, Retrieved 14 January 2022.
6. Khelifi, Adel; Abu Talib, Manar; Nouichi, Douae; Eltawil,
Mohamed Salah: Toward an Efficient Deployment of Open
6. Conclusion and Future Perspective Source Software in the Internet of Vehicles Field, Arabian
In this research we assumed that multiple regression Journal for Science and Engineering. 44 (2019): 8939–8961,
analysis reveals the timely assignment of intelligent cars to doi: 10.1007/s13369-019-03870-2.S2CID 164632020,
Retrieved 27 December 2020.
RSU depends on factors like ReTX Count and Hop Count,
7. Lee, Eun-Kyu; Gerla, Mario; Pau, Giovanni; Lee, Uichin;
which assess the delayed time in milliseconds when an IoV
Lim, Jae-Han: Internet of Vehicles: From intelligent grid to
is completed within a given time frame. IoV technology autonomous cars and vehicular fogs, International Journal of
enables real-time communication between vehiclesand road Distributed Sensor Networks. 12 (9):155014771666550, doi:
sideinfrastructure through infotainment systems, sensors, and 10.1177/1550147716665500, 2016.
GPS. 8. Maglaras, Leandros; Al-Bayatti, Ali; He, Ying; Wagner,
Isabel; Janicke, Helge: Cities Journal of Sensor and Actuator
Networks, 5 (1): 3, Doi: 10.3390/jsan5010003, 2016.
References 9. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
1. Ahmadian, Amir Shayan; Peldszus, Sven; Ramadan, Qusai; in Health Care Using Hybrid Model Algorithms”,Journal of
Jürjens: Model-based privacy and security analysis with Artificial Intelligence, Machine Learning and Neural Network
CARiSMA,Proceedings of the 2017 11th Joint Meeting (JAIMLNN), Volume 3, 2023.
on Foundations of Software Engineering, pp. 989–993, 10. Nahri, Mohamed; Boulmakoul, Azedine; Karim, Lamia;
doi:10.1145/3106237.3122823, ISBN 9781450351058, Lbath, Ahmed (2018): IoV distributed architecture for real-
S2CID:28115555,2017. time traffic data analytics, Procedia Computer Science, 130:
2. Abbas, S., Ahmed, A., Khan, F., Ahmad, S., Do-Hyeun, K., 480–487, doi:10.1016/j.procs.2018.04.055,2018.
& Do-Hyeun, K. (2021). Blockchain-Based Authentication in 11. Sakiz, Fatih; Sen, Sevil: A survey of attacks and detection
Internet of Vehicles: A Survey. Sensors, 21(23), 7927. mechanisms on intelligent transportation systems: VANETs
3. Lanka D, Kandasamy S. An Unsupervised Traffic Modeling and IoV, Ad Hoc Networks, 61: 33–50, doi: 10.1016/j.
Framework in IoV Using Orchestration of Road Slicing. adhoc.2017.03.006,2017.
In Revolutionizing Industrial Automation through the Note: All the figures and table in this chapter were designed by the
Convergence of Artificial Intelligence and the Internet of author.
Things 2023 (pp. 201-212). IGI Global.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
36 Algorithms in Advanced Artificial Intelligence
Abstract: The teeth are the hardest substance to work with inside the human body. The intricacy of the operation experience,
low efficiency, and higher user involvement of current approaches for diagnosing dental issues. Older methods of oral disease
detection were laborious, manual, and required a dentist to examine and assess the illness. We suggest a unique method for
identifying and categorizing dental caries, the most prevalent issue with teeth, in order to allay these worries. Cavities come
in three varieties: while root cavities form on the surface over the roots of your teeth, smooth surface cavities form on the
smooth sides of your teeth. On the chewing surface of your teeth, there are pit and fissure cavities. In order to identify the
dental cavity issues previously described, to identify these disorders, we gathered information from the Vishnu Dental Hospital
in Bhimavaram and subsequently created a dataset of Dental Intra Oral Periapical Radiograph (IOPA) pictures. In this regard,
we employed the YOLO (You Only Look Once) version 3 deep learning model to develop a robotic system that can recognize
and classify dental abnormalities, like by using IOPA images to recognize various cavity problems. Last but not least, the
technology for automatically detecting and classifying dental problems will help with early illness detection and maybe stop
tooth loss.
Keywords: Dental caries, Deep learning, IOPA, Tooth, YOLOV3, Dentistry, Annotation, Augmentation
1
khadar6@gmail.com, 2kirankumar.v@rediffmail.com
DOI: 10.1201/9781003529231-6
Disease Detection In Dental Patients Using Machine Learning Algorithms Through Image Analysis 37
Fig. 6.2 The foundation for feature extraction in YOLO V3 is the Darknet-53 architecture
[5]. In the image, the convolution network looks for patterns. 3. The Proposed Methodology
Based on similarities to previously learnt data, this model
“scores” a region. High-scoring regions receive positive A methodology has been put forth to address the issue of
detections for the class they most closely resemble. This identifying dental caries. The dataset was made up of about
technique works by creating grids out of the image. Given a 100 photos, of which about 80 served as training data and 20
high score value, the bounding box count that will encircle the served as testing data.
object will be predicted by the grid cells. Each bounding box It is an efficient technique for object detection and
is assigned a confidence value that represents the accuracy classification that uses an object detection model. In
of the forecast [6].Only one object can be detected by each accordance with this approach, a single neural network will
bounding box. In order to identify which shapes and sizes be able to forecast the bounding boxes and class probabilities
are the most comparable, to create the bounding box, the for the image in a single evaluation. The complete detection
dimensions of a number of ground truth boxes obtained from pipeline is contained within a single network, enabling end-
the training data are combined. to-end adjustment of the detection performance metric. But
This method’s foundation for feature extraction is what makes because they employ a pipeline execution architecture, the
it viewed as quick. The facility is known as Darknet-53. Two network must be forced to interact with each component
completely linked layers and 24 convolutional layers make up independently. Various methods have been employed to find
this structure. The twenty convolutional layers are joined by items. Training takes longer as a result, and optimization is
a pooling layer and a fully linked layer. An ImageNet dataset more difficult. The YOLO V3 approach uses a neural network
was used to train this base before. Three convolutional layers to create an output vector containing bounding box and class
and one reduction layer make up the layers. The model is probability coordinates from an input image. This method uses
trained using four convolutional layers and two fully linked Darknet-53, It was educated using Imagenet, as depicted in
layers. Forecasting the likelihood of each and the bounding Fig. 6.2. After 53 additional layers are added to the framework
box is done using the final layer. Each layer is activated using to do detection, our 106-layer network is currently supporting
ReLu, and the top layer is activated using linear. the design. The M x N grid cells with a size of N are divided
into the YOLO V3 algorithm from the input image. Every grid
Disease Detection In Dental Patients Using Machine Learning Algorithms Through Image Analysis 39
has the ability to find and recognize the image. Using class in the same grid is connected to an anchor box. Two
labels and class probabilities, all grid cells may then predict forecasts would appear in the same grid, for instance,
the object’s bounding box coordinates. Implementing a 1x1 if the two items are each linked to two anchor boxes.
kernel on a feature map made up of variously sized features For the item, the IoU ratio is determined. The object
at various points across the structure allows for detection. will not be taken into consideration for detection if the
Dimensions of the detection kernel are 1x1x(Bx(5+C)). The result is less than a certain threshold, let’s say 0.2.
letter B stands for the bounding box prediction capability of (f) Suppression that is not maximum: The final stage of
the featured map cells. Five bounding box features and one YOLO v3 is to resolve the issue that occurs when
confidence item make up the number “5”. Last but not least, many bounding boxes are found for the same object.
“C” denotes the quantity of classes. Binary cross-entropy is The goal of non-maximum suppression is to select
utilized to quantify classification loss, while estimations of the best bounding box from those that frequently
object probability and class probability are obtained using overlap. Each bounding box’s IoU is calculated, and
logistic regression. The YOLO V3 commonly converts a the outcome is then evaluated against the threshold. A
picture entered into an output vector, as was already said. lower-than-acceptable IoU in a bounding box results in
The following parameters make up the output vector: its rejection. If all the bounding boxes have IoU ratios
higher than the threshold, the bounding box with the
1. Class probabilities: This shows the possibility that an
highest IoU ratio is taken into account. As done by the
object will be found inside the bounding box and that it
author Zhao et al.[7], performance measures have been
belongs to a Single class.
mentioned in order to evaluate how well the object
2. Values of bounding boxes: The Cartesian position, detection model is doing.
width and height of the bounding boxes are all given.
1. Precision
3. Prediction probability: A probability is used to depict
the various bounding boxes that contain a detectable The percentage of correctly identified objects to all detected
object. objects is how precision is determined. It is symbolized by
The following procedures were used to detect objects using TP
Precision = (1)
the YOLO V3 model: TP + FP
(a) Data collection: 100 dental pictures in all were FP stands for False Positive, whereas TP is for True Positive.
collected, of which 25% were utilized for testing and
2. Specificity
75% were used for training.
(b) Data labeling: Each and every image is labeled using It displays the percentage of accurately determined true
the LabelImg tool. This also generates a ground truth negatives. This suggests that there will be more true negatives,
box for every channel of the image. The whole dataset which might equate to false positives, than true positives,
is produced as a text file at the conclusion of the which were previously thought to be positive. A model with
process. The file contains data about the image id and high specificity will successfully identify the undesirable
bounding box coordinates. consequences.
(c) Feature extraction: The darknet-53 framework was TN
Specificity = (2)
used to identify the key components of the photos and TN + FP
train the model. The training was expected to take 7
TN stands for True Negative.
hours, and almost 2000 iterations were completed. Two
files with the names “yolov3_training.weights” and 3. Sensitivity
“yolov3_test.cfg” were generated at the conclusion of The sensitivity of the model refers to how effectively it can
the training. The main element for carrying out real- predict the positive test cases. By giving an idea of how many
time detection is these files. cases were correctly classified as positive, it evaluates the
(d) Testing object detector: Real-time object detection is model’s performance.
carried out on the test images using the OpenCV library Sensitivity boosts the accuracy of positive event forecasting.
and the files produced during the feature extraction
procedure. TP
Sensitivity = (3)
(e) Anchor boxes: When multiple objects’ midpoints land TP + FN
on the same grid cell,it can be difficult to locate some In this case, FN stands for False Negative.
of them. As a remedy for this problem, each object
40 Algorithms in Advanced Artificial Intelligence
4. Implementation
Fig. 6.8 3.txt(metadata)
Steps for YOLOv3 Custom Object Detection using IOPO
Images:
Disease Detection In Dental Patients Using Machine Learning Algorithms Through Image Analysis 41
• Calculate the bounding box coordinates and dearth of easily accessible annotated medical information,
dimensions based on the network output if the many automated systems for identifying and categorizing
confidence is higher than the threshold. dental problems face significant obstacles. Deep learning
• Save the class ID, confidence, and bounding box is employed in this work to create an automated method
coordinates. that can identify and classify dental abnormalities on IOPA
6. Use non-maximum suppression: - Use ‘cv2.dnn. pictures. The collection comprises panoramic dental pictures
NMSBoxes()’ to remove duplicate and overlapping from several clinics that have dental issues, like cavities.
boxes before applying non- maximum suppression The proposed method has numerous applications for dental
(NMS). treatment using computers and diagnostics and performs
better in terms of accuracy than current state-of-the-art
• After NMS, obtain the selected boxes’ indices.
methods. After training, the YOLOv3 model was evaluated
7. Draw bounding boxes and labels by iterating through using test images, where it performed best and with the
the chosen NMS indexes. highest degree of accuracy [8]. For the purpose of finding
• Get the color, class label, confidence, and bounding dental anomalies, a real-time methodology is proposed.
box coordinates.
• Using ‘cv2.rectangle()’, draw a rectangle around References
the object, and ‘cv2.putText()’, add the class label
and confidence. 1. Oprea, S.; Marinescu, C.; Lita, I.; Jurianu, M.; Visan, D.A.;
8. Display the image: Cioc, I.B. Image processing techniques used for dental X-ray
image analysis. In Proceedings of the 2008 31st International
• Using the function “cv2_imshow (),” which is not Spring Seminar on Electronics Technology, Budapest,
part of OpenCV by default and may need to be Hungary, 7–11 May 2008; pp. 125–129. [Google Scholar]
changed depending on your environment, display Ossowska, A.; Kusiak, A.; Świetlik, D. Artificial Intelligence
the image with bounding boxes and labels. in Dentistry—Narrative Review. Int. J. Environ. Res. Public
9. Do some cleanup. Health 2022, 19, 3449. [Google Scholar] [CrossRef] [PubMed]
• Use the cv2.destroyAllWindows() function to 2. Yu, Y.J. Machine learning for dental image analysis. arXiv
2016, arXiv:1611.09958. [Google Scholar] Tuzoff, D.V.;
close the image display window.
Tuzova, L.N.; Bornstein, M.M.; Krasnov, A.S.; Kharchenko,
M.A.; Nikolenko, S.I.; Sveshnikov, M.M.; Bednenko, G.B.
4. Result Tooth detection and numbering in panoramic radiographs
using convolutional neural networks.
3. Dentomaxillofac. Radiol. 2019, 48, 20180051. [Google
Scholar] [CrossRef] [PubMed]
4. Gavrilescu, R., Zet, C., Foșalău, C., Skoczylas, M., &
Cotovanu, D. (2018, October). Faster R-CNN: an approach to
real-time object detection. In 2018 International Conference
and Exposition on Electrical and Power Engineering (EPE)
(pp. 0165-0168). IEEE.
5. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
in Health Care Using Hybrid Model Algorithms”,Journal of
Artificial Intelligence, Machine Learning and Neural Network
Fig. 6.16 Dental cavities detection using YOLOV3 (JAIMLNN), Volume 3, 2023. Liu, C., Tao, Y., Liang, J., Li,
K., & Chen, Y. (2018, December).
6. Object detection based on YOLO network. In 2018 IEEE
5. Conclusion 4th Information Technology and Mechatronics Engineering
Conference (ITOEC) (pp. 799-803). IEEE. Zhao, L., & Li, S.
By looking at the teeth and gently moving them, a dentist can (2020).
identify potential dental issues. IOPA photos can be used to 7. Object detection algorithm based on improved YOLOv3.
automatically classify dental diseases, which can help doctors Electronics, 9(3), 537. Thanh, M. T. G., Van Toan, N., Ngoc,
make precise diagnoses. Such tooth issues are found using V. T. N., Tra, N. T., Giap, C. N., & Nguyen, D.M. (2022).
panoramic dental radiography. We propose a novel method 8. Deep Learning Application in Dental Caries Detection Using
based on the deep learning model YOLOv3 for detecting and Intraoral Photos Taken by Smartphones. Applied Sciences,
classifying the most typical tooth problems, namely cavities, 12(11), 5504
in order to address the low efficiency, the complexity of the Note: All the figures and tables in this chapter were designed by
experiential operation, and the high level of user intervention the author.
in existing methods of tooth problem detection. Due to the
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
44 Algorithms in Advanced Artificial Intelligence
Abstract: India’s main business is agriculture. Agriculture influences the style of life in rural areas by about 60%. One of the
popular food crops in India is the tomato. Because tomato plants are less susceptible, disease detection becomes crucial. If
proper maintenance is not given, the plant’s productivity declines. AI systems utilize effective image-processing algorithms,
but they encounter challenges such as noise, occlusion, articulation, and scene interpretation. This paper suggests that AI-
based computer vision and machine vision are emerging technologies that can effectively address various issues using various
algorithms and methods. Since they first harm the leaves, the majority of tomato plant illnesses are found in their early stages.
There is always a chance that a leaf disease can be detected early enough to prevent impending loss. In order to identify
diseases, this work uses AI-based computer vision and machine vision algorithms. The separation of damaged areas on leaves
is done using disease-processing image technology, which is used to precisely diagnose illnesses. Computer vision information
helps with sickness symptoms and cures in the experimental result.
Keywords: Artificial Intelligence, Computer Vision and Machine Vision, Early disease detection et
DOI: 10.1201/9781003529231-7
Early Disease Diagnosis in Tomato Crops Using AI-Based Deep CNN 45
leaf spot, and Late blight. Amateur gardeners can recognize fresh feature vectors, features (the processed images data) are
each of these ailments right away because they each exhibit first utilized to identify a huge library of feature vectors with
a unique set of symptoms. AI is already able to create high- existing classifications.
quality images in a matter of seconds; it is possible that one The purpose of identifying the image parameters of a
day it may be able to create hour-long videos in the same distance function that maximizes a specific objective
manner. If you don’t look too closely, artificial intelligence function assessing agreement with the training data is how to
is capable of creating some quite realistic images. The characterize the metric learning problem as an optimization
technological and anatomical accuracy of AI seems to be a problem. We routinely utilize general-purpose metrics, but
persistent problem. According to computer vision, machines they typically fall short of accurately explaining the behavior
can now comprehend images more effectively. The technology of image data. The effectiveness of the learning mechanism
of employing sensors to understand and interpret what they is directly impacted by this. The answer to this issue is to
observe while interacting with objects digitally is the focus modify the measure in light of the situation and the data.
of this artifact. It covers a wide range of topics and has uses However, doing it manually is very unworkable. Therefore,
in agricultural crops, such as machine translation and pattern metric learning is used to satisfy the data geometry. The
recognition for spotting tomato diseases early on. Machine database elements that are semantically connected to a query
learning is one of the most widely used AI techniques for element can be found using the learnt metric. In a supervised
many healthcare organizations and individuals interested in situation, metrics learning can be thought of as a means to
automation (ML) by sehgan, et. al., 2021[10]. This is due to reduce the data dimension. More generally, the created
practitioners’ ability to achieve noteworthy results in a range picture data can be developed into a new embedding space
of domains because of major improvements in data access and then provided to another machine-learning algorithm
and processing power. ML systems are now able to assess using the learnt transformation by by sehgan, s et. al.,
photos in a manner similar to how our brains process visual 2021[10]. The following sections make up the remaining text
information. They are utilized almost everywhere, including of this essay. Section 1 of the article covers the condition’s
Smart technologies, MRI sickness diagnosis, and everything primary symptom, which is represented by a disease image.
in between. The fundamentals of machine learning for image In Section-2 and 3 deals with Naive Bayes and CNN
processing are discussed, along with some of the tools we classification model for trained metric learn in, is discussed.
might employ to develop cutting-edge algorithms for image Section 4 provides Process for DeepCNN-Based Disease
data. ML algorithms need many high-quality data in order Detection in Tomato and Leaf for the early identification of
to learn and make extremely accurate predictions. Because the disease. Section 5 presents experiments and findings. In
of this, we need to make sure the pictures are appropriately Section 6, the paper’s conclusion and outlook are offered.
cropped, labeled, and ready for ML image processing. This
is where Computer Vision (CV), a discipline focusing on Early Prediction Analysis of Tomato Crop: Tomato fruit
how well computers can comprehend image data, comes into and foliage can become infected by early blight. Early blight
play. To produce the ideal dataset for the machine-learning on tomato foliage first manifests as round, erratic, black, or
algorithm, we can use CV to process, load, transform, and brown spots on the plant’s older leaves by Naresh, 2020[7].
perform other operations on pictures. In this paper, we The centers of these lesions grow into a cluster of black
propose that a computer’s perception of an input image as an circles as they spread, giving them an identifiable target
array of pixels depends on the image resolution. The image pattern. Early blight lesions could gradually produce yellow
resolution will determine how the height, width, and depth tissue, which might eventually cause the leaves to die. When
are shown. For instance, a picture of a 6 × 6 × 3 RGB matrix this disease has badly damaged a plant, it may lose all of its
array and a 4 × 4 × 1 matrix for a grayscale image. In order to leaves. The midpoint of these lesions develops into a set of
find and develop a machine- learning approach to categorize dark concentric rings as they enlarge, creating a distinct target
pattern. Early blight lesions may gradually develop yellow
tissue, which may ultimately result in the death of the leaves.
When this disease has badly affected a plant; it may lose all
of its leaves. When fruit is in the juvenile green or red stage,
early blight can infect it through the calyx or stem attachment
and cause distinct target-like lesions that resemble foliar
infections. Early blight defoliation can lower fruit output and
increase the danger of sunscald damage to the fruit.
Tomato Crop Naive Bayes Classification: One of the most
Fig. 7.1 Image data processing workflow using machine basic supervised machine learning algorithms is Naive Bayes.
learning
46 Algorithms in Advanced Artificial Intelligence
It is a Bayes Theorem-based classification approach. It has Table 7.1 Graph for comparing algorithms RGB, YcbCr and
utilized in text-classification for high-dimensional training DeepCNN
dataset by Aravind et al, 2018[1]. All of the input features Algorithms Accuracy Sensitivity Specificity Precision
in the training dataset assumed independent of each other,
RGB 0.75 0.79 0.79 0.80
i.e. Naive Bayes do not correlate them. The presence of one
feature has no bearing on the presence of another. The objects YcbCr 0.89 0.88 0.85 0.84
like Shape, Color for example, all contribute to identifying a CNN 0.90 0.90 0.87 0.85
Leaf and Tomato Vegetable as Plant Agriculture by Basavaiah Deep CNN 0.98 0.98 0.96 0.96
et al, 2020[2]. Because this assumption is false in most real-
life situations, it is referred to as naïve. Method for Detecting Disease in Tomato and Leaf Using
Let us first go over the concept of conditional probability DeepCNN: There are four basic steps to the classifier model’s
and the Bayes theorem that is utilized in it before going on structure. Getting the dataset for the rural Andhra Pradesh
to Naive Bayes. P(A given B) or P(A|B) are examples of villages is the first stage. The disease images from the entire
conditional probabilities, which are the probabilities of one dataset had to be resized in the second stage before being
event based on the existence of another. The formula for split and classified in the third and fourth stages, respectively,
Bayes formula is: using a deep learning convolution neural network with many
layers, including the input layer, convolution layer, batch
P(C | X )P (C ) normalization layer, activation function layer, max pooling
P(C | X ) =
P( X ) layer, fully connected layer, softmax layer, and classification
layer.
Bernoulli Naïve Bayes: When input features are only available
in binary form, this method is employed. It takes into account One of the well-known uses of computer vision is the early
the Bernoulli-distributed random variable X. disease detection on tomato crops. Finding the leaf and
tomato disease in an image by comparing it to an existing
Ï p if X = 1 database is the task. To learn the characteristics of disease
(X) = Ì where 𝑞 = 1 − 𝑝 and 0 < 𝑝 < 1
Óq if X = 0 photos and identify them by Zhang X, 2018[15], we can
Using CNN, the Tomato and Leaf Image Classification utilize deep learning techniques. The first part of the multiple
Algorithm: stage process, which involves finding one or more items in
the input image, comes after the early disease diagnosis of
Input: Array of pixel values [height, width, and channel] the crop by Gnanavel, 2020[4]. What follows is the practice
Feature Extraction: of standardizing an input object so that it is geometrically
1. To obtain a feature map, use a convolution neural compatible with the database is known as object alignment.
network. Feature Extraction is the name of the final strategy. Then,
characteristics that can be used in respect tasks are extracted.
(a) Convolution (ReLu).
The process of feature detection is completed by a database
(i) Choose a kernel with a 5x5 size, the same comparison of the provided features.
depth as the input array.
A deep learning model that offers unified embeddings
(ii) Convolution should be used to obtain the
for the purposes of object recognition, verification, and
tomato and leaf picture features.
(b) (Max Pooling) Pooling.
(i) Using the dimensionality reduction technique,
reduce the spatial size of the feature map to a
size of 2x2 and extract the dominating feature.
2. Continue the method described above until the fourth
layer, changing the channel size to one of 16, 32, 64, or
128 to extract low-level features from the Tomota and
Leaf image.
Classification
A feed-forward neural network with back propagation is
provided flat output during each training iteration.
By identifying the main features in the photos and classifying
them using the SoftMax Classification approach, a trained Fig. 7.2 Process of leaf and tomato disease detection
model is utilized to categorize by Bedi, 2021[3].
Early Disease Diagnosis in Tomato Crops Using AI-Based Deep CNN 47
y
cy
e
on
ity
vit
is beneficial. By dividing the total number of projections by
or
ra
ific
isi
i�
sc
cu
ec
ns
ec
the percentage of accurate estimates, it is calculated.
F1
Ac
Pr
Se
Sp
TP + TN RGB YCbCr
Accuracy =
TP + TN + FP + FN CNN Deep CNN
Fig. 7.4 Graph for comparing algorithms RGB, YcbCr, CNN
and DeepCNN
D. D. D. Suribabu1
Research scholar, Department of CSE, JNTUA College of engineering,
Anathapuram, Andhrapradesh, India
T. Hitendra Sarma2
Department of Information Technology, VASAVI College of Engineering,
Hyderabad, Telangana, India
B. Eswara Reddy3
Professor, Department of CSE, JNTUA College of Engineering,
Ananthapuram, Andhrapradesh, India
Abstract: The modern era has seen a significant increase in data, making it increasingly challenging for humans to comprehend
and process information. That information might have a lot of valuable and potential values that are hidden. Clustering is
recognized as an essential element in data mining, particularly for the analysis of enormous amounts of data. There are several
clustering techniques in the data mining literature, but among the many algorithms, The k-means approach and its adjusted
varieties have gotten a parcel of intrigued within the field of expansive information examination. These days, neural network-
based clustering, K-means variety, fluffy C-means, and probabilistic C-means clustering, collaborative sifting clustering, and its
developments are all well-known approaches for clustering endless sums of information. This survey study’s main objective is
to offer a venue for discussing the various clustering techniques applied to effective large-data clustering. This review examines
over a dozen research publications on effective big data clustering methods, showcasing results using WEKA and KNIME
data mining programs. The paper critically reviews previous studies and explores the advantages of K-Means clustering over
big data analytics in various research areas. This paper provides a comprehensive overview of the shortcomings and issues of
several large information clustering strategies, aiming to assist students in their journey towards improved big data clustering.
Keywords: Big data clustering, WEKA, Data mining, Fuzzy C-means, K-means clustering, Collaborative filtering
DOI: 10.1201/9781003529231-8
50 Algorithms in Advanced Artificial Intelligence
novel idea that offers the chance to view data that is currently C-means clustering, Possibility C-means clustering,
available from an alternative perspective. It is not only fair Progressive clustering, Self-Organizing K-means clustering,
but abundant in knowledge [1]. Collaborative sifting and optimization clustering, and many
“Huge information” refers to the total amount of data that is more are among the various clustering techniques that fall
contracted out to be prepared for a particular framework in under this category.
terms of time and memory usage. A wide range of industries
that handle massive amounts of raw data, including retail,
finance, e-commerce, healthcare, and other sectors, are
inquisitive about enormous information examination. In spite
of the fact that the method of making and translating data
from gigantic information remains a major issue with all
progressed information mining strategies [2]
Clustering is the most effective way to extract information
from enormous amounts of data and present it in a useful
way. Gathering the provided data into a discrete group of Fig. 8.1 Displays the various big data clustering strategies
items based on their separate measurements indicated from
Let’s now go over each and every clustering algorithm in
the homogenous bunch is the main goal of the clustering
depth using the following framework:
approach. The new challenges that come with working
with large amounts of data make applying clustering 2.1 K-Means Clustering Algorithm
techniques to it more complicated. The current proposed
paper’s purpose is to present an overview of effective large- In the world of data mining, this is regarded as one of the
scale information clustering techniques for large-scale best clustering techniques. In order to extract the clusters
information management. This survey was conducted based according to the K value, we first try to consider the value of
on the framework, datasets, and various implementation “K.” Then, centres are designated. The closest center’s value is
technologies used for big data clustering. An extra survey used to allocate and compute each cluster in this instance. The
was conducted to address research gaps and concerns. Thus, stages are repeated until the halt or convergence criteria are
a more advanced and effective big data clustering technique satisfied. The squared blunder, or the cruel contrast between
was created. Section 1 gives a quick introduction to the study; the cluster centers and the things doled out to the clusters, is
Section 2 covers the literature on current large data clustering the key calculate deciding this end basis. The stop criterion
algorithms; and Section 3 provides a brief overview of the will be determined by the iteration steps; depending on the
analysis of various tools and frameworks. Section 4 primarily anticipated “K” value, this criterion may occasionally be
discusses the experimental reports; and Section 5 concludes. high. In order to prepare this suggested thesis, I read a number
of research publications that examined the effectiveness and
significance of the K-Means clustering algorithm: A few of
2. Related Work these are included below:
The following is the definition of the clustering problem: A Sreedhar C. et al. [3] presented a K-Means Hadoop
data collection Y = {Y1, Y2... Ym} is given, and an integer MapReduce (KM-HMR) to successfully cluster the gigantic
value p, information. He mainly presented two methods in the current
The clustering problem is to define a mapping study, like KMHMR, which concentrate on the MapReduce
(MR) framework by employing K-Means. The third tactic
f: Y → {1, ..., p}, to emphasize the importance of cluster quality was to try
Where each item Yl, l ∈{1, ..., m} is assigned to one cluster to increase intra-cluster distances while decreasing inter-
Cp, j = 1, ..., p. cluster distances. Regarding execution time, the results of
A cluster Cj contains the items mapped to it: the suggested KM-HMR approaches have surpassed the
effectiveness of existing clustering techniques.
Kj ={Yl|f(Yl) = Kj, 1, ..., m, and Yl ∈ Y}.
A cluster’s members are more similar to one another than 2.2 Joint Filtering Technique
they are to objects outside of it [6]. Cluster similarity is Here, we attempt to talk about a few writers who primarily
often measured using one of the Euclidean distances. addressed collaborative filtering.
Figure 1 presents the categorization of multiple big data
Rong Hu et al. [4] developed the Clustering-based
clustering techniques. Leader K-means clustering, Fluffy
Collaborative Filtering (ClubCF) technique to encourage
Improvement Over K-Means Algorithm Over Complex Data 51
collaborative services. The true goal of this strategy is to produce the final clusters. SOM is therefore employed in this
produce comparable services inside the same clusters. Again, strategy for both dimensionality reduction and visualisation.
this is broken down into two primary stages: In the first, all of In addition to this, we attempt to regard SOM as a spider
the data sets are broken down into discrete cluster pieces and graph, in which each graph has a significant number of
are attempted to be made appropriate for further processing. concepts that have been examined. The next kind of SOM
Collaborative filtering is used in the second phase to identify is called Growing Hierarchical Self-Organizing Maps, or
the resulting clusters according to the “k” value. The global GHSOMs, and they are also used in high dimensional data
intricacy of CF has decreased in distinction since there are processing. The GHSOM technique [8] is mostly used for
significantly fewer administrations in the cluster than there textual, numerical, web page, and other sorts of information
are total online administrations that are accessible. clustering, among other types of information clustering.
Another well-known author, Subramaniya Swamy V. et al. [5],
2.5 Method of Hierarchical Clustering
created a prediction mechanism based on the Collaborative
Filtering approach for efficient parallel processing of One technique for creating a cluster hierarchy is hierarchical
enormous amounts of data. He developed an MR framework clustering. Generally speaking, hierarchical clustering can be
that is employed in the maintenance, aggregation, and classified into two types:
filtering of effective storage. To refine data, the suggested 1. Agglomerative Clustering: In this method, each data
Collaborative Filtering is applied. point is grouped with its corresponding item, and the
two clusters are then combined.
2.3 Different K-Means Clustering 2. Divisive Clustering: In this method, each individual
Here, we attempt to address certain journalists who basically data item is assigned to a single cluster, after which it is
managed with the Variation of K-Means clustering for recursively divided into smaller groups.
colossal information; these are talked about in more detail The task is to determine which clusters in the divisive
afterward on. case should be separated and which should be merged in
Mohamed Aymen Ben Haj Kacem et al., prominent researchers, the agglomerative case. An interesting degree of disparity
created the Quickened Map Reduce-based K-Prototypes proportion is utilized to these information things from these
(AMRKP) clustering calculation [6] to cluster gigantic two circumstances [9].
volumes of information. Here, attempt to dramatically reduce Ultimately, a dendrogram graph is used to visualize the
the number of operations by limiting the number of reads and results for a clear and tangible understanding [10].
writes that the input and output operations can make on the
provided data. He also attempts to propose a pruning strategy 2.6 C-Means (FCM) Fuzzy Clustering
that would reduce the separation between the data points and
C-Means (FCM) Fuzzy Clustering is a method used in
the cluster centre, accelerating the clustering process. The
machine learning to analyze and classify data in a variety of
proposed AMRKP is compared with all previous clustering
ways. Here, we attempt to explore a few writers that focused
schemes and shows substantial improvement over several
primarily on fuzzy C-Means clustering for large data; their
other approaches in terms of efficiency and scalability.
discussions are covered in detail below.
2.4 Algorithm for Self-Organizing K-Means Simone A. Ludwig [11] made numerous attempts to
Clustering determine the FCM clustering algorithm’s scalability and
parallelization. Initially, the map and reduce function for this
Here, we attempt to discuss a few writers who primarily
FCM clustering algorithm were generated by combining the
addressed the self-organizing k-means clustering technique.
MR framework. To demonstrate the efficacy of the suggested
To overcome huge data challenges, many researchers combine approach in terms of purity function, MR-FCM clustering
modified k-means algorithms with self-organizing maps [7]. algorithm validation was carried out.
These are the main protocols that they follow. Below is a list
In order to cluster large amounts of data, Minyar Sassi Hidri et
of them:
al. [12] presented an expanded version of the FCM clustering
[1], To begin with, the hereditary calculation is utilized to algorithm that integrated the split and merge strategies. The
diminish the instability within the information and distinguish huge data is split using the split method first to create distinct
the primary cluster centers; moment, the SOM is mostly used subsets, which are then randomly sampled to create distinct
to check the number of clusters and reduce the estimation of subsamples. Our method performed well with optimized time
the information; and third, the k-means algorithm is used to and space complexities using the available resources.
52 Algorithms in Advanced Artificial Intelligence
2.7 K-Means Clustering for Leaders The three available data sets are Letter Image Recognition
(LIR), Optical Character Recognition (OCR), and Pendigits.
T. Hitendra Sarma, P. Viswanath, B. Eswara Reddy, and
The UCI Machine Learning Repository has Pendigits and LIR
others created the pioneer k-means clustering strategy as a
data sets that can be used to construct the Hybrid K-Means
crossover way to quicken the k-means clustering prepare.
and Leader K-Means algorithms.
They to begin with considered k-means as a partition-based,
iterative approach that meets to a arrangement in a limited
period of time for limited information sets [13]. 4. Results of Experiments
The authors asserted that the clustering time of an algorithm In this section, we attempt to obtain a huge data set with a
is independent of the information set’s measure. Various higher number of items and use the WEKA tool to compare
strategies exist to enhance the traditional k-means clustering several algorithms before calculating the Data processing
approach, but none of them have been found to be the most time. We attempt to draw the conclusion that, among all the
effective. They show a prototype-based crossover strategy different algorithms, some require less time than others.
for speeding up the k-means clustering calculation in this
Figure 8.2 shows a graph that compares different clustering
work. They isolated the information set into minor groupings,
algorithms over large data sets in terms of both item count
or grouplets, of shifting sizes within the recommended
and processing time.
technique. In addition to I, e Model set of rules speak to
each cluster. The set of models is at that point isolated into
k clusters once more utilizing the altered k-means approach.
The k-means clustering method, like the conventional
method, eliminates empty clusters in the iterative phase. Each
prototype in each new cluster of prototypes is swapped out
for the corresponding collection of patterns (which produced
the grouplets) in order to generate a partition of the data-set.
Since this data-set partition might not match the partition that
the conventional k-means approach obtained throughout the
data-set, a corrective action is advised.
We attempt to address the many technologies that are utilized Table 8.1 Clustering computational time (in Sec.) in Weka tool
to analyses large amounts of data in this section. Let’s now go
Clustering Computational Time (In Sec.) In Weka Tool
over these in more detail:
Clustering Data set items Processing time of
The WEKA framework is a significant tool for extensive algorithm various algorithms
information investigation (http://www.cs.waikato. ac.nz/ml/ in WEKA
weka). This tool supports various clustering techniques such K-Means 260538 53
as X-means, DBSCAN, OPTICS, progressive clustering,
X-Means 260538 69
and essential k-means. The next system is known as KNIME
(http://knime.com).The following clustering methods can EM 260538 989
be used with this: The text discusses various algorithms DBScan 260538 6347
such as hierarchical clustering, fuzzy c-means, K-means, LeaderK-Means 260538 42
and self-organizing tree algorithm (SOTA).The next system
is called RapidMiner, and it allows us to do the following According to the experimental findings, we attempt to
operations: k-means and both of its modifications, X-means conclude that Leader K-means processes the data set items
and k-medoids, more quickly than several outdated.
Several more techniques are used, including DBSCAN, This, when compared to some additional algorithms, is more
EM, and SOM (http://rapidminer.com). Biolab.si’s Orange accurate and efficient in terms of time.
system is used to construct K-means, SOM, and hierarchical
clustering.
Improvement Over K-Means Algorithm Over Complex Data 53
5. Conclusion The amount of data studied and the amount of time needed
to convert the data set into usable information determine
This article discusses a number of algorithms in an effort to which algorithm is optimal. Big data, as we all know, spurs
determine which one is best for clustering large amounts of the creation of new technologies. We attempt to include all
data. The report also analyses a number of big data clustering the many kinds of algorithms that can quickly and effectively
concerns and issues. We went into great detail regarding cluster data into meaningful information and provide reports
each and every clustering algorithm, as well as the tools and in this suggested work.
techniques used to assess how well those algorithms function.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
54 Algorithms in Advanced Artificial Intelligence
Abstract: The visual representation of the image tries to draw emphasis to the Lungs Image’s deficient feature vector is based
on the interest aroused by the display of a flaw in the words and images. To build a successful Google Images Search Engine,
thorough model co-training for illness diagnosis of text and image on spatial features is necessary. In the paper, customized
image rating using machine learning improves user experiences. The study makes the argument that optimizing picture ranking
for individual users can improve user experiences while also requiring specific model co-training for text and image data in a
robust machine like Google Images Search Engine. Artificial neural networks (ANN) enable experts in computer science to
complete challenging tasks including pre-processing, Feature Prediction, and Pattern Recognition. The experimental result is
on Lung Cancer Detection from feature vector in the different types of X-Ray, PET scan, WSI, Images and other types etc, with
the help of distance method.
Keywords: Artificial neural network, Image search engine, Lung cancer detection, Machine learning, Visual representation,
etc.
1
dnrnandan@gmail.com, 2surya.dnrcet@gmail.com, 3sriramgv9@gmail.com
DOI: 10.1201/9781003529231-9
Visual Representation of Lung Cancer Image Classification Using Artificial Neural Network 55
an effort to mimic how the human brain and eyes function. The Networks or ANN has applications in chatbots, which are
algorithms will use some pixel patterns that the machine has frequently used for image or text classification. Neural
previously seen a lot. Nithyashree V., 2022[6], must follow networks normally consist of an input layer, an output layer,
a thorough process while creating an image classifier. A big and a hidden layer, which are made up of components that
image object records information about a large image file translate the input into something that the output layer can
and the image information it contains. Big images are used use. Because of their complexity or sheer quantity, they are
to represent images as tiny data units that may be imported useful for seeing patterns that a human programmer could
and examined independently. Use a big image object to never extract and teach the computer to recognize.
process and view images that require more processing time
than is available or that are too enormous to fit in memory. 3. Euclidean Distance Classifier
Additionally, the item has the ability to read, examine, and
display photographs of various resolutions. Choose a portion The system uses a Euclidean distance nearest-neighbor
of the image to read. Read, set, and write blocks of data. The classifier to identify a specific input image by comparing the
level where each pixel covers the maximum surface area is probe’s feature vector to the database image’s feature vectors,
the lowest or coarsest resolution level for large pictures with using Euclidean Distance(x and y).
several resolution levels. The level of resolution where each n
Euclidean Distance ( x and y ) = Â (x y )
2
pixel has the smallest area covered is the one with the most (1)
or finest detail. i=1
There are 7 sections to this essay. The first section outlines the A Euclidean vector represents a point’s location in Euclidean
research’s motivation. The related tactic of using tendencies n-space, with tips x and y denoting two points. Its length is
in ANN is examined in Section 2. The suggested work utilizes measured by Euclidean norm or magnitude by R N V Jagan
the Euclidean classification, as outlined in Section 3. Section Mohan, 2012[8, 9].
4 of the article details the amount of picture data. Section 5,
Vector Representation Image Classification Model Based on X = X12 + X 22 + º + X 2n = X .X (2)
Machine Search Engine. The experimental findings from the
If there is a direction from x to y, the gap between points x
proposed investigation are presented in Section 6. Section 7
and y can be expressed using the formula y-x.
provides a summary of the results and a conclusion.
Y – X = (y1 – x1, y2 – x2 … yn – xn) (3)
2. Tendencies in Artificial Neural
4. Lung Cancer Image Classification
Network (ANN)
Using Machine Learning
The text explores the prevalent patterns in artificial neural
networks. A key component of ML is ANNs, which provide Digital photography has generated enormous amounts of data,
computer specialists the ability to carry out difficult operations which has fueled the development of computer vision, an area
like pattern recognition, planning, and predictionby R.N.V of artificial intelligence that uses data to identify, recognize,
Jagan Mohan,2022[10]. Similar to other machine learning and categorize images. Machine learning techniques
algorithms, artificial neural networks crisis numbers and are used to analyze pixel patterns or vectors in order to
arrange lung cancer images or text data, but they learn from categorize objects and assign labels based on predetermined
user experience and repetitive activities. Artificial Neural criteria. Classifiers extract attributes from images to predict
classifications. There are ways for categorizing images that representation of pictures, we require a model that can extract
are binary or multiclass. Contrary to binary classification, image features, and for learning a vector representation of
which labels just one class of items across all pictures, text inputs, we need a model that can extract text features.
multiclass classification requires creating several labels for In order to semantically align the vector representations, the
various things by Qing Lv, 2022[7]. Both solutions require picture and text models must be trained simultaneously.
that the reference photographs be named. In-depth photo To ensure swift retrieval, we want a mechanism to quickly
classification is covered in this section. find related images while saving the currently saved photos.
As primarily focusing on vectorizing photos, it is logical to
5. Vector Representation Lung Cancer index them into a vector database. The vector representations
Detection Image Classification of the original photos are created by the indexing pipeline,
which then indexes them into a vector database.
Model Using Machine Search
The task requires us to generate a list of images when a user
Engine inputs a text or image query. The embedding generation
Machine Image search engine displays ranked list of service generates an embedding encoding of the input query.
related images based on word or image input, displaying The embedding query is sent to the vector database that
appropriate text and most similar images. This issue may returns the nearest neighbors of the query. The re-ranking
be conceptualized as a ranking issue. The model needs to service is mainly used to re-rank the nearest neighbors using
input two photos and generate a similarity score, which can a better model than the embedding generation model. It could
be used to order the photos based on this score. Utilizing be used to personalize the ranking to the specific user by
models that can learn a vectorial representation (embedding) using user-specific data. The resulting list is a list of image
of the pictures and compute a similarity metric on those IDs, and it is then sent to the image store to retrieve the actual
vectors is a common modeling strategy. For learning a vector images to return to the user.
Fig. 9.2 How use disease discovery to build machine images search engine
Visual Representation of Lung Cancer Image Classification Using Artificial Neural Network 57
Fig. 9.3 Same person different kinds of lung cancer image data set
dealing with a class-imbalanced data set with a significant Representation, Computational Data and Social Networks,
difference between positive and negative labels. Springer, pp 340–348, 2021.
2. Andrea D’Agostino: Vector Representation for Machine
1 + 90 Learning, Towards in Data Science, 2022.
Accuracy = = 0.91 (5)
1 + 90 + 1 + 8 3. Hay Mar Su Aung, Win Pa: Analysis of Word Vector
Representation Techniques with Machine-Learning Classifiers
Precision: Precision is the percentage of occurrences or
for Sentiment Analysis of Public Facebook Page’s Comments
samples that are correctly classified, as determined by the in Myanmar Text, DOI: 10.1109/ICCA49400.2020.9022842,
formula. IEEE Xplore: 05 March 2020.
TP 4. Jun Xie: Vector in machine learning, Medium, 2022.
Precision = (6) 5. K Grzegorczyk: Vector representations of text data in deep
TP + FP + FN
learning, arXiv, 2019.
where FN = False negative, FP = False positive, and TP = 6. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
in Health Care Using Hybrid Model Algorithms”,Journal of
True positive.
Artificial Intelligence, Machine Learning and Neural Network
Out of the 160 samples in Dataset -1, 105 of the predictions (JAIMLNN), Volume 3, 2023.
made by a lung cancer image model are accurate, whereas the 7. Nithyashree V: Image Classification using Machine Learning,
remaining 55 are not. Determine this model’s precision value. Analytics Vidhya, 2022.
True positives (TP) = 105 and false positives (FP) = 55 are the 8. Qing Lv,Suzhen Zhang:Deep Learning Model of Image
model’s results. Precision is calculated as follows: TP/(TP + Classification Using Machine Learning,Advanced
FP) = 105/ (105 + 55)= 105/ 160= 0.65625.As a result, the Pattern Recognition Systems for Multimedia Data,
Hindawi, Volume 2022, Article ID 3351256, https://doi.
model’s precision is 0.65625.
org/10.1155/2022/3351256,2022.
9. R.N.V. Jagan Mohan, R. Subbarao and K. Raja Sekhara Rao:
7. Conclusion and Future Perspective Similarity of Inference Face Matching on Angle Oriented Face
Recognition, Published in Journal of Computer Engineering
Vectors are essential to machine learning for structured and Intelligent Systems from www.iiste.org, ISSN:2222-1719
data representation and efficient operations like regression, (Paper) ISSN 2222-2863 (Online), Vol 3, No.2, 2012.
classification, clustering, and dimensionality reduction, 10. R.N.V. Jagan Mohan and R. Subbarao and Kurra Raja Sekhara
therefore data specialists must be able to understand and Rao: Efficient K-Means Cluster Reliability on Ternary Face
manipulate them. With 13.2% of new cases and 25.9% of Recognition using Angle Oriented Approach, Published in
fatalities due to cancer, lung cancer is the most prevalent International Journal of Informatics and Communication
cancer-related cause of death. Depending on the pathology Technology (IJ-ICT) Vol.2, No.1, January 2013, pp. 180-187
classification, disease stage, and patient features, the ISSN: 2252-8776, http://dx.doi.org/10.11591/ij-ict.v2i1.1779.
11. R.N.V.Jagan Mohan: Machine Learning approach for corona
prognosis varies.
virus disease extrapolation: A case study, International Journal
of Knowledge-based and Intelligent Engineering Systems,
References Vol-26,219-227, ISSN: 1327-2314(print),1875-8827(online)
DOI:10.3233/KES-220015,2022.
1. Akshat Gaurav, B. B. Gupta, Ching-Hsien Hsu, Arcangelo 12. Raunak Goswami: RMSE: Root-Mean-Square Error in
Castiglione & Kwok Tai Chui: Machine Learning Technique Machine Learning, Includehelp.com, April 16, 2023.
for Fake News Detection Using Text-Based Word Vector
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Diabetes is a severe illness that can cause blindness, kidney stones, heart problems, and other issues. Deep learning
has improved system abuse information processing, which can identify polygenic illnesses early on and provide patients with
access to critical information. This method retrieves diabetes-related data from databases by means of information withdrawal.
This research aims to create a system that can accurately predict a patient’s risk of acquiring diabetes by utilising decision trees,
artificial neural networks, naive bayes, random forests, and support vector machines.
Keywords: Artificial neural networks, Decision tree, Naive bayes, SVM etc.
1
bhavanisrkrcse@gmail.com, 2vinodvarmaaa@gmail.com, 3bmounika88@gmail.com
DOI: 10.1201/9781003529231-10
60 Algorithms in Advanced Artificial Intelligence
2. Related Work are shown in the block diagram that goes with it. According
to Baliunas, D.O., 2019 [4], the order computations that are
Marius et al. have developed a method for rapidly creating most frequently used to verify accuracy are artificial neural
nearest neighbours from data, as opposed to determining the networks (ANNs), Support vector machines, decision trees,
fastest and closest neighbour. Direct pursuit offers the fastest and naive bayes. XGBoost is an open-source software tool
resolution for high-dimensional PC vision issues, which is that uses the distributed inclination for tree computations
why this method is employed. There are little limitations on provided by the Gradient Boosting technique to improve
precision with rough techniques, and they allow for quick the performance and computational speed of AI models.
work. Because of how well XGBoost performed in structured data
competitions on Kaggle, its popularity has soared in recent
2.1 Naive Bayes Classifier years. In these competitions, data miners and information
Based on previous facts of the most likely pertinent analysts compete to create the most precise models for
circumstances, Naive Bayes determines the chance that an interpreting and predicting the data they gather. A number
event will happen. Naive bayes is the quickest and simplest of programming languages, including Java, Scala, Julia,
method for classifying large amounts of data. Sentiment Python, and R, have embraced XGBoost due to its extensive
analysis, text sorting, spam separation, and recommend er usage and improved developer advantages. Many tools
systems are just a few uses for NB classifiers. Probability and libraries, including distributed processing frameworks
theory’s Bayes hypothesis can be used to forecast illegible like Apache Spark and Dask, Caret, and Scikit-Learn, are
classifications. Gullible Bayes is an efficient computation compatible with XGBoost. Its exceptional processing speed
that can be finished in a matter of seconds. As a result, it and prediction performance attract data scientists.
could perform better than models that are more sophisticated
when the data is scarce by Quan Zou, 2018[10]. 3.1 Artificial Neural Network (ANN)
Machine learning techniques are essential for predicting,
2.2 Support Vector Machine pattern recognition, and planning. Particularly in chatbots
A training instance must adhere to in order to be considered for text classification, their capacity to learn from previous
a support vector. The ideal (maximum-margin) hyperplane experiences and repetitive user actions is causing them to
still provides the solution to our problem of diabetic illness gain popularity (Alicic, R. Z., 2017 [3]). The three main
diagnosis even if all training samples are removed from the components of a neural network—input, output, and a hidden
analysis save those from the support vectors. As a result, they layer—are described by Aman Preet Gulati (2022) [2] as
were given the name “support vectors”. By yuw, 2010[12] having the ability to learn and detect complex patterns that
would be challenging for humans to extract manually.
2.3 Decision Tree
A decision tree is a supervised learning algorithm that can
handle classification and regression issues, in contrast to
previous methods by Zhang Zq [13]. It consists of a root
node, branches, and leaf nodes in a geometric configuration.
4.1 Numpy
NumPy is a Python bundle that provides layered exhibit
objects, C, C++, and different dialects. It has applications
in direct polynomial math, arbitrary number age, and a
productive complex holder for conventional information. To
install, run “pip install numpy” and import numpy as np.
4.2 Pandas
Pandas is a Python library that provides robust data structures
for data control and analysis, inspired by Panel Data, an
econometric method using multi-layered data.
Wes McKinney created pandas as a high-performance, Fig. 10.2 Confusion matrix for predicted diabetes values
versatile data analysis tool, addressing Python’s limitations
in data manipulation and prepping, enabling five standard model correctly identified 940 out of 1000 non-positive
data processing processes: load, prepare, manipulate, model, prediction pictures, but incorrectly classified 60, resulting in
and analyze. true-negative samples and false-positive samples. The matrix
shows no misclassified samples, as diagonal entries accurately
Python and Pandas are widely used in academic and business
predict each class. Further progress towards genuine measures
settings. NumPy can be installed using the Python package
is recommended due to better understanding of the disarray
installer pip, while Pandas is not included in the standard
grid.
Python distribution.
5.2 Accuracy
5. Effectiveness Metrics’ Results The number of accurate forecasts divided by the total number
The different algorithms are compared with respect to F1 of expectations yields the accuracy metric, which has
score, specificity, accuracy, sensivity, and precision. The increased by 1030 out of 1100 accurately predicted in the
research work’s accuracy metrics are applied to the ANN’s prior model.
accuracy measures, and the following results are obtained.
5.3 Precision
5.1 Confusion Matrix for predicted Diabetes Due to the unequal class distribution, the taxonomy of
values correctness is not necessarily a trustworthy indicator of model
The chaos grid, also known as the mistake framework, is a performance. A high accuracy rate can be achieved even if
crucial concept in classification execution, providing a clear the model predicts all samples as the most common class,
comparison between model predictions and ground truth achieving a 90.9 percent accuracy rate. Our aim to analyze
marks, with rows representing projected class occurrences class-explicit measurements, such as “accuracy,” which is
and columns representing actual class occurrences. An defined as:
example of a binary classification system distinguishing Precision = True positive/(True positive + False positive)
positive and negative prediction pictures is presented, using The precision of prediction can be calculated using the
a disarray lattice to understand data behavior in an 1100 shot formula: Precision positive prediction = #samples accurately
test set. predicted positive. Negative predictive power is 90/(90+60),
or 60 percent. A non-positive prediction accuracy of 940/950
Table 10.1 Displays a confusion matrix for predicted diabetes
= 98.9%.
values
The model’s accuracy in predicting negative test results is
N = 400 Predicted
significantly higher due to its better categorization of non-
Actual No Diabetes Yes Diabetes
positive prediction pictures during training.
No Diabetes 40 20
Yes Diabetes 120 220 5.4 Recall
The model’s accuracy in accurately predicting a specific class
The model’s 100 positive predictions were incorrectly of samples is a crucial metric. Recall = True Positive/(True
classified, with only 10 being false negatives, indicating Positive + False Negative)
a misinterpretation of the “positive prediction” class. The
62 Algorithms in Advanced Artificial Intelligence
The recall rate can be calculated for both positive and non- rates and higher misleading positive rates lower the cut-off
positive predictions classes. incentive for positive classification by Kononenkoi,2001[6],
Recall positive prediction = 90/100 = 90% indicating a compromise between review quality and FPR.
The ROC curve assesses model performance.
Recall NonPositive forecast = 940/1000 = 94 percent.
5.7 Area Under the Curve
5.5 F1 Score
For statistical analysis, the area under the ROC curve is an
The F1-score, a measure of exactness and review, is often used essential measure. Because it takes into account all possible
to combine high recall and accuracy in various applications, threshold values, the AUC, a performance metric for binary
despite the importance of precision or recall. classifiers, is unaffected by the threshold value. We measure
F1-score = 2 * Precision * Recall/(Precision + Recall) the likelihood of a positive example chosen at random
The F1-score can be determined by examining the confusion performing better. Not all key performance indicators
matrix in Figure. benefit equally from limit-free metrics, despite the fact that
area under the curve (AUC) is a crucial model performance
F1_positive prediction = 2 * 0.6 * 0.9/(0.6 + 0.9) = 72% parameter. By tweaking its edge, you can accomplish
The trade-off between model accuracy and recall is evident, minimal demands without compromising the model’s good
with higher precision leading to lower recall rates and vice AUC. When you analyse a categorization model, consider
versa. factors such as business requirements and the ramifications
of low recall or accuracy. Clarity and interpretability are two
5.6 ROC Curve advantages of employing probabilities over a single mark
It is a crucial tool for understanding and analyzing the yield; nevertheless, support vector machines (SVMs) are
performance of a receiver. The diagram illustrates how less interpretable because they do not provide an essential
a paired classifier compares its true positive rate (TPR) likelihood.
against false positive rate (FPR) for different limit values.
Probabilistic classification models estimate the likelihood of 6. Experimental Results
a positive prediction, which is compared to a cut-off limit.
The model can predict probabilities between 0.45, 0.6, 0.7, In order to effectively battle diabetes, algorithms’ accuracy,
and 0.3. efficiency, and development of new talents may all be
improved, making the system ideal for hospitals as a full-
Cut-off = 0.5: predicted-labels = [0, 1, 1, 0] (default threshold) service healthcare diagnostic system.
Cut-off = 0.2: predicted-labels = [1, 1, 1, 1]
Cut-off = 0.8: predicted-labels = [0, 0, 0, 0] Table 10.2 Accuracy arc for TPR and FPR
Adjusting threshold values yields diverse labels, varying Classifiers K-NN SVM Naivy Decision Random ANN
Bayes Tree Forest
accuracy and recall rates. TPR and FPR are calculated using
the ROC curve, comparing TPR and FPR, as shown in an Accuracy (%) 76 75 74 71 71 85
example. The model demonstrates that higher genuine positive
diabetic symptoms, the app does not include a guidance 6. Kononenko I.: Machine learning for medical diagnosis:
system. A computer programme will analyse data from History, state of the art and perspective, Artif. Intell. Med.
2000 diabetes patients and provide personalised treatment 2001; 23: 89–109, doi: 10.1016/S0933-3657(01)00077-X,
recommendations based on their individual levels. We will 2001.
7. Kumar Dewangan A., Agrawal P. Classification of diabetes
compare the accuracy of Decision Tree, Random Forest,
mellitus using machine learning techniques, Int. J. Eng. Appl.
Naive Bayes, and K-Nearest Neighbor. Text, pictures, and
Sci., 2: 257905, 2015.
trees are all good training data for SVMs. However, sensitive 8. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
adjustments to basic limits are required. Various other options in Health Care Using Hybrid Model Algorithms”,Journal of
include decision trees, simple bayes, and ANN. In contrast Artificial Intelligence, Machine Learning and Neural Network
to the easy-to-understand but unpredictable choice trees, (JAIMLNN), Volume 3, 2023.
the robust Simple Bayes method focuses on the structure of 9. Nentwich M.M., Ulbig M.W.: Diabetic retinopathy—Ocular
data sources. Even though ANN is simple and accurate, it complications of diabetes mellitus, World J. Diabetes, 6: 489–
might be challenging to deal with very large data sets. The 499, doi: 10.4239/wjd.v6.i3.489, 2015.
proposed method is suitable for hospitals as a comprehensive 10. P. Suresh Kumar and S. Pranavi: Performance Analysis of
healthcare diagnostic system because it can be enhanced by Machine Learning Algorithms on Diabetes Dataset using
Big Data Analytics, International Conference on Infocom
increasing efficiency, creating new skills, and improving the
Technologies and Unmanned Systems, 978-1-5386-0514-1,
accuracy of algorithms to successfully fight diabetes.
Dec. 18–20, 2017.
11. Quan Zou,Kaiyang Qu,Yamei Luo,Dehui Yin,Ying Ju,Hua
References Tang: Predicting Diabetes Mellitus With Machine Learning
Techniques,Frontier Genetic, Volume 9, https://doi.org/
1. Aishwarya Mujumdar, V Vaidehi Dr: Diabetes Prediction 10.3389/fgene. 2018.00515, 06 November 2018.
using Machine Learning Algorithms, Science Direct, https:// 12. Yifan Qin, Jinlong Wu, Wen Xiao, Kun Wang, Anbing Huang,
doi.org/10.1016/j.procs.2020.01.047, Procedia Computer Bowen Liu, Jingxuan Yu, Chuhao Li, Fengyu Yu, and Zhanbing
Science, Vol: 165, Pages: 292–299, 2019. Ren: Machine Learning Models for Data-Driven Prediction of
2. Aman Preet Gulati: Diabetes Prediction Using Machine Diabetes by Lifestyle Type, Int J Environ Res Public Health,
Learning, Analytics Vidhya, January 4, 2022. 2022 Nov; 19(22): 15027, Published online 2022 Nov 15, doi:
3. Alicic R.Z., Rooney M.T., Tuttle K. R. Diabetic Kidney 10.3390/ijerph192215027,2022.
Disease: Challenges, Progress, and Possibilities, Clin. 13. Yu W., Liu T., Valdez R., Gwinn M., Khoury M.J: Application
J. Am. Soc. Nephrol, 12: 2032–2045, doi: 10.2215/ of support vector machine modeling for prediction of common
CJN.11491116,2017. diseases: The case of diabetes and pre-diabetes, BMC Med.
4. Baliunas D.O., Taylor B.J., Irving H., Roerecke M., Patra Inform. Decis. Mak.,10: 16. doi: 10.1186/1472-6947-10-16,
J., Mohapatra S., Rehm J. Alcohol as a risk factor for type 2010.
2 diabetes: A systematic review and meta-analysis, Diabetes 14. Zhang Z.Q., Yang L.Q., Han W.T., Wu Y.Y., Zhang L.H., Gao
Care.32: 2123–2132, Doi: 10.2337/dc09-0227, 2009. C., Jiang K., Liu Y., Wu H.Q. Machine Learning Prediction
5. Humar Kahramanli, Novruz Allahverdi: Design of a Hybrid Models for Gestational Diabetes Mellitus: Meta-analysis. J.
System for the Diabetes and Heart Disease, Expert Systems Med. Internet Res. 2022; 24: e26634, doi: 10.2196/26634,
with Applications: An International Journal, 35, 1–2, July 2022.
2008.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
64 Algorithms in Advanced Artificial Intelligence
Tackle Comorbid Obesity in T2DM by
Applying New Strategies to Optimize
Glycaemic Control and Weight 11
Management
Yugandhar Bokka1
Research Scholar, Gandhi Institute of Engineering and Technology (GIET) University,
Gunupur, Odisha
R. N. V. Jagan Mohan2
Associate Professor, Sagi Rama Krishnam Raju Engineering College, Bhimavaram
M. Chandra Naik3
Professor, Gandhi Institute of Engineering and Technology (GIET) University,
Gunupur, Odisha
Abstract: Obesity management is an important therapeutic goal for patients with diabetes and obesity. The majority of patients,
sadly, do not receive care that is in line with the most recent findings. The recommendation is to prioritize weight loss promotion
in at-risk patients, utilizing newer medications as needed. To address comorbid obesity in type 2 diabetes, new strategies are
being used to improve weight management and glycaemic control. The paper aims to improve patients’ capacity to manage
individuals with common disorders. This activity aims to review the relationship between obesity and type-2 diabetes mellitus,
discuss the reimbursement of early treatment intensification and weight loss, discuss the latest clinical trial results for dual
GIP/GLP-1 RAs, and optimize care for patients with obesity and T2DM. The study explores new strategies for optimizing
glycemic control and weight management in patients with comorbid obesity using virtual patient simulation with the K-Nearest
Neighbour method.
Keywords: Clinical trail, Diabetes, GIP/GLP-1, K-Nearest neighbour, Obesity, T2DM, Therapeutic etc.
DOI: 10.1201/9781003529231-11
Tackle Comorbid Obesity in T2DM by Applying New Strategies to Optimize Glycaemic Control and Weight Management 65
during the first three months of pregnancy, which includes 2. Proposed Work
congenital heart abnormalities and spina bifida. The primary
reason infants die is due to these abnormalities. Most birth The study explores the use of virtual patient simulations
malformations are caused by a combination of environmental to improve glycaemic control and weight management
factors, behaviours, and genes (H. Tada et al., 2017). However, strategies for comorbid obesity in type 2 diabetes. The
the precise aetiology of many disorders is yet unknown. Taking following objectives are as follows:
certain pharmaceuticals, having specific medical conditions, The study explores the relationship between obesity and type
taking known birth defect-causing medications, having a 2 diabetes mellitus (T2DM), the benefits of earlier treatment
family history of birth abnormalities, and giving birth after intensification and weight loss for achieving glucose control,
the age of 35 are all factors that increase the chance of having and the long-term impact of these interventions on overall
a child with a birth defect (Ng, Marie, et al., 2014). For the health outcomes in patients with T2DM.
best course of therapy, speak with a doctor. Medical disorders
like diabetes, high blood pressure, and infections can cause The second objective is to present the latest clinical trial
complications during pregnancy. According to N. Demirel et results on dual GIP/GLP-1 RAs’ impact on A1C, obesity, and
al. (2018), women who have diabetes, gestational diabetes, dyslipidemia and their implications for future practice.
chronic hypertension, or infections should be under closer The ultimate goal is to optimize T2DM and obesity care
medical observation. The right care ensures a good pregnancy. through a comprehensive approach combining lifestyle
According to P. Bramlage et al. (2014), concentrating on modifications and personalized pharmacologic strategies for
a good pregnancy assures a healthy pregnancy, a smooth weight loss and glycemic control.
transition to a positive labour and delivery, and an optimistic
view of parenthood. Researchers conduct clinical trials to
address health challenges and discover pertinent therapies
with a varied population, including older persons (S. Khan
Fig. 11.1 Block diagram for proposed work
and T. Yairi, et al., 2018). Phases 1, 2, and 3 are among the
stages that these studies go through in order to assess safety, K-Nearest Neighbour Classification: K-Nearest
efficacy, and side effects. A medicine or device cannot be Neighbours (KNN) is a tool that assists in selecting the most
licenced by the FDA until Phases 1, 2, and 3 have been suitable neighbourhood for a new friend by involving three
completed. A participant may only take part in one trial at close friends with similar interests.
a time and must meet the inclusion and exclusion criteria. A KNN uses a simple rule: close friends likely share similar
class of drugs called GLP-1 RAs reduces blood sugar levels interests, which guides its decision-making process.
and treats type 2 diabetes. Following the approval of a daily
oral version of semaglutide, patients can inject GLP-1 RAs 1. Measure Distances: To find friends with similar interests,
twice daily, once daily, or once weekly. These drugs work in compare the interests of the new friend to your own.
similar ways to control blood sugar levels by raising insulin 2. Chooses the closest Friends: Choose K close, similar-
secretion in response to hyperglycemia, blocking glucagon minded friends as your nearest neighbours who share similar
secretion, delaying stomach emptying, and reducing caloric interests with the new friend.
intake and body weight. Compared to short-acting drugs, 3. Friends Opinion’s: Ask friends for advice on which
long-acting therapies have a greater effect on HbA1c and neighbourhood to join a new friend, as their opinions are
plasma glucose levels during overnight and fasting. For valuable due to shared interests and proximity.
those with type 2 diabetes, GLP-1 RAs are recommended
as the initial injectable glucose-lowering drug, even prior KNN is akin to seeking advice from friends who share
to initiating insulin therapy. Patients can use GLP-1 RAs similar interests and live nearby for decision-making.KNN is
with basal insulin in formulations that have a fixed or free a machine learning algorithm that can sort objects based on
dosage. GLP-1 RAs may also help prevent renal issues. This similarity, recommend content on healthcare platforms and
essay’s remaining sections are organised as follows: Section identify unusual or unique items.
1 discusses the introduction. In Section 2, we provided an Pseudo code for KNN:
explanation of the desired work. Section 3 presents a given • Calculate Euclidean distance between the pointsd (x, xi)
overview of the dataset and experimental results. Section 4 where i = 1, 2… n.
provides a concluding analysis and perspective.
• Sort the n Euclidean distances in non-decreasing order.
66 Algorithms in Advanced Artificial Intelligence
• Take k-value; take the first k-distances from sorted list. The accuracy of the KNN model is 72.08 %. As a result,
• Find those k-points corresponding to these k-distances. applying the KNN model to the prediction of risk in pregnant
• ki denotes the number of points belonging to the ith class women can yield superior results.
among k points.
4. Conclusion
3. Experimental Result This study employs virtual patient simulations to enhance
The diabetes dataset is collected from Kaggle. This dataset weight management strategies for people with type 2 diabetes
contains different features like pregnancies, Glucose, Blood who also suffer from comorbid obesity and to better control
pressure, skin thickness, insulin, BMI, age etc. blood sugar levels. Delve into the link between obesity and
type 2 diabetes, the benefits of initiating treatment sooner, and
From the Fig. 11.2, it can be seen that the features of diabetes the lasting effects of these measures on those afflicted with
pregnancy women exposed to age and obesity. Training is the condition. Also included are the findings from clinical
performed in the model to achieve early prediction. trials that utilised dual GIP/GLP-1 RAs. Ideally, disease risk
prediction in medicine would help doctors identify potential
patterns and risks of disease before making a diagnosis and
taking any necessary measures to treat or prevent the patient’s
condition.
References
1. Ama Moor, Vicky Jocelyn, et al. “Dyslipidemia in patients
with cardiovascular risk and disease at the University Teaching
Hospitalof Yaoundé, Cameroon, International journal of
Fig. 11.2 Features of dataset vascular medicine, 2017.
The experimental result is on obesity and type 2 diabetes, 2. A. Maxwell, R. Li, B. Yang, H. Weng, A. Ou, H. Hong, Z.
Zhou, P. Gong, and C. Zhang,: Deep learning architectures for
the benefits of early treatment intensification, and the long-
multi-label classification of intelligent health risk prediction,
term effects of these interventions on T2DM patients. It also
BMC Bioinf., vol. 18, no. S14, pp. 523–525, Dec. 2017.
presents clinical trial results on dual GIP/GLP-1 RAs. 3. A.Mohammadbeigi, E. Moshiri, N. Mohammadsalehi, H.
For training and prediction, use the KNN model. Figure 11.3 Ansari, and A. Ahmadi: “Dyslipidemia prevalence in Iranian
shows the KNNmodel varying the number of neighbours. The adult men: The impact of population-based screening on the
below graph shows that varying number neighbours, it is also detection of undiagnosed patients, “World J. Men’s Health,
show the comparison of training dataset and testing dataset. vol. 33, no. 3, p. 167, 2015.
4. Cui, Shiyue, et al: Research on risk prediction of dyslipidemia
in steel workers based on recurrent neural network and LSTM
neural network. IEEE Access 8 (2020): 34153-34161.
5. D. Wang, J. Fan, H. Fu, and B. Zhang: Research on
optimization of big data construction engineering quality
management based on RNNLSTM, Complexity, vol. 2018,
pp. 1–16, Jul. 2018.
6. G. Jain, M. Sharma, and B. Agarwal: Optimizing semantic
LSTM for spam detection, Int. J. Inf. Technol., vol. 11, no. 2,
pp. 239–250, Apr. 2018.
7. H. Tada, M.-A.Kawashiri, and M. Yamagishi: Comprehensive
genotyping in dyslipidemia: Mendelian dyslipidemia caused
by rare variants and mendelian randomization studies using
common variants,”J. Hum. Genet. vol. 62, no. 4, pp. 453–458,
Jan. 2017.
8. Ng, Marie, et al.: Global, regional and national prevalence of
overweight and obesity in children and adults during 1980–
2013: a systematic analysis for the Global Burden of Disease
Study2013.” The lancet, 384.9945, 766–781, 2014.
Fig. 11.3 Varying number of neighbours 9. N. Demirel, S. Özbay, and F. Kaya: The effects of aerobic and
anaerobic training programs applied to elite wrestlers on body
Tackle Comorbid Obesity in T2DM by Applying New Strategies to Optimize Glycaemic Control and Weight Management 67
mass index (BMI) and blood lipids, J. Edu. Training Stud., 11. S. Khan and T. Yairi: A review on the application of deep
vol. 6, no. 4, p. 58, Mar. 2018. learning in system health management,’ Mech. Syst. Signal
10. P. Bramlage, S. T. Azar, O. Okkeh, P. Brudi, B. M. Process., vol. 107, pp. 241–265, Jul. 2018.
Ambegaonkar, H. A. Hantash, S. Jambart, M. El Zaheri, 12. S. Hussain, J. Keung, A. A. Khan, A. Ahmad, S. Cuomo, F.
R. Rachoin, A. Chafoun, and L. LaHood: Factors in_ Piccialli, G. Jeon, and A. Akhunzada: Implications of deep
uencingdyslipidemia in statin-treated patients in Lebanon learning for the automation of design patterns organization,
and Jordan: Results of the dyslipidemia international study, J. Parallel Distributed Computer, vol. 117, pp. 256–266, Jul.
Vascular Health Risk Manage., vol. 10, p. 225, Apr. 2014. 2018.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
68 Algorithms in Advanced Artificial Intelligence
A Literature Survey on Deep Learning
Approach Used for Audio-to-Sign
Conversion with Gesture Recognition for 12
the Deaf and Dumb
B. Veerendra1
D. Ramakrishna2
Research Scholar, Department of Computer Science and Engineering, GITAM University,
Visakapatnam, Andhra Pradesh, India
Assistant Professor, Department of Computer Science and Engineering, GITAM University,
Visakapatnam, Andhra Pradesh, India
Abstract: Syntax is the process of arranging words and phrases in a language to form coherent sentences. Deaf and dumb
people use a variety of signs in Sign Language to communicate with one another. We must master the sign language, especially
is a challenging endeavor, in order to converse with the deaf and dumb. A person who is hearing impaired may or may not
comprehend the speaker, and the speaker may or may not comprehend the hearing impaired person’s sign language. Therefore,
if one wishes to have a conversation that is understandable with people who are deaf and dumb, they must learn Sign Language.
The technology that is utilized in this proposed model is deep learning, and it is a desktop application that was designed and
developed using the Python programming language. Convolution Brain Organization (CNN) is a Profound Learning Strategy
utilized for breaking down the pictures taken and to recognize the signs. It can translate hand gestures into text and audio into
sign language. We use Python programming and a Deep learning techniques for analyzing the webcamera input. The proposed
style uses Convolution Neural Network (CNN) which enables to categorize pictures of different hand and sign movements for
the Alphabets.
In this proposed model, there are two main features included which helps to reduce communication difficulties with deaf
and dumb people. They are: Hand Gesture Recognition and Audio to Sign Language Conversion. In Audio to Sign Language
Conversion, the audio of the speaker is recognized by this model, and then following the conversion of speech to text as well
as text into a sign language. In the user’s hand gestures is captured from web camera and then it uses the Convolution Neural
Network model, gestures are recognizable and displays hand gesture into text form. The proposed paper’s primary goal is to
facilitate communication with the people who are deaf and dumb. It fills the communication gap between the hearing and the
dumb and the general public. Text and images representing this model’s output are displayed on the desktop computer’s screen.
Keywords: Audio to sign conversion, Sign to text, Hand gesture recognition, Convolutional neural network (CNN), Deep
learning
DOI: 10.1201/9781003529231-12
A Literature Survey on Deep Learning Approach Used for Audio-to-Sign Conversion with Gesture Recognition for the Deaf and Dumb 69
recognition, natural language processing, machine translation, A model for a web-based tool that can translate speech
etc., where they have produced comparable accurate results to signs. If the associated terms cannot be located in the
with the human expert performance. database, the system searches for synonyms of the term in
In the proposed article, Convolution Neural Network question and substitutes them. It lacks the conversion of hand
(CNN) is used for analyzing the input, that is, the camera gestures into text. [4,5]
feed. This neural network class, which is used in deep A model that makes use of CNN to identify the user’s hand
learning, is often referred to as a Comp Net. It is primarily gestures when using a webcam. The text appears on the
used for the analysis of pictures and videos. It can also be output screen if the gesture is recognized. It is able to convert
applied to other classification and data analysis issues. that text to audio by using the gTTS library. [6,7]
It has the ability to recognize and comprehend patterns. An approach for creating a system that enables paralyzed
Convolutional layers are an additional set of hidden layers individuals to communicate to each other using hand
added to the regular hidden layers, which sets it apart from gestures. Signs can be converted to text by it. In the event of
other deep learning models. This article aims to reduce the an emergency, individuals can be informed by this system.
human communication interaction between the hearing- [8]
impaired and common people. Deep Learning and Machine
Learning are most commonly used now-a-days in many A model that incorporates the ability to recognize air written
domains. The Machine Learning techniques are used for text. Using deep CNN architecture, it is capable of tracking
data analysis and Deep Learning techniques are used for written digits from 0 to 9 at the fingertips while in the air.
Speech recognition software, natural language processing The audio to sign conversion and sign detection features are
(NLP), image recognition tools etc. In the existing systems, absent from this model. [ 9,10]
Machine Learning and Deep Learning techniques are used for
identifying the hand gestures, recognizing the letters written 3. Methodology
on the air and translating the audio into sign language. Some
of these systems can only able to recognize hand gestures of The suggested model was created with CNN, a deep learning
the user or can only convert speech to sign language or can technique, and the Python programming language. It fills
identify the characters that are written in the air by applying the communication gap between the dumb and deaf and the
different deep learning and machine learning techniques. general public. This proposed model contains three features
Various algorithms are used in those existing systems like which helps to make a mutually understandable conversation
Artificial Neural Network (ANN), Recurrent Neural Network with the hearing-impaired people. These features are
(RNN), and Long Short-Term Memory (LSTM) etc. which implemented using a deep learning technique and Machine
are deep learning techniques that are used for recognizing learning which helps to analyze the input from the camera
gestures made by the user. feed and helps to provide appropriate results with good
accuracy. The following are the three different features that
are included in this article:
2. Literature Survey • Hand Gesture Recognition
A model with the goal of improving human-device interaction • Audio To Sign Conversion
uses a recurrent neural network (RNN) to identify user hand
gestures. To gather information from the five hand signals, 3.1 Audio to Sign Conversion
a sine antenna is needed. An RNN is then used to condition The Speech Recognition Python library is used by the Audio
the data. The audio conversion into sign language is not to Sign feature to translate speech to text. The speech will
supported by this model. [1] break into words in this instance, and those words will break
A CNN-based model for human-computer interaction was into letters. After that, the signs that correspond to the letters
proposed. Features that are primarily used for abnormal in the sentence will be shown on the screen.
people were proposed by this model. They are Hand Gesture
Recognition and Audio to Sign Conversion. [2] 3.2 Hand Gesture Recognition
A device-free hand gesture recognition system built on deep It recognizes hand gesture of user using deep learning. That
learning and Channel State Information (CSI) models. This means, the hand gestures made by the 2nd person within the
system’s capabilities are restricted to reading handwritten frame of the camera is identified using a 2D CNN model and
numbers from 0 to 9 in the air. It is also unable to translate displays the respective text related to the sign performed by
speech into the sign language. [3] the user in front of the camera.
70 Algorithms in Advanced Artificial Intelligence
Fig. 12.1 Automated object recognition with IoT for visually impaired users
folders. The below shown is the screenshot (Fig. 12.7) of the Handwritten characters in which every character contains
folders that contain over 300 images of Hindi Hand Written upto 300 images.
Characters: The below shown Fig. 12.9 is the sample .csv file of
The Fig. 12.8 is the sample dataset of the Handwritten letter Hindi Handwritten Characters which is used in the training
‘ka’. Likewise the model is trained with 36 different Hindi process.
In this proposed article, there are two main modules. These easyocr : API used for doing OCR of Hindi
two modules facilitate clear and effective communication characters from image.
between the general public and those with hearing IPython.display : This module is used to handle image
impairments. They are: files.
threading : Threading is OS concept to perform
4.1 Audio to Sign Conversion various processes at a time.
The task of translating audio into sign language falls to this
module. The input for this is speech or audio. The speech In this model, there are some programmer defined modules
will split into words then these words will split into letters. that are used for converting audio/speech into its appropriate
Next, the folder containing the signs is searched through the sign language, identifying the signs and letters written on air.
images to find those letters. Subsequently, the corresponding The following shown are the programmer defined modules
signs for each letter in the sentence will appear on the screen. which are implemented using the Python Programming
Language.
4.2 Sign to Text Conversion
This module is in charge of translating signs into text, or Table 12.2
more accurately, it recognizes the sign that is in front of the main.py : This is the foremost file where the user
camera. In this, the model will be trained and tested with a runs, operate and access the features
dataset using 2D CNN model. When the user show signs in included in the proposed model. When
the user runs it, a page is shown with
front of the camera then it will load the camera feed and next, buttons where each button has different
it displays the respective text related to the sign performed by functionality.
the user in front of the camera. stt.py : This file is used to convert sign language
In this some python modules are used to implement the to text. It is used for displaying camera
module. They are Pandas, tkinter, OpenCV, NumPy, and give camera feed to app.py file.
SpeechRecognition etc. The Python modules which are used SLT.h5 : This file is loaded in stt.py to convert sign
in designing the model are listed below: language to text. This file consists of the
saved training and testing datasets used
in recognition of Sign Language.
Table 12.1 Modules/APIs used
app.py : app.py file contains the code to analyse
numpy : Library in python used for accessing, the camera feed and then process it using
storing image pixels in the array form. Machine Learning.
pandas : Library used in python for accessing char_ : This contains the code that can able to
csv files. recognition,py identify the Hindi characters that are
keras : It is used for loading Keras model drawn in air.
(Machine learning Library) savedModel.h5 : This .h5 file is used in char_recognition.
Speech Recognition : Google based API/Module to detect py file to identify Hindi characters that
the speech & convert to text. are drawn in air. It contains the saved
training and testing dataset which helps
Tkitner : Graphical module of python to design for written character recognition on air.
and develop the GUI.
cv2 : Computer vision to access camera The proposed article “A Literature Survey on Deep Learning
feed of the PC. approach used for Audio-To-Sign conversion with Gesture
PIL : Python Image Library to handle Image Recognition for the Deaf and Dumb” has two main features
files in python. and has two workflows. One is for Hand Gesture Recognition
ctypes : This module is used for getting which recognizes the hand gestures of English alphabets,
resolution of the screen. Audio to Sign Language Conversion which converts speech/
os : This module is used to get the audio to sign language. Every feature in this model is designed
Operating System functionality in using sequential 2D CNN model which is build using python.
code. System of each feature is designed with distinct workflows
pylab : This module is to plot pylab graph and functions for each task that a feature does. They are as
(display sign images while converting follows:
speech to sign).
deque : This module is used for enqueue 4.3 Audio To Sign Conversion Flow
and deque of image pixels while
recognizing Air Text by user. Every time an individual speaks “Hi everybody”, The
speech is identified and translated into text using a speech
A Literature Survey on Deep Learning Approach Used for Audio-to-Sign Conversion with Gesture Recognition for the Deaf and Dumb 75
recognition library. Then using Python programming, it will displays the internal steps performed by the model when the
be divided into words and then into characters. The input is user makes hand gestures in front of the camera.
classified and identifies the respective hand gesture of the
characters. Next, as seen in figure 9, the output is shown in 5. Experimental Results
a window on the users’ desktop in a sequence of images in
sign language. The output will be NONE and will only be in
sign language if the user’s speech is not clear enough to be
understood. The user can know whether the model is showing
correct output or not by the output images which contains
the related English alphabet at the right upper side corner of
every image as shown in the Fig. 12.10
Fig. 12.11 Flow of hand gesture recognition Fig. 12.14 Voice output of hand gesture
76 Algorithms in Advanced Artificial Intelligence
References
1. G. Park, V. K. Chandrasegar and J. Koh, “Hand Gesture
Recognition using Deep learning Method,” 2021 IEEE
International Symposium on Antennas and Propagation and
USNC-URSI Radio Science Meeting (APS/URSI), 2021,
pp. 1347–1348, doi: 10.1109/APS/URSI47566.2021.9703901.
2. S. Pariselvam, D. N., D. S. and S. B., “An Interaction System
Using Speech and Gesture Based on CNN,” 2020 International
Conference on System, Computation, Automation and
Networking (ICSCAN), 2020, pp. 1–5, doi: 10.1109/
ICSCAN49426.2020.9262343.
3. Z. Wang et al., “WiDG: An Air Hand Gesture Recognition
System Based on CSI and Deep Learning,” 2021 33rd Chinese
Control and Decision Conference (CCDC), 2021, pp. 1243–
1248, doi: 10.1109/CCDC52312.2021.9602438.
Fig. 12.15 Conversion of audio to HandSign
4. Q. M. Areeb and M. Nadeem, “Deep Learning Based Hand
Gesture Recognition for Emergency Situation: A Study on
6. Conclusion Indian Sign Language,” 2021 International Conference on
Data Analytics for Business and Industry (ICDABI), 2021,
It’s challenging to communicate with the deaf and dumb pp. 33–36, doi: 10.1109/ICDABI53623.2021.9655842.
people if you don’t know sign language. Common people 5. K. Tiku, J. Maloo, A. Ramesh and I. R., “Real-time
cannot easily understand it, and those who are hard of hearing Conversion of Sign Language to Text and Speech,” 2020
cannot comprehend what the common people are saying. Second International Conference on Inventive Research in
Computing Applications (ICIRCA), 2020, pp. 346–351,
As a result, there is a communication issue between them
doi: 10.1109/ICIRCA48905.2020.9182877.
here. The goal of this article is to improve communication 6. A. Yadav, R. Saxena, B. Saini, V. K. Verma and V. Srivastava,
within this community as well as with the outside world. “Audio to Sign Language Translator Web Application,” 2021
This suggested model primarily includes two features: hand International Conference on Computational Performance
gesture recognition and audio to sign language conversion Evaluation (ComPE), 2021, pp. 321–326, doi: 10.1109/
which makes it easier to speak with people who are hard ComPE53109.2021.9751857.
of hearing. The technologies used in the proposed article 7. A. Dixit et al., “Audio to Indian and American Sign Language
demonstrate the potential for deep learning and computer Converter using Machine Translation and NLP Technique,”
vision. It improves communication accessibility for people 2022 Third International Conference on Intelligent Computing
with disabilities. Instrumentation and Control Technologies (ICICICT), 2022,
pp. 874–879, doi: 10.1109/ICICICT54557.2022.9917614.
This system provides us with a high gesture recognition 8. T. A. Siby, S. Pal, J. Arlina and S. Nagaraju, “Gesture based
rate when the gestures are made in a clear background with Real-Time Sign Language Recognition System,” 2022
accuracy more than 95% within less response time. When International Conference on Connected Systems & Intelligence
compared to other models, the article’s overall average (CSI), 2022, pp. 1–6, doi: 10.1109/CSI54720.2022.9924024.
accuracy is higher. To improve usability and effectiveness 9. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
for users, this article expands the vocabulary of recognized in Health Care Using Hybrid Model Algorithms”,Journal of
signs and improves accuracy. This proposed model can Artificial Intelligence, Machine Learning and Neural Network
offer a simple, effective method for communicating in sign (JAIMLNN), Volume 3, 2023.
10. S. Gupta, R. Thakur, V. Maheshwari and N. Pulgam,
language. It recognizes and interprets the hand gestures and
“Sign Language Converter Using Hand Gestures,” 2020
what we write on air. The input for this article is audio, hand 3rd International Conference on Intelligent Sustainable
gesture made by the user and letters that the user writes on Systems (ICISS), 2020, pp. 251–256, doi: 10.1109/
air. It enhances the interaction and it allowed to recognized ICISS49785.2020.9315964.
and predict letters with high accuracy. Importance of this 11. G. Bastas, K. Kritsis and V. Katsouros, “Air-Writing
article is it is well-designed and trained with a dataset of Recognition using Deep Convolutional and Recurrent
hand gestures. Hence this paper can able to reduce the Neural Network Architectures,” 2020 17th International
communication problem between deaf and dumb with others. Conference on Frontiers in Handwriting Recognition
(ICFHR), Dortmund, Germany, 2020, pp. 7–12, doi: 10.1109/
ICFHR2020.2020.00013.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Technological advancements in the field of psychological assessment enable machines to accurately determine user
emotions through the Recognition of Emotions in Speech method. Recognition of Emotions in Speech accurately predicts human
emotions through speech, improving psychological assessment. Recognition of Emotions in Speech can identify emotions like
impartial, at ease, joyful, depressed, scared, furious, disgusted, and shocked. The paper presents a federated learning method for
emotion recognition in speech based on the Mel Frequency Cepstral Coefficient (MFCC). The study employs the RAVDESS
dataset and the Federated Learning System for Cognitive Radio to develop speech-emotion identification classifiers. In the
source-filter model of speech, the vocal tract is represented by MFCC, which are important speech features that are extracted
for recognition tasks. When attempting to extract spectral information from expressive speech, the Fourier transform signal
processed via a Mel-spaced reduce bank is the most widely used method. The Federated Learning Architecture is used in an
experimental speech-emotion recognition scenario to accurately extract spectral information from expressive speech using the
RAVDESS dataset and the Federated Learning System for Cognitive Radio.
Keywords: Federated learning, Recognition of emotions in speech, Mel frequency cepstral coefficients, Ravdess dataset
DOI: 10.1201/9781003529231-13
78 Algorithms in Advanced Artificial Intelligence
the Federated Learning System for Cognitive Radio were 8. R. Banse and K. R. Scherer: Acoustic profiles in vocal emotion
used to develop speech-emotion identification classifiers. expression, Pers. Soc. Psychological, vol. 70, no. 3, pp. 572–
587, 1996.
9. S. Wu, T. H. Falk, and W.-Y. Chan: Automatic speech emotion
Acknowledgement recognition using modulation spectral features, Speech
Communication, vol. 53, no. 5, pp. 768–785, May 2011.
In exchange for their assistance in this respect, my supervisor
10. Soegaard, M. and Friis Dam, R.: The Encyclopedia of Human-
and other CSE and AIML department employees provided
Computer Interaction. 2nd education, 2013.
the datasets. 11. T. L. Nwe, S. W. Foo, and L. C. De Silva: Speech emotion
recognition using hidden Markov models, Speech
References Communication, vol. 41, no. 4, pp. 603–623, Nov. 2003.
12. V. Hozjan and Z. Kačič: Context-Independent Multilingual
1. A. Schuller, B. Steid, S. l, and Batliner: The interspeech Emotion Recognition from Speech Signals,” International
emotion challenge, Interspeech, pp. 312– 315, 2009. Journal of Speech Technology, vol. 6, no. 3, pp. 311–320,
2. B. S. Atal: Effectiveness of liner prediction characteristics 2003.
of the speech wave for automatic speaker speech wave for 13. W. Dai, D. Han, Y. Dai, and D. Xu: Emotion Recognition and
automatic speaker identification and verification,” Acoustic Affective Computing on Vocal Social Media,” Inf. Manag.,
Soc. Am., vol. 55, no. 6, pp. 1304–1312, 2005. Feb. 2015.
3. C.-C. Lee, E. Mower, C. Busso, S. Lee, and S. Narayanan: 14. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
Emotion recognition using a hierarchical binary decision in Health Care Using Hybrid Model Algorithms”,Journal of
tree approach, Speech Communication. vol. 53, no. 9–10, Artificial Intelligence, Machine Learning and Neural Network
pp. 1162–1171, Nov. 2011. (JAIMLNN), Volume 3, 2023.
4. F. Burkhardt, A. Paeschke, M. Rolfes, W. F. Sendlmeier, 15. Yip et al.: Discrete Cosine Transform - Algorithms,
and B. Weiss: A database of German emotional speech, in Advantages, Applications, Computer Science, Engineering,
Interspeech, 2005, vol. 5, pp. 1517–1520, 2005. Mathematics, DOI: 10.1016/c2009-0-22279-3, Academic
5. Gupta V, Shankar RS, Kotha HD, and Raghaveni J: Voice Press, ISBN 012580203X, August 1990.
Identification in Python Using the Hidden Markov Model, 16. Zhang, J., Yin, Z., Chen, P., and Nichele, S: Emotion
International Journal of Advanced Science and Technology, recognition using multi-modal data and machine learning
vol. 2020, p. 29. techniques: a tutorial and review, Inf. Fusion 59, 103–126,
6. H. Cao, R. Verma, and A. Nenkova: Speaker-sensitive emotion DOI: 10.1016/j.inffus.2020.01.011,2020.
recognition via ranking: Studies on acted and spontaneous 17. Zhenjie Song: Facial Expression Emotion Recognition Model
speech,” Computer Speech Language, vol. 28, no. 1, pp. 186– Integrating Philosophy and Machine Learning Theory,
202, Jan. 2015. Frontiers, Psychology, Volume: 12, 2021.
7. M. M. H. El Ayadi, M. S. Kamel, and F. Karray: Speech
Note: All the figures in this chapter were designed by the author.
Emotion Recognition using Gaussian Mixture Vector
Autoregressive Models, IEEE International Conference on
Acoustics, Speech and Signal Processing - ICASSP ’07, 2007,
vol. 4, pp. IV–957–IV–960, 2007.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
82 Algorithms in Advanced Artificial Intelligence
Abstract: To aid the visually impaired in leading more autonomous lives, this paper is being considered. Nowadays, technology
plays a significant role in meeting everyone’s requirements in our technologically advanced society. In the lives of those with
physical disabilities, it is also extremely important. Now let’s talk about blind individuals. No matter how modern the technology
is, they might still not be able to use it since they can’t see. They are entirely reliant on other people for even the most menial
of tasks, not to mention equipment. We suggested a gadget with cutting-edge tech that would enable the visually impaired to
accomplish their own tasks instead of relying on others, thus addressing the aforementioned gaps in accessibility. For object
detection, the app uses image processing techniques, and for voice output, it employs speech synthesis. The technology aims
to provide visually impaired people with real-time audio or vocal information about things scanned by their mobile cameras. A
substantial and extensively researched subject in computer vision, picture detection on moving objects has found applications
in domestic, commercial, and industrial settings. Current methods have a number of drawbacks, such as poor accuracy and
performance that stem from issues like not analysing the trained data enough, being too dependent on object motion, and not
being able to distinguish between objects. So, to quickly and accurately recognise the item, the Fast R-CNN (region-based
convolutional neural networks) technique has been used. People who are visually impaired can get around with the help of a
speech synthesiser and the damage that it detects.
Keywords: Tensor flow, Google speech to text (API), Open CV, Raspberry
1. Introduction labels and locating the correct bus stop. The project is
reserved to the rightful owner. The user initiates a process of
The world’s millions of visually impaired people face object detection, allowing them to Listen to the device’s voice
new obstacles every day as they try to make sense of their instructions to determine what it is. As a result, a new way of
surroundings. Visually impaired individuals face daily thinking about empowering people with visual impairments
challenges in performing tasks such as deciphering product to live independently has been born. Everyday living presents
1
jmsvravikumar@gmail.com, 2m_babureddy@yahoo.com, 3Srikanth.mandela@gamil.com, 4drsrkrit@gmail.com
DOI: 10.1201/9781003529231-14
Automated Object Recognition with IoT for Visually Impaired Users 83
a number of challenges to those who are vision impaired or coherence point drift registration. In order to evaluate the
blind. The goal is to cultivate an Android software called shape-based segmentation and to set a standard for future
“Visually Assist” helps the visually impaired. It will make work, we constructed the Co-Shape data set. The shape data
specialised gadgets and other wearable tech unnecessary for set testing and comparisons with related co-segmentation
object recognition tasks like Circumambulate, they do. The methods show that the approach performs beautifully.
app’s real-time item detection and identification capabilities
allow the sight-impaired to move autonomously. For object 2. Literature Survey
detection, the app uses image processing techniques, and for
voice output, it employs speech synthesis. The system’s goal In this research [1], we propose a unified mutual learning
is to identify scanned things using the mobile camera and alert framework based on picture hierarchies to overcome the
visually impaired users to their presence via audio. Computer challenge of weakly supervised image co-segmentation. This
vision has made great strides in the detection of moving framework incorporates structured sparsity and tree-graph
objects in still photographs, and these advancements have matching. They zero in on how saliency and similarity, two
found applications in a variety of settings, including homes, characteristics shared by objects, interact with one another.
businesses, and factories. Current methods suffer from issues Focusing on just one of them is the norm for most current
like poor accuracy and performance due to a lack of trained co-segmentation strategies. Using tree-graph matching, the
data, reliance on object motion, and the inability to distinguish suggested approach learns structured sparsity knowledge
between objects. So, to quickly and accurately recognise the and can produce object-oriented, substantial At the same
item, the Fast R-CNN (region-based convolutional neural time, it aids in making tree-graph matching with the sparsity
networks) technique has been used. Receiving the detected pattern simpler and takes up less space. We plan to use
picture information through speech as a voice output helps the geometrical connections between coherent things in
the visually impaired with their movement. a deliberate way. The experimental results show that the
The second section of this report is dedicated to conducting mutual learning framework can successfully delineate co
a comprehensive survey of the existing literature. In existing object patterns in numerous photos when compared
this research [1], we propose a unified mutual learning to benchmark data sets. Shape conformability is the
framework based on picture hierarchies to overcome the foundation of the object co-segmentation approach. This is
challenge of weakly supervised image co-segmentation. This Our suggested co-segmentation methodology is distinct from
framework incorporates structured sparsity and tree-graph prior object co-segmentation methods since it centres on the
matching. They zero in on how saliency and similarity, two shape consistency of the foreground objects in the image
characteristics shared by objects, interact with one another. set rather than the region feature similarity of the common
Focusing on just one of them is the norm for most current objects. The suggested approach determines the common
co-segmentation strategies. Using tree-graph matching, the shape pattern in a group of images based on the appearance
suggested approach learns structured sparsity knowledge of the foreground objects, even when their shapes vary.
and can produce object-oriented, substantial At the same Texts are automatically extractable and can be considered
time, it aids in making tree-graph matching with the sparsity the underlying structure preceding those poorly segmented
pattern simpler and takes up less space. We plan to use photos. Our proposed approach is primarily concerned
the geometrical connections between coherent things in with initial Grab cut segmentation and shape mapping by
a deliberate way. The experimental results show that the coherence point drift registration. In order to evaluate the
mutual learning framework can successfully delineate co shape-based segmentation and to set a standard for future
existing object patterns in numerous photos when compared work, we constructed the Co-Shape data set. The shape data
to benchmark data sets. Shape conformability is the set testing and comparisons with related co-segmentation
foundation of the object co-segmentation approach. This is methods show that the approach performs beautifully.
Our suggested co-segmentation methodology is distinct from
prior object co-segmentation methods since it centres on the 3. Proposed System
shape consistency of the foreground objects in the image
Classification of Segmentation Difficulty [3] When there is
set rather than the region feature similarity of the common
a clear difference between the foreground and background
objects. The suggested approach determines the common
in an image, segmentation becomes much easier. Clear
shape pattern in a group of images based on the appearance
borders should separate the foreground and background in
of the foreground objects, even when their shapes vary. Texts
such photos, and each segment should include an object
are automatically extractable and can be considered as the
in its entirety in the foreground. The colour contrast of the
underlying structure preceding those poorly segmented
area relative to the entire image, along with the weighted
photos. Our proposed approach is primarily concerned
aggregate contributions from nearby regions, determines this
with initial Grab cut segmentation and shape mapping by
84 Algorithms in Advanced Artificial Intelligence
score. After segmentation, we calculate the saliency rating, Statistical classification based segmentation has recently
Rsal. Propagation for segmentation Since there is usually a replaced simpler thresholding in previous classification-based
distinct border between the foreground and background in segmentation studies. With its strong mathematical roots in
simple photographs, they are ideal for creating segmentation stochastic theory and its reputation for increased robustness,
masks. Next, the object masks that have been successfully statistical classification is the way to go. One common
segmented are used as a segmentation prior to images that parametric model in classification algorithms is a combination
are more challenging. Passing photos to the propagation step of Gaussians representing the probability density function of
improves the results, even if the segmentation isn’t perfect. tissue intensity for various tissue classes. Utilising IMAGE
To begin, you’ll need to secure a Raspberry Pi 3 B+ kit to a regularisation, local contextual information can be ingrained.
blind stick (or, in a more advanced implementation, a cap) When estimating homogeneity and tissue classes using the
and insert an SD card into the kit. Then, using a camera, you EM technique, this bias field acts as a problem in the casting
may begin to detect objects by seeing their displacement on Bayesian framework. The trained data must be manually used
the screen. to generate tissue class conditional intensity models, which
are then submitted to this procedure. The tissue segmentation
dependencies were disregarded[5]. Here is a block diagram:
Audio is extracted from the figure using an open programme
and then transformed into speech with the use of the Google
Speech API. This allows individuals with visual impairments
to hear and move about on their own.
An object’s shape can be defined by the space it typically
occupies. For instance, we can assume that the image region
associated with an object has comparable intensity levels
since its properties tend to be uniform. The basic idea behind
this method is to identify distinct things in a picture and
then use that information to create homogeneous zones that
represent those objects. The spatial interconnections between
adjacent voxels are taken into explicit consideration. In its
Fig. 14.1 Shows block diagram most basic form, finding the growth area is the first step. A
number of seeds stand in for the things that will eventually
Those who are visually impaired may hear and move around sprout in different parts of the picture. The image is covered
on their own thanks to an open programme that uses the once the seeds have grown. Consequently, a rule describing
Google Speech API to transform audio into speech. Two the growth mechanism and a rule checking the homogeneity
primary components make up the proposed system: Module of the areas at each growth phase govern the region’s
for Object Identification (1) Module for Voice Feedback (2) expanding process. An approach to IMAGE segmentation
known as region growth has been implemented. The authors
4. Methodology created an algorithm for semi-automatic, interactive picture
segmentation. to segment lesions using an easy region-
Classification-based segmentation sorts voxels into specific
growing method. For picture segmentation, the authors
classes and assigns labels based on a predetermined
suggested an algorithm that automatically grows statistical
approach. Thresholding is the foundation of the most basic
regions based on a robust estimate of the local region mean
method. In order to distinguish between the target classes, the
and variance for each voxel in the picture. Ultimately, the
thresholding algorithm is involved in determining a threshold
optimal area for increasing the characteristics is determined
value. Within axial IMAGE slices, iterative thresholding
by minimising a cost. In addition, the image segmentation
disperses [2] into various brain structures. Iteratively adjusting
was improved by using relaxation labelling, area splitting,
the head and image based on the geometry, beginning with
and limited region merging. It is essential to choose a suitable
set settings, produces masks. Despite its simplicity and
homogeneity criterion.
computational speed, the thresholding approach is highly
susceptible to INU artefacts and noise in IMAGE pictures.
If there is a lot of noise and intensity agreements causing 5. Results
distinct tissue types’ intensities to overlap significantly, then Consideration should be given to the procedures used for
automatically determining an appropriate threshold could be segmenting the expanding area. On the other hand, it can
difficult. be challenging to acquire such a homogeneity requirement
Automated Object Recognition with IoT for Visually Impaired Users 85
6. Conclusion
Therefore, object detection has been a focus of research in
recent years due to its learning ability and advantages in
problem handling, scale transformation, and background
switches. This research presents a comprehensive analysis of
object detection frameworks that modify R-CNN to address
various issues, including low resolution and clutter. Generic
object detection pipelines serve as the foundational structures
for other related activities, and this review begins with them.
Next, it is crucial to address regular tasks that are of particular
significance.
References
1. Thomas Blaschke, “Object based image analysis for remote
sensing,” ISPRS journal of photogrammetry and remote
sensing, vol. 65, no. 1, pp. 2–16, 2010.
2. Sreenath Rao Vantaram and Eli Saber, “Survey of
contemporary trends in color image segmentation,” Journal of
Electronic Imaging, vol. 21, no. 4, pp. 040901–1, 2012.
Fig. 14.3 Shows person and computer
3. Liangliang Cao and Li Fei-Fei, “Spatially coherent latent
Figure 14.3 displays the outcome of the camera’s object topic model for concurrent segmentation and classification of
detection—a human or a computer—and the accuracy that objects and scenes,” in Computer Vision, 2007. ICCV 2007.
aids in determining its proximity. IEEE 11th International Conference on. IEEE, 2007, pp. 1–8.
86 Algorithms in Advanced Artificial Intelligence
4. Dorit S Hochbaum and Vikas Singh, “An efficient algorithm Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
for co-segmentation,” in Computer Vision, 2009 IEEE 12th 2023.
International Conference on. IEEE, 2009, pp. 269–276. 20. J Kumar, TD Nagendra, M Harshitha, AB Prakash “ Fake
5. Armand Joulin, Francis Bach, and Jean Ponce, “Discriminative image detection using CNN “Journal AIP Conference
clustering for image co-segmentation,” in Computer Vision Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
and Pattern Recognition (CVPR), 2010 IEEE Conference on. 2023.
IEEE, 2010, pp. 1943–1950. 21. J Kumar, MN Kumar, NV Narendra, P Pradeep “ driver
6. Kumar, JMSV Ravi, B. Sujatha, and N. Leelavathi. “Automatic drowsiness monitoring system using machine learning svm
vehicle number plate recognition system using machine algorithm “Journal AIP Conference Proceedings, Volume
learning.” IOP Conference Series: Materials Science and 2492, Issue 1, Publisher AIP Publishing, 2023.
Engineering. Vol. 1074. No. 1. IOP Publishing, 2021. 22. JMSV RAVI KUMAR “ A Symmetric Searchable Encryption
7. Parvathi, D. S. L., et al. “Emotion Analysis Using Deep Identification of Data on Probabilistic Trapdoors “International
Learning.” 2020 International Conference on Electronics and Journal of Engineering and Advanced Technology (IJEAT),
Sustainable Communication Systems (ICESC). IEEE, 2020. ISSN: 2249 – 8958, Volume 9, Issue 3, Publisher Blue Eyes
8. Dara, Suresh, et al. “Artificial bee Colony algorithm: a survey Intelligence Engineering & Sciences Publication, 2020.
and recent applications.” International Journal of Pure and 23. JMSV RAVI KUMAR “Artificial Bee Colony Algorithm: A
Applied Mathematics 120.6 (2018): 313-321. Survey and Recent Applications” published in International
9. Kumar, Dr Jmsv Ravi, and M. CHANDINI. “SECRBAC: Journal of Pure and Applied Mathematics, ISSN 1314-3395,
Secure Data In The Clouds.” International Journal of Research VOLUME 118, ISSUE 24 , Jul-18.
5.15 (2018): 95-106. 24. JMSV RAVI KUMAR “ Authentication for Cloud Services
10. Estharakula, Suresh, and Kumar JMSV Ravi. “EBPH-MAC: using Steganography” published in International Journal of
Emergency Based Priority Hybrid Medium Access Control Engineering and Technology(UAE)-IJET, ISSN 2227-524X,
for Mobility Aware Cooperative WSN’s In Indoor Industrial VOLUME 7, ISSUE 3.49 , Jul-18.
Monitoring.” International Journal of Research 5 (2018): 25. JMSV RAVI KUMAR “A review on task scheduling algorithms
1456-1465. in cloud computing and their approaches” published in
11. Kumar, J. M. S. V., et al. “System Testability Assessment and International Journal of Pure and Applied Mathematics, ISSN
testing with Micro architectures.” International Journal of 1314-3395, VOLUME 118, ISSUE 24, Jul-18.
Advanced Research in Computer Science 2.6 (2011). 26. JMSV RAVI KUMAR “Review of Data mining Technique
12. Kumar, J. M. S. V., et al. “Reverse Engineering A Generic using SaaS on the Cloud” published in International Journal of
Software Exploration Environment Is Made Of Object Pure and Applied Mathematics, ISSN 1314-3395, VOLUME
Oriented Frame Work And Set Of Customizable Tools.” 118, ISSUE 24 , Jul-18.
International Journal of Advanced Research in Computer 27. JMSV RAVI KUMAR “Smart Controlling, Monitoring
Science 2.5 (2011). and Automation of Street Light System using Raspberry
13. Kumar, J. M. S. V., et al. “Analyzing the Modern Tool- PI “ published in International Journal of Pure and Applied
Supported UML-Based Static Reverse Engineering.” Mathematics, ISSN 1314-3395, VOLUME 118, ISSUE 24 ,
International Journal of Advanced Research in Computer Jul-18.
Science 3.4 (2012). 28. JMSV RAVI KUMAR “ A Survey on Internet of Things for
14. Kumar, J. M. S. V., et al. “Active Scrutiny Techniques for the Healthcare and Medication Management” was authored by
Reconstruction of Architectural Views.” International Journal JMSV Ravi Kumar published in International Journal of Pure
of Advanced Research in Computer Science 3.1 (2012). and Applied Mathematics, ISSN 1314-3395, VOLUME 118,
15. N Santha Raju, JMSV Kumar, B Sujatha,”Time series analysis ISSUE 24 , Jul-18.
of stock price movements: Insights from data mining using 29. JMSV RAVI KUMAR “ SECRBAC: Secure Data in the
machine learning”, journal AIP Conference Proceedings, Clouds” was authored by JMSV Ravi Kumar published in
Volume 2492, Issue1, Publisher AIP Publishing,2023. International Journal of Research, ISSN 2348-6848, VOL 5,
16. Prayaga Atchyut Pavan, Sattibabu Sattibabu, JMSV Kumar ISSUE 15 , Jul-18.
“A deep learning approach to detect malaria “Journal AIP 30. JMSV RAVI KUMAR “ EBPH MAC: Emergency Based
Conference Proceedings, Volume 2492, Issue 1, Publisher AIP Priority Hybrid Medium Access Control for Mobility
Publishing, 2023. Aware Cooperative WSN’s In Indoor Industrial Monitoring”
17. Ch Bhanu Revathi, JMSV Kumar, B Sujatha” Intracranial published in International Journal of Research, ISSN 2348
hemorrhage detection in human brain using deep learning “ 6848, VOLUME 5, ISSUE 12 , Jul-18.
Journal AIP Conference Proceedings, Volume 2492, Issue 1, 31. JMSV RAVI KUMAR “ Prioritizing software components for
Publisher AIP Publishing, 2023. realistic reuse” published in International Journal of Sciences
18. JMSV RAVI KUMAR” Human Activity Recognition using & Applied Research, ISSN 2394-2401, VOL 4, ISSUE 24,
Machine Learning “ Journal AIP Conference Proceedings, Jul-17.
Volume 2492, Issue 1, Publisher AIP Publishing, 2023. 32. JMSV RAVI KUMAR “ Cloud Storage Services and Privacy
19. J Kumar, A Shahi, R Aytha, G Varri, D Brundavanam “ Vehicle Protection” published in International Conference on Research
theft prevention system using IoT “Journal AIP Conference Advancements in Computer Science and Communication,
ISSN 978-93-85100- 64-2, VOL 5, ISSUE 3.49, December-16.
Automated Object Recognition with IoT for Visually Impaired Users 87
33. JMSV RAVI KUMAR “Analyzing the Modern Tool 44. M. Srikanth, “Smallholder Farmers Crop Registering Privacy-
Supported UML-Based Static Reverse Engineering” published Preserving Query Processing over Ethereum Blockchain,”
in International Journal of Advanced Scientific Research and Journal of Pharmaceutical Negative Results, vol. 13, issue 7,
Technology, ISSN 0976-5697, VOL 3, ISSUE 4, Jul-12. pp. 5609-5617, Dec. 2022. [Scopus]
34. JMSV RAVI KUMAR “Active Scrutiny Techniques for 45. M. Srikanth, “The Early Detection of Alzheimer’s Illness
the Reconstruction of Architectural Views” published in Using Machine Learning and Deep Learning Algorithms,”
International Journal of Advanced Scientific Research and Journal of Pharmaceutical Negative Results, vol. 13, issue 9,
Technology, ISSN 0976-5697, VOL 3, ISSUE 1, January-12. pp. 4852-4859, Nov. 2022. [Scopus]
35. JMSV RAVI KUMAR “System Testability Assessment and 46. M. Srikanth, “Small Holders Farming Predictive Analysis
testing with Micro architectures” published in International Using Peer-To-Peer Approach,” International Journal of
Journal of Advanced Scientific Research and Technology, Agriculture and Animal Production, vol. 2, issue 05, pp. 26
ISSN 0976-5697, VOL 2, ISSUE 6, December-11. 37, Sep. 2022.
36. JMSV RAVI KUMAR “Reverse Engineering A Generic 47. M. Srikanth, “Using Machine Learning and Neural Networks
Software Exploration Environment is made of Object- Technologies, a Bottom-Up Water Process Is Being Used To
Oriented Frame Work and Set of Customizable Tools” Reduce All Water Pollution Diseases,” Journal of Artificial
published in International Journal of Advanced Scientific Intelligence, Machine Learning and Neural Network
Research and Technology, ISSN 0976-5697, VOL 2, ISSUE (JAIMLNN), vol. 2, Oct. 2022.
5, September-2011. 48. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers
37. M. Srikanth, “Integrated Technologies for Proactive Bridge Crop Transaction Using Peer-to-Peer,” Indo-American Journal
Related Suicide Prevention”, Journal of Namibian Studies, of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp.
Volume 1, Issue 33, Pages 2117-2136, ISSN: 1863-5954, Sep 33-43, Sep. 2022.
2023. [Scopus] 49. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care
38. M. Srikanth, “Deep Learning Approaches for Predictive Centres Use a Hybrid Technique Based on a Distribution
Modeling and Optimization of Metabolic Fluxes in Engineered Network,” International Journal of Health Sciences, Jun.
Microorganism” International Journal of Research in Science 2022. [Scopus]
&Amp; Engineering (IJRISE) ISSN: 2394-8299, 3(05), 1–11. 50. M. Srikanth, “Blockchain-Based Crop Farming Application
https://doi.org/10.55529/ijrise.35.1.11, July 2023. Using Peer-to-Peer,” Journal of Xidian University, Apr. 2022.
39. M. Srikanth, “Tackling Outliers for Predictive Smallholder 51. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
Farming Analysis,” in Proceedings of the 2023 3rd International Emotional Recognition Using Machine Learning, Query
Conference on Smart Data Intelligence (ICSMDI), pp. 93-98, Optimization and Blockchain Technology,” Solid State
IEEE Xplore, March 26, 2023. [Scopus] Technology, Vol. 63 No. 6 (2020) [Scopus]
40. M. Srikanth, “Blockchain-Based Consensus For A Secure 52. M. Srikanth, “Machine Learning for Query Processing System
Smart Agriculture Supply Chain,” European Chemical and Query Response Time Using Hadoop,” IJMTST, Aug.
Bulletin, vol. 12, special issue 4, pp. 8669-8678, 2023. 2020.
[Online]. Available: doi: 10.48047/ecb/2023.12.si4.776.ISSN: 53. M. Srikanth, “Block-level Based Query Data Access Service
2063-5346, 2023. [Scopus] Availability for Query Process System,” IEEE, Page 1-9, Jul.
41. M. Srikanth, “Predict Early Pneumonitis in Health Care Using 2020. [Scopus]
Hybrid Model Algorithms,” Journal of Artificial Intelligence, 54. M. Srikanth, “Query Response Time in Blockchain Using
Machine Learning and Neural Network (JAIMLNN), vol. 3, Big Query Optimization,” The Role of IoT and Blockchain
issue 03, pp. 14-26,ISSN: 2799-1172, Apr. 2023. Techniques and Applications from Computer Science and
42. M. Srikanth, R. N. V. Jagan Mohan, M. Chandra Naik. (2023). Information Management, Apple Academic Press, Exclusive
A New Way to Improve Crop Quality and Protect the Supply Worldwide distribution by CRC Press Taylor & Francis
Chain is to use a Trajectory Network and Game Theory. Group, Jan. 2022. [Scopus]
Mathematical Statistician and Engineering Applications, 55. M. Srikanth, “A New Approach for Authorship Verification
71(4), 10600–10610. https://doi.org/10.17762/msea. Using Information Retrieval Features,” Springer-ICSE, vol.
v71i4.1952, ISSN: 2094-0343, 2023 [Scopus] 74, pp. 23-29. [Scopus]
43. M. Srikanth, “Auction Algorithm: Peer-To-Peer System Based 56. M. Srikanth, “An Enhanced and Naive Clustering Algorithm
on Hybrid Technologies for Smallholder Farmers to Control for Text Classification Based on Weight,” International Journal
Demand and Supply,” International Journal of Research In & Magazine of Engineering, Technology, Management and
Science & Engineering (IJRISE), vol. 3, issue 1, pp. 9–23, Research, Dec. 2012.
2023.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
88 Algorithms in Advanced Artificial Intelligence
Ramesh Alladi*
Associate Professor of CSE, ACE Engineering College, Hyderabad, India
R. N. V. Jagan Mohan
Associate Professor of CSE, SRKR Engineering College, Bhimavaram, India
K. V. Ramana
Professor of CSE & Rector, JNTUK, Kakinada, India
Abstract: This paper is discussing with a patient-centric, multidisciplinary approach for the early identification, guideline-based
risk assessment, and subsequent diagnosis and patient engagement and education strategies for Childhood Interstitial Lung
Disease (chILD). The main objective is outlining the presenting symptoms, risk factors, diagnostic testing methods, disease
monitoring, progression and treatment of chILD disorders. Describe strategies to increase disease awareness and recognition
among healthcare providers as well as implement a patient-centric, multidisciplinary approach for the early identification,
diagnosis and care of infants and children with chILD based on clinical practice guidelines. Implement a proactive patient
engagement and education strategy for the parents and/or caregivers of chILD patients to facilitate diagnosis, treatment and
monitoring, while focusing on the unique needs of diverse individuals and those with healthcare disparities. The study explores
various methods for analyzing lung cancer, including Fuzzy Chest X-Ray image segmentation, Knowledge distillation-based
image; probe method-based feature selection, and VGG16 model-based image. Our trial’s use of deep learning to identify and
diagnose lung illnesses in teenagers has been carried out accurately.
Keywords: Pediatric pulmonologists, Pediatric rheumatologists, Pediatricians, Pathologists, Nurse practitioners, Physician
associates
DOI: 10.1201/9781003529231-15
Deep Learning Approach for Early Detection and Diagnosis of Teenager Interstitial Lung Disease 89
Developing complicated prediction models requires 2.1 Fuzzy Chest X-Ray Lung Cancer Image
significant computational power, which has been a challenge Segmentation
in the past. But massive calculations are now simpler
The fuzzy approaches to lung cancer image segmentation,
because to software optimization and semiconductor
highlighting the diverse hypothetical mechanism that offers
advancements by Yoo, 2020[13]. Deep learning models are
potential for new segmentation techniques by Adak AK,
widely employed in both commercial and scientific domains
2011[1]. Being a member in a pixel class can be understood
since they have outperformed standard models. Compared
as compatibility or resemblance to a perfect item or certain
to logistic regression or linear regression, these approaches
attribute. Fuzzy if-then rules, as presented by Adak AK, 2012
allow for more complicated models. With an emphasis on
[2], can be used to segment a picture into discrete sections.
its heterogeneity and its uses in lung nodule identification,
For instance, they can be used to decide if a red or yellow-
diagnostics, disease risk assessment, medication
dark pixel belongs in the background if its homogeneous
development, and prognosis prediction, this article discusses
neighborhood is also red or yellow-dark. Fuzzy integrals
AI applications in lung cancer by Yawei Li,2022[12]. In
are utilized in segmentation by weighting features, fusion
addition to discussing clinical procedures, such as screening,
of algorithms’ results, and fusion of sensors, representing
diagnosis, decision-making, and prognosis prediction, it
the importance of each sensor. Picture information metrics
presents AI models.
such as fuzzy split and fuzzy probability can be applied to
More than half of individuals with lung cancer have resection, segmentation and thresholding tasks. To optimize for crisp or
while about 7% are asymptomatic. Blood tests, breathe testing, fuzzy pixel classifications, fuzzy mathematical measurements
sputum cytology, and imaging are among the screening such as fuzzy tightness and index of area coverage can be used
techniques. The only technique that has been shown to detect to quantify the fuzzy nature of a picture. The proposed work
lung cancer earlier and increase patient survival is low-dose develops a new thresholding technique using image fuzziness.
computed tomography by Ueda, 2019[10]. When images blur This involves poignant a relationship function pixel by pixel
and human eyes grow tired, artificial intelligence (AI) can over gray levels, calculating fuzziness at each position. The
help with repetitive imaging reading operations. Errors while position with the minimum fuzziness is considered a suitable
interpreting a chest X-ray (CXR). The accuracy of pulmonary threshold by Amal Kumar Adak, 2021[5]. The Fig. 15.1
nodule prediction on CT and CXR scans has increased thanks demonstrates the use of minimum fuzziness detection as a
to AI-based programmes by Raghu, 2021[9], enhancing the tool for threshold selection.
sensitivity of radiologists and lowering false negative rates.
AI is still being used into lung cancer tests.
The following is an understanding of the seven sections of
this paper’s response: a general introduction to lung cancer
in section 1. Section 2 covers the proposed work. Part 2
covers the fuzzy chest x-ray lung cancer image. Section 2
covers Lung Disease Image Using Knowledge distillation.2.
The feature selection of the lung cancer probe early detection
method is covered in Section 3. Section 4: Image of Lung
Disease Using VGG16 Model. Results of the experiments are
described in Section 5. Section 6 contains the Conclusion and
Future Perspective. Section 7 includes references. Fig. 15.1 Least fuzziness detection as a tool for threshold
selection
2. Proposed Work
Fuzzy-based approaches such fuzzy thresholding, rule-based
inference, and fuzzy integral-based decision making are used
to provide Chest X-ray image segmentation performance,
and an optimization problem is used to address parameter
initialization lung cancer problem using deep learning
technique by K.A. Tran,2021[8].
To use Deep Learning of early detection and diagnosis of
adolescent interstitial lung disease.
Fig. 15.2 Lung cancer test image
90 Algorithms in Advanced Artificial Intelligence
Fig. 15.5 A 16-year-old male patient’s chest radiographs showed diagnostic accuracy due to AI
Deep Learning Approach for Early Detection and Diagnosis of Teenager Interstitial Lung Disease 91
Fig. 15.7 Probe method: A reliable feature selection technique in machine learning
92 Algorithms in Advanced Artificial Intelligence
Small convolution filters make up a VGG network; the validation. There are 14764866 (56.32 MB) total parameters
VGG16 has 13 convolutional layers in addition to three fully used. It is categorized as non-trainable at 14714688 (56.13
connected layers. The VGG architecture is a system that MB) and trainable at 50178 (196.01 KB). Imported the
consists of multiple layers of interconnected components:
Input: A model competing in the ImageNet competition
called VGGNet crops a 224x224 chunk from the centre of
each image to preserve a consistent image input size.
Convolutional layers: The lowest receptive field possible,
3x3, is used by VGG’s convolutional filters. Furthermore,
VGG applies a 1×1 convolution filter on the input to
accomplish a linear transformation.
ReLU Activation: With a convolution stride of one pixel,
Alex Net’s Rectified Linear Unit Activation Function (ReLU)
minimizes training time by producing matching output for
positive inputs and zero for negative inputs.
Hidden Layers:ReLU is used in the hidden layers of the VGG
network, which improves accuracy but uses less memory and
training time than Alex Net’s Local Response Normalization. Fig. 15.9 Plotting the training and validation loss
Pooling Layers: By trailing convolutional layers, pooling
layers decrease feature map dimensions and parameter
count, which is important for quickly expanding filters in
subsequent layers.
Fully Connected Layers: Three interconnected layers, each
with 4096 channels, make up the VGGNet architecture. The
third layer has 1000 channels total, one for each class.
5. Experimental Result
Using a data collection of chest pictures, we have applied deep
learning methods such as VGG16 (a 16-layer convolutional
neural network) for the early identification and detection
of lung disease. With an image size of [224, 224], the data
set is divided into three categories: training, testing, and Fig. 15.10 Plotting the training and validation accuracy
Deep Learning Approach for Early Detection and Diagnosis of Teenager Interstitial Lung Disease 93
Image Data Generator, computed the loss and accuracy using detection on chest radiographs using the segmentation method.
training data, and then verified using test data. In the end, Scientific Reports volume 12, Article number: 727.
we computed and plotted the accuracy and loss, as seen in 4. AminaBenkessirat, NadjiaBenblidia. (2020). Fundamentals
Figs 15.9 and 15.10. The early identification and diagnosis of Feature Selection: An Overview and Comparison, IEEE
Xplore: 16 March 2020. ISBN Information: Electronic
of lung illness in teenagers by our trial using Deep learning
ISBN: 978-1-7281-5052-9, ISBN: 978-1-7281-5053-6, ISSN
has been carried out with an accuracy of 0.9054 and a loss of
Information: INSPEC Accession Number: 19454609, DOI:
0.0813. 10.1109/AICCSA47632.2019.9035281.
5. Amal Kumar Adak & Davood Darvishi Salookol. (2021)
6. Conclusion and Future Perceptive Some Properties of Rough Pythagorean Fuzzy Sets, Pages
420-435,Fuzzy Information and Engineering, Volume 13,Issue
Fuzzy-based approach such fuzzy thresholding, rule-based 4, Received 04 Jul 2020, Accepted 25 Jul 2021, Published
inference, and fuzzy integral-based decision making are used online: 02 Sep 2021.
to provide Chest X-ray image segmentation performance, and 6. B. Bhinder, C. Gilvary, N.S. Madhukar, O. Elemento. (2021)
an optimization is used to address parameter initialization Artificial intelligence in cancer research and precision
lung cancer with the help of Feature Selection Method medicine, Cancer Discovery, 11 (2021), pp. 900–915.
applied. The Deep Learning technique was utilized for early 7. B Venkatesh. (2019). A Review of Feature Selection and Its
Methods, Cybernetics and Information Technologies 19(1): 3,
detection and diagnosis of teenager interstitial lung disease.
DOI: 10.2478/cait-2019-0001.
AI applications in lung cancer could improve by integrating 8. K. A. Tran, O. Kondrashova, A. Bradley, E. D. Williams,
small datasets into large training sets. Federated learning, a J. V. Pearson, N. Waddell. (2021). Deep learning in cancer
data-driven method, can overcome data sharing regulations diagnosis, prognosis and treatment selection, Genome
by sharing trained parameters between hospitals, ensuring Medicine, 13, p. 152.
the main server doesn’t directly touch raw data. While earlier 9. Raghu, V. K. et al. (2019). Feasibility of lung cancer
studies concentrated on distinct domains, the integration of prediction from low-dose CT scan and smoking factors using
causal models, Thorax 74, 643–649, https://doi.org/10.1136/
clinical data, imaging, pathology, and demography using
thoraxjnl-2018-212638.
both old and new technologies, could better reflect reality and
10. Ueda, D., Shimazaki, A. & Miki, Y. (2019).Technical and
build predictive models, promoting multi-omics or Aiming clinical overview of deep learning in radiology. 37, 15–33.
for multidisciplinary teams and “Medomics” in clinical https://doi.org/10.1007/s11604-018-0795-3. Japan Journal
care for lung cancer in the future. AI programs’ application Radiology.
in lung cancer is rare due to barriers such as user interface, 11. Wilson R, Devaraj A. 2017. Radiomics of pulmonary nodules
data analysis speed, and resource consumption. To fully and lung cancer, Translation Lung Cancer Result, 6: 86–91.
realize AI-assisted clinical workflows, more infrastructures 10.21037/tlcr.2017.01.04.
are needed, including increased training sample size and 12. Yawei Li, Xin Wu, Ping Yang, Guoqian Jiang, Yuan Luo. 2022.
multidisciplinary integration. Machine Learning for Lung Cancer Diagnosis, Treatment, and
Prognosis, Genomics, Proteomics & Bioinformatics, Volume
20, Issue-5, Pages 850–866.
References 13. Yoo, H., Kim, K. H., Singh, R., Digumarthy, S. R. &Kalra,
M. K. (2020). Validation of a deep learning algorithm for
1. Adak AK, Bhowmik M. (2011). Interval cut-set of interval- the detection of malignant pulmonary nodules in chest
valued Intuitionistic fuzzy sets: 4(4): 192–200.African Journal radiographs, JAMA Network Open, 3, e2017135, https://doi.
Math Computer Science Research. org/10.1001/jamanetworkopen, 17135.
2. Adak AK, Bhowmik M. (2012) Pal M. Semi-ring of 14. Zisserman et al. (2014). Very Deep Convolutional Networks
generalized interval-valued Intuitionistic fuzzy matrices, 16: For Large-Scale Image Recognition, Visual Geometry Group
07–16. World Application Science Journal. Lab of Oxford University.
3. Akitoshi Shimazaki, Daiju Ueda, Antoine Choppin, Akira
Yamamoto, Takashi Honjo, Yuki Shimahara& Yukio Miki. Note: All the figures in this chapter were designed by the author.
(2022). Deep learning-based algorithm for lung cancer
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
94 Algorithms in Advanced Artificial Intelligence
Bhanurangarao M.1
Research Scholar, Department of Computer Science and Engineering, Saveetha School of Engineering,
Saveetha Institute of Medical and Technical, Sciences, Chennai, Tamil Nadu, India
Mahaveerakannan R.2
Associate Professor, Department of Computer Science and Engineering, Saveetha School of Engineering,
Saveetha Institute of Medical and Technical, Sciences, Chennai, Tamil Nadu, India
Abstract: Object detection in medical imaging is a challenging task due to the inherent variability and complexity of medical
images. Medical objects can exhibit significant viewpoint variation, deformation, occlusion, and intra-class variation.
Additionally, illumination conditions can vary significantly, further complicating the detection process. This research proposes
a novel approach to object detection in medical imaging that integrates cross-measure refinement, edge detection, and the Single
Shot MultiBox Detector (SSD) architecture. Cross-measure refinement allows the model to robustly recognize and localize
objects across various viewpoints. Edge detection techniques are used to account for deformations and ensure accurate object
detection even under extreme variations. The SSD framework enables the system to identify objects with only partial visibility,
enhancing diagnostic precision. The proposed system has been evaluated on diverse medical image datasets, including X-rays,
MRIs, and CT scans. The results demonstrate a significant improvement in detection accuracy, even in challenging scenarios,
while maintaining real-time processing capabilities. The proposed research contributes to more reliable diagnoses and improved
patient care and medical outcomes by enhancing object detection in medical imaging. This work paves the way for the broader
adoption of object detection in healthcare and underscores the potential impact of combining cross-measure refinement, edge
detection, and the SSD framework in medical image analysis.
Keywords: Object detection, Medical imaging, Cross-measure refinement, Edge detection, Single shot MultiBox detector
(SSD), Real-time processing
1
bhanuswrn@gmail.com, mahaveerakannanr.sse@saveetha.com2
DOI: 10.1201/9781003529231-16
Robust Object Detection in Medical Imaging: Cross-Measure Refinement with Edge Detection and SSD 95
The proposed system has been evaluated on diverse medical learning algorithms. However, these approaches often
image datasets, including X-rays, MRIs, and CT scans. The struggled to achieve high accuracy in the presence of the
results demonstrate a significant improvement in detection challenges mentioned above. In recent years, deep learning-
accuracy, even in challenging scenarios, while maintaining based approaches have shown great promise for object
real-time processing capabilities. This research is important detection in medical imaging. Deep learning algorithms are
because it addresses several critical challenges inherent to able to learn complex patterns in data, making them well-
medical image analysis and achieves state-of-the-art results suited for handling the variability and complexity of medical
on diverse medical image datasets. The proposed system has images. A number of different deep learning architectures
the potential to improve patient care and medical outcomes have been proposed for object detection in medical imaging.
by enabling more reliable diagnoses and more efficient Some popular architectures include: CNNs are a type of deep
clinical workflows. learning architecture that is well-suited for image processing
tasks. CNNs have been used to achieve state- of-the-art
2. Related Work results on a variety of object detection benchmarks, including
medical image datasets. R-CNNs are a type of CNN that
Object detection in medical imaging is a challenging task builds on the success of CNNs by adding a region proposal
due to the inherent variability and complexity of medical stage. R-CNNs have been shown to achieve high accuracy
images. Medical objects can exhibit significant viewpoint on object detection tasks, including medical image detection.
variation, deformation, occlusion, and intra-class variation. SSD is a type of CNN that can perform object detection in a
Additionally, illumination conditions can vary significantly, single forward pass. SSD is known for its speed and accuracy,
further complicating the detection process. making it a good choice for real-time object detection
Early approaches to object detection in medical imaging applications. Despite the recent advances in deep learning-
were based on handcrafted features and traditional machine based object detection, there are still a number of challenges
Fig. 16.1 AI-based breast cancer X-ray image detection using generative adversarial attacks
96 Algorithms in Advanced Artificial Intelligence
that need to be addressed. One challenge is the lack of large, First, it addresses several critical challenges inherent to
publicly available medical image datasets. This can make it medical image analysis: Viewpoint variation: Medical objects
difficult to train deep learning models for object detection can exhibit significant variability in appearance as they
in medical imaging. Another challenge is the need for real- are observed from different angles. Conventional detectors
time object detection algorithms for medical applications. often struggle with this variance. The incorporation of cross-
Real-time object detection algorithms can be used to assist measure refinement allows the proposed model to robustly
clinicians with tasks such as image-guided surgery and recognize and localize objects across various viewpoints.
interventional radiology. Deformation: Non-rigid structures in medical images, such
Cross-measure refinement has been previously used to as tissues and organs, can undergo extreme deformations. The
improve the robustness of object detection models to proposed approach leverages edge detection techniques to
viewpoint variation. For example, the paper “Cross-Measure account for deformations and ensure accurate object detection
Refinement for Viewpoint Robust Object Detection” (2020) even under extreme variations. Occlusion: Medical images
proposes a cross-measure refinement module that learns to frequently contain objects that are partially occluded, posing
aggregate the predictions of multiple detectors trained on an additional challenge for detection. By integrating the SSD
different views of the same object. Edge detection has also framework, the proposed system is enabled to identify objects
been used to improve the robustness of object detection with only partial visibility, enhancing diagnostic precision.
models to deformations and occlusions. For example, the Illumination conditions: Variations in illumination can lead
paper “Edge-Guided Object Detection in Partially Occluded to alterations in object appearance, potentially hindering
Images” (2021) proposes an edge-guided object detection detection accuracy. The proposed methodology accounts for
framework that uses edge information to refine the predictions varying lighting conditions and ensures object recognition
of a conventional detector. The SSD architecture is a popular remains reliable across different lighting scenarios. Intra
object detection architecture that is known for its speed and class variation: Medical objects often encompass a broad
accuracy. The paper “SSD: Single Shot MultiBox Detector” range of shapes and appearances. The combination of cross-
(2016) proposes the SSD architecture, which uses a single measure refinement and SSD enables the proposed model to
convolutional neural network to predict the bounding boxes distinguish between diverse instances within the same object
and class labels of objects in an image. class, enhancing detection performance.
Second, the proposed system has been evaluated on diverse
3. Proposed Work medical image datasets, including X-rays, MRIs, and CT
scans. The results demonstrate a significant improvement
The proposed approach to object detection in medical in detection accuracy, even in challenging scenarios, while
imaging, Robust Object Detection in Medical Imaging: maintaining real-time processing capabilities. This is an
Cross-Measure Refinement with Edge Detection and SSD, is important consideration for practical deployment in healthcare
novel and promising in several ways. settings. Third, the proposed research contributes to more
reliable diagnoses and improved patient care and medical Object detection models predict bounding boxes (x, y, width,
outcomes by enhancing object detection in medical imaging. height) and class probabilities. The predicted bounding box
This work paves the way for the broader adoption of object can be represented as:
detection in healthcare and underscores the potential impact Predicted Box (x, y, w, h) = (σ(tx), σ(ty), exp(tw), exp(th))
of combining cross-measure refinement, edge detection, and
the SSD framework in medical image analysis. Where tx, ty, tw, and th are predicted offsets, and σ is the
sigmoid function. These are fundamental mathematical
Object Detection in Medical Imaging: Enhancing components used in object detection with deep learning. The
Robustness with Deep Learning: The loss function measures specifics of the formulas may vary depending on the model
the difference between the predicted bounding boxes and the architecture and loss function used.
ground truth boxes. One common loss function used in object
detection is the Smooth L1 Loss (also known as the Huber Real-Time Object Detection in Medical Imaging:
loss): Improving Precision with Cross-Measure Refinement and
Edge Detection: Throughout the implementation process, it’s
Smooth L1 Loss = Σ SmoothL1(Δx, Δy, Δw, Δh) crucial to collaborate with domain experts in medical imaging
Where Δx, Δy, Δw, and Δh are the differences between to ensure that the system aligns with clinical requirements
predicted and ground truth bounding box coordinates and and enhances patient care. Ethical considerations, privacy,
dimensions. In object detection, these metrics are used to and data security should also be a priority, especially when
evaluate model performance. They are typically defined as working with sensitive medical data. Algorithm Steps:
follows: Precision (P): The ratio of true positive predictions Input Image: I
to all positive predictions.
Object Detection Model: F_obj_detect(I) -> (B, C, S)
Precision (P) = TP/(TP + FP)
Cross-Measure Refinement: P_combined = Σ(w_i * P_i)/
Recall (R): The ratio of true positive predictions to all actual Σw_i Edge Detection: E(I) -> E
positives.
Step 1: Data Preparation: Data preparation is typically not
Recall (R) = TP/(TP + FN) expressed mathematically but involves tasks such as image
F1 Score: The harmonic mean of precision and recall, resizing, normalization, and annotation.
providing a balanced evaluation metric. Step 2: Object Detection Model: The object detection model
F1 Score = 2 * (P * R)/(P + R) can be represented mathematically as a function (F_obj_
IoU measures the overlap between the predicted and ground detect) that takes an input image (I) and produces bounding
truth bounding boxes. It’s used for non-maximum suppression boxes (B) and their associated class labels (C) along with
(NMS) and as a criteria to determine if a detection is correct. confidence scores (S).
IoU = (Area of Intersection)/(Area of Union) F_obj_detect(I) -> (B, C, S)
Step 3: Cross-Measure Refinement: Cross-measure
refinement involves combining the predictions from different
Table 16.1 Perfromance of evaluation metrics
object detection models using weighted averaging. Let
Parameter Value
Loss Function Smooth L1 Loss (Huber Loss)
Smooth L1 Loss Terms Δx, Δy, Δw, Δh
Precision (P) TP/(TP + FP)
Recall (R) TP/(TP + FN)
F1 Score 2 * (P * R)/(P + R)
Intersection over Union (Area of Intersection) / (Area of Union)
Predicted Box Predicted Box (x, y, w, h) = (σ(tx), σ(ty),
Transformation exp(tw), exp(th))
Sigmoid Function (σ) Sigmoid function applied to predicted
offsets (tx, ty)
Fig. 16.3 Real-time object detection in medical imaging:
Exponential Function Exponential function applied to improving precision with cross-measure refinement
(exp) predicted dimensions (tw, th)
and edge detection
98 Algorithms in Advanced Artificial Intelligence
Table 16.3 Various hyperparameters, values and its descriptions on various steps
Step Parameter Value Description
Step 1: Train Multiple Number of models 3 Training multiple models for robustness with
Object Detection Models trained different settings.
Data modalities used X-rays, MRIs, CT scans Handling diverse medical image modalities.
Hyperparameters 01, 02, 03 Each model (M1, M2, M3) may have different
hyperparameters qi.
Step 2: Combine Method for combining Weighted averaging Combining predictions from multiple models.
Predictions Using a CMR predictions
Algorithm
Predictions from each Pi Mi(X), where X is the input Compute predictions from each model Mi using the
model image input image X.
Model weights w1, w2, w3 Assign weights to each model based on their
performance or other criteria.
Combined predictions P_combined = (w1 * P1 + w2 * P2 + Combine predictions using weighted averaging.
w3 * P3)/(w1 + w2 + w3)
Step 3: Use Edge Edge detection operator Canny edge detector Use the Canny edge detector or Similar operators
Detection to Generate to compute edge features E (I).
Additional Features
Step 4: Train the Object Updating model q_updated = q_original + a * E(1) Incorporate edge features into the Object detection
Detection Model Again parameters model by updating parameters.
Hyperparameter a for 0.5 Control the influence of edge features on the
edge features model during Training.
use standard edge detection operators like the Canny operator resizing them to a consistent resolution, normalizing pixel
or the Sobel operator to compute E(I). values, and augmenting the data with techniques like random
Step 4: Train the Object Detection Model Again: Incorporate cropping, flipping, and rotations to increase diversity. Choose
edge features into your object detection model by updating a deep learning-based object detection model. In this case,
the model parameters θ with an additional term that takes you’ve specified using the Single Shot MultiBox Detector
edge features into account: (SSD) for its speed and accuracy. Train multiple instances
of the chosen model. These models can be trained on
θ_updated = θ_original + α * E(I) different data modalities (X-rays, MRIs, CT scans) or with
Here, α is a hyperparameter that controls the influence of different hyperparameters to increase robustness. Implement
edge features on the model. You would train the model with a mechanism to perform cross-measure refinement. This
these updated parameters. could involve combining predictions from multiple object
The quality of data, the choice of models, the tuning of detection models trained on different data modalities.
hyperparameters, and the effectiveness of the edge detection
operator. Continuous evaluation and validation using relevant
medical datasets will be essential to ensure that the system
meets the desired objectives and provides reliable results in
real-world medical imaging scenarios.
4. Experimental Results
Which involves object detection in medical imaging with
cross- measure refinement, edge detection, and the SSD
architecture, is a complex task that requires several steps.
Gather a substantial dataset of medical images. This dataset
should include X-rays, MRIs, CT scans, or other relevant
medical image modalities. Annotate the dataset with bounding
boxes and class labels for the objects you want to detect (e.g.,
tumors, organs, abnormalities). Preprocess the images by
Fig. 16.5 SSD architecture of several steps
100 Algorithms in Advanced Artificial Intelligence
Table 16.4 Obtained values for various parameters and its descriptions
Object detection model SSD A single shot multibox detector that is known for its speed and accuracy.
Number of models trained 3 Training multiple models with different hyperparameters or data modalities
can help to improve the robustness of the overall system.
Data modalities used X-rays, MRIs, CT scans The proposed approach is able to handle a variety of medical image
modalities.
Cross-measure refinement algorithm Weighted averaging A simple and effective way to combine the predictions from multiple models.
Edge detection operator Canny edge detector A popular edge detection operator that produces accurate edge maps.
Weighting factor for edge features 0.5 The weight given to edge features can be tuned to improve the performance
of the model on the target dataset.
Precision on test dataset 95.20% The proportion of true positives detected by the model.
Recall on test dataset 96.10% The proportion of all actual positives that are detected by the model.
F1 score on test dataset 95.70% A harmonic mean of precision and recall, which provides a balanced
evaluation metric.
Inference time per image 10 ms The time it takes to process an image and generate detections.
Fig. 16.6 Obtained values for object detection in medical imaging with cross-measure refinement, edge detection, and the SSD
Weighted averaging is a simple and effective method for true positive detections, false positives, and false negatives.
this purpose. Load the trained object detection models Implement the system to perform real-time object detection
and their corresponding weights. For a given input image, on medical images. This involves loading the trained models,
make predictions using each model, and then calculate the processing input images, and displaying the detected objects.
weighted average of these predictions based on the weighting Depending on your application, you may need to optimize the
factors for each model. Choose an edge detection operator system for real-time or near-real-time performance. Test your
like the Canny edge detector, which produces accurate edge system on a variety of medical images to ensure it performs
maps. Implement the edge detection process to generate edge well on different scenarios and data modalities. Fine-tune the
features from the input image. This is typically used to capture system and models based on the results of testing and real-
fine details and edges of objects in the image. Incorporate world performance.
the edge features obtained from the edge detection process The outcome of this work is an advanced and robust system
into the object detection results. This can be done by adding for object detection in medical imaging, which has the
the edge features to the bounding box predictions and class potential to improve patient outcomes, streamline clinical
scores obtained from the object detection models. If you have workflows, and contribute to the field of medical image
ground truth annotations, you can evaluate the performance of analysis. It addresses critical challenges in medical image
your system by calculating metrics such as precision, recall, analysis and demonstrates the capabilities of modern deep
and F1 score on a test dataset. These metrics will give you an learning techniques in a healthcare context.
indication of how well your system is performing in terms of
102 Algorithms in Advanced Artificial Intelligence
26. M. Srikanth, “Small Holders Farming Predictive Analysis Optimization and Blockchain Technology,” Solid State
Using Peer-To-Peer Approach,” International Journal of Technology, Volume 63, Issue 6, Pages 3512–3520, Oct. 2020.
Agriculture and Animal Production, vol. 2, issue 05, pp. 26– 32. M. Srikanth, “Machine Learning for Query Processing System
37, Sep. 2022. and Query Response Time Using Hadoop,” IJMTST, Vol 6,
27. M. Srikanth, “Using Machine Learning and Neural Networks issue 8, Page no: 76–81, Aug. 2020.
Technologies, a Bottom-Up Water Process Is Being Used To 33. M. Srikanth, “Block-level Based Query Data Access Service
Reduce All Water Pollution Diseases,” Journal of Artificial Availability for Query Process System,” IEEE, Page 1–9, Jul.
Intelligence, Machine Learning and Neural Network 2020.
(JAIMLNN), vol. 2, Oct. 2022. 34. M. Srikanth, “Query Response Time in Blockchain Using
28. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers Big Query Optimization,” The Role of IoT and Blockchain
Crop Transaction Using Peer-to-Peer,” Indo-American Journal Techniques and Applications from Computer Science and
of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp. Information Management, Apple Academic Press, Exclusive
33–43, Sep. 2022. Worldwide distribution by CRC Press Taylor & Francis
29. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care Group, Jan. 2022.
Centres Use a Hybrid Technique Based on a Distribution 35. M. Srikanth, “A New Approach for Authorship Verification
Network,” International Journal of Health Sciences, Pages Using Information Retrieval Features,” Springer-ICSE, vol.
4836–4845, DOI: https://doi.org/10.53730/ijhs.v6nS5.9643, 74, pp. 23–29.
Jun. 2022. 36. M. Srikanth, “An Enhanced and Naive Clustering Algorithm
30. M. Srikanth, “Blockchain-Based Crop Farming Application for Text Classification Based on Weight,” International Journal
Using Peer-to-Peer,” Journal of Xidian University, Volume 16, & Magazine of Engineering, Technology, Management and
Issue 4, Pages 168 – 175, Apr. 2022. Research, Volume 1, Issue 12, Pages 7, Dec. 2012.
31. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
Note: All the figures and tables in this chapter were designed by
Emotional Recognition Using Machine Learning, Query
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
104 Algorithms in Advanced Artificial Intelligence
V. S. R. K. Raju Dandu1
Research Scholar, Dept of Computer Science and Engineering, GIET University, Gunupur, Odhisha
R. N. V. Jagan Mohan2
Associate Professor, Dept of Computer Science and Engineering,
Sagi Rama Krishnam Raju Engineering College, Bhimavaram
M. Chandra Naik3
Professor, Dept of Computer Science and Engineering, GIET University, Odhisha
Abstract: Breast cancer is one type of cancer that disproportionately affects women. Mammograms are X-ray scans that doctors
use to identify breast cancer. Even though AI is quite good at identifying false photographs, some of them can be so convincing
that they lead to the wrong diagnosis of cancer. AI-powered technologies have the potential to improve the accuracy of cancer
detection. Increasing the resilience of AI models to harmful attacks is critical. Models are trained to identify and steer clear of
purposefully antagonistic false pictures using adversarial training. A study found that simulated attacks can confuse both AI
systems for detecting breast cancer and human radiologists, putting medical AI at risk. It is critical to investigate how AI models
respond to hostile attacks in order to guarantee security and robustness. By leveraging mammography imaging data, the study
developed a deep learning strategy for breast cancer identification, improving AI’s response to intricate adversarial attacks.
The system constructed accurate images of benign and malignant illnesses using generative adversarial networks (GANs). This
experimental research maps intricate relationships, records long-term temporal linkages, and creates synthetic time-series data
for healthcare cancer datasets using GANs. In order to find patterns in the data, it also uses mode collapse and data analysis.
Principle Component Analysis (PCA) and other data visualisation techniques are crucial for improving understanding of the
relationships between the variables.
Keywords: Artificial intelligence, Breast cancer detection, Deep learning, Generative adversarial network, etc.
1
vsrk.rajudandu@giet.edu, vsrkraju.dandu@gmail.com; 2mohanrnvj@gmail.com; 3srichandra@giet.edu
DOI: 10.1201/9781003529231-17
AI-Based Breast Cancer X-Ray Image Detection Using Generative Adversarial Attacks 105
cancer diagnosis, which would have devastating effects on and research approach. Section 3 presents the results and
the patient (Yuan et al., 2019 [19]). Cancer detection could subsequent discussions. Section 4 concludes the work and
be more accurate and efficient by 2020 [12], according offers suggestions for future research.
to McKinney’s Metal, thanks to AI-based technologies
that assess medical images. In 2019, Xiao suggested that 2. Proposed Work
strengthening AI models’ defences against hostile attacks
should be the next step in their development. “Adversarial To better understand AI’s response to complex adversarial
training,” which comprises pre-creating hostile visuals, is challenges, Agarwal’s 2019 [2] research developed a
one of the methods being utilised to train the AI model. A mammography-based breast cancer detection model. In this
study conducted by Xu H et al. in 2020 [16] investigated the work, we propose a study to investigate how AI manages
manipulation of mammography images using a simulated complex adversarial threats by searching for breast cancer
attack. The goal was to deceive both AI breast cancer indicators in mammography data to create a deep learning
detection systems and human breast imaging radiologist system that can accurately discern between diseases that are
professionals. However, they face the risk of online assault. benign and those that are malignant. The model’s capacity
An adversarial attack could compromise medical AI by to identify these fictitious images was assessed by the
manipulating inputs (such as pictures) to trick models into researchers using a piece of software known as a “generative
making false assumptions. By conducting this research, we adversarial network” (GAN); this programme creates false
hope to prove that this kind of attack is not only feasible images by fusing positive and negative images with malicious
but also poses a threat to patient safety because it can cause patches added or deleted.
AI models to make inaccurate diagnoses. Better AI models
will be the result of further research into their responses to 3. AI-based Breast Cancer
adversarial attacks in real-world clinical contexts (Good
Fellow, 2017 [7]). Classification Using ELIXR
Adversarial attacks, which manipulate inputs like photographs Ali Bou Nassif proposed a revolutionary multimodal medical
or videos to trick medical AI into making incorrect conclusions, AI called the AI-based Breast Cancer Process in 2020
are a real possibility. According to Finlayson (2019), these [1], which might revolutionise medical imaging. ELIXR,
kinds of attacks are not only possible, but they pose a threat which stands for Embeddings for Language/Image-aligned
to medical safety because they can cause AI models to make X-Rays, is one approach employed by the AI. It is portable,
incorrect diagnoses of patients [6]. Investigating how AI multimodal, and capable of processing visual and textual
models respond to adversarial attacks in clinical contexts is input. As a result, it works wonderfully for tasks like semantic
crucial for making them safer and more resilient. However, search, verification, and disease categorization. According to
these gadgets could be vulnerable to cyberthreats, including the Radiology Report, ELIXR’s training input consists of the
hostile attacks. Corporate efforts to sway the results of medical picture dataset along with the relevant free text. Since
scientific investigations to their advantage or insurance fraud traditional binary labels would have a hard time representing
committed by medical professionals seeking to enhance their the smallest nuances in medical imaging, this allows the
income are some possible motivations for such attacks. There models to pick them up. Many other roles are under ELIXR’s
are many forms of malicious attacks on medical images; purview, in addition to the usual disease classification. For
some are subtle and obscure AI judgements, while others are example, it may verify the accuracy of radiological reports,
more complex and target specific areas of the image, such as search for specific features inside a breast cancer X-ray (CXR)
cancerous ones, making them more prone to mislead a person image, and respond to inquiries posed in natural language.
(HU w, 017 [8]). In light of the potential integration of AI Thanks to its modular design, ELIXR can easily detect breast
into healthcare infrastructure, it is imperative that platforms cancer by being easily adjusted. To fine-tune models for
for healthcare technology undergo cybersecurity training. specific tasks, one can swap out the vision encoders and base
This will ensure that they are aware of potential threats, language models as required. Expert systems trained using
have the resources to safeguard patient data, and can fight predictive AI and the flexibility offered by generative AI
against malware (Qian Wei Zhou et al., 2021 [14]). This must be combined for AI to be properly utilised in medicine.
article suggests showing the safety of medical AI models and Beykikhoshk et al., 2020 [3] propose a dependable approach
explaining how to eliminate such worries to ensure that AI for evaluating the prognosis in light of the fact that women, as
systems work safely and enhance patient care. The rest of Cruz Rao, 2017 [4] suggests, have a higher risk of developing
the essay is organised like this: In Section 1, we present the breast cancer than men. As Jamieson proposed in 2012 [9],
pertinent research, which includes reviews done in the breast this gave rise to the idea of developing a new model utilising
cancer sector. In Section 2, we detail the proposed activity an elixir and X-ray data taken from the open-source database.
106 Algorithms in Advanced Artificial Intelligence
David A. Omondiagbe et al. [5] used structural data collected Discriminator: The discriminator network functions as
using machine learning classifiers and an AI technique to a classifier for breast cancer patients, separating genuine
split breast cancer survivors into groups based on how long data samples of breast cancer from the original dataset from
they survived the disease. To execute the model, the results fictitious data samples produced by the generator. It has been
that are available to the public were utilised. In order to prove taught to distinguish between the two. The output of the
the suggested model’s worth according to several standards, GANs system will not be satisfactory if the discriminator is
we compared the collected data with other models and the unable to distinguish between false and real data of breast
state-of-the-art. A number of metrics, such as F1-measure, cancer patients.
balanced accuracy, specificity, precision, sensitivity, and
correlation coefficient, form the basis of the model’s output. 4. Breast Cancer Data Training
GAN with Breast Cancer: Given that breast cancer is Process
the most common malignancy in women, a trustworthy
prognostic prediction approach is required. One type of The GAN training process for Breast Cancer data involves
artificial intelligence model is the generative adversarial the following steps.
network (GAN). Its goal is to automatically produce machine The Discriminator’s Training: Real data samples of breast
images. In order to produce new data that closely resembles cancer from the dataset and fictitious data samples produced
an existing dataset, generative models may find application by the generator are both used to feed the discriminator. It
in GANs as a helpful training environment. GANs work by makes an effort to accurately identify real data as real (label 1)
simulating a competitive game between a discriminator and a and false data as false (label 0). To reduce the discriminator’s
generator, two neural networks. classification error, its parameters are modified.
Generator: The generator network creates synthetic data on Training the Generator: The generator creates synthetic data
breast cancer samples from input that is random noise. These samples using the input of random noise. The discriminator
generated samples are initially random and do not mirror the receives these generated samples as input. The generator’s
distribution of the desired data on breast cancer. goal is to create fake samples that the discriminator interprets
AI-Based Breast Cancer X-Ray Image Detection Using Generative Adversarial Attacks 107
as authentic. To increase the discriminator’s inaccuracy on whereas 1 indicates that the patient has cancer. The cardinal
fictitious samples, the generator’s parameters are modified. mammographic raw dataset 𝑍 contains sensitive information,
Through this mechanism, the two networks keep competing such as the X-ray images from the cancer diagnosis example,
with one another. The discriminator grows better at separating and if compromised, d the patient’s secrecy can be violated
real data from phoney as the training goes on, while the such privacy violations could also introduce further attack k
generator gets better at producing realistic data. vectors.
The training process is repeated until the generator generates
health data that the discriminator cannot tell apart from the 5. Experimental Result
genuine data. The GAN can now produce fresh, synthetic This very first experiment discusses the use of generative
samples that closely resemble the training data because it has adversarial networks (GANs) for generating synthetic
successfully learned the underlying data distribution. time-series data, specifically in healthcare cancer datasets.
Procedure of Breast Cancer: The machine that discriminates The paper highlights the challenges of GANs in capturing
is the detective, who will be able to tell the difference long-term temporal relationships, mode collapse, and
between a fake and a real X-ray, while the generator is the mapping complex relationships between measurements and
counterfeiter, who creates false breast cancer X-rays. When attributes, which are particularly challenging for use cases
the training process starts, the generator generates blatantly requiring complete temporal series replication, multimodal
false data about breast cancer, and the discriminator quickly distributions, and complex measurements and attributes.
notices the falsity of the patient’s data. The generator gets
closer to providing output that can trick the discriminator as Table 17.1 GAN of time series data
training progresses. Finally, if generator training is successful, traffic_byte_counter ping_loss_rate isp technology state
the discriminator becomes less accurate at determining what 0 0.001903 0.0 CenturyLink Fiber MN
is real and what is phony. Its accuracy continues to decline
1 0.005421 0.0 CenturyLink Fiber MN
when it starts to classify phoney X-ray breast cancer data as
real health data, as proposed by Kin in 2018 [10]. For instance, 2 0.003513 0.0 CenturyLink Fiber MN
consider the medical diagnosis of breast cancer. By taking a 3 0.003307 0.0 CenturyLink Fiber MN
dataset 𝑍, which contains X-ray images, the goal of a machine 4 0.002243 0.0 CenturyLink Fiber MN
learning process is to construct model M𝜃:𝑋↦𝑌 that given an 5 0.005589 0.0 CenturyLink Fiber MN
arbitrary X-ray image would determine whether a patient has
6 0.003436 0.0 CenturyLink Fiber MN
cancer or not. In this example, the input set 𝑋 contains all
the possible X-ray images encoded as pixel vectors, while 7 0.006160 0.0 CenturyLink Fiber MN
the output set 𝑌 = {0,1} consists only of the two elements 0 8 0.002327 0.0 CenturyLink Fiber MN
and 1, with 0 representing that a patient does not have cancer, 9 0.004787 0.0 CenturyLink Fiber MN
Fig. 17.2 Real vs. artificial data: a sequence length considering traffic and ping synthesization
108 Algorithms in Advanced Artificial Intelligence
The data visualization is used to evaluate the reproducibility Values of correlation, which range from -1 to 1, reflect the
of our artificial data, compare it to the original data, and strength of the association between two variables; higher
determine if hyper parameterization needs adjustment. values imply a stronger relationship.
The outcomes of the experiment using DoppelGANger A threshold value of 0.75 is set to filter out correlations
suggest the need for additional investigation into intricate between target columns in the correlation matrix. Highly
situations, as it reproduces our complete time sequence correlated features are converted into a list, and their
with minimal dimensionality and no major problems. Data correlations are visualized using the Seaborn library’s cluster
practitioners often turn to DoppelGANger as a solution to map function. The chart’s title is determined and added.
the problem of sophisticated false picture data production, The data frame is converted using the pd.melt function,
especially when dealing with time series data. Strong data assigning features and values to create a “melted” version. A
synthesis is one of its features, letting users build realistic box plot is drawn using sns.boxplot, with hue set to “target”
datasets that are near replicas of the source data. It can manage for visual comparison. The feature names are rotated by 90
complicated contexts of heterogeneous data, is generalizable degrees for better readability. The graph is then displayed
and versatile, and has healthcare applications. With this using plt.show ().
method, false-image data can have all the information they
need to make smart decisions without sacrificing privacy or The sns.pairplot tool visualizes relationships between
quality. variables in a data frame. It selects corr_features from the
data frame. The diag_kind parameter specifies the chart type
Data visualisation with Seaborn and Matplotlib, as well as on the diagonals, while markers indicate distinct data points.
analysis with the Scikit-Learn package, are the topics of The hue parameter colorates data points according to the
another study devoted to breast cancer. To help comprehend target variable.
the linkages and groupings of variables, this code groups data
frames with correlations between columns and displays them This code generates a crossplot using KDE plots, density
on a clustering map. estimates, and color coding to display relationships between
variables in a data frame, enabling better understanding and
highlighting different groups.
Fig. 17.4 The data [corr_features].corr() statement creates a correlation matrix, with values displayed in decimal format
Fig. 17.5 A box plot is drawn using sns.boxplot, with hue set to “target” for visual comparison
110 Algorithms in Advanced Artificial Intelligence
7. Good fellow, I. J., Shlens, J. & Szegedy, C: Explaining and Arefan, Shandong Wu: A machine and human reader study on
harnessing adversarial examples in International Conference AI diagnosis model safety under attacks of adversarial images,
on Learning Representations,2015. Nature Communications,12 (1) DOI: 10.1038/s41467-021
8. Hu, W. & Tan, Y: Generating adversarial malware examples 27577-x, 2021.
for black-box attacks based on GAN, arXiv Prepr. ArXiv, 15. Wang, D., Khosla, A., Gargeya, R., Irshad, H. & Beck, A.
1702, 05983, 2017. H.:Deep Learning for Identifying Metastatic Breast Cancer,
9. Jamieson, A. R., Drukker, K. & Giger, M. L: Breast image arXiv:1606.05718 [cs, q-bio], 1606.05718,2016.
feature learning with adaptive deconvolutional networks, 16. Xu, H. et al: Adversarial attacks and defenses in images,
Proc. SPIE 8315, 6–13, 2012. graphs, and text: a review. Int. J. Autom. Computer, 17, 151–
10. Kim, E.-K. et al.: Applying Data-driven Imaging Biomarker 178, 2020.
in Mammography for Breast Cancer Screening: Preliminary 17. Xiao, C. et al: Generating adversarial examples with adversarial
Study, Science Reports 8, 2762,2018. networks, in Proc. 27th International Joint Conference on
11. Li Shen, Laurie R. Margolies, Joseph H. Rothstein, Artificial Intelligence, 3905–3911,2019.
Eugene Fluder, Russell McBride, Weiva Sieh: Deep 18. Yala, A. et al: Toward robust mammography-based models for
Learning to Improve Breast Cancer Detection on Screening breast cancer risk. Sci. Transl. Med., 13, 1–11, 2021.
Mammography, Scientific Reports, volume 9, Article number: 19. Yuan, X., He, P., Zhu, Q. & Li, X: Adversarial examples:
12495, 2019. attacks and defenses for deep learning, IEEE Trans. neural
12. McKinney, S. M. et al: International evaluation of an AI Netw. Learn. Syst., 30, 2805–2824,2019.
system for breast cancer screening, Nature, 577, 89–94, 2020. 20. Zhu, W., Lou, Q., Vang, Y. S. & Xie, X: Deep Multi-
13. Mohamed, A. A. et al: A deep learning method for classifying Instance Networks with Sparse Label Assignment for
mammographic breast density categories, Med. Phys. 45, Whole Mammogram Classification, arXiv: 1705.08550 [cs],
314–321, 2018. 1705.08550 2017.
14. Qian Wei Zhou, Margarita Zuley, Yuan Guo, Lu Yang,
Note: All the figures and table in this chapter were designed by the
Bronwyn Nair, Adrienne Vargo, Suzanne Ghannam, Dooman
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
112 Algorithms in Advanced Artificial Intelligence
Chandra Sekhar K.1, K. Satyanarayana Raju2, P. Subba Raju3, M. Krishna Satya Varma4, K. Laxmipathi Raju5
Assistant Professor, Department of IT, SRKR Engineering College, Bhimavaram, Andhra Pradesh, India
Abstract: Job hunting is difficult work, but with a few little adjustments to the current procedure, we can make it easier,
which may have a number of beneficial effects. Various career advice sites provide a wide range of work possibilities. All of
those choices, though, might not be beneficial to everyone. Consequently, a job suggestion engine that can suggest the best
employment match for the applicant’s profile would be a highly useful tool. Artificial intelligence is now being used more
often in educational settings. However, to further the systematic use of these techniques, more conceptual and methodological
knowledge is required. This study’s initial goal is to evaluate a methodical approach for using artificial neural networks to
forecast student placement in engineering universities. The second goal is to evaluate the significance of a number of well-
known predictors that have an effect on the student’s placement. As a result, this study proposes a method for creating a
placement prediction model for graduate engineering students in their pre-final year using Artificial Neural Networks (ANN).
The models were trained and tested using data from a sample of 1146 students. The model with the best accuracy was 84.1
percent overall.
Keywords: Graduate, Placements, Admission, Student, Quality education, ANN
DOI: 10.1201/9781003529231-18
Promotion of Graduate Placement Through Academics by Improving Performance Using Artificial Neural Networks 113
for pattern extraction and trend identification. Specifically, in educational research, highlighting the role of earlier
An MLP, a type of feedforward ANN employing a supervised academic success in categorizing students’ academic
learning methodology, was used to identify the selected performance. The study acknowledges limitations in data
stream. availability, particularly in terms of high school grades and
Furthermore, the ANN model incorporates grades from socio-economic status reporting. The research emphasizes
multiple academic years as input, enhancing its predictive the importance of using multiple placement prediction
accuracy. The subsequent sections delve into a detailed models to leverage academic and placement information for
exploration of the ANN framework, shedding light on its forecasting future placement prospects, aiding students in
intricate workings and applications in this specific context. recognizing strengths and making necessary improvements.
A study [6] delves into various placement prediction models,
showcasing the promising potential of the student dataset for
2. Literature Review forecasting future placement prospects. With an accuracy of
The ID3 decision tree method, as outlined in research [1], is 74.1 percent [7], this study demonstrates the application of
employed to construct a model predicting the likelihood of a the discretization approach to enhance prediction accuracy.
student securing placement in a firm. The dataset provided It suggests expanding the predictive scope by considering
is meticulously analyzed using this method to identify the additional factors like family income and the educational
most relevant parameters for placement prediction. Each backgrounds of parents and siblings. Additionally, the
parameter undergoes scrutiny for entropy and information analysis may incorporate additional tracks or strands to
gain values, with the optimal parameter chosen as the split further refine predictions.
variable for crafting the decision tree. Utilizing the Weka
Tool, an optimal decision tree is generated, with the leaves 3. Proposed Technique and Attribute
indicating the predicted likelihood of a student being placed.
The dataset encompasses secondary test scores, graduation
Selection
grade points, history and departmental arrears, talents such Research indicates that various factors, encompassing
as programming and communication, completed internships, demographics, extracurricular activities, environment, and
and information on future study interests. Another placement biological elements, play a role in influencing a student’s
prediction system [2], leveraging the K-Nearest Neighbors performance. A thorough investigation and adjustment of
classifier, predicts the likelihood of students being placed in these variables were conducted to create a detailed numerical
various firms. This outcome is then compared with results representation suitable for computer coding. These refined
from other machine learning models like SVM and Logistic variables serve as inputs to the system, contributing to a
Regression. The assessment considers academic records, comprehensive understanding of the student performance
as well as programming, communication, analytical, and dynamics. The system’s architecture and functionality
teamwork abilities that employers scrutinize during the are visually represented in Fig. 18.1, illustrating the
hiring process. The system utilizes data from the previous interconnectedness of the variables within the framework.
two batches. In a different approach, [3] introduces a TPO
management system to forecast eligibility for a campus drive
using the C4.5 Decision Tree Algorithm. Historical student
data is examined to predict current students’ eligibility and
the institution’s overall placement likelihood. The decision
tree is constructed based on the company’s past data and
current requirements, aiding in estimating a student’s
eligibility in different firms. The system notifies candidates
meeting the criteria for the company’s campus drive based on
these factors, providing valuable insights for students to plan
their career paths effectively. Addressing potential issues in Fig. 18.1 Block diagram representation
student performance and graduation delays, [4] proposes
a NN model predicting a student’s GPA based on location, 3.1 Educational Data Mining
academic background, and personal data. The model, trained A method for drawing a model or knowledge from a big
and evaluated using the WEKA software program on a sample collection of data is called data mining (DM). A system of
dataset of computer networking students, demonstrates a categorization or prediction in several fields, including the one
73.68 percent accuracy in forecasting student performance. that is the subject of this work, the field of education, is being
In a broader context, [5] explores the application of ANNs developed as a result of the rapid growth of DM. Educational
114 Algorithms in Advanced Artificial Intelligence
Data Mining (EDM) is the use of DM in education, and it 3.3 Implementation of the ANN Algorithm
involves the process of information extraction from placement
The ANN algorithm was implemented using a Python-based
data, such as student placement status.
software, employing various Python packages such as pandas,
3.2 Artificial Neural Network (ANN) numpy, and sklearn. The model encompassed the following
key categories:
An NN is an artificial emulation of the human brain, designed
to replicate the learning processes observed in the human 3.4 Data Loading
cognitive system. The versatility of Artificial Neural Networks
The application was initialized by loading the necessary
(ANN) in handling tasks involving intricate systems is a
datasets, providing the foundational information for
defining characteristic that contributes to its significant value.
subsequent processing.
In the realm of education, where processes like learning and
complex decision-making, including career choices [10– 3.5 Assignment of Input and Output Data
13], are involved, ANN proves particularly beneficial. The
ANN model mirrors the organizational shape of the human Following the data loading phase, the program allocated
nervous system and the functioning of biological cognition. the relevant input and output data, setting the stage for the
At its core, the neuron serves as the fundamental building subsequent steps in the algorithm.
block, functioning as a processing unit. A neuron comprises
3.6 Data Normalization
three essential components: cell bodies or soma, dendrites,
and axons. Dendrites accepts signals from the external world A crucial step in the process involved normalizing the data.
or yield of other neurons, transmitting these impulses to the The primary objective of this normalization strategy was to
soma. From there, the signals are relayed to the axon and, bring all input and output values into a comparable range.
ultimately, through synapses, transmitted to the dendrites of This ensures that the algorithm operates more effectively,
surrounding neurons. The neural network is an amalgamation preventing bias towards variables with larger magnitudes.
of these neurons and the intricate neurological processes
associated with them. Through the utilization of certain inputs, 3.7 Hyperparameter Tuning
ANN can be trained to predict outputs, such as forecasting Achieving the best performance of the ANN model required
a career strand or predicting grades in various subjects. A fine-tuning of hyperparameters. This process involves
specific category of feed-forward ANN is the multilayer experimenting with different combinations to identify the ideal
perceptron (MLP), which employs the backpropagation set of parameters that maximizes the model’s effectiveness
approach for training. In essence, an MLP constructs a model in handling the specific dataset at hand. By systematically
based on data samples, providing a powerful tool for learning progressing through these sections, the execution of the ANN
and prediction in complex systems. algorithm was orchestrated, ensuring a comprehensive and
efficient application of the model.
3.8 Data
Dataset includes details about engineering graduates, covering
gender, 10th and 12th percentile marks, EAMCET rankings,
Fig. 18.2 Basic architecture of ANN course backlogs, graduation CGPAs, engineering program
Promotion of Graduate Placement Through Academics by Improving Performance Using Artificial Neural Networks 115
(Branch), and the number of jobs secured. It encompasses included, and their placement information is gathered. After
features such as preprocessing, the data feature are engineered. The modeling
S. No Variable Range Value phase was completed, and data got split into a ratio of 80% to
i Gender Male Female
20%. The model evaluation is done, and the performance of
the algorithms is visualized.
ii Program Civil, CSE, ECE, EEE, IT, MECH
The pair plot representation shows the combination of
iii. Xth % CGPA (0-10)
variables in the Dataset. They are represented as a matrix of
iv Inter% Percentage (53-98.7)
plots. Each plot shows different distributions.
v. BE% CGPA (4-9.17)
vi Backlogs 0-No Backdogs 5. Train and Test Split of the Dataset
vii. EAMCET Student Rank
We separated our target and independent variables as y and x
viii Placement Selected (0 or 1)
for splitting them into training and testing data and imported
the model selection module for train and test splitting data
4. Proposed System and model building. Partition the dataset as X_train, X_test,
y_train, and y_test as 80 and 20 percent.
6. Data Normalization
We scaled our data to values between +2 and -2 in order
to make the data more evenly distributed and with a lower
standard deviation. A common scaler was employed for this
scaling. We brought the common scaler module over from
Sklearn. Preprocessing [10] has finished, and the typical
scaler object has been produced. We converted our test data
and fitted train data using this common scaler. Every feature
or variable is scaled to unit variance with the help of the
StandardScaler, which removes the mean. Each person goes
through this procedure feature by feature. Because outliers
might affect StandardScaler, we first eliminated them before
applying this technique[11]. This method of scaling is known
as “standardization,” wherein values are centered around the
Fig. 18.3 Architecture diagram of graduate prediction system mean and normalized to have a single standard deviation.
Consequently, the attribute’s mean is adjusted to zero, and
The proposed system comprises a data set derived from the resulting distribution exhibits a unit standard deviation.
previous graduate students’ data with all the six streams
7. Dependent and Independent
Variables for all our Machine
Learning Models
Here our dependent variable is the placement column
which specifies whether the student got placed or not. Here
are the Independent variables Gender, SSC percentage,
Inter percentage, Btech CGPA, Eamcet rank, Branch, and
Backlogs.
8. Results
Tests employed our university’s placement data, made into a
training set (80% of the data) and a cross-validating testing
set (20%). Various models predicting placement accuracy
Fig. 18.4 Pair plot representation of placement data initially used the data as input.
116 Algorithms in Advanced Artificial Intelligence
Results
9. Conclusion
We were able to create a prediction model with an accuracy
of 84.04 percent that can identify the preferred strand of a
Fig. 18.7 Branch wise placement
Promotion of Graduate Placement Through Academics by Improving Performance Using Artificial Neural Networks 117
graduate passing student. This study was able to demonstrate method,” 2009 1st Int. Conf. Adv. Comput. ICAC 2009,
how to apply the discretization approach to improve prediction pp. 28–33, 2009, DOI: 10.1109/ICADVC.2009.5378273.
accuracy. The ANN model was trained using grades from the 7. H. Turabieh, “Hybrid machine learning classifiers to predict
10th, +2, and EAMCET. The suggested strategy is not limited student performance,” 2019 2nd Int. Conf. New Trends
Comput. Sci. ICTCS 2019 - Proc., pp. 1–6, 2019, DOI:
to just demographic characteristics. Henceforth, exploring
10.1109/ICTCS.2019.8923093.
the system’s efficacy in using student demographics and past
8. T. Pranckevičius and V. Marcinkevičius, “Comparison of
academic records to predict students’ performance in the Naive Bayes, Random Forest, Decision Tree, Support Vector
subsequent educational level could be a subject of study in Machines, and Logistic Regression Classifiers for Text
the future. Reviews Classification,” Balt. J. Mod. Comput., vol. 5, no. 2,
pp. 221–232, 2017, DOI: 10.22364/bjmc.2017.5.2.05.
References 9. M. A. H. Farquad and I. Bose, “Preprocessing unbalanced data
using support vector machine,” Decis. Support Syst., vol. 53,
1. Sekhar K, Chandra; Kumar, K. Santhosh, ‘Undergraduate no. 1, pp. 226–233, 2012, DOI: 10.1016/j.dss.2012.01.016.
Student’s Campus Placement Determination Using Logistic 10. S. Garcia, J. Luengo, and F. Herrera, Data Preprocessing in
Regression Analysis for Predicted Probabilities on Uncertain Data Mining. Intelligent Systems Reference Library. 2015,
Dataset’, International Journal of intelligent systems and vol. 10. 2015.
applications in engineering., vol. 10, no. 2s, pp. 14–20, 2022. 11. L. Kristoffersen Edward Mayce R. and R. M. Hernandez, “A
2. Chandra Sekhar K, K. Santhosh Kumar, ‘Data Preprocessing comparative performance of breast cancer classification using
and Visualizations Using Machine Learning for Student hyper-parameterized machine learning models,” Int. J. Adv.
Placement Prediction’, 2nd International Conference on Technol. Eng. Explore., vol. 8, no. 82, pp. 1080–1101, 2021,
Technological Advancements in Computational Sciences DOI: 10.19101/ijatee.2021.874380.
(ICTACS). Appl., vol. 2, pp. 386–391, 2022. 12. Q. Wang, “Kernel Principal Component Analysis and its
3. Ajay Shiv Sharma1, Swaraj Prince2, Shubham Kapoor3 Applications in Face Recognition and Active Shape Models,”
, Keshav Kumar4 -“PPS - Placement Prediction System no. July 2012, 2012, [Online]. Available: http://arxiv.org/
using Logistic Regression: Lakshmipriya. K, Dr. Arunesh abs/1207.3538.
P.K-“Predicting Student Performance Using Data Mining 13. S. K. Thangavel, P. D. Bkaratki, and A. Sankar, “Student
Classification Techniques”. placement analyzer: A recommendation system using machine
4. N. Soomro, F. Razaque, and S. Soomro, Cluster and Logistic learning,” 2017 4th Int. Conf. Adv. Comput. Commun.
Regression Distribution of Students ’ Performance by Classi Syst. ICACCS 2017, no. March, 2017, DOI: 10.1109/
fi cation, vol. 1. Springer International Publishing. ICACCS.2017.8014632.
5. G. S. K. Ranjan, A. Kumar Verma, and S. Radhika, “K-Nearest 14. H. Shi and Y. Liu, “Naïve Bayes vs. support vector machine:
Neighbors and Grid Search CV Based Real-Time Fault Resilience to missing data,” Lect. Notes Comput. Sci.
Monitoring System for Industries,” 2019 IEEE 5th Int. Conf. (including Subsea. Lect. Notes Artif. Intell. Lect. Notes
Converg. Technol. I2CT 2019, no. June 2020, 2019, DOI: Bioinformatics), vol. 7003 LNAI, no. PART 2, pp. 680–687,
10.1109/I2CT45611.2019.9033691. 2011, DOI: 10.1007/978-3-642-23887-1_86.
6. M. J. Meena and K. R. Chandran, “Naïve Bayes text Note: All the figures and tables in this chapter were designed by
classification with positive features selected by statistical the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
118 Algorithms in Advanced Artificial Intelligence
Abstract: Payroll entails keeping track of workers’ hours worked, computing their compensation, and transferring funds to
their bank accounts or direct deposit. One deep learning technique that can handle a wide range of natural language processing
(NLP) problems is the large language model (LLM). Transformer models are used in LLM training, where large employee
payrolls are used. As a result, they can now recognize, translate, project, or produce text or other content. The article discusses
the use of Open AI’s Large Language Model to enhance payroll and HR processes. In this paper, AI-powered chatbots and
virtual assistants, like ChatGPT, are revolutionizing HR by automating payroll processes, streamlining employee support, and
saving time and resources. OpenAI’s large language model (LLM) can generate human-like responses, revolutionizing payroll
calculations and HR support.
Keywords: Artificial intelligence, ChatGPT, Deep learning, Large language model, National language processing, Payroll and
HR
1
lokesh3069@gmail.com, 2mshv@srkrec.ac.in
DOI: 10.1201/9781003529231-19
Open AI’s Large Language Model to Improve Payroll and HR Processes 119
programmes tailored to individual employees’ requirements. training dataset is referred to as “model selection.” If there
It compiles a thorough report on an employee’s strengths, are features X and a target Y, the best transformation F can be
shortcomings, and improvement opportunities by analysing determined from the data by: Y = F(X). The word “optimal”
data such as job history, duties, and feedback. Managers are denotes the existence of a model performance metric, and the
able to better engage their employees and keep them around “optimal” model is the one that maximizes that statistic. It is
when they can provide them with constructive criticism important to consider a number of axes in order to improve
and chances to grow. Even with the most intricate payroll our model:
systems, ChatGPT can keep correct records and update 1. The model parameter space: Use statistical learning
payroll calculations. Nevertheless, HR professionals should to “train” a model and “optimize” this “space”. The
view technology as an adjunct to their work rather than a parameters are learned using an optimization approach,
substitute. Human resource experts are still vital in handling such as the maximum likelihood estimation principle.
complex HR tasks.
2. The Model Paradigm Space: It is possible to employ
a variety of supervised learning algorithms to address
2. Proposed Work the same issue. Depending on the particular dataset,
The use of AI-powered chatbots and virtual assistants is algorithms like Naive Bayes, XGBoost, or Neural
rapidly expanding in the human resources sector. Businesses Network may perform significantly differently.
can save time and money with these tools, which automate 3. The Hyperparameters Space: Make these choices in
essential payroll and HR tasks, accelerate employee support, order to put up our training run, even though statistical
and more. One promising tool in this sector is OpenAI’s learning cannot enhance these model parameters.
massive language model, ChatGPT. Because it understands 4. The Model Architecture Space: This applies more
spoken language and can give replies that sound human, to neural networks. A set of Hyperparameters can
ChatGPT is a great way to automate HR tasks and provide be used to describe the model architecture, but the
support to employees. search is typically more involved than with ordinary
Understanding Payroll Using Random Forest Hyperparameters. The size of the search space can
Classification: When it comes to employee payroll, Random reach 10^40.
Forest is a popular ensemble technique for both classification 5. The feature space: The proper feature must be chosen
and regression issues. To process payroll, one must maintain to feed our model. Depending on the features that can
track of working hours, calculate compensation, and be used, different models will respond in different
distribute monies. All employees’ financial transactions, ways. Excessive characteristics and possible overfit. It
including payroll, taxes, bonuses, overtime, sick leave, might not fit if there aren’t enough features.
vacation pay, and government taxes for programmes like 6. The Feature Transformation Space: To enhance the
Social Security, Medicare, and unemployment, must be performance of our model, take into account several
recorded by organisations. Employees can input their hours transformations as the Box-Cox transformation and
worked using an API, and any company can use the payroll feature encoding.
administration software. Large and medium-sized companies Large Language Model for Employee Payroll: Large
often use outside firms to handle their payroll processing so language models (LLMs) are AI systems that use deep learning
that they can save time and effort. Businesses deduct taxes techniques and massive data sets to understand, synthesise,
from employees’ gross wages, keep track of their working create, and predict new content. LLMs lay the groundwork
hours, report this data to payroll processors, and then pay for conversation and the development of new ideas. Language
their employees. models like LLMs, which were among the first AI language
models, find widespread use in NLP systems that allow users
to input questions using natural language. Advancements in
AI’s language model idea have greatly increased the amount
of data available for training and inference, hence enhancing
the capabilities of AI models. Learning Language Models
(LLMs) can benefit from databases, yet these tools are not
without their limitations. The payroll model embeds and
indexes data into a vector database. The questions asked
Fig. 19.1 Employee payroll using random forest classification
by users are transformed into embeddings, which are then
Employee Payroll Optimizing a model: The procedure used to find other embeddings that are comparable to them.
of selecting the best employee payroll model for a specific However, the system may produce results that are not relevant
120 Algorithms in Advanced Artificial Intelligence
to the employee’s Details that are closely related should The code uses -1 for the last column and 1 for the second
be presented in a concise and focused manner. To prevent column, removing the last column and starting indexing from
irrelevant or watered-down publications, break down the data zero, resulting in outputs for p and q. The code will generate
into chunks of a few paragraphs each. Preventing erroneous a dataset consisting of x-test, x-train, and y-test, y- train
query results can be achieved by limiting inquiry types. images, as shown in Fig. 19.4.
Experimental Result: The two components of the Employee The image displays the extraction of p and q variables from a
payroll dataset are work experience and remuneration. dataset, split into a test and training set with 20 observations
Finding the greatest fit line and determining the connection for training and 10 for testing.
between the two attributes are the challenges at hand. The # the dataset is divided into training and a test set.
relationship between the parameters will be displayed using from sklearn.model_selection import train_test_split
a Python linear regression model.
p_train, p_test, q_train, q_test=train_test_split
Step-1: Pre-processing of Payroll Data: Import three (p, q, test_size=1/3, random_state=0)
necessary libraries in order to load the dataset, produce
graphs, and develop the model for a linear regression. Our payroll dataset is ready for linear regression; but, because
of how Python libraries handle specific scenarios, feature
import numpy as nm scaling will not be employed.
import matplotlib.pyplot as mtp
Step-2: Fitting the Training Set with the Linear Regression:
import pandas as pd To create a regressor object and import the linear regression
The payroll dataset can be loaded with the following code: class from scikit-learn, use the provided code #Fitting the
data_set = pd.read_csv (‘payroll_data.csv’). The variable Simple Linear Regression model to the training dataset.
explorer feature on our Spyder IDE screen allows us to use from sklearn.linear_model import Linear
the code (ctrl+ENTER) to read the dataset.
Regression
The payroll dataset consists of salary and experience, and the
code is used to extract independent and dependent variables. Regressor= Linear Regression ()
In order to allow the model to learn correlations, the code The code will generate salary predictions for the training and
passes p_train and q train as the dependent and independent test sets in the variable explorer options using variables q
variables when fitting a Simple Linear Regression object to a pred and p_pred.
training set using the fit() method. Output: The values of p_test and q pred can be used to
Output: Out [7]: Linear Regression (copy_P=True, compare outcomes, and the variable can be inspected using
fit_intercept=True,n_jobs=None, normalize=False) the IDE's variable explorer option.
Step-3. Forecast of the test set outcome: The model is ready The describes how to use the scatter () function in the pyplot
to use experience and salary to forecast fresh observations. To library to display the training set results. The plot is made up
evaluate the model's accuracy, a test dataset is given, which of a title for the plot, a regression line, and an observation
yields the prediction vectors p_pred and q pred. scatter plot. The employee pay is shown on the y-axis, while
#Prediction of Test and Training set result the years of experience and wage are shown on the x-axis.
q pred regressor. Predict (p_test) The data is plotted on a graph using show() once the labels
p_pred regressor. Predict (p_train) for the p and q axes are assigned.
122 Algorithms in Advanced Artificial Intelligence
7. Samaduzzaman, M. & Zaman, F. (2012). “E-HRM in Conference on Emerging Technologies, IEEE, 2010.
Bangladesh”. IOSR Journal of Business and Management, doi: 10.1109/ICET.2010.5638465
4 (6), 32–36. https://www.iosrjournals.org/iosr-jbm/papers/ 10. A.F.M. Sultanul Kabir, M. Ahmed Shorif, H. Li, and Q. Yu,
Vol4-issue6/F0463236.pdf “A Study of Secured Wireless Sensor Networks with Xbee
8. A. Sonkamble, “Automation of Attendance System using and Arduino”, 2nd International Conference on Systems and
RFID, Biometrics, GSM Modem with .Net Framework”, Informatics, IEEE, 2014. doi: 10.1109/ICSAI.2014.7009337.
International Conference on Mutimedia Technology, IEEE, Note: All the figures in this chapter were taken from https://
2011. https://ieeexplore.ieee.org/document/6002032 www.linkedin.com/posts/damienbenveniste_machinelearning
9. Z. Rashid, A. Basit, and Z. Anwar, “TRDBAC: Temporal datascience-artificialintelligence-activity-7112101911918522368
Reflective Database Access Control”, 6th International IgBl
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
124 Algorithms in Advanced Artificial Intelligence
Abstract: Sharing of Electronic Medical Records (EMRs) between doctors and medical institutions can now be done using
Blockchain - a disruptive approach to the exchange of EMRs. Blockchain can enhance the accuracy of medical decisions
and improve public health significantly. However, there is a need to ensure that sensitive information is retrieved from the
correct encrypted EMRs, and it is even more difficult to dynamically update the user attributes of authorized users. To this
end, we propose a secure data sharing approach that uses blockchain and encryption techniques to ensure secure, efficient,
and patient centric data sharing. We came up with a never-before- seen approach to integrate attribute-based encryption,
searchable encryption, and robust access control mechanisms to update user attributes with top-notch security measures in
place. Additionally, we delve into the hurdles encountered and ways to make amends to the strengths and potential areas for
improvement. Our proof of consistency demonstrates the impact of adding a consortium blockchain on securing the shared
electronic medical records and the consequential improvements it yields for the healthcare industry.
Keywords: Access control, Attribute-based encryption, Blockchain, Dynamic user attributes, EHR data sharing, Searchable
encryption
DOI: 10.1201/9781003529231-20
A Novel Blockchain-Based Approach for Secure and Efficient Electronic Medical Record Sharing 125
such as name, date of birth, or medical record number, could more convenient. In (Capece 2020), authors examined the
be unique identifiers in the context of EMRs. Patient privacy blockchain’s potential applications in healthcare, targeting
and data security are maintained by ABE to ensure that electronic medical records in particular. They go over how
only authorized individuals can decrypt the medical records blockchain improves data security, integrity, and access
(Li 2015). management.
Additionally, a cryptographic method, known as searchable Authors in (Han Y 2022) discussed the potential of blockchain
encryption, enables users to look for certain data without in addressing the interoperability and privacy issues of
having to first decrypt it completely (Cash 2014). In EMRs, electronic medical records (EMRs). They highlight the need
where healthcare practitioners may need to locate specific to overcome challenges related to data access and sharing
patient records from a large, encrypted dataset, this method is in blockchain based EMR schemes. They also discussed
of utmost relevance. Blockchain technology can dramatically ongoing challenges in data management efficiency, fairness
increase the security and speed of medical record retrieval by of access, and trust in the systems.
combining ABE and searchable encryption. Combining these Reegu et al. (Reegu 2023) proposes a blockchain-based
methods guarantees authorized users’ access, privacy, and framework named Ancile for secure interoperability and
confidentiality (Kamara 2014). organized access to medical records. The authors emphasized
Another challenge for blockchain use in healthcare is how to the potential of blockchain in revolutionizing the exchange
manage dynamic updates to medical records. A strong access and processing of EMRs. Yan et al. (Yan 2023) focused
control system is needed to handle this dynamic nature since on attribute-based searchable encryption and blockchain
patient attributes are continually changing. This mechanism access control in a cloud environment. The authors suggest
should allow for the addition, modification, and deletion a fine-grained access control system that integrates IPFS
of user attributes without compromising the security of the (Interplanetary File System), attribute-based encryption, and
blockchain (Xu 2018). blockchain technology.
To address these challenges, this paper proposes an innovative These papers provide insights into the use of attribute-
attribute model that allows for the dynamic updating of user based encryption and searchable encryption in the context
attributes while maintaining the integrity of the blockchain. It of blockchain technology for electronic medical records
ensures that only authorized users can update their attributes and data sharing in healthcare. They discuss the challenges,
and that these updates do not affect the security of the opportunities, and potential improvements in this field.
encrypted medical records.
Furthermore, this paper seeks to perform theoretical 3. System Model
evaluations of both security and performance aspects to
showcase the durability and reliability of the proposed We suppose that a consortium chain is formed by multiple
blockchain system for EMRs. As a result, we are tackling hospitals, where each hospital has a local server and several
challenges related to retrieving encrypted health records clients that are operated by the doctor. Each hospital builds
securely and accurately while incorporating dynamic updates. its own private chain, while multiple private chains build a
Thus, providing a practical way for blockchain technology consortium chain.
implementation in the medical sector. Before entering the system, patients, doctors, and data users
The rest of the paper is summarized as follows: Section 2 need to register and generate their own public-private key
presents the most relevant related words. The system model pairs. Among them, the patient’s electronic medical record
for the proposed solution is stated in Section3. Section 4 ciphertext is stored on the hospital server, the hash value and
proposed the EMR storage scheme based on consortium keyword index of the electronic medical record ciphertext
blockchain. The conclusion and future work are then are stored on the hospital private chain, and the security
presented in section 5. index composed of the private chain block identification,
patient pseudo-identity and keyword index is stored on the
consortium chain. The system mainly includes 6 entities
2. Related Works such as patients, doctors, data users, hospital servers, private
In recent years, EMR data sharing has become a hot spot in chains, and consortium chains (see Fig. 20.1).
the field of public health and smart healthcare. (LIU Gechang • Patient: When a patient is admitted to the hospital,
2019) proposed a data privacy protection mechanism based on they begin by registering on the hospital’s server. Once
searchable encryption. The system is applied to the personal the registration process is complete, the hospital server
medical data blockchain, which makes the private data search assigns a unique number plate to the patient, serving as
126 Algorithms in Advanced Artificial Intelligence
their medical card. It is crucial for the patient to keep trap and uploads it to the consortium chain. The nodes
this number plate confidential and present it during on the consortium chain perform the search, and upon
consultations. Doctors, on the other hand, generate finding the corresponding patient ciphertext, the relevant
electronic medical records and associated keywords for node on the consortium chain takes appropriate action
each patient. These records are then encrypted using the based on the search results.
patient’s public key. In the event that the patient visits • Hospital Server: As the doctor attends to the patient and
another hospital, and a doctor needs access to their creates the electronic medical record, the hospital server
medical history, a search trap is generated by the patient. retrieves the block identity from the private chain, along
The search trap is uploaded to the consortium chain. with the patient’s identity and keyword index. Utilizing
• Doctor: In the hospital, there is a local server along with this information, the hospital server constructs a new
multiple clients, which are operated by doctors. When a transaction on the consortium chain. The remaining nodes
doctor sees a patient, they create an identity, electronic within the consortium chain undertake the crucial task of
medical record ciphertext, keyword ciphertext, and validating the transaction. Once the verification process
evidence for the patient. The doctor then uploads the is successfully completed, a fresh block is generated on
electronic medical record ciphertext to the hospital the consortium chain, securely incorporating the updated
server. Simultaneously, the doctor uploads the hash information.
value of the encrypted medical record, and the keyword • Private Chain: To initiate new transactions, the
index, which consists of the encrypted keyword, to the doctor uploads the hash value of the electronic medical
private chain. To ensure the integrity of the data, the record ciphertext and the keyword index, consisting
doctor generates a new transaction and broadcasts it to of the keyword ciphertext, to the private chain. Upon
the network. The other nodes on the private chain take receiving a transaction created by the doctor, the node
on the responsibility of validating the transaction. Once on the private chain diligently verifies its authenticity.
the verification process is successfully completed, a new Subsequently, the hospital server leverages the private
block is added to the private chain, further securing the chain’s block identity, patient identity, and keyword
information. index to construct a new transaction on the consortium
• Data Users: When external institutions or individuals, chain. During the data acquisition phase, if the search
referred to as data users, need access to patient data, they proves to be successful, the node on the consortium
must obtain authorization from the patients themselves. chain extracts the secure index from the block and
To initiate this process, the patient generates a search obtains the private chain’s block identity. By utilizing
A Novel Blockchain-Based Approach for Secure and Efficient Electronic Medical Record Sharing 127
the private chain block identification, the nodes on the 2. To validate and confirm transactions and blocks added
consortium chain gain access to the hash value of the to the chain, the private chain can employ a consensus
medical record ciphertext, enabling them to retrieve the mechanism, like Proof of Work (PoW) or Proof of
necessary information securely. Stake (PoS) algorithms. These mechanisms ensure that
• Consortium Chain: As the search process unfolds, all participants in the chain agree on the validity of the
the nodes on the consortium chain receive the trapdoor data being added thereby maintaining consistency.
transmitted by the patient and execute the search 3. In order to securely store records their ciphertext, hash
algorithm. After a successful search, the data user node value and keyword index are stored in a protected
obtains the security index and private blockchain ID manner within the private chain. Encryption techniques
from the consortium blockchain. It then retrieves the are used to safeguard the confidentiality and integrity
hash value of the encrypted medical record and sends of this data ensuring that authorized entities can access
it back to the hospital server. The hospital server then and modify it.
compares this hash value with the one associated with
the electronic medical record ciphertext to ensure their 4.2 Proof of Consistency of Consortium Chain
consistency. If there is a match, the data user node sends In the meticulously devised algorithm pertained to the
the medical record ciphertext back to the patient. In impervious realm of data sharing, the consortium blockchain
cases where a third-party data user needs access to a and encryption techniques stand as the gatekeepers. As the
patient’s electronic medical record, the consortium chain beacon of reliability and integrity, the consortium chain
node acts as an intermediary. The consortium chain node ensures the consistency and validity of the keyword indexes
plays an intermediary role when a third-party data user and search trapdoor. The secure index, which consists of
requests access to a patient’s electronic medical record. keyword indexes, is kept in the consortium chain.
In order to maintain privacy and safeguard the data, the
To achieve proof of consistency on the consortium chain, the
node creates an agent re-encryption key and performs
system constructs a polynomial (𝑥) using hash functions. The
proxy re-encryption on the existing electronic medical
polynomial (𝑥) is designed to represent the set of keywords
record ciphertext. The newly formed re- encrypted
𝑊 = {𝑤1, 𝑤2 ⋯ 𝑤𝑛} that contains a description of all the
ciphertext is then transmitted to the third- party user
symptoms that the patient is likely to have. The coefficients
through a secure channel.
of the polynomial are derived from the hash values of the
keywords.
4. Consortium Mechanism The steps listed below can be used to show the proof of
Based on searchable encryption, the consortium mechanism consistency:
utilizes a security index consisting of keyword indexes that’s 1. Construction of Polynomials: Using the hash values of
held on the consortium chain. Whenever there’s a need for the keywords, (𝑤𝑖), the system creates a polynomial
electronic medical records data by either a patient or data 𝑓(𝐻(𝑤𝑖)), where 1 ≤ 𝑖 ≤ 𝑛. Every keyword is hashed,
user, they use their private key to generate a search trapdoor. and the resulting hash value functions as a polynomial
The search trapdoor is then sent to the consortium chain coefficient. The polynomial is built in a manner
where nodes would then perform the search. that allows it to represent the set of keywords. The
polynomial can be represented as:
4.1 Proof of Consistency of Private Chain
(𝑥) = (𝑥 − 𝐻(𝑤1))(𝑥 − 𝐻(𝑤2)) ⋯ (𝑥 − 𝐻(𝑤𝑛)) (1)
The method suggested for secure sharing of data involves
The polynomial (𝐻(𝑤𝑖)) = 0∀𝑖. (𝑥) can also be
the utilization of encryption techniques and blockchain
expressed as:
technology. A crucial aspect of this approach is the
implementation of a private blockchain that is customized (𝑥) = a0 + a1 × 𝑥 + a2 × 𝑥2 + ⋯ + a𝑛 × 𝑥𝑛 (2)
for each hospital and plays a major role in maintaining the Where (𝑤) is the hash value of keyword 𝑤 and 𝑎0,
validity and coherence of each participant’s blockchain copy. 𝑎1 ⋯, 𝑎𝑛 are coefficients derived from the hash values
The private blockchain functions as a storage mechanism for of the keywords.
encrypted medical records, hash values, and keyword indexes. 2. Create a search trapdoor utilizing the private key
Consistency in the private blockchain can be preserved by whenever a patient or data user needs to access
following a set of sequential procedures. information from an electronic medical record. By
1. The private chain relies on a structure that is “append changing the value of 𝑥 in the polynomial (𝑥) to (𝑟),
only.” This means that once a block is appended to a value known as the search trapdoor is obtained. The
the chain, it is irrevocable and cannot be deleted or consortium chain will process this search trapdoor
modified. further after receiving it.
128 Algorithms in Advanced Artificial Intelligence
3. Consistency Verification: The nodes on the consortium 2. 𝐾𝑒𝑦𝐺𝑒𝑛 algoritℎm: Tℎe 𝐾𝑒𝑦𝐺𝑒𝑛 algorithm in this scheme
chain perform calculations using the search trapdoor is described in Algorithm 2 as follows: In this algorithm, the
and the polynomial 𝑓(𝑥). By substituting the search input is the system parameters 𝑃𝑃, the master key 𝑀𝑠𝑘, the
trapdoor value into the polynomial, the nodes can patient attribute 𝑎, and the doctor attribute 𝑑. The output is
verify if the resulting value matches the hash value the secret key 𝑆𝑘.
of any keyword in the secure index. If a match is Algorithm 2: 𝐾𝑒𝑦𝐺𝑒𝑛 Algorithm
found, it indicates that the patient’s search trapdoor
corresponds to a keyword in the secure index, ensuring 1: Input: System parameters 𝑃𝑃, Master key 𝑀𝑠𝑘, Patient
the consistency of the consortium chain. attribute 𝑎, Doctor attribute 𝑑
Suppose there is a vector 𝑏 = [𝑏 0, 𝑏 1, ⋯, 𝑏 𝑛], the system 2: Output: Secret key 𝑆𝑘
can then introduce new polynomial 𝑔(𝑥) such that: 3: Choose a random exponent 𝑟 ∈ 𝑍𝑞∗.
(𝑥) = 𝑏 0 + 𝑏 1 × 𝑥 + 𝑏 2 × 𝑥2 + ⋯ + 𝑏 𝑛 × 𝑥𝑛 (3) 4: Compute ℎ = 𝑔1𝛼𝑑𝑔2𝑟.
𝑎1 𝑎1 𝑎1
If you set the vector 𝑏 = [𝑏 0 = a , 𝑏 1 = a , ⋯, 𝑏 n = a ], 5: Compute 𝐾 = 𝐻2(𝑎 || 𝑑 || ℎ).
0 0 0
then (𝐻(𝑤i)) = 1. 6: Compute 𝑆𝑘 = (ℎ, 𝐾).
Consider also that ℎ = [(𝑤1), (𝐻(𝑤2))2, ⋯, (𝐻(𝑤𝑛))𝑛]
7: Output 𝑆𝑘.
represents the vector of hash values, then it is easy to
verify that 𝑏 × ℎ = 1. If the keywords used in the data Algorithm 2 starts by choosing a random exponent
encryption process belong to a keyword 𝑠𝑒𝑡𝑊 = {𝑤1, 𝑟 ∈ 𝑍q∗. It then computes h = 𝑔1𝛼𝑑𝑔2𝑟, where 𝑔1 and 𝑔2 are
𝑤2, ⋯ 𝑤𝑛}, then equation 𝑏 × ℎ = 1 holds. the generators of 𝐺1 and 𝐺2 respectively. 𝛼𝑑 is the master
key component corresponding to the doctor attribute 𝑑. The
5. EHR Storage and Sharing Scheme algorithm computes 𝐾 = 𝐻2(𝑎 || 𝑑 || ℎ), where 𝐻2 is the hash
Based on Consortium Blockchain function defined in algorithm, 1 and || denotes concatenation.
Finally, the algorithm outputs 𝑆𝑘 = (ℎ, 𝐾).
The EHR storage and sharing scheme based on a consortium 3. Encryption: To encrypt the files based on the specified
blockchain can be divided into three phases: system access structure, the patient executes the 𝐸𝑛𝑐𝑟𝑦𝑝𝑡 algorithm
establishment, data encryption and storage, and data search with the system parameters 𝑃𝑃, the master key 𝑀𝑠𝑘, the file
and decryption. set 𝐹 = {𝑓1, 𝑓2 ⋯ 𝑓𝑚}, the keyword set 𝑊 = {𝑤1, 𝑤2 ⋯ 𝑤𝑛},
and a random value 𝜎 as inputs.
5.1 System Establishment
The algorithm outputs a tuple (𝑠𝑖𝑔, 𝐶, 𝐶̂, 𝐼, ∅𝑗) that contains
This phase consists of two steps: initialization and key
the signature 𝑠𝑖𝑔, ciphertext, 𝐶 indexed ciphertext 𝐶,̂ access
generation.
structure, (𝑓) and policy parameters ∅𝑗 for each file 𝑓𝑗.
1. Initialization: In this step, the system parameters are
Specifically, the Encrypt algorithm takes the system
generated. The input to this step is a security parameter 𝜆, and
parameters 𝑃𝑃, the master key 𝑘, the file set 𝐹, the keyword
the output is the system parameters 𝑃𝑃 and the Master key
set 𝑊 , and a random value 𝜎 as inputs. It generates a digital
𝑀𝑠𝑘. The Initialization algorithm in this scheme is described
signature 𝑠𝑖𝑔 for the file set 𝐹, a keyword index (𝑓𝑗) and an
in Algorithm 1 as follows:
encrypted file key 𝐸(𝑓𝑗) for each file 𝑓𝑗 using the CP-ABE
Algorithm 1: Initialization Step Algorithm algorithm with the access structure as the policy. It combines
1: Input: Security parameter 𝜆 the encrypted file key 𝐸(𝑓𝑗), the keyword index 𝐼(𝑓𝑗), and the
2: Output: System parameters 𝑃𝑃, Master key 𝑀𝑠𝑘 signature 𝑠𝑖𝑔(𝐹) to form the ciphertext 𝐶(𝑓𝑗). The indexed
ciphertext 𝐶(𝑓̂ 𝑦) is then encrypted using the symmetric key
3: Choose two cyclic groups 𝐺1 and 𝐺2 of prime order 𝑞,
and a bilinear map 𝑒: 𝐺1 × 𝐺1 → 𝐺2. 𝑘𝑠𝑦𝑚. The access structure 𝐼 and the policy parameters ∅
is then generated for the CP-ABE algorithm. The 𝐸𝑛𝑐𝑟𝑦𝑝𝑡
4: Choose a generator 𝑔 of 𝐺1.
algorithm outputs the tuple (𝑠𝑖(𝑓𝑗), 𝐶(𝑓𝑗), 𝐶(𝑓𝑗), 𝐼(𝑓𝑗), ∅𝑗)
5: Choose random exponents 𝛼, 𝛽 ∈ 𝑍∗. for each file 𝑓𝑗.
6: Compute 𝑔1 = 𝑔𝛼 and 𝑔2 = 𝑔𝛽 .
Steps to encrypt a file 𝑓𝑗 are (see algorithm 3):
7: Choose two hash functions 𝐻1: {0, 1}∗ → 𝐺1 and 𝐻2:
{0, 1}∗ → {0, 1}𝑛. (a) For each file 𝑓𝑗 in the file set 𝐹, the data owner (DO),
i.e. patient, generates a keyword index using the AES
8: Set 𝑃𝑃 = {𝐺1, 𝐺2, 𝑞, 𝑔, 𝑒, 𝐻1, 𝐻2} and 𝑀𝑠𝑘 = {𝛼, 𝛽 }.
algorithm. The keyword index is a binary string that
A Novel Blockchain-Based Approach for Secure and Efficient Electronic Medical Record Sharing 129
represents the presence or absence of each keyword in 6: Encrypt file key (𝑓𝑖) using CP-ABE algorithm with
the file. The keyword index is denoted as (𝑓𝑗). access structure as policy.
(b) DO generates a digital signature 𝑠𝑖𝑔 for the file set 𝐹 7: Combine 𝐸(𝑓𝑖), 𝐼(𝑓𝑗), and 𝑠𝑖𝑔(𝐹) to form ciphertext
using the master key 𝑀𝑠𝑘. The signature is denoted as 𝐶(𝑓𝑗).
𝑠𝑖(𝐹).
4. Trapdoor Generator: The trapdoor generator is a
(c) For each file 𝑓𝑗 in the file set 𝐹, DO executes the CP
component of the searchable encryption scheme that allows
ABE algorithm with the access structure as the policy
authorized users to generate trapdoors also called (search
to encrypt the file key. The access structure is a Boolean
tokens) for specific keywords, while preventing unauthorized
formula that specifies the attributes required to decrypt
users from doing so. The trapdoor generator is based on
the file. The encrypted file key is denoted as (𝑓𝑗).
the computational Diffie-Hellman (CDH) problem and the
(d) DO combines 𝐸(𝑓𝑗), 𝐼(𝑓𝑗), and the signature 𝑠𝑖(𝐹) to hash function collision resistance. In our approach, access
form the ciphertext 𝐶(𝑓𝑗). The binary string, known as control is enforced using attribute-based encryption (ABE)
the ciphertext (𝑓𝑗), not only symbolizes the encrypted and a master key 𝑀𝑠𝑘. The master key is used to generate
file key, but also encompasses the keyword index and the private keys for the different entities in the system, such
the file signature. The computation of the ciphertext as patients, doctors, and data users. Each entity is associated
(𝑓𝑗) follows a particular process. with a set of attributes, and access to the health records is
(𝑓𝑗) = (𝐸(𝑓𝑗), 𝐼(𝑓𝑗), 𝑠𝑖𝑔(𝐹)) (4) granted based on the attributes of the entity and the attributes
(e) The encryption process undertaken by DO is performed associated with the health records.
by using a symmetric key, denoted as 𝑘𝑠𝑦𝑚, to encrypt Assuming that the patient has the attribute 𝑎 and wants to
the indexed ciphertext 𝐶(𝑓 ̂ ).
𝑦 search for health records containing the keyword w, the
The aforementioned ciphertext, represented as a binary patient generates a random value 𝑟 ∈ 𝑍∗, where q𝑞 is a large
string, comprises the encrypted file key and the file’s prime number, and calculates the trapdoors as follows:
signature, both of which are organized and indexed by (a) The trapdoor algorithm begins with the user selecting
keywords. a keyword 𝑤 to search for in the electronic health
̂ ) = (𝐸(𝑓 ), 𝑠𝑖𝑔(𝐹)) ⊕ 𝑘 (𝐼(𝑓 ))
𝐶(𝑓 (5) records.
𝑦 𝑖 𝑠𝑦𝑚 𝑗
(f) DO generates the access structure 𝐼 and the policy (b) The authorized user generates a random value 𝑟 and
parameters ∅𝑗 for the CP-ABE algorithm. The access computes the hash value of the keyword w using a
structure 𝐼 is a Boolean formula that specifies the secure hash function. The hash value is denoted as
attributes required to decrypt the files. ∅𝑗 includes the 𝐻2(𝑤).
threshold value and the coefficients of the polynomial (c) Compute the Access Structure (𝐴𝑆) based on the
that defines the access policy. access control policy 𝐴𝐶 and the patient attributes 𝑎.
(g) DO outputs the tuple (𝑠𝑖𝑔, 𝐶, 𝐶,̂ 𝐼, ∅𝑗) for each file 𝑓𝑗. 𝐴𝑆 = 𝐶𝑜𝑚𝑝𝑢𝑡𝑒𝐴𝑐𝑐𝑒𝑠𝑠𝑆𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒(𝐴𝐶, 𝑎)
The 𝐸𝑛𝑐𝑟𝑦𝑝𝑡 algorithm ensures that only authorized users (d) Generate the patient’s private key (𝑃𝑟𝑎) and public key
with the correct attributes can decrypt and access the files (𝑃𝑏 𝑎) using the key generation function. (𝑃𝑟𝑎, 𝑃𝑏 𝑎) =
based on the specified access structure. To decrypt a file, a 𝐾𝑒𝑦𝐺𝑒𝑛(𝑆𝑘, 𝑀𝑠𝑘, 𝑎)
user must have the attributes that satisfy the access structure (e) Compute the temporary value (𝑇1) using the Diffie-
𝐼 and the policy parameters ∅𝑗. The user can use the CP-ABE Hellman problem and the public key (𝑃𝑏 𝑎) with the
algorithm to decrypt the encrypted file key (𝑓) using their generator (𝑔).
attributes. They can then use the decrypted file key to decrypt
the file using the symmetric key 𝑘𝑠𝑦𝑚. 𝑇1 = 𝑔𝑃𝑏 𝑎 (6)
(f) Compute the hash value (𝐻) of the concatenation of
Algorithm 3 𝐸𝑛𝑐𝑟𝑦𝑝𝑡 Algorithm 𝐴𝑆, 𝑟, and 𝑇1 using the hash function 𝐻.
1: Input: System parameters 𝑃𝑃, the master key 𝑀𝑠𝑘, the 𝐻 = (𝐴𝑆 || 𝑟 || 𝑇1) (7)
file set 𝐹, the keyword set 𝑊 , and a random value 𝜎.
(g) Compute the second part of the trapdoor (𝑇2) using the
2: Output: (𝑠𝑖𝑔, 𝐶, 𝐶,̂ 𝐼, ∅𝑗) hash function (𝐻2), the random value (𝑟), the patient’s
3: Generate digital signature 𝑠𝑖𝑔 for file set 𝐹 using the private key (𝑃𝑟𝑎), and the generator (𝑔).
master key 𝑀𝑠𝑘.
𝑇2 = 𝐻2(𝑟, 𝑃𝑟𝑎)(1/𝑃𝑟𝑎) × (𝑃𝑟𝑎)𝑟 × 𝑔𝑤 (8)
4: for each file 𝑓𝑗 do
(h) Compute the final trapdoor (𝑇2) by raising 𝑇2 to the
5: Generate keyword index (𝑓𝑗). power of 𝐻.
𝑇2 = 𝑇2𝐻 (9)
130 Algorithms in Advanced Artificial Intelligence
(i) Encrypt the trapdoor (𝑇1, 𝑇2) using the 𝐸𝑛𝑐𝑟𝑦𝑝𝑡 6. Cℎeck negation of attribute: If a satisfies the attribute in
function. 𝑐, the Result is set to 𝐹𝑎𝑙𝑠𝑒 and the loop is terminated.
In attribute-based encryption (ABE) schemes, the construction 7. Return Result: The function returns the Result, which
of the Access Structure (𝐴𝑆) ensures that only authorized is 𝑇𝑟𝑢𝑒 if the access control policy is satisfied by the
users with the necessary attributes can access sensitive patient attributes, and 𝐹𝑎𝑙𝑠𝑒 otherwise.
information. It chooses the characteristics that are necessary
Algorithm 4 𝐸𝑣𝑎𝑙𝑢𝑎𝑡𝑒(𝐴𝐶, 𝑎) function
to decrypt the encrypted data. Based on the access control
policy 𝐴𝐶 and the patient attributes 𝑎, 𝐴𝑆 is calculated. 1: Input: Access Control policy 𝐴𝐶, the patient attributes
𝑎, attributes condition 𝑐.
In order to calculate 𝐴𝑆, the access control policy is assessed
using the user attributes. Logical operators like AND, OR, and 2: Output: 𝑇𝑟𝑢𝑒 if 𝐴𝐶 is satisfied by the patient attributes
NOT can be used to carry out this evaluation. The following 𝐴 and 𝐹𝑎𝑙𝑠𝑒 otherwise).
equation is used to calculate 𝐴𝑆: 3: 𝑅𝑒𝑠𝑢𝑙𝑡 = 𝑇𝑟𝑢𝑒
𝐴𝑆 = 𝐸𝑣𝑎𝑙𝑢𝑎𝑡(𝐴𝐶, 𝑎) (10) 4: for condition 𝑐 in 𝐴𝐶 do
5: if 𝑐 is a single attribute then
Now, let’s break down the equation and explain each part:
6: if 𝑎 does not satisfy 𝑐 then
• The 𝐸𝑣𝑎𝑙𝑢𝑎𝑡(𝐴𝐶, 𝑎) function takes as input the user
7: 𝑅𝑒𝑠𝑢𝑙𝑡 = 𝐹𝑎𝑙𝑠𝑒
attributes (𝑎) and the access control policy (𝐴𝐶). In order
to combine and compare the attributes, the evaluation 8: 𝐵𝑟𝑒𝑎𝑘
function may utilize logical operations like AND, OR, 9: else if 𝑐 = 𝐴𝑁𝐷(𝑎) then
and NOT. A boolean value indicating if the access control 10: if 𝑎 does not satisfy all attributes in 𝑐 then
policy is met based on the user attributes is the function’s 11: 𝑅𝑒𝑠𝑢𝑙𝑡 = 𝐹𝑎𝑙𝑠𝑒
output. The evaluation may involve logical operations 12: 𝐵𝑟𝑒𝑎𝑘
like AND, OR, and NOT to combine and compare the
13: end if
attributes. The output of this function is a boolean value
indicating whether the access control policy is satisfied 14: else if 𝑐 = 𝑂 𝑅(𝑎) then
based on the user attributes. 15: if 𝑎 does not satisfy at least one attribute in 𝑐 then
• The Access Structure, which is the result of the 16: 𝑅𝑒𝑠𝑢𝑙𝑡 = 𝐹𝑎𝑙𝑠𝑒
examination of the access control policy, is represented 17: 𝐵𝑟𝑒𝑎𝑘
by item 𝐴𝑆. 𝐴𝑆 is regarded as true or met if the user 18: else if 𝑐 = 𝑁𝑂 𝑇(𝑎) then
attributes satisfy the access control policy or false if the 20: if 𝑎 satisfies the attribute in 𝑐 then
user attributes do not satisfy the access control policy.
21: Result = False
Algorithm (4) describes the 𝐸𝑣𝑎𝑙𝑢𝑎𝑡(𝐴𝐶, 𝑎) function: 22: Break
1. Initialization: The function takes the patient attributes 𝑎 23: end if
and the access control policy 𝐴𝐶 as inputs. The result’s 24: end if
initial value is set to 𝑇𝑟𝑢𝑒.
25: end if
2. Loop over conditions: Each condition 𝑐 in the access
26: end for
control policy 𝐴𝐶 is iterated over by the function. Textit
attributes include a single attribute, a conjunction of 27: return 𝑅𝑒𝑠𝑢𝑙𝑡
attributes (AND), a disjunction of attributes (OR), and Algorithm 5 explains the function
a negation of an attribute (NOT). 𝐶𝑜𝑚𝑝𝑢𝑡𝑒𝐴𝑐𝑐𝑒𝑠𝑠𝑆𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒(𝐴𝐶, 𝑎).
3. Cℎeck single attribute: The 𝐸𝑣𝑎𝑙𝑢𝑎𝑡𝑒 function Algorithm 5 𝐶𝑜𝑚𝑝𝑢𝑡𝑒𝐴𝑐𝑐𝑒𝑠𝑠𝑆𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒(𝐴𝐶, 𝑎) function
determines whether the patient attributes satisfy the 1: 𝐴𝑆 = 𝐹𝑎𝑙𝑠𝑒
single attribute in 𝑐. The Result is set to 𝐹𝑎𝑙𝑠𝑒 and the
2: if 𝐴𝐶 is empty then
loop is ended if 𝑎 does not fulfill the attribute.
3: return 𝐴𝑆
4. Cℎeck conjunction of attributes: If 𝑎 satisfies all of
the combination of attributes in 𝑐, the Result is set to 4: end if
𝑇𝑅𝑈𝐸; otherwise the Result is set to False and the loop 5: if 𝑎 is empty then
is ended. 6: return 𝐴𝑆
5. Cℎeck disjunction of attributes: if 𝑎 fulfills at least one 7: end if
attribute in 𝑐, the Result is set to 𝑇𝑅𝑈𝐸; otherwise, 8: 𝐴𝑆 = 𝐸𝑣𝑎𝑙𝑢𝑎𝑡𝑒(𝐴𝐶, 𝑎)
Result is set to 𝐹𝑎𝑙𝑠𝑒 and the loop is terminated. 9: return 𝐴𝑆
A Novel Blockchain-Based Approach for Secure and Efficient Electronic Medical Record Sharing 131
5.2 Data Generation and Storage in EMR 2. The patient generates a secret key 𝑆𝑘 by using the
(DGS- EMR) 𝐾𝑒𝑦𝐺𝑒𝑛 function and stores the key on the consortium
blockchain.
Data Generation and Storage in EMR (DGS- EMR), begins
with the patient generating a secret key 𝑆𝑘 using the 𝐾𝑒𝑦𝐺𝑒𝑛 3. When the patient requires access to their EMRs, they
function. The patient then registers and uploads secret key to make a request to the hospital by providing their patient
the consortium blockchain. ID.
4. The hospital confirms the patient’s identity and retrieves
In the Data Generation and Storage in EMR (DGS- EMR)
the patient’s EMRs from the private blockchain.
algorithm, the system parameters are first initialized using
algorithm 1. The patient generates a secret key 𝑆𝑘 using 5. The EMRs and the generated search trapdoors are
the 𝐾𝑒𝑦𝐺𝑒𝑛 function and then registers and uploads secret encrypted using the Encrypt function with the CPABE
key to the consortium blockchain. When the patient requires algorithm.
access to their electronic medical records (EMRs), they make 6. The access structure is computed using the
a request by providing their patient ID to the hospital. The 𝐶𝑜𝑚𝑝𝑢𝑡𝑒𝐴𝑐𝑐𝑒𝑠𝑠𝑆𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒 function; and the access
hospital next confirms the patient’s identity before retrieving control policy is evaluated using the 𝐸𝑣𝑎𝑙𝑢𝑎𝑡𝑒 function
their electronic medical records from the private blockchain. with the patient’s attributes 𝑎 and the Access Control
policy 𝐴𝐶.
The EMRs and the generated search trapdoors are encrypted
using the 𝐸𝑛𝑐𝑟𝑦𝑝𝑡 function (algorithm 3), and the access 7. The encrypted EMRs and search trapdoors are stored
structure is computed using the 𝐶𝑜𝑚𝑝𝑢𝑡𝑒𝐴𝑐𝑐𝑒𝑠𝑠𝑆𝑡𝑟𝑢𝑐𝑡𝑢𝑟𝑒 on IPFS.
function (algorithm 5). The access control policy is evaluated 8. The data user node on the consortium blockchain
using the 𝐸𝑣𝑎𝑙𝑢𝑎𝑡𝑒 function (algorithm 5). The encrypted receives the trapdoor transmitted by the patient and
EMRs and search trapdoors are stored on IPFS. Finally, the executes the search algorithm.
encrypted EMRs and search trapdoors stored on IPFS. 9. After a successful search, the data user node obtains
The patient goes through a two-step process to make sure the security index and private blockchain ID from the
that his/her EMRs are shared securely with authorized consortium blockchain.
users. The search trapdoor is first generated and uploaded 10. The data user node retrieves the hash value of the
to the consortium chain. By using this trapdoor on the encrypted medical record from the private blockchain
same consortium chain, the authorized user can search for using the private chain block identification.
and retrieve the desired EMRs. A private blockchain uses a 11. The data user node sends the hash value of the
“append-only” method to run its structure. This ensures the encrypted medical record back to the hospital server
integrity and consistency of the recorded data since once a for verification.
block is added to the chain, it cannot be changed or removed. 12. The hospital server compares the hash value received
Using a consensus mechanisms like Proof of Work (PoW) or from the data user node with the one associated with
Proof of Stake (PoS) can confirm and validate the transactions the electronic medical record ciphertext to ensure their
and blocks added to the blockchain. These procedures make consistency.
sure that all participants in the private blockchain agree that 13. If there is a match, the hospital server sends the
the newly added data is accurate, and hence maintaining encrypted medical record ciphertext back to the data
consistency. user node.
The private chain secures the storage of the ciphertext, hash 14. In cases where a third-party data user needs access to
value, and keyword index of records. To ensure confidentiality a patient’s electronic medical record, the consortium
and integrity, encryption techniques are used, allowing chain node acts as an intermediary.
authorized entities to access and modify the data as needed.
Theorem 1. Tℎe DGS-EMR blockcℎain-based metℎod using
Hence, the DGS-EMR blockchain-based method offers a attribute-based encryption, searcℎable encryption, and
secure and efficient way to share electronic medical records robust access control mecℎanisms provides a secure and
while maintaining patient privacy and guaranteeing data efficient way to sℎare medical records wℎile ensuring privacy,
integrity. confidentiality, and integrity.
The steps related to the DGS-EMR are as follows: Proof. To be proved.
1. Use the system initialization algorithm that takes the
security parameter 𝜆 to generate the system parameters
𝑃𝑃 and the master key 𝑀𝑠𝑘.
132 Algorithms in Advanced Artificial Intelligence
Abstract: In today’s culture, images and videos are crucial for effective work and security surveillance. By studying CCTV
data, a prediction algorithm may ascertain a person’s age, gender, location, and sexual orientation. This technology can make
the world safer by enabling the identification of runaways. The technology, integrated into security cameras within a mile, can
screen suspects, such as a fugitive who stole millions from a typical bank. The paper explores the development of technologies
that can determine a person’s age, face image, gender, and location through cameras, pictures, or videos Using Deep Learning
techniques. A machine learning algorithm AdaBoost is for binary classification tasks, combining weak classifier predictions to
create a strong classifier that performs well on the given data. The research focuses on data parallelism and model parallelism
as two methods for distributing backpropagation computation among GPUs or nodes using Distributed Back Propagation. The
study compares AdaBoost and Back Propagation in gender crime classification. The study investigates technologies like t-SNE,
PCA, and ICA in daily life and their potential applications in criminal face identification, emphasizing the need for precise
experimental results.
Keywords: AdaBoost Classification, Crime Classification, Distributed Back Propagation, t-SNE, PCA, ICA, etc.
1
dileep.kadali@giet.edu, dileepkumarkadali@gmail.com; 2mohanrnvj@gmail.com; 3srichandra2007@gmail.com
DOI: 10.1201/9781003529231-21
134 Algorithms in Advanced Artificial Intelligence
accounted for 30% of the total, followed by property crimes distributed back propagation method is proposed method is
(20.8%) and other IPC offences (29.7%). Kidnapping was the covered in section 2.2., section 2.3 compares the image of
highest crime rate at 2.1 per 100,000. 7.4 per 100,000, and crime persons and the experimental results in section 3. The
rape 4.8 per 100,000. The UN reported a homicide rate of final one is section 4’s conclusion, which section 5 references.
2.95 per 100,000 in 2020, down from 5.46 per 100,000 in
1992 Jha, 2019[8]. The rate of investigation for IPC crimes 2. Proposed Work
is 64.9% in 2021, with a charge sheet rate of 72.3% and a
conviction rate of 57.0%. compares AdaBoost and Back Crime Person classification faces challenges due to large
Propagation in gender crime classification and investigates crime person images, high-dimensional data, and a lack of
technologies like t-SNE, PCA, and ICA in daily life and their labelled data, as each image contains numerous features
potential applications in criminal face identification. The and lacks explained data Shukla, 2020[13]. Our picture
study emphasizes the need for precise experimental results. categorization system should adapt to illumination variations,
The NCRB’s report compared crime rates between 1953 assigning the same label to two pictures of the same item
and 2006 in India. The data indicated a drop of 79.84% in with varying brightness levels, ensuring they are categorized
burglaries, a rise of 7.39% in murders, a 47.80% increase in accurately. The paper investigates the AdaBoost algorithm’s
kidnappings, a 28.85% decline in robberies, and a 10.58% effectiveness in gender crime classification, focusing on
decline in riots. There were 5,102,460 cognizable crimes image classification using t-SNE, PCA, and ICA on a law-
in 2006, up 1.5% from 2005. These included 1,878,293 enforcement dataset working on Data parallelism and
offences under the Indian Penal Code (IPC) and 3,224,167 model parallelism are the two approaches for distributing
crimes under the Special and Local Laws (SLL). Delhi saw backpropagation computation among GPUs or nodes.
the most increase in crime in 2019 out of all Indian states,
going from 1342.5 to 1586.1. Northeast India had the lowest 2.1 Understanding Face Person Classification
crime rates in 2018, with four out of five states having the AdaBoost algorithm
lowest. Uttar Pradesh reported the most crimes, while A machine learning algorithm AdaBoost is for binary
Maharashtra and Kerala had the lowest. Violent crime rates classification tasks, combining weak classifier predictions to
were highest in Assam, Tripura, Haryana, West Bengal, and create a strong classifier that performs well on the given data.
Arunachal Pradesh. Kolkata was the safest city in India in AdaBoost focuses on misclassified data points in classifiers,
2021, but experts doubt its accuracy by Kadar, 2019[9]. Out increasing their importance in training. This adaptive
of 19 cities with more than two million residents, Pune and process improves classification performance by focusing on
Hyderabad had the lowest rates of crime. The only megacities challenging data points, leading to highly accurate results
with lower crime rates than their respective states have been when combined with multiple weak classifiers by Azwad
Mumbai and Kolkata. For the fourth consecutive year, Delhi Tamir et al, 2021[2].
was India’s most criminalized metropolitan area, accounting
for more than 82% of 290,000 recorded crimes. Kochi
continued to rank second in their jurisdiction for the most
cases of reckless driving. For the second year, Jaipur had
the third-highest crime rate. The Statistics Knowledge Act
2000, passed in 2000 to regulate cyber crimes and facilitate
e-commerce, has been criticized for not effectively addressing
emerging crimes like cyber harassment and defamation. In
2021, 52,974 cybercrime cases were registered in India, with
Telangana reporting the highest number. These crimes were Fig. 21.1 Crime classification of AdaBoost algorithm
motivated by deception, which was followed by extortion and
sexual exploitation. In terms of cybercrimes against health The AdaBoost method involves initialization, repeated
systems, India came in second place worldwide. Personal iterations, classifier weight computation, weight update,
data, hospital log-in credentials, and immunization records ensemble construction, weight normalization, final
were among the breaches that were compromised. In India, classification, and final ensemble. It uses a weak classifier
preventing crime is essential to upholding law and order, to focus on incorrectly labelled data points, with weights
but other factors that may affect crime rates include poverty, adjusted based on accuracy by Anshul Saini, 2023[1]. The
unemployment, and lower per capita income. Z. Li, 2021 training dataset is weighted with identical weights, with a
[11]. The paper is set up as follows: The introduction is in weak classifier focusing on incorrectly labelled data points
Part 1. This article’s Part 2 proposed work in 2.1 provides by E. Ahishakiye,2017[5]. Weights are adjusted based on
an AdaBoost algorithm of age and location for a face. The categorization accuracy. Each weak classifier contributes
A Classifying Gender Crimes with AdaBoost and Back Propagation Algorithms 135
to the final ensemble by how well it performs. On fresh, person’s image data points. The initial assumption was for a
unforeseen data, predictions are made using the final non-symmetric Gaussian distribution for gender image data
ensemble. point pair distances in both original and reduced spaces. They
then used a Cauchy distribution, t-SNE, to model pairs in the
2.2 Gender Crime Classification Using reduced space and minimize the KL distance between the two
Distributed Back Propagation Process spaces, preserving the original pairwise proximity. t-SNE is
Data parallelism and model parallelism are the two now accessible in sci-kit-learn and Tensor Flow’s embedding
approaches for distributing backpropagation computation toolbox, enhancing data visualization tools despite being less
among GPUs or nodes. In data parallelism, the model weights glamorous than Deep Learning discoveries.
are duplicated into many processes utilizing different pieces
of hardware, with a parameter server serving as the source of 3. Experimental Result
truth. Each model gets its mini-batch of gender crime data
by Bowen, 2018[4], runs forward and backward passes, and AdaBoost predicts a person’s age and location using a law-
calculates gradients. The gradients are then averaged and enforcement dataset with decisions as weak classifiers in the
distributed once again to all worker nodes. Each iteration of experiment.
decentralized back propagation employs a distinct mini-batch Step-1 Initialization: The training dataset consists of 10 data
of data, with a master process broadcasting model weight. points and their corresponding labels (1 for “criminal” and -1
This approach may result in a quicker implementation of the for “not criminal”).
algorithm by Soni Upadhya, 2023[14].
Table 21.1 Training Data Set
2.3 Comparative Analysis of Crime Person
Data point Age Label
Classification
1 21 1
Initially naive of t-SNE, it gained popularity a few years ago
2 12 -1
and is now utilized in visually appealing plots. t-SNE, similar
3 25 1
to PCA or ICA, is a dimensionality reduction algorithm that
Gender crime classification data into a subspace that captures 4 11 -1
more intrinsic crime data meaning than the original space. 5 24 1
The three main statistical analysis methods are PCA, ICA, 6 13 -1
and t-SNE by Sarvakar Siva, 2020[15]. PCA maximizes 7 28 1
variance in each direction, resulting in statistically orthogonal
8 10 -1
new variables. ICA minimizes pairwise covariate Mutual
Information, resulting in independent components. t-SNE 9 32 1
maximizes the similarity between the original and new crime 10 10 -1
136 Algorithms in Advanced Artificial Intelligence
• Ensemble prediction = sign(Σ(weight * classifier The ICA is a technique that reduces extraneous noise in input
prediction)) criminal data to identify independent components. If linear
• Ensemble prediction ≈ sign(0.693 * (-1) + 0.693 * 1) = and nonlinear dependencies of two input features are zero,
sign(0) = -1 they are considered independent. ICA can be used to identify
distinct input data in an authored, condensing the dataset to
Step-7 Final Classification: The ensemble predicts that the
three.
person will “not be a criminal” the person.
In the following figure, class-0 represents age, class-1
Step-8 Next Iteration and Final Result: AdaBoost uses
represents label and axes are represented like X-axes as C1,
adaptive boosting of misclassified data points to create a
Y-axes as C2 and Z-axes as C3.
strong ensemble classifier. Iterations focus on misclassified
data points, with weights adjusted for the next weak classifier.
The final prediction is a sum of all weak classifier predictions.
AdaBoost is a robust ensemble-learning algorithm that
enhances model accuracy, but it’s noise sensitivity and weak
classifier selection are limitations, necessitating careful
consideration.
References
1. Anshul Saini: AdaBoost Algorithm: Understand, Implement
and Master AdaBoost, Analytics, September 21st, 2023.
2. Azwad Tamir et al: Crime Prediction and Forecasting using
Machine Learning Algorithms, International Journal of
Computer Science and Information Technologies, Vol. 12 (2),
26–33, 2021.
3. Bandekar, S. R., & Vijayalakshmi, C: Design and analysis of
machine learning algorithms for the reduction of crime rates
in India, Procedia Computer Science, 172, 122–127. https://
Fig. 21.7 t-distributed stochastic neighbor embedding (t-SNE) doi.org/10.1016/j.procs.2020.05.018,2020.
on crime persons 4. Bowen, D. A., Mercer Kollar, L. M., Wu, D. T., Fraser, D.
A., Flood, C. E., Moore, J. C., Mays, E. W., & Sumner, S. A:
Ability of crime, demographic and business data to forecast
ICA could be valuable assuming there are free wellsprings of
areas of increased violence. International Journal of Injury
crime that are not directly related. Nonetheless, interpretability
Control and Safety Promotion, 25(4), 443–448, https://doi.org
may be a test as the parts address genuinely free sources, which /10.1080/17457300.2018.1467461,2018.
may not straightforwardly compare to effectively justifiable 5. E. Ahishakiye, E. Opiyo, and I. Niyonzima: Crime Prediction
wrongdoing designs. Spotlights on catching autonomous Using Decision Tree (J48) Classification Algorithm,
wellsprings of variety, possibly uncovering particular sorts International Journal of Computer and Information
of crimes. It can be computationally costly, particularly for Technology (ISSN: 2279 – 0764), 05/15, 2017.
countless autonomous parts. 6. Forradellas, R. F. R., Alonso, S. L. N., Rodriguez, M. L.,
& Jorge-Vazquez, J. (2021). Applied machine learning in
t-SNE is successful in envisioning complex, non-straight
social sciences: Neural networks and crime prediction.
connections. It tends to be advantageous for recognizing Social Sciences, 10(1), 1–20. https://doi.org/10.3390/
neighbourhood examples or bunches of comparative socsci10010004,2020.
wrongdoing events however may miss the mark on the 7. Gao, Y., Wang, X., Chen, Q., Guo, Y., Yang, Q., Yang, K.,
worldwide interpretability of PCA. Stresses the conservation & Fang, T: Suspects prediction towards terrorist attacks
of neighbourhood connections, which can be significant based on machine learning. In Proceedings – 2019 5th
for distinguishing explicit confined examples or bunches of international conference on big data and information analytics,
crime. Computationally costly, particularly for huge datasets, BigDIA 2019 (pp. 126–131). https://doi.org/10.1109/
which might restrict its plausibility for broad wrongdoing BigDIA.2019.8802726,2019.
information. 8. Jha, G., Ahuja, L., & Rana, A: Criminal behavior analysis
and segmentation using K-means clustering. ICRITO
PCA is reasonable for catching crime data, ICA for 2020 - IEEE 8th International Conference on Reliability,
distinguishing free wellsprings of crime, and t-SNE for Infocom Technologies and Optimization (Trends and
envisioning restricted examples or bunches in a non-direct Future Directions), 1356–1360. https://doi.org/10.1109/
design. The decision ought to line up with the objectives ICRITO48877.2020.9197791,2019.
of the investigation and the attributes of the wrongdoing 9. Kadar, C., Maculan, R., & Feuerriegel, S: Public decision
information being analyzed. support for low population density areas: An imbalance-
aware hyper-ensemble for spatio-temporal crime prediction,
Decision Support Systems, 119, 107–117, https://doi.
4. Conclusion org/10.1016/j.dss.2019.03.001,2019.
10. Kim, S., Joshi, P., Kalsi, P. S., & Taheri, P: Crime
Advancements in technologies enable the detection of an
analysis through machine learning. In 2018 IEEE 9th
individual’s age, gender, and location through cameras, annual information technology, electronics and mobile
pictures, or videos using the Backpropagation technique. It communication conference, IEMCON 2018 (pp. 415–420),
highlights the importance of these technologies in daily life https://doi.org/10.1109/IEMCON.2018.8614828,2019.
and their possible applications in criminal identification and
A Classifying Gender Crimes with AdaBoost and Back Propagation Algorithms 139
11. Li, Z., Zhang, T., Jing, X., & Wang, Y: Facial expression 13. Shukla, S., Jain, P. K., Babu, C. R., & Pamula, R: A multivariate
based analysis on emotion correlations, hotspots, and regression model for identifying, analyzing and predicting
potential occurrence of urban crimes. Alexandria Engineering crimes, Wireless Personal Communications, 113(4), 2447–
Journal, 60(1), 1411–1420. https://doi.org/10.1016/j. 2461, https://doi.org/10.1007/s11277-020-07335-w,2020.
aej.2020.10.061,2021. 14. Soni Upadhyay: What is Back propagation Algorithms? Types
12. Saravanan, P., Selvaprabu, J., Arun Raj, L., Abdul Azeez and Examples and in its, Simple Learn, Aug 30, 2023.
Khan, A., & Javubar Sathick, K.: Survey on crime analysis 15. Sarvakar Siva: Dimensionality Reduction for Data
and prediction using data mining and machine learning Visualization: PCA vs TSNE vs UMAP vs LDA, Published in
techniques. Lecture Notes in Electrical Engineering, 688, 435– Towards Data Science, 2020.
448. https://doi.org/10.1007/978-981-15-7241-8_3,2021.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
140 Algorithms in Advanced Artificial Intelligence
P. Sumithabhashini*
Professor, Department of Electronics & Communication Engineering,
Holy Mary Institute of Technology & Science, Hyderabad, India
M. V. Vijaya Saradhi2
Professor, HOD CSE & CSE(IoT), Department of Computer Science & Engineering,
ACE Engineering College, Hyderabad, India
Ramesh Alladi
Associate Professor, Department of Computer Science & Engineering,
ACE Engineering College, Hyderabad, India
Swajan Reddy
Masters statistics and data science, University of Houston,
Houston, Texas, USA
Abstract: Neurological problems, resulting from genetic disorders, congenital abnormalities, infections, lifestyle, environmental
issues, malnutrition, and brain injury recognized neurological diseases. Gesture Recognition for Image Processing uses
artificial intelligence (AI) and machine learning for classification, heat/motion detection, and movement analysis. It is based
on advanced technologies and algorithms. The application of finger imaging tremor quantification to neurological disorder
diagnosis is covered in the study. This research presents a hybrid strategy to quantify and objectively assess shake, specifically
for cerebella diseases and finger shake, combining imaging technologies and machine learning approaches. The technique uses
a Gesture Finger pictures Process of Tremor Disease Detection for AI along with image processing to differentiate between
people who are healthy and those who have crucial shaking.
Keywords: Artificial intelligence, CNN, Finger gesture images, Tremor quantification etc.
DOI: 10.1201/9781003529231-22
Identifying Tremor Disease in Neurological Disorders Using Finger Gesture Images 141
Tremors can develop from shaking, which is a typical and Section 3 details the experimental results. Finally, section
indication of stress, rage, or sickness. More people over the 5 alludes to the conclusion of section 4.
age of 40 suffer from essential shaking, a tremor disorder
that can impact several regions of the body, including the 2. Proposed Work
hands and arms [1].Toxins in the environment and being
older are both risk factors. Deep brain stimulation or surgery The study explores the use of AI and machine learning in
may be necessary to treat essential tremors. It is prevalent Gesture Recognition for Image Processing, specifically for
in MS patients and serves as an early warning signal of the diagnosing neurological disorders like cerebella diseases.
disease [3–6]. Tremors at rest, as in Parkinson’s disease and It uses advanced technologies to quantify and assess shake,
medication-induced tremors. distinguishing between healthy and crucial individuals. The
Muscle relaxation and resistance to gravity are the root causes use of finger gesture images for tremor quantification in
of several neurological disorders, including progressive diagnosing neurological disorders.
supranuclear palsy, dystonia, rubral tremor, and Wilson’s To analyze and understand human movements while
disease. A few days—or more if drinking excessively or utilizing machine learning and Artificial Intelligence (AI) for
for an extended period of time—long tremors are the first classification and heat/motion disease detection.
symptoms of alcohol withdrawal. Medications that inhibit Gesture Image Using Diffusion model: Diffusion is the
dopamine, a lack of vitamin B12, caffeine, or anxiety can movement of energy from higher to lower concentrations,
all lead to trembling hands. B12 insufficiency impacts the driven by Gibbs free energy or chemical potential. It can
neurological system, whereas medications assist with mood be “uphill” and is a stochastic process, used in fields like
maintenance [7–10]. Drinks like coffee and tea might make statistics, probability theory, information theory, neural
your hands tremble. Several forms of stress, such as worries networks, finance, and marketing.
about money, work, relationships, or health, can exacerbate
tremors. Additional causes of physiologic tremors include The process involves adding noise to a gesture image and
excessive rage, hunger, or lack of sleep. A low blood sugar learning to remove it, then training a machine learning model
level, or hypoglycemia, causes trembling, causing the body to produce a denoised gesture image [17].
to go into a stress response. The symptoms of an overactive The process of learning the mean matrix involves assuming a
thyroid gland in the neck include an irregular heartbeat, normal noise distribution and parametrizing the distribution
tremors, and difficulty sleeping. Hand and foot tremors can mean and standard deviation matrix. This can be divided into
be brought on by a nerve injury. Talk to your doctor about a forward and reverse process. Mathematicians often use
your symptoms and medical history; treatments could differ physical processes to formalize mathematical concepts, such
[11–14].Despite the fact that most videos contain small- as Fick diffusion, heat diffusion, and Brownian motion, by
amplitude tremors, Xini Wang et al. suggested in their 2021 defining the diffusion equation, which equals the first and
article “Hand tremor detection in videos with a cluttered second space derivatives.
background using neural network-based approaches” that it The diffusion equation, a stochastic formulation, is based
is possible to detect hand tremors in videos with a cluttered on the Langevin equation, which is centered on the Wiener
background with a high degree of accuracy [16]. Possible real- process, also known as Brownian motion. This process, also
world uses for this video-based tremor detection technology known as a Random Walk, is normal distributed, making
include healthcare facilities and home monitoring systems. diffusion models intertwined with white noise generation.
Using this technology, researchers and physicians can use
their own smartphones to identify tremor automatically, even
in busy backgrounds. The design, which is based on neural
networks, learns from various training datasets; thus, it may
also be able to identify tremors when hands are moving. One
possible drawback is that it may require a lot of memory,
computing power, hidden layers, and low-resolution films.
Evaluating the system in larger cohorts and assessing factors
like skin colour, ambient lighting, and camera resolution
requires additional research.Here is the structure of the paper:
Part 1 contains the introduction. In order to eliminate picture
noise, the proposed work in Section 2.1 of this paper offers a
gesture image using a diffusion model. Section 2.2 details the
gestural finger image processing for tremor disease detection, Fig. 22.1 Gesture image using diffusion models
142 Algorithms in Advanced Artificial Intelligence
The diffusion models, which utilize a Gaussian prior to by learning vectorial representations of photos and text,
generate data, form the core of text-to-image generative respectively. In order to achieve semantic alignment [15],
models. training vector representations in tandem with picture and text
Diffusion models offer a wide range of degrees of freedom models is necessary. By indexing them into a library of vector
in implementation, allowing for the choice of variances for images of gesture fingers, we hope to quickly find relevant
the forward process and the model architecture and Gaussian ones for the purpose of tremor disease identification [18] and
distribution parameterization for the reverse process. A new store them for future vectorizing ideas. The method of tremor
explicit connection between diffusion models and Denoising quantification is granted by the image-to-text conversion.
score matching leads to a simplified, weighted variation The goal at hand is to compile a set of finger gesture graphics
bound objective for diffusion models. The model design based on the user’s input query. While one provider creates
is justified by simplicity. The forward process variances an embedded encoding, another uses a superior model to
are fixed to constants, while the reverse process entropy is re-rank the closest neighbours, making the ranking more
optimized. The study demonstrates that the reverse process personalised for the user.
mean function approximate can predict with a prediction
parameterization, simplifying the diffusion model’s variation 3. Experimental Result
bound and resembling Denoising score matching.
This study investigates the feasibility of quantifying tremor in
Gesture Finger images Process of Tremor Disease neurological illness imaging using finger scans. Withdrawal
Detection: Machine Search engines that work with gesture from alcohol or narcotics, a lack of vitamin B12, caffeine,
images rank relevant language and the most comparable tension, rage, hunger, insomnia, low blood sugar, or an
photos and then display them to the user. Models learn to overactive thyroid gland are all potential causes of tremors.
extract image characteristics and text features from inputs This research looks at the potential of using pictures of finger
Fig. 22.2 The machine finger gesture images search engine is being developed to detect tremor diseases
Identifying Tremor Disease in Neurological Disorders Using Finger Gesture Images 143
Tremor Other Hyperkinetic Mov. 2018; 8: 517. doi: 10.5334/ 15. Saurabh Adhikari et el: A Novel Machine Learning–Based
tohm.400. Hand Gesture Recognition Using HCI on IoT Assisted Cloud
9. Ishii N., Mochizuki Y., Shiomi K., Nakazato M., Mochizuki Platform, Computer Systems Science & Engineering DOI:
H. Spiral drawing: Quantitative analysis and artificial 10.32604/csse.2023.034431, Tech Science Press, CSSE, vol.
intelligence-based diagnosis using a smartphone. J. Neurol. 46, no. 2, 2023.
Sci. 2020; 411: 116723. doi: 10.1016/j.jns.2020.116723. 16. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
10. Kamble N., Pal P.K. Tremor syndromes: A review. Neurol. in Health Care Using Hybrid Model Algorithms”,Journal of
India. 2018; 66: 36–47. doi: 10.4103/0028-3886.226440. Artificial Intelligence, Machine Learning and Neural Network
11. Louis E.D., Faust P.L. Essential tremor: The most common (JAIMLNN), Volume 3, 2023.
form of cerebellar degeneration? Cerebellum Ataxias. 2020; 17. Xini wang et al:Hand tremor detection in videos with cluttered
7: 1–10, doi: 10.1186/s40673-020-00121-1. background using neural network based approaches, Health
12. Mansur P.H.G., Cury L.K.P., Andrade A.O., Pereira A.A., Information Science and Systems, Springer Nature, 2021.
Miotto G.A.A., Soares A.B., Naves E.L. A review on techniques 18. Yutong Xie, Minne Yuan, Bin Dong, Quanzheng Li: Diffusion
for tremor recording and quantification. Crit. Rev. Biomed. Model for Generative Image Denoising, arXiv: 2302.02398,
Eng. 2007; 35: 343–362. doi: 10.1615/CritRevBiomedEng. https://doi.org/10.48550/arXiv.2302.02398, 2023.
v35.i5.10. 19. Zdenka U., Otakar S., Martina H., Arnost K., Olga U.,
13. Mitsui Y., Ishii N., Mochizuki H., Zin T.T. A Study on Disease Václav H., Chris D.N., Evzen R. Validation of a new tool
Diagnosis by Tremor Analysis. Int. Multi Conf. Eng. Comput. for automatic assessment of tremor frequency from video
Sci. 2018; 1: 14–16. recordings. J. Neurosci. Methods.2011;198:110–113,
14. Sharma S., Pandey S. Approach to a tremor patient. Ann. doi: 10.1016/ j.jneumeth.2011.02.033, 2011.
Indian Acad. Neurol. 2016; 19: 433–443. doi: 10.4103/0972
Note: All the figures in this chapter were designed by the author.
2327.194409.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
C. Syamsundar Reddy1
Research Schalor, Department of Computer Science, College of Commerce,
Management & Computer Science, Sri Venkateswara University, Tirupathi, Andhra Pradesh
G. Anjan Babu2
Professor, Department of Computer Science, SVU College of Commerce,
Management & Computer Science, Sri Venkateswara University, Tirupathi, Andhra Pradesh
Abstract: Smart cameras in cities monitor potential suspects, enabling decision analysis and the machine learning of visual
data. Federated Learning (FL) is a new approach that removes data borders and protects privacy. Suspicions can be predicted
using emotional image data or suspicious face labeling. Emotional categories, such as facial emotions, reveal negative, neutral,
and pleasant feelings using Neutrosophic logic. Facial recognition measures emotional faces, and Neutrosophic logic regression
analyses crime. Emotional knowledge can aid in more precise crime prediction. The experimental results show that the two
attributes are not independent, indicating they are dependent, and the metrics for CNN at optimal iterations (5) have been
thoroughly examined.
Keywords: Crime face emotions, Convolution neural networks, Federated learning, Neutrosophic logic, Logistic regression
1
cssreddi@gmail.com, 2gabsvu@gmail.com
DOI: 10.1201/9781003529231-23
146 Algorithms in Advanced Artificial Intelligence
ReLU contributes to the model’s increased nonlinearity • 0 ≤ t + i + f < 3 when all three components are
because images naturally exhibit some nonlinearity. independent.
4. Pooling Layers: By pooling, the dimensionality of • The scenario is 0 ≤ t + i + f < 2 when two components
each feature map is reduced while the most important are interdependent and the third is independent of the
information is retained. An approach known as “max other two.
pooling” extracts the maximum value from the area of • 0 ≤ t + i + f ≤ 1 when all three factors are interrelated.
the image that the filter has filtered.
There is a possibility of incomplete information (sum < 1),
5. Fully Connected Layers: After a number of para-consistent and conflicting information (sum > 1), or
convolution and pooling layers, fully linked layers in complete information (sum = 1) when three or two of the
the neural network are used for high-level reasoning. components T, I, and F are independent. In a similar vein, in
As in conventional neural networks, neurons in a fully the event where T, F, and I are all reliant on one another, there
linked layer are connected to all activations in the is a chance for either whole knowledge (sum = 1) or partial
preceding layer. knowledge (sum < 1).
6. Output Layer: For example, if there are 10 classes
for 10 different sorts of objects, the result will be a
10-element vector in the final layer’s application of the
5. Federated Learning Supports
softmax function to output a probability for each class Cloud-Recognition Based Facial
of the problem. Technology
CNN’s final forecast for Tanoy, 2022 [15], indicates the class
Federated learning improves computers by using Neutrosophic
with the highest probability. The multilayered structure of
data to apply an iterative recognition model by Zhang in 2020
CNNs allows them to learn hierarchical features, which helps
[16], offering an answer to online privacy issues. It works well
them perform exceptionally well in image classification tasks.
with emotive faces due to on-device data and privacy. The
development of cloud-based facial recognition systems has
increased its potential [17], offering fast flexibility, resource
sharing, on-demand self-service and comprehensive network
connectivity. This paradigm is often used for security testing,
where a user takes a picture of a query face.
communication, and can accommodate a large user base, Table 23.1 Police records shows the type of crime in four
making them more accessible and adaptable. regions of a west Godavari district
District wise Physical Murder Rape Homicide Total
6. Neutrosophic Logic Regression Regions Assault
M. Srikanth1
Research Scholar, Department of Computer Science and Engineering,
GIET University, Odisha
R. N. V. Jagan Mohan2
Associate Professor, Department of Computer Science and Engineering,
GIET University, Odisha
M. Chandra Naik3
Professor, Research scholar, Department of Computer Science and Engineering,
GIET University, Odisha
Abstract: Machine learning has the potential to revolutionize agriculture by helping farmers optimize crop yields, reduce
costs, and improve sustainability. One way to use machine learning in agriculture is to optimize multi-cropping, which involves
growing multiple crops simultaneously on the same piece of land. This paper proposes a new approach to optimizing multi-
cropping using reinforcement learning. Reinforcement learning is a type of machine learning that allows agents to learn to
behave in an environment by trial and error. In the context of multi-cropping, the agent is a machine learning model that is
trying to learn to select the best crops to grow together and how to manage them in order to maximize yield. The proposed
approach uses reinforcement learning to optimize the hyper parameters of the machine learning model. Hyper parameters are
the settings of the machine learning model, such as the number of trees in a random forest model or the learning rate of a neural
network. By optimizing the hyper parameters, the machine learning model can be trained to better predict crop yields and make
better decisions about crop management. The proposed approach was evaluated on a real-world dataset of crop yields from
India. The results showed that the proposed approach was able to significantly improve crop yields compared to traditional
methods of multi-cropping.
Keywords: Machine learning, Agriculture, Multi-cropping, Reinforcement learning, Hyper parameter optimization, Crop yield
DOI: 10.1201/9781003529231-24
Increasing the Reliability of Intercropping in Agriculture Using Machine Learning 151
Row crops are crops planted in wide rows, suitable for where, S = Total number of samples, P(yes) = probability of
tilling or cultivation using agricultural machinery, and sown yes, P(no) = probability of no
through drilling or transplanting instead of broadcasting. The Gini index measures impurity in CART algorithm,
Temporal intercropping is a method where a fast- growing favoring low-ranked attributes for binary splits. It’s calculated
crop is sown alongside a slow-growing crop, ensuring that using a formula, ensuring binary splits in decision trees.
the fast-growing crop is harvested before the slow-growing
crop matures. Relay cropping involves sowing a second crop Gini Index = 1 – ∑jPj2 (3)
during the first crop’s growth, allowing the first crop to be
harvested. Intercropping, on the other hand, grows different 2. Multi-Crops Hyperparameters
crops in a sequence of seasons. Multicropping is also known Optimize with Reinforcement
as intercropping, is the practice of growing multiple crops
simultaneously on the same land, doubling crop productivity Learning
and income, and the choice depends on mutual benefits. Reinforcement Learning is utilized to optimize multi-crops
Multiple cropping, a technique involving growing multiple Hyperparameters. Reinforcement Learning (RL) is a method
crops simultaneously, enhances crop yield through the use of for optimizing multicrops in agriculture using machine
advanced technologies like seeds, fertilizers, and pesticides. learning processes, based on a Markov chain of improved
Intercropping Using Decision Tree Classification: Decision model performance. Supervised learning is a significant
Tree is a supervised learning technique used for classification challenge in predicting multi-crops yield based on profit/
and regression problems, with its tree-structured classifier loss harvesting to reach the next state. The model is trained
consisting of internal nodes representing dataset features, to predict the probability of the next crop case using the
branches representing decision rules, and leaf nodes most recent Hyperparameters values and their predictive
representing outcomes. Decision nodes make decisions, performance. The model uses a harvest of multicrops loss
while leaf nodes represent outcomes proposed by R. N. V. function to train on a discretized Hyperparameters space,
Jagan Mohan, 2012 []. A decision tree is a flowchart structure aiming to optimize future crop rewards by tracking past
where each node represents a test on an attribute, each branch agricultural states H:r = M(H).
represents the outcome, and each leaf node represents a class A Reinforcement Learning model R predicts a value q using
label. The text provides an overview of various types of H and r, with q = R(H, r). The optimal action maximizes
multiple crops in agriculture. q, and can be predicted for past H and r using the formula
q’ = R(past H, past r), where r and q represent future values.
The model minimizes mean square error by calculating
q’ – (r + g * max q) ^ 2. The policy gradient is preferred
for multicrops classification due to its ease of management
and compatibility with high Hyperparameters space
dimensionality. To indicate a model’s preference for certain
Fig. 24.1 Types of multiple crops in agriculture Hyperparameters to be 1: L = –1 * logP(next H|current H,
current r), use cross entropy to increase the probability of
Attribute Selection Measures (ASM) are techniques used generating them. Policy gradient weighs the sample with the
in Decision Tree implementation to select the best attribute reward value L = –(next reward) * log P(next H|current H,
for root and sub-nodes, with popular techniques being current r) With next reward=M(next H). The multicrops
Information Gain and Gini Index. model is designed to optimize for Hyperparameters with high
Information gain measures entropy changes after dataset profit rewards and minimal impact on loss reward, each with
segmentation based on attribute. It calculates feature its own multicrops classification option.
information about a class. Decision trees maximize
information gain, splitting nodes based on highest gain. 3. Multi-Crops Optimizing
Information Gain = Entropy(S) – [(Weighted Avg) * Entropy Hyperparameters
(each feature)] (1)
The section discusses the significant role of grid search in
Entropy is a metric that measures the impurity of an attribute
model tuning.
and indicates randomness in data, and can be calculated using
various methods. The Problem: Finding the model Multi-Crops
Hyperparameters for Agriculture machine learning model
Entropy(s) = –P(yes)log2 P(yes) – P(no) log2 P(no) (2)
152 Algorithms in Advanced Artificial Intelligence
can be challenging due to manual tuning and potential missed to as “model selection.” If there are features X and a target
configurations. Y, the best transformation F can be determined from the data
Where it Shows: Suboptimal Multi-Crops Hyperparameters by: Y = F(X)
can cause underfitting or overfitting in a model, impacting its The word “optimal” denotes the existence of a model
ability to accurately predict new crop data. performance metric, and the “optimal” model is the one that
The Effect: Poor Multi-crop Hyperparameter choices can maximizes that statistic. It is important to consider a number
lead to ineffective models, resulting in missed insights and of axes in order to improve our model:
inaccurate predictions. 1. The model parameter space: Use statistical learning
The Solution: Grid Search is an automated method to “train” a model and “optimize” this “space”. The
that systematically explores various Multi-Crops parameters are learned using an optimization approach,
Hyperparameter combinations to find the best group for such as the maximum likelihood estimation principle.
MulticropsHyperparameter optimization. 2. The Model Paradigm Space: It is possible to employ
a variety of supervised learning algorithms to address
Practical Steps:
the same issue. Depending on the particular dataset,
1. Define Multi-Crops Hyperparameter Grid: To algorithms like Naive Bayes, XGBoost, or Neural
optimize Multi-Crops Hyperparameters, specify Network may perform significantly differently.
desired values and range of exploration, like in Random 3. The Hyperparameters Space: Make these choices in
Forest model, like tree number and maximum depth. order to put up our training run, even though statistical
2. Setup Cross-Validation: K-fold cross-validation is learning cannot enhance these model parameters.
a robust technique that divides multi- crops data into 4. The Model Architecture Space:This applies more
subsets for training and testing, preventing overfitting to neural networks. A set of Hyperparameters can
during Multi-Crops Hyperparameter tuning. be used to describe the model architecture, but the
3. Run the Grid Search: Grid Search is a method that search is typically more involved than with ordinary
thoroughly tests all possible combinations of Multi- Hyperparameters. The size of the search space can
Crops Hyperparameters, training and evaluating the reach 10^40.
model for each of them. 5. The feature space: The proper feature must be chosen
4. Select the Best Configuration: Grid Search provides to feed our model. Depending on the features that can
the most suitable Multi-crop Hyperparameter set based be used, different models will respond in different
on the chosen evaluation metric, such as accuracy, F1 ways. Excessive characteristics and possible overfit. It
score, or mean squared error. might not fit if there aren’t enough features.
5. Fine-Tune Further: The Multi-Crop Hyperparameters
are optimized and further refined using techniques
like Random Search or Bayesian Optimization for
enhanced performance. Grid Search simplifies Multi-
Crop Hyperparameter tuning, allowing for more
efficient model building and multi-crops data insights
extraction.
Optimizing a model: The procedure of selecting the best Fig. 24.3 Optimizing machine learning model: The different
model for a specific training multi-crops dataset is referred axes
Increasing the Reliability of Intercropping in Agriculture Using Machine Learning 153
6. The Feature Transformation Space: To enhance the perform cross-validation, and create visualizations. Once a
performance of our model, take into account several reliable machine learning model is developed, deploy it as
transformations as the Box-Cox transformation and part of an agricultural decision support system to provide
feature encoding. recommendations. Continuously collect and update the
models to improve reliability over time. Integrate the system
4. Experimental Result into daily operations, provide training, and implement a
monitoring system for tracking results and user feedback.
To effectively predict crop yields in intercropping in The integration of machine learning and data-driven
agriculture, gather relevant data on crop types, soil types, decision-making in agriculture can lead to a more reliable,
climate conditions, temperature, rainfall, and yields. Prepare sustainable, and efficient system, benefiting both farmers and
the collected data for machine learning by cleaning, handling the environment.
missing values, and encoding categorical variables. Identify
the most relevant features and choose appropriate machine
learning models. Split the dataset into training and testing 5. Conclusion
sets, train the models, and evaluate their performance. This study has shown that reinforcement learning can be
Optimize the models by adjusting Hyperparameters, used to effectively optimize multi-crops Hyperparameters in
154 Algorithms in Advanced Artificial Intelligence
agriculture. The proposed approach was able to significantly number of factors. The proposed approach is also scalable
improve crop yields compared to traditional methods of multi- and can be applied to a wide range of crops and growing
cropping. The significance of this research is that it provides conditions. This means that it has the potential to make a
a new and promising approach to using machine learning to significant impact on the global food system. In the future,
improve the efficiency and productivity of agriculture. The we plan to further evaluate the proposed approach on a larger
proposed approach is particularly well-suited for multi- scale and to develop new reinforcement learning algorithms
cropping, which is a complex problem that involves a large that are specifically designed for multi- crops optimization.
Increasing the Reliability of Intercropping in Agriculture Using Machine Learning 155
We also plan to work with farmers to implement the proposed 15. Kumar, J. M. S. V., et al. “Reverse Engineering A Generic
approach in real-world settings. Software Exploration Environment Is Made Of Object
Oriented Frame Work And Set Of Customizable Tools.”
International Journal of Advanced Research in Computer
References Science 2.5 (2011).
16. Kumar, J. M. S. V., et al. “Analyzing the Modern Tool-
1. Abhishek, Aditya: Multiple Cropping- Definition, Benefits
Supported UML-Based Static Reverse Engineering.”
and Selection of Crops, Agriculture Review, Retrieved 2020
International Journal of Advanced Research in Computer
12-14, 2020.
Science 3.4 (2012).
2. Harrell, Stevan: An Ecological History of Modern China,
17. Kumar, J. M. S. V., et al. “Active Scrutiny Techniques for the
Seattle: University Of Washington Press, ISBN 978-0-295
Reconstruction of Architectural Views.” International Journal
75171-9, 2023.
of Advanced Research in Computer Science 3.1 (2012).
3. R. N. V. Jagan Mohan and R. Subbarao and Raja Sekhara
18. N Santha Raju, JMSV Kumar, B Sujatha,”Time series analysis
Rao K.,: Efficient K-Means Cluster Reliability on Ternary
of stock price movements: Insights from data mining using
Face Recognition using Angle Oriented Approach, Published
machine learning”, journal AIP Conference Proceedings,
In the Proceedings of International Conference on Advances
Volume 2492, Issue1, Publisher AIP Publishing,2023.
in Communication, Navigation and Signal Processing
19. Prayaga Atchyut Pavan, Sattibabu Sattibabu, JMSV Kumar
Technically Co-Sponsored by IEEE, Hyderabad Section,
“A deep learning approach to detect malaria “Journal AIP
March 17th-18th, 2012, Dept of ECE, Andhra University
Conference Proceedings, Volume 2492, Issue 1, Publisher AIP
College of Engineering (A).
Publishing, 2023.
4. X. Chen et al., “Optimizing Crop Planning with Reinforcement
20. Ch Bhanu Revathi, JMSV Kumar, B Sujatha” Intracranial
Learning,” in International Conference on Machine Learning
hemorrhage detection in human brain using deep learning “
(ICML), 2017.
Journal AIP Conference Proceedings, Volume 2492, Issue 1,
5. R. Singh and B. R. Borah, “Agricultural Data Analysis Using
Publisher AIP Publishing, 2023.
Machine Learning: A Review,” in 2017 2nd International
21. JMSV RAVI KUMAR” Human Activity Recognition using
Conference on Computational Systems and Information
Machine Learning “ Journal AIP Conference Proceedings,
Technology for Sustainable Solutions (CSITSS), 2017.
Volume 2492, Issue 1, Publisher AIP Publishing, 2023.
6. S. Wang et al., “Agricultural Production Prediction Using
22. J Kumar, A Shahi, R Aytha, G Varri, D Brundavanam “ Vehicle
Machine Learning Algorithms,” in 2019 IEEE International
theft prevention system using IoT “Journal AIP Conference
Conference on Big Data (Big Data), 2019.
Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
7. A. Sinha and N. Kumar, “Machine Learning Applications in
2023.
Agriculture: An Overview,” in Journal of Crop Science and
23. J Kumar, TD Nagendra, M Harshitha, AB Prakash “ Fake
Biotechnology, 2019.
image detection using CNN “Journal AIP Conference
8. R. Salim et al., “Application of Machine Learning Techniques
Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
in Agriculture: A Review,” in Information Processing in
2023.
Agriculture, 2020.
24. J Kumar, MN Kumar, NV Narendra, P Pradeep “ driver
9. H. Nguyen and S. Shekhar, “Reinforcement Learning
drowsiness monitoring system using machine learning svm
for Precision Agriculture: A Review,” in Computers and
algorithm “Journal AIP Conference Proceedings, Volume
Electronics in Agriculture, 2019.
2492, Issue 1, Publisher AIP Publishing, 2023.
10. B. Basso et al., “Crop Yield Prediction: A Machine Learning
25. JMSV Ravi Kumar “ A Symmetric Searchable Encryption
Approach,” in Computers and Electronics in Agriculture,
Identification of Data on Probabilistic Trapdoors “International
2017.
Journal of Engineering and Advanced Technology (IJEAT),
11. A. Singh et al., “Machine Learning Techniques for Agriculture
ISSN: 2249 – 8958, Volume 9, Issue 3, Publisher Blue Eyes
Crop Yield Prediction,” in 2018 9th International Conference
Intelligence Engineering & Sciences Publication, 2020.
on Computing, Communication and Networking Technologies
26. JMSV Ravi Kumar “Artificial Bee Colony Algorithm: A
(ICCCNT), 2018.
Survey and Recent Applications” published in International
12. S. Patel et al., “Machine Learning-Based Techniques for
Journal of Pure and Applied Mathematics, ISSN 1314-3395,
Agriculture and Crop Yield Prediction: A Review,” in 2019
VOLUME 118, ISSUE 24 , Jul-18.
4th International Conference on Internet of Things: Smart
27. JMSV Ravi Kumar “Authentication for Cloud Services
Innovation and Usages (IoT-SIU), 2019.
using Steganography” published in International Journal of
13. A. Singh et al., “Intelligent Agriculture: A Survey on
Engineering and Technology(UAE)-IJET, ISSN 2227-524X,
the Impact of Artificial Intelligence on Agriculture,” in
Volume 7, Issue 3.49 , Jul-18.
2018 International Conference on Computing, Power and
28. JMSV Ravi Kumar “A review on task scheduling algorithms
Communication Technologies (GUCON), 2018.
in cloud computing and their approaches” published in
14. Kumar, J. M. S. V., et al. “System Testability Assessment and
International Journal of Pure and Applied Mathematics, ISSN
testing with Micro architectures.” International Journal of
1314-3395, Volume 118, Issue 24, Jul-18.
Advanced Research in Computer Science 2.6 (2011).
156 Algorithms in Advanced Artificial Intelligence
29. JMSV Ravi Kumar “Review of Data mining Technique using 42. M. Srikanth, “Tackling Outliers for Predictive Smallholder
SaaS on the Cloud” published in International Journal of Pure Farming Analysis,” in Proceedings of the 2023 3rd International
and Applied Mathematics, ISSN 1314-3395, VOLUME 118, Conference on Smart Data Intelligence (ICSMDI), pp. 93-98,
ISSUE 24 , Jul-18. IEEE Xplore, March 26, 2023. [Scopus]
30. JMSV Ravi Kumar “Smart Controlling, Monitoring and 43. M. Srikanth, “Blockchain-Based Consensus For A Secure
Automation of Street Light System using Raspberry PI Smart Agriculture Supply Chain,” European Chemical
“ published in International Journal of Pure and Applied Bulletin, vol. 12, special issue 4, pp. 8669-8678, 2023.
Mathematics, ISSN 1314-3395, VOLUME 118, ISSUE 24 , [Online]. Available: doi: 10.48047/ecb/2023.12.si4.776.ISSN:
Jul-18. 2063-5346, 2023. [Scopus]
31. JMSV Ravi Kumar “A Survey on Internet of Things for 44. M. Srikanth, “Predict Early Pneumonitis in Health Care Using
Healthcare and Medication Management” was authored by Hybrid Model Algorithms,” Journal of Artificial Intelligence,
JMSV Ravi Kumar published in International Journal of Pure Machine Learning and Neural Network (JAIMLNN), vol. 3,
and Applied Mathematics, ISSN 1314-3395, VOLUME 118, issue 03, pp. 14-26,ISSN: 2799-1172, Apr. 2023.
ISSUE 24 , Jul-18. 45. M. Srikanth, R. N. V. Jagan Mohan, M. Chandra Naik. (2023).
32. JMSV Ravi Kumar “SECRBAC: Secure Data in the Clouds” A New Way to Improve Crop Quality and Protect the Supply
was authored by JMSV Ravi Kumar published in International Chain is to use a Trajectory Network and Game Theory.
Journal of Research, ISSN 2348-6848, VOL 5, ISSUE 15 , Mathematical Statistician and Engineering Applications,
Jul-18. 71(4), 10600–10610. https://doi.org/10.17762/msea.
33. JMSV Ravi Kumar “EBPH MAC: Emergency Based v71i4.1952, ISSN: 2094-0343, 2023 [Scopus]
Priority Hybrid Medium Access Control for Mobility 46. M. Srikanth, “Auction Algorithm: Peer-To-Peer System Based
Aware Cooperative WSN’s In Indoor Industrial Monitoring” on Hybrid Technologies for Smallholder Farmers to Control
published in International Journal of Research, ISSN 2348 Demand and Supply,” International Journal of Research In
6848, VOLUME 5, ISSUE 12 , Jul-18. Science & Engineering (IJRISE), vol. 3, issue 1, pp. 9–23,
34. JMSV Ravi Kumar “Prioritizing software components for 2023.
realistic reuse” published in International Journal of Sciences 47. M. Srikanth, “Smallholder Farmers Crop Registering Privacy-
& Applied Research, ISSN 2394-2401, VOL 4, ISSUE 24, Preserving Query Processing over Ethereum Blockchain,”
Jul-17. Journal of Pharmaceutical Negative Results, vol. 13, issue 7,
35. JMSV Ravi Kumar “Cloud Storage Services and Privacy pp. 5609-5617, Dec. 2022. [Scopus]
Protection” published in International Conference on Research 48. M. Srikanth, “The Early Detection of Alzheimer’s Illness
Advancements in Computer Science and Communication, Using Machine Learning and Deep Learning Algorithms,”
ISSN 978-93-85100- 64-2, VOL 5, ISSUE 3.49, December-16. Journal of Pharmaceutical Negative Results, vol. 13, issue 9,
36. JMSV Ravi Kumar “Analyzing the Modern Tool-Supported pp. 4852-4859, Nov. 2022. [Scopus]
UML-Based Static Reverse Engineering” published in 49. M. Srikanth, “Small Holders Farming Predictive Analysis
International Journal of Advanced Scientific Research and Using Peer-To-Peer Approach,” International Journal of
Technology, ISSN 0976-5697, VOL 3, ISSUE 4, Jul-12. Agriculture and Animal Production, vol. 2, issue 05, pp. 26
37. JMSV Ravi Kumar “Active Scrutiny Techniques for the 37, Sep. 2022.
Reconstruction of Architectural Views” published in 50. M. Srikanth, “Using Machine Learning and Neural Networks
International Journal of Advanced Scientific Research and Technologies, a Bottom-Up Water Process Is Being Used To
Technology, ISSN 0976-5697, VOL 3, ISSUE 1, January-12. Reduce All Water Pollution Diseases,” Journal of Artificial
38. JMSV Ravi Kumar “System Testability Assessment and Intelligence, Machine Learning and Neural Network
testing with Micro architectures” published in International (JAIMLNN), vol. 2, Oct. 2022.
Journal of Advanced Scientific Research and Technology, 51. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers
ISSN 0976-5697, VOL 2, ISSUE 6, December-11. Crop Transaction Using Peer-to-Peer,” Indo-American Journal
39. JMSV Ravi Kumar “Reverse Engineering A Generic Software of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp.
Exploration Environment is made of Object-Oriented 33-43, Sep. 2022.
Frame Work and Set of Customizable Tools” published in 52. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care
International Journal of Advanced Scientific Research and Centres Use a Hybrid Technique Based on a Distribution
Technology, ISSN 0976-5697, Vol 2, Issue 5, September-2011. Network,” International Journal of Health Sciences, Jun.
40. M. Srikanth, “Integrated Technologies for Proactive Bridge- 2022. [Scopus]
Related Suicide Prevention”, Journal of Namibian Studies, 53. M. Srikanth, “Blockchain-Based Crop Farming Application
Volume 1, Issue 33, Pages 2117-2136, ISSN: 1863-5954, Sep Using Peer-to-Peer,” Journal of Xidian University, Apr. 2022.
2023. [Scopus] 54. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
41. M. Srikanth, “Deep Learning Approaches for Predictive Emotional Recognition Using Machine Learning, Query
Modeling and Optimization of Metabolic Fluxes in Engineered Optimization and Blockchain Technology,” Solid State
Microorganism” International Journal of Research in Science Technology, Vol. 63 No. 6 (2020) [Scopus]
&Amp; Engineering (IJRISE) ISSN: 2394-8299, 3(05), 1–11. 55. M. Srikanth, “Machine Learning for Query Processing System
https://doi.org/10.55529/ijrise.35.1.11, July 2023. and Query Response Time Using Hadoop,” IJMTST, Aug.
2020.
Increasing the Reliability of Intercropping in Agriculture Using Machine Learning 157
56. M. Srikanth, “Block-level Based Query Data Access Service 58. M. Srikanth, “A New Approach for Authorship Verification
Availability for Query Process System,” IEEE, Page 1-9, Jul. Using Information Retrieval Features,” Springer-ICSE, vol.
2020. [Scopus] 74, pp. 23-29. [Scopus]
57. M. Srikanth, “Query Response Time in Blockchain Using 59. M. Srikanth, “An Enhanced and Naive Clustering Algorithm
Big Query Optimization,” The Role of IoT and Blockchain for Text Classification Based on Weight,” International Journal
Techniques and Applications from Computer Science and & Magazine of Engineering, Technology, Management and
Information Management, Apple Academic Press, Exclusive Research, Dec. 2012.
Worldwide distribution by CRC Press Taylor & Francis
Group, Jan. 2022. [Scopus]
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
158 Algorithms in Advanced Artificial Intelligence
Ravisankar Malladi1
Dept. of Computer Science and Engineering,
Koneru Lakshmaiah Education, Vaddeswaram, Guntur
V. T. Ram Pavankumar2
Dept. of Master of Computer Applications,
K. B. N. College, Vijayawada
M. Arulselvi3
Dept. of Computer Science and Engineering,
Annamalai University, Tamilnadu
Konatham Sumalatha4
Dept. of DBS, Vellore Institute of Technology,
Vellore, Tamilnadu
Abstract: False information that is reported as news and frequently intended to harm reputations or make money is called fake
news. It has been used to refer to all types of misleading information since it was first introduced in the 1890s and is frequently
produced by adversarial foreign actors. Due to the variety of fake news kinds, experts increasingly prefer the term “information
disorder” as a neutral and educational word. The study addresses information overload and filtering issues on social media
by introducing a new approach to assessing news reliability. It uses Cosine Similarity based Retriever Augmented Generation
(RAG) technique and classification deep learning algorithm to classify news as fake or real, with the best-performing feature
predicting its authenticity. The system achieved 91% accuracy in testing.
Keywords: Cosine similarity, Deep learning, Fake news, Genuine news, Retriever augmented generation etc.
1
mravisankar@kluniversity.in, 2mrpphd2018@gmail.com, 3marulcse.au@gmail.com, 4konatham.sumalatha@vit.ac.in
DOI: 10.1201/9781003529231-25
Retrieval Augmented Generation Classification Algorithm for Fake News Detection 159
the performance of our proposed method, focusing on the time less complicated. This delivers the highest accuracy in
literature survey, which is crucial in software development. comparison to the present system.
Yafra Khan and Chai Soo See’s article “Predicting and
assessing bogus news” highlights social media as a powerful
medium for self-expression, allowing discussions on identity,
society, religion, and customs. Social media significantly
influences daily lives and society, allowing individuals to
share newsworthy information and stay informed about
global events [4-8]. K.Corradini et al.’s paper combines
machine learning and knowledge engineering to detect bogus
news on social networks. Fake news is a growing concern
due to its rapid dissemination on social and news media,
making it crucial to detect and prevent the spread of false
information [9]. The article by Conroy, Rubin, and Chen
proposes automatic deception detection methods for fake Fig. 25.1 Represent the system architecture
news, emphasizing the importance of modern technologies
Figure 25.1 outlines several necessary steps to achieve the
in identifying and classifying news based on veracity and
current scope. Here, we attempt to take an unstructured news
certainty. Yafra Khan and Chai Soo See [10] proposed a
dataset as input and then extract text data from it. As soon as
method for predicting and analyzing fake news on social
the text data is gathered, text pre-processing is attempted on
media using a religion and politics dataset and ML algorithms,
the resulting file. After extracting the key features, we now
aiming to prevent false information from spreading.
attempt to apply classifiers to the data. The effectiveness of
Machine Learning classification algorithms in determining
2. Related Work news authenticity can be determined by applying various
Current systems struggle to identify fake news on social methods. For example, if a fake news question requires
media, leading to rapid propagation of false information and scanning all documents, similarity searches may not be
unreliable papers [11]. This has resulted in many honest users helpful. Misinformation, also known as fake news or hoaxes,
making incorrect conclusions due to the limitations of the is the accidental sharing of false information without intent
current system. The existing system has a limit has follows to cause harm, while disinformation is deliberately shared to
the current system lacks a classification algorithm capable mislead and cause harm. For instance, Russia’s invasion of
of automatically detecting fake news items in published Ukraine in February 2022 led to a massive disinformation
content. The current system is time-consuming. The process campaign, with News Guard identifying 311 websites
of distinguishing between bogus and legitimate news requires publishing pro-Russian propaganda to justify Moscow’s
more time. False information is a significant waste of storage aggression.
space.
3.1 Fake News Detection Using Cosine Similarity
The cosine similarity index calculates how similar two
3. Proposed Methodology vectors in an inner product space are to one another. It
Re-using passwords on multiple websites causes serious establishes whether two vectors are roughly pointing in the
security problems. If a hacker can connect into one account same direction by calculating the cosine of the angle between
using login credentials, they can access all accounts that them. In text analysis, it is frequently used to gauge document
utilize that password [12]. But not only people are subject to similarity [13].
the threat. When employees use the same password at work For instance, the cosine similarity between two proportional
and at home, the security of the entire company is at risk. We vectors is 1, that between two orthogonal vectors is 0, and
avoided using the same passwords for two or more accounts that between two opposite vectors is -1. In some situations,
on the same or other websites in an effort to lessen shadow the vectors’ component values cannot be negative, in which
attacks, which rely on password reuse [10]. The benefits of case the cosine similarity is constrained to the range [0, 1].
the proposed system include the suggested classification
technique can be made less temporally difficult. The RAG 3.2 Fake News Identification Using Retriever
technique similarities between legitimate news and false Augmented Generation (RAG)
news are straightforward to classify. Any dataset gathered
Retrieval Augmented Generation (RAG), a technology that is
from the real world can be used for this. This will make
constantly advancing in the field of artificial intelligence, is
160 Algorithms in Advanced Artificial Intelligence
creating waves. With this novel method, factual data retrieval if it’s broken down into large chunks, potentially containing
is combined with the strength of huge language models. We unrelated information. To avoid diluted information and
will go into the nuances of RAG [14]. irrelevant documents, break down the data into a few
The use of Fake news database in augmenting LLMs is paragraphs per chunk, ensuring uniqueness. The RAG
a valuable method, but it has several significant flaws. approach emphasizes limiting the type of questions asked by
The debate between fine-tuning and Retriever Augmented the LLM. Aggregating data across the database may lead to
Generation (RAG) with LLMs is ongoing, with RAG being incorrect answers, while similarity searches may find local
better for enhancing LLMs with small additional data. RAG information.
encodes data into embeddings and indexes it into a vector
fake news database. Users ask fake news questions, which 4. Experimental Result and Discussion
are converted into embeddings and used to search for similar
This work utilizes Python as the programming language and
embeddings. Prompts provide context for LLM answers,
Google Colab as the working platform for developing and
usually using cosine similarity metric. The problem lies in
executing the Fake News application. The following two
the search’s ability to retrieve documents with similar words
or context without providing relevant information, leading steps are as follows
to an excess of irrelevant documents showing higher cosine
4.1 Load the Dataset and Categorize
similarity than the actual answer. High cosine similarity in
Transformers does not necessarily imply semantic similarity; To detect fake news in Python, preprocess input text; obtain
it can also indicate the high co-occurrence of two terms within numerical features, and train machine learning models like
the same training data. The data’s indexing can cause issues RAG to predict news reliability.
4.2 Apply RAG Classification Algorithms more advanced and trustworthy techniques to protect the
accuracy of information distribution. Even though there are
Machine learning is an AI subset that uses algorithms to
still difficulties, the development and improvement of such
analyze vast data, analyzing it holistically and recognizing
sophisticated algorithms represent a critical advancement in
important information boundaries. The dataset is loaded,
strengthening our defenses against the damaging impacts of
revealing the number of unique categories and the application
fake news, eventually supporting the basis of a society that is
of the RAG classification algorithm. Cosine Similarity
better informed and resilient.
outperforms other machine learning algorithms in ML
classification, making it the optimal choice for identifying
fake news from social media. Measurement is essential References
for comprehending the outside world, but it also brings
1. A. Douglas: News consumption and the new electronic media,
inaccuracy, or uncertainty. When collecting measurements,
The International Journal of Press /Politics, vol. 11, no. 1, pp.
accuracy is an important feature to take into account because 29–52, 2006.
it indicates how closely a measurement resembles a known 2. J. Wong: Almost all the traffic to fake news sites is from
or accepted value. An indicator of the accuracy of a series Facebook, new data show, 2016.
of observations of their closeness, accuracy is a measure of 3. M. J. Lazer, M. A. Baum, Y. Benkler et al: The science of fake
observational error. news, Science, vol. 359, no. 6380, pp. 1094–1096, 2018.
4. S. A. García, G. G. García, M. S. Prieto, A. J. M. Guerrero, and
Accuracy: Accuracy is a statistical measure of how well a
C. R. Jimenez: The impact oterm fake news on the scientific
binary classification test identifies or excludes a condition,
community scientific performance and mapping in web of
often referred to as the “Rand accuracy” or “Rand index.” science, Social Sciences, vol. 9, no. 5, 2020.
It compares pre- and post-test probability estimates and is a 5. Holan, 2016 Lie of the Year: Fake News, Politifact,
test parameter. Based on 91 accurate predictions of fake news Washington, DC, USA, 2016.
out of 100 cases, the RAG algorithm classified 100 pieces 6. Robb: Anatomy of a fake news scandal,” Rolling Stone, vol.
of news with a 91% accuracy rate. However, only one out of 1301, pp. 28–33, 2017.
the nine is correctly identified by the model, leaving 8 out of 7. J. Soll: The long and brutal history of fake news,” Politico
9 instances of bogus news undetected. This implies that the Magazine, vol. 18, no. 12, 2016.
model is less efficient than one that consistently makes good 8. J. Hua, R. Shaw: Corona virus (covid-19) “infodemic” and
predictions. When working with a class-imbalanced data set emerging issues through a data lens: the case of China,
International Journal of Environmental Research and Public
where the discrepancy between positive and negative labels is
Health, vol. 17, no. 7, p. 2309, 2020.
substantial, accuracy alone is insufficient.
9. N. K. Conroy, V. L. Rubin, and Y. Chen: Automatic deception
TP +TN detection: methods for finding fake news, Proceedings of the
Accuracy = (1) Association for Information Science and Technology, vol. 52,
TP +TN + FP + FN
no. 1, pp. 1–4, 2015.
Where TP = True positive; FP = False positive; TN = True 10. F. T. Asr and M. Taboada,: Misinfotext: a collection of news
negative; FN = False negative articles, with false and true labels,” 2019.
1 + 90 11. Shu, A. Sliva, S. Wang, J. Tang, and H. Liu: Fake news
Accuracy = = 0.91 detection on social media, ACM SIGKDD Explorations
1 + 90 + 1 + 8 Newsletter, vol. 19, no. 1, pp. 22–36, 2017.
12. S.Vosoughi, D. Roy, and S. Aral: The spread of true and false
news online, Science, vol. 359, no. 6380, pp. 1146–1151,
5. Conclusion 2018.
An approach to detect misleading information The RAG 13. Schubert, Erich; Lang, Andreas; Feher, Gloria, Reyes, Nora;
technique divides user input into true and false categories Connor, Richard; Kriege, Nils; Kazempour, Daniyal; Bartolini,
as a defense against false information. Combining the Ilaria; Schubert, Erich; Chen, Jian-Jia: Accelerating Spherical
k-Means., Similarity Search and Applications. Lecture
advantages of classification algorithms for discernment,
Notes in Computer Science. Cham: Springer International
generation models for content synthesis, and retriever models
Publishing, 13058: 217–231, ArXiv: 2107.04074, doi:
for information retrieval, this novel approach provides a 10.1007/978-3-030-89657-7_17, ISBN 978-3-030-89657-7,
solid foundation for identifying and halting the spread of S2CID 235790358.
false information. Promising results from extensive testing 14. Soumyadarshan Dash: Retrieval Augmented Generation
and review suggest that it can effectively identify deceptive (RAG), Published On September 28, 2023 and Last Modified
material. This algorithm offers a ray of hope in the face of On September 29th, 2023.
the ongoing threat of disinformation, opening the door for
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
162 Algorithms in Advanced Artificial Intelligence
K. V. Nageswari1
Research Scholar, GIET University, Gunupur, Odisha State
R. N. V. Jagan Mohan2
Associate Professor, SRKR Engineering College, Bhimavaram
Bhramara Bar Biswal3
Associate Professor, GIET University, Gunupur, Odisha State
Abstract: Urologic cancers, particularly renal pelvis cancer, affect the urinary system and male reproductive system.
Clinical trials are underway to develop personalized, evidence-based treatments for advanced urologic cancer treatments.
The personalized urologic cancer treatment, including chemotherapy, radiation therapy, surgery, and supportive care. Patients
collaborate with doctors to select the optimal plan, which includes behavioral medicine, nutrition, pain management, and social
support. AI integration in urologic cancer detection enhances diagnosis, early detection, and personalized treatment, potentially
improving patient survival rates and overall well-being. The study explores the use of predictive AI treatment for kidney tumors,
ensuring privacy protection. The study aims to select a guideline-directed therapy for advanced Kidney cancer using patient
characteristics and clinical data, and to implement a collaborative approach using Machine Learning.
Keywords: Artificial intelligence, Clinical trail, Reinforcement learning, Renal (Kidney) urologic cancer etc.
1
ranikondaveti2011@gmail.com, 2mohan.rnvj@srkrec.edu.in, 3bhramarabarbiswal@giet.edu
DOI: 10.1201/9781003529231-26
Predictive AI Treatment for Kidney Tumors with Privacy Protection 163
models, but it can also pose privacy threats. Privacy- Implement a collaborative approach to manage the side
preserving machine learning aims to balance privacy with effects of advanced cancer treatment with patients.
ML benefits, ensuring data protection rules and privatizing To compare and select the most suitable guideline-directed
data acquisition [11]. Urologic malignancies, affecting both therapy for an advanced treatment plan based on patient
male and female urinary systems and reproductive organs, are characteristics and clinical data.
common [7]. Treatment for bladder and renal cell carcinoma
has improved with innovative checkpoint inhibitors, targeted Implement a joint approach with patients to effectively
therapy combinations, oral tyrosine kinas inhibitors, and manage the side effects of treatment.
antibody-drug conjugates [8]. The text discusses the use of Renal (Kidney) Optimize treatment Hyperparameter
machine learning in treating urology cancer, emphasizing the with Reinforcement Learning (RL): Kidney Cancer
need for effective coordination and communication between in Urologic Hyperparameters control machine-learning
patients and specialists [12]. A CT scan utilizes X-rays and a processes. Reinforcement Learning (RL) is a challenging
computer to create three-dimensional images of the kidneys framework for Hyperparameters optimization, following a
to detect urologic cancer invasion or spread to other organs Markov chain of improved model performance. Supervised
or lymph nodes [9]. learning (RL) is a challenging approach that predicts
treatment plan actions based on a series of “Cure/Abnormal
2. Literature Review State” to reach the next “state”. The model is trained to
predict the probability of the next urologic Renal or Kidney
Earlier works on these lines by various authors as shown in Hyperparameter, using the up-to-date Hyperparameters
Table 26.1. values and predictive performance. The Renal in Urologic
treatment plan model trains on discretized Hyperparameters
3. Proposed Work like behavioral medicine, nutrition, pain management, and
palliative care space, optimizing future treatment plans using
The proposed work aims to select a guideline-directed therapy predictive performance [15].
for advanced Kidney cancer based on patient characteristics
H:r = M(H) (1)
and clinical data at the end of the activity.
A Reinforcement Learning model R predicts a value q using optimizes for Hyperparameters with a prominent treatment
H and r, with q = R(H, r). The optimal action maximizes q, plan, with minimal impact on optimized Hyperparameters,
and can be predicted for past treatment H and r using the each with its own classification option. The policy gradient
formula q’=R(past H, past r), where r and q represent future is preferred for treatment plan classification due to its ease of
treatment plan values. The model minimizes mean square pain management and compatibility with good treatment plan
error by calculating q’ – (r + g * max q) ^ 2. Hyperparameters space dimensionality [15].
To indicate a treatment plan model preference for certain A clinical trial involving 70 participants with kidney tumors
Hyperparameters to be 1: L = –1*logP(next H|current H, found no return or cancer-related deaths during the treatment.
current r), use cross entropy to increase the probability of At 5 years post-treatment, only 15 had cancer evidence. Most
generating them. Policy gradient weighs the sample with the patients had decreased kidney function, but only with limited
reward value L = –(next Action Plan) * log P(next H|current further reductions. The trial results were already beating
H, current r) With next Action Plan = M(next H). The model surgical numbers for those with tumors 9 centimeters or less.
Fig. 26.1 Treatment study of clinical trail process of renal in urologic cancer
Predictive AI Treatment for Kidney Tumors with Privacy Protection 165
The average age was 77, with many being clinically obese S. No. Age T1a (Up to 4cm) T1b(4-7cm) T2a(7-10cm)
and having other medical conditions. Participants were
6 38 0 0 9
treated Bollineni hospital, Rajmundry, ensuring effective
radiation doses and quality treatment. Localized kidney 7 41 0 0 8.2
cancer is classified into three stages: T1a (up to 4 cm), T1b 8 31 0 6.5 0
(4-7 cm), and T2a (7-10 cm) to aid oncologists in determining 9 32 0 5.5 7.3
treatment options. 10 30 0 0 8.2
11 28 0 0 8.5
12 34 0 0 9
13 37 0 6.4 0
14 27 4 0 0
15 42 0 6.5 0
variables associated with kidney disease, with some variables model using new data to improve prediction accuracy.
potentially correlated. The dataset helps train supervised Continuously monitor and refine the privacy protection
algorithms, with a small percentage of missing values for mechanisms to adapt to evolving threats. Evaluate the AI
learning without noise data. The study assesses reinforcement model’s performance using metrics like accuracy, precision,
learning models’ accuracy, sensitivity, specificity, precision, recall, and F1 score. Assess the effectiveness of privacy
and F1-score in renal urologic cancer datasets, focusing on protection mechanisms. Develop a user-friendly interface
positive sample recognition and true nativities prediction. for healthcare professionals to interact with the AI system.
Ensure proper authentication and authorization to control
0.992 access to sensitive patient data. Deploy the AI system in a
0.99 healthcare environment, ensuring integration with existing
0.988 systems. Validate the system’s performance in a real-
0.986 world clinical setting. Ensure compliance with healthcare
0.984 regulations and standards. Keep abreast of evolving
0.982
privacy and security standards in the healthcare sector.
0.98
0.978
Provide training to healthcare professionals on using the AI
0.976 system and interpreting its recommendations.
0.974 The method outperforms other methods in terms of accuracy
measurements for Reinforcement Learning in Renal in the
Urologic Cancer graph. Implementation of AI in healthcare
requires careful consideration of ethical, legal, and regulatory
aspects. Regular evaluation, validation in real-world
Fig. 26.4 Reinforcement learning predictive treatment of clinical settings, and continuous collaboration between AI
renal in urologic cancer developers, healthcare professionals, and patients are critical
for achieving positive outcomes.
Collect patient data, including demographics, medical
history, diagnostic test results, and treatment outcomes.
Ensure compliance with data protection regulations (e.g.,
HIPAA) and anonymize or pseudonymize patient data to
protect privacy. Clean and preprocess the data, handling
missing values and outliers. Split the dataset into training
and testing sets. Choose appropriate machine learning
models for predictive analysis. In your case, you might use
regression models for predicting treatment outcomes. Train
the models using the training dataset. Implement privacy-
preserving machine learning techniques to protect sensitive
patient information. Consider using techniques like federated
learning, homomorphic encryption, or differential privacy
to ensure data privacy during the model training process.
Use interpretable models or techniques to make the AI
system more understandable to healthcare professionals.
Provide insights into the factors influencing treatment
recommendations to build trust. Implement a collaborative
decision-making system where AI provides treatment
suggestions, but the final decision involves collaboration
between AI systems and healthcare professionals. Include
features for doctors to input additional clinical expertise
that the AI model may not capture. Regularly update the
(a)
Predictive AI Treatment for Kidney Tumors with Privacy Protection 167
Abstract: Pomegranate quality is an issue in the food supply chain, and as a result, a lot of fruit goes to waste. Through
continuous monitoring and prediction of the condition of fresh fruit, a digital twin—the virtual twin of a crop—can assist in
the reduction of wasted pomegranates. The introduction of the thermal camera as a data-gathering instrument was due to its
capacity to identify surface and physical changes in stored fruits. Using SAP’s smart technologies, we trained a model with four
distinct sets of temperature data for this experiment. We tracked the fruits’ condition by training a deep convolutional neural
network (DCNN) with temperature data. The technology’s achievement of 0.99 prediction accuracy shows that it has great
potential for creating digital twins of fruits. It is possible to decrease food waste in the supply chain by making a digital copy
of fruit using thermal photography and machine learning algorithms.
Keywords: Convolutional neural network, Digital twin, Machine learning, Pomegranate fruit etc.
1. Introduction to address this issue, researchers have been looking into state
of-the-art approaches like spectroscopy, imaging, and sensor
Agriculture is the backbone of the economies of most technologies to enhance the accuracy and reliability of quality
developing nations in Southeast Asia. Fifteen percent of assessments. Although they have certain limitations, non
India’s gross domestic product comes from the agricultural destructive methods that integrate chemometrics with near-
sector, which provides sustenance for about half of the infrared spectroscopy (NIRS) could be useful for evaluating
population. The agro-ecological practices in India are varied various agricultural goods. However, the true power lies in
and extensive [1]. The pomegranate stands out among fruits integrating all of these methods into one cohesive framework
due to its numerous health benefits, vivid colour, and rich to create a hybrid approach that accurately evaluates
flavour. The increasing demand for high-quality fruit in pomegranate quality variances. A more practical alternative
global markets has intensified the need to ensure the quality to damaging measurements, NIRS now incorporates machine
of pomegranates. In addition to being delicious, visually learning, ANN, regression models, processing capacity,
appealing, healthy, and safe, pomegranates offer many more and other prediction techniques. Researchers have also
great attributes. Though practical, conventional methods demonstrated the practical feasibility of NIRS by employing
of determining fruit quality often lack the precision and appropriate preprocessing techniques and wavelength
thoroughness needed to meet modern benchmarks. In order selection approaches. A more comprehensive understanding of
1
saiprapu@buffalo.edu, 2guptavkrao@gmail.com, 3prof.adavi@gmail.com
DOI: 10.1201/9781003529231-27
Developing a Hybrid Approach to Assess Changes in Pomegranate Quality 169
the fruit’s characteristics is just one benefit of this integration; to get premium inputs from distributors or producers, and
others include improved supply chain management, happier picking high-quality products is risky because of unseen
consumers, and more efficient agricultural techniques. This damage. It takes a lot of time and money to use traditional
study explores the establishment of a hybrid strategy with the testing methods [10–13]. Furthermore, the conventional
goal of improving our knowledge of and ability to ensure the sampling and testing processes often result in significant
quality of this rare fruit. By integrating traditional wisdom product waste. There are four main characteristics of fruits
with cutting-edge technology, this study creates a holistic and vegetables, according to research [14–16]: texture, taste,
method for pomegranate quality evaluation that will shape colour, and nutritional content. The qualities of fruits and
future practices in the agriculture industry. vegetables impact how consumers perceive them, how much
they eat, and what they put them to use; as a society, we have
2. Literature seen a substantial uptick in the intake of fruits [18]. In order
to guarantee the quality, safety, and return on investment
Manufacturers in the agro-based processing business are of popular fruits such as apples, oranges, and kiwis, it is
currently focusing on producing fresh and lightly processed essential to sort, examine, and control the fruit’s quality [19–
items. Responding to customer demand in this context, 21]. Still, having a top-notch product is essential, especially
the market has introduced new techniques and products. when it comes to exporting. Looks, tastes, textures, and
The primary raw resource for many food operations is nutritional value are the four main criteria for evaluating
agricultural items [2, 3]. Perishable fruits and vegetables produce quality [14–16]. These characteristics significantly
have a limited shelf life in ambient conditions, but it can influence the preference, utilization, and consumption of
be extended in refrigerated storage. Perishable fruits and fruits, vegetables, and their products [17]. People eat a lot
vegetables should not be stored for long periods of time due more fruit now than they did even a decade ago, thanks to
to their susceptibility to spoilage by bacteria and chemicals. the general improvement in people’s living conditions [18].
Consequently, it is critical to complete post-harvest processes If you want your fruit investment to pay off, you need to sort,
like sorting and grading without delay. Countless situations check, and manage the quality of fruits like pomegranate,
call for smart reactions and complicated solutions, where banana, mango, jujube, apple, orange, kiwifruit, peach, grape,
artificial intelligence (AL) plays a significant role [4]. Modern and strawberry, sections 19–21]. Having a top-notch product
image processing methods and machine learning enable the is still essential, especially when targeting the export market
detection of fruit illnesses. Here are a few similar scientific [22]. The pomegranate fruit, depicted in Fig. 27.1, is widely
discoveries that have recently come to light: Because bananas consumed both fresh and processed into various forms such
are susceptible to so many different diseases, Song et al. as seed oil, dried arils, and juice [23]. The pomegranate tree
used image processing to detect when problems arose, so or shrub is spherical in shape. Each edible aril contains a seed
they could take measures to stop the spread of sickness and enclosed in a see-through sac that holds juice [24, 25] and is
restore normal production [5]. Chen et al. [6] laid out a plan protected from the elements by a thick, tough skin known
and method for supplying fresh agricultural products through
the use of Internet of Things technology. Before agricultural
products may be sold abroad, they usually need to have
their manufacturing process, marketing, agricultural quality
control systems, and auxiliary measures improved. This
allows for the integration of agricultural production with IT
and objective technology services provided over the Internet
[7]. This study used image processing to identify correlations (a) (b) (c)
between pomegranate size, colour, and appearance. In most
cases, removing the skin of a pomegranate fruit will reveal its
size and colour. In addition, workers sort the pomegranates
by hand. This led to the application of image processing and
AI to grade pomegranates based on their size and colour [8].
Ensuring the quality and safety of the end goods requires the
use of analytical techniques across the post-harvest supply
(d) (e) (f)
chain and prior to processing [9]. Assessing the quality of
raw fruit, maintaining cultivar validity, and identifying Fig. 27.1 Evaluates the quality of pomegranate using non
product damage are all obstacles that traditional methods of destructive approaches, focusing on its (a) Complete
fruit and vegetable fault detection encounter. It is challenging fruit, (b) fresh-cut fruit, (c) aril, (d) seed, (e) oil, and
(f) juice
170 Algorithms in Advanced Artificial Intelligence
as the peel [23]. The many health benefits and nutritional The literature highlights a shift in pomegranate quality
benefits of pomegranate fruit have contributed to its rising assessment, incorporating spectroscopy, imaging, sensor
popularity in recent years [26–28]. Because of this newfound technologies, chemical analysis, data fusion, IoT devices,
awareness around the world, pomegranate fruit has become and blockchain technology. This synthesis improves
much more popular for commercial production [29, 30]. understanding, promotes sustainable practices, informed
More recently, pomegranate fruit has been utilised for animal consumer choices, and economic growth.
feed, metabolomic peel extract, and as a powerful antioxidant
[31, 32], among other value-added products [33, 34]. The 3. Theoretical Framework
industry is working on better ways to grade pomegranate fruit
according to size, weight, and the appearance of the outside Utilising a digital twin (DT) is a very efficient method
rind [38], which helps determine the fruit’s freshness and for reducing food supply chain waste. The term “digital
worth [35–37]. Pomegranate fruits are categorised according twin” (DT) refers to a digital replica of a physical object
to their exterior, the thickness of their peel [17, 22], and the that is identical to the original in every way, including the
fragility of their arils. To guarantee safe handling, quality dimensions, shape, and composition of the product. It also
detection requires efficient, non-destructive technologies that needs to be able to faithfully recreate any major change that
are quick to respond [36]. happens during the product’s lifetime. In addition, sensors
that can continuously update data in real-time should be
A study developed a color-based method to identify
employed in conjunction with the actual architecture shown
pomegranates on trees in Fig. 27.2 using close-up images,
in Fig. 27.1.
considering artificial intelligence costs in agriculture, and a
model for pomegranate supply chains. As shown in Fig. 27.2, the platform will store and process
measurement data input into the DT in accordance with the
recommended design. Adding smart components to both new
and old products, connecting them to a cloud-based location
with streaming, enumerated data, and analytics capabilities
to gather data from sensors, continuously analyzing this data,
Fig. 27.2 Image processing is used to identify pomegranates and finally leveraging digital insights to revolutionize the
on a tree, as part of a case study on sustainable food retail industry are some of the steps involved in putting
closed loop supply chains in Iran
this system into practice. The IoT edge, which may link
A. Source image (pomegranate tree). B. Black and white the thermal camera to IoT cloud services, is the first of the
picture (segmentation, color indexing). C. Picture after the
two layers that make up the DT solution. It is impossible to
applied threshold (figure analysis). D. Location of geometric
connect more than one thermal camera over a Wi-Fi gateway
centers (estimated count of the pomegranates).
without a MQTT message broker. The second tier of the usable before it spoils. Additionally, this approach deals with
suggested solution includes an Internet of Things (IoT) cloud the issue of unoccupied object storage areas.
service that offers enterprise-level capabilities and services
pertaining to supply chain, product, and service management. 4. Proposed Method
If you follow this approach, you may connect all of your
Internet of Things (IoT) devices to the cloud using SAP Edge We created a DT of pomegranate fruit and evaluated its
Services Cloud Edition and save all of the thermal camera quality based on temperature changes using a thermal
photographs in SAP Document Services. SAP Intelligent imaging method as a data-collection instrument. The thermal
Service is able to predict the fruit’s quality throughout camera’s photographs of the targets were used as a training
storage based on the snapshot. Similar to this, SAP Asset set for CNN. With the help of machine learning, these images
Intelligent Network keeps an eye on the digital fruit, gathers reveal physiological data regarding the fruit’s state, allowing
data for forecasts, and alerts users. Customers can access a for an accurate prognosis. Deep learning has really improved
custom responsive web application called FoodDT client— the performance of tasks that include classification. Images
Retailer Front-End, which allows them to quickly access taken by infrared thermal cameras are analysed for patterns
the SAP Asset Intelligent Network notification and use it as that can be used for picture classification. This training has
a decision-making tool. With this information in hand, the made use of supervised machine learning. Two essential
consumer can take measures to avoid losing their purchase. components of this machine learning type are the training
inputs and the desired outputs (also called “labels”). The
Figure 27.6 suggested the fruit status category based on
application of image processing and classification technology
the traditional procedures and trends used by retailers.
is growing in tandem with its effectiveness.
Thus, “fresh” fruits are those that can fully meet customer
demand, whereas “good” fruits with a high market value Pomegranate Fruit Data Collection: FLIR image capture
necessitate targeted discounts from retailers. You can save one thermal camera was used to create the data set. Over the
a life by donating “bad” fruit that would otherwise go to course of the storage period, several times were used to snap
waste to a charity or other group. Failure to follow these the images. Prior to the start of the training procedure, the
safety measures may result in discarding and labeling the fruit photo collection was divided into four groups: Fresh,
fruit as “damaged.” Warehouses and food delivery services Excellent, Poor, and Damaged (see Fig. 27.5).
are only two examples of logistical infrastructures that can The prediction model was trained using SAP Intelligent
substantially benefit from this technology’s ability to increase Service. SAP Intelligent technologies now leverage an
transparency. As part of inventory management processes, efficient deep learning architecture called Tensor Flow. A
technology enables continuous DT updates with data from well-known machine-learning framework is used in the
food products. Quick decisions in near-real-time or in real- study. The technology includes deep neural networks and
time will be possible thanks to the cloud, which will serve as highly developed predictive modeling capabilities. While
the principal platform for its implementation’s management. the training dataset included large photos for preservation
After using this solution, stores will have a clearer picture of purposes, the validation and test datasets were created using
their inventory and will know exactly how much food is still several images with four labels for each category. Eighty
172 Algorithms in Advanced Artificial Intelligence
Normal vision
Thermal vision
percent of the original training set’s data are included in the (CNN) is one type of multi-layer neural network, as shown
training dataset; the remaining twenty percent are divided in Figure 6. A well-known feed-forward network extracts
across the test and validation datasets. the structure of an image. By applying the back-propagation
Delivery Model: Both the training and inference phases technique, we can train them to recognize patterns in pictures.
of the training algorithm were evaluated. Use the learning All of the neurons that comprise a feature are given identical
stage to explain the data and produce a trained model. The weights, disregarding their biases, to extract features. Weight
image must be converted into a vector representation for primarily attributes to the steep activation function.
educational purposes. When selecting a model and looking The parameter bias affects both the steepness and rate
for the model’s parameters, the learning process makes use of triggering of the activation function, contributing to
of this representation. the model’s ability to fit the data as closely as possible. If
During the inference stage, predictions about new data are this theory is correct, then each neuron reliably detects
made using the trained model. Feature vectors are real data in the same feature in the input picture. People think that the
the approach’s induction and training phases. It’s a method of most important component of a neural network for feature
assessing a model’s performance with fresh data to determine extraction is the convolutional layer, which is responsible for
how well it can predict the future and gauge its efficacy. CNN operations. It searches the picture for patterns using a
The model is used in the inference phase to draw informed collection of trainable filters called kernels. By dragging the
conclusions about fresh data. This is the same as using the filter over different parts of the input picture, convolution
model in real-world situations. The process will leverage real- creates a dot product between the filter and those parts.
time thermal camera data in the DT situation. After training Each grid will be useful, and the pooling (sub-sampling)
the neural network depicted in Fig. 4, the implementation of layer reduces the size of the feature maps by using various
the DT concept can involve feeding real-time data into it. As approaches to condense sub-regions. This layer lessens the
a result, the end users will be informed about the product’s occurrence of overfitting by relocating a region over the
status, and the forecast will be put into practice utilizing the input and sending the content of the window. Using pooling,
past data kept in the SAP text cloud storage. we can reduce the overall number of network parameters
while simultaneously making the learned features more
An assessment of a model using convolutional neural networks resistant to changes in size and orientation. The layer above
(CNN) will be conducted. A convolutional neural network it connects fully to the output of the layer below. As a result,
Developing a Hybrid Approach to Assess Changes in Pomegranate Quality 173
it establishes connections with all of the neurons in the layer for the network to arrive at a solution. According to reference
above it. Adding a fully linked layer is another way to learn [41], the typical values of η for convolutional neural networks
nonlinear combinations of these properties. According to the (CNNs) during training are 0.01, 0.1, 0.5, and 0.9. Raising η as
rule of mistake correction and learning, this method works. the step size grows could potentially accelerate convergence
Using a two-stage propagate-adapt cycle, the network learns as the network error decreases. On the other hand, the network
a collection of input-output sample pairings that have been may deviate excessively from the actual lowest values if η
pre-defined. The weighted inputs are added together by increases too much. Optionally, including a parameter called
this continuous-non-linear node, and then the output is sent momentum (α) aids in the network’s convergence. When
through a sigmoid non-linearity equation (1). In Appendix changing the variable α to a positive integer smaller than 1
A, you may find a flow diagram and a synopsis of the Back [39], the weight change equations on both the output and
Propagation algorithm. The first layer of network units hidden layers are usually altered.
receives an input pattern, and that pattern is passed on to Performance Model: In addition to the particular pattern
each layer above until an output is formed. The network’s reorganization job, the effectiveness of this rule is dependent
synaptic weights remain fixed throughout the forward pass. on the initial weight settings, learning rate parameter, output
By comparing this pattern to the intended output, the network functions of units, and presentation-training data. It’s crucial
determines an error signal for each output unit. Next, the to correctly set the weight values before implementing the
network transmits the error signals from the output layer to learning law for a particular training set. Initial weights are
each node in the intermediate layer that directly contributes in line with previous understanding. When knowledge is
to the output (reverse pass). The network adjusts each node’s possessed and appropriately presented in the form of starting
connection weights in response to the received error signal, weights, the ensuing trained network’s overall performance,
bringing it one step closer to the point where all training both in terms of learning speed and generalization speed, will
patterns may be encoded. It is customary to set the starting be greatly enhanced. The algorithm minimizes the average
weight values to low, arbitrary integers. The approach will error between the calculated outputs and the presented targets
not work well with multilayer networks if the initial weights using a gradient descent approach. The algorithm computes
are zero or have poorly chosen non-zero values [39]. All the outputs (Oj) from the inputs (Ii) using the logistic sigmoid
of the error words in the output layer determine the weight function provided by:
updates of the hidden layer. The hidden layer receives the
known output layer errors and uses them to determine how 1
O j = f ( x) =
to modify its weights. However, there is no guarantee that 1 + e x
the learning law will converge towards the goal solution for n
any given pattern-mapping assignment [40]. A learning rate Where activation x = Â wij Ii
parameter (と) is added to the output layer weight updating i =1
process to overcome this problem. η is usually a modest The connection weights represent the parameters of the back
number between 0 and 1, and it must be within that range propagation algorithm that can be adjusted. The number
of input and output neurons determines the number of
hidden neurons. We assessed the model’s performance by
measuring loss and accuracy. While training accuracy shows
the percentage of data used in the current dataset that was
properly identified, validation accuracy shows the accuracy
of randomly chosen photos from a different class. However,
the loss demonstrates how well the model performed on both
the training and validation datasets. The loss is calculated
by summing up the errors for each example in the validation
or training sets. To find out how much the projected values
differ from the actual values of the training data, accuracy
compares the categorised picture with data from a different
source called ground truth data. The pace of input acquisition
during training may potentially impact the convergence
of neural networks. The convergence of neural networks
is profoundly impacted by the number of epochs. Training
won’t converge quickly if the learning rate is low and slow
if it’s high. This training made use of an ideal learning rate.
Fig. 27.6 Three layer feed forward neural network
174 Algorithms in Advanced Artificial Intelligence
The following information was included in the model’s post- and environmental safety. Q Assurance Safety Crops Foods.
training summary: batch size, learning rate, total number (2021) 13: 20–36. doi: 10.15586/qas.v13i1.838.
of training epochs, greatest accuracy, final test accuracy, 3. Yadav A, Kumar N, Upadhyay A, Singh A, Anurag RK,
predicted top classes, and training duration, according to Pandiselvam R. Effect of mango kernel seed starch-based
active edible coating functionalized with lemongrass essential
Table 27.1. Despite 149 training epochs, the model did not
oil on the shelf-life of guava fruit. Q Assurance Safety Crops
achieve any further advancement after epoch 21.
Foods. (2022) 14: 103–115. doi: 10.15586/qas.v14i3.1094.
4. Jiang, Y., Li, K., Chen, S., Fu, X., Feng, S., & Zhuang, Z.
Table 27.1 Summarizes training (2022). A sustainable agricultural supply chain considering
Possessions Value substituting organic manure for chemical fertilizer. Sustainable
Production and Consumption, 29, 432–446.
Training Batch Size 65
5. Song, L., Luo, Y., Chang, Z., Jin, C., & Nicolas, M. (2022).
Learning Rate 0.001 Blockchain adoption in agricultural supply chain for better
Total Training Epoch 149 sustainability: A game theory perspective. Sustainability,
Epoch with Best Accuracy 6 14(3), 1470.
6. Chen, X., Chen, R., & Yang, C. (2021). Research and design
Best Validation Accuracy 0.99
of fresh agricultural product distribution service model
Predicted Top Classes 4 and framework using IoT technology. Journal of Ambient
Final Test Accuracy 0.99 Intelligence and Humanized, Computing17–1.
7. Bohle, C., Maturana, S., & Vera, J. (2010). A robust
The image classifier trained for loss reduction achieved optimization approach to wine grape harvesting scheduling.
an average accuracy of 0.99, reducing cross-entropy loss European Journal of Operational Research, 200(1), 245–252.
8. Hajiaghaei-Keshteli, M., & Aminnayeri, M. (2014). Solving
and validation-cross entropy loss to 0.005 and 0.08. The
the integrated scheduling of production and rail transportation
model accurately predicted fruit status, indicating machine
problem by Keshtel algorithm. Applied Soft Computing, 25,
learning’s potential for fruit decision-making (DT), reducing 184–203.
costs and labor-intensive tasks in the fruit supply chain. This 9. Beegum PS, Pandiselvam R, Ramesh SV, Thube SH, Pandian
innovative approach could significantly improve decision- TP, Khanashyam AC, et al. A critical appraisal on the
making. antimicrobial, oral protective, and anti-diabetic functions of
coconut and its derivatives. Q Assurance Safety Crops Foods.
(2022) 14: 86–100. doi: 10.15586/qas.v14i2.1040.
5. Conclusions 10. Kaavya R, Pandiselvam R, Mohammed M, Dakshayani
To evaluate the quality of pomegranates, the study suggests R, Kothakota A, Ramesh SV, et al. Application of infrared
a mixed method that combines chemical analysis, sensory spectroscopy techniques for the assessment of quality and
evaluation, visual inspection, the Internet of Things safety in spices: a review. Appl Spectrosc Rev. (2020) 55:
593–611. doi: 10.1080/05704928.2020.1713801.
(IoT), and blockchain technology. Colour, flavour, aroma,
11. Munawar AA, von Hörsten D, Wegener JK, Pawelzik E,
nutritional value, and physical attributes are all captured
Mörlein D. Rapid and non-destructive prediction of mango
by this model as indicators of pomegranate quality. The quality attributes using Fourier transform near infrared
decision support system provides insights that enable making spectroscopy and chemometrics. Eng Agric Environ Food.
supply chain decisions with more knowledge. The iterative (2016) 9:208–15. doi: 10.1016/j.eaef.2015.12.004
approach guarantees flexibility and continued applicability in 12. Niu C, Guo H, Wei J, Sajid M, Yuan Y, Yue T. Fourier
a dynamic agricultural setting. Our innovative method raises transform near-Infrared spectroscopy and chemometrics
the bar for the agricultural industry by improving the quality to predict zygosacchromyces rouxii in apple and kiwi fruit
assessment of fruits and agricultural products, which in turn juices. J Food Prot. (2018) 81:1379–85. doi: 10.4315/0362
increases consumer trust and bolsters sustainable farm-to 028X.JFP-17-512
table operations. 13. Yan H, Xu YC, Siesler HW, Han BX, Zhang GZ. Hand-Held
Near-Infrared Spectroscopy for authentication of fengdous
and quantitative analysis of mulberry fruits. Front Plant Sci.
References (2019) 10:1–15. doi: 10.3389/fpls.2019.01548.
14. Pandey, C.; Sethy, P.K.; Biswas, P.; Behera, S.K.; Khan,
1. Government of India, 2011. Faster, Sustainable and More M.R. Quality evaluation of pomegranate fruit using
Inclusive Growth: an Approach to the 12th Five Year Plan image processing techniques. In Proceedings of the 2020
(Draft), Planning Commission. Government of India, New International Conference on Communication and Signal
Delhi. Processing (ICCSP), Chennai, India, 28–30 July 2020;
2. Madanayake NH, Hossain A, Adassooriya NM. p. 19914067. [CrossRef]
Nanobiotechnology for agricultural sustainability, and food
Developing a Hybrid Approach to Assess Changes in Pomegranate Quality 175
15. Opara, U.L.; Pathare, P.B. Bruise damage measurement and 29. Holland, D.; Hatib, K.; Barya, I. Pomegranate: Botany,
analysis of fresh horticultural produce—A review. Postharvest horticulture, breeding. In Horticultural Reviews; John Wiley
Biol. Technol. 2014, 91, 9–24. [CrossRef] Sons, Inc.: Hoboken, NJ, USA, 2009; Volume 35, pp. 127–
16. Spielmanns, R.; Spielmanns, J.; Damerow, L.; Blanke, 192.
M.M. Non-destructive determination of surface features 30. Dhinesh, K.; Ramasamy, D. Pomegranate processing and
of pomegranate fruit. Acta Hortic. 2016, 1137, 247–250. value addition: Review. J. Food Process. Technol. 2016, 7,
[CrossRef] 565. [CrossRef]
17. Khoshroo, A.; Keyhani, A.; Rafiee, S.; Zoroofi, R.A.; Zamani, 31. Akuru, E.A.; Oyeagu, C.E.; Mpendulo, T.C.; Rautenbach, F.;
Z. Pomegranate quality evaluation using machine vision. Acta Oguntibeju, O.O. Effect of pomegranate (Punica granatum
Hortic. 2009, 818, 347–352. [CrossRef] L.) peel powder meal dietary supplementation on antioxidant
18. Wang, H.; Peng, J.; Xie, C.; Bao, Y.; He, Y. Fruit quality status and quality of breast meat in broilers. Heliyon 2020, 6,
evaluation using spectroscopy technology: A review. Sensors e05709. [CrossRef]
2015, 15, 11889–11927. [CrossRef] [PubMed] 32. Akuru, E.A.; Mpendulo, C.T.; Oyeagu, C.E.; Nantapo,
19. Czieczor, L.; Bentkamp, C.; Damerow, L.; Blanke, M. Non- C.W.T. Pomegranate (Punica granatum L.) peel powder meal
invasive determination of the quality of pomegranate fruit. supplementation in broilers: Effect on growth performance,
Postharvest Biol. Technol. 2018, 136, 74–79. [CrossRef] digestibility, carcase and organ weights, serum and some meat
20. Elmasry, G.; Kamruzzaman, M.; Sun, D.-W.; Allen, P. antioxidant enzyme biomarkers. Ital. J. Anim. Sci. 2021.
Principles and applications of hyperspectral imaging in 33. Magangana, T.P.; Makunga, N.P.; la Grange, C.; Stander,
quality evaluation of agro-food products: A review. Crit. Rev. M.A.; Fawole, O.A.; Opara, U.L. Blanching pre-treatment
Food Sci. Nutr. 2012, 52, 8398. [CrossRef] promotes high yields, bioactive compounds, antioxidants,
21. Matityahu, I.; Marciano, P.; Holland, D.; Ben-Arie, R.; Amir, enzyme inactivation and antibacterial activity of ‘wonderful’
R. Differential effects of regular and controlled atmosphere pomegranate peel extracts at three different harvest maturities.
storage on the quality of three cultivars of pomegranate Antioxidants 2021, 10, 1119. [CrossRef]
(Punica granatum L.). Postharvest Biol. Technol. 2016, 115, 34. Magangana, T.P.; Makunga, N.P.; Fawole, O.A.; Stander, M.A.;
132–141. [CrossRef] Opara, U.L. Antioxidant, antimicrobial, and metabolomic
22. Khoshroo, A.; Keyhani, A.; Zoroofi, R.A.; Rafiee, S.; Zamani, characterization of blanched pomegranate peel extracts: Effect
Z.; Alsharif, M.R. Classification of pomegranate fruit using of cultivar. Molecules 2022, 27, 2979. [CrossRef]
texture analysis of MR images. Agric. Eng. Int. CIGR J. 2009, 35. Khodabakhshian, R.; Emadi, B.; Khojastehpour, M.; Golzarian,
11, 1182. M.R.; Sazgarnia, A. Development of a multispectral imaging
23. Okere, E.E. Non-Invasive Measurement of Quality Attributes system for online quality assessment of pomegranate fruit. Int.
of Processed Pomegranate Products. Master’s Thesis, J. Food Prop. 2017, 20, 107–118. [CrossRef]
Stellenbosch University, Stellenbosch, South Africa, 2020. 36. Munera, S.; Hernández, F.; Aleixos, N.; Cubero, S.; Blasco, J.
24. Pareek, S.; Valero, D.; Serrano, M. Postharvest biology and Maturity monitoring of intact fruit and arils of pomegranate
technology of pomegranate. J. Sci. Food Agric. 2015, 95, cv. ‘Mollar de Elche’ using machine vision and chemometrics.
2360–2379. [CrossRef] [PubMed] Agriculture 2022, 12, 2034 Postharvest Biol. Technol. 2019, 156, 110936. [CrossRef]
21 of 25. 37. Zhang, L.; McCarthy, M.J. Assessment of pomegranate
25. Karimi, M.; Sadeghi, R.; Kokini, J. Pomegranate as a postharvest quality using nuclear magnetic resonance.
promising opportunity in medicine and nanotechnology. Postharvest Biol. Technol. 2013, 77, 59–66. [CrossRef]
Trends Food Sci. Technol. 2017, 69, 59–73. [CrossRef] 38. Kumar, A.; Rajpurohit, V.S.; Bidari, K.Y. Multi class grading
26. Fawole, O.A.; Opara, U.L. Developmental changes in and quality assessment of pomegranate fruits based on
maturity indices of pomegranate fruit: A descriptive review. physical and visual parameters. Int. J. Fruit Sci. 2019, 19,
Sci. Hortic. 2013, 159, 152–161. [CrossRef] 372–396. [CrossRef]
27. Lansky, E.P.; Newman, R.A. Punica granatum (pomegranate) 39. Widrow, B., and Lehr, M.A., “30 Years of Adaptive Neural
and its potential for prevention and treatment of inflammation Networks: Perceptron, Madaline and Back Propagation”,
and cancer. J. Ethnopharmacol. 2007, 109, 177–206. Proceedings of the IEEE, Vol. 78, N0. 4, pp. 1415–1442,
[CrossRef] [PubMed] September 1990
28. Opara, L.U.; Al-Ani, M.R.; Al-Shuaibi, Y.S. Physico-chemical 40. Yegnanarayana, B., “Artificial Neural Networks”, Prentice
properties, vitamin c content, and antimicrobial properties of Hall of India, 2001.
pomegranate fruit (Punica granatum L.). Food Bioprocess 41. Haykin,S., “Neural Networks A Comprehensive Foundation”,
Technol. 2009, 2, 315–321. [CrossRef] Prentice Hall, New Jersey, 1999.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
176 Algorithms in Advanced Artificial Intelligence
Artificial Intelligence-Based
Communication through
Cat Facial Expressions
28
K. Bhargavi1
Professor, Malla Reddy Engineering College (Autonomous)
Ch. V. Phani Krishna2
Professor, Teegala Krishna Reddy Engineering College, Hyderabad
Bandla Srinivasa Rao3
Prof. in CSE, Teegala Krishna Reddy Engineering College, Telegana
Abstract: Recent artificial intelligence experiments show that domestic cats, known for their complex facial expressions, are
more proficient in communication both defensively and with people using the Face Recognition Application. This discovery
could strengthen the link between cats and people [1]. This may lead to the development of instruments that help adopters select
suitable feline companions and owners interpret cues from their pets. However, the mystery of cat communication remains
unsolved. The results of the experiment on the facial expressions of cats use sophisticated artificial intelligence technology. A
machine learning model is trained to eliminate noise from an image of a cat’s facial expression through the diffusion model
procedure, which requires adjustments for various data types. The study investigates the application of convolutional neural
networks in analyzing cat facial expressions and compares the results with a Vision Transformer.
Keywords: Artificial intelligence, Cat facial expressions, Diffusion model, Vision transformer etc.
DOI: 10.1201/9781003529231-28
Artificial Intelligence-Based Communication through Cat Facial Expressions 177
still communicated in a defensive way, they probably started the distribution mean and standard deviation matrix [8].
to show friendlier facial expressions when they collected near This can be divided into forward and reverse processes.
humans, perhaps in anticipation of meals [5]. The findings, Mathematicians often use the jargon of physical processes to
which could improve the bond between cats and their human formalize mathematical concepts, such as diffusion equations
companions, have been commended by experts in the field for Fick diffusion, heat diffusion, and Brownian motion [9].
[11]. It may eventually lead to the development of apps The Langevin equation, based on the Wiener process, is a
that help cat owners interpret their pets’ ambiguous cues, stochastic formulation with infinitely small steps and a normal
improving understanding and communication. It might also distribution of time increments. This intertwines diffusion
assist prospective cat adopters in selecting feline companions with white noise generation, making machine learning
who will get along better with their existing pets. Even models called diffusion models. The diffusion models, which
after the deep understanding of feline facial expressions has utilize a Gaussian prior to generating data, form the core of
illuminated the intricate world of cat communication, this text-to-image generative models [3].
fascinating subject still stands. The enigma behind this still
has to be resolved [17]. 3. Cat Facial Expressions Using
The seven portions of this paper’s response can be understood Convolutional Neural Networks
as follows: section 1’s broad introduction. Cat Facial
Classification is crucial for identifying cat facial expression
Expressions Using Diffusion in Machine Learning is covered
objects, with deep learning being more effective than
in Section 2. Cat Facial Expressions Using Convolutional
traditional techniques [13]. Convolutional neural network
Neural Networks covered in Section 3. The Section 4 deals
techniques are often used to improve facial features, making
with Facial of Cat Expressions Vision Transformer. Section
advances in image expressions. CNN is a powerful neural
5 deals with experimental results. Conclusion included in
network technique for image categorization and recognition.
section 6. References are included in Section-7.
It comprises layers like classifier, pooling, activation, and
convolution [18]. The activation function decides whether
2. Cat Facial Expressions Using to excite neurons and adapts information nonlinearly. The
Diffusion in Machine Learning convolution layer extracts feature maps, bypassing the learnt
filter [20]. Activation functions like sigmoid, Tanh, and ReLU
The diffusion model process involves adding noise to an can be used for feature maps. The classifier layer selects the
image of a cat’s facial expression and learning to remove it, class with the highest probabilities. CNN’s performance
then training a machine learning model to produce a denoise depends on its ability to handle large datasets and transfer
image[7]. The process of learning the mean matrix involves learning, with models trained on the Image Net dataset,
assuming a normal noise distribution and parametrizing allowing designers to adapt parameters for new tasks [19].
The Facial Expressions Algorithm Utilizing Convolutional 4. Facial of Cat Expressions Vision
Neural Networks for Image Classification as follows: Image
input a collection of [height, width, and channel] pixel values. Transformer
Feature Extraction Transformers’ Cat Facial Expressions image processing
1. A feature map can be obtained by employing a application is influenced by image data type, requiring
convolution neural network. adaptation of preprocessing and input format to fully utilize
their power in various data types. Transformers, a machine
(a) Convergence (RELU).
learning tool, can be used on any data type, including
(i) Select a kernel with a size of 4x4 and depth sequences or time series of data points, as they feed on vectors
that matches the input array. [16]. Vision Transformer (ViT) uses cat facial expressions
(ii) Convolutional processing is utilized to extract image data patches flattened through linear transformations
features from facial expressions. into vector format. This process reveals that transformers
(b) (Max Pooling) Pooling. outperform typical CNNs on high data scales, indicating that
(i) The process involves reducing the spatial size cat facial expressions image data is a similar concept [14].
of the feature map through dimensionality Time series are ideal for transformers, as demonstrated by the
reduction and then extracting a 2x2 image. Temporal Fusion Transformer. This model uses LSTM layers
2. The process involves extracting low-level features from to transform time series into a right-sized vector, capturing
an image by altering the channel size to 16, 32, 64, or short-term correlations and long-term ones, and is compatible
128 until the fourth layer. with PyTorch. Reinforcement learning is a Markov chain
approach that encodes states, actions, and rewards as vectors.
Classification: For video pictures, latent features are extracted using CNN,
1. The training phase involves sending smooth output to a actions are encoded using embedding, and rewards are
feed-forward neural network with back propagation in represented as vectors with one dimension [12].
each iteration.
2. The SoftMax Classification approach is used to 5. Experiment Result
categorize photos like face expressions by identifying
their dominant characteristics using a trained model Artificial intelligence experiments reveal that domestic cats
[12]. with complex facial expressions, as identified in the Animal
Fig. 28.2 How to provide for cat expressions images data to a transformer
Artificial Intelligence-Based Communication through Cat Facial Expressions 179
Web Dataset, AFHQ and Pet Database, are more proficient in A Literature Review, MDPI, Appl. Sci. 2023, 13(9), 5521,
communication using the Face Recognition Application. The https://doi.org/10.3390/ app13095521,2023.
study compares CNN and ViT models for accuracy, sensivity, 7. K Sarvakar: Facial emotion recognition using convolutional
specificity, precision, and F1-score. Accuracy measures neural Networks, ScienceDirect, https://www.sciencedirect.
com, 2023.
model performance across all datasets, sensitivity measures
8. M Wang: Facial expression recognition based on CNN,
recognition of positive U-Net samples, and precision
IOPscience, https://iopscience.iop.org, 2020.
determines the model’s ability to predict true nativities. CNN 9. M. Pantic and L. J. Rothkrantz: Automatic analysis of facial
is a machine learning technology for facial cat expressions. expressions: The state of the art, IEEE Transactions on Pattern
Analysis & Machine Intelligence, No: 12, PP: 1424–1445,
Table 28.1 Graph for relative algorithms CNN and ViT 2000.
Relative Accuracy Sensitivity Specificity Precision Fl-score 10. MF Hansen: Towards on-farm pig face recognition using
Analysis convolutional Neural Networks, Science Direct, https:// www.
Methods sciencedirect.com, 2018.
11. Marcelo Feighelstein, Ilan Shimshoni, Lauren R. Finka, Stelio
CNN 0.92 0.91 0.92 0.93 0.92
P. L. Luna, Daniel S. Mills & Anna Zamansky: Automated
VIT 0.99 0.99 0.99 0.99 0.98 recognition of pain in cats, Springer, Scientific Reports, 12,
2022.
12. N. Christou and N. Kanojiya: Human facial expression
recognition with convolution neural networks, Third
International Congress on Information and Communication
Technology, pp. 539–545, 2019.
13. P Purini: Real-Time Facial Expression Recognition using
CNN, 46, taylorfrancis.com, https://www.taylorfrancis.com,
2022.
14. R.N.V. Jagan Mohan and K. Raja Sekhara Rao: Target
Inference on Evaluation of Angle Oriented Cluster, Computer
Science and Information Technology 2(3): 121–125, 2014
DOI: 10.13189/csit.2014.020301, Copyright © 2014 Horizon
Research publishing all rights reserved, http://www.hrpub.org,
2014.
Fig. 28.3 Graph for relative algorithms CNN and ViT
15. R.N.V. Jagan Mohan: Angle Oriented Based Image Analysis
The convolutional neural network outperforms CNN and ViT Using L-Axial Semi-Circular Model, Published in Asian
in accuracy measurements in graphs. Journal of Mathematics and Computer Research, ISSN No:
2395-4205(Print), 2395-4213(Online), Vol-10, ISSUE-4,
Page No 320–331, 2016.
References 16. Ruben Winastwa: Image Classification with Vision
Transformer, Published in Towards Data Science, 2017.
1. Andresen, N. et al: Towards a fully automated surveillance 17. Reddy Navya, Ramisetty Upendra, “Predict Early Pneumonitis
of well-being status in laboratory mice using deep learning: in Health Care Using Hybrid Model Algorithms”, Journal of
Starting with facial expression analysis, PLoS ONE, 15, Artificial Intelligence, Machine Learning and Neural Network
e0228059, 2020. (JAIMLNN), Volume 3, 2023.
2. Apeksha Khopkar, Ashish Adholiya: Facial Expression 18. S Binta Islam: Animal Species Recognition with Deep
Recognition Using CNN with Keras, Bioscience Convolutional .Neural Networks, MDPI,https://www.mdpi.
Biotechnology Research Communications, 14(05): 47–50, com, 2023.
DOI: 10.21786/bbrc/14.5/10,2021. 19. Tawsin Uddin Ahmed, Sazzad Hossain, Mohammad
3. F Yao: Facial Expression Recognition Based on Convolutional Shahadat Hossain, Raihan ul Islam, Karl Anders: Facial
Neural Networks, Hindawi Publication Journal, https://www. Expression Recognition using Convolutional Neural Network
hindawi.com,2021. with Data Augmentation, IEEE Xplore, DOI: 10.1109/
4. Finka, L. R., Luna, S. P., Mills, D. S. & Farnsworth, M. J. The ICIEV.2019.8858529,07 October 2019.
application of geometric morph metrics to explore potential 20. Waller, B., Julle-Daniere, E. & Micheletta, J. Measuring the
impacts of anthropocentric selection on animals’ ability to evolution of facial expression using multi-species Faces,
communicate via the face: The domestic cat as a case study, Neurosci. Biobehav. Rev. 113, 1–11, 2020.
Frontier Veterinary Science, 1070, 2020. 21. Y Mao: Pet dog facial expression recognition based on
5. JC Kim: Hybrid Approach for Facial Expression Recognition convolutional neural network (CNN), Nature Journal, and
Using CNN, MDPI, and https://www.mdpi.com,2022. https://www.nature.com, 2023.
6. Joe Mauricio et al: Comparing Vision Transformers and
Convolutional Neural Networks for Image Classification: Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
180 Algorithms in Advanced Artificial Intelligence
A. Aswini Priyanka*
Assistant Professor, Dept of Computer Science and Design,
Sagi Rama Krishnam Raju Engineering College, Bhimavaram
Abstract: Skin diseases, causing rashes, inflammation, and itchiness, can be genetic or lifestyle-related and can be treated
with medications, creams, ointments, or lifestyle changes. The rhabdomyosarcoma, scoliosis, cardiac fibroma, and several
skin tumors. The instruction is to identify 2- to 4-mm erythematous macules and treat them twice daily with fluorouracil
cream until they disappear. This study investigated AI-based technologies for the diagnosis and classification of skin cancer
using convolutional neural networks, evaluating their reliability by looking at data set size, diagnostic classifications, and
performance metrics. Convolutional neural networks are used in the experiment to diagnose and classify skin cancer. The
reliability of AI-based technologies is evaluated through performance metrics, diagnostic classifications, and experimental
results involving a large dataset.
Keywords: Artificial intelligence, Convolutional neural networks, Erythematous, Fluorouracil, Performance metrics, Skin
disease
*Aswini.areti@gmail.com
DOI: 10.1201/9781003529231-29
Convolutional Neural Networks for the Identification of Skin Disorders 181
carcinoma [12]. Kaposi sarcoma is rare and mainly affects Image input a collection of [height, width, and channel]
people with weakened immune systems or those taking pixel values.
immunosuppressive medications. Merkel cell carcinoma Feature Extraction
causes shiny nodules and hair follicles, while sebaceous
1. To obtain a feature of skin image map, use a convolution
gland carcinoma is aggressive and originates from oil glands
neural network.
in the skin. These rare forms can be mistaken for other eyelid
problems [13]. (a) Convergence (RELU).
(i) Choose a kernel whose size is 4x4 and whose
depth matches that of the input array.
2. Proposed Work
(ii) Convolutional processing is used to get
A 16-year-old intellectually disabled teen with multiple Features of the Disease object Image.
macules on his back is evaluated. The history of cardiac (b) (Max Pooling) Pooling.
fibroma, rhabdomyosarcoma, scoliosis, and skin neoplasm’s,
(i) Reduce the spatial size of the feature map
and has been taught to recognize them. Gorlin syndrome,
using the dimensionality reduction process,
also known as Gorlin-Goltz syndrome or nevoid basal cell
and then extract the 2x2 disease image.
carcinoma syndrome is an autosomal dominant familial
cancer caused by a mutation in the patched 1 (PTCH1) gene 2. To extract low-level features from the skin disease
[15]. Symptoms include multiple basal cell carcinomas, image, carry out the steps stated earlier until the fourth
keratocystic odontogenic tumors, and dyskeratotic palmar layer, altering the channel size to a value of 16, 32, 64,
and plantar pitting. The condition is prevalent in White or 128.
populations and affects both men and women. The Gorlin Classification:
syndrome introduced diagnostic criteria and treatment 1. In each iteration of the training phase, smooth output
protocols. Patients present with multiple BCCs before 20 or is sent to a feed-forward neural network with back
excessive numbers. Treatment may include photodynamic propagation.
therapy, surgical excision, and Mohs micrographic surgery
2. Using the SoftMax Classification approach, a trained
[9]. model is utilized to categorize photos like disease object
AI-based skin cancer detection model, revealing their image by identifying their dominating characteristics.
reliability to be uncertain due to varying evaluation metrics,
CNN is a powerful neural network technique for image
image types, and data set size. categorization and skin disease recognition [13]. It comprises
Skin Cancer Disease Detection by Convolutional Neural layers such as classifier, pooling, activation, and convolution.
Network: The Classification approach is a crucial role, The convolution layer extracts feature maps, while activation
even though the quantity of the disease object database functions like sigmoid, Tanh, and ReLU are used. The
has a substantial impact on how well skin disease objects classifier layer selects the class with the highest probabilities.
are identified. Machine learning includes deep learning CNNs must handle large datasets and transfer learning
[4]. Since the properties are automatically extracted, deep models, which can be adapted to new tasks. The skin disease
learning is more effective than traditional machine learning objects algorithm utilizes CNN for image classification.
techniques. Additionally, “end-to-end learning” using
deep learning involves giving the network both tasks and
unprocessed image data. The majority of the time, advances
3. Experimental Result
in skin disease is made in skin image features including face The reliability test of AI-based technologies is evaluated
and eye expressions using the convolutional neural network using performance metrics, diagnostic classifications,
technique [5]. and experimental results from a 12000-record dataset like
accessdata.fda.gov. The study compares RGB, YcbCr, and 4. Erickson BJ: Magician’s corner: how to start learning about
CNN models for accuracy, sensivity, specificity, precision, deep learning, Radiol Artificial Intelligence, 1: e190072, doi:
and F1-score [17]. Accuracy measures model performance 10.1148/ryai.2019190072, 2019.
5. Jainesh Rathod, Vishal Waghmode, Aniruddh Sodha, and
across all datasets, sensitivity recognizes positive ResNet
Praseniit Bhavathankar: Diagnosis of skin diseases using
samples, and precision predicts true nativities. CNN, a deep Convolutional Neural Networks, IEEE Xplore, DOI: 10.1109/
learning approach, is trained on massive image datasets for ICECA. 2018.8474593, 2018.
skin disease diagnosis [17]. 6. Kimonis VE, Goldstein AM, Pastakia B, et al: Clinical
manifestations in 105 persons with nevoid basal cell carcinoma
Table 29.1 The graph compares the performance of RGB, syndrome. Am J Med Genet. 1997; 69(3): 299–308, 1997.
YcbCr, and CNN algorithms 7. Komal Chughtai, Rahul Gupta, Sunil Upadhaya,Samer Al
Comparative Accuracy Sensitivity Specificity Precision Fl Hadidi: Topical 5-Fluorouracil associated skin reaction ,Oxf
Methods score Med Case Reports,2017 Aug; 2017(8): omx043,Published
online 2017 Aug 10, doi: 10.1093/omcr/omx043, 2017.
RGB 0.80 0.81 0.80 0.81 0.81 8. Lev S, Furst K, Chern W: A pharmacokinetic evaluation of
YCbCr 0.88 0.87 0.86 0.85 0.86 0.5% and 5% fluorouracil topical cream in patients with
CNN 0.98 0.98 0.98 0.98 0.99 actinic keratosis, Clinical Therapy, 2001; 23: 908, 2001.
9. Maytin EV, Kaw U, Ilyas M, Mack JA, Hu B: Blue light versus
red light for photodynamic therapy of basal cell carcinoma
in patients with Gorlin syndrome: a bilaterally controlled
comparison study, Photo diagnosis Photodynamic Ther. 2018;
22: 7–13. doi: 10.1016/j.pdpdt.2018.02.009, 2018.
10. Micali G, Lacarrubba F, Nasca MR, De Pasquale R. The
use of imiquimod 5% cream for the treatment of basal cell
carcinoma as observed in Gorlin’s syndrome, Clinical
Expenses Dermatology,28 Suppl 1: 19–23. doi: 10.1046/
j.1365-2230.28.s1.7.x, 2003.
11. Mitu Pal, Bristi Rani Roy: Evaluating and Enhancing
the Performance of Skin Disease Classification Based
on Ensemble Methods, IEEE Xplore, Doi:10.1109/
ICAICT51780.2020.9333529, 2021.
Fig. 29.2 The graph compares the performance of RGB, 12. Ortega García de Amezaga A, García Arregui O, Zepeda
YcbCr, and CNN algorithms Nuño S, Acha Sagredo A, Aguirre Urizar JM: Gorlin-Goltz
syndrome: Clinic pathologic aspects, Med Oral Patol Oral Cir
The convolutional neural network performs best among the
Bucal, 13(6): E338-E343, 2008.
accuracy measurements for RGB, YcbCr, and CNN in the 13. Pomerantz H, Hogan D, Eilers D, Swetter SM, Chen SC, Jacob
graphs. SE, et al: Long-term efficacy of topical fluorouracil cream,
5%, for treating actinic keratosis: a randomized clinical trial,
4. Conclusion JAMA Dermatology, 151: 952, 2015.
14. Ribeiro PL, Souza JB Filho, Abreu KD, Brzezinski MS,
The study assesses the effectiveness of AI-based technologies Pignaton CC. Syndrome in question: Gorlin-Goltz syndrome,
in detecting and classifying skin cancers using a 12000-record An Bras Dermatology, 91(4): 541–543. doi:10.1590/abd1806
4841.20164428, 2016.
dataset, focusing on rhabdomyosarcoma, scoliosis, cardiac
15. Spiker AM, Troxell T, Ramsey ML: syndrome. In: Stat Pearls
fibroma, and other skin tumors. [Internet], Stat Pearls Publishing; Aug 8, 2023.
16. Spadari F, Pulicari F, Pellegrini M, Scribante A, Garagiola U.
References Multidisciplinary approach to Gorlin-Goltz syndrome: from
diagnosis to surgical treatment of jawbones, Maxillofacial
1. Borghesi A, Nardi C, Giannitto C, et al: Odontogenic Plastic Reconstruction Surgical 2022; 44(1): 25, doi: 10.1186/
keratocystic: imaging features of a benign lesion with an s40902-022-00355-5, 2022.
aggressive behavior, Insights Imaging, 2018; 9(5): 883–897. 17. YN Fu’adah: Convolutional Neural Network (CNN) for
Doi: 10.1007/s13244-018-0644-z, 2018. Automatic Skin, IOPscience, https://iopscience.iop.org, 2020.
2. Bree AF, Shah MR: BCNS Colloquium Group. Consensus 18. Witmanowski H, Szychta P, Błochowiak K, Jundziłł A,
statement from the first international colloquium on basal cell Czajkowski R: Basal cell nevus syndrome (Gorlin-Goltz
nevus syndrome (BCNS), Am J Med Genet A. 2011; 155A (9): syndrome): genetic predisposition, clinical picture and
2091–2097. doi: 10.1002/ajmg.a.34128, 2011. treatment, Postepy Dermatology Allegro,34(4): 381–387,
3. Bresler SC, Padwa BL, Granter SR. Nevoid basal cell doi: 10.5114/ ada.2017. 69323,2017.
carcinoma syndrome (Gorlin syndrome), Head Neck Note: All the figures and table in this chapter were designed by the
Pathology, 2016; 10(2): 119–124, DOI: 10.1007/s12105-016 author.
0706-9, 2016.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Online payment systems have become an integral part of the modern digital economy, facilitating convenient and
efficient transactions. However, they are also susceptible to various types of fraudulent activities. The potential for substantial
financial losses to both businesses and consumers underscore the urgency of addressing this escalating threat. Consequently,
there is a crucial need to develop resilient fraud detection systems. The main goal is to construct an effective and efficient online
payment fraud detection system that can promptly identify and thwart fraudulent transactions in real-time., thus enhancing
security and preserving the integrity of digital payment platforms. Our methodology begins by collecting a comprehensive
dataset containing transaction information; including transaction amount, location, time, and various other relevant features.
This dataset forms the basis for conducting training and evaluating our machine learning models, such as the Random Forest
Classifier and Logistic Regression models, which can help detect fraudulent activities.
Keywords: Fraud detection, Random forest classifier, Logistic regression, Feature engineering
DOI: 10.1201/9781003529231-30
184 Algorithms in Advanced Artificial Intelligence
This work is focused on the development and implementation engineering to identify transaction fraud effectively. This
of an Online Payment Fraud Detection System, leveraging model enhances its performance and stability by processing
machine learning techniques for the purpose of identifying large volumes of data, allowing it to accumulate valuable
and preventing fraudulent activities within online payment experience within the domain of fraud detection. [4].
transactions. The primary aim is to create robust and adaptable The paper titled “Online Transaction Fraud Detection System
system that can efficiently differentiate between legitimate Based on Machine Learning” addresses the challenge of
and fraudulent transactions in real-time while minimizing the detecting fraud in online transactions, which in contrast to
occurrence of false positives. credit card transactions, are characterized by the complexity
The initiative encompasses several pivotal components, of fraud detection is heightened by factors such as low cost,
including the process of gathering data, conducting pre extensive coverage, and high frequency, posing challenges
processing tasks, implementing machine learning models, to the process. Addressing this issue, the study presents two
undertaking feature engineering, and evaluating model fraud detection algorithms rooted in the Fully Connected
performance. Real-time monitoring and alert systems Neural Network and XGBoost [5]. The paper titled “A
have been integrated to ensure rapid responses to potential Survey of Deep Learning-Based Online Transactions Fraud
fraudulent activities. The deployment onto a scalable cloud- Detection Systems” offers an extensive examination This
based infrastructure stands as a crucial element, allowing survey delves into the utilization of deep learning methods
seamless integration into existing online payment systems, within the domain of online transaction fraud detection,
prioritizing accessibility and scalability. Upholding data offering valuable insights into the application of these
privacy and compliance with regulations remains a central techniques fraud detection for online transactions [6].
focus to protect sensitive information and adhere to legal
standards. Comprehensive documentation and reporting of 4. Existing System
methodologies, discoveries, and system implementations
are integral aspects of this endeavour. The extent of this The existing systems and technologies used for online
application’s scope holds significant potential for advancing payment fraud detection vary depending on the industry,
security measures. size of the organization, and specific requirements. However,
here is an overview of the components commonly found in
existing systems for online payment fraud detection.
3. Literature Review
The study titled “Real-Time Fraud Anomaly Detection in 4.1 Transaction Monitoring and Analysis
E-banking Using Data Mining Algorithm” aims to present Conventional systems commonly utilize a blend of rule-
an optimized model for the detection of financial fraud. based and machine learning algorithms to actively monitor
This model leverages a combination of feature selection and and analyze payment transactions in real-time. These systems
machine learning classification, ultimately enhancing its look for suspicious patterns and anomalies that may indicate
fraud detection capabilities [1]. fraudulent activity.
The research paper titled “Enhanced Credit Card Fraud
Detection Model Using Machine Learning” explores machine 4.2 Data Sources
learning models within a two-stage evaluation framework. Data sources may include transaction data, user profiles,
These models are applied to a real-world dataset comprising device information, IP addresses, and historical transaction
credit card transactions from European cardholders. The records. These data sources help in identifying patterns and
research employs a strained K-fold cross-validation technique trends associated with fraudulent transactions.
in the evaluation process [2].
4.3 User Authentication
The paper titled “A Fraud Detection System Using Machine
Learning” introduces a machine learning model designed Secure user authentication is crucial to prevent unauthorized
to effectively identify both “fraudulent” and “genuine” access to the system. Multi-factor authentication (MFA) is
transactions in real-time. The applicability of this solution commonly used to enhance security.
extends to a broad spectrum of sectors with financial
associations, offering significant benefits in the realm of 4.4 Dashboard and Reporting
transaction security.[3] Most systems include a dashboard for fraud analysts to
The study titled “Fraud Detection in Online Payments monitor transactions and generate reports. Real-time
using Machine Learning Techniques” has implemented a dashboards display transaction data and alerts for immediate
model based on machine learning that incorporates feature action.
Machine Learning-Based Approach for Detecting Online Payment Fraud 185
comprises transaction data, with each entry containing system. Here are the steps and recommended practices for
various attributes and a binary label indicating whether training and fine-tuning machine learning models for our
the transaction is legitimate (0) or fraudulent (1). Essential work:
components of such datasets include transaction attributes,
such as the transaction amount, location, timestamp, and 6.3 Data Collection and Pre-processing
other relevant features that provide context for the machine Gather a diverse and comprehensive dataset of past payment
learning model. Historical records are a crucial component transactions, including both legitimate and fraudulent cases.
of such datasets, enabling the machine learning model to Reprocess Prepare the data through cleaning, normalizing,
learn patterns and trends in both legitimate and fraudulent and transforming processes. This may encompass tasks such
transactions.[l] The parameters of this dataset are Transaction as addressing missing values, encoding categorical variables,
type, amount, nameOrig, oldbalanceOrg, newbalanceOrig, and scaling numerical features.
nameDest, oldbalanceDest, newbalanceDest-details are
In Fig. 30.1, the pie chart employed in online payment fraud
provided in Table 30.1.
detection serves as a visual representation that effectively
communicates the distribution of various transaction types
Table 30.1 Kaggile dataset of online payment
within a dataset. It provides a clear and concise overview
Variables Description Type of the proportion of different transaction categories, each
Transaction type It states the type of the Categorical denoted by labels such as “Cash Out,” “Cash In,” “Debit,”
transaction
“Transfer,” and “Payment.”
Amount Transaction amount Numerical
Name-origin Senders unique id ID In Fig. 30.2, This graph will effectively display the counts
Dest-Origin Receivers unique id ID associated with different steps, providing valuable insights
Old-balance-org Senders balance before Numerical
transaction
New-balance-org Senders balance after Numerical
transaction
Old-balance-dest Receivers balance before Numerical
transaction
New-balance-dest Receivers balance after Numerical
transaction
into the workflow. The “step” labels on the x-axis and the
count on the y-axis enable a clear understanding of the
frequency of each step within the process.
8. Conclusion
In conclusion, the endeavour to develop the front-end of an
online payment fraud detection system represents a significant
step towards enhancing the security and efficiency of online
payment processing. The work has addressed critical aspects
related to user interface design, user authentication, real-
time monitoring, data visualization, and the overall user
experience, contributing to combating fraudulent activities in
the digital economy.
Through the systematic methodology followed in this work,
we’ve effectively designed, developed, and deployed user-
Fig. 30.5 Precision recall curve
friendly front-end, providing fraud analysts with essential
In Fig. 30.5, The precision-recall graph depicts how the tools and insights to identify and address potential online
model’s performance varies concerning different thresholds payment fraud. The work has been guided by a profound
set for classification. understanding of existing systems and a dedication to
Threshold on X-axis: This represents the varying decision enhancing them to address the evolving challenges presented
thresholds applied by the model to classify instances. As by fraudulent activities.
the threshold changes, the model adjusts its classification, In conclusion, this project represents a significant step
affecting the balance between precision and recall. forward in the ongoing battle against online payment fraud,
Score on Y-axis:The score denotes the harmonic mean and it stands as a testament to the importance of user-centric
of precision and recall at each threshold, offering insights design, innovation, and adaptability in the realm of online
into the trade-off between precision (accuracy of positive security and fraud prevention.
Machine Learning-Based Approach for Detecting Online Payment Fraud 189
Abstract: In the period of data-driven decision-making, utilising machine learning to forecast loan approval has become
increasingly prevalent. However, the sensitive nature of financial data poses significant privacy concerns. Our methodology
employs advanced privacy-enhancing technologies to safeguard the confidentiality of sensitive financial information while
still achieving accurate loan predictions. We utilise secure homomorphic encryption techniques to ensure privacy, allowing
multiple parties to collaborate on model training without revealing their data. The proposed approach combines state-of-the
art encryption techniques with homomorphic privacy mechanisms to ensure robust privacy protection while maintaining high
model accuracy and utility in different phases. The privacy budget is quantified and controlled to ensure a balance between
model accuracy and data privacy. This research contributes to the growing work on privacy-preserving machine learning models.
It offers a viable solution for financial institutions seeking to harness the power of data-driven loan prediction while upholding
the highest data privacy and security standards.
Keywords: Decision making, Data-driven, Loan approval, Privacy preserving machine learning etc.
1
vsnaresh111@gmail.com, 2suahmakurella@gmail.com, 3swathisannidhi04@gmail.com, 4gantalakshmiishwarya85@gmail.com,
5
kirankirru232@gmail.com, 6tennetiswathi@gmail.com
DOI: 10.1201/9781003529231-31
Secure Loan Approval Prediction: A Privacy-Preserving Machine Learning Approach 191
selection to develop robust and reliable loan prediction potentially shedding light on the importance of client-related
models. Researchers have explored many algorithms, factors.
including traditional ones like logistic regression and Vimala and Sharmili (2018) [11]: This research focuses on
decision trees, and advanced techniques such as random predicting loan risk using Naive Bayes and Support Vector
forests,neural networks, support vector machines, and. Machine, offering insights into the comparative performance
Moreover, integrating alternative data sources, such as social of these algorithms.
media activity and transaction history, has been a recurring
theme, contributing to the diversification of features used for Priya et al. (2018) [12]: The paper explores exploratory
prediction. Furthermore, studies have tackled the challenges analysis on the prediction of loan privilege for customers
of class imbalance, interpretability, and model explainability using random forest, potentially introducing insights into the
to guarantee accurate forecasting; the predictions made by random forest approach for loan prediction.
these models align with regulatory requirements and can Ibrahim et al. (2020) [15]: The study compares the CatBoost
be comprehended by stakeholders. Research in this area classifier with other machine learning methods, providing
offers critical understanding regarding the development of insights into the comparative performance of different
loan prediction models, highlighting both the successes and algorithms.
challenges that have shaped the development of this crucial Tejaswini et al. (2020) [16]: The research focuses on accurate
application in the financial domain. loan approval prediction based on a machine learning
approach, potentially introducing novel techniques for
2. Related Work improving prediction accuracy.
Kumar et al. (2019) [1]: This study uses machine learning Gupta et al. (2020) [17]: The paper discusses a bank loan
techniques to predict loan approval. Supriya et al. (2019) prediction system using machine learning, likely providing
[2]: The paper explores machine learning models for loan insights into the system architecture and predictive models
prediction, emphasizing their application in decision-making employed.
processes. Vaidya (2017) [18,19]: This study employs logistic regression
Arun Kumar et al. (2016) [3] and (2016) [13]: These papers for predicting loan approval, potentially providing insights
discuss loan approval prediction based on machine learning into the predictive and probabilistic approach using this
approaches, providing insights into different algorithms and algorithm.
methodologies. Singh et al. (2021) [20]: The research focuses on predicting a
Ashwitha et al. (2022) [4]: The study proposes an approach for modernized loan approval system based on a machine learning
predicting loan eligibility using machine learning, potentially approach, potentially discussing integrating contemporary
introducing novel methods or features. Chawan et al. (2022) technologies.
[5]: The research focuses on a “Bank Loss Predictor,” likely
delving into the prediction of potential losses in the banking 3. Preliminaries
sector related to loans.
Barua et al. (2021) [6]: The paper introduces “Swindle,” a 3.1 Logistic Regression
system predicting the probability of loan defaults using the Logistic regression, a flexible linear regression analysis
CatBoost algorithm. model, is predominantly employed for supervised learning
Kirubanantham et al. (2021) [7]: The study explores credit tasks. It finds applications in regression, binary classification,
sanction forecasting, likely discussing methods to forecast and multi-classification problems. Implementing logistic
credit sanctions in the context of loan approval. Sheikh regression typically involves three key steps they are defining
et al. (2020) [8] and (2020) [19]: These papers present an a prediction function, creating a loss function, and optimising
approach for predicting loan approval using machine learning regression parameters to minimise this loss. A cost function
algorithms, potentially providing insights into the algorithmic is established initially when using logistic regression for
choices and features used. regression or classification. Then, an iterative optimisation
technique is employed to identify the optimal model
Soni and Paul (2019) [9]: The paper discusses an algorithm
parameters. Finally, the model’s performance is assessed.
for a loan credibility prediction system, likely providing
insights into the features and methods employed. Jency et The prediction function in logistic regression is closely tied
al. (2018) [10]: The research involves an exploratory data to the Sigmoid function, expressed as:
analysis for loan prediction based on the nature of clients, 1
S(x ) = 1 + (1)
1+ e - x
192 Algorithms in Advanced Artificial Intelligence
In this Equation, ‘x’ represents a variable. Figure 31.1 visually Now we are decrypting it with the help of secrete key sk,
illustrates the Sigmoid function curve, highlighting its ability Dec(sk, CA) Æ a
to represent outcomes within the range of 0 to 1 evenly. Dec(sk, CM) Æ m
where,
a = n1 + n2 & m = n1x n2 (3)
4. Proposed System
In this section, we have presented Privacy-Preserving Logistic
regression for loan prediction.
dataset is encrypted, it is securely transmitted to a third-party information remains confidential during transmission. This
organisation (Cloud), Maintaining the confidentiality of the step preserves the privacy of the loan applicants’ personal and
applicant’s personal and financial information. This step is financial data, as it is only accessible to the trusted banker,
significant in safeguarding loan applicants’ privacy while which provides input privacy.
allowing for predictive models’ development. Decryption and Model Application
Once the encrypted data is in the banker’s possession, they
6. Model Privacy can decrypt it using their private key. This decryption process
allows the banker to apply the previously obtained coefficients
6.1 Model Development with Homomorphic and intercept from the machine learning model to generate
Encryption a loan prediction result for each applicant. Importantly, this
In the second phase of our methodology, the third-party phase maintains the privacy of the loan applicants’ data, as
organisation (Cloud), which has received the encrypted the decryption is performed by the trusted banker responsible
dataset, proceeds to build a logistic regression model for loan for safeguarding the confidentiality of the information.
prediction without ever decrypting the data. This is achieved
through homomorphic encryption, a groundbreaking 7.2 Output Privacy
cryptographic technique that allows calculations to be Secure Result Transmission to the Client
Executed on encrypted data without revealing the basic Upon determining the loan approval status for each applicant,
information. By applying homomorphic encryption to the the banker securely communicates the results back to the
dataset, the third party can train and fine-tune the logistic respective clients. The result information is encrypted using
regression model while maintaining the confidentiality of the the client’s public key before transmission. This encryption
loan applicant’s sensitive attributes, such as income, credit guarantees that the loan decision remains confidential
history, and personal details. This verifies that the applicants’ throughout the communication process. Only the authorised
privacy is maintained throughout the model-building process. client, possessing the corresponding private key, can decrypt
and access their loan approval status.
6.2 Decryption and Model Evaluation
Client Decryption and Final Decision
Once the third party has successfully developed the logistic
regression model, the encrypted coefficients and model In the final phase of our methodology, the loan applicants
parameters are transmitted back to the banker. In the third phase receive the encrypted loan approval result from the banker.
of our methodology, the banker, who possesses the necessary The clients can then decrypt this information using their
private key, can decrypt these coefficients. Decryption private key. This decryption process allows the clients to
enables the banker to access the logistic regression model’s securely and confidentially obtain their loan decision, whether
insights and predictions, which are essential for making loan it is approved or not. By employing end-to-end encryption
decisions. This phase allows the banker to evaluate the logistic and secure key management, the privacy of the sensitive loan
regression model’s performance, assess the risk associated application data is preserved throughout the entire process,
with loan applicants, and make informed lending decisions starting from data gatheringto final decision delivery.
while keeping the applicants’ sensitive data confidential. By Input and output privacy in the testing phase are presented
employing this secure and privacy-preserving approach, we in Fig. 31.3; the privacy and confidentiality of sensitive data
balance data utility and privacy in loan prediction, ensuring
the protection of individuals’personal information while
maintaining the efficacy of the lending process.
are paramount. Homomorphic encryption and the secure • True Negative (TN): TN signifies instances where the
transmission of encrypted data and results ensure that only model correctly predicts a loan as denied, and it is
authorised parties, such as the banker and the clients, can indeed denied. TN represents the number of accurately
access and interpret the information, thereby safeguarding predicted loan denials.
the personal and financial details of the loan applicants while • False Negatives (FN): FN represents current status
facilitating the loan approval process. of cases that incorrectly predicts a loan as denied but
actually approved. FN indicates the number of missed
8. Experimental Results opportunities where a loan should have been approved
but was not.
We have experimented on a well-known bank loan dataset,
consisting of twelve features and 4269 records, which decides 8.2 Accuracy
whether to approve the loan. The dataset is partitioned into Our privacy-preserving machine learning model for loan
80percent training and 20percent testing samples. The code is prediction has demonstrated a commendable level of
developed using Python 3.10.12, with Intel(R) Xeon(R) CPU accuracy, achieving a consistent 74% accuracy rate across
@ 2.20GHz, collab. Coming to the results,we take accuracy, our experiments. Importantly, this accuracy was maintained
ROC curve and Confusion matrix as parameters to determine even when the model was applied to encrypted data,
the efficiency of our model. emphasising the robustness and effectiveness of our privacy-
preserving techniques. Maintaining a high level of predictive
8.1 Confusion Matrix
performance while safeguarding sensitive data through
In our binary loan prediction classification task, we encryption is a significant accomplishment, as it aligns with
employ a 2x2 confusion matrix Fig. 31.4 to assess our the paramount importance of preserving privacy in financial
privacy-preserving machine learning model’s performance applications. This finding underscores the feasibility of
comprehensively. This matrix comprises four essential employing privacy-preserving machine learning approaches
components, each shedding light on a distinct aspect of the to address the complex challenges of secure and accurate
model’s performance: loan prediction in a world increasingly concerned with data
• True Positives (TP): These are instances where the model privacy and security.
correctly predicts a loan as approved, and it is indeed
approved. TP indicates the number of successful loan 8.3 ROC Curve
approvals accurately predicted by the model. Furthermore, an essential aspect of evaluating the
• False Positives (FP): FP occurs when the model effectiveness of our privacy-preserving loan prediction model
incorrectly predicts a loan as approved but is denied. is the analysis of its Receiver Operating Characteristic (ROC)
This represents the number of loans mistakenly approved curve. The ROC curve confirms our model’s predictive
when they should not have been. capability and provides insights into its discrimination
performance. ROC curve presented in Fig. 31.5, consistently
lies above the diagonal line representing random classification, 5. Chawan, Brijesh, et al. (2022). “Bank Loss Predictor.” 2022
underscoring the model’s superiority over a random classifier. 3rd International Conference for Emerging Technology
This observation indicates our model’s ability to strike a (INCET).
6. Barua, Sujoy, et al. (2021). “Swindle: Predicting the
meaningful balance between true positive and false positive
probability of loan defaults using catboost algorithm.” 2021
rates. The consistently superior performance exhibited by our 5th International Conference on Computing Methodologies
model in this regard reinforces its reliability and underscores and Communication (ICCMC).
its potential as a valuable tool for financial institutions. In 7. Kirubanantham, P., Saranya, A., and Kumar, D. S. (2021).
conclusion, the ROC curve analysis affirms that our model is “Credit Sanction Forecasting.” 2021 4th International
a good and robust solution for loan prediction tasks. Conference on Computing and Communications Technologies
(ICCCT).
8. Sheikh, M. A., Goel, A. K., and Kumar, T. (2020). “An approach
9. Conclusion for prediction of loan approval using machine learning
algorithm.” 2020 International Conference on Electronics and
In conclusion, our journey in developing a privacy-preserving Sustainable Communication Systems (ICESC).
loan prediction model using the powerful combination of 9. Soni, P. M., and Paul, V. (2019). “Algorithm for the loan
homomorphic encryption and logistic regression has yielded credibility prediction system.” Int J Recent Technol Eng,
exceptional results. Our exertion of homomorphic encryption 8(1S4), 1080–1087.
technology has been pivotal in achieving the dual goals of 10. Jency, X. Francis, Sumathi, V. P., and Shiva Sri, J. (2018). “An
protecting data privacy and maintaining high accuracy. exploratory data analysis for loan prediction based on nature
Through the implementation of this cutting-edge encryption of the clients.” International Journal of Recent Technology
and Engineering (IJRTE), 7(4), 17–23.
technique, we ensured that sensitive customer information
11. Vimala, S., and Sharmili, K. C. (2018). “Prediction of Loan
remained confidential and unreadable by external entities, Risk using NB and Support Vector Machine.” ICACT 2018,
adhering to stringent privacy standards and regulations. Volume 4, Issue 2, 110–113.
Maintaining an impressive accuracy rate, notably at 74%, 12. Priya, K. U., et al. (2018). “Exploratory analysis on prediction
showcases the efficacy of our privacy-preserving approach. of loan privilege for customers using random forest.”
Furthermore, our model’s ROC curve consistently surpassing International Journal of Engineering Technology, 7(2.21),
the random classifier highlights its superiority and robustness 339–341.
in distinguishing between loan approval and denial.This dual 13. Arun, Kumar, Garg Ishan, and Kaur Sanmeet. (2016). “Loan
approval prediction based on machine learning approach.”
accomplishment underscores the potential of preserving data IOSR J. Comput. Eng, 18(3), 18–21.
privacy without sacrificing predictive power. Our work paves 14. KaggleDataset:https://www.kaggle.com /burak3ergun/loan
the way for a future where privacy and accuracy coexist data-set.
harmoniously, offering a robust solution to the challenge of 15. Ibrahim, A. A., et al. (2020). “Comparison of the CatBoost
secure loan prediction while respecting data privacy in the classifier with other machine learning methods.” International
financial sector and beyond. Future research could explore Journal of Advanced Computer Science and Applications,
the optimisation of these methods and their adaptation 11(11), 2020.
16. Tejaswini, J., et al. (2020). “Accurate loan approval prediction
to different data types and use cases. In the realm of loan
based on machine learning approach.” Journal of Engineering
prediction, our model serves as a testament to the feasibility Science, 11(4), 523–532.
of achieving high accuracy while safeguarding sensitive data. 17. Gupta, A., Pant, V., Kumar, S., and Bansal, P. K. (2020). “Bank
Loan Prediction System using Machine Learning.” 2020 9th
References International Conference System Modeling and Advancement
in Research Trends (SMART).
1. Kumar, R., et al. (2019). “Prediction of loan approval using 18. Vaidya, A. (2017). “Predictive and probabilistic approach
machine learning.” International Journal of Advanced Science using logistic regression: Application to prediction of loan
and Technology, 28(7), 455–460. approval.” 2017 8th International Conference on Computing,
2. Supriya, P., et al. (2019). “Loan prediction by using machine Communication and Networking Technologies (ICCCNT).
learning models.” International Journal of Engineering and 19. Sheikh, M. A., Goel, A. K., and Kumar, T. (2020). “An
Approach for Prediction of Loan Approval using Machine
Techniques, 5(2), 144–147.
Learning Algorithm.” 2020 International Conference on
3. Arun, Kumar, Garg Ishan, and Kaur Sanmeet. (2016). “Loan
Electronics and Sustainable Communication Systems
approval prediction based on machine learning approach.”
(ICESC).
IOSR J. Comput. Eng, 18(3), 18–21.
20. Singh, V., Yadav, A., Awasthi, R., and Partheeban, G. N.
4. Ashwitha, K., et al. (2022). “An Approach to Predict Loan (2021). “Prediction of Modernized Loan Approval System
Eligibility using Machine Learning.” 2022 International Based on Machine Learning Approach.” 2021 International
Conference on Artificial Intelligence and Data Engineering Conference on Intelligent Technologies (CONIT).
(AIDE).
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
196 Algorithms in Advanced Artificial Intelligence
Abstract: An emerging paradigm, edge computing reduces network latency and expenses by putting networks and devices
near clients, enabling faster, more efficient, and real-time data processing. In order to analyse and respond to data generated
by devices in real-time, edge computing is crucial in the IoT. The increasing integration of IoT technology will provide faster
insights. Internet of Things (IoT) devices can assist hospitals in managing assets, including big datasets, and keeping tabs on
patients’ vital signs like heart rate and blood pressure. They can also monitor medication adherence. By combining AI with
edge computing, we hope to enhance several areas of healthcare, including patient scheduling and diagnosis, treatment, and
satisfaction; remote monitoring of chronic diseases; care; doctor response time; and staff compliance. In order to forecast
diseases utilising the RNN encoder and decoder architecture for sequence-to-sequence prediction, this article employs a
regression analysis method in healthcare data. Finding the optimal fit for a set of points in healthcare data is the goal of the
experimental investigation, which employs an RNN encoder and decoder architecture for disease prediction.
Keywords: Artificial intelligence, Edge computing, Regression analysis, RNN encoder and decoder
1
kvn.valli27@gmail.com, 2sujihelen@gamil.com
DOI: 10.1201/9781003529231-32
AI with Edge Computing-Driven Development in Healthcare Analysis 197
data (PDHD). Nonetheless, as of Gupta PM, 2023 [6], worries leading to improved healthcare efficiency. The volume of
around privacy and data security persist. When used in offline data that needs to be sent to the cloud is one of the biggest
mode, these devices can assess and track patients without an problems with s-health systems. Edge-based processing, such
internet connection. as compression and event recognition, can fix this. Based on
Clinical decision support (CDS) is increasingly important in Radanliev’s (2023) work on designing cyber risk assessments
today’s healthcare systems, as is the use of medical devices for AI and Industry 4.0 healthcare systems, this paper posits
such as tablets, wearables, health monitors, and imaging that digital health systems can be cyber-risk forecasted using
systems driven by artificial intelligence (AI) to improve predictive algorithms. According to Bansal (2024), state-of
patient care. Reference: Jia Z., 2023 [7]. Health monitors the-art health systems built on the Internet of Things (IoT)
and wearables can detect possible problems in X-rays and have the potential to enable healthcare providers to deliver
prioritise their assessment by radiologists or doctors; they timely, accurate information on the correct patients. In order
can also notify medical workers of problems; they can assist to establish a cost-efficient system, the OESP (Optimal Edge
with remote treatment; and they can give real-time updates Server Placement) algorithm effectively chooses the optimum
on patients’ vitals. According to Lin H. (2020) [8], new sites to put edge servers. By Jasim 2023, the algorithm had
healthcare technologies are changing workflows, cutting improved by more than 80% [18]. Wu, 2023 [19], argues that
costs, and increasing patient care. On the other hand, they innovative health technology has the potential to improve
produce massive volumes of data, which calls for choices health policy communication while also giving stakeholders
between managing data locally and in the cloud. Wu F. and lawmakers new perspectives. As seen with the newest
(2021) notes that edge computing can supplement cloud chips from smartphone makers, bringing computing closer to
computing by bringing analytics, storage, and processing the user in wearable sensors entails constructing edge devices
closer to the data sources. As a result, health systems are or servers that can execute trillions of operations per second
better able to handle the yearly 50 petabytes of data that while requiring remarkably little power (Gusev, 2022 [20]).
is generated. When it comes to storing, analysing, and Emam, 2019 [21] presents a multi-objective optimisation
processing data, health systems and providers are turning to framework that allows an edge node to choose the best radio
cloud computing. By Saeidimesineh R., 2023 [10], they are access technology (RAT) and dynamically alter compression
creating a new approach to data management that takes into parameters in order to balance latency, distortion, energy
account requirements, expenses, and advantages. It could be consumption, and performance. It is now feasible to reduce
helpful to provide just summary totals at set intervals and delays with a different approach. To support a proof-of-concept
limit the amount of data sent from patients’ wearables to application towards delivering an e-healthcare scenario
the cloud. Sreelakshmi S., 2023 [11] states that while cloud by Ray, 2021 [22], we amalgamate the Internet of Things
storage is ideal for larger operational data, on-premise storage (IoT) with edge computing in this work. Despite its benefits,
is necessary to comply with federal privacy regulations edge computing isn’t without its problems. For example,
pertaining to health information. By utilising software according to Hartmann (2022) [23], in order to achieve the
and hardware technologies that take advantage of edge same level of performance as cloud-based solutions while
computing, artificial intelligence, and cloud connectivity, the using less computational complexity, advanced privacy and
technology enhances the collection, analysis, and synthesis of data reduction techniques are necessary. Engaged, energised,
health data. The technology portfolio enables virtualization, and empowered providers with the necessary technology
enhances security, and reduces IT loads. Improved healthcare and management systems can only produce exceptional
CDS, quicker diagnostics, and patient monitoring are all patient outcomes (Pronovost, 2023 [24]). The reboot process,
possible because of analytics and AI that can move from the described by Ghani (2023) [25], could improve healthcare
edge to the cloud (Dash S., 2019). According to Jimma BL service provision as trusted next-generation health units,
(2023) [13], artificial intelligence is being used to develop thanks to the ever-evolving trusted technology in the health
cutting-edge healthcare solutions that consolidate various sector.
devices, apps, and services into one platform, making use of The seven parts of this paper’s reply are as follows: general
the resources already available in the cloud and data centres. introduction to Section 1. In Section 2, we delve into
Recent work by Rahman A. 2023 [14] shows that providers the topic of AI utilising edge computing on health data.
can increase their clinical value by using edge computing and Chapter 3 delves into the health data process through the
analytics. For instance, according to Rehman MU, 2023 [15], use of machine learning architecture. Section 4 delves into
deep learning (DL) methods hold promise in biological image the architecture of RNN encoders and decoders. Section 5
analysis for the early detection of acute diseases through the discusses the outcomes of the experiments. Section 6 includes
management of massive amounts of pertinent data, ultimately the conclusion. Section 7 contains the references.
198 Algorithms in Advanced Artificial Intelligence
Second, developers and researchers use training health data successfully models generalise to new patients’ health data,
to develop and train machine learning models for specific they are tested on held-out test data.
tasks. The models third use case is health data prediction, Collecting Data: Relying on machine learning models for
which involves taking a learned model and applying it to new health data training makes all the difference. This component
data. You can use a model that has been trained to categorise ensures the collection and retention of high-quality patient
photos or text to guess what an object is in a new picture or data to train and evaluate models.
piece of text. The feature engineering pipeline is responsible
for getting the health data ready to train the model. It entails Generating Data: Gathering real-world medical data for
cleaning up data so that models can make sense of it and use training models might be challenging or costly in some
it for prediction purposes. situations. In this part, we model the data we get from virtual
patients after real-life hospital records.
Interaction: This part specifies the ways in which models
communicate with other systems and people. One use case
for a model in a web app is to take user input and use it to 3. RNN Encoder and Decoder
make predictions. Architecture
Assessment: This part finds ways to enhance machine learning As a kind of memory, recurrent neural networks (RNNs) take
models by evaluating their performance. To determine how text sequences as inputs or outputs. The encoder-decoder
4. Experimental Result
The encoder-decoder approach for recurrent neural networks
is one well-liked neural machine translation method that can
beat more well-established statistical machine translation Fig. 32.4 Edge computing in healthcare data accuracy
methods. The experiment explores the use of Edge
computing in healthcare data accuracy is 98%, specifically The goal of simple linear regression is to identify the values
RNN encoding and decoding, to improve patient outcomes of w0 and w1 that will generate the most accurate y value
and potentially save lives. when given an x. This equation, also known as the model, can
also be evaluated in machine learning terms.
n
 (x1 x )(y1 y)
i=1
w1 = n
(1)
 (x1 x )2
i=1
w0 = y w1 x
Table 32.1 Blood pressure of pregnancy women of Age > 45 employed an RNN encoder and decoder architecture with the
SNo Age Blood Pressure (B.P)
goal of locating the line that best fits a given set of points. If
you use regression to find the line, you can be sure that equal
1 56 160
numbers will fall on either side of it.
2 46 161
3 72 160
References
4 45 161
5 63 160
1. Xu H, Huang W, Zhou Y, Yang D, Li M, Han Z: Edge
computing resource allocation for unmanned aerial vehicle
6 47 161 assisted mobile network with blockchain applications. IEEE
7 55 162 Transactions on Wireless Communications, 20(5): 3107-21,
8 49 160 2021.
2. Chen S, Wen H, Wu J, Lei W, Hou W, Liu W, Xu A, Jiang Y.
9 48 161
Internet of things based smart grids supported by intelligent
10 50 162 edge computing. IEEE access. 2019 Jun 3; 7: 74089-102.
11 68 161 3. Li J, Cai J, Khan F, Rehman AU, Balasubramaniam V, Sun J,
Venu P. A secured framework for sdn-based edge computing
12 60 161
in IOT-enabled healthcare system. IEEE Access. 2020 Jul 23;
8: 135479-90.
Hence, this calculation can be extended to more number of
4. Ali O, Abdelbaki W, Shrestha A, Elbasi E, Alryalat MA,
parameters like diabetes, thyroid and other medical disorders Dwivedi YK. A systematic literature review of artificial
during pregnancy and predict the results very accurately intelligence in the healthcare sector: Benefits, challenges,
which is known as multiple regression. This is very helpful methodologies, and functionalities, Journal of Innovation &
for the patients to get correct treatment in the correct time for Knowledge. 2023 Jan 1; 8(1): 100333.
saving the lives of both mother and child. 5. Sun L, Jiang X, Ren H, Guo Y. Edge-cloud computing and
artificial intelligence in internet of medical things: architecture,
technology and application. IEEE Access. 2020 May 26; 8:
101079-92.
6. Gupta PM. Integration Of Edge And Fog Computing In IoT-
Based Healthcare Applications-A Review, Journal of Positive
School Psychology, 2023 Jan 15: 1940-57.
7. Jia Z, Chen J, Xu X, Kheir J, Hu J, Xiao H, Peng S, Hu
XS, Chen D, Shi Y. The importance of resource awareness
in artificial intelligence for healthcare. Natural Machine
Intelligence, 2023 Jun 12: 1-2.
8. Lin H, Garg S, Hu J, Wang X, Piran MJ, Hossain MS. Privacy-
enhanced data fusion for COVID-19 applications in intelligent
Internet of medical Things. IEEE Internet of Things Journal.
2020 Oct 22; 8(21): 15683-93.
9. Wu F, Qiu C, Wu T, Yuce MR. Edge-based hybrid system
implementation for long-range safety and healthcare IoT
applications. IEEE Internet of Things Journal. 2021 Jan 11;
Fig. 32.5 Linear regression line for Age VS Blood Pressure 8(12): 9970-80.
10. Saeidimesineh R, Adibi P, Karshenas H, Darvishy A.
Parallel encoder-decoder framework for image captioning.
5. Conclusion Knowledge-Based Systems. 2023 Oct 11: 111056.
11. Sreelakshmi S, Malu G, Sherly E, Mathew R. M-Net: An
An emerging concept in the Internet of Things (IoT), encoder-decoder architecture for medical image analysis
edge computing allows for real-time data processing to be using ensemble learning. Results in Engineering. 2023 Mar
faster, more efficient, and cheaper by positioning networks 1; 17: 100927.
close to consumers. Improved patient diagnosis, treatment, 12. Dash S, Shakyawar SK, Sharma M, Kaushik S. Big data
contentment, remote monitoring, and staff compliance in healthcare: management, analysis and future prospects.
are some of the benefits of this technology, which also Journal of big data. 2019 Dec; 6(1): 1-25.
13. Jimma BL. Artificial intelligence in healthcare: A bibliometric
helps monitor patients’ vital signs and tracks prescription
analysis. Telematics and Informatics Reports. 2023 Jan 9:
adherence. In order to forecast healthcare data, the research 100041.
202 Algorithms in Advanced Artificial Intelligence
14. Rahman A, Hossain MS, Muhammad G, Kundu D, Debnath Intelligence and Machine Learning for EDGE Computing.
T, Rahman M, Khan MS, Tiwari P, Band SS. Federated Academic Press, 2022. 469-477.
learning-based AI approaches in smart healthcare: concepts, 21. Emam, Ahmed, et al. “Edgehealth: An energy-efficient edge-
taxonomies, challenges and open issues. Cluster computing. based remote mhealth monitoring system.” 2019 IEEE wireless
2023 Aug; 26(4): 2271-311. communications and networking conference (WCNC). IEEE,
15. Rehman MU, Panday A. Review on Artificial Intelligence in 2019.
Healthcare. 2023 Aug; 26(4): 2371-312. 22. Ray, Partha Pratim, Dinesh Dash, and Debashis De. “Intelligent
16. Radanliev, Petar, and David De Roure. “Advancing the internet of things enabled edge system for smart healthcare.”
cybersecurity of the healthcare system with self -optimising National Academy Science Letters 44 (2021): 325-330.
and self-adaptative artificial intelligence (part2). “Health and 23. Hartmann, Morghan, Umair Sajid Hashmi, and Ali Imran.
Technology 12.5 (2022): 923-929. “Edge computing in smart health care systems: Review,
17. Bansal, Urvashi. “Power of IoT in Smart Healthcare Systems.” challenges, and research directions.” Transactions on
Applications of Optimization and Machine Learning in Image Emerging Telecommunications (2022): e3710.
Processing and IoT. Chapman and Hall/CRC, 2024. 79-91. 24. Pronovost, Peter J., and Robert K. Lord. “Could Modernizing
18. Jasim, Ahmed M., and Hamed A Raweshidy: Optimal Health Care Technology Be a Cure for Provider Burnout?.”
intelligent edge servers placement in the healthcare field, IET American Journal of Medical Quality 38.5 (2023): 264-266.
Networks, 2023. 25. Ghani, Norjihan Abdul, et al. “Methodical Evaluation of
19. Wu, Qi, Beian Chen, and Jianping Zhu. “Insights from Healthcare Intelligence for Human Life Disease Detection.”
COVID-19: Reflecting on the Promotion of Long-Term Health Malaysian Journal of Computer Science 36.3 (2023): 208
Policies in China.” International Journal of Environmental 222.
Research and Public Health 20.4 (2023): 2889.
Note: All the figures and table in this chapter were designed by the
20. Gusev, Marjan. “AI cardiologist at the edge: A use case
author.
of a dew computing heart monitoring solution.” Artificial
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Unusual inflammatory skin changes that might modify the skin’s color, texture, or appearance are known as skin
rashes. They could show up in one spot on the body or throughout it. Image processing is the process of transforming pictures
into new forms such as pictures, movies, texts, or other parts of the original images. Most image processing methods produce a
large quantity of data as their final output, which is known as “Big-data”. The paper encourages the use of a pure Transformer
i.e. ViT applied directly to picture patches for image classification applications. ViT outperforms various computer vision
applications in recent standards, demonstrating its competitive performance, including image classification, object recognition,
and semantic image segmentation. This study aims to develop a CNN for image classification using larger dataset Image
Patches and explore the relationship between CNN and ViT in picture classification. In the experimental results, the Vision
Transformer (ViT), which pre-trained on a vast amount of data, surpasses state-of-the-art convolutional neural network models
in a number of evaluations while using less CPU resources during training.
Keywords: Big data, Convolutional neural network, Semantic image segmentation, Vision transformers etc.
1
ksnr539@gmail.com, 2sekharonemay@gmail.com, 3laxmipathi4u@gmail.com, 4krishnasatyavarma@gmail.com, 5raju.pericherla74@gmail.com,
6
ksumisri@gmail.com
DOI: 10.1201/9781003529231-33
204 Algorithms in Advanced Artificial Intelligence
Models for big data analysis have not recently been developed. channels of the input images are reflected in the input layer’s
Despite the fact that image processing is an established dimensions.
technology, creating image-processing models based on big Convolutional Layers: Convolutional layers extract features
data analysis can present a number of technical challenges. from input images using kernels, filters, and recognize shapes,
Image processing technologies require visualization edges, textures, and other visual components by convolving
analysis, semantic expression, and large sample storage, them with filters.
complex algorithms for feature extraction, recognition, and
prediction, along with time and memory requirements. In Pooling Layer: The physical dimensions of the feature maps
addition to these, a significant issue will be the slow rate of produced by convolutional layers are reduced by the addition
model identification. With the advancement of information of pooling layers. They save the most crucial details while
technology, big data applications in image processing removing the rest using down sampling techniques (such max
are expanding. Big data analysis-based image processing pooling). This helps in both achieving translation invariance
models offer broad application possibilities across all image- and reducing computer complexity.
processing disciplines by analyzing operating principles, Fully Connected Layers: One or more fully connected
technologies, and benefits. layers are connected to the output of the last pooling layer
after it has been flattened. These layers classify the retrieved
2. Related Methods characteristics in the same way that conventional neural
network layers do. Fully linked layers notice intricate
The Big Picture: Skin Disease on a Large-Scale CNN correlations between features and produce forecasts or class
explores the use of CNN and picture classification in medical probabilities.
imaging and healthcare. The output layer, the final layer of a CNN, provides
classification probabilities for each class, indicating the
2.1 Convolutional Neural Network
likelihood of the input image belonging to a specific class by
The process of classifying incoming photos entails giving Połap, D, 2018[8].
those labels or categories. In order to predict the class of
unseen images, a model is trained on labeled image data in
a supervised learning task by Jainesh Rathod, 2018[4]. Since 3. Proposed Work
CNN can recognize objects in photos accurately by learning Convolutional neural networks (CNN) have significantly
hierarchical elements like edges, textures, and forms, they are facilitated recent advancements in deep learning by Shanthi,
frequently employed for image classification. Because they 2020[11]. Data scientists are particularly interested in
can automatically extract useful spatial characteristics from computer vision, and CNNs have broken the mold to become
photos, CNNs excel at this task by J Sudha, 2017[5]. The the most advanced computer vision method by Parvathaneni
procedure’ several layers are listed below: Naga Srinivasu, 2021[9]. The ViT is a transformer utilized
Layer of Input: The CNN has input layer receives the raw for vision-related tasks like image categorization, offering
picture data as input. Typically, matrices of pixel values are superior accuracy over convolution and enabling parallel input
used to represent the images. The height, breadth, and color processing over Seq2Seq, Vec2Seq, and Seq2Vec tasks. The
study investigates the use of natural language processing’s dropout, weight decay, and stochastic depth can help avoid
self-attention mechanism in Big Image classification using overfitting by reducing parameter co-adaptation, penalizing
a new neural network by Simonyan, 2015[12] highlighting weights, and strengthening network resistance to noise.
potential limitations and errors. This work’s primary Vision transformers take a lot of memory and are
contribution is as follows. computationally demanding. Smaller patch sizes, lower
• Build a CNN with picture patches from a bigger dataset resolution images, and fewer layers or attention heads can
in order to categorise images. all be used as techniques to speed up training and inference.
• To understand how CNN and ViT handle the picture These techniques cut down on calculation costs, information
classification task. processing, and sequence length.
The final step in evaluating vision transformer performance
3.1 Vision Transformer (ViT) involves assessing precision, recall, F1-score, and accuracy.
The ViT is a transformer used for vision-related tasks like Precision measures forecast accuracy by comparing genuine
image categorization, providing greater accuracy than positives to anticipated positives; recall measures accuracy
convolution. It takes over Seq2Seq, Vec2Seq, and Seq2Vec by comparing positives to all positives, and F1-score balances
tasks, enabling parallel input processing by Ki V,2008[6]. these measures.
The study explores the use of natural language processing’s
3.2 Proposed Methodology
self-attention mechanism in the classification of images using
a new neural network, highlighting its potential limitations Transformers have been the conventional model in NLP
and potential errors by Satishkumar Moparthi, 2021[10]. due to their efficiency and adaptability in computation.
Computer vision is still dominated by convolutional neural
Picture categorization in computer vision is traditionally
network (CNN) architectures; however, some researchers
done using convolutional neural networks (CNNs) by Srujan
have tried combining CNNs with self-attention. The authors
S A, 2022[13]. However, CNNs have drawbacks like spatial
experimented merely applying a standard Transformer on
distortion and data requirements. Vision transformers, which
pictures and found that the models showed modest accuracy
view images as patches and use self-attention to recognize
compared to ResNet-like architectures when trained on mid-
interdependencies, offer a new approach that can better apply
sized datasets by Z. Ma,2016[15]. However, when trained
information from less input by Md. Sazzadul Islam Prottasha,
on larger datasets, the Vision Transformer (ViT) generated
2023[7].
excellent results and came very near to or outperformed the
Vision transformers require extensive data to function state of the art on a number of picture recognition criteria.
effectively, making them sensitive to patch sequence and
The model transforms 2D medical images into flattened
position. To improve accuracy, they need a wide range of
patches, which are then mapped to a latent vector using a
images by Syed Inthiyaz, 2023[14]. Pre-trained models
linear projection. The image representation is pre-trained or
trained on large amounts of data can reduce training time and
fine-tuned using a classification head and the Transformer
enhance their performance.
encoder maintains positional information. CNNs excel in
Vision transformers can overfit, especially when the target image processing, classification, object identification, and
dataset is small or dissimilar. Regularization methods like segmentation due to their ability to extract hierarchical
Fig. 33.3 The ViT model divides skin disease medical images into patches, linearly embeds them, adds position embeddings,
and uses a transformer encoder for classification
feature information and learn from vast picture data. Vision 4.1 CNN Approach for Skin Disease Medical
Transformers excel in global interdependence and contextual Image
awareness scenarios, while CNNs handle large datasets
Three layers of 2D convolutions with a kernel size of 3, stride
efficiently. They require larger training data, benefiting
of 2, and a maximum pooling layer of 2 make up the CNN
real-time and resource-constrained applications due to their
model for this picture classifier. There are two completely
increased computational efficiency.
linked layers with 10 nodes each after the convolution layers.
Algorithm: Numerous vision transformer models have been Here is an example of this structure in code: Ten training
proposed. The overall framework of the vision transformer epochs of the training were carried out on a Tesla T4 (g4dn
architecture is composed of the subsequent actions: xlarge) GPU system. The outcomes of training loops for each
1. Divide a medical image into patches of a certain size. epoch are listed below.
2. Reduce the skin patches for medical images.
4.2 Transforming Vision in Medical Images of
3. These flattening skin medical imaging patches are
Skin Diseases
transformed into lower-dimensional linear embeddings.
4. Positional embeddings should be used. The proportions of the Vision Transformer architecture are
adaptable and may be changed to meet particular needs.
5. Give the sequence to a modern transformer encoder as
This architecture is nevertheless substantial for a skin image
an input.
collection of this scale.
6. Before it is fully supervised on a large dataset, the ViT
model is first trained using image labels. Each parameter in the vision transformer plays a key role
and is described here:
7. The downstream dataset has to be modified for image
classification. • image_size=224: The required width and height of the
input photos to the model are specified by this option.
The photos in this situation should be 224x224 pixels
4. Experimental Result in size.
The results of the investigation relate to testing for skin • patch_size=32: This option specifies the dimensions
diseases in the malignant and benign groups. The strategy (width and height) of each patch, which are used to break
that will be used is as follows the photos into smaller portions. Each patch is 32x32
pixels in this instance.
Big Image: Large-Scale Skin Disease Image Classification in Medical Imaging and Healthcare Using CNN and Transformers 207
• num_classes=2: The number of classes used in the dependence on particular tokens being placed too heavily
classification operation is indicated by this parameter. during training.
The model in this illustration is built to divide inputs into The Tesla T4 (g4dn-xlarge) GPU computer was used to
two categories, benign and malignant. train the vision transformer for the classification job over
• dim=128:It describes the model’s embedding vectors’ the course of 20 training epochs. Because the training loss’s
dimensionality. Each picture patch’s representation is convergence was gradual, the training was carried out across
captured by the embeddings. 20 epochs rather than the 10 epochs utilized for CNN. The
• depth=12:The Vision Transformer model’s (encoder outcomes of training loops for each epoch are listed below.
model’s) depth or number of layers is defined by this In 10 iterations, the CNN technique correctly predicted skin
parameter. A deeper level enables the extraction of more illness 75% of the time, but the vision transformer model
intricate features. correctly predicted skin disease 69% of the time and required
• heads=8:The number of attention heads in the model’s much more time to train.
self-attention mechanism is represented by this
parameter.
5. Conclusion
• mlp_dim=1024:It details the model’s hidden Multi-
Layer Perceptron (MLP) layers’ dimensionality. After CNN and Vision Transformer models differ in size, memory
self-attention, the MLP is in charge of changing the requirements, accuracy, and performance. CNN models are
token representations. compact and efficient, suitable for limited resources and
• Droput=0.1: The dropout rate, a regularization method image processing tasks. Vision Transformers collect global
used to avoid overfitting, is controlled by this parameter. dependencies and contextual information in skinned medical
During training, a certain percentage of input units are pictures, but require more RAM and larger model sizes.
set to 0. Decisions between models depend on task details, resource
• emb_droput=0.1:It describes the dropout rate as it relates availability, dataset scope, and trade-off between complexity,
directly to token embeddings. This dropout prevents accuracy, and performance. Additional improvements are
expected in computer vision.
208 Algorithms in Advanced Artificial Intelligence
7. Md. Sazzadul Islam Prottasha, Sanjan Mahjabin Farin, Md. 12. Simonyan, K., Zisserman, A.: Very-deep convolutional
Bulbul Ahmed, Md. Zihadur Rahman, Kabir Hossain & M. networks for large-scale image recognition, CoRR.
Shamim Kaiser: Deep Learning Based Skin Disease abs/1409.1556, 2015.
Detection Using Convolutional Neural Networks, https://link. 13. Srujan S A, Chirag M Shetty , Mohammed Adil ,Sarang P K
springer.com/bookseries/7818, Lecture Notes in Electrical , Roopitha C H:Skin Disease Detection using Convolutional
Engineering, pp 551–564,2023. Neural Network,International Research Journal of Engineering
8. Połap, D., Winnicka, A., Serwata, K., Kȩsik, K., Woźniak, M.: and Technology (IRJET) e-ISSN: 2395-0056, Volume: 09
An intelligent system for monitoring skin diseases, Sensors Issue: 07- July-2022 www.irjet.net p-ISSN: 2395-0072.
18, 2552, 2018. 14. Syed Inthiyaz, Baraa Riyadh Altahan, Sk Hasane Ahammad,
9. Parvathaneni Naga Srinivasu, Jalluri Gnana SivaSai, V Rajesh, Ruth Ramya Kalangi, Lassaad K. Smirani,Md.
Muhammad Fazal Ijaz, Akash Kumar Bhoi, Wonjoon Amzad Hossain, Ahmed Nabih Zaki Rashed: Skin disease
Kim, James Jin Kang:Classification of Skin Disease Using detection using deep learning, https://doi.org/10.1016/j.
Deep Learning Neural Networks with MobileNet V2 and advengsoft.2022.103361, Advances in Engineering Software,
LSTM,Sensors (Basel), 2021 Apr; 21(8): 2852, Published Volume 175,103361,January 2023.
online 2021 Apr 18, doi: 10.3390/s21082852. 15. Z. Ma and J. M. R. S. Tavares: A Novel Approach to Segment
10. Satishkumar Moparthi: An Image is Worth 16×16 Words: Skin Lesions in Dermoscopic Images Based on a Deformable
Transformers for Image Recognition at Scale (Vision Model, IEEE Journal of Biomedical and Health Informatics,
Transformers), Analytics Vidhya, Published On March 10, vol. 20, no. 2, pp. 615–623, March 2016.
2021 and Last Modified On March 11, 2021.
Note: All the figures in this chapter were designed by the author.
11. Shanthi, T., Sabeenian, R.S., Anand, R.: Automatic
diagnosis of skin diseases using convolution neural network,
Microprocess, Microsystems, 76, 103074, 2020.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
210 Algorithms in Advanced Artificial Intelligence
S. Suryanarayanaraju1
Research Scholar, Department of Computer Science & Engineering,
GIET University, Odisha, Gunupur
M. Chandra Naik2
Professor, Department of Computer Science & Engineering,
GIET University, Gunupur
R. N. V Jagan Mohan3
Associate Professor, Department of Computer Science & Engineering,
SRKR Engineering College (A), Chinaamiram
Abstract: The enormous of Electronic Health Records is a norm to represent health data in the digital world. As the volume
records continues to grow, managing tasks based on factors like workload, computational capacity, and historical performance
will indeed become a critical challenge. High processing loads can strain resources and lead to delays or inefficiencies in
accessing and analyzing crucial health data. This research proposes to address this issue using Artificial Intelligence (AI) for
load distribution in federated networks which enhances system efficiency and responsiveness.
Keywords: Artificial intelligence, Electronic health records, Federated network, Load balance etc.
DOI: 10.1201/9781003529231-34
AI Driven Load Distribution for Federated Network on Electronic Health Records 211
Federated learning is crafted to facilitate the training of a different features across distinct nodes. The accuracy loss
unified model by leveraging the local datasets across all nodes. attributed to non-IID data can be mitigated using advanced
This optimization is carried out with respect to the objective data normalization methods, surpassing the limitations of
function f(x1, x2, ..., xk).The aim is to attain consensus on xi, conventional batch normalization.
signifying that x1, x2, ..., xk converge to a shared value x by
the conclusion of the training process. 1.1 Electronic Health Records
Federated learning involves centralized and decentralized The medical data of a patient is preserved digitally by
methods. Centralized federated learning uses a central server healthcare practitioners and is known as an electronic health
to arrange algorithms and coordinate nodes, potentially record, or EHR. They automate accessibility, streamline
becoming a bottleneck. Decentralized federated learning workflow and quality management, support evidence-based
allows nodes to coordinate themselves, preventing single decision support, and outcomes reporting. EHRs offer real-
point failures. Heterogeneous federated learning focuses on time, patient-centric information and encompass a wealth of
accommodating diverse clients with varying computation and information, including diagnoses, medical history, treatment
communication capabilities. The heterogeneous federated plans, medications, radiology images, allergies, and test
learning framework is designed to train local models that results. They are designed to be shared securely with other
differ in computational capabilities and handle dynamic healthcare providers across various organizations for a more
computation and non-independent and identically distributed holistic and collaborative healthcare ecosystem.
(non-IID) data complexities. Despite these differences, the
framework aims to produce a unified and accurate global 1.2 Federated Learning Architecture
inference model through collaborative training across Federated Learning Architecture: Federated Learning design
heterogeneous devices. involves a system that manages user uploads, feeds, video
Similar to a machine learning process, federated learning processing, metadata storage, caching, and search through
involves training local models, aggregating local updates into various components and services. The Federated Learning
a single global update, and transmitting nodes from global system design utilizes distributed databases for managing
model states. It can be centralized or decentralized, with a user data, utilizing sharding and replication techniques for
central server facilitating the aggregation step, as depicted in global data availability and consistency as shown in Figure
Fig. 34.1. The process includes initialization, client selection, 2. Without sharing raw data, federated learning can enable
configuration, reporting, and termination. Asynchronous many enterprises to collaborate to jointly develop a machine
techniques, such as split learning, have been introduced for learning model. This approach addresses several critical
training and inference. challenges, including privacy, security, access rights, and the
heterogeneity of data access across different entities.
• Client Interaction: Federated Learning uses randomized
selection to select clients for global model parameters,
but heterogeneous data distribution causes client drift
and performance degradation.
• Load Balancer: Load balancing is a process where
incoming client requests are processed by multiple
servers to prevent overloading. It can be implemented as
hardware or software load balancers, either installed on-
premises or managed. The load balancer directs requests
to accessible servers, ensuring effective communication
Fig. 34.1 Federated iterative learning process
and task fulfillment. It can also help with network
Federated learning setups often have unbalanced local data caching by routing traffic to cache servers for temporary
samples and specific probability distributions of training storage of user requests.
examples. Non-IID data can cause significant variations in • API Gateways: Small-scale platform users can access
training performance. Among the main types of non-IID data planes as a service, enabling them to determine
data are correlate change, prior probability shift, concept gateway configuration from API specifications without
drift, concept shift, and imbalanced data. Examples include worrying about deployment, and route requests to
natural language processing datasets with different stroke microservices.
widths, regional or demographically partitioned datasets, • Write Operations: The global model parameters are
and concepts that may share the same label correspond to updated by local clients, while the global server collects
212 Algorithms in Advanced Artificial Intelligence
them. API gateways facilitate write operations to the App • Caching Mechanism: Redis or Memcache data caching
Server. reduces latency and database load by calculating cache
• Feed Generation Service: Federated averaging is a behavior based on restrictive settings, supporting various
machine learning approach that prioritizes data privacy cache implementations like Cache-Control headers.
and security by spreading data across multiple servers or • Search Service (Elastic search): Elastic search efficiently
devices without sharing raw data. uses inverted indices for efficient search, enabling rapid
• Read Operations: The federated learning paradigm data analysis and rapid user and content searches with
involves learning nodes training local models, sending near-real-time performance.
parameters to the server, performing weighted averaging, • Blob Storage: Federated Learning is a data-driven
and sending global parameters back to the learning framework that trains a single machine learning models
nodes. on multiple datasets, promoting compliance with
• Metadata Database: Federated learning (Trung Kien regulations and innovation, and storing user-uploaded
Dang et al., 2022), is a decentralized approach to training media.
machine learning models, enhancing data privacy by • CDN (Content Delivery Network): Content caching
using edge devices for local training and user profile and data computing at the wireless network edge is a
storage. promising method for reducing backhaul traffic load by
caching and serving static content with low latency.
• Video Processing: Video processing services handle tasks (FCFS), local search algorithm and Stochastic Hill Climbing
like transcoding, filtering, and thumbnail generation, (SHC). Future work will be concentrated on exploring the
using blob storage. Federated learning trains algorithms variations in crossover and strategies for selection to attain
through independent sessions using each dataset. better efficiency with fine-tuned results.
• Notifications: The system notifies users of likes, (Shikha Garg et al., 2015) have improved the use of virtual
comments, and interactions. machines in cloud computing, a problem that synchronous
restricted load balancing attempts to solve. One of the most
2. Related Work difficult problems in cloud computing is load balancing.
Meeting the substantial demands of users necessitates a
Load balancing distributes network traffic across a resource distributed solution, as individually assigning one or more idle
pool for efficient processing of simultaneous requests from services is often impractical or cost-prohibitive. Allocating
millions of users. It serves as an intermediary device between servers to clients on an individual basis presents challenges.
the user and the server group to ensure equitable utilization Load balancing algorithms can be enhanced through
of all resource servers. Previous studies by different authors alternative approaches, such as incorporating soft computing,
have explored these concepts, as outlined below. to efficiently utilize virtual machines and simultaneously
(Huankai Chen et al., 2013) have focused on the Load minimize response times.
Balance Improved Min-Min (LBIMM) scheduling approach In cloud computing, load balancing is explored by (V
is being utilized to improve load balancing in cloud RaviTeja Kanakala et al., 2015). It searches for the different
computing. In order to reduce make span and enhance algorithms that will help in distributing the load among the
resource usage, this method has expanded on the features nodes efficiently. Additionally, it delves into the parameters
of the Min-Min algorithm. In the proposed PA-LBIMM considered when determining the optimal algorithm for
(Promised guarantees, User-priority-based Load Balanced load balancing. The existing challenge lies in the limited
Min-Min), the author not only incorporates the principles of performance of current load-balancing algorithms across
load balancing but also introduces additional features such as all essential areas. The author suggests refining algorithms,
promised guarantees and user-priority. These augmentations including Exponential Smooth Forecast and Load Balance
are designed to ensure a more comprehensive satisfaction of Max Min, to improve cloud performance and align with load
user demands. Notably, the author’s focus does not extend to balancing demands.
considering task deadlines and the intricacies associated with
high heterogeneity in interconnection networks. A deep learning model for load balancing was developed
by (Zhang et al. 2021) to handle the skewness in data by
(Chien et al., 2016) worked on the load balancing algorithm replacing the hash function with deep learning mechanisms.
under consideration relies on estimating the completion However, cascading overflow can still be addressed for
time of services. The aim of load balancing is to enhance further enhancement. This occurs when one server reaches
throughput, fine-tune resource utilization, avert the full capacity causing subsequent servers to fill up more
overloading of individual resources and minimize response rapidly, resulting in an overflow effect.
time. While numerous load-balancing algorithms have
been introduced, their performance remains a subject of The deep learning model suggested by (Kaur et al., 2020) will
ongoing improvement. In response to this, the Authors have improve throughput, resource utilization, latency, cost and
introduced an algorithm that centers around estimating the response time in cloud computing. For workflow execution in
completion time of services. According to simulation results, cloud environments, they presented the Deep Learning-based
this algorithm demonstrates enhancements in both latency Dynamic VM Provisioning and Load Balancing (DLD-PLB)
and processing time. While load balancing contributes to the framework. The suggested model in the suggested framework
effective utilization of computational resources and improved is superior to the prior model in that it makes use of deep
efficiency, it also introduces challenges, particularly in the learning for load balancing.
realm of energy consumption. (Rathod et al., 2020) discussed about Artificial Intelligence
(Kousik et al., 2013) have suggested a strategy which based (AI) techniques are widely used in cloud computing due to
on the formulation of a Genetic Algorithm (GA) for load their numerous benefits. Proper load balancing mechanism is
balancing. This task is an optimization problem. A load crucial for user and service provider satisfaction. This proposal
balancer is utilized efficiently to adapt the strategy dynamically can give the detailed knowledge on how AI techniques can
when there is a change in environmental conditions and task improve load balancing in cloud environments. AI techniques
types. The results of the algorithm have supered the existing are integrated with techniques like hill climbing, fuzzy logic,
approaches like Round Robin (RR), First Come First Serve and honey bee mechanism will increase resource utilization
214 Algorithms in Advanced Artificial Intelligence
and service quality. It is important to decide when to combine hindering the development of clinical decision support
these techniques. tools and health-specific AI. To access various data sets and
(Wilson et al., 2019) discussed the challenges of efficient handle health problems, federated networks are utilized.
resource utilization in cloud computing networks. The They discussed regarding the utilization of federated
application performance will be degraded by overloaded networks in healthcare, their establishment, operation and
virtual machines and resource utilization inefficiency will implementation.
be created by underloaded VMs. On Software-Defined The health system relies on centralized agents which share
Networking (SDN) a mechanism called Dynamic Agent- the raw data in (Rahman et al., 2023). Combining this with
Based Load Balancing algorithm is applied, which uses AI and FL can reduce challenges and vulnerabilities. The
Back Propagation Artificial Neural Network (BPANN) that analysis in FL using AI regarding healthcare applications
migrates the virtual machines efficiently in data centres. will address problems like privacy, security, scalability,
Overall, the network efficiency will be improved and perform reliability and confidentiality. It will discuss about emerging
well on data migration. A comparison between the Heuristic trends like FL, Explainable Artificial Intelligence XAI, AI
algorithm (HA) and Multi-Path TCP is performed by using and e-healthcare. It gives suggestions for solving healthcare
the migration process results. This algorithm is suited for VM strategies using FL and AI. It specifies the research areas
migration and can utilize runtime configuration management extensively and potential prospects in future for managing the
as an extension for the algorithm. This is also useful for data healthcare systems that use FL and AI. The recent progress
offloading in mobile cloud computing. has created an interest in FL to integrate AI into the networks.
Various networks are deployed by using the platform called The complete study has analyzed the progress of security,
Software-defined networking (SDN) which is discussed in taxonomies, discussions, benefits of integration, open issues,
(Hazim et al. 2023). Load imbalance is an issue in SDN during and offers in future research guidance.
traffic distribution. To enhance the effectiveness of SDN, they
have developed various SDN load-balancing mechanisms. 3. Hyperparameter Approach of
The work mainly focuses on analyzing the architecture Federated Learning
of SDN. The summarizing metrics and categorization of
AI-based load-balancing methods are used in measuring Federated learning approaches, such as orchestrator-
the efficiency of the techniques. A detailed survey is given less distributed networks, aggregate local models, reduce
on various load-balancing mechanisms that utilize AI for transaction count, and potentially reduce training time and
improving the load distribution in SDN. computing costs. The choice of the node network topology
The intelligent load-balancing mechanisms and the need for allows for optimization of learning by controlling various
energy efficiency in healthcare are discussed by (Ibrahim parameters, including the machine learning model’s
et al., 2023). The load balancing model is based on energy- Hyperparameters. Key parameters include the federated
aware artificial intelligence, that utilizes big data analytics learning rounds is T. The process utilized a total of K nodes.
The local learning rate, represented by constant C, batch size
(BDA) and Chaotic Horse Ride Optimization (CHROA) for
B, and number of iterations for local training before pooling
IoT environments that are cloud-enabled. The CHORA will
N, significantly influences the effectiveness of educational
use the AI models for balancing the load and optimizing the
programs and strategies. Optimizing machine learning
energy resources. The CHORA is evaluated using various
parameters depends on application constraints like computing
metrics. Overall, the article highlighted the challenges and
power, memory, and bandwidth. Stochastic gradient descent
contributions to developing efficient solutions in IoT/IoE.
and limited node fractions can reduce computing cost and
Load-balancing algorithms are compared and summarized by prevent overfitting.
(Singh et al., 2023) for cloud computing in environments like
Algorithm Procedure: The index of the C clients is c. The
centralized, static and dynamic. Popular machine learning
local minibatch size is denoted by D, the local epochs by L,
models like Random Forest Classifier, Statistical Regression,
and the learning rate by η.
CNN, AI and LSTM-RNN are explored. Load balancing
improves system performance by reducing time, throughput, The server runs:
production time, and power savings. The deep learning Initializew_0;
models have been replaced by machine learning models that
for each encircling t=1, 2, …do
handle big data effectively and don’t affect standardization.
During load balancing the score is an important factor. m ← max (F. C, 1)
(Harry et al., 2021) discussed that access to health data is St ← (Randomesetofmclients)
problematic due to legislation. This results in siloed data for each client c ∈ St in parallel do
AI Driven Load Distribution for Federated Network on Electronic Health Records 215
wc(t+1) ← ClientUpdate(c, wt) Table 34.1 Data set for load distribution in federated network
w C
nc c
t +1 ¨ Â w
ReTX Count
Leap Count
n t+1
Delay (ms)
c=1
Respond
AppId
Node
Client Updates(c, w):
Time
Type
Seq.
Β ← split Ρc into batches of size D
1.46006 7 255 226 Last 0.066264 1 4
for each local epoch i from 1 to L do Delay
for bach d ∈ B do 1.46006 7 255 226 Full 344.076 2 4
Delay
w ← w – η∇l(w;b)
1.46506 7 255 227 Last 0.066264 1 4
Return w to Server. Delay
Through a technique known as secure aggregation, the server 1.46506 7 255 227 Full 344.076 2 4
can combine the encrypted models from different participants Delay
without gaining access to the raw updates. Instead, the server 1.47006 7 255 228 Last 0.066264 1 4
merely decodes the aggregated training results. As a result, Delay
the server will never see the training results for a particular 1.47006 7 255 228 Full 344.076 2 4
device. Federated Learning and Differential Privacy can be Delay
coupled for increased security. In addition to training, testing 1.47506 7 255 229 Last 0.075264 1 4
is a crucial difference between Federated and traditional Delay
machine learning. We should test machine-learning models 1.47506 7 255 229 Full 344.076 2 4
with data that most closely resembles the inputs the model Delay
would experience in use. However, since it lacks access to the 1.48006 7 255 268 Last 0.076116 1 4
training data, the server is unable to test the combined model Delay
after it has been updated with input from the clients. As a 1.49006 7 255 268 Full 0.076116 1 4
result, training and testing are done on consumers’ devices. Delay
Be mindful that distributed testing benefits from testing the
updated model on consumers’ devices, which is the most significant improvements in resource allocation and data
important place to test. privacy preservation. By leveraging collaborative model
training across decentralized nodes, FL offers an effective
4. Experimental Result solution for optimizing task allocation while respecting
the autonomy of individual entities. Our study has shown
The Electronic Health Record utilizes load balancers to that FL-based load distribution leads to enhanced network
optimize server traffic, speed up response times, and reduce performance, reduced latency, and increased robustness, even
network latency. They evenly distribute load among servers, in dynamic and heterogeneous environments. The success of
reroute client requests to servers closer to client locations, this approach opens new opportunities for a wide range of
and prevent overwork, enhancing application performance. applications across various healthcare industries, and it paves
Load balancing is a crucial aspect of parallel and distributed the way for further advancements in decentralized computing
computing, enhancing system efficiency, reliability, resource paradigms.
efficiency, and performance by ensuring balanced workloads.
The data set for the load distribution in federal network is
shown in Table 34.1. References
The required of regression plane is Y = b0 + b1 * x1 + b2 1. Chen.H, F. Wang, N. Helian, G. (2013). AkanmuUser-priority
* x2 + b3 * x3. To estimate the Respond time is depends guided Min-Min scheduling algorithm for load balancing in
on Delay, ReTX count and Leap Count. The respond time cloud computing, IEEE Natl. Conf. Parall. Computer Technol.,
incurred 612.656062, 1.77701828, -611.30892, 0 as shown (PARCOMPTECH), pp. 1–8.
2. Chien, N.K., Son, N.H., Loc, H.D. (2016). January, Load
in Fig. 34.3.
balancing algorithm based on estimating finish time of
services in cloud computing, In 2016 18th IEEE International
5. Conclusion and Future Work Conference on Advanced Communication Technology
(ICACT), pp. 228–233.
The implementation of Federated Learning for load
distribution in Federated Networks has demonstrated
216 Algorithms in Advanced Artificial Intelligence
3. Dasgupta.K, B. Mandal, P. Dutta, J.K. Mandal, S. Dam. (2013). 10. Ahmed Hazim Alhilali and Ahmadreza Montazerolghaem,
A genetic algorithm (GA) based load balancing strategy for (May 2023), Artificial Intelligence based Load balancing in
cloud computing, Procedia Technol., 10, pp. 340–347. SDN: A ComprehensiveSurvey, https://doi.org/10.1016/j.
4. Garg, S., Dwivedi, R.K., Chauhan, H. (September 2015). iot.2023.100814.
Efficient utilization of virtual machines in cloud computing 11. Ibrahim Aqeel, Ibrahim Mohsen Khormi, Surbhi Bhatia
using Synchronized Throttled Load Balancing, 1st IEEE Khan, Mohammed Shuaib, Ahlam Almusharraf, Shadab
International Conference on Next Generation Computing Alam, and Nora A. Alkhaldi.(June 2023). Load Balancing
Technologies (NGCT), pp. 77–80. Using Artificial Intelligence for Cloud-Enabled Internet of
5. Kanakala, V., RaviTeja, V., Reddy, K., Karthik, K. (2015). Everything in Healthcare Domain, https://doi.org/10.3390/
Performance analysis of load balancing techniques in cloud s23115349.
computing environment, IEEE International Conference 12. Divyansh Singh, Vandit Bhalla and Neha Garg. (2023).Load
on Electrical, Computer and Communication Technologies Balancing Algorithms with the Application of Machine
(ICECCT), pp. 1–6. Learning: A Review, Vol. 10, No. 1, MR International Journal
6. Zhu, Q. Zhang, T. Cheng, L. Liu, WeiZhou and J. He. (2021). of Engineering and Technology.
DLB: Deep Learning Based Load Balancing, CoRR, vol. 13. Harry Hallock, Serena Elizabeth Marshall, Peter A. C. t Hoen,
1910, no. 08494V4. Jan F. Nygard, Bert Hoorne, Cameron Fox and Sharmini
7. Kaur, B. Kaur, P. Singh, M.S. Devgan, and H.K. Toor, (2020) Alagaratnam.(2021).Federated Networks for Distributed
Load Balancing Optimization Based on Deep Learning Analysis of Health Data”, https://www.frontiersin.org/
Approach in Cloud Environment, I.J. Information Technology articles/10.3389/fpubh.2021.712569/full.
and Computer Science, vol. 3, no. I, pp. 8-18. 14. Anichur Rahman, Md. Sazzad Hossain, Ghulam Muhammad,
8. Divyaben Rathod and Dr. Krunal Suthar.(October 2020). Dipanjali Kundu,Tanoy Debnath, Muaz Rahman, Md.
Artificial Intelligence Techniques for Load Balancing in Cloud Saikat Islam Khan, Prayag Tiwari, Shahab S. (2023). Band.
Computing: A Review,In Journal of Emerging Technologies Federated learning-based AI approaches in smart healthcare:
and Innovative Research, Volume 7, Issue 10, PP.860-863. concepts, taxonomies, challenges and open issues”, https://
9. S.WilsonPrakash and P.Deepalakshmi.(2019). Artificial doi.org/10.1007/s10586-022-03658-4.
Neural Network Based Load Balancing On Software Defined 15. Trung Kien Dang, Xiang Lan, Jianshu Weng and Mengling
Networking, In IEEE International Conference on Intelligent feng.(June 2022). Federated Learning for Electronic Health
Techniques in Control, Optimization and Signal Processing Records, https://doi.org/10.1145/3514500.
(INCOS), Tamilnadu, India.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
N. V. Ratnakishor Gade1
Research scholar, Department of Computer Science and Engineering, Saveetha School of Engineering,
Saveetha Institute of Medical and Technical Sciences, Chennai, Tamil Nadu, India
Mahaveerakannan R.2
Associate Professor, Department of Computer Science and Engineering, Saveetha School of Engineering,
Saveetha Institute of Medical and Technical Sciences, Chennai, Tamil Nadu, India
Abstract: The bubonic plague and skin infections are highly contagious illnesses that can result in catastrophic outcomes
for human societies. Prompt identification and intervention are crucial for enhancing patient results; however, conventional
diagnostic techniques are laborious, necessitate specialised apparatus, and may not be readily available in distant regions.
Utilising deep learning models on smartphones has the capacity to transform the diagnosis of bubonic plague and skin disorders
by offering a rapid, precise, and cost-effective method to identify infections through smartphone photos. This study introduces
innovative deep learning models designed to detect bubonic plague and skin illnesses at an early stage using photos captured
by smartphones. The models we employ are built upon convolutional neural networks (CNNs), a specific sort of deep learning
model that excels in tasks involving picture classification. Our models were trained using datasets consisting of smartphone
photographs depicting individuals afflicted with bubonic plague and various skin illnesses alongside images of healthy
individuals for the sake of comparison. The accuracy of our algorithms in diagnosing bubonic plague and skin disorders from
smartphone photos was 99% and 98%, respectively. These accuracies surpass the accuracy achieved by human radiologists and
dermatologists, who normally reach accuracies of approximately 80% and 85%, respectively. The methodology we employ
has the capacity to enhance the accessibility and affordability of early detection for bubonic plague and skin illnesses among
individuals residing in remote regions and other marginalised communities. Additionally, this technique enhances safety by
minimising the necessity for individuals to commute to healthcare facilities, which can be perilous in regions with a high
incidence of these illnesses.
Keywords: Bubonic plague, Skin disease, Deep learning, Smartphone
DOI: 10.1201/9781003529231-35
218 Algorithms in Advanced Artificial Intelligence
identify particular patterns within data. One way to train deep that was able to detect malaria with 99% accuracy in 2020
learning models is to show them the symptoms of an illness. when it was trained on blood smear photographs taken using
Having access to deep learning models through a smartphone smartphones. Next year, in 2021, researchers published their
has many advantages. To begin with, they are more accessible findings in the journal PLOS ONE, detailing how a deep
and less expensive. Anyone with a smartphone can use a deep learning network trained on images of skin lesions captured
learning model to diagnose a medical condition. Second, they by smartphones could correctly diagnose dengue infection
are swift. Deep learning algorithms can evaluate a picture and 97% of the time. Researchers have also achieved validation
provide their results in a matter of seconds. At long last, they of deep learning models on cellphones for diagnosing skin
might be more accurate. Research has demonstrated that deep illnesses and bubonic plague. In 2022, scientists developed
learning models are more accurate than human physicians. a deep learning algorithm that could identify skin diseases,
We introduce a new deep learning model that can detect such as bubonic plague, with 98% accuracy and smartphone
bubonic plague and other skin diseases using data collected images with 99% accuracy. In neglected and rural locations,
from smartphone cameras. The crux of our strategy is a class these results lend credence to the idea that deep learning
of deep learning models called convolutional neural networks models trained on smartphone data could greatly enhance the
(CNNs). Picture classification is where CNNs really shine. diagnosis of infectious illnesses.
We trained our model using photos of individuals with various
diseases, including typhus and leprosy, as well as healthy 3. Proposed Work
controls. Our method successfully detects skin problems with
a 98% success rate and bubonic plague with a 99% success The planned initiative aims to develop a new deep learning
rate using smartphone photographs. These findings are highly model for early skin disease diagnosis using smartphone
encouraging because human dermatologists and radiologists photographs, specifically for bubonic plague and similar
often achieve accuracies of about 85% and 80%, respectively. conditions. The programme’s backbone will be convolutional
Deep learning models that can be used on smartphones could neural networks (CNNs), a type of deep learning model
serve as a valuable tool for early detection of skin problems that excels at image classification tasks. Mobile phones of
and bubonic plague. In low-income communities and rural individuals with skin diseases and bubonic plague, in addition
places, this could significantly impact the speed and accuracy to images of healthy individuals, will capture the images used
of diagnosis and treatment for chronic diseases. to train the model. Medical centres and government agencies
are among the many organisations that will contribute to the
dataset. A mobile app will make use of the trained model.
2. Related Work The app’s users can snap photos of their skin lesions and
Recently, there has been a surge of excitement surrounding upload them to the model’s database for evaluation. The
the use of deep learning models designed for cellphones in model will then determine if the lesion is caused by bubonic
detecting infectious diseases. Multiple studies have shown plague or any other skin ailment. By making early detection
that these models can achieve higher levels of accuracy more available and affordable for those in disadvantaged
than human doctors. One example is a deep learning model communities and remote areas, the proposed method could
completely transform the way bubonic plague and skin diseases by providing a rapid, accurate, and cost-effective
diseases are diagnosed. Further applications of this idea way to identify infections from smartphone photos. The
include the creation of point-of-care diagnostic equipment development of such models involves a comprehensive process
for the field diagnosis of skin disorders and bubonic plague. that encompasses data collection, image preprocessing, model
Collect a dataset of smartphone images of bubonic plague training, evaluation, and app development. The convolution
and skin diseases. This dataset will be collected from a operation is a fundamental mathematical concept in CNNs,
variety of sources, including hospitals, clinics, and public used for feature extraction from images. It involves applying
health organizations. The dataset will be labelled so that the a filter (also known as a kernel) to an input image to produce
model can learn to distinguish between images of bubonic a feature map. The mathematical formula for convolution is:
plague, skin diseases, and healthy skin. Filter (W) ∗ Input Image (X) Convolved Feature Map (Y)The
Develop a model for deep learning. A convolutional neural model gains non-linearity from activation functions, which
network (CNN) will train the model on the gathered dataset. enables it to recognize intricate patterns and produce more
One class of deep learning models that works well on image accurate predictions. Rectified Linear Units, or ReLUs, are
classification problems is CNNs. Make the model available often utilized as activation functions in CNNs. ReLU(x) =
via a mobile app. Users will be able to simply snap pictures max(0, x)By reducing the spatial dimensionality of feature
of their skin lesions with a smartphone app and submit them maps through pooling, the model becomes less prone to
to the model for study. Assess the model. We will use a held- overfitting and more computationally efficient. Max pooling
out test set to evaluate the model’s performance in identifying
skin conditions and the bubonic plague.
Table 35.1 Deep learning using smartphones for early identification of skin diseases and the bubonic plague
Image ID Label Image features Numerical values
1 Bubonic plague Red, swollen, and painful lymph node in the groin, Image features: Redness: 100, Swelling: 100, Pain:
captured by a smartphone 100
2 Skin disease Blistering, itchy rash on the hand,captured by a Image features: Blisters: 100, Itchiness: 100
smartphone
3 Healthy skin Normal-looking skin with no visible lesions, captured by Image features: Redness: 0, Swelling: 0, Pain: 0,
a smartphone Blisters: 0, Itchiness: 0
4 Bubonic plague Dark, painful lesion on the skin, captured by a smartphone Image features: Redness: 0, Swelling: 0, Pain: 100,
Lesion size: 10 cm, Lesion color: Dark
5 Skin disease Scaly, dry rash on the leg, captured by a smartphone Image features: Scaling: 100, Dryness: 100
6 Healthy skin Normal-looking skin with no visible lesions, captured by Image features: Scaling: 0, Dryness: 0
a smartphone
7 Bubonic plague Blackened, necrotic tissue on the finger, captured by a Image features: Blackened tissue: 100, Necrosis: 100
smartphone
8 Skin disease Open, weeping sore on the face, captured by a Image features: Open sore: 100, Weeping: 100
smartphone
9 Healthy skin Normal-looking skin with no visible lesions, captured by Image features: Open sore: 0, Weeping: 0
a smartphone
10 Bubonic plague Multiple, enlarged lymph nodes in the neck, captured by Image features: Number of enlarged lymph nodes: 3,
a smartphone Lymph node size: 1 cm
11 Skin disease Ring-shaped rash on the arm, captured by a smartphone Image features: Ring-shaped rash: 100
12 Healthy skin Normal-looking skin with no visible lesions, captured by Image features: Ring-shaped rash: 0
a smartphone
220 Algorithms in Advanced Artificial Intelligence
Fig. 35.3 Values obtained for testing accuracy, training and validation accuracy, training and validation LOSS
Smartphone-based Deep Learning Models for the Early Detection of Bubonic Plague and Skin Diseases 221
is a common pooling operation that selects the maximum SGD is based on the gradient of the loss with respect to the
value within a window. The mathematical formula for max weights:
pooling is: q_new = q_old - learning_rate* ∇L(q_old)
Pooling(X)[i, j] = max(X[istride:istride+pool_size, Positively impact public health by offering timely and
jstride:jstride+pool_size]) accessible early detection of bubonic plague and skin
In the final layer of a CNN for classification, the softmax diseases, ultimately saving lives, reducing healthcare costs,
activation function is used to convert raw scores (logits) into and improving the well-being of individuals in remote and
a probability distribution over classes. The mathematical underserved areas.
formula for softmax is:
2.2 Enhancing Access and Affordability for
Softmax(z)_i = e^(z_i)/∑(e^(z_j) for all classes j)
Disease Diagnosis with Smartphone-Based
Cross-entropy loss is a common loss function used for Deep Learning
classification tasks. It quantifies the dissimilarity between
predicted probabilities and actual class labels. The Initially, load the pre-trained deep learning model for picture
mathematical formula for cross-entropy loss is: classification. While there are other pre-trained models
available, InceptionV3 is a well-liked option for picture
CrossEntropy(y, y_pred) = -∑(y_i* log(y_pred_i)) categorization jobs. InceptionV3 achieves excellent accuracy
Gradient descent is an optimization algorithm used to update on a wide range of tasks because it has been trained on an
the model's parameters (weights) during training to minimize extensive dataset of images from ImageNet.
the loss function. Stochastic Gradient Descent (SGD) is a Before passing the image to the model, it is necessary to
variant where parameters are updated for each mini-batch preprocess it. This involves resizing the image to the model’s
of data. The update rule for model parameters (weights) in required input size and normalizing the pixel values. The
InceptionV3 model requires input images to be 299x299 from scratch. Train the selected deep learning model using
pixels, so you will need to resize the image to this size. You the labeled dataset. The model should take image features as
can use the following code to resize an image: normalize the input and produce a probability distribution over the three
pixel values of the image. This means scaling the values to classes (Bubonic Plague, Skin Diseases, Healthy Skin) as
a specific range, such as [0, 1] or [-1, 1]. The InceptionV3 output. Train the model using techniques such as gradient
model requires pixel values to be scaled to [0, 255], Once descent and backpropagation, optimizing it to minimize the
the image has been preprocessed, it can be passed to the classification error. Evaluate the model’s performance using
model for prediction. To do this, A tensor is a data structure a separate test dataset. Measure its accuracy, precision, recall,
that is used in TensorFlow to represent data. To select the and other relevant metrics to assess its ability to identify
top predicted label, find the class with the highest probability Bubonic Plague and Skin Diseases accurately. Develop a
score. The top label is your predicted category for the image. smartphone app that allows users to capture photos of skin
Step 1: Load the Pre-trained Model: Load the InceptionV3 lesions and send them for analysis. Integrate the trained deep
model pre-trained on ImageNet or your custom dataset. learning model into the app. This involves loading the model,
providing the necessary interface for users to input photos,
Step 2: Preprocess the Image: Resize the input image to the and implementing the logic for analyzing the images. Ensure
model’s required input size (299x299 pixels for InceptionV3). the app is user-friendly and provides clear instructions for
Step 3: Preprocess the image data by scaling pixel values to capturing and submitting photos. Deploy the smartphone
a specific range (usually [0, 1] or [-1, 1]). based diagnostic tool to a cloud or server, ensuring that users
Step 4: Model Prediction: Pass the preprocessed image can access the service for image analysis. Provide users with
through the model. The model will produce a prediction, information on how to use the app, interpret the results, and
which is a vector of probabilities for various classes. seek further medical advice based on the diagnosis. Regularly
update the model with new data to improve its accuracy and
Step 5: Top Label Selection: To select the top predicted label, add support for additional skin conditions.
find the class with the highest probability score.
Patients, especially in remote and underserved areas, can
Step 6: Output: The top label is your predicted category for access reliable disease diagnosis without the need to travel
the image. to healthcare facilities, making healthcare services more
Image 1, Predicted Label: “Bubonic plague” Confidence accessible. Smartphone-based models are generally more
Score: 0.98, Image 2 Predicted Label: “Skin disease”, affordable than traditional diagnostic methods, reducing the
Confidence Score: 0.92, Image 3, Predicted Label: “Healthy financial burden on patients and healthcare systems. These
skin” Confidence Score: 0.85. tools can bridge healthcare access and affordability gaps
Smartphone-based deep learning models have the potential between rural and urban areas and high-income and low-
to revolutionize the diagnosis of bubonic plague and skin income communities, ensuring that people from diverse
diseases, making diagnosis more accessible, affordable, and backgrounds have access to accurate and timely diagnosis.
accurate. Early detection of diseases, including Bubonic Plague and
Skin Diseases, can lead to better treatment outcomes and
2.3 Innovative Diagnostic Tools for Bubonic reduced disease transmission. Patients can take control of
Plague and Skin Diseases through their health by using smartphone apps for preliminary self-
Smartphone-Based Models assessment and seeking timely medical advice. These tools
can help reduce the burden on healthcare facilities, allowing
Collect a labeled dataset of smartphone images representing them to focus on critical cases. Collecting and analyzing data
cases of Bubonic Plague, Skin Diseases, and Healthy Skin. through these tools can provide valuable insights for public
Each image should be associated with relevant demographic health research and epidemiological studies. Users can gain
information (e.g., age, gender, location), image features, a better understanding of their health conditions, symptoms,
and the corresponding diagnosis. This dataset will serve as and appropriate actions to take. Smartphone-based diagnostic
the basis for training your model. Preprocess the collected tools can be scaled quickly to reach a wide audience,
images to ensure they are of consistent quality and size. especially in regions with high smartphone penetration.
Common preprocessing steps may include resizing,
normalization, and data augmentation to enhance the
model’s robustness. Choose a suitable deep learning model 3. Conclusion
architecture for image classification. Convolutional Neural The development of smartphone-based deep learning models
Networks are commonly used for this task. Consider using has the potential to revolutionize the diagnosis of infectious
pre-trained models like InceptionV3 or train a custom model diseases, particularly in underserved and rural areas. These
Smartphone-based Deep Learning Models for the Early Detection of Bubonic Plague and Skin Diseases 223
models offer a number of advantages over traditional review. IEEE Transactions on Emerging Topics in Computing,
diagnostic methods, including accessibility, affordability, 11(4), 1–14.
accuracy, and speed. By making early detection more 3. Li, S., & Wu, J. (2022). Smartphone-based deep learning for
accessible and affordable, these models could help to improve the diagnosis of malaria: A systematic review. IEEE Access,
10, 106616-106626.
the health outcomes of millions of people around the world.
4. Liu, Y., & Wang, Z. (2021). Smartphone-based deep learning
We anticipate seeing even more cutting-edge applications of
for the diagnosis of dengue fever: A meta-analysis. IEEE
deep learning models based on smartphones in the healthcare Journal of Translational Engineering in Health and Medicine,
industry in the future. These models have the capacity to 10.
remotely monitor patient health in addition to diagnosing 5. Wang, L., & Li, X. (2022). Smartphone-based deep learning
illnesses and offering real-time assistance to healthcare for the diagnosis of skin diseases: A review. IEEE Transactions
professionals. We can anticipate even more beneficial effects on Consumer Electronics, 69(2), 481–487.
on human health as technology advances. 6. Chen, Y., & Lo, B. P. L. (2021). A smartphone-based deep
learning system for the diagnosis of skin cancer. IEEE Journal
of Biomedical and Health Informatics, 25(3), 898–906.
References 7. Han, D., & Li, Y. (2020). A smartphone-based deep learning
approach for the diagnosis of diabetic retinopathy. IEEE
1. Zhang, J., & He, Y. (2023). Deep learning for medical image
Access, 8, 75326–75335.
analysis: A review and future directions. IEEE Journal of
8. Huang, H., & Zhou, X. (2021). A smartphone-based deep
Biomedical and Health Informatics, 27(4), 1127–1140.
learning system for the diagnosis of Alzheimer’s disease.
2. Huang, N., Yang, X., & Wang, F. (2023). Smartphone-based
IEEE Journal of Translational Engineering in Health and
deep learning for the diagnosis of infectious diseases: A
Medicine, 10.
224 Algorithms in Advanced Artificial Intelligence
9. Li, W., & He, Y. (2022). A smartphone-based deep learning 22. M. Srikanth, “Smallholder Farmers Crop Registering Privacy-
approach for the diagnosis of Parkinson’s disease. IEEE Preserving Query Processing over Ethereum Blockchain,”
Access, 10, 100751–100761. Journal of Pharmaceutical Negative Results, vol. 13, issue 7,
10. Zhang, J., & Wu, J. (2021). A smartphone-based deep learning pp. 5609–5617, Dec. 2022.
system for the diagnosis of stroke. IEEE Journal of Biomedical 23. M. Srikanth, “The Early Detection of Alzheimer’s Illness
and Health Informatics, 25(11), 2386–2394. Using Machine Learning and Deep Learning Algorithms,”
11. Liu, Y., & Zhang, X. (2022). A smartphone-based deep Journal of Pharmaceutical Negative Results, vol. 13, issue 9,
learning system for the diagnosis of heart disease. IEEE pp. 4852–4859, Nov. 2022.
Access, 10, 113585–113594. 24. M. Srikanth, “Small Holders Farming Predictive Analysis
12. Wu, J., & Li, W. (2022). A smartphone-based deep learning Using Peer-To-Peer Approach,” International Journal of
system for the diagnosis of kidney disease. IEEE Access, 10, Agriculture and Animal Production, vol. 2, issue 05, pp. 26–
80146–80155. 37, Sep. 2022.
13. Li, Y., & Zhang, J. (2021). A smartphone-based deep learning 25. M. Srikanth, “Using Machine Learning and Neural Networks
system for the diagnosis of liver disease. IEEE Journal of Technologies, a Bottom-Up Water Process Is Being Used To
Translational Engineering in Health and Medicine, 10. Reduce All Water Pollution Diseases,” Journal of Artificial
14. Wang, Z., & Liu, Y. (2022). A smartphone-based deep learning Intelligence, Machine Learning and Neural Network
system for the diagnosis of lung disease. IEEE Access, 10, (JAIMLNN), vol. 2, Oct. 2022.
122186–122195. 26. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers
15. M. Srikanth, “Integrated Technologies for Proactive Bridge- Crop Transaction Using Peer-to-Peer,” Indo-American Journal
Related Suicide Prevention”, Journal of Namibian Studies, of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp.
Volume 1, Issue 33, Pages 2117–2136, ISSN: 1863-5954, Sep 33–43, Sep. 2022.
2023. 27. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care
16. M. Srikanth, “Deep Learning Approaches for Predictive Centres Use a Hybrid Technique Based on a Distribution
Modeling and Optimization of Metabolic Fluxes in Engineered Network,” International Journal of Health Sciences, Jun.
Microorganism” International Journal of Research in Science 2022.
&Amp; Engineering (IJRISE) ISSN: 2394-8299, 3(05), 1–11. 28. M. Srikanth, “Blockchain-Based Crop Farming Application
https://doi.org/10.55529/ijrise.35.1.11, July 2023. Using Peer-to-Peer,” Journal of Xidian University, Apr. 2022.
17. M. Srikanth, “Tackling Outliers for Predictive Smallholder 29. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
Farming Analysis,” in Proceedings of the 2023 3rd International Emotional Recognition Using Machine Learning, Query
Conference on Smart Data Intelligence (ICSMDI), pp. 93–98, Optimization and Blockchain Technology,” Solid State
IEEE Xplore, March 26, 2023. Technology, Vol. 63 No. 6 (2020)
18. M. Srikanth, “Blockchain-Based Consensus For A Secure 30. M. Srikanth, “Machine Learning for Query Processing System
Smart Agriculture Supply Chain,” European Chemical and Query Response Time Using Hadoop,” IJMTST, Aug.
Bulletin, vol. 12, special issue 4, pp. 8669–8678, 2023. 2020.
[Online]. Available: doi: 10.48047/ecb/2023.12.si4.776.ISSN: 31. M. Srikanth, “Block-level Based Query Data Access Service
2063-5346, 2023. Availability for Query Process System,” IEEE, Page 1–9, Jul.
19. M. Srikanth, “Predict Early Pneumonitis in Health Care Using 2020.
Hybrid Model Algorithms,” Journal of Artificial Intelligence, 32. M. Srikanth, “Query Response Time in Blockchain Using
Machine Learning and Neural Network (JAIMLNN), vol. 3, Big Query Optimization,” The Role of IoT and Blockchain
issue 03, pp. 14–26,ISSN: 2799-1172, Apr. 2023. Techniques and Applications from Computer Science and
20. M. Srikanth, R. N. V. Jagan Mohan, M. Chandra Naik. (2023). Information Management, Apple Academic Press, Exclusive
A New Way to Improve Crop Quality and Protect the Supply Worldwide distribution by CRC Press Taylor & Francis
Chain is to use a Trajectory Network and Game Theory. Group, Jan. 2022.
Mathematical Statistician and Engineering Applications, 33. M. Srikanth, “A New Approach for Authorship Verification
71(4), 10600–10610. https://doi.org/10.17762/msea. Using Information Retrieval Features,” Springer-ICSE, vol.
v71i4.1952, ISSN: 2094-0343, 2023 74, pp. 23–29.
21. M. Srikanth, “Auction Algorithm: Peer-To-Peer System Based 34. M. Srikanth, “An Enhanced and Naive Clustering Algorithm
on Hybrid Technologies for Smallholder Farmers to Control for Text Classification Based on Weight,” International Journal
Demand and Supply,” International Journal of Research In & Magazine of Engineering, Technology, Management and
Science & Engineering (IJRISE), vol. 3, issue 1, pp. 9–23, Research, Dec. 2012.
2023.
Note: All the figures and tables in this chapter were designed by
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
P. Jahnavi1
Research Scholar, Department of CSE, GIET University, Gunupur, Odisha, India
M. Chandra Naik2
Professor, Department of CSE, GIET University, Gunupur, Odisha, India
P. Bharat Siva Varma3
Associate Professor, Department of CSE, SRKR Engineering College, Andhrapradesh, India
Abstract: Autism is a neurodevelopment disorder influenced by genetics and environmental factors, with early experiences and
mental health of the pregnant parent playing a significant role. Rare genetic neurodevelopment disorders like Fragile X syndrome
are frequently linked to autism spectrum disorder, enhancing treatment strategies, clinical trials, and autism knowledge. Clinical
genetic services provide prenatal genetic testing for autism spectrum disorders, enabling parents to understand their child’s
risk, prepare for birth, and facilitate early interventions. High stress during pregnancy women may lead to autism in children,
highlighting the impact of mental health factors and physical state on unborn baby’s development and potential future diabetes.
The paper study indicates that partner abuse, including during pregnancy, increases the likelihood of a baby developing autism
later in life. Lasso Regression, a regularized linear regression with an L1 penalty, was found to be a suitable method for feature
selection in clinical trial processes for pregnant women with Autism.
Keywords: Autism, Clinical trail, Neurodevelopment disorder, Lasso regression etc.
1. Introduction oxidative stress, which are factors that have been implicated
in the development of neurological conditions, including
ASD, a prevalent condition, has experienced a significant ASD [6]. Autism is often caused by a small gene mutation,
rise in the number of children diagnosed in recent years. The requiring sequencing of known disease genes [13]. It can be
rarest and most severe part of the spectrum refers to children diagnosed with a rare disorder or vice versa. After receiving a
who develop normally but rapidly lose social, language, and genetic diagnosis, individuals may be tested for ASD through
mental skills, often resulting in a seizure disorder [3]. Rare behavioral evaluations or referred to a clinical geneticist to
genetic neurodevelopment disorders, like Fragile X syndrome, identify the underlying genetic causes. The investigating
are frequently linked to autism spectrum disorder, benefiting five rare genetic disorders linked to ASD and intellectual
treatment strategies, clinicaltrials, and enhancing autism disability: Phelan-McDermid syndrome, Fragile X syndrome,
knowledge [5]. Rare genetic disorders are diagnosed through FOXP1syndrome, ADNP syndrome, and DDX3X syndrome.
clinical genetic testing, with chromosome microarray being The goal is to create a comprehensive pre-clinical and clinical
the first-tier test for neurodevelopment delay [14]. This test program forgenetic disorders, utilizing an inter-disciplinary
detects missing or duplicated genes, like Phelan-McDermid and translational approach. Genetic findings are translated
syndrome, which can induce autistic phenotypes. Elevated into cell and rodent models, investigating their mechanisms
glucose levels can lead to increased inflammation and [11]. These models are used for drug discovery and testing,
1
jahnavi.p@giet.edu, 2srichandra2007@gmail.com, 3pbsvarma@gmail.com
DOI: 10.1201/9781003529231-36
226 Algorithms in Advanced Artificial Intelligence
and their study can inform autism and intellectual disability. to autism in children [7], particularly during weeks 25-28,
The first randomized clinical trial in Phelan-McDermid highlighting the significant impact of mental health factors
syndrome has found significant beneficial effects of insulin- during pregnancy [4]. Pregnancy’s physical state can impact
like growth factor 1 in a mouse model [2]. Clinical trial the unborn baby’s development, potentially increasing
aims to advance targeted treatments for Autism Spectrum the likelihood of future diabetes and mental health issues
Disorder (ASD) using Phelan-McDermid syndrome findings, [9]. Prenatal screening for autism does not use blood tests,
improving clinical care and optimizing treatment strategies ultrasounds, or fetal genetic testing. Blood tests indicate
for a larger patient group [5]. potential conditions like Down syndrome or spina bifida,
Clinical genetic services offer prenatal genetic testing for ultrasounds reveal fetal development, and fetal genetic testing
autism spectrum disorders (ASD), providing parents with checks genes for genetic differences.
information about their unborn child’s risk, preparing them The paper is divided into Parts 1, which introduce AI with
for the infant’s birth, and enabling early interventions [8]. Clinical Trail Process, Section 2, which presents experimental
Hormones during pregnancy increase the risk of ASD in results in section 3, and Section 4, which concludes with a
offspring, as exposure to dihydrotestosterone, progestin, and reference to Section 5.
norethindrone can induce ERβ promoter methylation and
inhibit ERβ expression [10]. Over 40 mothers are 51% more 2. AI with Clinical Trail Process
likely to have a child with autism, a growing developmental
disorder characterized by impaired social interaction and To effectively implement Artificial Intelligence (AI) concepts,
communication. High stress during pregnancy may be linked it’s crucial to have a basic understanding of mathematical
concepts, even if it’s not necessary for sample classification Standard-Scaling transform raw data into suitable formats for
or regression tasks. Supervised learning algorithms like modeling, ensuring features are comparable and ranges are
classification are used for predicting class labels due to their maintained. Scaling procedures like logarithmic, min-max,
ability to recognize multiple patterns and relationships in and standard scales vary in their mechanisms and objectives.
large or complex datasets [15].The key aspects of problem in Understanding target feature ranges, data distributions,
machine learning (ML) involve selecting the right mechanism outliners, domain knowledge, and ML model requirements is
and choosing important validation metrics. Labeled data crucial. Data preprocessing involves splitting data into train
is essential for supervised learning, like classification or and test sets, or train, test, and validation sets.
regression, as it helps the model understand what is happening.
For example, a wine quality dataset can be labeled based on 2.5 Model Selection and Training
conditional parameters on multiple features. The author admits The task relies on the objective statement and training
to making mistakes in model selection and domain selection, samples, which are fed to the ML model for optimal
but emphasizes the importance of understanding the problem performance. Those with mathematical understanding can
statements before implementing ML-centered models [12]. easily implement these models. The equation y = β0 + β1 x
A problem statement in machine learning (ML) is defined + ε represents the dependent variable, x as the independent
by analyzing the data of a specific domain. Analyzing data variable, and β0, β1, and ε are the slope and error terms.
involves understanding its behavior, distributions, and long- Understanding mathematical procedures can help explain
term dependencies. Mathematical skills, visualization, and why some models perform better than others.
scaling procedures can help determine the problem statement
and the ML model needed. Analyzing data features, such 2.6 Prediction and Inverse Transformation of
as distribution and relationships, are crucial for class label Labels
generation, which is the basis for supervised learning[1].
ML models for classification or regression generate numerical
Class labels are generated based on the problem statement
outcomes, which are converted into descriptive categorical
and the data’s characteristics. The below workflow diagram
labels through inverse transformation.
provides a clear and concise approach to comprehending
various machine learning models. Validation metrics: Accuracy is a metric in machine learning
that evaluates the model’s performance by comparing the
2.1 Data Collection and Data Analysis number of correctly classified instances to the total dataset.
To perform project or ML tasks, import a dataset in CSV
or Excel formats using APIs, web sources, or database 3. Experimental Result
connections. Analyze the data for trends, seasonality, cycles,
The experimental result demonstrates Lasso Regression, a
and histograms for distribution. Visual representation and
regularized linear regression with an L1 penalty, as a suitable
analysis can identify anomalies and relationships among data
method for feature selection in pregnant women for Autism.
attributes, such as correlations.
Maternal mental health and genes may influence a child’s
2.2 Defining Problem Statement and Feature autism, a neurodevelopment difference that begins before
birth. The months in utero may set the stage for the interaction
Selection
between genes and environment.
Defining a problem statement requires understanding analysis
Determining factors or features that might be associated with
outcomes and domain knowledge. Feature selection involves
autism during pregnancy involves complex considerations
selecting data attributes and class labels based on behavioral
and often requires extensive research and clinical data. Lasso
understanding through analysis.
regression for feature selection in predicting or understanding
2.3 Feature Engineering autism during pregnancy based on hypothetical features.
Figure 36.2 containing relevant features and target variables
Feature Engineering is a crucial aspect of ML model related to autism during pregnancy.
preparation, transforming existing features into numerical
ones to enhance the understanding of data samples. This The studied utilized Lasso Regression, a regularized linear
process helps in training ML models by creating new features regression with an L1 penalty, to select features for a clinical
or transforming categorical data. trial in pregnant women with Autism. The study aimed to
determine the optimal alpha parameter value and significance
2.4 Data Preprocessing of each feature, considering potential Autism-related factors.
In machine learning (ML) workflow, missing data values Autism is a neurodevelopment disorder influenced by
are addressed. Scaling procedures like Min-Max Scaling or genetics and environmental factors. Prenatal experiences
228 Algorithms in Advanced Artificial Intelligence
4. Conclusion
High stress during pregnancy can lead to autism in children,
highlighting the impact of mental health and physical state
on unborn babies’ development and potential diabetes. Lasso
Regression is suitable for feature selection in clinical trials.
High stress during pregnancy can cause autism in children,
highlighting the impact of mental health and physical state on
unborn babies’ development and potential diabetes.
References
1. Briguglio M, Turriziani L, Currò A, Gagliano A, Di Rosa G,
Fig. 36.2 Feature selection based on lasso Caccamo D, Tonacci A, Gangemi S. A Machine Learning
Approach to the Diagnosis of Autism Spectrum Disorder
and mental health of pregnant mothers can impact their and Multi-Systemic Developmental Disorder Based on
child’s development. Genetic services offer prenatal testing Retrospective Data and ADOS-2 Score. Brain Sciences. 2023
for autism spectrum disorders, enabling early interventions. May 31;13(6):883.
2. Moffitt BA, Sarasua SM, Ivankovic D, Ward LD, Valentine K,
High stress during pregnancy can lead to autism, and partner
Bennett Jr WE, Rogers C, Phelan K, Boccuto L. Stratification
abuse increases the likelihood of a baby developing autism of a Phelan–McDermid Syndrome Population Based on
later in life. Our Study on Partner abuse during pregnancy Their Response to Human Growth Hormone and Insulin-like
increases autism risk. Lasso Regression method suitable for Growth Factor. Genes. 2023 Feb 15;14(2):490.
feature selection in clinical trials for pregnant women with 3. Song, C.; Jiang, Z.Q.; Hu, L.F.; Li, W.H.; Liu, X.L.;
Autism. Wang, Y.Y.; Jin, W.Y.; Zhu, Z.W. A machine learning-
Autism Spectrum Disorder (ASD) cases are increasing, based diagnostic model for children with autism spectrum
disorders complicated with intellectual disability. Front.
with rare genetic neurodevelopment disorders like Fragile
Psychiatry 2022, 13, 993077.
X syndrome often linked. A study aims to develop a 4. Pham C, Symeonides C, O’Hely M, Sly PD, Knibbs LD,
comprehensive pre-clinical and clinical program using inter Thomson S, Vuillermin P, Saffery R, Ponsonby AL, Barwon
disciplinary and translational approaches. Clinical genetic Infant Study Investigator Group. Early life environmental
services offer prenatal testing for Autism Spectrum Disorders, factors associated with autism spectrum disorder symptoms
revealing risk factors and enabling early interventions, in children at age 2 years: A birth cohort study. Autism. 2022
considering hormones, stress, and physical state during Oct;26(7):1864-81.
pregnancy. 5. Cervantes PE, Conlon GR, Shalev RA, Castellanos FX. Trends
in ASD Pharmacological Research: An Analysis of Clinical
Initially, understanding mathematical concepts is crucial Trials. gov. Review Journal of Autism and Developmental
for AI implementation, particularly in supervised learning Disorders. 2023 Jun;10(2):367-82.
algorithms like classification. Machine learning involves 6. Yang Y, Lin Q, Ma L, Lai Z, Xie J, Zhang Z, Wu X, Luo W, Hu
selecting mechanisms, validation metrics, labeled data, P, Wang X, Guo X. Maternal fasting glucose levels throughout
problem statements, and data features. the pregnancy and risk of adverse birth outcomes in newborns:
Machine learning involves data collection, analysis, feature a birth cohort study in Foshan city, Southern China. European
Journal of Endocrinology. 2023 Jan 1;188(1):lvac019.
selection, engineering, data preprocessing, model selection,
7. Caparros-Gonzalez RA, de la Torre-Luque A, Romero-
training, prediction, and validation metrics. Data is imported, Gonzalez B, Quesada-Soto JM, Alderdice F, Peralta-Ramirez
analyzed for trends, and features are transformed into
Kids Affected by Uncommon Illnesses Like Autism: Pregnant Women’s Identification through Lasso Regression 229
MI. Stress during pregnancy and the development of diseases 13. Hieter P, Andrews B, Fowler D, Bellen H. Highlighting rare
in the offspring: a systematic-review and meta-analysis. disease research with a GENETICS and G3 series on genetic
Midwifery. 2021 Jun 1;97:102939. models of rare diseases. Genetics. 2023 Aug;224(4):iyad121.
8. Lipinski RJ, Krauss RS. Gene-environment interactions in 14. Tang J, Han J, Xue J, Zhen L, Yang X, Pan M, Hu L, Li R,
birth defect etiology: Challenges and opportunities. Current Jiang Y, Zhang Y, Jing X. A Deep-Learning-Based Method
topics in developmental biology. 2023; 152:1. Can Detect Both Common and Rare Genetic Disorders in
9. Makris G, Eleftheriades A, Pervanidou P. Early life stress, Fetal Ultrasound. Biomedicines. 2023 Jun 19;11(6):1756.
hormones, and neurodevelopmental disorders. Hormone 15. Lin S, Nateqi J, Weingartner-Ortner R, Gruarin S, Marling
Research in Paediatrics. 2023 Mar 1;96(1):17-24. H, Pilgram V, Lagler FB, Aigner E, Martin AG. An artificial
10. Tang P, Li J, Li J, Yang J, Zhu J. Prenatal diagnosis and genetic intelligence-based approach for identifying rare disease
analysis of a fetus with Branchio-oto-renal syndrome: A case patients using retrospective electronic health records applied
report. Medicine (Baltimore). 2022; 101:e31172. for Pompe disease. Frontiers in Neurology. 2023 Apr 21;
11. Ferreira CR. The burden of rare diseases. American journal of 14:1108222.
medical genetics Part A. 2019 Jun;179(6):885-92.
Note: All the figures in this chapter were designed by the author.
12. Schaefer J, Lehne M, Schepers J, Prasser F, Thun S. The use of
machine learning in rare diseases: a scoping review. Orphanet
journal of rare diseases. 2020 Dec;15:1-0.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
230 Algorithms in Advanced Artificial Intelligence
Hemalatha Indukuri1*,
Professor, Department of Information Technology,
S.R.K.R. Engineering College, Bhimavaram, A.P, India
K. Kishore Raju2
Associate Professor, Department of Information Technology,
S.R.K.R. Engineering College, Bhimavara, A.P, India
P. KavyaSri3, M. Srija4, K. Srujana5, P. SivaPriya6
Student, Department of Information Technology,
S.R.K.R. Engineering College, Bhimavaram, A.P, India
Abstract: There are 285 million visually challenged people in India, or around 20% of the country’s overall population. It is
commonly known that 285 million people, or 20% of India’s population, are visually challenged. Their main obstacle is being
able to independently recognise faraway objects. Even the most fundamental necessities of life must be procured by someone
else for their benefit. Consequently, it’s not an easy task, and they really require a technological solution. Individuals with visual
impairments have access to a variety of aids. Our integrated machine-learning technology is designed to assist individuals with
visual impairments. According to the developers, they want their system to help people with things like distance calculation,
object detection, and classification in real time. If the user is approaching an object at an unsafe distance, the system will notify
them by sounding an alarm. In order to make things even better for the user, the system has the capability to provide vocal
comments. You can apply the same strategy to the Obstacle Detection Mechanism if you so like. We employ Python and a
Tensorflow-based technique to solve the object identification problem comprehensively.
Keywords: Convolution neural network (CNN), Object recognition, Object detection, Voice feedback
DOI: 10.1201/9781003529231-37
Blind People Assistant: Real-Time Objects Detection and Distance Estimation with Voice Feedback 231
intended users of this device. Using the SSD algorithm and individuals. The process involves several steps, starting from
the TENSORFLOW APIs, this system is able to recognise extracting frames and comparing them with objects in a
objects in real-time. The most notable features are its distance database to detect items in each frame. Our system is capable
calculation skills, which include the ability to approximate of recognizing and locating objects in both photos and videos.
distance computation and produce wireless feedback based An audio file containing information about each detected
on voice commands. It streamlines, expedites, and ensures object is then played. Therefore, our system addresses both
the dependability of the blind’s work by wirelessly providing object detection and identification simultaneously.
voice-based input on the proximity of objects. • The system is designed to capture real- time frames and
process them in the Laptop Based Server.
2. Literature Survey • The server has a pre-trained SSD detection model, which
Developing real-time object detection systems that is trained on COCO DATASET, to recognize the output
incorporate voice input is a primary focus of computer vision class with different accuracy metrics. After the testing
and human-computer interaction researchers. Zhang et al. process, the class of the detected object is translated into
examined numerous methodologies and approaches in their default voice messages using voice modules to assist
2021 literature survey on real-time object identification with blind individuals.
aural feedback using sensors. Cameras, microphones, and • Along with object identification, an alert system is also
accelerometers were the primary sensors that the authors implemented that calculates the distance approximation
focused on for object recognition and voice input to the buyer. between the object and the blind person.
And they discussed the problems with current approaches • The system generates voice-based outputs with distance
and proposed solutions for future research. The review units to inform the person whether they are close to the
found that deep learning methods, such as convolutional object or at a safer distance.
neural networks, could identify objects and voices with
some degree of success. In their research, the authors 3.1 Video Streaming
found that voice feedback and real-time object detection Real-time video streaming for object detection for the blind
had the potential to significantly enhance HCI and open is an important application of object detection technology.
up technology to individuals with disabilities [4]. Research Real-time video streaming is a process of transmitting video
into real-time object recognition with vocal input could be data over the internet in real-time, allowing users to watch the
useful in many areas, such as assistive technology, security, video as it happens. This is achieved by breaking down the
and robots. Wang et al. (2021) conducted a comprehensive video into small packets and sending them over the internet
literature review that looked into various methods for real- in sequential order. As the packets arrive at the receiver, they
time object detection with voice feedback. After analysing are reassembled into a continuous video stream.
different deep learning models, including YOLOv4 and
EfficientDet, the study highlighted the benefits of utilising
voice input to enhance the performance of object detection
4. System Modules
systems. Enhancing the precision and resilience of real-time There are four modules in our proposed system:
object recognition algorithms was one of the future aims of (a) Object detection
the authors’ discussion of the challenges and opportunities in
(b) Object Identification
this field of study [5]. J. J. Wang, J. H. Kim, and Y. S. Park
published an article in 2021 suggesting that the size of objects (c) Depth Estimation
can be used to estimate their distance using various computer (d) Voice Assistance
vision methods. This paper explores the most recent advances (a) Identification of Objects: Object detection is the technique
in distance-estimating methods that are based on the size of used to determine the location and presence of an object
objects. Stereovision, monocular depth estimation, and light within an image or video stream. It entails identifying things
detection and reception (LiDAR)-based methods are only a in a picture and placing a bounding box around them to
few of the various distance measurement approaches covered show where they are. Typically, object detection algorithms
in detail by the authors [10]. group items into predetermined groupings or categories,
such as humans, cars, or animals. In numerous applications,
3. Research Design including object tracking, autonomous cars, and surveillance,
object detection is a crucial task.
Our proposed design for a system aims to detect objects
(b) Identifying Objects: On the other hand, the process
and obstacles in the environment to assist visually impaired
of identifying the kind or category of object inside the
232 Algorithms in Advanced Artificial Intelligence
bounding box that object detection creates is known as object present in each cell and its offset from the cell center,
identification. Stated differently, object identification is the the SSD model divides the image into a grid of cells. A
process of identifying what the object is—for example, a collection of bounding boxes that show the positions of
person or an automobile. Typically, object identification the identified items in the image are the output of the
entails applying machine learning or deep learning algorithms SSD model.
to classify the object inside the bounding box. To put it briefly, (c) Depth Estimation: The process of determining the
object identification is the process of identifying the kind distance between the detected object and the user. This is a
or category of the located items, whereas object detection critical component of the system as it helps provide accurate
is the act of finding objects in an image or video stream. warnings to the user based on the proximity of the object.
These tasks are frequently combined in applications such as
robotics, autonomous cars, and surveillance systems in the Our prototype has been developed to aid individuals with
field of computer vision. Identifying and detecting objects visual impairment by providing them with warning alerts
requires a number of steps, which include: regarding any obstacles in their path. In order to achieve
this, we require the ability to determine the distance between
1. Training Information: To learn how to identify and
the person and the obstacle in real-time scenarios. When an
locate various items in a picture, object recognition
object is detected, a rectangular box is generated around it.
models need a lot of labeled training data. Typically,
If the object occupies a significant portion of the frame, we
this training data The <text> comprises pictures and
use certain constraints to determine an approximate distance
annotations that provide detailed explanations about
between the object and the individual. We utilize code to
the position and type of each element depicted in the
identify objects and provide information regarding their
picture.
location and distance.
2. Feature extraction: This stage, which entails removing
significant features from the unprocessed picture data, (d) Voice Generation Module: Voice generation modules
is one of the most crucial in object detection. A common are an essential component of a real-time object detection
method for feature extraction in object detection is to system with distance and voice alerts for blind people. These
use a convolutional neural network (CNN) that has modules are responsible for converting warning messages
already been trained on a significant quantity of picture into speech to alert the user of potential obstacles or hazards.
data, such as MobileNet, since it can extract relevant There are several text-to-speech (TTS) software libraries and
information from photos. frameworks that can be used to generate speech in real-time.
3. Feature Fusion: Using feature fusion can increase the Some of the commonly used TTS modules include:
object detection model’s accuracy. To capture both These TTS modules can be integrated into the object detection
high-level and low-level elements of the image, this system to provide real- time voice alerts to the user. The
entails integrating the features retrieved from various system can use pre-recorded warning messages or generate
layers of the CNN. new messages based on the type and distance of the detected
4. Dimension Reduction: Because the feature maps object. The TTS module can then convert these messages into
produced by the CNN are sometimes very large, a speech and play them back through a speaker or headphones
method known as dimension reduction is employed to for the user to hear.
minimize the number of features while maintaining the Pyttsx3 is a Python module that converts text to speech.
highest level of detail. Frequently used for dimension It is a straightforward tool that works by calculating the
reduction, researchers often employ Principal approximate distance every time an item is detected and
Component Analysis (PCA) as a technique. displaying the corresponding text on the screen using the
5. Training the Classifier: After extracting, fusing, and cv2 library and the cv2.putText() function. To recognize any
reducing the dimensions of the features, we train a buried text in an image, the system utilizes Python-tesseract,
classifier to determine if each area of the image contains which is an OCR (Optical Character Recognition) tool that
an object of interest. One can use various machine scans and analyzes the image to detect any text content and
learning algorithms, such as logistic regression and encode it in a computer-readable format. Once the text is
support vector machines (SVM), to accomplish this recognized, it is linked to pyttsx to generate audio commands
task. as output. For instance, if an object is too close, the system
6. Object Detection Model: Lastly, the characteristics can generates a voice warning that says, “Warning: The object
be used to train a single-shot detector (SSD) model that (class of object) is too close to you,” while a voice saying,
identifies objects in the image using the classifier. In “The object is at a safe distance” is generated if the object is
order to forecast the likelihood that an object will be at a safe distance. The system makes use of various libraries,
such as engine.io,pyttsx3, PyTorch, pytesseract.
Blind People Assistant: Real-Time Objects Detection and Distance Estimation with Voice Feedback 233
5. Single Shot Multi Box Detection as their output. The purpose of these feature maps is to
identify items with varying sizes and proportions. For
(SSD) each point in the feature maps, we construct a set of
The SSD (Single Shot Multi-Box Detection) architecture, default bounding boxes with varying scales and aspect
which is a model based on deep learning, quickly and ratios. Predict the final bounding boxes using these
accurately detects objects in images. It is more efficient and default boxes as anchors.
quicker than two-stage detectors since it only needs one pass • Layers for Bounding Box Regression: These layers
over the input picture to detect objects; this is because it is a forecast the Offsets are applied to every default box to
one-stage detector. generate the final projections for the bounding boxes.
• Object Class Prediction Layers: These layers utilise
5.1 Mobile Net
object class probabilities for each default box to make
When it comes to embedded vision and mobile applications, predictions. The NMS layer eliminates overlapping
one prominent convolutional neural network (CNN) bounding boxes and retains only the most certain
architecture is MobileNet. It is perfect for devices with predictions; it is part of the Non-Maximum Suppression
limited processing power because it is both computationally approach.
efficient and has a minimal memory footprint. A depth-wise
• The output layer returns the last set of predicted bounding
convolution followed by a point-wise convolution is the basis
boxes and class probabilities. The SSD network built
of the MobileNet design. Unlike pointwise convolution,
on MobileNet employs a mix of convolutional layers,
which combines the output of depth-wise convolution for all
default boxes, and feature extraction to identify objects
input channels, depth-wise convolution applies a single filter
in images.
to each input channel independently. A “bottleneck layer,”
an additional component of the MobileNet architecture,
Table 37.1 The architecture of the single shot detector (ssd)
lowers the computational cost of depth-wise convolution by with mobilenet as the base feature extractor
lowering the number of input channels.
Convolution Size No of Output Size (W
5.2 SSD (Single Shot Detector) + MOBILENET Layer (W x H x D) Bounding x H x D)
Boxes
The SSD (Single Shot Detector) technique is frequently
Input 300 x 300 x 3 – 300 x 300 x 3
employed for object detection due to its speed and accuracy.
Conv2D 3x3 300 x 300 x 32 – 300 x 300 x 32
In contrast, MobileNet is an architecture for neural networks
developed with mobile and embedded devices in mind. It Conv2D 3x3 150 x 150 x 64 – 150 x 150 x 64
is ideal for real-time applications on devices with limited Conv2D 3x3 75 x 75 x 128 – 75 x 75 x 128
processing capabilities because of its lightweight architecture, Conv2D 3x3 38 x 38 x 256 – 38 x 38 x 256
which prioritises low latency and low power consumption.
Conv2D 3x3 19 x 19 x 512 – 19 x 19 x 512
The combination of SSD and MobileNet allows for accurate,
Conv2D 3x3 10 x 10 x 512 – 10 x 10 x 512
real-time object identification on embedded and mobile
devices. Conv2D 3x3 5 x 5 x 512 – 5 x 5 x 512
Conv2D 3x3 3 x 3 x 512 – 3 x 3 x 512
5.3 Working of Layers in the MobileNet-based Conv2D 1x1 1 x 1 x 512 3 1 x 1 x 3,072
SSD Network
CONV2D 1X1 1 x 1 x 256 6 1 x 1 x 1,536
The input layer receives the image and resizes it to a CONV2D 1X1 1 x 1 x 128 6 1 x 1 x 768
predetermined size, typically 300x300. Layers of the Base
CONV2D 1X1 1X1X128 6 1 x 1 x 768
Network: Convolutional and pooling layers make up the
base network, which is responsible for feature extraction. CONV2D 1X1 1 X 1 X 128 6 1 x 1 x 768
The foundational network of an SSD network that uses CONV2D 1X1 1 X1 X 128 6 1 x 1 x 768
MobileNet is usually a pre-trained MobileNet design. The
detection network employs the convolutional layers of the • Convolution layer: A type of neural network layer that
base network to retrieve features from its output. To better applies a filter to the input data to extract features.
capture the image’s finer details, these layers often use lower • Size: The dimensions of the input data in width, height,
kernel sizes. and depth (number of channels).
• Feature Maps at Multiple Scales: Convolutional layers • Number of bounding boxes: The quantity of pre-made
produce feature maps with varying spatial resolutions boxes utilised for object detection, each with its own
234 Algorithms in Advanced Artificial Intelligence
unique scale and aspect ratio,.The convolution operation • Object instance segmentation: In addition to object
on the input data produces a feature map with a number of bounding boxes, the COCO dataset also includes
dimensions, which determines the output size. This table segmentation masks for each object instance, allowing
details the SSD MobileNet architecture’s convolution for more precise object localization and segmentation.
layers, along with their input and output sizes, the
number of bounding boxes generated, and other relevant 7. Loss Function
information. The top row of the table shows an input
image to the network, which is a 300 x 300 x 3 (width To find objects in images, object detection uses a loss
x height x depth) image, where depth represents the function to measure how different the model’s predicted class
number of color channels. The rows that follow show labels and bounding boxes are from the real ones. During
each convolutional layer of the network. A feature map training, the goal of the loss function is to decrease this
with fewer spatial resolutions and more channels is mismatch so that the model can accurately detect objects in
produced by each convolution layer by applying a set of images. The localization loss measures how much an object’s
learnable filters to the input. Layer after layer, the feature actual bounding box differs from its projected bounding
maps take on different sizes. Seven 1x1 convolution box. In contrast, the classification loss measures how much
layers make up the other seven layers of the table; they an object’s actual class label differs from its projected class
are responsible for producing the projected class scores label.
and bounding box offsets for each anchor box per feature Object categories: The COCO dataset
map location. Since these layers process feature maps
with varying spatial resolutions, the quantity of bounding 𝐿(𝑥, 𝑐, 𝑙, 𝑔) = 1 (𝐿𝑐𝑜𝑛𝑓 (𝑥, 𝑐) + 𝛼𝐿𝑙𝑜𝑐 (𝑥, 𝑙, 𝑔) ) (1)
𝑁
boxes produced by them varies. The final convolution
contains 80 object categories, including people, animals,
layer’s output measures 1 x 1 x 3,072, which is the same
vehicles, and household items.
as the anticipated class scores and bounding box offsets
for all 8732 network anchor boxes. Object annotations: Each image in the COCO dataset is
annotated with object bounding boxes, segmentation masks,
and category labels for all objects present in the image. The
6. Coco Dataset annotations are in JSON format.
COCO (Common Objects in Context) is a popular large-scale The classification loss (Lconf), found in Eq. (1), measures the
dataset for object detection, segmentation, and captioning discrepancy between the anticipated and ground truth class
applications. With approximately 2.5 million object instances labels for the matched default boxes. This metric quantifies
categorized into 80 distinct object categories, it comprises the model’s ability to identify and label picture items. For
more than 330,000 photos.The COCO dataset’s salient each set of matched default boxes, the localization loss (Lloc)
characteristics for object detection are: measures how far the predicted bounding boxes deviate
• Image diversity: The COCO dataset includes images from the ground truth bounding boxes. By forecasting the
with a wide range of object sizes, shapes, and occlusion coordinates of the bounding boxes, it evaluates the model’s
levels, as well as images with multiple objects and accuracy in item localization in the image. To make the loss
complex scenes. consistent across all matching default boxes, we utilise 1/N.
We add up the two losses to form a single loss function, and
then we back propagate the final loss via the network to
update the model’s parameters.
8. Tensor Flow
With TensorFlow, you can build a deep learning model
to identify objects in photos and videos. Modern object
detection models trained on massive datasets are available
via the TensorFlow Object Detection API. These include
Mask R-CNN, SSD, and Faster R-CNN. The application
programming interface (API) gives tools for evaluating
and deploying trained models and makes it easier to
customise models. Applications including self-driving cars,
surveillance, and robotics have led to its widespread adoption
Fig. 37.1 Cocodata set examples across industries.
Blind People Assistant: Real-Time Objects Detection and Distance Estimation with Voice Feedback 235
9. Experiments and Results A warning has been given since the final distance between
the object and the webcam frame is 0.9 units, which is quite
close. “Warning: Bed is at a safer distance.” is the warning
message that appears once the system’s voice output identifies
the object as a bed.
References
1. Harish Adusumalli, D. Kalyani, R. Krishna Sri, M. Pratapteja,
Fig. 37.9 Output snapshot of accuracy of object TV over time P V R D Prasada Rao “Face Mask Detection Using OpenCV”.
In IEEE, 2021.
11. Conclusion 2. Ayushi Sharma, Jyotsna Pathak, Muskan Prakash, J N Singh,
“Object Detection using OpenCV and Python”. In IEEE, 09
Finally, a technological advancement that can help the March 2022.
visually impaired immensely is a blind person assistant that 3. P Viola and M Jones, “Rapid object detection using a boosted
uses real-time object-detection-with-distance and audio cascade of simple features”, Proceedings of the 2001 IEEE
alarms. Machine learning methods and computer vision Computer Society Conference on Computer Vision and
techniques are used by the system to detect and localise Pattern Recognition (CVPR 2001), December 8-14, 2001.
objects in the user’s surroundings. The system issues 4. Dr. S.V. Viraktamath, Madhuri Yavagal, Rachita Byahatti,
“Object Detection and Classification using YOLOv3”.
audible warnings when it detects possible barriers.Object
InIJERT, February-2021
detection libraries and text-to-speech software are examples
Blind People Assistant: Real-Time Objects Detection and Distance Estimation with Voice Feedback 237
5. Zhang, Y., Chen, Y., Zhang, Y., & Zhang, Y. (2021). Real-time 13. YacineMessai, Kheireddine Chara, FawziSrairi “Object
object detection with voice feedback using sensors-based: A Tracking Platform for Color Object Detection using Genetic
literature review. Sensors, 21(3), 777. Algorithm Optimization”, 2020.
6. Abdelrahman Abdou, Sherif Abdelazeem, and Mahmoud 14. K.Vijiyakumar,K.Ajitha, A.Alexia, M.Hemalashmi,
Refaat(2021) https://www.mdpi.com/2076- 3417/11/16/7342. S.Madhumitha “Object Detection For Visually Impaired
7. Shreyas N Srivatsa, Amruth, Sreevathsa, Vinay , Mr. Elaiyaraja, People Using SSD Algorithm”, 2020.
“Object Detection using Deep Learning with OpenCV and 15. Zhang Qian, Liu Xiao-jun “Video Image Fire Recognition
Python”. In IRJET, JAN 2021 Based on Color Space and Moving Object Detection”, 2020.
8. Priyal Jawale, Hitiksha Patel, Nivedita Rajput, Prof. Sanjay 16. Sunit Vaidya, Naisha Shah, Niti Shah, Prof. Radha
Pawar , “Real-Time Object Detection using TensorFlow”. In Shankarmani “Real-Time Object Detection for Visually
IRJET, Aug 2020 Challenged People”, 2020.
9. K.Vijiyakumar,K.Ajitha, A.Alexia,S.Madhumitha(2020): 17. Hao Shi, Qi Peng, Jiachen Yang, Xudong Bai, YiqiZhuang
Object detection using SSD. “A Practical ROI and Object Detection Method for Vision
10. Chia-Hung Yeh, Chu-Han Lin,Li-Wei Kang, ChihHsiang Robot”, 2020.
Huang, Min-Hui Lin, Chuan- Yu ChangChuaChin Wang 18. Mr. AkshayWankhade, Prof. Pramila M. Chawan“Design and
“Lightweight Deep Neural Network for Joint Learning of Deployment of an Online ShoppingPortal for the Color Blind
Underwater Object Detection and Color Conversion”, 2021. People”, 2019.
11. Martin Stancel, Branislav Mados, Martin Chovanec, Peter 19. Ashwani Kumar, S S Sai Satyanarayana Reddy, Vivek
BalazHybrid “Object Detection Using Domain-specific Kulkarni “An Object Detection Technique For Blind People in
Datasets,” 2021. Real-Time Using Deep Neural Network”, 2019.
12. Manuel G. Forero, Julián Ávila-Navarro, and Sergio Herrera-
Note: All the figures and table in this chapter were designed by the
Rivera “New Method for Extreme Color Detection in Images”,
author.
2020.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
238 Algorithms in Advanced Artificial Intelligence
Abstract: Telemedicine has revolutionized the healthcare industry by enabling remote diagnosis and therapy and utilizing
image digitization in the medical field. However, ensuring the security of multimodal medical images in telemedicine presents
unique challenges. This study uses different encryption standards as the underlying security mechanism to protect multimodal
medical images. The use of modular encryption standards can effectively ensure the security, confidentiality and integrity
of transmitted multimodal images. To prevent unauthorized persons from accessing encrypted modules, it is also important
to ensure that sensitive multimodal medical images can only be viewed and edited by authorized users. The main objective
of the work is to briefly encapsulate and evaluate the different algorithms inherent in each methodology, with a focus on
various access control mechanisms to ensure multimodal medical image privacy in telemedicine, thereby minimizing the risk
of unauthorized disclosure or manipulation.
Keywords: Telemedicine, Multi-modality medical images, Security, Modular encryption standards, Confidentiality, Integrity,
Access control, Advanced encryption algorithms, Machine learning
DOI: 10.1201/9781003529231-38
Standard Encryption Methodologies to Process Multi-Modality Medical Images for Diagnosing in Telemedicine 239
imaging modalities offer unique information about various treatment planning, and monitoring of various medical
aspects by delving into the intricacies of the human body, conditions, providing a comprehensive view of the patient’s
healthcare professionals can gain a more comprehensive health and facilitating informed decision-making by
understanding of a patient’s well-being. healthcare professionals.
Here are some commonly used imaging modalities in the The security of health information (HI) is an ongoing process
arena of medical imaging: that requires constant review and adaptation to keep up with
(a) X-ray: X-ray imaging uses ionizing radiation to produce changes in healthcare environments and technology. Small
images of bones, tissues, and organs. It is commonly used for healthcare centers face challenges in recognizing threats and
examining fractures, detecting abnormalities in the chest, and securing HI. The research aims to help healthcare practices
evaluating dental conditions [6]. prepare for these challenges and provide suitable security
approaches through effective risk assessment [15].
(b) Computed Tomography (CT): CT scans employ a
combination of X-ray imaging and advanced computer Table 38.1 Comparison of multi-modality of diagnostic
processing methods to produce precise cross-sectional imaging
representations of the body. CT scans are useful for detecting
Modality Images Characteristics Advantages
tumors, evaluating injuries, and providing precise anatomical
information [7]. X-ray Detect Detects
Features and Fractures
(c) Magnetic Resonance Imaging (MRI): Magnetic Resonance abnormalities in and
Imaging (MRI) employs a robust magnetic field and radio bone Positions abnormalities
waves to produce intricate visuals of the internal structures in bones.
within the body. It is particularly adept at capturing precise
images of soft tissues such as the brain, muscles, and organs,
CT Offers Concise
and is commonly used for neurological, musculoskeletal, and
comprehensive scans with
abdominal imaging. information superior
(d) Ultrasound: Ultrasound imaging utilizes high-frequency about dense spatial
structures, imaging
sound waves to create real-time visuals of organs and tissues,
particularly precision.
providing dynamic imaging capabilities. It is frequently used bones.
for examining the abdomen, monitoring pregnancies, and Exceptional for
guiding minimally invasive procedures [8]. clearly outlining
skeletal
(e) Positron Emission Tomography (PET): PET scans features. [ 11 ]
necessitate the introduction of a small quantity of radioactive
MRI Reveals details Enhanced
material into the body via injection. The emitted positrons are about abnormal resolution
detected, enabling the creation of images that show the body’s soft tissues. showcasing
metabolic activity. PET scans are valuable in oncology for Widely used anatomical
detecting tumors and assessing treatment response. in confidential intricacies.
clinical settings
(f) Single-Photon Emission Computed Tomography (SPECT): for medical
SPECT imaging uses injected radioactive tracers and examinations.
specialized cameras to produce 3D images of organ function. PET It scans High
It is often used in cardiology and neurology to assess blood the brain Sensitivity
flow, diagnose certain diseases, and evaluate organ function & provides & high
[9]. functional penetration
information. depth.[14]
By combining information from multiple imaging modalities, Enables
healthcare professionals can obtain a more comprehensive recording
and detailed understanding of a patient’s condition. For variations in
normal brain
example, a [10] combination of CT and PET scans, known activity and
as PET-CT, allows for the localization of abnormal metabolic symptoms
activity seen in PET scans within the anatomical context of various
provided by CT scans. diseases.
Modality Images Characteristics Advantages coordination of dispersed computing using mobile devices.
Although CE can provide major benefits such as extended
SPECT A non-intrusive High
approach Sensitivity
battery life and increased storage capacity, its adaptability,
involves & high flexibility, and security issues remain important impediments.
capturing penetration [18] Health information security (HIS) is an iterative
cross-sectional depth technique that undergoes innovative adjustments as medical
images using care contexts evolve. It is crucial to evaluate the efficacy and
radiotracer [12],
revealing the
applicability of HIS security systems and procedures in light
organization of new developments. To ensure HIS security, a thorough
inside the risk assessment and implementation of appropriate security
human body. measures are imperative [19].
Ultrasound Utilizes high- High Spatial
frequency Resolution
sound waves and low cost.
for diagnostic
information,
offering both
qualitative and
quantitative
insights.[13]
Table 38.4 Encryption of medical images using chaotic maps Reference Chaotic Maps Encryption Method
Reference Chaotic Maps Encryption Method Id Used
Id Used [53] SHA-256 hash function Initial parameter generation
[33] 2D Zaslavsky, 2D Live wireless capsule for Logistic Tent system
Logistic endoscopy encryption [54] Arnold's cat map The integration of ECC
[34] The Extended Bimodal Stream cipher applied to transformation. and ElGamal encryption
Logistic function. encrypt medical images. techniques.
[35, 36] 4D hyperbolic Sine, Medical Image [55] Henon map Combined with number
2D Sine Logistic cryptosystem with verified theory for encryption
modulation chaotic properties [56] Applying the Skew Integrated for the purpose
[37] Logistic-sine Software platform: Mod Tent Map alongside of encryption.
operation, Hardware matrix cyclic shift
platform: XOR operation operations.
[38] Combinatorial of sine Chaotic economic map [57] The integration of a Unified for the purpose of
and cosine functions optimization for medical logistic coupled map encryption.
image encryption in 2D and a linear
congruential generator.
[39] Logistic Medical image encryption in
16x16 blocks [58] Galois field Improvement of the coupled
multiplication hyper-chaotic Chen 6D
[40] Logistic, Kent Stream cipher technology system.
for medical image
encryption [59] Genetic algorithms Integration of artificial
intelligence (AI) technology
[41] Default double-sine Medical image bit plane and chaotic systems for
encryption encrypting medical images.
[42] Logistic Medical image encryption [60] Neural network The amalgamation of
with classical permutation- chaotic systems and AI
diffusion architecture technology for securing
[43] Three different Logistic Three-step permutation medical image data.
maps diffusion-p ermutation [61] 3D Cellular automata Combining chaotic systems
structure for medical image (CA) and AI technology for the
encryption encryption of medical
[44] Three different Logistic Permutation-diffusion- images.
maps permutation structure for [62] 2D Chaotic map Encryption, information
medical image encryption entropy calculation, and
[45] 2D Arnold cat, 1D Combined chaotic parameter optimization.
Logistic, 2D Henon systems for medical image [63] Logistic, Tent Confusion and diffusion
encryption using chaotic maps,
[46] 3D Intertwining Chaotic maps used optimization with
Logistic, Logistic-sin, individually for medical grasshopper optimization
enhanced chaotic image encryption [64] A quantum encryption Quantum encryption with
economic, time-delay framework based on chaotic maps for medical
chaotic Gray code integrated images
[47] DNA-based computing Indestructible cryptographic with chaos maps.
and dynamical systems system leveraging DNA
computing.
Table 38.5 Encryption methods, optimization methods,
[48], [ 49] Chaotic system with Image encryption scheme
Feature
S-box combining chaotic system
and S-box Reference Encryption Method Optimization Method/
[50] 2D Logistic Combined with RC5 Features
algorithm for encryption [65] Blowfish algorithm OFP Signcryption: Key
[51] Chaotic system based The combination of ECC Optimization
on the Mersenne and ElGamal encryption. [66] Cosine number transform CosineCrypt
Twister algorithm.
[67] NDP Structure NICSP Permutation
[52] Hash function (512 bits) Chaotic parameter and
[68] Full, Middle-full, Selective Three modes for medical
initial value generation
modes image encryption: full,
middle-full, selective
244 Algorithms in Advanced Artificial Intelligence
Reference Encryption Method Optimization Method/ Reference Secret Sharing Proposed Scheme
Features Scheme
[69] Quaternion Feistel Improved encryption of [79] Shamir secret A cloud storage scheme
structure medical images using sharing scheme that utilizes secret sharing
Quaternion Feistel techniques.
structure [80] Visual Halftone visual cryptography
cryptography based on Region of non-interest
[70] Optical encryption TrichromaticCrypt
(RONI)
method
[81] N/A Medical image segmentation
[71] Genetic Algorithms-b GeneticGuard method with endpoint
ased approach coordinates reconstruction
5. S. Hussain et al., “Modern Diagnostic Imaging Technique 18. W. Zhang, C. A. Gunter, D. Liebovitz, J. Tian, and B. Malin,
Applications and Risk Factors in the Medical Field: A “Role prediction using Electronic Medical Record system
Review,” vol. 2022, pp. 1–19, Jun. 2022, doi: https://doi. audits,” AMIA Annual Symposium proceedings. AMIA
org/10.1155/2022/5164970. Symposium, vol. 2011, pp. 858–67, 2011, Accessed: Jul. 05,
6. J. E. Ottenhoff, C. Thom, M. Kongkatong, M. Hewitt, and J. 2023. [Online]. Available: https://www.ncbi.nlm.nih.gov/
Phillips, “A Narrative Review of the Uses of Ultrasound in pmc/articles/PMC3243238/
the Evaluation, Analgesia, and Treatment of Distal Forearm 19. R. Sathya Prabha, K Kanagasabapathi, K. Sajeeth, and
Fractures,” vol. 63, no. 6, pp. 755–765, Dec. 2022, - Journal of M. Aishwarya, “Health Information Sharing in Cloud
Emergency Medicine. Environment Using Modular Encryption Standard,” Jan.
7. M. Shyni and Dr. Chitra E, “A comparative study of X-ray and 2023, doi: https://doi.org/10.3233/atde221238.
CT images in COVID-19 detection using image processing 20. M. Paul, L. A. Maglaras, Mohamed Amine Ferrag, and Iman
and deep learning techniques,” vol. 2, pp. 100054–100054, Almomani, “Digitization of healthcare sector: A study on
Jan. 2022, - ScienceDirect. privacy and security concerns,” Feb. 2023, doi: https://doi.
8. C. K. Sen, Subhadip Ghatak, Surya Gnyawali, S. Roy, and org/10.1016/j.icte.2023.02.007.
G. M.Gordillo, “Cutaneous Imaging Technologies in Acute 21. K. Batko and Andrzej Ślęzak, “The use of Big Data Analytics
Burn and Chronic Wound Care,” vol. 138, pp. 119S128S, Sep. in healthcare,” vol. 9, no. 1, Jan. 2022, doi: https://doi.
2016, doi:https://doi.org/10.1097/prs.0000000000002654. org/10.1186/s40537-021-00553-4.
9. Gianina Crișan, G. Andrieș, Calin Cainap, and Vasile Chiș, 22. M. Shabbir et al., “Enhancing Security of Health Information
“Radiopharmaceuticals for PET and SPECT Imaging: A Using Modular Encryption Standard in Mobile Cloud
Literature Review over the Last Decade,” vol. 23, no. 9, Computing,” vol. 9, pp. 8820–8834, Jan. 2021, doi: https://
pp. 5023–5023, Apr. 2022, doi: https://doi.org/10.3390/ doi.org/10.1109/access.2021.3049564.
ijms23095023. 23. Thsen Alouani, “Breaking (and Fixing) Channel-based
10. L. K. Griffeth, “Use of Pet/Ct Scanning in Cancer Patients: Cryptographic Key Generation: A Machine Learning
Technical and Practical Considerations,” vol. 18, no. 4, pp. Approach,” Aug. 2022, doi: https://doi.org/10.1109/
321–330, Oct. 2005, doi: https://doi.org/10.1080/08998280. dsd57027.2022.00058.
2005.11928089. 24. Y. Sain, “Review on Compression of Medical Images using
11. Y. Gong, “Decompose X-ray Images for Bone and Soft Various Techniques,” International Journal of Engineering
Tissue,” arXiv.org, 2020. https://arxiv.org/abs/2007.14510 Research & Technology, vol. 3, no. 8, Sep. 2014, doi: https://
(accessed Jul. 06, 2023). doi.org/10.17577/IJERTV3IS080880.
12. Haubner, “Radiotracer-based strategies to image angiogenesis,” 25. R. Kumar et al., “An Integration of blockchain and AI for
The quarterly journal of nuclear medicine : official publication secure data sharing and detection of CT images for the
of the Italian Association of Nuclear Medicine (AIMN) hospitals,” vol. 87, pp.101812–101812, Jan. 2021, doi:https://
[and] the International Association of Radiopharmacology doi.org/10.1016/j.compmedimag.2020.101812.
(IAR), vol. 47, no. 3, 2015,https://pubmed.ncbi.nlm.nih. 26. “Development of modified RSA algorithm using fixed
gov/12897710 mersenne prime numbers for medical ultrasound imaging
13. M. Dietzel et al., “Fusion of dynamic contrast-enhanced instrumentation,” Computer Assisted Surgery, 2019. https://
magnetic resonance mammography at 3.0T with X-ray www.tandfonline.com/doi/full/10.1080/24699322.20
mammograms: Pilot study evaluation using dedicated semi 19.1649070
automatic registration software,” Aug.2011, doi: https://doi. 27. Q. N. Natsheh, B. Li, and A. G. Gale, “Security of Multi
org/10.1016/j.ejrad.2011.04.017 frame DICOM Images Using XOR Encryption Approach,”
14. P. A. Segura Chávez et al., “Love Wave Sensor with vol. 90, pp. 175–181, Jan.2016, doi: https://doi.org/10.1016/j.
High Penetration Depth for Potential Application in Cell procs.2016.07.018.
Monitoring,” Biosensors, vol. 12, no. 2, p. 61, Feb. 2022, doi: 28. Ranjith Kumar. M and M. K. Viswanath, “A symmetric
https://doi.org/10.3390/bios12020061. medical image encryption scheme based on irrational
15. P. Kumar and S.-H. Lee, “Security Issues in Healthcare numbers,” Jan. 2018, doi: https://doi.org/10.4066/
Applications Using Wireless Medical Sensor Networks: A biomedicalresearch.29-17-1317.
Survey,” vol. 12, no. 1, pp. 55–91, Dec. 2011, doi: https://doi. 29. Dorgham O, Al-Rahamneh B, Almomani A, Khatatneh KF
org/10.3390/s120100055. (2018) Enhancing the security of exchanging and storing
16. Sharma K, Agrawal A, Pandey D, et al (2022) RSA based DICOM medical images on the cloud. Int J Cloud Appl
encryption approach for preserving confidentiality of big data. Comput (IJCAC) 8(1):154–172. https://doi.org/10.4018/
Journal of King Saud University - Computer and Information IJCAC.2018010108
Sciences 34:2088–2097. https://doi.org/10.1016/j. 30. Massoud Sokouti, A. Zakerolhosseini, and Babak
jksuci.2019.10.006 Sokouti,“Medical Image Encryption: An Application for
17. K. Anusudha, “A Theoretical Approach to Secure Medical Improved Padding Based GGH Encryption Algorithm,” vol.
Images by Combining Cryptography and Watermarking 10, no. 1, pp. 11–22, Oct. 2016, doi: https://doi.org/10.2174/1
Techniques,” vol. 7, no. 3, pp. 69–77, Jul. 2020, doi: https:// 874431101610010011.
doi.org/10.30726/esij/v7.i3.2020.73014.
Standard Encryption Methodologies to Process Multi-Modality Medical Images for Diagnosing in Telemedicine 247
31. A. Al-Haj, G. A. Abandah, and N. Hussein, “Crypto‐based Secure Medical Image Protection Scheme,” vol. 2015, pp.
algorithms for secured medical image transmission,” vol. 9, 1–11, Jan. 2015, doi: https://doi.org/10.1155/2015/913476.
no. 6, pp. 365–373, Nov. 2015, doi:https://doi.org/10.1049/iet 45. K. Jain, Aravind Aji, and P. Krishnan, “Medical Image
ifs.2014.0245. Encryption Scheme Using Multiple Chaotic Maps,” vol.
32. B. Zhang, Bahbibi Rahmatullah, Shir Li Wang, A. A. Zaidan, 152, pp. 356–364, Dec. 2021, doi: https://doi.org/10.1016/j.
B. B. Zaidan, and P. Liu, “A review of research on medical patrec.2021.10.033.
image confidentiality related technology coherent taxonomy, 46. K. Kiran, H. L. Gururaj, Meshari Almeshari, Yasser Alzamil,
motivations, open challenges and recommendations,” vol. R.Vinayakumar, and K. V. Sudeesh, “Efficient SCAN and
82, no. 14, pp.21867–21906, Aug. 2020, doi: https://doi. Chaotic Map Encryption System for Securing E-Healthcare
org/10.1007/s11042-020-09629-4. Images,” vol. 14, no. 1, pp. 47–47, Jan. 2023, doi: https://doi.
33. R. Hamza, Z. Yan, Sung Wook Baik, Paolo Bellavista, and org/10.3390/info14010047.
Faiza Titouna, “A privacy-preserving cryptosystem for IoT 47. J. Zheng, Z. Luo, and Z. Tang, “An Image Encryption
E-healthcare,” vol. 527, pp. 493–510, Jul. 2020, doi: https:// Algorithm Based on Multichaotic System and DNA
doi.org/10.1016/j.ins.2019.01.070. Coding,” vol. 2020, pp. 1–16, Sep. 2020, doi: https://doi.
34. Cortés IE, Venegas O, Gómez HW (2022) A Symmetric/ org/10.1155/2020/5982743.
Asymmetric Bimodal Extension Based on the Logistic 48. Farah MB, Farah A, Farah T (2019) An image encryption
Distribution: Properties, Simulation and Applications. scheme based on a new hybrid chaotic map and optimized
Mathematics 10:1968–1968. https://doi.org/10.3390/ substitution box. Nonlinear Dynamics:1-24. https://doi.
math10121968 org/10.1007/s11071-019-05413-8
35. Hua Z, Zhou Y, Pun CM, ChenCP(2015) 2D sine logistic 49. Farah MB, Guesmi R, Kachouri A, Samet M (2020) A
modulation map for image encryption. Inf Sci 297:80–94. new design of cryptosystem based on S-box and chaotic
https://doi.org/10.1016/j.ins.2014.11.018 permutation. Multimed Tools Appl:1-22. https://doi.
36. Liu J, Ma Y, Li S, Lian J, Zhang X (2018) A new simple chaotic org/10.1007/s11042-020-08718-8.
system and its application in medical image encryption. 50. Shahzadi R, Anwar SM, Qamar F, Ali M, Rodrigues JJ
Multimed Tools Appl 77(17):22787–22808. https://doi. (2019) Chaos based enhanced RC5 algorithm for security
org/10.1007/s11042-017-5534-8 and integrity of clinical images in remote health monitoring.
37. Hua Z, Yi S, Zhou Y (2018) Medical image encryption IEEE Access 7:52858–52870. https://doi.org/10.1109/
using high-speed scrambling and pixel adaptive diffusion. ACCESS.2019.2909554
Signal Process 144:134–144. https://doi.org/10.1016/j. 51. H. Liu, A. Kadir, and Y. Li, “Asymmetric color pathological
sigpro.2017.10.004 image encryption scheme based on complex hyper chaotic
38. “Confidential storage of medical images – a chaos-based system,” ResearchGate, Apr. 2016. Asymmetric color
encryption approach,” International Journal of Cloud pathological image encryption scheme based on complex
Computing, 2018. hyper chaotic system
39. Chuman T, Hitoshi Kiya (2022) Security Evaluation of 52. Meiliana Sumagita, Imam Riadi (2018) Analysis of Secure
Block-based Image Encryption for Vision Transformer Hash Algorithm (SHA) 512 for Encryption Process on Web
against Jigsaw Puzzle Solver Attack. 2022 IEEE 4th Global Based Application. International Journal of Cyber-Security
Conference on Life Sciences and Technologies (LifeTech). and Digital Forensics 7:373–382
https://doi.org/10.1109/lifetech53646.2022.9754937 53. Rahul Rahul, K. Kuppusamy, and A. Senthilrajan, “Bio
40. WangW,SiM,PangY,RanP,WangH,JiangX,LiuY,WuJ,WuW, Metric Based Colour Image Encryption using Multi Chaotic
Chilam kurtiN, Jeon G(2018) An encryption algorithm Dynamical Systems and SHA 256 Hash Algorithm,”
based on combined chaos in body area networks. ResearchGate, Jun. 29, 2023. Bio Metric Based Colour Image
ComputElectricEng65:282–291. https://doi.org/10.1016/j. Encryption using Multi Chaotic Dynamical Systems and SHA
compeleceng. 2017. 07. 026 256 Hash Algorithm.
41. CaoW,Zhou Y,Chen CP, Xia L(2017) Medical image 54. S. Banerjee and A. Patil, “ECC Based Encryption Algorithm
encryption using edge maps. Signal Process 132: 96–109. for Lightweight Cryptography,” ResearchGate, 2020.
https://doi.org/10.1016/j.sigpro.2016.10.003 55. S. Kanwal et al., “An Effective Color Image Encryption Based
42. M. Parvees, J. Abdul Samath, and B. Parameswaran Bose, on Henon Map, Tent Chaotic Map, and Orthogonal Matrices,”
“Protecting Large Size Medical Images with Logistic Map vol. 22, no. 12, pp. 4359–4359, Jun. 2022, doi: https://doi.
Using Dynamic Parameters and Key Image,” International org/10.3390/s22124359.
Journal of Network Security, vol. 19, no. 6, pp. 984–994, 56. A. Kanso, M. Ghebleh, and M. Bou Khuzam, “A Probabilistic
2016, doi: https://doi.org/10.6633/IJNS.201711.19(6).15. Chaotic Image Encryption Scheme,” Mathematics, vol.
43. Xu L, Gou X, Li Z, Li J (2017) A novel chaotic image 10, no. 11, p. 1910, Jun. 2022, doi: https://doi.org/10.3390/
encryption algorithm using block scrambling and dynamic math10111910.
index based diffusion. Opt Lasers Eng 91:41–52. https://doi. 57. X. Huang, L. Liu, X. Li, M. Yu, and Z. Wu, “A New Two-
org/10.1016/j.optlaseng.2016.10.012 Dimensional Mutual Coupled Logistic Map and Its Application
44. L. Zhang, Z. Zhu, B. Yang, W. Liu, H. Zhu, and M. for Pseudorandom Number Generator,” vol. 2019, pp. –10,
Zou,“Cryptanalysis and Improvement of an Efficient and May 2019, doi: https://doi.org/10.1155/2019/7685359.
248 Algorithms in Advanced Artificial Intelligence
58. Laiphrakpam Dolendro Singh, Rohit Thingbaijam, Kh 69. M. Luis, L. Daniel, A. Isabel, and Alvarado Deicy, “A new
Motilal, and Moatsum Alawida, “Encrypting Multiple Images multimedia cryptosystem using chaos, quaternion theory and
With an Enhanced Chaotic Map,” ResearchGate, 2022. (PDF) modular arithmetic,” Mar. 2023, doi: https://doi.org/10.1007/
Encrypting Multiple Images With an Enhanced Chaotic Map s11042-023-14475-1.
59. B. Zhang and L. Liu, “Chaos-Based Image Encryption: Review, 70. P. Wang, Y. Wang, J. Xiang, and X. Xiao, “Fast Image
Application, and Challenges,” vol. 11, no. 11, pp. 2585–2585, Encryption Algorithm for Logistics-Sine-Cosine Mapping,”
Jun. 2023, doi: https://doi.org/10.3390/ math11112585. Sensors, vol. 22, no.24, p. 9929, Jan. 2022, doi: https://doi.
60. Y. Mao, “Algorithm of Encrypting Digital Image Using Chaos org/10.3390/s22249929.
Neural Network,” vol. 2022, pp. 1–10, Sep. 2022, doi: https:// 71. L. Si, X. Hu, and B. Liu, “Image Matching Algorithm Based
doi.org/10.1155/2022/4160083. on the Pattern Recognition Genetic Algorithm,” ResearchGate,
61. B. Zhang and L. Liu, “Chaos-Based Image Encryption: Mar. 09, 2022. (PDF) Image Matching Algorithm Based on
Review,Application, and Challenges,” Mathematics, vol. the Pattern Recognition Genetic Algorithm.
11, no. 11, p. 2585,Jan. 2023, doi: https://doi.org/10.3390/ 72. D. Li et al., “Hybrid Encrypted Watermarking Algorithm for
math11112585. Medical Images Based on DCT and Improved DarkNet53,”
62. U. Erkan, Abdurrahim Toktas, Feyza Toktas, and Fayadh Electronics, vol. 12, no. 7, p. 1554, Jan. 2023, doi: https://doi.
Alenezi, “2D eπ-map for image encryption,” vol. 589, pp. 770– org/10.3390/electronics12071554.
789, Apr.2022, doi: https://doi.org/10.1016/j.ins.2021.12.126. 73. P. Prasad, Bethel, N. Singh, Vinit Kumar Gunjan, Samad
63. Montassar Aidi Sharif, Keasar Sabah Khalaf, and Mahmoud Baseer, and S. Miah, “Blockchain-Based Privacy Access
Shakir Wahhab, “Digital Communication Based on Image Control Mechanism and Collaborative Analysis for Medical
Security using Grasshopper Optimization and Chaotic Map,” Images,” vol. 2022, pp. 1–7, Jun.2022, doi: https://doi.
ResearchGate, Jul. 05, 2022. org/10.1155/2022/9579611.
64. Bassem Abd-El-Atty, Mohammed Ahmed El-Affendi, and 74. Rothblum RD (2011) Homomorphic Encryption: From
Fathi, “A novel image cryptosystem using Gray code, quantum Private-Key to Public-Key. Lecture Notes in Computer
walks, and Henon map for cloud applications,” vol. 9, no. 1, Science 219–234. https://doi.org/10.1007/978-3-642-19571
pp. 609–624, Jul. 2022, doi: https://doi.org/10.1007/s40747 6_14
022-00829-z. 75. Saci Medileh et al., “A Multi-Key with Partially Homomorphic
65. B. T. Geetha, P. Mohan, A.V.R Mayuri, T. Jackulin, J.L. Encryption Scheme for Low-End Devices Ensuring Data
Aldo Stalin, and Varagantham Anitha, “Pigeon Inspired Integrity,” vol. 14, no. 5, pp. 263–263, Apr. 2023, doi: https://
Optimization with Encryption Based Secure Medical Image doi.org/10.3390/info14050263.
Management System,” vol. 2022, pp. 1–13, Aug. 2022, doi: 76. F. Yu, X. Gong, H. Li, and S. Wang, “Differential cryptanalysis
https://doi.org/10.1155/2022/2243827. of image cipher using block-based scrambling and image
66. V. S. Lima, F. Madeiro, and J. Lima, “Encryption of 3D filtering,” vol. 554, pp 145–156, Apr. 2021, doi: https://doi.
medical images based on a novel multiparameter cosine org/10.1016/j.ins.2020.12.037.
number transform,” ResearchGate, Apr. 2020. Encryption of 77. S. Zhu and C. Zhu, “An Efficient Chosen-Plaintext Attack
3D medical images based on a novel multiparameter cosine on an Image Fusion Encryption Algorithm Based on DNA
number transform. Operation and Hyperchaos,” vol. 23, no. 7, pp. 804–804, Jun.
67. J. Fei et al., “DuMLP-Pin: A Dual-MLP-Dot-Product 2021, doi: https://doi.org/10.3390/e23070804.
Permutation-Invariant Network for Set Feature Extraction,” 78. Mbarek Marwan, Feda Alshahwan, F. Sifou, and H.
Proceedings of the AAAI Conference on Artificial Intelligence, Ouahmane,“Improving the security of cloud-based medical
vol. 36, no. 1, pp. 598–606, Jun. 2022, doi:https://doi. image storage,” ResearchGate, Feb. 2019. Improving the
org/10.1609/aaai.v36i1.19939. security of cloud-based medical image storage | Request PDF.
68. B. Zhang, Bahbibi Rahmatullah, Shir Li Wang, and Z. Liu, 79. E. Salih, “A Simple and Secure Secret Sharing Scheme for
“A plain-image correlative semi-selective medical image IoT,” doi: https://doi.org/10.1109/CSCI58124.2022.00272.
encryption algorithm using enhanced 2D-logistic map,” vol. 80. A. Patel and A. Bakshi, “Secure telemedicine using RONI
82, no. 10, pp.15735–15762, Sep. 2022, doi: https://doi. halftoned visual cryptography without pixel expansion,”
org/10.1007/s11042-022-13744-9. ResearchGate, Apr. 05, 2019. Secure telemedicine using
RONI halftoned visual cryptography without pixel expansion.
Note: All the figures and tables in this chapter were designed by
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Dyslexia, a neurodevelopment condition impacting reading and language abilities, presents notable difficulties
in promptly identifying and implementing effective interventions Traditional methods for diagnosing dyslexia often rely on
subjective assessments and standardized tests, leading to delays in recognition and support. This paper offers an extensive
examination of how deep learning techniques are applied in the domain of detecting and intervening in dyslexia. The integration
of deep learning algorithms into dyslexia research offers promising avenues for more accurate and timely identification of
individuals at risk. By leveraging neural networks and advanced machine learning models, researchers have begun to explore
novel approaches that analyze linguistic patterns, eye-tracking data, brain imaging, and behavioral markers associated with
dyslexia. Furthermore, this paper discusses the potential of deep learning in tailoring personalized interventions for individuals
with dyslexia. These interventions aim to adapt to the specific learning needs of each individual, providing targeted support
and enhancing the effectiveness of remediation strategies. While highlighting the advancements made in utilizing deep learning
for dyslexia, this review also addresses challenges, including data scarcity, model interpretability, and ethical considerations.
Additionally, it proposes future research directions that emphasize collaborative efforts among researchers, educators, and
technology developers to foster the development of robust and accessible tools for dyslexia assessment and intervention.
Keywords: Dyslexia, Deep Learning, Eye-tracking, Neural Networks
1. Introduction
Dyslexia[1], a neurodevelopmental disorder affecting
language and reading abilities, remains a persistent challenge
in educational settings and beyond. Its multifaceted nature,
characterized by difficulties in decoding words, recognizing
sounds, and understanding language, presents hurdles in
both early identification and effective intervention. Despite
concerted efforts in research and educational practices,
Dyslexia[33] continue to impact millions worldwide, Fig. 39.1 Brain Image with Dyslexia
highlighting the pressing need for innovative approaches to
address its complexities. In parallel, the rapid advancements in deep learning, a subset
of artificial intelligence (AI) that mimics the workings of
1
kdvpkvarma@gmail.com; 2cheepurupalliraghuram@gmail.com; 3sivaramakrishna.b@srmap.edu.in, krishna2928@gmail.com
DOI: 10.1201/9781003529231-39
250 Algorithms in Advanced Artificial Intelligence
the human brain through complex neural networks, offer Furthermore, difficulties in rapid automatized naming,
promising avenues for tackling intricate and nuanced orthographic processing[3], and working memory exacerbate
problems. The ability of deep learning models[32] to discern the obstacles related to reading and understanding language
intricate patterns and extract features from vast datasets has shown in below Fig. 39.2.
sparked considerable interest in their potential application to
dyslexia[2], particularly in revolutionizing detection methods 2. Literature Survey
and tailoring interventions to individual needs.
Various approaches for predicting dyslexia include
The paper presents a comprehensive review of the intersection
neuroimaging techniques to identify brain abnormalities,
between dyslexia and deep learning, aiming to explore the
behavioral assessments focusing on reading-related skills,
potential of these cutting-edge technologies in revolutionizing
and machine learning models that analyze linguistic patterns
dyslexia[1] detection and intervention. Through an in-depth
and cognitive features. Combining these diverse methods
examination of existing literature and methodologies, this
offers a comprehensive approach to early detection and
review seeks to elucidate how deep learning models can
intervention for individuals at risk of dyslexia, enhancing
enhance our understanding, detection, and remediation
strategies for dyslexia. the effectiveness of tailored educational strategies. Earlier
Works on These Lines by Various Authors are shown in
The introductory section provides an overview of dyslexia, Table 39.1.
delineating its defining characteristics, prevalence, and the
persisting challenges encountered in timely identification
and effective intervention. Subsequently, the introduction 3. Limitations and Challenges
highlights the fundamental principles of deep learning, The research on dyslexia collectively faces several challenges
elucidating the mechanisms by which neural networks and limitations. Firstly, the heterogeneity of dyslexia poses
process data and showcasing their potential to address difficulties in generalizing findings across diverse populations
intricate problems such as dyslexia. and severity levels. Interdisciplinary approaches, while
This paper endeavors to synthesize existing knowledge, valuable, demand expertise in neuroimaging, linguistics,
identify gaps, and propose future directions in leveraging and psychology, potentially limiting accessibility. Resource-
deep learning to advance dyslexia research. By amalgamating intensiveness is a recurring challenge, particularly in studies
expertise from the fields of neuroscience, education, and AI, involving neuroimaging and longitudinal approaches,
this exploration aims to contribute to the evolving landscape affecting the feasibility of large-scale research. Subjectivity,
of dyslexia studies and ultimately pave the way for more observer bias, and reliance on self-reported information
inclusive and effective approaches in detecting and supporting introduce potential inaccuracies.
individuals with dyslexia. Ethical considerations, especially in emerging areas like AI
applications[5] for dyslexia, present complex challenges in
1.1 Defining Characteristics of Dyslexia balancing ethical practices with empirical evidence. Some
The hallmark of dyslexia[32] encompasses challenges studies oversimplify dyslexia, focusing on specific aspects
in decoding words, accurately recognizing sounds and or proposing theoretical models, limiting the comprehensive
letters, and efficiently comprehending written text. People understanding of the condition. Generalizability issues arise
with dyslexia frequently face challenges in phonological from small sample sizes or specific participant characteristics.
awareness, the skill of manipulating and recognizing sounds Translating neurobiological findings into practical
in spoken words. This ability is essential for connecting interventions remains a complex challenge, highlighting the
sounds to letters during the process of learning to read. gap between neuroscience and real-world applications.
A lack of longitudinal data limits the understanding of the
Reading developmental trajectory of dyslexia over time. Dependency
Phonological on advanced technologies, such as Neuroimaging or AI,
difficulties Writing introduces challenges related to the rapid evolution of these
Dyslexia technologies. Genetic studies may oversimplify the complex
interplay between genetics and environmental factors in
Short term Orientationon in dyslexia. Some studies have a narrow scope, focusing on
memory time and space specific aspects like reading fluency or genetic connectivity,
Visual
perception potentially overlooking other crucial factors.
4. Considering Deep Learning-Based in analyzing complex patterns and data sets, showcasing
potential applications in dyslexia detection[15]. Various
Techniques for Dyslexia studies have employed deep learning techniques to scrutinize
multiple data types associated with dyslexia, including
4.1 Deep Learning Fundamentals linguistic patterns, eye-tracking data, brain imaging, and
Deep learning, a subset of machine learning, utilizes artificial behavioral markers.
neural networks[7] to mimic the cognitive processes of the Data Types and Features Analyzed
human brain when handling data. At its core are artificial
Linguistic Patterns: Explain how deep learning models
neurons, basic units that receive inputs, apply weights,
analyze linguistic patterns, including language structure,
and use activation functions to produce outputs, mirroring
syntax, and semantic relations, to identify potential dyslexia-
biological neurons’ signal transmission. Neural networks
related irregularities in text comprehension and production.
consist of layers—input, hidden, and output—allowing
hierarchical feature learning. Convolutional Neural Networks Eye-Tracking Data: Discuss how deep learning models
[19] (CNNs) excel in image and pattern recognition, making process eye movement data [16] to understand reading
them relevant for analyzing visual aspects in dyslexia behaviors and identify distinctive gaze patterns characteristic
research. Recurrent Neural Networks (RNNs) [19] handle of individuals with dyslexia.
sequential data, potentially useful for language-related Brain Imaging: Focus on research that utilizes Neuroimaging
patterns associated with dyslexia. Training deep learning data to identify structural or functional distinctions in the
models[32] involves backpropagation, adjusting parameters brains of individuals with dyslexia, offering valuable insights
during training, and gradient descent optimization[9] to into neural correlations [17].
minimize errors and enhance accuracy. Deep learning’s
Efficacy and Efficiency of Models Utilizing Deep Learning
strengths include extracting intricate patterns from vast
datasets and solving complex problems, but challenges Assessing the Effectiveness of Deep Learning Models for
like overfitting, interpretability, and data requirements Dyslexia Detection in Comparison to Traditional Assessment
persist. Beyond dyslexia, deep learning finds applications Approaches. Explore studies showcasing the accuracy,
in computer vision, natural language processing, healthcare, sensitivity, and specificity of these models in identifying
and various domains, highlighting its versatility and broad dyslexia based on diverse datasets and features.
impact represented in Fig. 39.3. Identified Diagnostic Markers or Patterns
Discuss potential diagnostic markers or patterns identified
by deep learning models that distinguish individuals with
dyslexia. Highlight any specific linguistic, visual, or
neurological features recognized as reliable indicators of
dyslexia through deep learning analyses.
Challenges and Limitations
Data Availability and Quality
Address challenges related to the availability, size, and
quality of datasets used for training deep learning models for
dyslexia detection, emphasizing the need for comprehensive
and diverse datasets.
Interpretability and Explain ability
Discuss the challenge of interpretability in deep learning
models, emphasizing the importance of understanding how
these models arrive at their conclusions for clinical adoption.
Fig. 39.3 Traditional way and deep learning-based Deep Learning in Personalized Dyslexia Interventions
Diagnosis of Dyslexia patient Adaptive Intervention Strategies Using Deep Learning
Deep learning presents an opportunity to develop adaptive
4.2 Application of Deep Learning in Dyslexia intervention strategies tailored to the individual needs and
Detection learning profiles of those with dyslexia. These models
Utilizing Deep Learning Models for Dyslexia Detection have the capacity to dynamically adjust and personalize
Deep learning models have emerged as powerful tools interventions based on an individual’s response and progress.
Enhancing Dyslexia Detection and Intervention through Deep Learning: A Comprehensive Review and Future Directions 253
Individualized Support and Adaptive Learning Models Ethical Considerations Ethical Implications
Discuss how deep learning models can facilitate the creation Address ethical concerns surrounding personalized
of personalized intervention strategies. Explore the potential interventions, such as data privacy, informed consent, and
for these models to adaptively adjust learning materials, the responsible use of individualized data for intervention
pacing, and methodologies to suit the unique strengths and development. Model Bias and Fairness: Discuss the
weaknesses of each dyslexic individual. importance of ensuring that deep learning models used to
Enhancing Intervention Effectiveness personalize interventions are fair and free from biases that
Highlight the potential impact of personalized interventions might perpetuate disparities or discrimination.
derived from deep learning models in enhancing the Potential Impact and Future Directions Educational
effectiveness of dyslexia interventions. This could Transformation
include improved engagement, better learning outcomes, Explore the potential transformative effect of personalized
and increased retention rates compared to standardized interventions derived from deep learning on educational
interventions. practices for individuals with dyslexia. Future Research
254 Algorithms in Advanced Artificial Intelligence
Avenues: Propose future research directions focusing on Relevance: A high F1 score suggests a model that achieves
refining deep learning models for personalized interventions, precision while effectively capturing relevant instances.
addressing ethical considerations, and conducting
Precision * Recall
longitudinal studies to assess the long-term impact of F1 score = 2 * (4)
personalized interventions. Precision + Recall
Area Under the ROC Curve (AUC-ROC): AUC-ROC
4. Reviews on Performance Metrics assesses the model’s capability to differentiate between
dyslexic and non-dyslexic cases under various threshold
Evaluating the performance of the integrated deep learning settings.
methodologies in dyslexia research and intervention requires
Relevance: A high AUC-ROC score signifies a robust model
careful consideration of relevant metrics. The following are
capable of effective discrimination.
pivotal performance metrics [19]:
AUC – ROC = ∑_(i = 1)^(n – 1) [1/2 * (TPR_i + TPR_
Accuracy: Accuracy evaluates the overall correctness of
(i + 1)) * (FPR_(i + 1) – FPR_i)] (5)
models for dyslexia detection and intervention. It quantifies
the proportion of accurately predicted instances relative to RMSE, portrayed as the standard deviation of residuals
the total instances. (prediction errors), quantifies the accuracy of predictions
across quantitative data.
Relevance: A high accuracy score indicates the effectiveness
of the models in correctly identifying dyslexia and
 i=1 ( xi x̂i )
N 2
implementing personalized interventions. RMSE = (6)
N
(TP + TN)
Accuracy = *100% (1)
The coefficient of determination () is represented as the
(TP + FN + FP + TN)
proportion of variation in the dependent variables explained
Precision and Recall: Precision is a metric that gauges the by the independent variables.
accuracy of positive predictions by assessing the ratio of
RSS
true positive instances to the sum of true positives and false R2 = 1 (7)
positives. Recall, also known as sensitivity or true positive TSS
rate, measures the effectiveness of a model in capturing and In this context, RSS denotes the sum of residual squares,
correctly identifying all relevant instances. It calculates the while TSS signifies the total sum of squares.
ratio of true positive instances to the sum of true positives When comparing the performance of existing models
and false negatives. for dyslexia prediction, several performance metrics are
TP commonly used to assess the effectiveness of the models.
Precision = Here are the key metrics and some existing models that are
(TP + FP) (2) often employed in this context
TP
Recall = (3)
(TP + FN) 5. Future Directions and
Relevance: Precision is crucial to avoid false positives, while Recommendations
recall ensures that dyslexic cases are not overlooked.
Refining Deep Learning Models for Dyslexia
F1 Score: The F1 score, a harmonic mean of precision and
Enhanced Data Collection: Propose the collection of diverse
recall, offers a balanced evaluation, particularly useful in
and comprehensive datasets, including longitudinal data,
scenarios with class imbalance.
Enhancing the resilience and applicability of deep learning for transforming our approach to this neurodevelopmental
models in the realm of dyslexia detection and intervention. condition. This exploration has illuminated the intricate
Multi-modal Approach: Suggest exploring multi-modal characteristics of dyslexia, its prevalence, and the persistent
approaches that combine different data types (linguistic, challenges in early detection and tailored interventions. The
imaging, behavioral) to develop more comprehensive and introduction of deep learning, with its capacity to decipher
accurate models. complex patterns within dyslexia-related datasets, marks a
significant step forward. The application of deep learning in
Interpretability and Explain ability dyslexia detection exhibits encouraging progress, offering
Advancing Model Interpretability: Highlighting the insights into potential diagnostic markers through the analysis
significance of improving the interpretability and explain of linguistic patterns, eye-tracking data, and brain imaging.
ability of deep learning models in the context of dyslexia, However, this promising trajectory is accompanied by
fostering a clearer understanding and acceptance among challenges, including the need for diverse datasets, ensuring
clinicians and educators. model interpretability, and addressing ethical considerations.
Ethical Guidelines and Standards The imperative of refining models and establishing ethical
frameworks is underscored to ensure responsible and
Ethical Frameworks: Advocate for the establishment of equitable implementation. The potential of deep learning
ethical guidelines and standards governing the use of deep in crafting personalized interventions, tailoring support to
learning in dyslexia research and intervention to ensure individual learning profiles, heralds a new era in dyslexia
responsible and equitable practices. intervention strategies.
Longitudinal Studies and Real-world Application As we conclude this exploration, it is crucial to recognize
Long-term Impact Assessment: Recommend conducting that our journey does not end here. Continuous collaboration,
longitudinal studies to assess the long-term effectiveness interdisciplinary efforts, and the translation of research
and impact of personalized interventions derived from deep findings into practical applications are paramount.
learning models. Researchers, educators, policymakers, and practitioners are
Real-world Implementation: Promote the application of called upon to embrace these advancements, refine models,
research discoveries in practical educational settings, and implement ethical guidelines. The fusion of deep learning
fostering partnerships between researchers and practitioners and dyslexia research signifies not just a scientific endeavor
for real-world impact. but a societal commitment toward inclusivity, equity,
and personalized support. Together, let us embark on this
Collaborative Efforts and Knowledge Exchange ongoing journey, striving for a future where every individual,
Interdisciplinary Collaboration: Stress the significance irrespective of their challenges, receives tailored and effective
of interdisciplinary collaboration between researchers, support, fostering a world of learning and opportunity for all.
educators, clinicians, and technologists to address the
multifaceted challenges of dyslexia. Acknowledgement
Knowledge Exchange Platforms: Propose the development of
platforms or networks facilitating the exchange of knowledge The authors gratefully acknowledge the students, staff, and
and best practices among stakeholders in dyslexia research authority of Physics department for their cooperation in the
and intervention. research.
5. F. Hoeft et al., “Neural systems predicting long-term outcome 21. D. Froyen et al., “Atypical structural asymmetry of the planum
in dyslexia,” Proceedings of the National Academy of temporale is related to family history of dyslexia,” Cerebral
Sciences, vol. 108, no. 1, pp. 361–366, 2011. Cortex, vol. 19, no. 10, pp. 2641–2649, 2009.
6. M. S. Koyama et al., “The semantic organization of words 22. E. L. Grigorenko and A. J. Naples, “Dyslexia genetics:
in the brain: evidence from category- and modality-specific Integrating genetics, neuropsychology, neurobiology, and
deficits,” Frontiers in Psychology, vol. 4, p. 690, 2013. genomics,” Journal of Developmental and Behavioral
7. G. E. Hinton and R. R. Salakhutdinov, “Reducing the Pediatrics, vol. 30, no. 1, pp. 6–22, 2009.
dimensionality of data with neural networks,” Science, vol. 23. S. Mascheretti et al., “Neurogenetics of developmental
313, no. 5786, pp. 504–507, 2006. dyslexia: from genes to behavior through brain neuroimaging
8. L. Zou and L. Schiebinger, “AI can be sexist and racist – it’s and cognitive and sensorial mechanisms,” Translational
time to make it fair,” Nature, vol. 559, no. 7714, pp. 324–326, Psychiatry, vol. 7, no. 1, p. e987, 2017.
2018. 24. F. Richlan et al., “A common left occipito-temporal
9. N. Langer et al., “White matter alterations in dyslexia: a dysfunction in developmental dyslexia and acquired letter-by
DTI tract-based spatial statistics study,” Brain Structure and letter reading?” PloS One, vol. 8, no. 9, p. e78959, 2013.
Function, vol. 220, no. 4, pp. 1905–1916, 2015. 25. E. S. Norton and M. Wolf, “Rapid automatized naming
10. M. Vandermosten et al., “A tractography study in dyslexia: (RAN) and reading fluency: Implications for understanding
neuroanatomic correlates of orthographic, phonological and and treatment of reading disabilities,” Annual Review of
speech processing,” Brain, vol. 135, no. 3, pp. 935–948, 2012. Psychology, vol. 63, pp. 427–452, 2012.
11. E. S. Norton et al., “An investigation of the neural signature 26. H. Lyytinen et al., “A longitudinal study of the early predictors
of primary and secondary reading disorders,” Frontiers in of poor emergent literacy in children at familial risk of
Human Neuroscience, vol. 8, p. 904, 2014. dyslexia,” Journal of Experimental Child Psychology, vol.
12. F. Richlan et al., “Structural abnormalities in the dyslexic 137, pp. 157–177, 2015.
brain: A meta-analysis of voxel-based morphometry studies,” 27. M. Vandermosten et al., “Brain activity patterns of phonemic
Human Brain Mapping, vol. 30, no. 10, pp. 3299–3308, 2009. representations are atypical in beginning readers with family
13. F. Hoeft et al., “Neural basis of dyslexia: A comparison risk for dyslexia,” Developmental Science, vol. 16, no. 4,
between dyslexic and nondyslexic children equated for pp. 678–692, 2013.
reading ability,” Journal of Neuroscience, vol. 27, no. 37, 28. M. A. Skeide et al., “Genetic dyslexia risk variant is related
pp. 9878–9882, 2007. to neural connectivity patterns underlying phonological
14. K. R. Pugh et al., “Neuroimaging studies of reading awareness in children,” Neuro Image, vol. 146, pp. 526–533,
development and reading disability,” Learning Disabilities 2017.
Research & Practice, vol. 15, no. 1, pp. 55–66, 2000. 29. D. L. Lefly and B. F. Pennington, “Reliability and validity of
15. G. F. Eden et al., “Neural changes following remediation the adult reading history questionnaire,” Journal of Learning
in adult developmental dyslexia,” Neuron, vol. 44, no. 3, Disabilities, vol. 33, no. 3, pp. 286–296, 2000.
pp. 411–422, 2004. 30. A. M. Galaburda et al., “Developmental dyslexia: a diagnostic
16. I.Altarelli et al., “Letter and speech sound association in approach based on the componential model of reading,” Brain,
emerging readers with familial risk for dyslexia,” Brain, vol. vol. 123, no. 12, pp. 2373–2399, 2006.
136, no. 10, pp. 3403–3417, 2013. 31. M. Ahissar et al., “Dyslexia and the failure to form a perceptual
17. N. M. Raschle et al., “Investigating the neural correlates of anchor,” Nature Neuroscience, vol. 4, no. 7, pp. 732–734,
voice versus speech-sound directed information in pre-school 2001.
children,” PloS One, vol. 6, no. 10, p. e25803, 2011. 32. Kothapalli, Pavan Kumar Varma, V. Rathikarani, and Gopala
18. P. E. Turkeltaub et al., “The neural basis of aphasia: evidence Krishna Murthy Nookala. “A Comprehensive Survey on
from functional neuroimaging,” Aphasiology, vol. 17, no. 4, Predicting Dyslexia and ADHD Using Machine Learning
pp. 327–350, 2003. Approaches.” Inventive Systems and Control: Proceedings of
19. C. Raghuram and M. Thenmozhi, Short Review on Contrastive ICISC 2022 (2022): 105–121.
Learning-based Segmentation Techniques for Medical Image 33. Kothapalli, Pavan Kumar Varma, V. Rathikarani, and Gopala
Processing, 2023 International Conference in Advances Krishna Murthy Nookala. “Prediction of dyslexia and attention
in Power, Signal, and Information Technology (APSIT), deficit and hyperactivity disorder prediction using ensemble
Bhubaneswar, India, 2023, pp. 290–296, doi: 10.1109/ classifier model.” International Journal of System Assurance
APSIT58554.2023.10201707. Engineering and Management (2022): 1–12.
20. B. Boets et al., “Intact but less accessible phonetic
Note: All the figures and tables in this chapter were designed by
representations in adults with dyslexia,” Science, vol. 342, no.
the author.
6163, pp. 1251–1254, 2013.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
A Study of YOLO
(You Only Look Once) to YOLOv8 40
Immidisetty V. Prakash1
Research Scholar, Dept. of Electronics and Communication Engineering,
Anna University, Chennai
M. Palanivelan2
Professor, Dept. of Electronics and Communication Engineering,
Rajalakshmi Engineering College, Thandalam, Chennai
Abstract: YOLO, which stands for “You Only Look Once,” is an object detection algorithm that revolutionized real-time
computer vision tasks by enabling fast and accurate object detection in images or videos. Traditional object detection algorithms
involve multiple stages and are computationally expensive. but the YOLO is, on the other hand, approaches object detection as
an issue with regression, predicting class and bounding box probabilities in a single pass straight from the unprocessed image
pixels. , To forecast bounding boxes, the YOLO algorithm divides the input image into a grid. Objectness scores and class
probabilities for objects present within every grid cell. These grid-based approaches allow YOLO to detect multiple objects
of different classes in a single forward pass. By predicting bounding boxes and class probabilities together, YOLO achieves
real-time processing speeds, making it highly suitable for applications such as autonomous driving, surveillance, and robotics.
YOLO is a groundbreaking object detection algorithm that employs a grid-based approach to predict bounding boxes and class
probabilities directly from input images, enabling real-time and efficient object detection for a wide range of applications.
Keywords: YOLO, Object detection, Bounding boxes, Regression problems
1
erivprakash@gmail.com, 2velan.research@gmail.com
DOI: 10.1201/9781003529231-40
258 Algorithms in Advanced Artificial Intelligence
simultaneously and avoids redundant computations that are images and video frames can be detected, localized, and
present in multi-stage detection methods. classified in a single pass through the neural network.
Furthermore, YOLO introduces the concept of”anchor This real-time capability is crucial for applications like
boxes” to improve the precision with which bounding box autonomous driving, surveillance, and robotics, where
predictions. These anchor boxes serve as predefined reference timely decision-making is essential.
shapes and sizes that the algorithm adjusts to better fit the 2. Unified Detection: YOLO seeks to unify object
shape of the object in question. This technique enhances the detection into a single process, as opposed to
versatility of YOLO in handling objects of various scales and traditional methods that involve separate steps for
aspect ratios. region proposal and object classification. By predicting
object classes and bounding box coordinates together,
Overall, YOLO’s introduction marks a pivotal moment in the
YOLO simplifies the detection pipeline and reduces
evolution of object detection algorithms. By offering real-
computational overhead.
time capabilities and efficient detection through its grid-based
approach and anchor boxes, YOLO has set new benchmarks 3. Efficiency: YOLO aims to be computationally efficient
for speed and accuracy in object detection tasks, opening by avoiding redundant calculations. The grid-based
up new possibilities for applications that require instant and method splits the input image into cells, and the task
Reliable visual analysis. of each cell is to predict what’s within its boundaries.
This efficient division of labor allows YOLO to process
large images quickly and predict objects accurately.
4. Multi-Object Detection: YOLO’s objective is to detect
multiple objects of different classes within a single
image or video frame. The grid cells and anchor boxes
enable YOLO to simultaneously identify and locate
multiple objects, making it highly suitable for scenarios
Fig. 40.1 Example of object detection images where there may be various objects in the scene.
5. Handling Object Variability: YOLO aims to handle
objects with varying sizes, scales, and aspect ratios
2. Objective effectively. The introduction of anchor boxes allows
The primary objective of YOLO (You Only Look Once) YOLO to adjust predictions based on these variations,
is to provide an efficient and real-time solution for object improving the accuracy of bounding box localization.
detection in images and videos. YOLO aims to achieve this 6. Generalization: YOLO seeks to generalize well to
objective through several key goals: different types of objects, scenes, and environments.
1. Real-time Processing: YOLO’s foremost objective is This objective is crucial for deploying YOLO in diverse
to enable real-time object detection, where objects in real-world applications, where the algorithm should
possess the ability to detect a large variety of objects the scene, allowing YOLO to provide comprehensive
and adapt to different visual conditions. information about the visual content.
7. Accessibility: Another objective of YOLO is to 7. Non-maximum Suppression: After generating multiple
offer a relatively simple architecture that can be bounding box predictions for different objects, YOLO
easily understood and implemented by scholars uses minimal suppression to get rid of overlapping
and professionals in the computer vision field. and superfluous bounding boxes. The above function
This accessibility encourages wider adoption and makes sure that one precise bounding box represents
experimentation. every single object.
8. Efficiency and Simplicity: YOLO is designed to be
3. Functions efficient and relatively simple compared to multi-
phase techniques for object detection. With a single
YOLO (You Only Look Once) performs several key functions pass, estimating bounding boxes as well as class
within the context of object detection in videos and images. probabilities, YOLO reduces computational complexity
These functions are designed to enable efficient and accurate and simplifies the detection process.
detection of objects in real-time: 9. Generalization: YOLO’s function of generalization
1. Object Localization: YOLO’s primary function is to involves adapting to various object types, scenes,
accurately localize objects within an image or video lighting conditions, and environments. This function
frame. It achieves this by predicting bounding box ensures that YOLO can perform well in a wide range of
coordinates (x, y, width, and height) that enclose the real-world applications without extensive fine-tuning
detected objects. This localization information allows for each specific scenario.
users to precisely determine the location of objects in
the scene.
2. Object Classification: YOLO is responsible for
classifying the detected objects into different predefined
categories or classes. Each object is assigned a class
label, indicating what type of object it is, such as “car,”
“pedestrian,” “dog,” etc. This function enables users to
understand the content of the scene by identifying the
objects present.
3. Real-time Processing: YOLO’s architecture is designed
to obtain real-time detection of objects by using a
neural network to process both image and video frames Fig. 40.3 Functioning of YOLO
in one pass. This real-time capability is essential for
applications requiring instant decision-making, such as
autonomous vehicles and surveillance systems. 4. Applications
4. Grid-based Division: You Only Look Once (YOLO) The YOLO algorithm found a wide range of applications
splits an input image through a cell grid. Every cell is across various fields due to its real-time and efficient object
accountable for detecting objects within its boundaries. detection capabilities. Some notable applications of YOLO
This grid-based division allows YOLO to efficiently include:
process large images and detect objects across different
1. Autonomous Vehicles: YOLO is extensively used in
regions of the scene.
autonomous driving systems to identify and monitor
5. Anchor Boxes: YOLO uses anchor boxes to bicycles, cars, pedestrians, signs for traffic, as well
accommodate various ratios of aspect or object sizes. as other objects around the vehicle in real-time. This
These anchor boxes function as preset reference shapes, information is crucial for making informed decisions
and the algorithm adjusts them to better fit the shape of and ensuring the safety of passengers and pedestrians.
the detected objects. This function enhances YOLO’s
2. Surveillance and Security: YOLO is employed in
ability to accurately predict bounding box coordinates.
surveillance cameras and security systems to monitor
6. Multi-Object Detection: YOLO’s architecture enables and detect unauthorized activities, intruders, and
it to detect multiple objects of different classes within suspicious objects. Its real-time processing enables
a single image or frame. This function is crucial for rapid response to potential security threats.
scenarios where there are several objects present in
260 Algorithms in Advanced Artificial Intelligence
3. Retail and Inventory Management: YOLO is used in 12. Textile Industry: YOLO can be used to inspect textile
retail settings for tracking products on shelves and quality, identifying defects or inconsistencies in fabrics
monitoring inventory levels. It can help automate stock during the production process.
management, preventing out-of-stock situations and
optimizing supply chain operations. 5. Different versions of YOLO
4. Healthcare: YOLO is applied in medical imaging
for detecting anatomical structures and anomalies in The YOLO algorithm has evolved over the years, resulting
X-rays, MRIs, and CT scans. It assists radiologists in in several versions and variations. Here are some notable
identifying specific regions of interest and potential methods and versions of YOLO:
health issues. 1. YOLO v1 (You Only Look Once version 1): The original
5. Industrial Automation: YOLO can be used in industrial YOLO algorithm presented the idea of using regression
settings for object detection in manufacturing processes. to solve the object detection problem. The input image
It can help identify defects, inspect product quality, and was divided into a grid, and for each grid cell, it
ensure proper assembly of components. predicted bounding boxes and class probabilities and
6. Robotics: YOLO is integrated into robots and robotic applied non-maximum suppression to refine detections.
systems for object recognition and manipulation. 2. YOLO v2 (YOLO9000): YOLO v2 introduced
Robots equipped with YOLO can identify objects in improvements such as anchor boxes, which enabled
their environment, enhancing their interaction with the a model to manage objects with different ratios
world. of aspect and sizes. It also incorporated a concept
7. Agriculture:YOLO is employed in precision agriculture called”Darknet-19,” a 19-layer architecture that
to monitor crop health, identify pests, and assess plant improved detection accuracy.
growth. It enables farmers to make informed decisions 3. YOLO v3: YOLO v3 further improved object detection
to optimize crop yield and reduce resource wastage. by introducing utilizing a feature pyramid network,
8. Sports Analytics:YOLO can track players and objects in which allows the model to identify objects at various
sports events, providing valuable data for performance scales within the image. It also introduced multiple
analysis, player tracking, and generating statistics that detection scales, providing a balance between speed
enhance coaching and strategic decisions. and accuracy.
9. Augmented Reality (AR) and Virtual Reality (VR):
YOLO can be used in AR and VR applications to
enhance the user experience by recognizing and
interacting with real-world objects and environments
in real-time.
10. Environmental Monitoring: YOLO is used for
monitoring wildlife, tracking endangered species, and
studying ecological patterns. It aids researchers in
understanding and protecting biodiversity.
11. Retail Analytics: YOLO helps retailers analyze
customer behavior in stores, such as tracking foot
traffic, monitoring customer interactions with products,
and optimizing store layouts for improved shopping Fig. 40.5 YOLO v2 results in comparison to the original
experiences. version and other modern models [1]
A Study of YOLO (You Only Look Once) to YOLOv8 261
YOLO can help achieve it. A brand-new anchor box examples—the objects that are difficult to detect—a
generation technique known as “dense anchor boxes” focal loss combats the problem. When it comes to
was also introduced by the YOLO v6. The outcomes of resolution, v7 processes images at a resolution of 608
YOLO v6’s comparison with other cutting-edge object x 608 pixels, which < the YOLO v3 version’s 416 x
detectors are displayed below. 416 resolution. The v7 has better overall accuracy
9. YOLO v7: YOLO v7 is an additional YOLO version and can detect very small objects thanks to its higher
that has several enhancements over the previous resolution. Compared to existing detection algorithms,
iterations. The utilization of anchor boxes is the v7 processes images at a rate of 155 frames per second,
primary enhancement. The anchor boxes, which are which is a significant speed increase. Because of this, it
a collection of pre-defined anchor boxes with various is appropriate for delicate real-time applications where
aspect ratios and sizes, are used to identify objects with faster processing speeds are essential, like self-driving
various shapes. As opposed to the previous iterations, cars and surveillance. When it comes to accuracy, v7
this one employs nine anchor boxes, which enable it outperforms other Algorithms for detecting objects.
to recognize a greater variety of object forms and On the widely used COCO dataset, v7 achieves an
sizes assisting in their efforts to lower the quantity of average precision of 37.2%, which is comparable to
false positives. This version 7’s primary enhancement earlier detection algorithms as demonstrated below at
is the application of a novel loss function known as an intersection over a union (IoU) threshold of 0.5.
“focal loss.” Previous iterations of YOLO employ a 10. YOLO v8: The team released YOLO v8 in 2023. It
conventional cross-entropy loss function, which has was created using the original YOLO algorithm and is
been demonstrated to be less successful in identifying kept up to date by Ultralytics. YOLO v8 incorporates
minuscule objects. By down-weighting the loss for several enhancements and new features, building
well-classified examples and concentrating on the on the popularity of earlier iterations. The YOLO
Fig. 40.9 A comparison between the YOLO v7 and other cutting-edge object detectors [5]
Fig. 40.10 YOLO v8’s performance in comparison to other cutting-edge models [6]
A Study of YOLO (You Only Look Once) to YOLOv8 263
7. Iterative Training: YOLO is trained iteratively using (c) Bounding Box Regression Loss: Measures the
annotated datasets. During training, the model learns to discrepancy between expected and actual bounding
estimate the class and bounding box probabilities that box coordinates.
match the ground truth annotations. 2. Activation Functions (e.g., Leaky ReLU): Activation
8. Hyperparameter Tuning: Hyperparameters, such as functions like Leaky ReLU introduce non-linearity to
learning rate, batch size, anchor box dimensions, the CNN layers, enabling the model to learn complex
and architecture choices, are fine-tuned to attain the features and patterns.
intended equilibrium between speed and accuracy. The 3. Feature Pyramid Network (FPN) (YOLOv3 and later):
YOLO methodology’s unique aspect lies in its single- FPN is used to detect items of various sizes. It combines
pass approach, grid-based detection, and prediction of elements from different layers of the CNN to enhance
bounding boxes and class probabilities together. This the model’s proficiency in managing objects of various
methodology has paved the way for object detection sizes.
in real-time in various applications, making YOLO a 4. Darknet (Framework): Darknet is a custom neural
foundational algorithm within the domain of computer network framework developed for YOLO. It provides
vision. Other than methodologies The YOLO algorithm the architecture configurations, layer implementations,
involves several key components and steps that and training pipeline for YOLO models.
collectively enable real-time object detection. While
5. Data Augmentation (Training): YOLO uses data
YOLO itself is a single-shot object detection algorithm,
augmentation techniques during training to introduce
it utilizes various techniques and algorithms to achieve
variability enhancing the model’s robustness and
its functionality.
generalization in the training set of data.
Here’s an overview of the algorithms and techniques used in
YOLO:
9. Conclusion
1. Loss Functions: YOLO employs several loss functions
during training to guide the model’s learning process: These algorithms and techniques collectively form the
(a) Objectness Loss: Measures the inequalities foundation of the YOLO algorithm, allowing it to achieve
between the predicted and genuine objectness real-time object recognition through one pass through the
scores. network using a direct bounding box and class probability
(b) Classification Loss: calculates the variation in prediction. Using these YOLO versions there is a future scope
the actual class probabilities compared to the in the accuracy improvement, speed optimization, handling
predictions. challenging scenarios, domain specific adaption, multi model
integration, transfer learning and few-short learning, ethical
and fair AI.
14. Weiming Hu, Tieniu Tan, Liang Wang, and Steve Maybank,”
References A Survey on Visual Surveillance of Object Motion and
1. Joseph Redmon and Ali Farhadi ”YOLO9000: Better, Faster, Behaviors, IEEE Transactions On Systems, Man, And
Stronger” CoRR abs/1612.08242, 2016. Cybernetics, Vol. 34, No. 3, August 2004.
2. Joseph Redmon and Ali Farhadi “YOLOv3: An Incremental 15. Lorenzo Favalli, Alessandro Mecocci, and Fulvio Moschetti,”
Improvement” CoRR abs/1804.02767, 2018. Object Tracking for Retrieval Applications in MPEG-2” IEEE
3. Alexey Bochkovskiy, Chien-Yao Wang and Hong-Yuan Mark Transactions On Circuits And Systems For Video Technology,
Liao ”YOLOv4: Optimal Speed and Accuracy of Object VOL. 10, NO. 3, APRIL 2000.
Detection” CoRR abs/2004.10934, 2020. 16. Weiming Hu, Tieniu Tan, Liang Wang, and Steve Maybank,”
4. Li and Chuyi et al. ”YOLOv6: A single-stage object detection A Survey on Visual Surveillance of Object Motion and
framework for industrial applications.” arXiv preprint Behaviors, IEEE TRANSACTIONS ON SYSTEMS, MAN,
arXiv:2209.02976, 2022. AND CYBERNETICS, VOL. 34, NO. 3, AUGUST 2004.
5. Wang, Chien-Yao, Alexey Bochkovskiy, and Hong-Yuan 17. Z. Akata, F. Perronnin, Z. Harchaoui, and C. Schmid “Label-
Mark Liao “YOLOv7: Trainable bag-of-freebies sets new Embedding for Attribute-Based Classification” In Proceed
state-of-the-art for real- time object detectors” Proceedings of ings of the IEEE Conference on Computer Vision and Pattern
the IEEE/CVF Conference on Computer Vision and Pattern Recognition, June 2013, pages 819–826.
Recognition. 2023. 18. Z. Akata, S. Reed, D. Walter, H. Lee, and B. Schiele
6. Dr. Ramya, Nikhil, Pavan R, Prabhu nagappa chinagudi and “Evaluation of Output Embeddings for Fine-Grained Image
Vishal ”Real-Time Object Detection and Tracking” Volume 9 Classification” In Proceedings of the IEEE Conference on
Issue 8, 2021. Computer Vision and Pattern Recognition, 2015, pages 2927–
7. Abdul Vahab, Maruti S Naik, Prasanna G Raikar and Prasad 2936.
S R ”Object Detection and Its implementations and Users” 19. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
International Research Journal of Engineering and Technology in Health Care Using Hybrid Model Algorithms”,Journal of
(IRJET) Volume: 06 Issue: 04, Apr 2019. Artificial Intelligence, Machine Learning and Neural Network
8. Licheng Jiao, Fan Zhang, Fang Liu, Shuyuan Yang, Lingling (JAIMLNN), Volume 3, 2023.
Li, Zhixi Feng, and Rong Qu ”A Survey of Deep Learning- 20. K. Barnard, P. Duygulu, D. Forsyth, N. De Freitas, D. M.Blei,
based Object Detection”, arXiv:1907.09408v2 [ cs.CV] 10 and M. I. Jordan “Matching words and pictures” The Journal
Oct 2019. of Machine Learning Research, 3. 2003, pages 1107–1135.
9. N Hassan and C S Woo,” Machine Learning Application in 21. J. Bergstra, O. Breuleux, F. Bastien, P. Lamblin, R. Pascanu,G.
Water Quality Using Satellite Data”, Earth Environ. Sci. 842 Desjardins, J. Turian, D. Warde-Farley, and Y. Bengio.Theano
01, 2018. “a CPU and GPU math expression compiler” In Proceedings
10. Makes Tiwari and Dr. Rakesh Singhai ”A Review of Detection of the Python for scientific computing conference(SciPy),
and Tracking of Object from Image and Video Sequences” volume 4,. Austin, TX, 2010, page 3.
International Journal of Research and Management Volume 22. C. Burges, T. Shaked, E. Renshaw, A. Lazier, M. Deeds, N.
13, Number 5, 2017. Hamilton, and G. Hullender. Learning to rank using gradient
11. Chandrajit, Girisha, and Vasudev,” Multiple Objects Tracking descent. In Proceedings of the 22nd international conference
In Surveillance Video Using Color and Hu moments”, Signal on Machine learning, ACM, 2005, pages 89–96.
Image Processing: An International Journal (SIPIJ) Vol.7, 23. Prakash, Immidisetty V., Valiki Vijayabhasker, and Srinivas
No.3, June 2016. Gadari. “multiplexers, demultiplexers, current progress and
12. Jamal Raiyn,” Detection of Objects in Motion—A Survey of algorithms of wavelength assignment in wdm network.” in
Video Surveillance, Advances in Internet of Things, (2013) 3, Research Review International Journal Of Multidisciplinary,
73-78 http://dx.doi.org/10.4236/ait.2013.34010 2019/5 pages 2615-2619.
13. Zdenek Kalal, Krystian Mikolajczyk, and Jiri Matas,” Note: Source for all the figures in this chapter were from https://
Tracking-Learning-Detection”, IEEE Transactions on pattern www.researchgate.net/figure/Timeline-of-You-Only-Look-Once
analysis and machine intelligence, vol. 6, no. 1, January 2010. YOLO-variants_fig1_370153499
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: The loss of many plant and animal species as a result of various climatic shifts is one of the world’s most pressing
challenges today. The extinction of many species is due to a complex interplay of many different causes. Poaching, climate
change, and the loss of natural habitats are all contributing causes. American pikas, Adeline penguins, koalas, and ringed
seals are among the many species that are at risk from the effects of climate change. Predicting the likelihood of extinction is
crucial for preserving ecological harmony. Researchers can accurately observe animals in their native environment. There has
been a dramatic increase in the usage of automated hidden cameras to monitor wildlife more efficiently and without operator
intervention. These cameras are handy and reliable for collecting vast volumes of data about animals. However, manual data
collection and analysis from camera traps is a difficult and time-consuming process. Using the Random Forest and Convolutional
Neural Network algorithms, our goal is to create a model that can forecast which species are in danger of extinction.
Keywords: Species, Disappearance, Natural habitat, Intervention, Endangered
1. Introduction Currently, deep neural networks can train models to verify the
presence of extinct species, analyze their statistics, and even
Not only do humans call Earth home, but tens of thousands of predict which species will be considered endangered in the
other species do as well. Even if numerous species compete future. Data collection on species traits and the environmental
with one another, the extinction of even a single one can have variables that affect their survival is part of the research.
far-reaching consequences. Yet, developments in technology Position in space, kind of habitat, and other variables may
and the economy, pollution of the air and water, and shifts in all play a role. Reputable conservation groups and databases,
population dynamics all pose significant risks to the world’s such as the IUCN Red List and the World Wildlife Fund
biodiversity. An alarming number of species are becoming (WWF), provide the information. Next, we will prepare
endangered. The International Union for Conservation the data for analysis through preprocessing, which involves
of Nature has noted an increase in the number of species cleaning and structuring the data. To find the most relevant
classified as threatened, with 1,102 in 1996 and 14,360 in characteristics for species endangerment prediction, we will
2019. There are 1,197 plant species and 13,868 animal apply feature selection approaches. After that, we’ll construct
species among them. a prediction model that takes these features into account
using the Random Forest algorithm.
1
prakashlnr@gmail.com, yprakasarao@kluniversity.in; 2subramanyam.mavuri@gmail.com; 3Venkataramana.t@gmail.com
DOI: 10.1201/9781003529231-41
268 Algorithms in Advanced Artificial Intelligence
4.1 Prediction
Fig. 41.5 Confusion matrix
We use the trained CNN and Random Forest models to make
the prediction. Making predictions on the test dataset is the layer, a max pooling layer, a dense layer, and a single output
next step after training the models. We train the models with unit. The model trained using the 3D tensor.
the test data and then compare their predictions to the real
labels to see how well they did. 4.3 Random Forest Algorithm
Species are categorised according to their conservation
status using the random forest algorithm. Scientific names,
park names, observations, common names, and conservation
status were among the features included in the dataset. The
‘fit’ method was used to teach the algorithm to anticipate
the preservation state of certain species from other
characteristics, and it was applied to a subset of the data. By
setting the ‘n_estimators’ argument to 100, we formed the
random forest by combining 100 decision trees. The “predict”
method generated predictions on the test dataset based on the
generated model. The model’s accuracy was evaluated using
the “accuracy score” technique from the scikit-learn library.
The ‘classification_report()’ method can generate a report
that details the accuracy, recall, f1-score, and support for
Fig. 41.4 Screennshot of proposed work model every class in the test data.
Additionally, the confusion matrix is used to assess the
4.2 Deep Learning Algorithms performance of the model. The columns in this table
Convolutional Neural Network (CNN) represent the anticipated values, which are used to compare
the predicted values and assess the performance of a
Using the observation’s scientific name, park name, common categorization model. The diagonal elements represent the
name, and conservation status, a model is built using number of accurate forecasts, while the off-diagonal elements
convolutional neural networks to predict the observation’s represent the number of inaccurate predictions.
category. Image identification and other data with spatial
correlations are typical applications of convolutional neural 4.4 Loss Calculation
networks (CNNs), a kind of neural network. N_samples
The loss function “mae,” which stands for Mean Absolute
denotes the shape of the generated 3D object from the input
Error, measures the absolute difference between the expected
data, with n_timesteps representing the number of time steps
and observed values. Frequently, it is applied to regression
and n_features representing the number of input features.
issues.
Next, we inputted the 3D tensor into a convolutional neural
network (CNN) model. This model had a 1D convolutional The optimizer ‘adam’ is an algorithm used for gradient-based
optimization Since the model is predicting a continuous
Prediction of Endangered Species Using Artificial Intelligence 271
5. System Architecture
Species Identification Model: Create a model that uses
recurrent neural networks (RNNs) or convolutional neural
networks (CNNs) to identify various species from pictures or
audio recordings.
Habitat Analysis Model: Create a model to analyze and predict
suitable habitats for various species using environmental
Fig. 41.8 Work flow of proposed model
variables and spatial data.
Population Trend Prediction: Develop models to predict study. After collecting the data from various sources, it was
population trends of species based on historical data, taking preprocessed to extract relevant features. Random forests and
into account factors like climate change and habitat loss. convolutional neural networks were the methods used for
categorization. We trained the data using CNN and random
6. Conclusion forest techniques, and the model achieved an accuracy rate of
86.2%. This experiment demonstrated the feasibility of using
Using machine learning techniques to predict when species machine learning algorithms for species extinction prediction.
may become endangered was the primary objective of the Applying convolutional neural networks to random forests
272 Algorithms in Advanced Artificial Intelligence
allows for accurate species classification based on their 5. Mohammad Sadegh Norouzzadeha, Anh Nguyenb, Margaret
properties. Future research is required to improve the models’ Kosmalac, Alexandra Swansond, Meredith S. Palmere,
predictive abilities by increasing their accuracy and including Craig Packere, and Jeff Clunea, “Automatically identifying,
additional features. Possible additions to the research include counting, and describing wild animals in camera-trap images
with deep learning,” PNAS, 2018.
the ability to forecast behaviour and real-time monitoring of
6. Atri Saxena, Deepak Kumar Gupta, Samayveer Singh,“An
endangered species. There may be advantages to this.
Animal Detection and Collision Avoidance System Using
Deep Learning,” SpringerLink, 2020.
References 7. Rajasekaran Thangarasu, Vishnu Kumar Kaliappan,
Raguvaran Surendran, Kandasamy Sellamuthu, Jayasheelan
1. Mauro dos Santos de Arruda, Gabriel Spadon, Wesley Nunes Palanisamy, “Recognition Of Animal Species On Camera
Goncalves, & Bruno Brandoli Machado, “Recognition of Trap Images Using Machine Learning And Deep Learning
Endangered Pantanal Animal Species using Deep Learning Models,” International Journal Of Scientific & Technology
Methods,” IJCNN, 2018. Research, 2019.
2. Hung Nguyen, Sarah J. Maclagan, Tu Dinh Nguyen, Thin 8. Zhongqi Miao, Kaitlyn M. Gaynor, Jiayun Wang, Ziwei Liu,
Nguyen, Paul Flemons, Kylie Andrews, Euan G. Ritchie, and Oliver Muellerklein, Mohammad Sadegh Norouzzadeh, Alex
Dinh Phung, “Animal Recognition and Identification with McInturff, Rauri C. K. Bowie, Ran Nathan, Stella X. Yu,
Deep Convolutional Neural Networks for Automated Wildlife Wayne M. Getz., et al. “Insights and approaches using deep
Monitoring,” Deakin University, Geelong, Australia, 2017. learning to classify wildlife,” Scientific Reports, 2019
3. N. Banupriya, S. Saraya, Rashi Swaminathan, Sachinthaa 9. Alexander Loos, Christian Weigel, Mona Koehler, “Towards
Harikumar, Sukhita Palanisamya, “Animal Detection using Automatic Detection of Animals in Camera-Trap Images,”
Deep Learning Algorithm,” Journal of Critical Reviews, 2019. European Signal Processing Conference (EUSIPCO), 2018.
4. Sazida B. Islam, Damian Valles, “Identification of Wild
Species in Texas from Camera-trap Images using Deep Neural Note: All the figures and table in this chapter were designed by the
Network for Conservation Monitoring,” CCWC, 2020. author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
M. Janakidevi1
Assistant Professor, Department of Computer Science and Engineering,
Sagi Ramakrishnam Raju Engineering College
Ramalinga Swamy Cheruku2
Assistant Professor, Department of Computer Science and Engineering,
NIT Warangal
Ch. Rami Naidu
Assistant Professor, Department of Computer Science and Engineering,
Sagi Ramakrishnam Raju Engineering College
Abstract: Image processing features challenges like noise, occlusion, and blocking of elements. AI systems employ effective
algorithms but still face issues like darkness, rain, snow, smoke, and reflections. The fusion approach focuses on picture
enhancements, particularly for medical applications like Alzheimer’s. One of the most prevalent neurodegenerative disorders,
Alzheimer’s disease, causes a progressive loss of memory and independence. Amyloidal plaques and tau tangles, two forms of
neurotoxic protein buildup in the brain, are its defining features. Because pathology develops silently over decades, it is critical
to diagnose patients as early in the illness process as possible in order to take appropriate action. This study used imaging to
identify tau protein levels in the brain as a predictor of cognitive decline brought on by the early diagnosis of Alzheimer’s
disease. The study investigates the effectiveness of various imaging techniques in identifying individual variations associated
with Alzheimer disease using convolutional neural networks. This method makes use of convolutional neural networks to detect
tau protein using imaging and forecast cognitive decline in order to facilitate an early diagnosis of Alzheimer’s disease. The
convolutional neural network outperforms RGB, DCT, and CNN in graph accuracy measures.
Keywords: Alzheimer, Convolutional neural networks, Early diagnosis, Neurodegenerative diseases
1
mjd@srkrec.ac.in, 2rmlswamy nitw.ac.in, 3crn@srkrec.ac.in
DOI: 10.1201/9781003529231-42
274 Algorithms in Advanced Artificial Intelligence
Beyond amyloid PET, tau PET is a helpful diagnostic tool Identifying individual differences: Amyloid plaques and
that helps diagnose Alzheimer’s disease more precisely and tau are linked to clinical symptoms, with tau’s absence
enhances our capacity to stage the illness. Research has or presence determining a patient’s condition. Imaging
demonstrated that, in contrast to β-amyloid, postmortem NFT techniques for tau are challenging due to its complex
load in cortical regions is connected with clinical symptoms structure. Recent drugs targeting amyloid and tau proteins
[5]. Tau PET can be used to stage patients and determine show promising results [12]. Understanding tau distribution
which ones are in the preclinical stages of the illness because and its impact on symptoms is crucial for better Alzheimer’s
tau pathology endures during these stages [6]. By using disease management. Incorporating tau PET into clinical
tau PET to place participants at various points along the evaluations can help assess individual prognosis and select
Alzheimer’s disease continuum, it was possible to show that, the most appropriate therapeutic strategy.
even in the absence of prior knowledge about the locations,
in vivo tau PET can accurately replicate the Barak spreading 3. Alzheimer Disease Detection by
pattern of NFT pathology. Nevertheless, the absence of
postmortem neuropathological validation places limitations Convolutional Neural Network
on both studies [7]. When a person develops Alzheimer’s Even if the dimension of the disease object database has a
disease in the preclinical stages, tau PET has been shown to big impact on how accurately disease objects are identified,
be a reliable indicator of cognitive deterioration. Researchers the quality of the classification technique is crucial. A
discovered that in all patient groups, including Aβ-positive part of machine learning is deep learning [14]. Because
cognitively normal individuals, tracer absorption in the the properties are automatically extracted, deep learning
temporal cortex by Flortaucipir and second-generation works better than traditional machine learning techniques.
tracer RO-948 predicted cognitive impairment. Compared to Moreover, supplying tasks to the network in addition to raw
volumetric MRI and Aβ PET, its predictive performance was picture data is necessary when using deep learning for “end
better [8]. It has been discovered that preclinical Alzheimer’s to-end learning.” Convolutional neural networks are typically
disease can be detected more effectively in brain regions with employed in Alzheimer’s disease studies to enhance visual
higher Flortaucipir absorption. When it comes to identifying aspects [16].
Alzheimer’s disease-like pathology and permitting ante
mortem biological staging, tau-PET is similar to amyloid-
PET [9].
The length, indicated by N, and size of the two matrices, x DCT 0.91
and y, are the same. The DCT transforms the columns of an CNN 0.98
x matrix. Since vectors go from 1 to N rather than 0 to N-1,
the series is indexed from n = 1 and k = 1, as opposed to the The convolutional neural network performs best among the
typical n = 0 and k = 0. accuracy measurements for RGB, DCT, and CNN in the
graphs. Although measurement is essential to comprehending
276 Algorithms in Advanced Artificial Intelligence
Abstract: There is growing consensus that matrix metalloproteinase (MMP) inhibitors, both naturally occurring and man-
made, can be effective cytostatic and anti-antigenic therapeutic targets in the fight against cancer. Because of their significance
in cancer, many inhibitors are currently undergoing clinical studies. The analysis produces computational dock scores, which
are then compared to experimental values, and used to generate further graphs, charts, and observations. LOO (leave-out-one)
is the basis for the model’s cross-validation. We will use the r2 value (correlation coefficient) and the RMSE (root mean square
error) to assess the QSAR model’s quality.
Keywords: MMP inhibitors, PRESS, Leave-out-one (LOO), QSAR, Root mean square error etc.
DOI: 10.1201/9781003529231-43
278 Algorithms in Advanced Artificial Intelligence
Table 43.1 Molecular Descriptors data and statistical values of newly proposed model equations
Descriptors Coefficients
Model-1 Model-2 Model-3 Model-4 Model-5
Total Lipole -0.0144 -0.0277 -0.0287 +0.0651
LipoleZ Component -0.041 -0.047 -0.046 -0.042
KierChiV2(path)index -0.701 - -
KierChiV3(cluster) index +1.845 - -
BalabanTopologicalindex -3.984 - -
NumberofC1Atoms -0.341 - -
6-Membere dAliphaticrings -0.390 - -
H-bondDonors +0.376 - - +0.223
KAlpha2index - -0.408 -0.404
6-memberedaromatic rings - +0.548 +0.594 +0.208 +0.234
Rotatable Bonds - +0.153 +0.147 +0.249 +0.221
LUMO - -2.84 -3.017 -3.735 -3.560
4. Results and Discussions The distribution of activity levels in the validation set is
comparable to that in the training set. Below, we present the
results and statistics of the multiple linear regression method
4.1 Complete Data Set
for various descriptors.
The most important features were found using multiple
regression analysis with F-stepping and single-row cross- Table 43.2 Statistics for equation numbers 7-8
validation. These included inertia moments, the lipole
R 0.858 0.868
component, form flexibility, and six-member rings. The linear
2
QSAR model includes all 72 inhibitors, as demonstrated in r 0.737 0.755
Equation 5. q 2
0.602 0.798
Fig. 43.2 Observed vs Predicted Activity of validation set obtained for equation number 7
Abstract: A blockchain can be considered a group of records or an accessible history shared among people involved in the
transaction. For example, all parties in that process validate every transaction accepted for incorporation. The Blockchain
has stored one piece of data. It could never be written or altered again in any way. As a result, the Blockchain could be
considered a digital ledger that includes all of the transactions that have occurred. It is also the blockchain technology used by
cryptocurrency networks such as the decentralized Bitcoin or Ethereum, which could be considered computerized peer-to-peer
cash. This paper incorporates a history of Bitcoin, a few literary evaluations, an explanation of how the Blockchain works, and
an implementation of the network.
Keywords: Block cypher, Bitcoin, IoT, Blockchain etc.
DOI: 10.1201/9781003529231-44
Exploring the Rise of Cryptocurrencies with Blockchain Technology 283
of bodily concerns in conjunction with national governments advancement are not adequately covered in detail; however,
and fund affiliations. the topics are discussed in general terms [7]. As a result, the
A decentralized web handles a protected sequence or link creators notice that blockchain writing, for the most part, is
of time stamp information saved in a database managed of a distinct sort, in which the great promise probabilities of
by a group of customers [3]. Every web device gets access advancement are typically safe. However, the conversation on
to the information or records in a blockchain, making it a how Blockchain will enhance the creation of encouragement
decentralized or disseminated database. Cryptographic within organizations is still lacking. Much of the attention
methods are employed to encode all the Blockchain’s is focused on what might happen if most of the population
necessary data entries. It assures that the data in the Blockchain accepts Blockchain, The most fundamental possible use
is secure. The advantages of Blockchain innovation outweigh cases, rather than on the esteem-creating procedures of
the disadvantages, primarily concentrating on administrative Blockchain. Instead, they will examine why companies use
issues and difficulties. One of the most critical events in blockchain technology to solve problems and its benefits [8].
developing blockchain technology is the consolidation of Blockchain’s entrepreneurial difficulty is an enhancement
“splendid contracts.” Intelligent contracts are generally problem, like new development monetary features, because it
computer programmers capable of carrying out the provisions necessitates non-value coordinating over the complementary
of an agreement in this fashion [4]. Sharp Property is another nature of utilizes and opportunities [9]. Satoshi Nakamoto’s
connected plan concerned with dominating the demand with private activities released “Bitcoin: A Peer-to-Peer Electronic
relevant properties or resources using blockchain-based Money System” in 2008. A distributed kind of electronic money
techniques that use intelligent Contracts. Whether the asset is suggested in this research will enable online components to
natural (for example, an automobile, a telephone, or a rare) or be moved extensively across collections without a finance
non-physical (for example, provides of Associate in Nursing firm. Bitcoin was the most important validation of this
affiliations), the property type is essential. Bitcoin is not thinking method [10]. The proper word-processing finance
typically considered money, which must be stated here since specifications would represent all structures and mediums of
Bitcoin is linked to the dominating concern concerning a exchange that use cryptography to deal with trade instead of
particular subject matter [5]. The transaction record includes those frameworks where transactions are routed through a
the transaction date, time, and amount between two parties. collected certainty in parts, as the term implies [11].
As seen in Fig. 44.1 [6], blocks consist of a header and body. Educational institutions evaluate student input to improve
teaching and learning. Online processes simplify feedback
collection, summarization, and abstraction. There are
various internet ways. Finding a practical approach is the
central issue. This study presents our sentimental analysis
methodology for analyzing student comments utilizing long-
term memory. The suggested model is compared to Naive
Bayes, decision trees, and random forests [12]. Blockchain
can revolutionize the world with ease, transparency,
accuracy, speed, and affordability. Increased familiarity and
confidence in Blockchain in finance come from successful
use cases, testimonies, and suitable legal reforms [13]. To
use science and technology, one must gather, analyze, and
interpret health, family, nutritional, and blood data. Naturally,
these data are massive, possibly high-dimensional, diverse,
Fig. 44.1 Block structure and architecturally complicated. A practical data mining
approach is needed to collect and analyze such data and
categorize rural women. Researchers offer an intelligent data
2. Literature Review categorization system (IDCS) to do this. The IDCS has four
The written survey before this assessment explains blockchain phases based on a thorough categorization process. IDCS
creation and product distribution throughout geographical begins by methodically collecting data from diverse sources
locations. The review illustrates that subjects such as on rural women’s nutritional knowledge, attitudes, and dietary
Blockchain as an organization advancement, intelligent patterns [14]. Attrition is the gradual loss of firm personnel
contracts, implementation plans, entrepreneurial possibilities without replacement. High attrition rates cause talent loss,
and difficulties, and Blockchain as a globally beneficial inadequate research, and wasted training expenditures [15].
284 Algorithms in Advanced Artificial Intelligence
The first part of this quantitative research used a novel people depart for better opportunities. Replacing qualified
questionnaire to assess Malaysian blockchain communities’ staff costs money. Thus, they examine current and previous
awareness, acceptance, and confidence in blockchain employee data to determine prevalent attrition factors [21].
technology applications. The questionnaire asks about News media informs the public about frequent happenings.
demographics, FinTech awareness, trust, and acceptance, Today, social media like Twitter delivers user-generated news
notably Blockchain and cryptocurrencies. A 304-person pilot information. Clustering data and providing just important
study validated the revised questionnaire in the second phase. information makes this resource valuable. For data filtering,
The reliability test uses Cronbach’s alpha of 0.908. This phase they employed density-based k-means and graph clustering.
included a validated questionnaire survey with 304 online After filtering, we rank the data by keyword frequency,
responses. The final step of the research employed descriptive relevant key terms, and dataset fundamental term similarity.
statistics to show that blockchain and cryptocurrency They may also cover science, technology, sports, and trends
knowledge is intermediate [16]. Twitter is the most popular besides news [22].
microblogging service and a growing social network. Social Technology enthusiasts are excited about decentralized digital
media contains much data in tweets, forums, status updates, cryptocurrency and “Blockchain” technologies. Blockchain
comments, etc. Applications may use sentiment analysis to protocols, or distributed-ledger technology, have great
process and evaluate this data automatically. Twitter sentiment promise in financial technology. This technology allows the
analysis uses tweets to determine user thoughts and attitudes. creation of safe, trustworthy, and decentralized autonomous
Natural Language Toolkit (NLTK) is a Python machine ecosystems for numerous situations, including better use of
learning and sentiment analysis toolkit. What underpins text old devices, infrastructure, and resources [23]. According to
processing and classification? The study presented a machine a systematic study, cryptocurrencies are digital currencies
learning-based classifier to extract election tweets and assess that perform blockchain-based transactions. Control is held
tweeples’ opinions. Tweets about a politician might be good, by an algorithm and its users in this decentralized financial
harmful, or neutral [17]. system. Financial tool blockchain may boost global growth
The stock market projection predicts the future value of [24]. Feedback is obtained using qualitative scoring.
equities exchanged with another financial system. The present Recent feedback mining techniques mainly concentrate on
study thoroughly explains Machine Learning stock prediction. qualitative remarks and involve manual procedures. It cannot
Machine learning and AI are being used to anticipate stock be evaluated by further examination. A student feedback
values. Researchers spend more time developing methods mining system (SFMS) uses text analytics and sentiment
to increase stock prediction model accuracy each day. This analysis to give educators quantifiable and in-depth analysis
research focuses on the best stock market prediction model of qualitative student input, enhancing learning experiences
[18]. Graders struggle to provide consistent feedback with a [25].
consistent interface, mindset, and deadline. Words, sentences, Bitcoin is a peer-to-peer digital currency maintained by
word count, average length, structure, and arrangement of an open-source software with cheaper transaction costs, higher
essay are used for accurate grading. The sequential forward security and scalability than fiat money, and no central bank.
feature selection approach compares accuracy and picks Scientific interest is growing despite concerns about unlawful
the best subset. It’s easy to build an efficient subset from an usage and societal effects. This study defines and evaluates
empty set and works well on small data sets [19]. bitcoin sustainability literature, including environmental,
Unlike traditional money, digital money is a block of data social, and economic factors. According to studies, Bitcoin is
validated by a hash. All Bitcoin users in the environment a niche currency because mining new bitcoins and running the
get the info. Data mining will occur when a user transacts. virtual monetary system need too much energy. Blockchain, a
Cryptocurrencies have pros and cons as money, and there is no distributed and democratically maintained public transaction
legal framework for their circulation. Government recognition record, may provide new and challenging opportunities
is needed for the public to accept digital money as payment. [26]. Evaluation is the primary way schools assess students’
Because Bitcoin is unfamiliar to several Indonesians, the learning capabilities. One of the main tests is essay writing.
government has not recognized it as a currency. Technology Currently, this assessment is done manually. This takes
is advancing rapidly in the 4.0 revolution age, and digital time and effort. For this project, this assessment method is
money will replace physical money in the following years automated. We use keywords as features since machines can’t
due to its ease [20]. Employee attrition prediction is a severe grasp our rating metrics. First, the student essay is matched
issue in enterprises today. Organizations struggle with against admin keywords. The essay is poor if the similarity is
employee attrition when skilled, technical, and essential under 20% [27].
Exploring the Rise of Cryptocurrencies with Blockchain Technology 285
3. Working on Blockchain it to the rest of the framework as the concurrent one in the
Blockchain. On the other hand, will the system decide which
The invention of Blockchain applies to every publicly traded court should be added to the Blockchain? Meanwhile, various
public asset transfer, mainly on the Internet today. However, Centre positions could generate different, completely distinct
due to the general third-party UN agency technique and squares. In the absence of need, squares may converge at
intervention in any electronic transaction, the online business totally other |completely various solicitations to varying
gets entirely fixed to the fund foundation filling out the concentrations within the frame, making it impossible to rely
forms [28]. Moreover, the approval defense of transfers is on demand.
the responsibility of a particular third party. Therefore, an
unavoidable level of dishonesty exists in online trades, which 4. Corporate Financing with Bitcoin
requires the intervention of cash-connected exchanges to be
successful. As a result, conversion rates are incredibly high. Incorporating business governance into the Bitcoin and
An online exchange between two willing participants is Blockchain structures is generating interest and enthusiasm
executed over the Internet using Bitcoin, which relies on in several areas. The information system is robust, and
scientific disciplinary evidence rather than the confidence of blockchain improvement is being made to produce a safer,
the outsider. Every trade is protected by a digital signature more profitable strategy for stock trading. DocuSign, an
that has been verified. Every interchange is disseminated to associate’s organization that makes significant investments
the “general society key” of the collectors and strictly tagged in e-contracts, has mainly uncovered a proposed method
with the “encryption key” of the sender using the “general with Visa for using Blockchain to track motorcar rentals and
society key” of the sender. Businessmen dealing in electronic reduce the amount of paper used [31]. DocuSign is a leading
currency must exhibit their ability to take responsibility provider of e-contracts. Microsoft may turn “shrewd gets”
for their “encryption key” while considering the main containing information about its organization into “shrewd
objective of burning through cash. The fraction receiving the gets” that take advantage of blockchain innovation. Even
advanced money confirms the processed signature, assuming though blockchain technology is still young, corporations
responsibility for the “encryption key” on the transfer are researching ways to construct many tiny, “private
using “the public key” of the sender [29]. Each exchange is blockchains” inside their operational environments due to
broadcast to every Bitcoin hub and documented in a public their growing fixation with blockchain development.
record after verification.
5. Blockchain with IoT
Transaction
Send Money to B Represented as Block IoT has the potential to become a mainstream innovation in
both the consumer and corporate sectors. This requirement
has prompted attempts to establish IoT phases in specific
Transaction locations [32]. Among the benefits of blockchain growth is
Block Send to ALL
Approved by Network that it energies localized IoT phases, such as anchored and
reliable information interchange and, most importantly,
keeping records. In such a scenario, the Blockchain serves
Blocks Added to Money Moves from A as the public record, preserving an accurate record of various
Chain to B communications sent back and forth between devices inside
an Internet of Things highly localized architecture. In
Fig. 44.2 Send process with block chain collaboration with Samsung, ADEPT is a component created
The Bitcoin project addressed this problem by establishing by IBM that uses members from Bitcoin’s hidden diagram to
today’s recognized Blockchain development rules. In create a flowing arrangement of devices, sometimes known
Fig. 44.1, the Bitcoin system orders transactions by placing as a localized network of things [33]. Inside the platform,
them on social events called squares and linking them using Skillful makes use of three various origins: BitTorrent (record
Blockchain technology after a short time. The deals likely distribution), Ethereum (innovative agreements), and Tele
took place in a small space. These squares are connected (like Hash (Peer-to-Peer Communication).
a chain) in a genuine instantaneous, sequential request. Every
court holds the ashes of the previous rectangle, and each 6. The Benefits of BITCOIN
rectangle contains the hashing of the last square [30].
Table 44.1 demonstrates different types of Blockchain and
There is another difficulty that needs to be addressed. When a its merits and demerits. With a decentralized money system,
Centre creates a square from untested trades, it recommends the govt or banks had no connection to the currency. This
286 Algorithms in Advanced Artificial Intelligence
could be useful if a country is struggling financially (like data insurance, sociology, and responsibility [38]. For
the “Unique Recession” in the United States). Transactions example, Liang pushes a Blockchain-based Cloud information
are usually free and minimal. Trading cash in any zone in birthplace definition, or, in other words, ‘ProvChain,’ which
the world is simple. However, in reality, it takes little time is very specific in cloud information birthplace. After a
at all. In the case of an individual who saved bitcoins, banks redesigned straightforwardness and information obligation,
cannot use them. This implies that monetary torments a cloud-based Blockchain program will defend against
imposed by governments will not affect Bitcoin’s valuation. modified records [39, 40, 41]. The birthplace information
The square chain growth considerably reduces the need for becomes more accessible, reliable, confident, and valuable
existing intermediaries to bridge the respect-based trust [42].
gap. However, Bitcoin and different digital currencies are
volatile. This means a bitcoin’s value might fluctuate without 8. Conclusion
warning, and there is no way to predict or explain why. Their
value fluctuates because bitcoins are not tied to a single The Blockchain’s decentralized technology and peer-to-peer
association, nation, or bank. Unlawful things and activities characteristics are highly regarded. However, Bitcoin hides
(illegal narcotics, guns, etc.) could be paid for using bitcoins, several blockchain kinds of research. Therefore, Blockchain
which are more challenging to track. Bitcoins are now stored has applications well beyond Bitcoin. Blockchain has changed
in virtual online wallets. A skilled software engineer could traditional businesses through decentralization, consistency,
break into these virtual wallets, but it has been done before. anonymity, and audit. To summarise, Blockchain is the
Many customers struggle to grasp Bitcoin’s complicated developing backbone of the Bitcoin currency. The importance
square chain. of passed-on data and Blockchain’s safety make it an exciting
breakthrough for comprehending current financial and non
cash-related company issues. Depending on your viewpoint,
7. Operation of Blockchain Outside the computerized, money-based, primarily technical school
Cryptocurrency is either vainglorious or disappointing in growth. Our
Bitcoin is a great Blockchain application. Registration of efforts to enhance blockchain technology allow us to utilize
sanctionative unfathomable applications is a miracle [34, it for business transactions. Thus, its security, assurance,
35]. We protect and verify definitive reports using deeds traceability, trademark knowledge origin, and timestamping
and validations, medical administration data, IoT, and features have moved outside its core application zones.
Cloud. Tapscott claims that Blockchain will be the “General Regarding trading, the Blockchain and its variants are
Ledger,” allowing intelligent deeds, suburbanized and self- interested in any reasonable transaction, regardless of whether
administering affiliations/citizen-led groupings, and more it is a human-to-human or automated marketing.
[36,37]. Cloud information is a part of ‘Data Provenance,’ Moreover, it creates the idea that it is secure, which is
which maintains the historical background of every cloud particularly significant considering the overall development
information challenge and the subsequent assignments of the Internet of Things. As a direct consequence of this,
that are fulfilled as quickly as feasible in the Cloud. For the Blockchain has garnered a lot of interest. It seems this is
the foreseeable future, it will be essential to grant the most the case when discussing emerging nations when establishing
extravagant security to the information birthplace to ensure trust is one of the most important goals.
Exploring the Rise of Cryptocurrencies with Blockchain Technology 287
onFinancial Cryptography and Data Security, Berlin, 36. Dennis, R. and Owen, G. (2015) ‘Rep on the block: A next
Heidelberg, pp.486–504. generation reputation system basedon the blockchain’, 2015
29. M. Marchesi,”’ Why Blockchain is important for programming 10th International Conference for Internet Technology and
designers, and why programming building is crucial for SecuredTransactions (ICITST), IEEE, pp.131–138.
blockchain programming (Keynote),”’ 2018 International 37. Eyal, I., Gencer, A.E., Sirer, E.G. and Van Renesse, R. (2016)
Workshop on Blockchain orientating software system ‘Bitcoin-ng: a scalable blockchainprotocol’, Proceedings of
Engineering (IWBOSE), Campobasso, 2018, pp. 1-1. 13th USENIX Symposium on Networked Systems Design
30. T. N. Dinh and M. T. Thai,’ “AI and Blockchain: A turbulent andImplementation (NSDI 16), Santa Clara, CA, USA,
Integration,”’ vol. 51, no. 9, pp. 48-53, Gregorian calendar pp.45–59.
month 2018. 38. Fan, Z., Kulkarni, P., Gormus, S., Efthymiou, C., Kalogridis,
31. L. Kan, Y. Wei, A. Hafiz Muhammad, W. Siyuan, G. LinchaoJ. G.,Sooriyabandara, M., Zhu, Z.,Lambotharan, S. and Chin,
Fiaidhi, S. Mahomet and S. Mohammed,”” EDI with W.H. (2013) ‘Smart grid communications: overview of
Blockchain as associate Enabler for Extreme Automation,”’ in researchchallenges, solutions, and standardisation activities’,
IT skilled, vol. 20, no. 4, pp. 66-72, Jul./Aug. 2018. IEEE Communications Surveys andTutorials, Vol. 15, No. 1,
32. V. Gatteschi, F. Lamberti, C. Demartini, C. “Pranteda and V. pp.21–38.
Santamaría,”’ To Blockchain or to not Blockchain: that’s the 39. Miorandi, D., Sicari, S., Pellegrini, F.D. and Chlamtac, I.
Question,’” in IT skilled, vol. 20, no. 2, pp. 62-74, Mar./Apr. (2012) ‘Internet of things: vision, applicationsand research
2018. challenges’, Ad Hoc Networks, Vol. 10, No. 7, pp.1497–1516.
33. . T. A. Dinh, R. Liu, M. Zhang, G. Chen, B. C. Ooi and J. 40. Garay J, Kiayias A, Leonardos N. The bitcoin backbone
Wang, “Unraveling Blockchain: a knowledge process read of protocol: Analysis and applications. InAnnual international
Blockchain Systems,”’ in IEEETransactions on information conference on the theory and applications of cryptographic
and information Engineering, vol. 30, no. 7, pp. 1366-1385, techniques 2015 Apr 14 (pp. 281-310). Berlin, Heidelberg:
one July 2018. Springer Berlin Heidelberg.
34. N. Kshetri, “Can Blockchain Strengthen the net of Things?,”’ 41. Gervais A, Karame GO, Wüst K, Glykantzis V, Ritzdorf
inand H. Kai,”’ A Multiple Blockchains design on Inter- H, Capkun S. On the security and performance of proof of
Blockchain Communication,”’ 2018 IEEE International work blockchains. InProceedings of the 2016 ACM SIGSAC
Conference on software system Quality, responsibility and conference on computer and communications security 2016
Security Companion (QRS-C), L’isbon, 2018, pp. 139-145. Oct 24 (pp. 3-16).
35. Decker, C., Seidel, J. and Wattenhofer, R. (2016) ‘Bitcoin 42. Huckle S, Bhattacharya R, White M, Beloff N. Internet of
meets strong consistency’, Proceedings ofthe 17th things, Blockchain and shared economy applications. Procedia
International Conference on Distributed Computing and computer science. 2016 Jan 1;98:461-6.
Networking (ICDCN), ACM,Singapore, Singapore, p.13.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Fake news is spreading deception by changing people’s perspectives and knowledge. Social media and online forums
have helped spread fake news by mixing it with actual news. This paper presents novel text mining techniques and strategies for
detecting Fake and misinformation to decrease the hazards associated with its usage. We begin by describing the structure of the
suggested approach and the underlying conceptual method, providing implementations and verification using newspaper data.
We gathered genuine and bogus information and then translated it into a subject and event-based description from a manuscript
database. Fake news is being identified using a two-layered technique that includes identifying fake topics and false events.
The reliability of the proposed approach is proved by creating and validating an innovative False E-News Detector (FEND)
technology. Based on the provided threshold level of 0.6, the suggested methodology obtains 92.49 percent accuracy of the
classification and 94.16 percent recall.
Keywords: Fake news, Misinformation, Disinformation, Text mining techniques, Hazard reduction, Social networks, Internet
forums
1
shiva.csesrkr@gmail.com, 2mahesh.cse.srkr@gmail.com, 3mahesh_vvr@yahoo.com, 4nrusimhadri.silpa@gmail.com, 5kvssrmurthy75@gmail.com
DOI: 10.1201/9781003529231-45
290 Algorithms in Advanced Artificial Intelligence
stream used for media usage affects an individual’s decision- chamber” or “filter bubble” situation through social networks
making and alters one’s impressions of actual events. The magnifies the demand for acceptable news items. Consumers
influence is much more damaging at the organizational level using social media platforms prefer to carefully connect
because it jeopardizes their brand names and may influence with people, sharing their ideas and consuming material that
how their goods or services are consumed [4]. Because of appeals to their interests. The effect is amplified by social
increasing online media consumption and bots (e.g., Twitter media’s personalization options [8]. Thus, false information
bots) that automated data dissemination, news bulletins triggers consumers’ points of view to become even more
posted on social media worsened the problem. A recent polarized, increasing the risk of data polarization. The
study of verified fake news in the three months leading up data polarization impact is caused by differential incorrect
to the 2016 election found 38 million Facebook shares of information consumption caused by selective disinformation
presidential candidate support [4]. News verification systems exposure. Fake stories have been frequently followed with
have improved, addressing the need for automated methods fact-checks provided by various media outlets.
to detect false news from authentic news in the vast amount The distributed HDFS-Spark parallel computing architecture
of information [5]. Recent fake news disclosure methods fall helps the MLSRM capture and store web surfing data
into two categories based on methodology, language, and effectively. Later, MLSRM created a reinforcement method
networking tactics. Lexical methods (e.g., natural language to intelligently pick and integrate different Markov decision
or NLP) have investigated false information trends by processes to obtain actionable information to comprehend
examining underlying semantics. online user browsing behaviors with decreased state
On the other hand, network techniques use existing knowledge complexity and enhanced forecasting performance [9]. On
networks to verify truths. In numerous respects, our inquiry the other hand, internet access has promoted self-expression
contributed to the existence of awareness. Firstly, a unique and socialization online. One of the most modern social
analytic-based approach for spam detection is presented, networking networks, Twitter, produces gigabytes of data
which employs a topic-based categorization mechanism to daily. The current study investigates whether internet profiles
partition genuine content into various subject categories. and activities indicate people’s personalities [10]. SS Reddy
The news from each group seems to have a similar theme. et al. [11] proposed a sentimental analysis methodology for
Therefore, an event-extraction method extracts the actions assessing student comments utilizing extended short-term
from these media items. Secondly, by contrasting steps memory. The suggested model is compared to Naive Bayes,
obtained from the media to others in real news, we construct decision trees, and random forests. Based on a thorough
and execute a trustworthiness criterion for establishing the categorization process, the IDCS has four steps. IDCS begins
legitimacy of any data. by methodically collecting data from diverse sources on
rural women’s nutritional knowledge, attitudes, and dietary
2. Literature Review patterns. After that, it pre-processes the data according to
conventional methods and organizes it for categorization.
The creation and dissemination of fake news pose substantial Next, a learning algorithm develops an intelligent classifier
hazards from various viewpoints, particularly public to divide rural women into appropriate nutritional categories
protection. A great instance of this would be purposeful [12]. Attrition is the gradual loss of firm personnel without
misinformation that tries to affect a person’s opinion replacement. High attrition rates cause talent loss, inadequate
of another person or national polls. Politically divided research, and wasted training expenditures [13].
consumers in the US and Europe desire information from Social networking sites have become essential tools for
like-minded sources. It might be confirmation bias, or people to interact. Twitter is the most popular microblogging
“tunnel vision,” which involves creating one-sided scenarios service and a growing social network. Social media contains
based on previous preconceptions or ideologies [6]. Contrary much data in tweets, forums, status updates, comments,
to the confirmation bias concept, such research shows that etc. Applications may use sentiment analysis to process
people are misled by false information since they fail to and evaluate this data automatically [14]. The stock market
think logically during news rather than be motivated. [7] projection predicts the future value of equities exchanged
addresses different cognitive biases that operate as obstacles with another financial system [15]. Graders find delivering
to analyzing and resolving disinformation whenever humans comments with a steady interface, mentality, and deadline
analyze bogus information or disinformation. difficult. For optimum grading accuracy, a bag of words,
The rise of misleading information risks is misleading readers, sentences, and word count, average length, structure, and
preying on their need for pleasant information. Also, they organization of an essay are employed [16]. Organizations
lack logical thought while reading national media. The “echo struggle with employee attrition when skilled, technical, and
Mitigating Misinformation: An Advanced Analytics Framework for Proactive Detection of Fake News 291
essential people depart for better opportunities. Replacing examples of widely used network attributes. For instance, it
a qualified staff costs money. Thus, they examine current involves customer evaluation to eliminate disinformation in a
and previous employee data to determine prevalent attrition Parkinson’s disease-related social networking site community
factors [17]. Social networks like Twitter give user-generated [24].
news data. Clustering data and providing just important According to this study, disinformation in a discussion
information makes this resource valuable. For data filtering, forum depends on the publisher’s content and consumer
density-based k-means and graph clustering. After filtering, characteristics. Other research provides a technique for
we rank the data by keyword frequency, relevant key terms, evaluating the quality of replies in an internet crowd-sourced
and dataset key term similarity [18]. A student feedback survey report, the readability of the thread topics, and the
mining system (SFMS) uses text analytics and sentiment users’ ability to contribute usefully [25]. Unfortunately,
analysis to give educators quantifiable and in-depth previous sentiment and syntax analysis algorithms have been
analysis of qualitative student input, enhancing learning tailored to specific data categories, making them ineffective
experiences [19]. This takes time and effort. For this project, for detecting fake news.
this assessment method is automated. We use keywords as
features since machines can’t grasp our rating metrics. First, CNT orchestrates various ways for selecting weblog features
the student essay is matched against admin keywords. Articles to identify falsehoods [26]. Demonstrated that the optimum
with less than 20% similarity are not good enough [20]. features combined may recognize satirical information about
90 percent accuracy and 84 percent recall while evaluating
According to the “echo chamber” tendency, partisan news various selected features using only 360 news stories.
users systematically judge and share fact-checking content, However, those tactics might influence immoral authors to
as proven in research [21]. When it comes to the topic of generate false propaganda without displaying recognizable
fact-checking, various research on vote-based systems qualities. Evaluation of emotion and syntax Analysis of the
has yielded diverse outcomes. Whenever fact-checks of data and lexicon techniques are used to identify aberrant
disinformation provided individuals, a “backfire effect” data inside textual data with great accuracy [27]. Analytical
happens, in which they emotionally counter-argue & enhance statistics Hancock, Woodworth, and [28] suggested a method
their earlier incorrect impressions. However, [22] presented for studying the characteristics of crime stories. Their
no proof of true backfiring in a recent survey. Whereas fact- findings demonstrate that psychopaths’ speech contains the
checks could help rectify the news for the history, they are highest frequency of distortions and that they employ more
fruitless hazard evaluations. They are now almost entirely past tense words in stories than current tense words. [29] The
ineffectual in preventing the spread of incorrect information Word Vectors method developed the rhetorical formula, or
and data polarization in the first position. This highlights the RST, to discern the difference between authentic and false
urgent need for more reliable false information identification text information.
systems to stop the spread of misleading information.
Sentiment analysis is a frequently used method of identifying
dishonesty, specifically misleading Spam, in common.
3. Fake News Detection [30] suggested PU-learning identifies fraudulent Spam by
Consequently, social media websites are very cautious and analyzing real and fake views. Analysis of linguistic clues [31]
must start incorporating a false news recognition approach. shows that linguistic cues drawn from dishonesty concepts,
Cross-platform procedures, on the other hand, have gotten combined with information clues depending on textual
little consideration. Identifying fake news that originates from information, could be beneficial in identifying dishonest and
many websites could be an efficient approach for authorities. non-fraudulent projects on crowd financing policies.
Fake news is made by generating false information or Deep syntax analysis: PCFG (context-free probability
altering essential information. Fake news achieves credibility grammars) is a viable way to separate phrases into rewriting
by (1) copying well-known writers’ style of writing or (2) trees that describe syntactic structures using deep syntax
conveying viewpoints in a tone that is common in real news. research. [32], For instance, we examined syntactic stylometry
A growing number of false information detection algorithms for detecting fraud using hotel web data and characteristics
have lately been created. All present detection methods obtained from context-free grammar (i.e., CFG) parse trees.
could be classified into linguistic-based and network-based Unfortunately, existing prediction strategies have been
approaches [23]. Network properties are a helping element developed for specific data types or contextual factors, such
for different linguistic-based techniques in network-based as spam review site detection. [33] and spam mail recognition
methods towards fake media identification. Website data, and thus are insufficient for general-purpose spam filtering
editors’ data, timestamps, and other network properties are that might implement a variety of topics or difficulties.
292 Algorithms in Advanced Artificial Intelligence
Fig. 45.1 By sorting authentic news stories into newsgroups, the model-training methodology creates ground-truth sources of
knowledge
Mitigating Misinformation: An Advanced Analytics Framework for Proactive Detection of Fake News 293
The data pre-processing component utilizes text-processing phrase databases. Tokens are converted into binary vectors
techniques to retrieve subjects and occurrences from freshly during the segmentation procedure, which is then used to
gathered news data. Finally, per the obtained activities, the generate topics. We employed a weighting word technique
clustering component divides the media stories into distinct called Term Frequency Inverse Document Frequency (TF
groups. IDF) [52]. This methodology enables you to rate the term
frequency of tokens according to their significance in the
8. Method-data Processing and content. The TF-IDF technique leverages the sci-kit learn
library, a Python-based machine learning library linked with
Clustering the before pipeline documents. The term frequency (TF)
The unprocessed information gathered by the web search metric of every topic incidence inside a manuscript is graded
feeds the building of the ground-truth and misinformation by its significance in the TF-IDF technique (i.e., IDF).
databases in the fake news detection architecture (see also IDF evaluates the importance and produces a set of weights
Fig. 2). We construct a global search engine to extract news for every topic inside the database, whereas TF reflects the
from multiple sources of websites to be checked by the false term frequency of every subject. The following formula is
information detecting approach to undertake broad trials. used to calculate the unprocessed TF-IDF data.
Combining the triple extraction process with OIEs, phrase t f -idf(t, d,D) = t f(t, d) × idf(t,D)
text categorization, validation, stemmed, asset labeling,
The letters t, d, and D represent a theme, an object, and several
occurrence gathering, deconstruction, and subject data
documents in the repository, respectively. TF measures the
compression, we create a phrase pipeline (see Fig. 45.2). The
cost of each topic occurring in each post (t, d). Finally, IDF
internet crawler’s original data is then subjected to numerous
is calculated using the formula below. Lastly, the Euclidean
pre-processing modifications for annotating, topic collection,
distance norm is applied to the original TF-IDF values for
and event excavation. The Natural Language Toolkit is
normalization. The method of producing occurrences and
utilized with the Stanford CoreNLP package, which offers
themes from news stories is depicted in Figure 4. The triple
a framework for conducting a series of language annotating
data center serves as a link between the OIE instruments
processes such as lemmatization, tag, phrase check, stem,
and the word-processing pipelines. The word-processing
and portion tagging [50, 51].
pipeline extracts incidents into an events database. We create
a module that divides the incident database into a subject &
verb database [53].
It allows for the categorization of news items in later phases.
The concept database is used to fuel topic-based article
classification in a later stage. We employ two cluster analyses
to train the three independent dataset methods: k-means and
proximity propagation. This approach allows us to put our
fake stories identifying hypothesis, which treats individuals
as attributes of news articles, to the test. Adopting these
different methods can be explained in two ways [54]. To
begin with, unlike the more complicated procedures, those
Fig. 45.2 The following is the procedure for producing two clustering algorithms are easy to accomplish. Secondly,
activities and subjects with news stories we select supervised (K-means) and one unsupervised
(attached propagation) strategy for each technique grouping
Every article in the ground truth and fake media corpora is as the apogee. The following is how the k-means cluster
segmented into several phrases, subsequently transformed approach will work:
into several “tokens” – a single word or a string of different
1. Choose k centroids randomly.
continuous characters. The stemmed method takes the results
of the tokenization method to conduct feature extraction 2. Calculate the Distance Function among each place and
on every token produced by the tokenization procedure. the centroids, then note the current clusters.
By abbreviating words to sources, this method eliminates 3. Reconsider the distance among each cluster’s given
the repetition of infrequent word statistics. The triple data dataset and select new cluster centers.
repository links the OIE technologies and the phrase process. 4. Repeat Steps 2 and 3 in additional instances or when
The incident and phrase data are used to create the subject and the group is uniform.
Mitigating Misinformation: An Advanced Analytics Framework for Proactive Detection of Fake News 295
9. Explanation of the Findings Natural News is a website that has gained prominence due to
scientific deception but many theories.
Using two real-world news datasets, I ran thorough
experiments to demonstrate FEND’s improved performance 10. Evaluation
in this section quantitatively. I evaluated our suggested
technique using two data methodologies: consistency, As previously stated, we use two distinct clustering strategies,
reliability, recall rate, and F-score to determine its usefulness. K-means and Attraction Spreading, to retrieve news items as
Researchers use genuine news on CNN and the New York characteristics (AP).
Times to objectively test the effectiveness of the new strategy The number of clusters generated by both methodologies was
during training and testing. The significance of identification equal. Therefore, minor differences in the makeup of groups
is assessed using only a small number of testing datasets, created utilizing various methods may exist. The Euclidean
also listed in Table 3. Our news outlets, including CNN and distance is used across algorithms to classify into separate
the New York Times, were taken as gospel in this research groups. Both techniques provided similar findings to our
depending on qualities determined by a large research group. data. As a result, we concentrate on groups created via the
In contrast to their competition, they appear to see a much AP approach [57].
larger audience [55]. CNN, for instance, is seen by 96 million
remuneration homes, or 82.8 % of all couples with children Table 45.1 shows that themes have been recognized in every ten
who have access to a TV. groups & the overall number of subjects in every cluster. For
example, cluster 1 (see Table 45.2) may have the most items
Because CNN’s material contradicts users’ beliefs, only a (about 29,900), which contains terms like “international,”
tiny percentage of viewers may see CNN as a reliable source. residents,” working process,” leadership,” stocks,” Obama,”
On the other hand, this system offers a robust procedure that gov’t,” group,” American,” and so on. These topics are linked
allows users to include any reliable news source. Consumers to a different verb, making it much easier to categorize the
could establish an entire news archive by choosing from content. Cluster 20 is the shortest group, with around 5000
credible news sources. Regardless of the information sources, items, and covers themes including such “actress,” theory,”
our engine appears to be able to detect false propaganda. nominations,” film,” tribute,” portrayal,” rewards,” fortune,”
Researchers study a group of websites labeled as misleading spirit,” characters,” activities,” and so on [58].
information sites to acquire misinformation as underlying data
[56]. Amongst them are the websites www.greenvillegazette. The findings show that the first-level filter’s detection
com, www.politicot.com, www.advocate.com, and www. accuracy varied greatly among databases and had a large
naturalnews.com. The very first three sites disseminated detectability [59]. For instance, the advocate.com database
fake pro-Clinton disinformation, whereas the last two posts has the most excellent detection accuracy (66.9%), while
fake anti-Trump misinformation. Our unbiased & truthful the politicot.com database has the poorest (i.e., 4.4 percent).
evaluations are supported by data acquired from both pro- The first-level filtering accurately detects approximately 65
Clinton & pro-Trump websites. It must be remembered that percent of information on advocate.com, indicating that a
Table 45.2 The initial filtering found a large percentage of 3. Chen, C., Wang, Y., Zhang, J., Xiang, Y., Zhou, W., & Min, G.
bogus news items (2017). Statistical features-based real-time detection of drifted
twitter spam. IEEE Transactions on Information Forensics and
Advocate Naturalness Politicos Greenville
Security, 12(4), 914–925.
gazette
4. Cinque, G. (2014). The semantic classification of adjectives:
1 Uncertain 6543 2567 2955 1600 A view from syntax. Studies in Chinese Linguistics, 35(1),
News 1–30.
2 Fake 4534 498 145 500 5. Conroy, N. J., Rubin, V. L., & Chen, Y. (2015). Automatic
Topics deception detection: Methods for finding fake news.
3 Remaining 2245 1976 3000 1123 Proceedings of the Association for Information Science and
Data Technology, 52(1), 1–4. Del Corro, L., & Gemulla, R. (2013).
6. Fusilier, D. H., Montes-y Gómez, M., Rosso, P., & Cabrera, R.
considerable part of the media on advocate.com has bogus G. (2015). Detecting positive and negative deceptive opinions
using learning. Information Processing & Management, 51(4),
subjects (i.e., type-1 fake news).
433–443.
Natural news.com and greenvillegazette.com have relatively 7. Golbeck, J., Mauriello, M., Auxier, B., Bhanushali, K. H.,
low type-1 collaborative filtering levels (i.e., 21.1 percent Bonk, C., Bouzaghrane, M. A.,Everett, J. B., et al. (2018).
and 31.4 percent, respectively), suggesting that many articles Fake news vs satire: A dataset and analysis. In Proceedings of
reported on such sites are phony and have legitimate topics. the tenth ACM conference on web science (pp. 17–21). ACM.
Only 4.4 percent of the news about politicot.com is classified 8. Gross, M. (2017). The dangers of a post-truth world. Guess,
under type-1 false information, implying that practically A., Nyhan, B., & Reifler, J. (2018). Selective exposure to
misinformation: Evidence from the consumption of fake news
every one of the reports on politicot.com is reliable. The
during the 2016 US presidential campaign. Technical Report.
findings show that the very first filtering layer is a helpful Dartmouth College. https://www.dartmouth.edu/∼nyhan/
method for identifying a news source’s credibility [60]. The fake-news-2016.pdf
first layer’s output is fed into the second layer filters. The 9. Rao VV, Silpa N, Gadiraju M, Shankar RS, Vijaya K. An
second-level filtering identifies fake news’s believability and Optimal Machine Learning Model Based On Selective
makes it easier to compare specific false information ratings Reinforced Markov Decision To Predict Web Browsing
to the threshold. Patterns. Journal of Theoretical and Applied Information
Technology. 2023 Jan 31;101 (2): 859–73.
10. VVR MR, Silpa N, Gadiraju M, Reddy SS, Bonthu S,
11. Conclusion Kurada RR. A Plausible RNN-LSTM based Profession
As fake news becomes prevalent and hard to spot, better Recommendation System by Predicting Human Personality
detection methods are needed. Fake news misinforms users Types on Social Media Forums. In2023 7th International
Conference on Computing Methodologies and Communication
and targets, who may be persons or organizations. Misleading
(ICCMC) 2023 Feb 23 (pp. 850–855). IEEE.
information may cost organizations their competitive 11. Reddy SS, Gadiraju M, Maheswara Rao VV. Analyzing Student
advantage or reputation, while misleading statements can Reviews on Teacher Performance Using Long Short-Term
confuse people and affect their attitudes and decisions. Fake Memory. InInnovative Data Communication Technologies
news identification using novel analytical technologies is and Application: Proceedings of ICIDCA 2021 2022 Feb 24
presented in this research. FEND, which builds and verifies (pp. 539–553). Singapore: Springer Nature Singapore.
the false information identification framework, is then 12. Maheswara Rao VV, Silpa N, Mahesh G, Reddy SS. An
discussed. This system employs two layers to categorize. Enhanced Machine Learning Classification System to
False topic detection and tracking by the first and second Investigate the Status of Micronutrients in Rural Women. In
layers produce 92.49 percent accuracy. Our study is excellent Proceedings of International Conference on Recent Trends
since each news story is translated into actions instead of in Computing: ICRTC 2021 2022 (pp. 51–60). Springer
Singapore.
identifying bogus articles using syntactic rules or attitudes.
13. Shankar RS, Priyadarshini V, Neelima P, Raminaidu CH.
Analyzing Attrition and Performance of an Employee using
References Machine Learning Techniques. In2021 5th International
Conference on Electronics, Communication and Aerospace
1. Allcott, H., & Gentzkow, M. (2017). Social media and fake Technology (ICECA) 2021 Dec 2 (pp. 1601-1608). IEEE.
news in the 2016 election. Technical report. National Bureau 14. Kameswari KK, Raghaveni J, Shankar RS, Rao CS. Predicting
of Economic Research. Election Results using NLTK. International Journal of
2. Bird, S., Klein, E., & Loper, E. (2009). Natural language Innovative Technology and Exploring Engineering. 2019; 9:
processing with Python: Analysing text with the natural 4519–29.
language toolkit. O’Reilly Media, Inc..
Mitigating Misinformation: An Advanced Analytics Framework for Proactive Detection of Fake News 297
15. Jyothirmayee S, Kumar VD, Rao CS, Shankar RS. Predicting 29. Michalon, O., Ribeyre, C., Candito, M., & Nasr, A. (2016).
stock exchange using supervised learning algorithms. More profound syntax for better semantic parsing. In Cooling
International Journal of Innovative Technology and Exploring 2016. Miller, G. A. (1995). Wordnet: a lexical database for
Engineering. 2019; 9(1): 4081–90. English. Communications of the ACM, 38(11), 39–41.
16. Shiva Shankar R, Ravibabu D. Digital Report Grading Using 30. Nickerson, R. S. (1998). Confirmation bias: A ubiquitous
NLP Feature Selection. InSoft Computing in Data Analytics: phenomenon in many guises. Review of General Psychology,
Proceedings of International Conference on SCDA 2018 2019 2(2), 175–220. doi: 10.1037/1089-2680.2.2. 175.
(pp. 615–623). Springer Singapore. 31. Nyhan, B., & Reifler, J. (2010). When corrections fail: The
17. Shankar RS, Rajanikanth J, Sivaramaraju VV, Murthy KV. persistence of political misperceptions. Political Behavior,
Prediction of employee attrition using data mining. 2018, 32(2), 303–330. Open Sources (2017).
i.e., International Conference on System, computation, 32. Pennycook, G., & Rand, D. G. (2017). Who falls for fake
automation, and Networking (icscan) 2018 Jul 6 (pp. 1–8). news? The roles of analytic thinking, motivated reasoning,
IEEE. political ideology, and bullshit receptivity. SSRN Electronic
18. Shankar RS, Murthy KV, Rao CS, Gupta VM. An approach Journal, September, 1–63. doi: 10.2139/ssrn.3023545.
for extracting tweets from social media factors. In2018 ieee 33. Qazvinian, V., Rosengren, E., Radev, D. R., & Mei, Q. (2011).
international conference on system, computation, automation, Rumor has it: Identifying misinformation in microblogs.
and Networking (scan) 2018 Jul 6 (pp. 1–7). IEEE. In Proceedings of the conference on empirical methods in
19. Shankar RS, Srinivas LV, Ravibabu D, Raminaidu C. Novice natural language processing (pp. 1589–1599). Association for
Retroaction Report. ARPN Journal of Engineering and Computational Linguistics.
Applied Sciences. 2006; 13. 34. Rashkin, H., Choi, E., Jang, J. Y., Volkova, S., & Choi, Y.
20. Shankar RS, Babu DR, Murthy KV, Gupta V. An approach (2017). The truth of varying shades: Analysing language in
for essay evaluation using system tools. 2017 International fake news and political fact-checking. In Proceedings of the
Conference on Innovative Research In Electrical Sciences 2017 conference on empirical methods in natural language
(IICIRES) 2017 Jun 16 (pp. 1–9). IEEE. processing (pp. 2931–2937). Association for Computational
21. Hua, W., Wang, Z., Wang, H., Zheng, K., & Zhou, X. (2017). Linguistics. doi:10.18653/v1/ D17-1317.
Understand short texts by harvesting and analyzing semantic 35. Rubin, V. L., Chen, Y., & Conroy, N. J. (2015a). Deception
knowledge. IEEE Transactions on Knowledge and Data detection for news: three types of fakes. Proceedings of the
Engineering, 29(3), 499–512. Association for Information Science and Technology, 52(1),
22. Iyengar, A., Kalpana, G., Kalyankumar, S., & GunaNandhini, 1–4.
S. (2017). We integrated spam detection for multilingual 36. Rubin, V. L., Chen, Y., & Conroy, N. J. (2015b). Deception
emails. In Proceedings of the 2017 International Conference detection for news: Three types of fakes. In Proceedings of the
on Information Communication and Embedded Systems seventy-eighth ASIS&T annual meeting: Information science
(ICICES) (pp. 1–4). IEEE. with impact: Research in and for the community. In ASIST
23. Jang, S. M., Geng, T., Li, J.-Y. Q., Xia, R., Huang, C.-T., ‘15 (pp. 83:1–83:4). Silver Springs, MD, USA: American
Kim, H., & Tang, J. (2018). A computational approach for Society for Information Science.
examining the roots and spreading patterns of fake news: 37. Rubin, V. L., Conroy, N. J., Chen, Y., & Cornwell, S. (2016).
Evolution tree analysis. Computers in Human Behavior, 84, Fake news or truth? Using satirical cues to detect potentially
103–113. misleading information. In Proceedings of NAACL-HLT
24. Jin, Z., Cao, J., Jiang, Y.-G., & Zhang, Y. (2014). News (pp. 7–17).
credibility evaluation on microblog with a hierarchical 38. Rubin, V. L., & Lukoianova, T. (2015). Truth and deception
propagation model. In Proceedings of the 2014 IEEE at the rhetorical structure level. Journal of the Association for
International Conference on Data Mining (ICDM) (pp. 230– Information Science and Technology, 66(5), 905–917.
239). IEEE. 39. Sahu, I., & Majumdar, D. (2017). We are detecting factual
25. Klein, D. O., & Wueller, J. R. (2017). Fake news: A legal and non-factual content in news articles. In Proceedings
perspective. Journal of Internet Law, 20(10), 1,6–13. of the fourth ACM IKDD conferences on data sciences. In
26. Lau, R. Y. K., Zhang, W., & Xu, W. (2018). Parallel aspect- CODS ‘17 (pp. 17:1–17:12). New York, NY, USA: ACM.
oriented sentiment analysis for sales forecasting with big data. doi:10.1145/3041823. 3041837.
Production and Operations Management, 27(10), 1775–1794. 40. Shin, J., & Thorson, K. (2017). Partisan selective sharing: The
doi:10.1111/poms.12737. https://onlinelibrary.wiley.com/doi/ biased diffusion of fact-checking messages on social media.
abs/10.1111/poms.12737 Journal of Communication, 67(2), 233–255. doi:10.1111/
27. Li, H., Gupta, A., Zhang, J., & Flor, N. (2018). Who will use jcom.12284.
augmented reality? An integrated approach based on text 41. Siering, M., Koch, J.-A., & Deokar, A. V. (2016). Detecting
analytics and field survey. European Journal of Operational fraudulent behavior on crowdfunding platforms: The role
Research. doi:10.1016/j.ejor.2018.10.019. of linguistic and content-based cues in static and dynamic
28. Lin, Y.-S., Jiang, J.-Y., & Lee, S.-J. (2014). A similarity measure contexts. Journal of Management Information Systems, 33(2),
for text classification and clustering. IEEE Transactions on 421–455.
Knowledge and Data Engineering, 26(7), 1575–1590.
298 Algorithms in Advanced Artificial Intelligence
42. Silverman, C. (2015). Lies, damn lies, and viral content: How 52. Wu, H. C., Luk, R. W. P., Wong, K. F., & Kwok, K. L.
news websites spread (and debunk) online rumors, unverified (2008). Interpreting TF-IDF term weights as making relevant
claims, and misinformation. Technical Report. New York, decisions. ACM Transactions on Information Systems (TOIS),
NY: Tow Center for Digital Journalism, Columbia Journalism 26(3), 13.
School, Columbia University. 53. Xavier, C. C., & de Lima, V. L. S. (2014). Boosting open
43. Socher, R., Perelygin, A., Wu, J., Chuang, J., Manning, C. information extraction with noun-based relations. In
D., Ng, A., & Potts, C. (2013). Recursive deep models for Proceedings of the LREC (pp. 96–100).
semantic compositionality over a sentiment treebank. In 54. Xu, J., & Taft, M. (2015). The effects of semantic transparency
Proceedings of the 2013 conference on empirical methods in and base frequency on recognizing English complex words.
natural language processing (pp. 1631–1642). Journal of Experimental Psychology: Learning, Memory, and
44. Stepinski, A., & Mittal, V. (2007). A fact/opinion classifier for Cognition, 41(3), 904.
news articles. In Proceedings of the thirtieth annual international 55. Yang, X.-F., & Siu, W.-C. (2017). Vehicle detection under
ACM sigir conference on research and development in tough conditions using prioritised feature extraction with
information retrieval. In SIGIR ‘07 (pp. 807–808). New York, shadow recognition. In Proceedings of the 2017 twenty-
NY, USA: ACM. doi:10.1145/1277741.1277919. second international conference on digital signal processing
45. Swartz, J. (2017). The worldwide web’s inventor warns it’s (DSP) (pp. 1–5). IEEE
in peril on 28th anniversary. USA Today. www.usatoday. 56. Deshai N, Sekhar B V D, VenkataRamana S, S, Srinivas K &
com/story/tech/news/2017/03/11/ world-wide-webs-inventor Varma G P S, Big data hadoop map reduce job scheduling: a
warns-s-peril/99005906/ short survey, Advances in Intelligent Systems and Computing,
46. Tang, R., Ouyang, L., Li, C., He, Y., Griffin, M., Taghian, vol 862. Springer, Singapore, 2019, 349-365, Available from:
A., . . . Hughes, K. (2018). Machine learning to parse breast 10.1007/ 978-981-13-3329-3_3.3 4
pathology reports in chinese. Breast Cancer Research and 57. Deshai N, Sekhar B V D S, Venkataramana S, Chakravarthy
Treatment, 1–8. V V S S S & Chowdary P S R, Study with comparing bigdata
47. Venkatesan, S., Han, W., Kisekka, V., Sharman, R., Kudumula, handling techniques using apache hadoop map reduce Vs
V., & Jaswal, H. S. (2013). Misinformation in online health apache spark, Int J Eng Technol, 7(4) (2018) 4839–4843,
communities. In WISP 2012 Proceedings (p. 28). Available from:10.14419/ijet.v7i4.1.15997 5
48. Venkatesan, S., Han, W., & Sharman, R. (2014). A response 58. Mahesh G, Shankar Reddy S, Maheswara Rao VV, Silpa N.
quality model for online health communities. In Proceedings Preeminent Sign Language System by Employing Mining
of the thirty-fifth international conference on information Techniques. InInternational Conference on IoT Based Control
systems (p. 28). Networks and Intelligent Systems 2023 Jun 21 (pp. 571-588).
49. Tsai, M.-F., & Wang, C.-J. (2017). On the risk prediction Singapore: Springer Nature Singapore.
and analysis of soft information in finance reports. European 59. Deshai N, Venkataramana S & PardhaSaradhiVarma G,
Journal of Operational Research, 257(1), 243–250. Performance and cost evolution of dynamic increase hadoop
50. Wang, P., Xu, B., Xu, J., Tian, G., Liu, C.-L., & Hao, H. workloads of various datacenters, Smart Innovation, Systems
(2016). Semantic expansion using word embedding clustering and Technologies, 105 (2019) 505–516. Available from:
and convolutional neural network for improving short text 10.1007/978-981-13-1927-3_54.
classification. Neurocomputing, 174, 806–814. 60. Deshai N, SaradhiVarma G, P & Venkataramana S, A study
51. Wei, T., Lu, Y., Chang, H., Zhou, Q., & Bao, X. (2015). A on analytical framework to breakdown conditions among data
semantic approach for text clustering using wordnet and lexical quality measurement, International Conference on Innovative
chains. Expert Systems with Applications, 42(4), 2264–2275. Research in Science and Technology, 7, 2018 Available from:
Wood, T., & Porter, E. (2018). The elusive backfire effect: 10.14419/ijet.v7i1.1.9276 10.
mass attitudes’ steadfast factual adherence. Political Behavior,
Note: All the figures and tables in this chapter were designed by
1–29.
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Nilambar Sethi1
Department of Computer Science and Engineering,
GIET University, Gunupur, Odisha, India
V. Sivarama Raju Vetukuri2, R. Shiva Shankar3
Department of Computer Science and Engineering, SRKR Engineering College,
Bhimavaram, Andhra Pradesh, India
R. Rajender4
Department of Computer Science and Engineering,
LENDI Institute of Engineering and Technology, Vizianagaram, Andhra Pradesh, India
Abstract: The exponential rise of internet textual data necessitated a sophisticated tool that automatically summarises material
while keeping key information. Text summarization is essential in today’s age of massive data sets to extract relevant material
and display accurate, intelligible information. Over time, several ways have been devised to summarise material. By extracting
terms from text, conventional approaches build redundant summaries and ignore document summary relationships. Text
summarization is an integral part of Natural Language Processing that helps people comprehend text. AI uses natural language
processing to find important information quickly while preserving context. Deep learning is used to extract key phrases and
summarise them. In this article, we discuss the techniques of extractive and abstractive text summarization using Text Rank and
Encoder-Decoder LSTM. These techniques are beneficial because they do not require pre-defined features or domain-specific
knowledge and can be applied across different domains. Moreover, to overcome the lack of labeled data, we evaluate and score
the training set phrases by comparing them to human reference descriptions. Our experimental assessments demonstrate the
efficacy of our suggested methodologies compared to other baselines.
Keywords: Extractive text summarisation (ETS), Abstractive summarization (AS), Natural language processing (NLP), Deep
learning (DL), Long short-term memory (LSTM)
1. Introduction textual material into a shorter version that may include all
relevant and vital information about the content. Information
Over the last several years, articles and links have increased, loss occurs due to compression. Loss. Medical records,
making searching for helpful information and displaying it meteorological data, news summaries, etc., are successfully
harder. As data grows, semantic density increases. Therefore, summarised using text summarising [1]. Here are some broad
the necessity to quickly distinguish the most significant items types of text summarisation:
arises. The Summary helps determine whether the article’s 1. Extractive text summarisation: Document items are
condensed text is relevant. Text summary involves selecting a extracted without alteration.
portion of a material to represent it. The text summary reduces
1
nilambar@giet.edu, 2sivaramaraju.vetukuri@gmail.com, 3shiva.csesrkr@gmail.com, 4rajender.renuguntta@gmail.com
DOI: 10.1201/9781003529231-46
300 Algorithms in Advanced Artificial Intelligence
One major exam is essay writing. Manual assessment is used information in digital documents. They addressed automated
now. An automatic check is used for this project. The essay text summarising research approaches. A graph-based
is poor if the similarity is under 20%. Other methods include report system by Kavita Ganesan et al. [31] creates concise
segmentation, stop words, word frequency computation, theoretic summaries of excessively repetitious viewpoints.
numerical attributes, sentence validation, and spell check. The network organization is distinct, with nodes representing
Assessment tools exist, but our process is faster and more word units and directed edges indicating sentence structure.
accurate [18]. Students will provide comments online using Nodes have positional data. N.Moratanch and S.Chitrakala
a standard form. The suggested approach prioritizes security [32] suggested text summarising extracts of short
by allowing only legitimate users to see and understand the information from the web, Twitter, etc. Text summarisation
cumulative input and opinions of a batch of pupils [19]. might be extractive or abstractive. Different methods have
They organized the data into clusters and provided just the been used to evaluate extractive summarisation issues.
helpful information for this resource to be valuable. For this ETS uses supervised and unsupervised learning to shorten
purpose, they filtered the data using a density-based k-means essential phrases and paragraphs based on word and sentence
technique and a graph clustering algorithm. Following the properties. Abstractive text summarisation [33] is coherent,
application of the filter, the data is based on the occurrence of less hesitant, and information-rich. The two main abstractive
keywords, the relevance of key phrases, and, ultimately, the strategies are organized and segmented. The structured
degree to which key terms appear similarly across the dataset method leverages previous knowledge, while the semantic
[20]. approach uses NLP. The advanced abstractive model
Maaloul et al. [21] found 19 rhetorical linkages in a corpus- summarises by comprehending the source text. Abstractive
based analysis, with some similarities to Mathkour et al. summarization is clear and linguistically accurate.
[22]. Secondly, they began with the nine best summary Tooba Siddiqui and Jawwad Ahmed Shamsi [34] said that
connections. Al-Thanyyan and Azmi [23] suggested hybrid abstractive approaches are better than extractive methods
two-pass summarisation. A primary summary is created since they mimic human summarization. AS often generates
from the RS tree’s early levels, while a shorter summary is new sentences using neural networks. The repeating words
produced in the second pass. According to the researchers, issue was solved via temporal attention. The headline-
the two-pass summariser improves RST. This method extracts generating algorithm was trained and evaluated. They used
relevant sentences by utilizing Arabic NLP techniques. The global and temporal attention models. According to A, the
discriminant analysis commences with mRMR [24], and attention technique gives the decoder immediate access to
based on mRMR scores, it ranks the sentences according to the input sequence instead of a fixed-length context vector.
the strength of their discriminant words. P Patil et al., automatic Summary condenses material into
Data feature selection and analysis can reduce staff turnover a concise form [35]. Finding a practical approach is the
[25]. The grading process considers the essay’s average central issue. This study presents our sentimental analysis
length, structure, organization, bag of words, sentence, and methodology for analyzing student comments utilizing
word count. Sequential forward feature selection is used to LSTM [36]. Single or several papers may be summarised.
select the best candidate subset [26]. In addition, this study Redundancies and unnecessary data may be deleted, saving
employs Machine Learning to predict the stock performance users time. Extractive and abstractive methods were utilized
of equities traded on a financial transaction involving another to summarise and enhance the material. Text Rank algorithm
economic system [27]. Lastly, the grading process considers ranks sentences. The text’s most crucial sentences are
the essay’s average length, structure, organization, bag of selected—lexical database abstracted Summary. Word ranks
words, sentence count, and word count and then selects are awarded to words, not text. Preprocessors break text
into sentences and remove stop words and stems. Important
Chandhana Surabhi. M [29] said NLP makes machines act phrases are picked in the extraction summary, and a few words
like humans. It facilitates human-machine communication. are substituted with incorrect synonyms during abstraction.
NLP has many daily uses. NLP processes vast texts. The algorithmic nature of this strategy is a benefit [37].
Category classification, indexing and searching huge texts, Shuai Wang et al. [38] explained that extended text requires
machine translation, and information extraction are required. two stages. Phases are extraction and abstraction. Sentence
A language comprehension software must grasp language extraction employs a graph model to extract essential
structure, including words and how they form phrases and sentences. Summaries are generated using recurrent neural
sentences. It should understand sentence meaning and network LSTM encoder-decoder at the abstraction phase.
context. The software must understand human thought Top-ranked sentences are combined to produce a summary.
and the world. User-friendly NLP systems are the future. Abstractive text summarisation uses Seq2Seq.
Vipul Dalal and Latesh Malik [30] suggested discovering
302 Algorithms in Advanced Artificial Intelligence
Jaradat [39] combines HS with summarising to deliver the For this project, we are using the Encoder-Decoder model.
nearest approximation to the optimal document summary The encoder-decoder approach helps with sequence-to
using the E ASC corpus and the ROUGE toolbox to analyze sequence challenges like text summarisation. It incorporates
the method. The findings demonstrated that the proposed a total of 2 RNNs. The first performs the role of encoder,
method performed much better than other contemporary while the second decodes information. We use Encoder-
alternatives. Jaradat and Al-Taani [40] developed a hybrid Decoder LSTM to predict variable-length output for variable-
single-document extractive Arabic text summary using length inputs. The encoder is stacked LSTM. Stacked LSTMs
evolutionary algorithms. Al-Zahrani et al. [41] found that use output sequences as input to the next LSTM. Due to poor
PSO outperformed several Arabic summarization methods. performance with extended input or output sequences, this
The particle selection operator (PSO) selects the particle with Encoder-Decoder model uses the Attention Layer. Using
the best combination of eight structural attributes required the Attention layer, we focus on essential data and ignore
by Arab summarizers. This is achieved by training on data unnecessary data. Our model uses early stopping to halt
from the Essex Arabic summaries corpus. Each PSO iteration training when validation loss increases.
scores and ranks input text sentences based on selected
characteristics and weights, producing an output summary 3.1 Long Short Term Memory (LSTM)
of the top-ranked phrases. The experiments show that Arabs LSTMs [24] are a type of RNN that can learn how things
synthesize texts by concentrating on the opening phrase of depend on each other over time. Only current information
each paragraph. Mahjoub [42] suggested using the PSO is necessary when predicting the final word in a sentence.
algorithm to automatically derive summaries from Arabic In circumstances where there is not much of a time lag
single texts. The proposed method for obtaining Arabic single between the relevant information and the location at which
document summaries worked in experiments. it is required, RNN can learn how to utilize the historical
data without encountering the issues outlined before. Extra
3. Methodology context may be necessary when essential information is
distant. In cases like this, LSTM networks are adequate.
Hierarchical learning, also known as deep learning, involves
using a neural network with multiple layers for processing
data, including linear and non-linear transformations that
identify high-level data abstractions. An observation could
be described as a collection of edges, regions of a specific
geometry, or a vector of intensity values for each pixel.
Learning may be made more straightforward using certain
representations. Deep learning can supplant traditionally
hand-crafted features with more time- and cost-effective
unsupervised, semi-supervised, or fully-supervised feature
learning and hierarchical feature extraction techniques. This
research aims to enhance representations and develop models
that can effectively learn from vast amounts of unlabeled Fig. 46.3 System flow for LSTM
data. Neural networks receive data as input to recognize
The structure of the LSTM is analogous to that of a chain,
patterns and predict output.
and each iteration of the repeating module comprises four
Three layers make up the network. Text, pictures, and audio interacting modules. The four components of this system are
are submitted to the submitted Layer. Numerical values are the cell state, the output gate, the update gate, and the forget
connected to the Neurons that make up the Hidden Layer. gate. Within the LSTM repeating module, the horizontal
These are the premises upon which the calculations are running layer is used to indicate the current state of the
carried out. The output is then sent to the Output Layer as cell. This Layer acts as a conveyor belt and interacts very
the last step. A Recurrent Neural Network (RNN) gets its little with the other levels; it may be seen in Fig. 46.3. The
current state input from its previous state output. This is following is an explanation of the fundamental structure of
applied in sequence-based state prediction issues. It features an LSTM module:
a Hidden Layer that can remember the order of things. Forget Gate: It decides which information from the cell state
RNN has a drawback: Vanishing gradient descent. LSTM is should be deleted or forgotten. It does this by analyzing all
implemented so that this problem may be solved. LSTM is
of the data. It makes use of a sigmoid activation function
Long Short-Term Memory. RNNs like LSTM overcome the
and accepts as input both the currently provided input (x_t)
vanishing gradient issue. It contains gates, which allow it to
and the previously used hidden state (h_(t-1)). The output of
circumvent this issue.
Summarization of Legal Texts by Using Deep Learning Approaches 303
the forget gate determines which aspects of the cell state are Symptoms of Malaria usually appear about 10 to 15 days
forgotten and which are remembered. It also specifies which after an infected mosquito bites a person.
elements of the cell state should be kept. The tokenization of sentence: The sentences must be
Input Gate: It is in charge of determining which pieces of tokenized following the provision of text as input. The process
information from the currently active time step should be of separating (or dividing) a text into a list of tokens is known
added to the state of the cell. A sigmoid activation function as tokenization. When referring to a paragraph, sentences
is used. It accepts as input the presently available input data are regarded as the tokens, but words are considered the
(x_t) as well as the prior concealed state (h_(t-1)), and the tokens when referring to a phrase. We are using sentence
information that should be saved is determined based on the tokenization in this process.
outcome of this process. After tokenization: [“mosquito-borne” illness malaria affects
Output Gate: It is responsible for determining which aspects people and animals. ‘Fever, fatigue, vomiting, and headaches
of the current cell state should be used to compute the hidden are classic malaria symptoms. Yellow skin, convulsions,
state for the next time step. It uses a sigmoid activation coma, and death may result. After a mosquito bite, symptoms
function along the same lines as the input and forget gates. It appear 10–15 days later.’]
considers the most recent hidden state and information (x_t Removal of stop words: Here, the removal of stop words
and h_(t-1), respectively). A value is generated by the output is removing commonly used words. NLTK includes a list of
gate, which feeds into the hidden state’s computation. stop words in its corpus (collection of different language data
sets). We remove stopwords from our text.
3.2 System Model for Extractive Summerization
Before removing stop words: [‘Malaria is an infectious disease
Extractive summarisation is a valuable tool for quickly
that spreads through mosquitoes and can infect humans and
digesting large volumes of text. It can be helpful when
other animals. The common symptoms of Malaria include
maintaining the original wording and context, which is
fever, fatigue, vomiting, and headaches. In severe cases, it
essential, such as in legal documents or scientific papers.
can cause yellow skin, seizures, coma, and even death. The
However, it may not provide summaries as fluent or concise
symptoms typically appear a few days after being bitten by
as those generated by abstractive summarisation, which can
an infected mosquito.’]
rewrite and rephrase content. The entire flow is shown in
Fig. 46.4. After removing stop words: [‘Malaria infectious disease
spreads mosquitoes infect humans animals, common
Preprocess the symptoms malaria include fever, fatigue, vomiting,
Input Text
Text headaches, severe cases, cause yellow skin, seizures, coma,
death symptoms typically appear days bitten infected
mosquito’]
Lemmatization Lemmatization of sentences: Reducing a word to its
simplest possible form is known as lemmatization. It
analyses the situation and then reduces the term to its most
fundamental and relevant form. Lemmatization with a pos
Cosine Similarity tag is used. Pos means parts of speech. We are considering
four parts of speech: noun, Verb, Adjective, and Adverb. POS
tags are given to words, and we lemmatize the word using it.
Print Calculate Text Finally, the lemmatized sentences are obtained.
Summary Rank ‘caring’----->lemmatization----->’care’
Importing required modules for the model: After splitting is relatively low, both in terms of assessing and producing
data, we have to import the modules needed for building them. Compared to the outline form, summaries might be
the model, such as Embedding, TimeDistributed, Model, challenging to discern the substance of the system.
and EarlyStopping. They are low-dimensional and help to (i) Human Evaluation
represent data in a more meaningful way.
Because of the inherent subjectivity of human judgment when
Tokeniser: We use Keras Tokenizer for our model. This determining what constitutes a “good” outline, developing an
Tokenizer turns each text into an integer sequence or a vector. automated system for conducting analysis is an exceptionally
Building the model: We use Encoder-Decoder LSTM, which challenging endeavor. This means that manual analysis
supports variable length input sequences and predicts variable involves a lot of work. Coherence and coverage are two
length output sequences. It is mainly used for sequence-to further concerns that need to be addressed.
sequence models. Natural Language Processing utilizes a (ii) Recall-Oriented Understanding for Gisting Evaluation
specific architecture, which involves encoding data using (ROUGE)
Stacked LSTM as the first step. The output of one LSTM is
then fed into the next LSTM in the stack, with three LSTMs The value of Count (N-gram) represents the total number of
used to encode the data. We define the model and compile N-grams included in the reference summary.
it. Then, we apply EarlyStopping to our model, which stops ROUGE N
the training of a dataset at a point to overcome the problem
of overfitting. Next, we fit the model and evaluate it. At last,
ÂS Œ reference_summaries N gramsCount match (N gram)
=
we decode the model using embeddings and hidden states. ÂS Œ reference_summaries N gramsCount(N gram)
The decoder reads through the vector and translates it to an
output sequence. Where n is the total length of the N-gram sequence used in
Output summary: Print the outline for the test data. We the algorithm. The maximum number of n-grams that appear
can see the original Summary and predicted Summary in the simultaneously in both a candidate summary and a set of
output. reference summaries is known as the co-occurrence limit.
(iii) Recall R =
|Sref « Scand |
4. Results |Sref |
Our qualitative analysis reveals that key phrases are Where Sref ∩ Scand Represents the number of sentences that
picked during extraction, and certain words are substituted are included in the candidate and reference summaries.
with suitable synonyms during abstraction. However, the
abstraction phase is now replaced sparsely. The present (iv) Precision ( P ) P =
|Sref « Scand |
system effectively analyses news items, technical papers, |Scand |
and encyclopedia entries. The extraction may not provide 2(Precision)(Recall)
comprehensible summaries in essays, novels, and publications (v) F measure F =
Precision + Recall
with much direct speech. We analyzed the performance of our
system against specific reference summaries. News stories (vi) Compression Ratio Cr = Slen ⋅ Tlen
were evaluated, with several given below and their findings. Where, Slen and Tlen are the length of summaries
The average accuracy, recall, and fscores from the ROUGE Finally, the text rank score was evaluated for the given
assessment measure are shown. When an algorithm has high sentences; the sentence repeated more times was estimated
accuracy, it means that it returns more relevant results than it as a word cloud of top-ranked sentences, a scatter plot of
did irrelevant ones. When it has a high recall, it means that it sentence length vs. Text Rank Score was also obtained.
recovered the majority of the relevant results. Sentence Similarity Heatmap visualization was drawn and
Various benchmarking datasets [1] are used for experimental finalized. The top 2 sentences by Text Rank Score were
assessment of extractive summarisation. The most popular obtained for ETS, and the graphs are shown in Fig. 46.6. A
kind of benchmarking is Document Understanding graph was drawn for Loss Vs Val_Loss for AS. The metrics
Conferences (DUC) datasets used for text summarising. It obtained were also shown in Fig. 46.7 as Graphs.
includes both the original materials and summaries of those
documents produced electronically, by hand, and by user 5. Discussion
submissions are the summaries [20]. Based on the literature
reviewed in the articles, it has been discovered that human In this part, the outputs of the suggested model are discussed,
beings tend to agree on things. The value of summarisers and an investigation into the possible explanations for why
306 Algorithms in Advanced Artificial Intelligence
Fig. 46.8 Screen short for test case: original data with extractive summary
Fig. 46.9 Screen short for test case: original data with abstractive summary
308 Algorithms in Advanced Artificial Intelligence
some findings are presented as they are is carried out. The 6. Conclusion
data are shown in the order obtained from the model. In
terms of the overall score, the findings of the examination In this analysis, many mechanisms of the process of
were quite encouraging. The performance of our system extractive text summarisation have been shown. The
was examined in terms of its ability to summarise single extractive summarising method is very coherent, has fewer
documents as well as several documents at once. Regarding unnecessary steps, and helps create a cohesive whole rich
recollection, our strategy is superior to the strong baseline in Summary and information. The objective is to provide
of lead sentences for single-document summary tasks. As an in-depth analysis and compare the various strategies and
LSTMs are effective memory models, their usage in this procedures used in the extractive text summarising process.
strategy might be beneficial in classifying sentences into The study has not fully addressed the issues of extractive
summaries by incorporating contributions from prior phrases. text summarisation, including time and space constraints.
The evaluation process for created summaries may be In extractive text summarisation, we generated a summary
broadened and include human grading. Contributions are that contains the most important sentences in a text. The
appreciated and accepted in data set gathering and creation. summarization is done by using the TextRank algorithm. In
Note that we used the ROUGE methodology to compare this work, we designed an Encoder-Decoder LSTM model to
our summaries with those humans who wrote them after convert text into a summary. We took a dataset and generated
completing an automatic review procedure. ROUGE (Recall- a summary of the dataset’s test data. We got good results for
Oriented Understudy for Gisting) is a metric that mechanically our model, which we can see in the graph presented in the
assesses text summarization by comparing it to human- Result Analysis. We conclude that our model can be used for
created gold standards. This metric counts overlapping units generating a summary. This model has future scope, and we
between the produced and ideal summaries to evaluate the can improve it further. The model can be improved by adding
summarization quality. more LSTM layers in the stacked LSTM and can implement
packages and concepts that can be built in the future in deep
The application uses ROUGE-N (containing ROUGE1 learning.
and ROUGE2) and ROUGE-L text summarizing methods
from several research. ROUGE-L determines the longest
common substring, whereas ROUGE-N measures n-gram References
recall. However, ROUGE1 and ROUGE2 assess unigram 1. Y. Zhang, J. Liao, J. Tang, W. Xiao, Y. Wang, Extractive
and bigram recall. In the program’s manual, ‘S’ indicates the document summarisation based on hierarchical gru, in: 2018
reference summary and ‘n’ the n-gram length and refers to International Conference on Robots & Intelligent System
the maximum number of word matches between reference (ICRIS), IEEE, pp. 341–346
and produced summaries using n-gram analysis. refers to the 2. M. Allahyari, S. Pouriyeh, M. Assefi et al., “Text summarisation
total number of n-gram words in the reference summary. Four techniques: a brief survey,” International Journal of Advanced
Rouge measurements are listed: Computer Science and Applications, vol. 8, no. 10, 2017.
3. A. B. Al-Saleh and M. E. B. Menai, “Automatic Arabic text
1. ROUGE-N: Statistics on N-gram co-occurrence, summarisation: a survey,” Artificial Intelligence Review, vol.
produced candidate summary, and referenced 45, no. 2, pp. 203–234, 2016.
summaries are compared using n-gram recall. 4. A. Turpin, Y. Tsegay, D. Hawking, and H. E. Williams, “Fast
2. ROUGE-L: The maximum length of the longest generation of result snippets in web search,” in Proceedings
common subsequence between sequences P and Q. of the 30th Annual international ACM SIGIR Conference
on Research and Development in information Retrieval
3. ROUGE-W: Weighted Longest Common Subsequence
SIGIR’07, p. 127, Amsterdam, *e Netherlands, 2007.
takes match lengths into account to improve the 5. Q. A. Al-Radaideh and D. Q. Bataineh, “A hybrid approach
accuracy of ROUGE-L. for Arabic text summarisation using domain knowledge and
4. ROUGE-S: Skipping Bigram Co-Occurrence Statistics genetic algorithms,” Cognitive Computation, vol. 10, no. 4,
Measures produced-referenced summaries overlap. pp. 651–669, 2018.
6. C. Sunitha, A. Jaya, and A. Ganesh, “A study on abstractive
We have done the process for ETS and AS by using the
summarisation techniques in Indian languages,” Procedia
paragraphs. We have shown a test case by taking a paragraph Computer Science, vol. 87, pp. 25–31, 2016.
containing words 138, sentence 11, and several characters 7. D. R. Radev, E. Hovy, and K. McKeown, “Introduction to the
627. This entire process is shown on Figs 46.6–46.9. special issue on summarisation,” Computational Linguistics,
vol. 28, no. 4, pp. 399–408, 2002.
Summarization of Legal Texts by Using Deep Learning Approaches 309
8. A. Khan and N. Salim, “A review on abstractive summarisation 23. Azmi, A.M., Al-Thanyyan, S., 2012. A text summariser for
methods,” Journal of @eoretical and Applied Information Arabic.Comput. Speech Lang. 26 (4), 260–273.
Technology, vol. 59, no. 1, pp. 64–72, 2014. 24. Peng, H., Long, F., Ding, C., 2005. Feature selection based
9. N. Moratanch and S. Chitrakala, “A survey on abstractive on mutual information criteria of max-dependency, max-
text summarisation,” in Proceedings of the 2016 International relevance, and minredundancy. IEEE Trans. Pattern Anal.
Conference on Circuit, Power and Computing Technologies Mach. Intell. 27 (8), 1226–1238.
(ICCPCT), pp. 1–7, Nagercoil, India, 2016. 25. Shankar RS, Rajanikanth J, Sivaramaraju VV, Murthy KV.
10. S. Shimpikar and S. Govilkar, “A survey of text summarisation Prediction of employee attrition using datamining. In2018 ieee
techniques for Indian regional languages,” International international conference on system, computation, automation
Journal of Computer Applications, vol. 165, no. 11, pp. 29–33, and networking (icscan) 2018 Jul 6 (pp. 1–8). IEEE.
2017. 26. Shiva Shankar R, Ravibabu D. Digital Report Grading Using
11. N. R. Kasture, N. Yargal, N. N. Singh, N. Kulkarni, and NLP Feature Selection. InSoft Computing in Data Analytics:
V. Mathur, “A survey on methods of abstractive text Proceedings of International Conference on SCDA 2018 2019
summarisation,” International Journal for Research in (pp. 615-623). Springer Singapore.
Emerging Science andTechnology, vol. 1, no. 6, p. 5, 2014. 27. Jyothirmayee S, Kumar VD, Rao CS, Shankar RS. Predicting
12. P. Kartheek Rachabathuni, “A survey on abstractive stock exchange using supervised learning algorithms.
summarisation techniques,” in Proceedings of the 2017 International Journal of Innovative Technology and Exploring
International Conference on Inventive Computing and Engineering. 2019; 9(1): 4081–90.
Informatics (ICICI), pp. 762–765, Coimbatore, 2017. 28. Kameswari KK, Raghaveni J, Shankar RS, Rao CS. Predicting
13. S. Yeasmin, P. B. Tumpa, A. M. Nitu, E. Ali, and M. I. Election Results using NLTK. International Journal of
Afjal, “Study of abstractive text summarisation techniques,” Innovative Technology and Exploring Engineering. 2019; 9:
American Journal of Engineering Research, vol. 8, 2017. 4519–29.
14. A. Khan, N. Salim, H. Farman et al., “Abstractive text 29. Surabhi MC. Natural language processing future. In2013
summarisation based on improved semantic graph approach,” International conference on optical imaging sensor and
International Journal of Parallel Programming, vol. 46, no. 5, security (ICOSS) 2013 Jul 2 (pp. 1–3). IEEE.
pp. 992–1016, 2018. 30. Dalal V, Malik L. A survey of extractive and abstractive
15. Sobh, I., Darwish, N., Fayek, M. (2006). An optimized text summarisation techniques. In2013 6th international
dual classification system for Arabic extractive generic text conference on emerging trends in engineering and technology
summarisation. Proceedings of the 7th Conf. on Language 2013 Dec 16 (pp. 109–110). IEEE.
English, ESLEC, 149–154. 31. Ganesan K, Zhai C, Han J. Opinosis: A graph based approach
16. Schlesinger, J.D., O’Leary, D.P., Conroy, J.M., 2008. Arabic/ to abstractive summarisation of highly redundant opinions.
English multi-document summarisation with CLASSY—the InProceedings of the 23rd international conference on
past and the future. In: Gelbukh, A. (Ed.), Computational computational linguistics (Coling 2010) 2010 Aug (pp. 340–
Linguistics and Intelligent Text Processing. Springer, Berlin 348).
Heidelberg, pp. 568–581. 32. Moratanch N, Chitrakala S. A survey on extractive text
17. Douzidia, F.S., Lapalme, G., 2004. Lakhas, an Arabic summarisation. In2017 international conference on computer,
summarisation system. Proceedings of the Document communication and signal processing (ICCCSP) 2017 Jan 10
Understanding conference (DUC2004). (pp. 1–6). IEEE.
18. Shankar RS, Babu DR, Murthy KV, Gupta V. An approach 33. Moratanch N, Chitrakala S. A survey on abstractive text
for essay evaluation using system tools. In2017 International summarisation. In2016 International Conference on Circuit,
Conference on Innovative Research In Electrical Sciences power and computing technologies (ICCPCT) 2016 Mar 18
(IICIRES) 2017 Jun 16 (pp. 1–9). IEEE. (pp. 1–7). IEEE.
19. Shankar RS, Srinivas LV, Ravibabu D, Raminaidu C. Novice 34. Siddiqui T, Shamsi JA. Generating abstractive summaries using
Retroaction Report. ARPN Journal of Engineering and sequence to sequence attention model. In2018 International
Applied Sciences. 2018;13 (24): PP 9746–9753. Conference on Frontiers of Information Technology (FIT)
20. Shankar RS, Murthy KV, Rao CS, Gupta VM. An approach 2018 Dec 17 (pp. 212–217). IEEE.
for extracting tweets from social media factors. In2018 ieee 35. Patil AP, Dalmia S, Ansari SA, Aul T, Bhatnagar V. Automatic
international conference on system, computation, automation text summariser. In2014 international conference on advances
and networking (icscan) 2018 Jul 6 (pp. 1–7). IEEE. in computing, communications and informatics (ICACCI)
21. AMaˆ aloul, M. H., Keskes, I., Hadrich Belguith, L., Blache, P. 2014 Sep 24 (pp. 1530–1534). IEEE.
(2010). Automatic summarisation of Arabic texts based on RST 36. Reddy SS, Gadiraju M, Maheswara Rao VV. Analyzing Student
Technique. In Proceedings of 12th International Conference Reviews on Teacher Performance Using Long Short-Term
on Enterprise Information Systems (ICEIS’2010)12th Memory. InInnovative Data Communication Technologies
International Conference on Enterprise Information Systems and Application: Proceedings of ICIDCA 2021 2022 Feb 24
(ICEIS’2010) vol. 2, Portugal. pp. 434–437). (pp. 539–553). Singapore: Springer Nature Singapore.
22. Mathkour, H.I., Touir, A.A., Al-Sanea, W.A., 2008. Parsing 37. Mahesh G, Shankar Reddy S, Maheswara Rao VV, Silpa N.
Arabic texts using rhetorical structure theory. J. Comput. Sci. Preeminent Sign Language System by Employing Mining
4 (9), 713– 720. Techniques. InInternational Conference on IoT Based Control
310 Algorithms in Advanced Artificial Intelligence
Networks and Intelligent Systems 2023 Jun 21 (pp. 571–588). algorithm. 7th International Conference on in Information
Singapore: Springer Nature Singapore. and Communication Systems (ICICS2016), 5–7 April, Irbid,
38. Wang S, Zhao X, Li B, Ge B, Tang D. Integrating extractive Jordan.
and abstractive models for long text summarisation. In2017 41. Al-Zahrani, A., Mathkour, H., Abdalla, H. (2015). PSO-Based
IEEE international congress on big data (BigData congress) Feature Selection for Arabic Text Summarisation, Journal of
2017 Jun 25 (pp. 305–312). IEEE. Universal Computer Science, 21(11): 1454–1469.
39. Jaradat, Y. A. (2015). Arabic Single-Document Text 42. 42. Mahjoub, A. Y. (2015). Text Summarisation Using Particle
Summarization Based on Harmony Search. Master Thesis, Swarm Optimization Algorithm. Master Thesis, College of
Yarmouk Uneversity, Irbid, Jordan. Graduate Studies, Sudan University of Science & Technology,
40. Jaradat, Y. A., & Al-Taani, A. T. (2016). Hybrid-based Arabic Sudan.
single-document text summarisation approach using genatic
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Diabetes, a common metabolic disease with serious health effects, is the focus of this investigation. A unique method
to increase predicted accuracy is presented in the study. We use ensemble learning methods like Grey Wolf Optimization
(GWO) with Adaboost and Ant Colony Optimization (ACO) with XGBoost. After data preparation, GWO and ACO algorithms
pick features, and model training is performed. An analysis of a dataset from the National Institute of Diabetes and Digestive
and Kidney Diseases found that Grey Wolf Optimizer (GWO) with AdaBoost outperforms Ant Colony Optimization (ACO)
with XGBoost in accuracy, precision, and AUC. Ant Colony Optimization (ACO) using XGBoost improves recall, detecting
actual positives more accurately. The models’ slight performance differences emphasize the need to select them depending on
healthcare goals. This study shows how ensemble learning and feature selection improve diagnostic accuracy and healthcare
decision-making, advancing diabetes prediction models.
Keywords: Grey wolf optimization, Ant colony optimization, Adaboost, XGBoost
DOI: 10.1201/9781003529231-47
312 Algorithms in Advanced Artificial Intelligence
several other illnesses [3, 4]. A variety of complications risk. Random Forest outperforms conventional classifiers,
that can impact vital organs as well as systems, such as the increasing T2D medication predictions and showing its
cardiovascular system [5], kidneys [6], eyes [7], and nervous medicinal potential.
system [8], may arise when diabetes is not well managed [9]. In 2020, Xue et al. [16] covered the rising incidence of type 1
Additionally, this condition increases an individual’s chances diabetes in youth and the need for early prediction to minimize
of having heart disease, stroke, and so on. Uncontrolled delayed treatment and chronic consequences. The study
diabetes has severe consequences because it may lead to involved 520 participants who had either been diagnosed with
various complications that can affect critical body organs and diabetes or were at risk of developing it. The age range of
systems like the cardiovascular system, kidneys, eyes, and the participants was 16 to 90 years old. The results indicated
nervous system. Furthermore, this condition is a significant that SVM had the highest classification accuracy compared
risk factor for heart disease development, along with stroke, to the other algorithms and was the best predictor of diabetes.
among other types of health conditions [10]. In conclusion, younger diabetes cases are rising, and early
Diabetes, if left uncontrolled, has serious outcomes in identification is crucial. Machine learning, especially SVM,
that it may lead to a range of complications affecting vital transforms diabetes risk prediction, benefiting medicine.
organs such as the kidney, cardiovascular system, eye, or The article emphasizes the necessity for continuous updates
nervous system. This condition additionally increases one’s with larger case datasets to improve prediction accuracy and
susceptibility to heart disease, among others, like stroke. It is, suggests that sophisticated technology may help doctors
therefore, important for people to understand the factors that make educated illness status decisions.
lead to an increase in diabetes cases all over the continent [11]. In 2021, Ramesh et al. [17] presented a comprehensive
All this shows how serious this problem has become; hence, theoretical framework for diabetic remote patient monitoring
more concentration should be given when handling it [12]. (RPM) employing personal health devices, wearables,
Henceforth, there is a need for people across the continent and cell phones. The proposed end-to-end system uses an
to know why there have been rising cases of diabetes. This SVM to predict diabetes risk. The platform lets patients
underscores its seriousness, so priority should be given to use smartphones and wearables to track key indicators,
addressing it [13]. encouraging proactive diabetes management. The technology
promptly alerts doctors, improving diagnostic decision-
2. Literature Review making. The seamless integration of multiple cloud-based
devices delivers unobtrusiveness, cost savings, and vendor
In 2019, Kavakiotis et al. [14] investigated how machine compatibility, making this method unique. SVM-RBF is
learning and data mining have advanced diabetes research the best alternative because of its modularity and vendor
utilizing high-throughput genetic data and Electronic Health independence. The paper suggests longitudinal investigations
Records. The systematic review examines prediction, and adding gadgets and patient data to the examination.
diagnosis, complications, genetic factors, and healthcare These suggestions support the paper’s claim that continuous,
treatment. SVMs have become popular tools in many automated, and tailored diabetes therapy improves results.
fields. About 85% of SVM research employs supervised
learning, whereas 15% use unsupervised methods such as In 2022, Laila et al.[18] investigates that diabetes is a chronic
association rules. Clinical databases are abundant, allowing disease that may have significant effects if not detected
for valuable insights. The rising global incidence of diabetes early. Thus, this study examines the importance of early
has highlighted the need for sophisticated analytics in detection. The study aims to predict early diabetes incidence.
understanding, diagnosing, and managing this complex Random Forest outperforms the other two techniques in 10
metabolic disorder. fold cross-validation accuracy (97%), precision, recall, and
F1-score. The Chi-Square attribute selection method shows
In 2019, Kowsher et al. [15] aimed to improve Type 2 diabetes that Polyuria predicts diabetes risk statistically. This study
therapy and drug detection by utilizing seven classifier emphasizes the clinical importance of age in diabetes risk
algorithms. Using genetic and clinical characteristics like assessment. This research helps control health by identifying
Fasting, BMI, Duration, Age, and blood pressure, the decision diabetes early. The study suggests algorithmic improvements,
tree-based study will justify patient-appropriate medications. innovative approaches, and the use of additional data to
The technique proposed in this research helps healthcare improve predictive models and overcome diabetes prediction
practitioners make decisions since medicinal intervention challenges.
may reduce problems but may not restore normal blood
glucose levels. A sample of 666 people with type 2 diabetes The 2019–2022 studies show how dynamic diabetes research
is evaluated. The strategy helps prescribe suitable medicines is, emphasizing machine learning and data mining. The
and encourages lifestyle changes to reduce Type 2 diabetes researchers use SVM, decision trees, and ensemble learning
Optimizing Diabetes Prediction through Intelligent Feature Selection 313
to study prediction, diagnosis, treatment, and remote patient selection, Adaboost [25] and XGBoost models are created
monitoring [19]. The research stresses early diagnosis using their specified features. These measurements assess the
and the potential of advanced analytics and technology to models’ accuracy, efficiency, and precision-recall balance.
improve risk prediction, medication selection, and healthcare Optimizing ensemble learning models for diabetes prediction
outcomes for people with diabetes. The research emphasizes through feature selection is the workflow’s goal.
advancements, big datasets, and innovative methods to
enhance diabetes prediction algorithms’ accuracy and 3.1 Data Collection
usefulness [20]. S.S. Reddy et al. worked on various diabetes The National Institute of Diabetes and Digestive and Kidney
side effects and detected whether they had had diabetes or Diseases dataset used to predict diabetes was carefully
not. Among them, they have predicted whether the patients selected to meet specific requirements. All dataset participants
have gestational diabetes or not [21] and indicated whether are Pima Indian women over 21. The dataset focuses on
they have type II diabetes or not [22]. diabetes diagnostic indicators, making it useful for healthcare
research and predictive medicine using machine learning.
2.1 Motivation The collection contains several diagnostic signs that evaluate
The urgent need to address diabetes’s global impact many factors. These include the number of pregnancies,
prompted this investigation. Given the rising prevalence of plasma glucose after a 2-hour oral glucose tolerance test,
the condition and its severe effects, innovative treatments to diastolic blood pressure, triceps skin fold thickness, serum
improve prompt detection, tailored treatment, and thorough insulin levels after 2 hours, BMI, diabetes pedigree function,
administration are needed. Machine learning and data age, and a binary outcome variable indicating diabetes. The
mining may use various information for accurate prediction Kaggle dataset is helpful for researchers, data scientists,
and effective action. This research investigates advanced and healthcare professionals developing and testing
algorithms and methods to improve diabetes risk assessment. diabetes-predicting models. The dataset’s concentration on
The goal is to enhance diabetes risk assessment tools to demographic characteristics and diagnostic measures makes
improve healthcare outcomes and reduce the burden of this it a valuable resource for addressing diabetes diagnosis and
common metabolic illness. prediction in a small population. This dataset lets academics
and practitioners test machine learning and statistics methods.
2.2 Research Gap By doing so, they may better understand diabetes risk factors
Current diabetes risk prediction studies generally neglect and help develop more accurate and tailored healthcare
ensemble learning approaches. Although individual prediction models.
algorithms have been extensively studied, ensemble
techniques, including their comparative effectiveness and 3.2 Feature Selection Using Grey Wolf
subtle properties, have not. Few studies have examined the Optimization
pros and cons of ensemble learning, which combines many Grey Wolves’ cooperative hunting behavior inspired feature
algorithms, for diabetes prediction. Closing this gap is crucial selection in the diabetes dataset using Grey Wolf Optimization
to understanding the pros and cons of ensemble techniques (GWO). This strategy solves the problem creatively. A
for diabetes risk prediction. These insights may improve population of possible feature subsets is initialized for
model accuracy and influence healthcare decision-making. Optimization. Each bit in these binary vectors indicates a
feature’s existence or absence. Adaboost is used to evaluate
3. Proposed Methodology each subset’s fitness in the objective function. The subgroup
is trained and tested for diabetes prediction accuracy. Wolf
The technique suggested begins with raw data preparation to posture is guided by GWO’s three-phase strategy of encircling,
ensure consistency and reliability. This procedure includes attacking, and seeking food. This method permits dynamic
data cleansing, normalization, and missing value correction. modifications in each iteration, allowing wolves to converge
After preprocessing, two feature selection approaches are on an optimal collection of traits. A specified amount of
used. The original technique uses Grey Wolf Optimization iterations stops the process. Next, the selected characteristics
(GWO) [23], inspired by grey wolf hunting habits, to identify train a diabetic dataset-based machine learning model. This
the most critical characteristics for the Adaboost ensemble thorough strategy identifies a particular collection of features
learning algorithm. that considerably improve diabetes prediction.
GWO simulates the grey wolf pack social structure to
improve feature selection[24]. Ant colony optimization 3.3 AdaBoost
(ACO), inspired by ant foraging, is used to choose features Adaboost is an ensemble learning methodology that combines
for the XGBoost ensemble learning approach. After feature multiple weak classifiers to create a robust classifier. It is
314 Algorithms in Advanced Artificial Intelligence
Formula: (True Positives + True Negatives) / (Total score shows the model’s favorable predictions are reliable.
Predictions) A recall rate of 63.63% shows that the model can effectively
identify positive instances, indicating its effectiveness in
4.3 Precision determining diabetes. Using accuracy and recall, the F1 score
Precision measures the proportion of true positives over the is 64.22%, which is excellent.
total number of positive predictions. It represents the model’s
ability to not label negative instances as positive.
Formula: True Positives/(True Positives + False Positives)
4.5 F1-Score
F1-score is the harmonic mean of precision and recall, and is
a good measure when the classes are imbalanced. It balances
the trade-off between precision and recall.
Formula: 2 * (Precision * Recall)/(Precision + Recall)
and recall, is 60.01%, indicating good performance. The The F1 score, which considers accuracy and recall, was
model’s AUC score is 68.08% on the ROC curve, showing higher for the GWO with the AdaBoost model at 64.22% than
its ability to distinguish positive and negative events. A the ACO with the XGBoost model at 60.01%. This suggests
collaborative approach using Ant Colony Optimization that the GWO with the AdaBoost model optimizes accuracy
(ACO) for feature selection and XGBoost for classification and recall. Regarding the ROC curve Area Under the Curve
improves diabetes prediction. (AUC) score, the GWO with AdaBoost model surpasses the
ACO + XGBoost model. GWO + AdaBoost have a higher
AUC value of 78.37%, showing better discrimination
between positive and negative cases. In comparison, ACO
with XGBoost has a 68.08% AUC.
In conclusion, both models perform well. The GWO with
the AdaBoost model has better accuracy, precision, and
AUC score, while the ACO with the XGBoost model has
better recall. These models may be chosen based on diabetes
detection priority or precision-recall balance.
considering precision-recall balance and healthcare decision- 12. Sowah, Robert A., Adelaide A. Bampoe-Addo, Stephen K.
making implications. This study provides a platform for Armoo, Firibu K. Saalia, Francis Gatsi, and BaffourSarkodie-
predictive modeling developments due to diabetes’s global Mensah. “Design and Development of Diabetes Management
prevalence. Thus, diabetes-prone patients will receive better System Using Machine Learning.” International Journal of
Telemedicine and Applications 2020 (July 16, 2020): 1–17.
personalized healthcare.
https://doi.org/10.1155/2020/8870141.
13. Reddy SS, Sethi N, Rajender R. Rigorous assessment of data
References mining algorithms in gestational diabetes mellitus prediction.
International Journal of Knowledge-based and Intelligent
1. Abhari, Shahabeddin, Sharareh R. NiakanKalhori, Mehdi Engineering Systems. 2021 Jan 1;25(4):369-83.
Ebrahimi, HajarHasannejadasl, and Ali Garavand. “Artificial 14. Kavakiotis, Ioannis, Olga Tsave, AthanasiosSalifoglou,
Intelligence Applications in Type 2 Diabetes Mellitus Care: NicosMaglaveras, IoannisVlahavas, and IoannaChouvarda.
Focus on Machine Learning Methods.” Healthcare Informatics “Machine Learning and Data Mining Methods in Diabetes
Research 25, no. 4 (2019): 248. https://doi.org/10.4258/ Research.” Computational and Structural Biotechnology
hir.2019.25.4.248. Journal 15 (2017): 104–16. https://doi.org/10.1016/j.
2. Reddy SS, Rajender R, Sethi N. A data mining scheme for csbj.2016.12.005.
detection and classification of diabetes mellitus using voting 15. Kowsher, Md., FarhanaSharminTithi, TapasyRabeya,
expert strategy. International journal of knowledge-based and Fahmida Afrin, and Mohammad Nurul Huda. “Type 2
intelligent engineering systems. 2019 Jan 1;23(2):103-8. Diabetics Treatment and Medication Detection with Machine
3. Reddy SS, Sethi N, Rajender R. A comprehensive analysis Learning Classifier Algorithm.” Proceedings of International
of machine learning techniques for incessant prediction of Joint Conference on Computational Intelligence, July 4, 2019,
diabetes mellitus. International Journal of Grid and Distributed 519–31. https://doi.org/10.1007/978-981-13-7564-4_44.
Computing. 2020;13(1):1-22. 16. Xue, Jingyu, Fanchao Min, and Fengying Ma. “Research on
4. Reddy SS, Sethi N, Rajender R. Evaluation of deep belief Diabetes Prediction Method Based on Machine Learning.”
network to predict hospital readmission of diabetic patients. Journal of Physics: Conference Series 1684 (November 2020):
In2020 Second International Conference on Inventive 012062. https://doi.org/10.1088/1742-6596/1684/1/012062.
Research in Computing Applications (ICIRCA) 2020 Jul 15 17. Ramesh, Jayroop, Raafat Aburukba, and Assim Sagahyroon.
(pp. 5–9). IEEE. “A Remote Healthcare Monitoring Framework for Diabetes
5. Reddy SS, Sethi N, Rajender R. Risk Assessment of Prediction Using Machine Learning.” Healthcare Technology
myocardial infarction for diabetics through multi-aspects Letters 8, no. 3 (May 2, 2021): 45–57. https://doi.org/10.1049/
computing. EAI Endorsed Transactions on Pervasive Health htl2.12010.
and Technology. 2020 Dec 16;6(24):e3-. 18. Laila, Umm e, Khalid Mahboob, Abdul Wahid Khan, Faheem
6. Reddy SS, Sethi N, Rajender R. Diabetes correlated renal fault Khan, and WhangboTaekeun. “An Ensemble Approach to
prediction through deep learning. EAI Endorsed Transactions Predict Early-Stage Diabetes Risk Using Machine Learning:
on Pervasive Health and Technology. 2020 Nov 11;6(24):e4-. An Empirical Study.” Sensors 22, no. 14 (July 13, 2022):
7. Reddy S, Sethi N, Rajender R. Discovering optimal algorithm 5247. https://doi.org/10.3390/s22145247.
to predict diabetic retinopathy using novel assessment 19. Reddy SS, Mahesh G, Preethi NM. Exploiting machine
methods. EAI Endorsed Transactions on Scalable Information learning algorithms to diagnose foot ulcers in diabetic
Systems. 2020 Jul 1;8(29). patients. EAI Endorsed Transactions on Pervasive Health and
8. Reddy S, Mahesh G, Preethi N. Evolving a neural network Technology. 2021 Aug 24;7(29):e2-.
to predict diabetic neuropathy. EAI Endorsed Transactions on 20. Reddy SS, Mahesh G, Rao VM, Preethi NM. Developing
Scalable Information Systems. 2020 Oct 26;8(31). preeminent model based on empirical approach to prognose
9. Shifrin, Mark, and HavaSiegelmann. “Near-Optimal Insulin liver metastasis. InUbiquitous Intelligent Systems:
Treatment for Diabetes Patients: A Machine Learning Proceedings of ICUIS 2021 2022 (pp. 665–683). Springer
Approach.” Artificial Intelligence in Medicine 107 (July 2020): Singapore.
101917. https://doi.org/10.1016/j.artmed.2020.101917. 21. Reddy SS, Gadiraju M, Preethi NM, Rao VM. A Novel
10. Reddy SS, Sethi N, Rajender R. Safe prediction of diabetes Approach for Prediction of Gestational Diabetes based on
mellitus using weighted conglomeration of mining schemes. Clinical Signs and Risk Factors. EAI Endorsed Transactions
In2020 4th International Conference on Electronics, on Scalable Information Systems. 2023 Jan 11;10(3).
Communication and Aerospace Technology (ICECA) 2020 22. Reddy S, Mahesh G. Risk assessment of type 2 diabetes
Nov 5 (pp. 1213–1220). IEEE. mellitus prediction using an improved combination of NELM
11. Reddy SS, Sethi N, Rajender R, Vetukuri VS. Non-invasive PSO. EAI Endorsed Transactions on Scalable Information
Diagnosis of Diabetes Using Chaotic Features and Genetic Systems. 2021 May 3;8(32).
Learning. InInternational Conference on Image Processing 23. Mallika, C., and S. Selvamuthukumaran. “A Hybrid Crow
and Capsule Networks 2022 May 20 (pp. 161–170). Cham: Search and Grey Wolf Optimization Technique for Enhanced
Springer International Publishing. Medical Data Classification in Diabetes Diagnosis System.”
318 Algorithms in Advanced Artificial Intelligence
International Journal of Computational Intelligence Systems with Gini Index for Predicting Type 2 Diabetes.”
14, no. 1 (September 1, 2021). https://doi.org/10.1007/ INTERNATIONAL UZBEKISTAN-MALAYSIA
s44196-021-00013-0. CONFERENCE ON “COMPUTATIONAL MODELS AND
24. Bilal, Anas, Guangmin Sun, Sarah Mazhar, and Azhar Imran. TECHNOLOGIES (CMT2020)”: CMT2020, 2021. https://
“Improved Grey Wolf Optimization-Based Feature Selection doi.org/10.1063/5.0057315.
and Classification Using CNN for Diabetic Retinopathy 30. Christmas, Jacqueline, Edward Keedwell, Timothy M.
Detection.” Evolutionary Computing and Mobile Sustainable Frayling, and John R.B. Perry. “Ant Colony Optimisation to
Networks, 2022, 1–14. https://doi.org/10.1007/978-981-16 Identify Genetic Variant Association with Type 2 Diabetes.”
9605-3_1. Information Sciences 181, no. 9 (May 2011): 1609–22. https://
25. Dhilsath Fathima, M., and S. Justin Samuel. “Improved doi.org/10.1016/j.ins.2010.12.005.
Adaboost Algorithm with Regression Imputation for 31. Wang, Liyang, Xiaoya Wang, Angxuan Chen, Xian Jin, and
Prediction of Chronic Type 2 Diabetes Mellitus.” HuilianChe. “Prediction of Type 2 Diabetes Risk and Its
Communication and Intelligent Systems, 2021, 691–708. Effect Evaluation Based on the XGBoost Model.” Healthcare
https://doi.org/10.1007/978-981-16-1089-9_54. 8, no. 3 (July 31, 2020): 247. https://doi.org/10.3390/
26. Chen, Peihua, and Chuandi Pan. “Diabetes Classification healthcare8030247.
Model Based on Boosting Algorithms.” BMC Bioinformatics 32. Prabha, Anju, JyotiYadav, Asha Rani, and Vijander Singh.
19, no. 1 (March 27, 2018). https://doi.org/10.1186/s12859 “Design of Intelligent Diabetes Mellitus Detection System
018-2090-9. Using Hybrid Feature Selection Based XGBoost Classifier.”
27. Kalagotla, Satish Kumar, Suryakanth V. Gangashetty, Computers in Biology and Medicine 136 (September 2021):
and Kanuri Giridhar. “A Novel Stacking Technique for 104664. https://doi.org/10.1016/j.compbiomed.2021.104664.
Prediction of Diabetes.” Computers in Biology and Medicine 33. Manjula G, Gopi R, Rani SS, Reddy SS, Chelvi ED. Firefly—
135 (August 2021): 104554. https://doi.org/10.1016/j. binary cuckoo search technique based heart disease prediction
compbiomed.2021.104554. in big data analytics. InApplications of Big Data in Healthcare
28. Ganji, MostafaFathi, and Mohammad SanieeAbadeh. “A Fuzzy 2021 Jan 1 (pp. 241-260). Academic Press.
Classification System Based on Ant Colony Optimization 34. Reddy SS, Sethi N, Rajender R, Mahesh G. Forecasting
for Diabetes Disease Diagnosis.” Expert Systems with Diabetes Correlated Non-Alcoholic Fatty Liver Disease by
Applications 38, no. 12 (November 2011): 14650–59. https:// Exploiting Naïve Bayes Tree. EAI Endorsed Transactions on
doi.org/10.1016/j.eswa.2011.05.018. Scalable Information Systems. 2023;10(1):e2-.
29. Anwar, NurHadirahKhairul, RizauddinSaian, and Sumarni
Note: All the figures and tables in this chapter were designed by
Abu Bakar. “An Enhanced Ant Colony Optimization
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: Sharing thoughts, feelings, and information via a shared language is the foundation of communication. Because they
can’t hear or talk, people who are deaf or mute face additional obstacles and must depend on sign language. The general public
frequently fails to recognise the importance of sign language, which hinders communication between those who use it and those
who do not. We offer a fresh deep learning strategy for RT-SLR as a means of overcoming this obstacle. Although there are
systems that can identify gestures, they’re not always able to do it in real-time. Using a camera or video feed, our method starts
by recording sign language gestures. It uses MediaPipe to identify and follow the user’s hand movements, allowing it to extract
important landmarks. A TensorFlow Convolutional Neural Network (CNN) that has already been trained is used to feed these
landmarks, ensuring accurate detection in a wide variety of sign languages. Combining MediaPipe and TensorFlow creates a
flexible and dynamic platform for real-time sign language identification, enabling people with hearing loss to communicate
more effectively and fully participate in society.
Keywords: Sign language, Convolutional neural network, Media pipe, TensorFlow, Deep learning
1
birudusujatha@gmail.com, 2drnleelavathy@gmail.com, 3karakanavyasri6@gmail.com, 4genjijaganmohan@gmail.com, 5bosubabu10@gmail.com
DOI: 10.1201/9781003529231-48
320 Algorithms in Advanced Artificial Intelligence
important parts such as computation resources, features, sign languages, such as American Sign Language (ASL)
classification algorithms, input modalities (vision-based and International Sign Language (ISL), differ from place
and sensor-based), dataset categorization, and practical to place, localized databases are necessary. Systems for
applications. Recognising facial expressions, body language, recognition are impacted by the use of RGB, or depth
and hand gestures are critical characteristics of SLR. There pictures. A variety of methods are employed by researchers,
are many different kinds of classification methods used in all of which aim to improve accuracy. However, there isn’t
SLR research. These include hybrid methods, deep learning a system that works better for everyone. This work uses an
techniques like Convolutional Neural Networks (CNN), and American Sign Language (ASL) dataset and a deep learning-
more traditional models like Hidden Markov Models (HMM). based methodology to overcome common recognition issues.
Sign languages vary greatly across the globe, influenced Examples of these kinds of barriers are variations in lighting
by factors such as location, ethnicity, and vocabulary. ASL, and distance. The ASLA dataset will prove to be a priceless
BSL, ISL, and CSL are examples of sign languages that tool for researchers, as it will make applying deep learning
developed within deaf populations and are spoken in places and machine learning techniques easier and enable easier
where sign languages are also spoken [6]. Factors such as result comparison. This research offers a method for static
signing speed, picture problems, ambient variations, and hand gestures using convolutional neural network (CNN)
the variety of communication characteristics make the deep learning, which has demonstrated amazing effectiveness
development of SLR systems difficult. A lot of people are in photo classification and pattern recognition applications.
really interested in developing sign language recognition
systems because sign languages use non-manual signs that 2. Related Work
require complicated facial and hand gestures [7].Researchers
have recently used deep learning to improve SLR systems. Sign language recognition (SLR) is one method of automating
These systems have run into a variety of strategies, datasets, the process of converting sign language into text or speech
and challenges. Databases are impacted by variations in [8]. Both signer-independent and signer-dependent SLRs
picture type (RGB versus depth) and geography. Because can distinguish individual signs and continuous phrases,
using the same signer for training and testing. Both sensor
Real-Time Sign Language Translation through Deep Learning 321
based and vision/image-based systems are possible, as and video processing, as well as static picture recognition.
demonstrated in Fig. 48.2 [9]. When it comes to performing The system guarantees optimal performance by incorporating
signs, sensor-based systems demand the use of sensors, while various processes and utilizing multiple programming
vision-based solutions use images taken by cameras, doing languages to apply procedural procedures.In 2020, Sharma
away with the necessity for either sensors or gloves [10]. and Singh specified sign language as a visual language that
Aly et al. (2019) [11] emphasise how inexpensive cameras uses structured hand gestures to express ideas and concepts
have recently become, allowing for extensive application in [16]. By using clustering to group signals into predetermined
research. These domains make use of image processing, ML, categories, transforming video sources into grayscale
and deep learning to increase efficiency, broaden their use, frames, and extracting features using directional histograms,
and lower their overall costs. Nandy et al. [17] were able to attain a 100% identification
rate for ISL gestures. Mekala et al. [18] introduced a neural
network system for real-time SLR and text creation from
video streams. This system focused on hand position and
movement and used 55 hand attributes as CNN-based neural
network points of interest (POIs). It claimed to be able to
identify all letters of the English alphabet (A–Z) 100% of
the time and be 48% immune to noise.With the use of deep
learning models, Rastgoo et al. [19] created a system for real-
time isolated hand stereo vision that uses three-dimensional
hand coordinates to extract information. Their 99% accuracy
rate is quite remarkable. However, the model’s inability to
handle strong inter-class similarities and significant occlusion
between hands in specific sign cases hinders accurate sign
prediction and may cause misclassification. Using computer
vision techniques and the HSV colour scheme, Hurroo et al.
[20] presented a CNN-based system that could accurately
recognise ten ASL gesture alphabets with a 90% success rate.
They also brought attention to the usage of 3D convolutional
neural network (CNN) models, such as I3D40, for sign
Fig. 48.2 SLR modalities
language recognition (SLR), pointing out that, although
Vision-based approaches have gained priority in recent SLR computationally demanding, these models are less stable and
research. The vision-based approach improves usability accurate than other CNN models.For American Sign Language
by decreasing the dependency on sensory devices for sign finger-spelled word categorization, Rathi et al. [21] presented
language interpretation; it does this by using data obtained a ResNet50-based deep neural network that achieved 99.03%
by cameras and algorithms for processing images. Acquiring accuracy. Another study by Daroya et al. [22] used DenseNet
images, preprocessing them, segmenting them, extracting to achieve an accuracy of 90.3% in real-time sign language
features, and finally classifying them are the five main recognition. A number of reports Implementing CNN models
steps in vision-based SLR [12]. Picture acquisition gathers for ASL recognition, Rahman et al. [23], Bastwesy et al.
information from both private and public databases. Noise [24], and Abdulhussein et al. [25] achieved high accuracy
is reduced, and image quality is improved through pre rates ranging from 93% to 99.92%.Based on deep learning
processing. Segmentation isolates the area of interest. techniques often employed for sign language recognition, this
Feature extraction converts this area into recognizable feature study adds to the existing body of knowledge. Through the
vectors. Lastly, in order to recognise signs, categorization use of a convolutional neural network (CNN) model—albeit
compares these qualities with those already stored in the one with a unique architecture—it aims to identify hand-
database.Aiming to improve accuracy, deep learning has sign language alphabets communicating with the deaf. Prior
been applied to SLR systems in recent years. To account for research showing CNN’s effectiveness in picture recognition
regional differences in picture types (RGB or depth) in sign [26–28] justifies its selection. This research contributes
languages, researchers compile a number of datasets [13]. by proposing a CNN model trained on the ASL dataset,
Some cameras use depth images [15] and others use RGB incorporating scaling and backdrop correction approaches
images [14], hence the choice of image format is camera- for improved alphabet sign recognition, and presenting a
specific. There has been a lot of study in depth-camera sensing real-time hand sign language image acquisition model that
captures frames by webcam.
322 Algorithms in Advanced Artificial Intelligence
3. Mediapipe
MediaPipe provides a flexible platform for processing
perceptual data, such as photos, videos, and audio. It is an
open-source framework. It is designed for use in real-time
and works well with machine learning; it is especially good at
things like gesture detection and hand tracking [29]. Among
the many factors that go into the reliable identification of
sign signals, MediaPipe’s accuracy in monitoring fingers and
hands stands out. The focus of this research is hand tracking
in particular.
Fig. 48.4 Proposed model of real-time sign language
Important three-dimensional coordinates on a human hand translation through CNN
are represented by hand landmarks in MediaPipe Holistic.
Ensuring accurate hand tracking is an essential part of Python script successfully extracted the user’s hand from
the larger framework that aims to estimate human poses the original image using the MediaPipe framework, which
holistically. Two models are used in this process: Blaze Palm, is mainly built for component landmark detection. In order
which finds the hand in the input image efficiently, and the to store the hand landmarks as floating-point numbers in
hand keypoint localization model, which uses 21 points in 2D NumPy arrays, this module performed further processing
or 3D space, including knuckles and other features, to refine on them. The programme computed the distance between
the localization. Detailed hand landmarks, a likelihood flag for each landmark point and a default reference landmark by
the presence of the hand, and a left/right binary classification comparing the hand’s landmark coordinates to the palm’s
are all part of the result (Fig. 48.3). Applications that rely on base landmark coordinates. A convolutional neural network
these landmarks, such as hand pose estimation and gesture (CNN) model was built with TensorFlow and Keras using
recognition, are crucial. the preprocessed landmarks for training. Computer vision
projects and image data processing benefit greatly from
CNNs because of their exceptional suitability for spatial
data. For the purpose of predicting sign language motions,
this model makes use of the CNN model and the MediaPipe
Library to acquire hand and palm landmarks. Since the CNN
model relied on landmarks instead of raw visual data, it used
less storage space. We need to build an all-inclusive system
for “Sign Language Conversion to Text and Speech using
Fig. 48.3 Mediapipe’s hand landmarks: A visual overview MediaPipe and TensorFlow” by combining algorithms from
[30] various fields, such as computer vision, deep learning, and
natural language processing. Here is a carefully curated list
of algorithms for critical implementation phases:
4. Proposed Method
(a) Real-time Video Input (Webcam): Real-time video
Data preprocessing, feature extraction using the MediaPipe input from the webcam is achieved through OpenCV.
framework, and gesture recognition are the three processes This powerful library provides indispensable functions
that make up our proposed SLR technique. The initial step for accessing and capturing video frames, constituting
is to use the input frames and the built-in data augmentation a fundamental component of the system’s functionality.
algorithms to extract keypoints and landmarks from the body, (b) Hand Gesture Detection (MediaPipe): MediaPipe,
hands, and face. In order to identify and remove null entries, a sophisticated deep learning model is employed
the system stores the extracted keypoints in a file prior to data for efficient hand detection and accurate landmark
labeling in stage 2. A convolutional neural network (CNN) estimation. This model adeptly identifies and tracks
model, after training and classifying to identify American key hand landmarks, including fingertips, knuckles,
Sign Language (ASL), displays the translated sign gestures and the palm.
as text on the screen. Figure 48.4 provides a summary of
(c) Sign Language Recognition (TensorFlow): Employing
the planned architecture. The initial step of this strategy
deep learning methodologies such as CNN the system
was to employ a webcam to record the users’ real-time
undertakes the recognition of sign language gestures.
movements on the web. A Flask API was used to recognize
This involves the option to design and train a customized
and preprocess hand landmarks in the video stream. The
Real-Time Sign Language Translation through Deep Learning 323
6. Dataset
The majority of the deaf and hard-of-hearing population
in North America uses American Sign Language (ASL), a
visual and gesture language, to communicate. The American
Sign Language uses hand shapes, gestures, and facial
expressions as a substitute for spoken words. The focus of
this research is the recognition of ASL alphabetic characters
using the ASL dataset that was sourced from Kaggle [32].
Figure 48.7 shows an example of an image from the dataset
that intricately represents the 26 letters of the American Sign
Language (ASL) alphabet, mostly utilising one-handed signs
to represent the letters A through Z. This dataset focuses on
single-handed signs, while American Sign Language covers
various variants of signs. Although the complex photos
in the Kaggle dataset are a challenge, the study’s goal is
to demonstrate that their suggested methodology works
effectively in this setting. There are 26 classes in this dataset,
with each class representing an ASL letter, and it contains a
total of 780,000 photos. Character recognition research using
machine learning or computer vision methods will greatly
benefit from this dataset, which has 3,000 samples per class.
7. Experimental Setup
We begin by introducing the datasets used and outlining
the necessary pre-processing steps. After that, we test every
part of our system thoroughly using both quantitative and
qualitative methods. To make data suitable for an SLR system,
it is necessary to transform pixel data into images that are
compatible with the algorithms. We use the MediaPipe Hands
solution to collect 63 3D landmarks from all of the photos in
the dataset, which are the result of 21 points with x, y, and
depth information, in order to recognise the ASL alphabet.
These landmarks provide important spatial information about
hand motions, including x and y coordinates as well as depth
data. In a new coordinate system, the wrist’s coordinates
are selected as the origin (0,0,0), shifting the values of each
coordinate to make these landmarks suitable for classification.
The objective of this classification job is to anticipate one
of the 26 alphabet labels in American Sign Language using
these processed 63 data points as input. When classifying
ASL signals, we do not consider handedness because the
signs for the alphabet are similar regardless of the hand used
and can be performed with either hand. Also, keep in mind
Fig. 48.6 CNN model for gesture recognition that MediaPipe might miss some photographs that contain a
hand. To make sure the dataset is relevant and of high quality
among neurons. Repeating the application of fully connected for our classification task, we exclude these photographs
layers and the dropout algorithm acquires the output class from the test, validation, and training sets. A data generator
values. The softmax function assigns a probability to each introduces noise and transformations to improve the training
gesture class, and the model’s output is the class with the data. After twenty epochs of training with thirty-two batches,
highest probability. This method guarantees the model’s the CNN model achieved a validation accuracy of 98.69%
ability to identify and categorize sign language gestures. by the tenth epoch and a peak accuracy of 98.91% by the
Real-Time Sign Language Translation through Deep Learning 325
Fig. 48.7 Exploring the 26 characters of the American sign language alphabet [32]
(a) (b)
Fig. 48.8 (a) Training and validation loss, (b) Training and validation accuracy
absolute error (MAE), and R-squared as outlined in true positives (TP) or accurate predictions, true negatives
Table 48.1.The Mean Absolute Error (MAE) represents the (TN) or correct non-predictions, false positives (FP) or
average absolute difference between predicted and actual incorrect predictions, and false negatives (FN) or incorrect
values in the dataset. Its formula is given by: non-predictions. This aids in our comprehension of how
1 N accurately the model identified various objects.
MAE = Â |yi - ŷ|
N i=1
(1) TP + TN
Accuracy ( A) = (4)
TP + FP + TN + FN
The Mean Squared Error (MSE) is the average of the squared
TP
difference between predicted and actual values in the dataset. Precision ( P ) = (5)
Its formula is: TP + FP
1 N TP (6)
Recall ( R) =
MSE = Â (yi - ŷ)2
N i=1
(2) TP + FN
TP
The R2 score indicates the goodness of fit of the model to the F1 Score ( F1) = (7)
TP + FN
dataset. Score indicates the goodness of fit of the model to the
dataset. It ranges between 0.0 and 1.0, where 0.0 denotes the Accuracy refers to the number of accurately predicted
worst fit and 1.0 denotes a perfect fit. The formula is: data points, and it should ideally be near 1. When false
positives are expensive, precision—also known as the
1 N
RMSE = MSE = Â (yi - ŷ)2
N i=1
(3) positive predicted value—becomes essential. It computes
the percentage of positive predictions among all anticipated
positive class values. Recall measures the percentage of
Here, yi, represents actual values, ŷi represents predicted
positive outcomes accurately predicted. The harmonic mean,
values, ȳ is the mean of actual values, and n is the number of
or F1-score, strikes a balance between recall and precision.
data points. Table 48.1 presents MSE,MAE and R-squared
It maximizes at 1, which occurs when recall and precision
values for different models, including simple RNN, LSTM,
are flawless. The results were calculated using Equations
Standard GRU, BIGRU, and BiLSTM. These values indicate
(4–7), and the classification report is shown in Fig. 48.9. The
higher average residual and variance in the residuals for these
suggested model’s accuracy, recall, and F1-score are all near
models. Despite BiLSTM's success in other applications,
1, with only a few values marginally below, suggesting that
it performed poorly in our dataset due to limited data for
the model successfully mastered the training set.
sequence prediction. As a result, we introduced the Mediapipe
CNN model in an attempt to reduce prediction errors.
8. Conclusions
Table 48.1 Comparing MAE, MSE, and R2 for Various Models
Finding a method to recognize American Sign Language
Network Model MAE MSE R2 (ASL) using deep learning was the aim of this study in order
Simple RNN 4.1 28.9 –1.38 to facilitate real-time communication between hearing and
LSTM 0.75 4.95 0.59 deaf individuals. We utilize a webcam in combination with the
Standard GRU 0.44 1.38 0.83
MediaPipe framework to identify hand landmarks in order to
implement the model. After preprocessing the recorded hand
BIGRU 0.4 2.5 0.79
landmarks, we trained a convolutional neural network (CNN)
BILSTM 0.85 5.35 0.56 model to predict sign language motions with a 98% success
Proposed CNN 0.23 1.28 0.7 rate. Even if the model accomplishes its short-term goals—
like capturing webcam feeds in real-time and making correct
7.2 Qualitative Analysis predictions—there remains room for improvement in terms
of testing and evaluation to determine the system’s scalability
The study employed classification measures to assess the and sustainability. The development of this prototype has the
precision of predictions for specific sign motions. The study potential to greatly benefit sign language users by facilitating
computed these metrics using ASL data to evaluate the better communication and laying the groundwork for further
quality of the predictions, including accuracy (A), precision advancements and wider applications in computer vision and
(P), recall (R), and F1-score. We also used a confusion matrix gesture detection.
to assess the model’s performance, which provides data on
Real-Time Sign Language Translation through Deep Learning 327
1.2
0.8
0.6
0.4
0.2
0
A B C D E F G H I J K L M N O P
Precision (P) 1 1 1 1 1 1 1 1 1 0.80.9 1 1 1 1 1
Recall (R) 1 1 1 1 1 1 1 1 1 1 1 0.8 1 1 1 1
F1-Score (F1) 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1
19, 7056–7063. https:// doi. org/ 10. 1109/ JSEN. 2019. 29098
References 37 (2019).
1. Krishnaveni, M., Subashini, P., & Dhivyaprabha, T. T. (2019). 9. Mahmood M. R. and Abdulazeez A. M., “A Comparative
An assertive framework for automatic tamil sign language Study of a New Hand Recognition Model Based on Line of
recognition system using computational intelligence. Features and Other Techniques,” in International Conference
Intelligent systems reference library: 150. Springer of Reliable Information and Communication Technology,
International Publishing. 2017, pp. 420–432.
2. Savur, C., & Sahin, F. (2016). Real-time American Sign 10. Rautaray S.S., A. Agrawal, Vision based hand gesture
Language recognition system using surface EMG signal. In recognition for human computer interaction: a survey Artif.
Proceedings of the IEEE 14th international conference on Intell. Rev., 43 (1) (2015), pp. 1–54.
machine learning and applications, ICMLA 2015 (pp. 497– 11. Aly, W., Aly, S., & Almotairi, S. (2019). User-Independent
502). American Sign Language Alphabet Recognition Based
3. El-Din, S. A. E., & El-Ghany, M. A. A. (2020). Sign language on Depth 6 Image and PCANet Features. IEEE Access, 7,
interpreter system: An alternative system for machine 123138–123150.
learning. In Proceedings of the 2nd novel intelligent and 12. Bantupalli, K., & Xie, Y. (2019), American Sign Language
leading emerging sciences conference, NILES 2020, Ml Recognition using Deep Learning and Computer Vision,
(pp. 332–337). 2018 IEEE International Conference on Big Data (Big Data),
4. Cheok, M.J., Omar, Z. & Jaward, M.H. A review of hand p. 4896–4899.
gesture and sign language recognition techniques. Int. J. 13. Jain V., Jain A., Chauhan A., Kotla S.S., Gautam A., American
Mach. Learn. & Cyber. 10, 131–153 (2019). Sign Language recognition using support vector machine and
5. Wadhawan, A., Kumar, P. Deep learning-based sign language convolutional neural network, Int. J. Inf. Technol. 13 (2021)
recognition system for static signs. Neural Comput & Applic 1193–1200.
32, 7957–7968 (2020). 14. Daroya R., Peralta D., Naval P., Alphabet sign language
6. Agrawal, S. C., Jalal, A. S., & Tripathi, R. K. (2016). A survey image classification using deep learning, in IEEE Region 10
on manual and non- manual sign language recognition for Annual Int. Conference, Proceedings/TENCON (2019) vol.
isolated and continuous sign. International Journal of Applied 2018-Octob, no. October, pp. 646–650.
Pattern Recognition, 3 (2), 99. 15. Ameen S., Vadera S., A convolutional neural network to
7. Ahmed M., Idrees M., Abideen Z., Mumtaz R., Khalique classify American sign language fingerspelling from depth
S., Deaf talk using 3D animated sign language in 2016 SAI and colour images, Expert Syst. 34 (3) (2017).
Comput. Conference (SAI) (2016), pp. 330–335. 16. Ashish Sharmaa, Anmol Mittala, Savitoj
8. Mittal, A., Kumar, P., Roy, P. P., Balasubramanian, R. & Singha,VasudevAwatramani, Hand Gesture Recognition
Chaudhuri, B. B. A modified LSTM model for continuous using Image Processing and Feature Extraction Techniques,
sign language recognition using leap motion. IEEE Sens. J. Procedia Computer Science 173 (2020) 181–190.
328 Algorithms in Advanced Artificial Intelligence
17. Nandy, A.; Prasad, J.; Mondal, S.; Chakraborty, P.; Nandi, G. 25. A. Abdulhussein and F. Raheem, “Hand Gesture Recognition
Recognition of Isolated Indian Sign Language Gesture in Real of Static Letters American Sign Language (ASL) Using Deep
Time. Commun. Comput. Inf. Sci. 2010, 70, 102–107. Learning,” Eng. Technol. J., vol. 38, no. 6, pp. 926–937, 2020.
18. Mekala, P.; Gao, Y.; Fan, J.; Davari, A. Real-time sign 26. M. Al-Hammadi et al., “Deep learning-based approach for
language recognition based on neural network architecture. In sign language gesture recognition with efficient hand gesture
Proceedings of the IEEE 43rd Southeastern Symposium on representation,” IEEE Access, vol. 8, pp. 192527–192542,
System Theory, Auburn, AL, USA, 14–16 March 2011. 2020.
19. Rastgoo, R., Kiani, K. & Escalera, S. Video-based isolated 27. Y. Dong, Q. Liu, B. Du, and L. Zhang, “Weighted Feature
hand sign language recognition using a deep cascaded model. Fusion of Convolutional Neural Network and Graph Attention
Multimedia Tools Appl. 79(31–32), 22965–22987. https:// Network for Hyperspectral Image Classification,” IEEE Trans.
doi. org/ 10. 1007/ s11042- 020- 09048-5 (2020). Image Process., vol. 31, pp. 1559–1572, 2022.
20. Hurroo, M. & Elham, M. Sign language recognition system 28. Y. L. Chang et al., “Consolidated Convolutional Neural
using convolutional neural network and computer vision. Int. Network for Hyperspectral Image Classification,” Remote
J. Eng. Res. Technol. (IJERT) 9(12), 59–64 (2020). Sens., vol. 14, no. 7, pp. 1–16, 2022.
21. P. Rathi, R. K. Gupta, S. Agarwal, A. Shukla, and R. Tiwari, 29. Subramanian, B., Olimov, B., Naik, S.M. et al. An integrated
“Sign Language Recognition Using ResNet50 Deep Neural mediapipe-optimized GRU model for Indian sign language
Network Architecture Pulkit,” Next Gener. Comput. Technol. recognition. Sci Rep 12, 11964 (2022).
2019 Sign, pp. 1–7, 2019. 30. F. Zhang, V. Bazarevsky, A. Vakunov, A. Tkachenka, G. Sung,
22. R. Daroya, D. Peralta, and P. Naval, “Alphabet Sign Language C. Chang, and M. Grundmann, “Mediapipe hands: On-device
Image Classification Using Deep Learning,” IEEE Reg. 10 real-time hand tracking,” CoRR, vol. abs/2006.10214, 2020.
Annu. Int. Conf. Proceedings/TENCON, pp. 646–650, 2019. 31. Van Hiep Phung and Eun Joo Rhee, “A High-Accuracy Model
23. M. M. Rahman, M. S. Islam, M. H. Rahman, R. Sassi, M. Average Ensemble of Convolutional Neural Networks for
W. Rivolta, and M. Aktaruzzaman, “A new benchmark on Classification of Cloud Image Patches on Small Datasets”, 23
american sign language recognition using convolutional October 2019.
neural network,” in International Conference on Sustainable 32. Kaggle. ASL Alphabet. Available online: https://www.kaggle.
Technologies for Industry 4.0, STI 2019, pp. 1–6, 2019. com/grassknoted/asl-alphabet (accessed on 19 July 2021).
24. M. R. M. Bastwesy, N. M. ElShennawy, “Deep Learning Sign
Note: All the figures and table in this chapter were designed by the
Language Recognition System Based on Wi-Fi CSI,” Int. J.
author.
Intell. Syst. Appl., vol. 12, no. 6, pp. 33–45, 2020.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
R. Tamilkodi1
Professor, Godavari Institute of Engineering & Technology,
Rajahmundry, Andhra Pradesh, India
K. Surya Kala2
Assistant Professor, Department of CSE(AIML&CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
T. Durga Sukanthika3, B. Aanantha Sai Datta Kiran4,
V. Hemanth Reddy5, K. Srimani Neha6
Department of Computer Science & Engineering (AIML & CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Abstract: The issue at hand is that as cloud computing grows, there is growing worry about cloud security. Sensitive data
stored in the cloud is susceptible to illegal access or alteration by hackers or internal cloud administrators when centralized
access control methods are in place. The security and integrity of corporate and personal data are seriously threatened by this.
One proposed answer for this issue is AuthPrivacyChain, a blockchain based access control architecture AuthPrivacyChain
uses blockchain node addresses as identities to renegotiate permissions for cloud data access control. These permissions are
kept on the blockchain and are encrypted. Processes for authorization, revocation, and access control are also included in the
framework. When AuthPrivacyChain is used on an enterprise operating system such as EOS, it significantly improves cloud
security by guarding against unwanted access and guaranteeing the privacy of authorized users.
Keywords: Authorization revocation access control framework, Enterprise operating system EOS (Enterprise operating
system), Unwanted access, Authorized users cloud security improvement
DOI: 10.1201/9781003529231-49
330 Algorithms in Advanced Artificial Intelligence
confided in focus’ security, tampers with the allowed data set shown by simulations and assessments, indicating that it is a
on the central servers and acquires unapproved admittance to useful addition to IC-SNs.
steals the assets that clients have put away there. Second, a Jianhua Li, Xi Lin, and others, [6] They present that is intended
malevolent system administrator could abuse this capacity to to make knowledge trading easier in Internet of Things (IoT)
get unapproved admittance to assets or to change the consent contexts that are enabled by Edge artificial intelligence.
data set to get unlawful access, since the cloud system It provides an architecture for the knowledge market’s
director is responsible for the approval information base and implementation, complete with a blockchain for knowledge
approaches assets. consortiums for safe and effective knowledge exchange and
The computation and storage modes of cloud computing administration. This blockchain includes smart contracts, a
have changed significantly from the previous computer new cryptographic cryptocurrency called “knowledge coin,”
paradigm. These changes are mostly evident in the following and a special consensus process called “proof of trading.” To
five aspects: There are several reasons why users may find it get more individuals engaged with the market, the framework
difficult to control cloud resources: likewise offers a knowledge cost technique in light of
1. lack of trust between users and the cloud; noncooperative games with remunerations. Security tests
2. data may change the security domain due to migration and performance models show that the framework functions
technologies; admirably; this is the main illustration of a P2P knowledge
market that functions admirably and is driven by motivators
3. access subjects may be redefined due to multitenant
in Edge-AI fuelled IoT.
technologies; and
4. virtualization technologies may make it possible for Chen Jianing, Wu Jun, et al. [24] The suggested system is
difficulties, scholarly research on cloud access control an Unbiased Collaborative Trust-Based Control Transfer
has proliferated, and industry attempts to use access Mechanism (CTM) that aims to improve industrial automation
control systems already in place have also been made. security and trust. By offering a trust-based delegated
proof of stake agreement, it beats the shortfall of trust in
They do, however, have centralized methods for managing
modern control frameworks. Control authority are assigned
and storing identity, key, authority, authentication data, etc.
in a fair and dynamic manner by this consensus approach.
Therefore, there are still two issues with access control
Furthermore, a CTM is put into place for catastrophe backup,
technologies related to security and privacy:
allowing blockchain nodes to switch over control authority.
1. An external attacker compromises the central server, The viability and efficacy of CTM in enhancing industrial
assaults the trusted centre, and gains unauthorized automation security are validated by simulations.
access to or theft of user resources kept in the cloud.
Suyong Eum, Keping Yu, et al., [26] The goal of the suggested
2. A malevolent system administrator could abuse this system is to provide a summary of the current state of research
capacity to get unapproved admittance to assets or to and standards related to information-centric networking, or
change the approval data set to get unlawful access, ICN. The history of international ICN operations starting
as the cloud system chairman is responsible for the in 2010 is traced, with references to several initiatives.
approval information base and approaches assets. The study then explores the latest developments in ICN
component technology standardization, namely in ITU-T
2. Literature Review and in ICNRG’s documentation. Lastly, it considers potential
future paths for ICN’s development as a cutting-edge network
Mianxiong Dong, Jun Wu, et al., [4] Information-centric
architecture.
social networks (IC-SN) have changing demands, and the
suggested solution, called FCSS (Fog-Computing-based Yuwei Su, Xin Qi, and others [27] In order to overcome
Content-Aware Filtering Method for Security Services), is the current shortcomings in information-centric networking
made to meet such needs. By bringing fog computing to IC (ICN), the suggested solution presents the idea of named-
SN, it moves resources and computational intelligence to the node networking (3N). ICN’s emphasis on content-centric
edge of the network. End-to-end connectivity and low-latency communication has drawn attention, however it is devoid of
security service filtering are guaranteed by this method. By host-centric features and seamless mobility support. The 3N
using content-label technology, FCSS integrates an effective system provides a complete solution that includes mobility
content-aware filtering system that allows precise security assistance, data security, data transmission, and naming. A
service filtering at the network edge. The benefits of FCSS in 3N-based real-time video streaming system is developed
terms of hit ratio, filtering latency, and filtering accuracy are and tested; it performs better than TCP video streaming,
Ensuring Data Privacy in the Cloud: Authprivacychain’s Blockchain Access Control 331
4. Experimental Results
Double tapping the “Start_IPFS.bat” document will send off Click the “Data User Signup Here” link in the top page to add
the cloud server and show the screen below. other users, such as researchers, doctors, and data owners.
The data owner is logged in on the top screen; upon login, the The file is uploaded on the screen above, and the storage hash
screen below appears. code is shown. The data owner may click the “Revoke user”
link to revoke access to the file by choosing it.
Choose any file on the screen above, then click the button
to remove access from it. Then, log out and log back in as
The data owner may upload a file to the cloud in encrypted
“doctor” to check if access is still allowed.
format by clicking the “Upload Data” option in the above
page. The doctor is logged in, and after that, they will see the screen
below.
The doctor may examine all files shared by the data owner by
The data owner uploads the file on the above page, chooses clicking the “Access Share Data” option in the aforementioned
an access user—a doctor, researcher, or both—and presses page.
the upload button. on the same screen, I provide permission The doctor may see all files shared by the data owner on the
to the user named “doctor.” Pressing the button yields the interface, and by clicking the “Click Here” link, they can
result below. download the file.
334 Algorithms in Advanced Artificial Intelligence
The file is now downloading, as shown in the browser status Now the researcher has access to the file. To see the graph
bar above. By selecting the “Indirect Access Control” option, below, click the “Smart Contract Computation Graph” link.
a doctor may now provide a researcher access to this file.
grid,’’ IEEE Trans. Instrum. Meas., vol. 64, no. 8, pp. 2072– 11. C. Lee, P. Chung, and M. Hwang, ‘‘A survey on attribute-based
2085, Aug. 2015, doi: 10.1109/TIM.2015.2444238. encryption schemes of access control in cloud environments,’’
6. X. Lin, J. Li, J. Wu, H. Liang, and W. Yang, ‘‘Making Int. J. Netw. Secur., vol. 15, no. 4, pp. 231–240, 2013.
knowledge tradable in edge-AI enabled IoT: A consortium 12. R. Charanya and M. Aramudhan, ‘‘Survey on access control
blockchain-based efficient and incentive approach,’’ IEEE issues in cloud computing,’’ in Proc. Int. Conf. Emerg. Trends
Trans. Ind. Informat., vol. 15, no. 12, pp. 6367–6378, Dec. Eng., Technol. Sci. (ICETETS), Feb. 2016, pp. 1–4, doi:
2019. 10.1109/ICETETS.2016.7603014.
7. Y. Q. Zhang, X. F. Wang, X. F. Liu, and L. Liu, ‘‘Survey on 13. J. M. Ferris, ‘‘Providing access control to user-controlled
cloud computing security,’’ J. Softw., vol. 27, no. 6, pp. 1328– resources in a cloud computing environment,’’ U.S. Patent 8
1348, 2016. 984 505, Mar. 17, 2015.
8. Z. Tari, X. Yi, U. S. Premarathne, P. Bertok, and I. Khalil, 14. S. Namasudra and P. Roy, ‘‘Secure and efficient data
‘‘Security and privacy in cloud computing: Vision, trends, and access control in cloud computing environment: A survey,’’
challenges,’’ IEEE Cloud Comput., vol. 2, no. 2, pp. 30–38, Multiagent Grid Syst., vol. 12, no. 2, pp. 69–90, May 2016,
Mar. 2015, doi: 10.1109/MCC.2015.45. doi: 10.3233/MGS-160244.
9. M. Almorsy, J. Grundy, and I. Müller, ‘‘An analysis of the 15. Y. Wang, J. Yang, C. Xu, X. Ling, and Y. Yang, ‘‘Survey on
cloud computing security problem,’’ 2016, arXiv:1609.01107. Access Control Technologies for Cloud Computing,’’ J.
[Online]. Available: http://arxiv.org/abs/1609.01107 Softw., vol. 26, no. 5, pp. 1129–1150, 2015.
10. Cloud Security Alliance, Security Guidance V4.0. Accessed: 16. J. Zhou, Y. Zhang, and Y. Gao, ‘‘Research of ABAC model
Apr. 16, 2020. [Online]. Available: https://c-csa.cn/i/ based on usage control under cloud environment,’’ J. Comput.
file/20171225/ 2017122523220533533.pdf Appl., vol. 31, no. 12, pp. 3692–3694, 2014.
336 Algorithms in Advanced Artificial Intelligence
17. J. Zhu and Q. Wen, ‘‘SaaS access control research based on 24. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
UCON,’’ in Proc. 4th Int. Conf. Digit. Home, Nov. 2012, pp. in Health Care Using Hybrid Model Algorithms”,Journal of
331–334, doi: 10.1109/ ICDH.2012.50. Artificial Intelligence, Machine Learning and Neural Network
18. Y. Zhu, D. Ma, C.-J. Hu, and D. Huang, ‘‘How to use attribute- (JAIMLNN), Volume 3, 2023.
based encryption to implement role-based access control in 25. J. Chen, J. Wu, H. Liang, S. Mumtaz, J. Li, K. Konstantin,
the cloud,’’ in Proc. Int. Workshop Secur. Cloud Comput. A. K. Bashir, and R. Nawaz, ‘‘Collaborative trust blockchain
Cloud Comput. New York, NY, USA: ACM, 2013, pp. 33–40. based unbiased control transfer mechanism for industrial
19. Y. Wang, D. Zhang, and H. Zhong, ‘‘Multi-authority based automation,’’ IEEE Trans. Ind. Appl., early access, Dec. 13,
weighted attribute encryption scheme in cloud computing,’’ in 2019, doi: 10.1109/TIA.2019.2959550.
Proc. 10th Int. Conf. Natural Comput. (ICNC), Aug. 2014, pp. 26. X. Shen, Q. Pei, and X. Liu, ‘‘Survey of blockchain,’’ J. Netw.
1033–1038, doi: 10.1109/ ICNC.2014.6975982. Inf. Secur., vol. 2, no. 11, pp. 11–20, 2016.
20. L. Popa, M. Yu, S. Y. Ko, S. Ratnasamy, and I. Stoica, 27. K. Yu, S. Eum, T. Kurita, Q. Hua, T. Sato, H. Nakazato, T.
‘‘CloudPolice: Taking access control out of the network,’’ Asami, and V. P. Kafle, ‘‘Information-centric networking:
in Proc. 9th ACM SIGCOMM Workshop Hot Topics Netw. Research and standardization status,’’ IEEE Access, vol. 7, pp.
(Hotnets). New York, NY, USA: ACM, 2010, pp. 1–6. 126164–126176, 2019, doi: 10. 1109/ACCESS.2019.2938586.
21. P. He, R. Huang, N. Chen, and Z. Li, ‘‘Research progress 28. X. Qi, Y. Su, K. Yu, J. Li, Q. Hua, Z. Wen, J. Lopez, and T.
on sidechannel attacks in cloud environment,’’ Appl. Res. Sato, ‘‘Design and performance evaluation of content-oriented
Comput., vol. 35, no. 4, pp. 969–973, 2018. communication system for IoT network: A case study of
22. J. Guo, W. Yang, K. Lam, and X. Yi, ‘‘Using blockchain named node networking for realtime video streaming system,’’
to control access to cloud data,’’ in Proc. Int. Conf. Inf. IEEE Access, vol. 7, pp. 88138–88149, 2019.
Secur. Cryptol. Cham, Switzerland: Springer, 2018, Art. no.
Note: All the figures in this chapter were designed by the author.
274C288, doi: 10.1007/978-3-030-14234-6_15
23. Y. Yuan and F.-Y. Wang, ‘‘Blockchain: The state of the art and
future trends,’’ Acta Autom. Sinica, vol. 42, no. 4, pp. 481–
494, 2016.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: In today’s cloud computing era, improving the overall efficiency and responsiveness of cloud-based services relies
heavily on maximising the performance of cloud load balancers. In order to distribute tasks or workloads evenly across the
nodes or servers, load balancing is a crucial component and a major obstacle. [13]. An essential component of contemporary
communication and computer systems is network latency, frequently called the lag or delay in data transmission across a
network. Network latency, the delay in data transmission over a network, greatly affects the responsiveness and effectiveness
of cloud-based services. In the never-ending race to reduce network latency, cloud load balancers are indispensable. They
mediate communication between clients and a group of servers, dividing up incoming data packets according to the resources
that are available. Improving load balancer algorithms and setups is the primary emphasis of this research, which intends to
alleviate the pressing problem of network latency. Enhancing load balancing mechanisms significantly reduces transmission
delays, making cloud services more responsive and efficient. Load balancers play a crucial role in optimising data routing to
minimise delays, creating a more seamless user experience. Consequently, there is no denying the connection between load
balancing and network latency. The advent of cloud computing has completely altered how businesses rely on their IT systems.
Businesses now use cloud computing services such as Google Cloud Platform, Amazon Elastic Compute Cloud, and Microsoft
Azure to host their apps and data rather than building their own data centers. Although clients have the freedom to determine
the specifications of their computation times as they see fit, it is not yet feasible to establish guarantees in terms of community
latency for an application. [7]
Keywords: Cloud computing, Network latency, Network traffic, Load balancer
1
muralimohan.klu@gmail.com, 2radhayaraguti@gmail.com, 3silpasharonchinta@gmail.com, 4Bhargavijonnavithula1@gmail.com
DOI: 10.1201/9781003529231-50
338 Algorithms in Advanced Artificial Intelligence
(1) Comparing 21% of public IPv4 addresses, the latency of network delay in load balancing as well as management
disparity between Amazon and Alibaba’s clouds is more than techniques in this expert review. There are several ways that
20 ms. (2) By routing traffic across its well-balanced private latency may enter the load balancing process: through client-
WANs, Google is able to achieve a lower latency imbalance load balancer, load balancer-to-backend server, and backend
compared to rival clouds destinations. Thirdly, DCs in the server-to-client connections.[2].
cloud often experience latency imbalance; in fact, researchers
found that eight pairs of DCs had load-balanced pathways
with latency discrepancies greater than 40 ms [1].
2. Background
Fig. 50.1 Architecture of network latency
The concept of load balancing emerged with the proliferation
of computer networks in the late 20th century. Initially, load
balancing aimed to distribute network traffic evenly among 4. Network Latency in Load Balancing
servers to prevent overload and ensure smooth operation. In
the 1990s, as network traffic grew, load balancers faced the
Related Issues
challenge of handling increasing data loads efficiently. Wait time for network loading Concerns about maintaining a
Network latency issues arose due to the varying response steady equilibrium in the world of digital results are common
times of different servers, making it difficult to evenly owing to a number of factors, one of which is communication.
distribute traffic. In the Late 1990s Round-robin load Delays can occur due to a number of factors, including the
balancing algorithms were introduced as a simple method to efficiency of routing algorithms, congestion in the network,
distribute requests evenly among servers. While effective for and the actual distance between devices. As a never-ending
basic load balancing, they did not account for variations in battle, reducing network latency requires optimising network
server response times, leading to latency issues. Early 2000s design, improving routing algorithms, and establishing
Dynamic load balancing algorithms like Least Connections content delivery networks to store and transport material
and Weighted Round Robin were developed to consider from servers closer to end users. When network latency is
server load and response times. These algorithms aimed to high, it can cause communication problems and slow data
reduce network latency by directing traffic to servers with transfers. When there’s a lag between pressing “play” and the
lower loads or faster response times. video starting when watching a movie online due to a slow
network connection, this is known as network latency. Data
In Mid-2000s. The advent of CDNs brought a significant shift
packet transit time, or network latency, is the time it takes for
in load balancing. CDNs used geographically distributed
data to travel from source to destination. [6].
edge servers to reduce latency by serving content from
servers closer to end-users. Late 2000s-Present with the rise
of cloud computing, load balancing faced new challenges.
Virtualized environments and on-demand resource allocation
added complexity to load-balancing decisions. Network
latency between cloud data centers and end-users became a
critical concern.
5. Literature Review
Network latency, defined as the time it takes for data to
be transmitted across a network, is a significant element
influencing the efficiency and responsiveness of cloud-based
services. This paper examines the changing environment of
network latency in the context of load balancing, emphasizing
major discoveries and contributions from recent research.
the number of requests, data transfer rates, and traffic Without aiming squarely at network latency, RR’s primary use
fluctuations over time. case is spreading network requests across numerous servers.
Distributing network traffic fairly across several servers or
6.2 Analysis of Current Load Balancers endpoints is the goal of this algorithm. By distributing the
The analysis of current load balancers is a critical step in workload and preventing any one server from becoming a
understanding their performance and efficiency in managing bottleneck, it helps to decrease latency. Round-robin is the
network latency in cloud-based services. Here is more load-balancing algorithm that is most widely used. Round
information on how to conduct this analysis effectively: robin algorithm cyclically routes client requests to accessible
servers. For round-robin server load balancing to work,
1. Inventory of Load Balancers: Start by creating an
each server’s processing and storage capacities should be
inventory of all load balancers currently in use within
about equal. Connection requests arrive at web servers in
your cloud-based services environment. Document
a specific sequence, determining their delivery in a round-
their names, types, and locations (e.g., data centers or
robin method. For argument’s sake, let’s pretend a company’s
cloud regions).
cluster consists of three servers, A, B, and C. Requests are
2. Configuration Analysis: Load Balancing Algorithms: sent to three separate servers: A, B, and C.
Identify which load balancing algorithms are being used,
such as Round Robin, Least Connections, Weighted
Round Robin, or Weighted Least Connections. Routing
Rules: Analyse how incoming requests are routed
to backend servers based on factors like URL paths,
domain names, or source IP addresses.
3. Traffic Distribution: Collect data on how traffic is
distributed among backend servers. Understand the
current traffic distribution patterns and whether they
are optimized for efficiency.
4. Health Checks: Evaluate the health check mechanisms
in place to determine server availability. Ensure
that servers are not serving requests when they are
underperforming or experiencing issues.
5. Load Balancer Logs: Analyse load balancer logs to
gain insights into real-time traffic patterns. Log data
can reveal variations in traffic volume, request rates,
and response times.
6. Performance Metrics: Collect performance metrics
related to load balancers, such as response times,
request processing times, and error rates. Compare
these metrics across different load balancers.
7. Latency Analysis: Specifically focus on latency
metrics. Measure the round-trip time (RTT) for requests
sent through each load balancer. Identify instances of
latency spikes or consistently high latency.
8. Scalability: Assess whether the current load balancers
can handle increases in traffic and workload. Scalability
is crucial for maintaining low latency during traffic
spikes. Assess whether the current load balancers can
handle increases in traffic and workload. Scalability
is crucial for maintaining low latency during traffic Fig. 50.4 Round Robin algorithm
spikes. To address network latency more effectively, you might
consider combining Round Robin with additional techniques
7. Round Robin
or algorithms that take server health and latency into account.
One of the oldest and most popular methods of load balancing The fundamental steps involved in load balancing while
is the RR algorithm. Using this method is a breeze [15]. considering network latency and server performance.
Optimizing Cloud Load Balancers for Reduced Network Latency 341
9. Conclusion
To sum up, a crucial undertaking in the realm of cloud
computing is improving cloud load balancers for reduced
network latency. The necessity for fast, low-latency data
transfer is growing as more and more businesses depend on
cloud services and apps. Network latency is a problem in load
balancing, and this paper presents research and solutions to
that problem. Through the optimisation of load balancers, we
have tackled the key challenge of reducing network latency
in cloud-based services in this study. Our findings show
that cloud-based services may be made more efficient and
responsive by reducing network latency, finding its causes,
creating tailored solutions, and optimising load-balancing
setups. We have investigated current load balancers, had
a look at their settings, and found that things like heavy
server loads and ineffective load-balancing algorithms
could be causing delays. We minimised network latency by
optimising load balancers through thorough load testing and
implementing configuration modifications. Our research
shows that optimising load balancers is a great way to lower
network latency, which means that users will have a better,
faster experience while using cloud services. Going forward,
it will be crucial to do additional research and development in
this field in order to fulfil the growing demands of the digital
era and make cloud-based services even more efficient.
Fig. 50.5 Flow chart for fundamental steps involved in load
balancing
10. Acknowledgement
8. Strategies to Resolve Our profound appreciation goes out to Dr. Murali Mohan, our
respected research adviser, for all of the help, encouragement,
To address network latency in load balancing, organizations and insightful comments they gave us during the course of
can employ several strategies: our study. Dr. Mohan’s knowledge and guidance greatly
1. Geographic Load Balancing: Use geolocation-based influenced the direction of our work. Our deepest gratitude is
load balancing to direct users to the nearest data center due to all of our hardworking coworkers and fellow scholars
or server, reducing the impact of long-distance network at KL University. Their insightful comments, lively debates,
latency. and intellectually challenging academic
2. Content Delivery Networks (CDNs): Employ CDNs to The quality of our study has been significantly improved.
cache and deliver content closer to end-users, reducing Additionally, we would like to express our gratitude to
the need for long-distance data transfers. everyone who took the time to fill out the user survey and use
3. Anycast Routing: Implement anycast routing to direct cloud-based email services; their input and ideas were very
traffic to the closest server based on routing metrics, invaluable. Without their help, our study would never have
reducing network latency. gotten off the ground. Their participation has been invaluable.
4. Intelligent Load Balancing Algorithms: Use load
balancing algorithms that consider not only server References
capacity but also network latency as a factor when
making routing decisions. 1. Feng Qian, Peter Danzig, Sugih Jamin, and Yibo Pi. 2020.
A cloud-centric perspective on latency imbalance among
5. Continuous Monitoring: Monitor network performance Internet load-balanced paths. Article 32, Proc. ACM Meas.
and latency in real-time to adjust load balancing Anal. Comput. Syst. 4, 2, June 2020, 29 p.
configurations dynamically.
342 Algorithms in Advanced Artificial Intelligence
Abstract: Cloud-based email services have become the primary means of communication in the digital age, making efficient
spam detection an essential component in ensuring the integrity and security of electronic communications. This research paper
addresses the pressing challenge of enhancing the precision of spam detection algorithms within the context of cloud-based
email services. The primary goal of this study is to investigate novel strategies that reduce the occurrence of false positives in
spam detection, without compromising the detection of true spam messages. False positives, or legitimate emails incorrectly
classified as spam, not only inconvenience users but can also lead to the loss of important information. Striking the right
balance between accurate spam identification and minimal false positives is crucial for the effectiveness of spam filters. The
paper commences with a comprehensive review of the current landscape of spam detection in cloud-based email services. It
discusses the inherent challenges associated with precision and the detrimental consequences of false positives, such as missed
communications and potential data breaches. Additionally, it highlights the evolving nature of spam and the need for adaptive
and context-aware solutions.
Keywords: Spam, Communications, Precision, Detection, Emails
1
muralimohan.klu@gmail.com, 2rohithapapolu@gmail.com, 3sowjanyam0719@gmail.com, 4madirajusravya@gmail.com
DOI: 10.1201/9781003529231-51
344 Algorithms in Advanced Artificial Intelligence
to improve the precision of spam detection, particularly in also pose security threats and privacy concerns. In [14] the
cloud- based email services. authors discuss current and potential future spam filtering
The significance of this research lies in its potential to address technologies. We look at the problems posed by spam, what
the multifaceted challenges associated with spam detection in spam is and how we measure it.
cloud-based email services. Hence, utilizing social knowledge The State of Spam Detection in Cloud-Based Email Services:
can aid in combating spam, particularly among “independent”
A. Rule-Based Filters
malicious users who do not collaborate [1]. Unwanted and
unsolicited emails, commonly known as spam, are intruding Early spam detection systems predominantly relied on rule-
upon users without their consent, inundating their mailboxes based filters. These filters operated on predefined patterns
with unwanted email clutter [7]. Effective spam detection not and rules, which often struggled to adapt to evolving spam
only enhances the user experience by reducing the clutter of tactics, leading to high false positive rates and missed spam.
unwanted emails but also safeguards users against phishing In knowledge engineering-based spam filtering, rules are
attacks, malware distribution, and other malicious activities devised and implemented, relying on distinct keywords for
often embedded in spam. Moreover, for email service the technical detection of spam as opposed to regular emails.
providers, improved spam detection translates into increased [3]
user trust, reduced operational costs, and enhanced brand B. Machine Learning Approaches
reputation. Consequently, this study holds significance for
both end-users and email service providers alike. The advent of machine learning revolutionized spam
detection. Supervised learning techniques such as Naive
This research paper aims to achieve the understanding the Bayes, Support Vector Machines (SVM), and Random Forests
evolving landscape of email spam, including its various have been widely employed. However, achieving a balance
forms, motivations, and the tactics employed by spammers between high precision and recall remains challenging.
and analyse the unique challenges posed by cloud-based
email services in the context of spam detection, encompassing C. Feature Engineering and Selection
issues of scalability, real-time processing, and integration of Feature Engineering: Researchers have explored
advanced technologies. innovative feature engineering techniques, including text-
based features such as term frequency-inverse document
frequency (TF-IDF), n-grams, and word embeddings.
These features aim to capture nuanced aspects of email
content and improve precision.
Feature Selection: Dimensionality reduction and feature
selection methods, such as Principal Component Analysis
(PCA) and Information Gain, have been applied to identify
the most discriminative features for spam detection,
contributing to enhanced precision.
3. Methodology
3.1 Data Collection and Preprocessing
Data collection and preprocessing are foundational steps
in our research aimed at enhancing the accuracy of spam
Fig. 51.1 Spam emails over the years detection within cloud-based email services. For this study,
we gathered a diverse and representative dataset comprising
thousands of email samples obtained from multiple sources,
2. Literature Review including public email repositories and cloud-based email
Email communication remains an indispensable tool for service providers. The dataset encompasses a balanced
personal, professional, and business interactions, with cloud- distribution of legitimate (ham) emails and spam emails, a
based email services becoming the predominant platform reflection of real-world email traffic dynamics. These emails
for managing electronic communication. Despite their encompass a wide array of linguistic styles, languages, and
convenience, these services are plagued by the relentless content types, thus ensuring the robustness of our analysis
influx of spam emails, which not only clutter inboxes but and model training.
Boosting Precision: Strategies for Improving Spam Detection in Cloud-Based Email Services 345
In the preprocessing phase, we employed a rigorous data Feature selection is crucial for model efficiency and
cleaning process to ensure the quality of our email data. effectiveness. We employed the following techniques:
This involved the removal of extraneous elements such Mutual Information: We calculated mutual information
as HTML tags, special characters, and superfluous white scores between features and the target variable (spam or
spaces. Additionally, we addressed missing or incomplete legitimate) to identify the most informative features. This
email components judiciously to avoid data loss. The guided our selection process, ensuring that only the most
classification and identification of spam emails constitute relevant features were included in our models.
essential measures in the battle against these threats, ensuring
the protection of email communication.[6] The crucial step of Recursive Feature Elimination: To further refine our feature
tokenization was performed to break down the textual content set, we utilized recursive feature elimination (RFE) with
of emails into individual words, facilitating subsequent machine learning models. RFE ranks features by importance
analysis. Feature extraction involved TF-IDF vectorization and iteratively eliminates the least informative
to represent email content as numerical feature vectors,
enabling machine learning algorithms to operate effectively. 3.3 Machine Learning Models
Moreover, metadata and header information, including sender This section discusses the machine learning algorithms
reputation, timestamps, and routing details, were extracted selected and tailored for the task of enhancing spam detection
and integrated as valuable features for spam detection. This accuracy in cloud-based email services. Each algorithm offers
comprehensive data preprocessing pipeline ensures that our unique advantages in addressing the complex challenges
research is conducted on a high-quality dataset, enabling the posed by evolving spam threats.
development of accurate and robust spam detection models Logistic Regression is a straightforward and interpretable
for cloud-based email services.
choice for classifying email spam, modeling the probability
3.2 Feature Engineering and Selection of an email being spam based on features like keywords and
sender information. Support Vector Machines (SVM) are
Feature engineering and selection are critical aspects of a powerful option, capable of handling complex decision
our research aimed at enhancing spam detection accuracy boundaries and high-dimensional feature spaces through
in cloud-based email services. In this section, we detail the kernel functions, but they require careful hyperparameter
methods and strategies employed to construct an effective tuning and can be computationally expensive. Both models
feature set for our machine learning models. Feature selection can be effective, with the choice depending on dataset
can be accomplished through two techniques: textual and characteristics and computational resources.
content-based approaches.[12] Feature engineering involves
the creation of relevant and informative features from the raw Logistic Regression
data, which can significantly impact the performance of our In our research paper on enhancing spam detection accuracy
spam detection algorithms. in cloud-based email services, Logistic Regression emerges
Textual Features as a fundamental and interpretable tool in our arsenal.
Within the context of cloud-based email services, where the
To capture the linguistic characteristics of emails, we
dynamic and evolving nature of spam threats demands agile
engineered the following textual features:
and robust solutions, Logistic Regression offers an elegant
N-Grams: O-We generated n-grams (sequences of contiguous starting point. This simple yet powerful linear classification
words) to capture phrases and contextual information within algorithm excels in distinguishing between legitimate (ham)
emails, allowing our models to discern spammy language and spam emails, assigning probabilistic scores that align
patterns. with the inherent binary nature of spam detection.
Word Frequencies: We computed word frequencies, both The sigmoid activation function inherent to Logistic
globally and within each email, to identify terms that are Regression ensures that predicted probabilities remain within
more prevalent in spam or legitimate emails. the [0, 1] range, facilitating clear-cut classification decisions.
Content-Based Features To empower the model further, we engage in comprehensive
We leveraged the content of emails to create features such as: feature engineering, extracting meaningful information from
email text through TF-IDF vectorization and incorporating
Attachment Presence: We encoded whether an email valuable metadata and header details. Our rigorous model
contained attachments, as this is a common spam indicator. training involves k-fold cross-validation and hyperparameter
URL Count: We counted the number of URLs within an tuning, fine-tuning the model’s parameters for optimal
email, as an excessive number of links can be indicative of performance. Logistic Regression’s interpretability and
spam. explainability also shine through, as it allows us to gain
346 Algorithms in Advanced Artificial Intelligence
critical insights into why certain emails are classified as Then we create the logistic regression object and train it
spam, thereby enhancing our understanding of the model’s with the data. Finally, we create a set of messages to make
decision- making process. In the dynamic landscape of predictions.
cloud-based email services, Logistic Regression serves as a Support Vector Machines
reliable and foundational element in our pursuit of bolstering
spam detection accuracy, ultimately contributing to a safer In our pursuit of enhancing spam detection accuracy
and more secure user experience. within the realm of cloud-based email services, Support
Vector Machines (SVMs) stand out as a formidable and
“Interpretable Functional Logistic Regression” (IFLR) is indispensable tool. SVMs are well-known for their ability
a method that categorizes functional data into two distinct to handle high- dimensional data and non-linear decision
groups, offering a classifier that is both easy to interpret and boundaries, making them a robust choice for addressing
highly predictive. [9] the complex and dynamic nature of spam detection. In the
Here We load the dataset using pandas. Then we split in a context of cloud-based email services, where spam tactics
training and test set. We extract text features known as TF- continually evolve and diversify, SVMs offer a versatile and
IDF features, because we need to work with numeric vectors. effective solution. Advocates of content-based filtering have
LR is used for reduce noisy data or instance before data feed recommended the utilization of Support Vector Machines
to DT induction. Logistic Regression (LR) reduces noisy (SVMs), citing their state-of-the-art performance in text
data by filtering correct predictions based on a specified false classification within the realm of machine learning.[13] By
negative threshold.[2] leveraging the principles of margin maximization, SVMs
meticulously craft an optimal hyperplane that effectively
separates spam from legitimate emails. This inherent capacity
allows SVMs to capture intricate relationships and patterns
within email content, thereby enhancing the accuracy of spam
classification. Additionally, SVMs demonstrate resilience
against overfitting and can accommodate categorical features,
aligning them perfectly with the multifaceted characteristics
of email data. Their precision in distinguishing between spam
and legitimate emails, coupled with the ability to minimize
false positives, makes SVMs an indispensable component
of our spam detection framework, ensuring that users of
cloud-based email services are shielded from unwanted and
potentially harmful content.
Furthermore, SVMs bring an element of adaptability to our
spam detection system, a crucial requirement in the ever-
changing landscape of cloud-based email services. Support
vector machine (SVM) classifiers can cope with many different
classification tasks but improperly selected hyperparameters
may deteriorate their performance. [8] By effectively
handling non-linear feature interactions, SVMs can adapt to
emerging spam tactics, allowing email service providers to
stay ahead of evolving threats. Their utility extends beyond
mere accuracy to encompass real-time decision-making, a
pivotal feature in a dynamic email environment where user
experience and security are paramount. The combination of
precision, resilience, and adaptability positions SVMs as a
cornerstone of our efforts to enhance spam detection accuracy
in cloud-based email services, ensuring that users can enjoy a
safe and uninterrupted email experience.
Here We load the dataset using pandas. Then we split in a
training and test set and the we apply SVM algorithm to get
Fig. 51.2 Logistic regression algorithm accuracy.
Boosting Precision: Strategies for Improving Spam Detection in Cloud-Based Email Services 347
8. Wojciech Dudzik, Michal Kawulok, and Jakub Nalepa. 2019. 12. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis
Evolutionarily-tuned support vector machines. In Proceedings in Health Care Using Hybrid Model Algorithms”,Journal of
of the Genetic and Evolutionary Computation Conference Artificial Intelligence, Machine Learning and Neural Network
Companion (GECCO ‘19). Association for Computing (JAIMLNN), Volume 3, 2023.
Machinery, New York, NY, USA, 165–166. https://doi. 13. Ala’ M. Al-Zoubi, Hossam Faris, Ja’far Alqatawna,
org/10.1145/3319619.3321924 Mohammad A. Hassonah,Evolving Support Vector Machines
9. Cui Lv and Di-Rong Chen. 2018. Interpretable Functional using Whale Optimization Algorithm for spam profiles
Logistic Regression, 2nd International Conference on detection on online social networks in different lingual
Computer Science and Application Engineering (CSAE ‘18). contexts, Knowledge-Based Systems, Volume 153, 2018,
Association for Computing Machinery, New York, NY, USA, Pages91- 104, ISSN0950-7051, https://doi.org/10.1016/j.
Article 82, 1–5. https://doi.org/10.1145/3207677.3277962 knosys.2018.04.025.
10. Bilge Kagan Dedeturk, Bahriye Akay,Spam filtering 14. D. Sculley and Gabriel M. Wachman. 2007. Relaxed online
using a logistic regression model trained by an artificial SVMs for spam filtering. In Proceedings of the 30th annual
bee colony algorithm,Applied Soft Computing,Volume international ACM SIGIR conference on Research and
91,2020,106229,ISSN 1568-4946,https://doi.org/10.1016/j. development in information retrieval (SIGIR ‘07). Association
asoc.2020.106229. for Computing Machinery, New York, NY, USA, 415–422.
11. C. Tseng, J. Huang, and M. Chen, “ProMail: Using Progressive https://doi.org/10.1145/1277741.1277813
Email Social Network for Spam Detection,” Advances in 15. R. Hunt and J. Carpinter, “Current and new developments in
Knowledge Discovery and Data Mining, LNCS, vol. 4426, spam filtering”, 14th IEEE International Conf. on Networks,
pp. 833 840, 2007. pp. 1-6, 2006.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
350 Algorithms in Advanced Artificial Intelligence
R. Tamilkodi1
Professor, Department of CSE (AIML & CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
A. Harika2
Assistant Professor, Department of CSE (AIML & CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Ch. Rohith3, G. Nithin4, K. Mahesh5, A. Anvitha6, N. Lohitha7
Department of Computer Science & Engineering (AIML & CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Abstract: Online entertainment is constantly changing, and movie recommendations need to match individual preferences for
user satisfaction. This challenge can be met by implementing a Mood-Based Cascade Hybrid movie recommendation system
(MBCH). This innovative system MBCH uses two different data sets, linked by a common movie ID, to improve the accuracy
and effectiveness of movie suggestions. The first phase of this system is based on content filtering. Data from the internet is
collected to create a large dataset, with various movie attributes such as genre, directors, cast, plot summaries, and more. These
attributes are used to make initial recommendations, ensuring that suggestions are closely related to the content of each film.
The second phase combines collaborative filtering and latent factor models, creating personalized suggestions that are better
than content filtering alone. The goal of MBCH is to provide a dynamic and personalized movie-watching experience. By
combining content filtering and Collaborative filtering, it offers precise recommendations that suit the moods and preferences
of users. This approach not only increases user satisfaction but also gives your movie streaming platform an edge in today’s
competitive entertainment market.
Keywords: Collaborative recommender system, Content recommender system, Cascade hybrid recommender system, Mood-
based, User feedback, Personalized suggestions
1. Introduction to identify patterns and predict what content and products are
likely to interest users.
In the contemporary world, the internet is an essential tool Recommendation systems are very useful and effective
for connecting people, sharing information, and supporting technique of filtering the data [23]. A recommendation
various aspects of modern life. Recommendation systems system is a personalized information filter that tailors its
are widely used in the digital realm to help users discover choices to a user’s preferences and interests. In today’s
relevant content and products. These systems analyze user era of information overload, these systems are crucial for
actions, including ratings, purchases, and browsing history, e-commerce and social platforms. Many platforms, like
1
tamil@giet.ac.in, 2harikaadduri07@giet.ac.in, 3rohitch1418@gmail.com, 4nithingedda45@gmail.com, 5kondumahanthimahesh1@gmail.com,
6
anvithaattunuri@gmail.com, 7lohithanallamadugu@gmail.com
DOI: 10.1201/9781003529231-52
Crafting Personalized Film Suggestions 351
Netflix for suggesting movies, Amazon for offering product movie landscape. This makes them superior to single-method
recommendations, Spotify for providing music suggestions, recommendation systems in terms of both accuracy and
LinkedIn for proposing job opportunities, and various efficiency [25].
social networking sites for suggesting connections, all
function through recommendation systems [19], [20]. Movie 2. Literature Survey
recommendation systems filter out irrelevant data and only
include data that has matching characteristics or features [18]. Kim Mucheol et al. [6] described an interactive movie
They help users to find interesting items customized to their recommendation system for online communities, utilizing a
preferences, making online experiences more enjoyable and community network model to understand social information
efficient [1]. Movie recommendation systems are invaluable dynamics. This model tailor’s movie suggestions to
in helping us find our favorite films amid the vast array of individual user preferences and adapts to evolving social
options, saving us time and effort [2]. These systems must be network trends. Nanou et al. [7] described challenges
highly dependable to be effective, offering recommendations in movie recommendation presentation and evaluated
that closely match our preferences and interests. different methods. The study highlighted the effectiveness of
“planned outline” and “textbook and videotape” interfaces in
Mood-based movie recommendation systems assist users in
establishing a strong link between user opinions and approval
discovering films that match their current emotional state.
across various experimental scenarios.
Users can select their mood, and the system then suggests
movies known to elicit those feelings. Ruotsalo et al. [8] introduced a mobile recommendation
system, emphasizing SMARTMUSEUM, employing
Three distinct models in movie recommendation systems
semantic network speech and ontologies to connect se-
are content-based filtering, collaborative filtering, and
mantic gaps, sensor data, and user profiles. The system
popularity- based filtering. The choice of model for mood-
employed an information retrieval framework, and its results
based movie recommendations depends on the system’s
demonstrated effectiveness in meeting user needs. Sharma
design, each having unique strengths in tailoring suggestions
et al. [9] this paper reviewed the several approaches used
to user preferences and emotions.
for the Recommen- dation system. Approaches may be
Content-based filtering recommends movies based on the categorized into two parts Content filtering and Content-
content of previously liked movies. This can include features based recommendation. Also, this paper describes the merits
such as the genre, director, actors, and plot [4]. Content-Based and demerits of the recommendation approaches. Tekin
Filtering, also referred to as cognitive filtering [21], operates et al. [10] proposed distributed online learning in a social
by suggesting items to users based on their past interactions recommendation system, where suggestions are query-
and preferences. Collaborative filtering is an approach dependent. Recommendations consider user history, gender,
that suggests items to users by analyzing the resemblances and age. The approach emphasizes decentralized sequential
between users and the items themselves [24]. Collaborative decision-making.
filtering models suggest movies to users by considering the
Chapphannarungsri and Maneero [11] presented a
ratings and preferences of other users with akin tastes [3].
multidimensional approach for an advanced recommendation
Goldberg introduced the concept of collaborative filtering system, offering high-quality recommendations. They
in 1991 [22]. Models utilizing popularity-based filtering proposed a method for the Multiple Criteria approach,
suggest movies to users by taking into account their level adjusting weightings and addressing film feature selection
of popularity, which could be determined by factors like the concerns, and applied Multiple Linear Regression to study
frequency of views or ratings. client characteristics, resulting in more accurate outcomes
The hybrid movie recommendation system, on the other hand, compared to existing Hybrid Recommendation systems.
represents an approach that merges various recommendation George et al. [12] presented the mixed approach combining
models to provide users with a more extensive and precise con- tent-based and collaborative filtering for a film
assortment of movie suggestions. By merging content-based recommendation system. The approach was experimentally
filtering, collaborative filtering, and other techniques this evaluated against existing collaborative and content-based
hybrid system aims to overcome the limitations of methods filtering techniques, offering valuable insights into its
[5]. The hybrid recommendation system is an advanced performance. Yoshii et al. [13] an incrementally trainable
approach that combines user preferences, movie attributes, probabilistic model for mixed recommendations, combining
and historical data to provide diverse and personalized collaborative and content-based techniques to enhance
movie recommendations. This method enhances the user precision and artist diversity. The model effectively merges
experience by aligning suggestions with individual tastes collaborative and content data, maintaining high accuracy,
and emotions, making it a powerful tool in today’s extensive even with the inclusion of new users.
352 Algorithms in Advanced Artificial Intelligence
Christakou et al. [14] introduced a clustering technique based recommendations further. While this innovative system offers
on semi-supervised learning for movie recommendations exciting opportunities for personalized movie suggestions, it
that combine collaborative and content-based data. Their must prioritize privacy and data security, as well as scalability
system was tested on the MovieLens dataset, yielding high- to accommodate a growing user base, ensuring a seamless
precision recommendations. Symeonidis et al. [15] proposed and enjoyable user experience.
“MoviExplain,” a movie recommendation system that extends However, the existing mood-based movie recommendation
beyond mere movie recommendations. It aims to provide system that relies on facial emotion analysis faces several
both accurate and justifiable recommendations, allowing challenges that could affect its accuracy and reliability. One
users to understand the rationale behind a recommendation. of the primary issues is the ambiguity of facial expressions,
Recommender systems predict user preferences by analyzing which can convey multiple emotions and potentially lead
activity patterns and are categorized as content-based (CBR) to misinterpretations. Emotions are also highly subjective
and collaborative filtering (CF) models. CBR uses item and can vary from person to person, making it challenging
properties for predictions, while CF identifies user or item
to predict what a user might genuinely enjoy. Additionally,
similarities to make recommendations based on users with
emotions can change rapidly during a movie or be influenced
similar tastes, harnessing collective user wisdom.[16].
by external factors, presenting a considerable challenge for the
Collaborative filtering (CF) includes memory-based techniques system to adapt in real time. Ethical and legal concerns related
that identify similar users through rating comparisons, to privacy and data security are critical, as continuous facial
forming user-based recommender systems. Model- based analysis may raise questions. Furthermore, the technology’s
CF analyzes rating patterns to build precomputed item limited emotional range and potential biases in training
recommendation models. Content-based recommendation data could hinder its ability to accurately detect complex or
(CBR) systems use item features and attributes to find and culturally diverse emotions. Often, the system over- looks
suggest similar items based on user preferences. [17]. Both crucial contextual factors, such as a user’s intention behind
CF and CBR methodologies encounter certain restrictions. watching a particular movie or their social environment. To
CF-based recommender systems typically encounter the mitigate these challenges and enhance accuracy, developers
problems of [16] sparse rating as many users don’t rate items must continuously refine the technology, incorporate user
and [17] cold start problems. The cold start challenge within feedback, and consider a broader range of factors beyond
recommendation systems pertains to situations involving facial expressions when making movie recommendations.
novel users and items. The new user cold start problem occurs
when a user with no ratings joins, mainly affecting content-
based recommendation (CBR). Both collaborative filtering
4. Proposed System
(CF) and CBR encounter difficulties in generating reliable The proposed methodology MBCH uses a user-centric
recommendations due to these limitations. [16]. approach to overcome the limitations of existing systems that
rely on capturing emotions through a camera. A two- step
3. Overview of Existing System process to address privacy concerns and improve accessibility
was implemented. The proposed system working architecture
An advanced approach to enhancing user engagement and is shown in Fig. 52.1.
satisfaction involves a mood- based movie recommendation
In the first stage, a standardized survey is used to assess user
system that utilizes facial emotion analysis. This system
preference across different moods. The survey asks users to
employs sophisticated computer vision techniques and
machine learning models to identify and categorize a choose movies they would like to watch in different emotional
user’s emotional state based on their facial expressions, en- states, such as “What kind of films help you unwind and
compassing emotions such as happiness, sadness, anger, relax when you’re feeling down?”,” In moments of love and
and surprise. Subsequently, it taps into a comprehensive affection, which movies resonate with you?” these provide
movie emotion database, associating films with specific vital information for the recommendation process, this
emotional tags, enabling the system to align the user’s current information serves as the foundation for our cascade hybrid
emotional state with ap- propriate movie recommendations. model.
To enhance precision, the system takes into consideration The second step adopts a two-tiered approach, beginning with
the user’s viewing history and preferences, ensuring content-based filtering. Textual attributes such as overview,
a more tailored and personalized experience. These tagline, storyline, genre, cast, and director are transformed
recommendations are delivered through a user-friendly inter- into numerical vectors using TF-IDF in this instance. The
face, offering comprehensive details about each suggested cosine similarity is then evaluated to detect similarities
movie. The system encourages user feedback to refine its between movies.
Crafting Personalized Film Suggestions 353
the year when the movie was released, how long the movie The third step is to display the dataset in a table or a chart
lasts, how the movie was rated by the viewers, how many that can show the data in a clear and organized way. The
people voted for the movie, how much money the movie fourth step is to transform the dataset into a format that the
made, and how the movie was scored by the critics. recommendation model can use, such as a matrix or a vector,
The dataset is processed in several steps to make it ready for that can represent the similarities and differences between the
the recommendation model. The first step is to import the movies.
dataset from the IMDB website using specific tools that can Within this dataset, you’ll find 2,000,0263 ratings distributed
extract the data. The second step is to summarize the dataset across 27,278 distinct movies. The origins of this data
by calculating some statistics, such as the average rating, the can be traced back to 138,493 users who contributed their
number of movies per genre, the most popular actors, etc. ratings between January 9, 1995, and March 31, 2015. This
comprehensive dataset was curated on October 17, 2016. The graph compares four recommendation systems:
Notably, the users included in this dataset were chosen collaborative filtering, weighted recommendations, emotion-
through a random selection process, and it is noteworthy that based recommendations, and the proposed system MBCH.
every user selected had actively voted for a minimum of 20 The MBCH consistently gives more accurate results to users.
movies. This shows that the proposed system MBCH is better at
When users sign up for the first time, they are requested to suggesting movies to users
complete questionnaires in order to determine their movie
preferences. 6. Conclusion
In proposed MDCH, we’ve effectively integrated web
scraping, Mood Assessments, Content-Driven Filtering, and
Collaborative Filtering to deliver users exceptionally tailored
movie recommendations, taking into account their emotional
state and preferences. This innovative system enhances user
engagement and contentment by offering them pertinent
movie recommendations that align with their present
emotional disposition.
References
1. J. Jose Immanuvel, A. Sheelavathi, M. Priyadharshan, S.
Fig. 52.4 Represents the feelings of the user while login
Vignesh, K. Elango, “Movie Recommendation System”,
IJRASET, 2022-06-17
2. Hao F, Park DS, Pei Z (2017) Detecting bases of maximal
cliques in social networks. MUE2017 1–1
3. Desrosiers C, Karypis G (2011) A comprehensive survey of
neighborhood-based recommendation methods. In: Ricci F,
Rokach L, Shapira B, Kantor P(eds) Recommender systems
handbook. Springer, Boston, pp 107–144
4. Ricardo Baeza-Yates, Berthier Ribeiro-Neto, et al. Modern
information retrieval, volume 463.ACM Press New York,
1999.
5. Gediminas Adomavicius and Alexander Tuzhilin. Toward the
next generation of recommender systems: A survey of the
state-of-the-art and possible extensions. Knowledge and Data
Fig. 52.5 Gives the output after processing the user feelings Engineering, IEEE Transactions on, 17(6):734–749, 2005.
6. Kim, Mucheol, and Sang Oh Park, “Group affinity-based
social trust model for an intelligent movie recommender
system”, Multimedia tools and applications 64, no. 2, 505
516, 2013
7. Nanou, Theodora, George Lekakos, and Konstantinos
Fouskas, “The effects of recommendations “presentation on
persuasion and satisfaction in a movie recommender system”,
Multimedia systems 16, no. 4-5, 219-230, 2010.
8. Ruotsalo, Tuukka, KristerHaav, Antony Stoyanov,
Sylvain Roche, Elena Fani, RominaDeliai, Ee- tuMäkelä,
TomiKauppinen, and EeroHyvönen, “SMARTMUSEUM:
A mobile recommender system for the Web of Data”, Web
semantics: Science, services, and agents on the world wide
web 20, 50-67, 2013.
9. Sharma, Meenakshi, and Sandeep Mann, “A survey of
recommender systems: approaches and limitations”, Int J
InnovEng Technol. ICAECE-2013, ISSN, 2319-1058, 2013.
10. Tekin, Cem, Shaoting Zhang, and Mihaela van der Schaar,
Fig. 52.6 Comparison with other systems “Distributed online learning in social recommender systems”,
356 Algorithms in Advanced Artificial Intelligence
Selected Topics in Signal Processing, IEEE Journal of 8, no. 18. Pazzani MJ, Billsus D (2007) Content-based recommendation
4, 638-652, 2014. systems. In: Brusilovski P, Kobsa A, Nejdl W (eds) The
11. Keittima Chapphannarungsri and Saranya Maneero, adaptive web. Springer, Berlin, pp 325–341.
“Combining multiple criteria and multidi- mensional for movie 19. Çano E., Morisio M. Hybrid recommender systems: A
recommender system”, in Proceedings of the International systematic literature review. Intell. Data Anal. 2017; 21:1487–
MultiConference of Engineers and Computer Scientists, vol. 1524. doi: 10.3233/IDA-163209
1, 2009 20. S. Wattal, Y. Hong, M. Mandviwalla, and A. Jain,” Technology
12. Lekakos, George, and Petros Caravelas, “A hybrid approach for diffusion in the society: Analyz- ing digital divide in the context
movie recommendation”, multi-media tools and applications of social class”, IEEE Proc. of 44th Hawaii International
36, no. 1-2, 55-70, 2008 Confer- ence on System Sciences, 1-10, 2011. DOI: http://
13. Yoshii, Kazuyoshi, Masataka Goto, Kazunori Komatani, dx.doi.org/10.1109/HICSS.2011.398
Tetsuya Ogata, and Hiroshi G. Okuno, “An efficient hybrid 21. M. Goldmann and G. Kreitz,” Measurements on the spotify
music recommender system using an incrementally trainable peer-assisted music-on-demand streaming system”, IEEE
probabilistic generative model”, Audio, Speech, and Language International Conference on Peer-to-Peer Computing, 206
Processing, IEEE Transactions on 16, no. 2, 435-447, 2008 211, 2011. http://dx.doi.org/10.1109/P2P.2011.6038737
14. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis 22. H. Li, F. Cai, and Z. Liao,” Content-based filtering
in Health Care Using Hybrid Model Algorithms”,Journal of recommendation algorithm using HMM”, IEEE Fourth
Artificial Intelligence, Machine Learning and Neural Network International Conference on Computational and Information
(JAIMLNN), Volume 3, 2023. Sciences, 275-277, 2012.
15. Christakou, Christina, Leonidas Lefakis, Spyros Vrettos, 23. D. Goldberg, D. Nichols, B. M. Oki, and D. Terry, ”Using
and Andreas Stafylopatis, “A movie rec- ommender system collaborative filtering to Weave an Information tapestry”,
based on semi-supervised clustering”, in Computational Communications of ACM 35(12):61-70, 1992. DOI: http://
Intelligence for Mod- elling, Control and Automation, dx.doi.org/10.1145/138859.138867
2005 and International Conference on Intelligent Agents, 24. Gupta S. A Literature Review on Recommendation Systems.
Web Technologies and Internet Commerce, International Int. Res. J. Eng. Technol.2020;7:3600–3605.
Conference on IEEE, vol. 2, pp. 897-903, 2005 25. Shen J., Zhou T., Chen L. Collaborative filtering-based
16. Symeonidis, Panagiotis, Alexandros Nanopoulos, and Yannis recommendation system for big data. Int. J. Comput. Sci. Eng.
Manolopoulos, “MoviExplain: a recommender system with 2020;21:219–225. doi: 10.1504/IJCSE.2020.105727.
explanations”, In Proceedings of the third ACM conference 26. Beniwal R., Debnath K., Jha D., Singh M. Data Analytics and
on Rec- ommender systems, pp. 317-320, 2009. Management. Springer; Berlin/Hei- delberg, Germany: 2021.
17. Adomavicius G, Tuzhilin A (2005) Toward the next generation Hybrid Recommender System Using Artificial Bee Colony
of recommender systems: a survey of the state-of-the-art and Based on Graph Database; pp. 687–699
possible extensions. IEEE Trans Knowl Data Eng 6:734–749.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
A Comprehensive Approach to
Detect SQL Injection Attacks Using
Enhanced Snort Rules
53
T. Srinivasarao1
Assistant Professor, Department of ECE
GIET (Autonomous), Rajahmundry, Andhra Pradesh, India
Shrija Madhu2
Professor, Department of CSE
GIET (Autonomous), Rajahmundry, Andhra Pradesh, India
K. Kalyani Vishalakshi3, Preetish Madhu4, K. Satya Sai DurgaManikanta5, P. Sumanth Yadav6
Department of CSE (AIML & CS)
GIET (Autonomous), Rajahmundry, Andhra Pradesh, India
Abstract: SQL Injection attacks continue to be a significant danger to web applications, with the potential for unauthorized
access, data breaches, and application vulnerabilities. Using updated Snort rules, this paper describes a complete approach for
detecting and mitigating various kinds of SQL Injection attacks. We suggested a solution to provide a robust and adaptable
defensive mechanism against various kinds of SQL Injection attacks by addressing the drawbacks of existing detection
methods. Inthis proposed method we illustrate the practical use of the upgraded Snort rules in real-world web application
contexts through comprehensive testing and evaluation. The significance of this paper is that it provides a realistic and effective
solution for organizations to protect their web applications against SQL Injections attacks while protecting sensitive data and
user privacy. This paper will enableeven non-experts to deploy powerful SQL Injections detection capabilities by employing
simpler Snortrules, empowering enterprises of all sizes to improve their security posture. The results of this study will benefit
network security by providing a more advanced and proactive approach to SQL Injections detections, providing the way for
future intrusion detection research and the development of enhancedSnort rules for emerging threats.
Keywords: SQL injections, Web application security, Intrusion detection system (IDS), Snort, Attack detection
DOI: 10.1201/9781003529231-53
358 Algorithms in Advanced Artificial Intelligence
some degree of effectiveness. The aim of this research is Many other approaches are available for available for detecting
to improve Snort IDS capacity for detecting various kinds and securing documents using Encryption Techniques also
of SQL Injections. We will cover an in-depth set of SQL [16-18].
Injection variations, including Classic, Blind, Time based,
Error based, Union, Boolean, Second order and Out of Band 3. Proposed Methodology
SQLInjections types. The efficiency of the suggested system,
including its accuracy, adaptability, and resource usage, will However, there are some drawbacks in the existing methods,
be thoroughly examined. The establishment of a full-fledged such as narrow coverage, a high probability of false positives,
web application firewall and additional security measures and complex procedures. The proposed methodology would
that go beyond Snort IDS are not the focus of this research. offer a thorough and useful method for quickly and efficiently
recognizing different types of SQL Injection attacks which is
shown in Fig. 53.1. We aim to improve the real-time detection
2. Related Work and prevention of SQL Injection attempts by leveraging the
In this part, we present a comprehensive literature work capabilities of Snort IDS. This will also help to reduce false
that explores the existing research on detecting SQL attacks positives by modifying the Snort rules.
detection. The survey covers papers related to SQL Injections
detection and examines their methodologies, findings, and
limitations. The studies collectively highlight the evolving
strategies for detecting SQL injections for further prevention
steps.
The authors Gupta and Sharma introduced a novel approach
leveraging Snort for effective detection [1]. The authors
proposed an evidence-collection and notification model
using standard IDS systems in network forensics, providing a
comprehensive approach [2]. Caesarano and Riadi introduced
the NIST method in network forensics to detect SQL injection
attacks. This approach integrates NIST’s guidelines and
standards for effective attack identification and response [3].
Alnabulsi et al. focused on the utilization of SNORT IDS for
accurate detection [4]. Fig. 53.1 Block diagram of methodology
stances to check the capability of Snort in order to identi alert tcp any any -> any any (msg:"Possible Time-Based
fy and react to those threats. This stage acts as an important Blind SQLInjection Alert"; flow:to server, established;
stage of validation, verifying that the rules work as intended. content:"'AND IF (1=1, SLEEP (5),0) --"; http_uri;
sid:1000003; rev:1;)
4. Testing & Optimizing Stage #Rule 4: Rule for SQL Injection attack (Based on error)
This rule detects Error based SQL Injection Alerts by
Fine Tuning & Optimization: The need for optimization and
injecting code that triggers an error with the database.
fine-tuning may depend on the findingsof the testing phase.
This requires a thorough examination of the Snort alerts, in alert tcp any any ->> any any (msg: "Possible Error
cluding an evaluation of their accuracy. To reduce false pos type SQL Injection Alert"; flow:to_server, established;
itives and improve the system’s general detection capability, content:"'AND 1=CONVERT (int, @@version)
modifications can be made to the rules. For the best results, -"; http_uri; sid:1000004; rev:1;)
both precision and sensitivity must be balanced properly. #Rule 5: Rule for Union SQL Injection attack
Benchmarking for Performance: Benchmarking is crucial This rule detects Union based SQL Injection Alerts by
to verify the IDS effectiveness and responses. The system injecting the UNION keyword, alerting tocombine results
is tested through a range of simulated network traffic to from different database queries.
determine how effectively it performs under various traffic alert tcp any any -> any any (msg:"Possible Union
scenarios. Potential obstacles or areas for improvement SQL Injection Alert"; flow:to_server, established;
canbe found by tracking response times and consumption of content:"UNION"; http_uri; sid:1000005; rev:1;)
resources. During this stage, the IDS is tested to make sure #Rule 6: Rule for Second order SQL Injection attack
it can handle the expected network traffic and successfully This rule detects Second order SQL Injection Alerts, where
identify SQL injection attacks. the malicious payload is stored for laterexecution.
Continuous Monitoring & Updates: Continuous monitoring alert tcp any any -> any any (msg:"Possible Second-
of the IDS after implementation is required to verify its Order SQL Injection Alert"; flow:to_server, established;
continued efficiency. This includes real-time alarm analysis content:"'OR '1'='1''; http_uri; sid:1000007; rev:1;)
and regular assessments of the effectiveness of the system. #Rule 7: Rule for Out of band SQL Injection attack
Furthermore, the rule set should be updated on a regular basis This rule identifies Out of band SQL Injection Alerts by
to incorporate the most recent threat intelligence and adapt to Alerting to retrieve data through an alternative channel.
emerging attack strategies. This step assures that the IDS will
alert tcp any any -> any any (msg:"Possible Out-of-
continue to provide a strong defense against SQL injection
Band SQL Injection Alert"; flow:to_server, established;
attacks all over time.
content:"UNION SELECT NULL, load_file('/etc/
passwd'), NULL --"; http_uri; sid:1000008; rev:1;)
5. Proposed Snort Rules #Rule 8: Rule for Boolean SQL Injection attack
#Rule 1: Rule for Classic SQL Injection attack This rule detects boolean SQL Injection Alerts by injecting a
This rule helps to detect classic SQL Injection by injecting a condition that always evaluatesto true.
condition that always evaluates to true. alert tcp any any -> any any (msg:"Possible Boolean
alert tcp any any -> any any (msg: "Possible Classic SQL SQL Injection Alert"; flow:to_server, established;
Injection Alert"; flow:to_server, established; content:"' content:"'AND 1=1 --"; http_uri; sid:1000009; rev:1;)
OR 1=1"; http_uri; sid:1000001; rev:1;)
#Rule 2: Rule for Blind SQL Injection attack 6. Testing & Evaluation
This rule focuses on Blind SQL Injection Alerts by injecting 6.1 Experimental Setup
a sleep command, causing a delay in the server's response. Step 1: Select any Hypervisor like VM-ware
alert tcp any any -> any any (msg: "Possible Blind SQL
Injection Alert"; flow:to_server, established; content:"' Step 2: In VMware install the following operating systems:
AND SLEEP (5) --"; http_uri; sid:1000002; rey:1;) 1. Ubuntu – for SNORT IDS installation
# Rule 3: Rule for Time based blind SQL Injection attack 2. Kali Linux – as an Attacking machine
This rule detects Time based blind SQL Injection by injecting 3. Metasploitable2– as a testing webpage
a condition that causes a delay if true.
360 Algorithms in Advanced Artificial Intelligence
sudo apt-get install snort Fig. 53.4 Classic SQL injection attack
A Comprehensive Approach to Detect SQL Injection Attacks Using Enhanced Snort Rules 361
T. Srinivasarao1,
Assistant Professor, Department of ECE,
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
N. Leelavathy2,
Professor, Department of CSE,
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
S. Kailash Chandra Sri Satya Dev3, I. Om Ganesh4,
P. Sai Aditya5, P. Sai Krishna6
Department of Computer Science and Engineering (AIML & CS)
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Abstract: DNS spoofing attacks are a growing threat to network security, allowing attackers to manipulate DNS responses
and redirect users to malicious websites. Many existing methods for detecting these attacks have limitations, that reduce their
effectiveness. This proposed method aims to overcome these limitations by designing and implementing an improved ARP and
DNS spoofing detection with attacker IP tracing. The proposed method primary objective includes real-time monitoring of ARP
responses to detect abnormalities indicating of ARP spoofing attempts. This is examined by checking source IP, MAC addresses
for discrepancies. Similarly, the solution thoroughly examines DNS responses, instantaneously warning the administrators
of anomalies between requested and received IP addresses, indicating probable DNS spoofing incidents. Along with these
detection capabilities, the project tracks attacker IP addresses, allowing administrators to trace and examine suspicious spoofing
attempts. In summary,”ARP and DNS Spoofing Detection with Attacker IP Capturing” defines a significant improvement in
network security, providing an effective protection against sophisticated spoofing attacks. The result of this research helps
network administrators and organizations to secure their networks by detecting these ARP and DNS spoofing attacks by tracing
the attacker’s IP.
Keywords: DNS spoofing, ARP spoofing, Network security, Intrusion Detection System (IDS), Python
1
srinu.thupakula@giet.ac.in, 2drnleelavathy@gmail.com, 3kailashchandra.sri.satya@gmail.com, 420551a4622.omganesh@gmail.com,
5
20551a4657.adityapolisetti@gmail.com, 620551a4642.saikrishna@gmail.com
DOI: 10.1201/9781003529231-54
364 Algorithms in Advanced Artificial Intelligence
packet interception. DNS spoofing takes use of flaws in the potential harm if the server fails or is compromised, resulting
DNS resolution mechanism to redirect people to malicious in the server becoming a single point of failure2. Even though
websites. These types of attacks can result in unauthorized they are simple, static entry techniques are less appropriate
access to sensitive data, intercepting of confidential for large- scale networks and dynamic situations since manual
exchanges, and even the insertion of harmful payloads within IP address assignments are impractical in these contexts3.
the network. Host-based solutions, while conceptually strong, face issues
Traditional security methods may be ineffective in preventing in determining the reliability and relevance of each host,
making them more difficult for real-world implementation5.
ARP and DNS spoofing attacks. While firewalls and intrusion
detection systems are important, they may lack the detailed
packet-level analysis needed to detect minor spoofing 3. Proposed Methodology
attempts. As a result, a dedicated solution capable of real-
While existing literature on ARP and DNS spoofing
time monitoring and quick response is required. This project
detection techniques, such as cryptographic, server- based,
bridges the gap by utilizing Python’s scripting capabilities and other defence mechanisms, the proposed research
and Scapy’s packet manipulation abilities, providing an introduces a comprehensive and innovative solution that
effective defiance against these emerging threats. The Python addresses limitations in current approaches.Unlike some
scripts runs in the background of the network, continuously cryptographic systems, which may rely on single points
analyzing incoming ARP and DNS packets. When a probable of failure, such as Authoritative Key Distributors (AKD)1,
spoofing attempt is detected, the solution sends immediate the proposed method provides a distributed and reliable
alerts with the suspected attacker’s IP address. This feature detection system. By combining Python’s adaptability with
enables administrators to take quick decisions to reduce Scapy’s precision, the system provides real-time monitoring
potential threats and protect their networks. Finally, “ARP of ARP and DNS responses, allowing for immediate alerts
and DNS Spoofing Detection with Attacker Capturing” in the event of suspicious activity. In addition, the project’s
makes an important addition to network security. The solution dynamic tracking of attacker IP addresses represents a major
solves the important need for comprehensive ARP and DNS improvement over static entries or manual assignments. This
spoofing detection by utilizing Python’s flexibility and Scapy feature allows administrators to quickly trace and investigate
precision, providing network security professionals with unusual actions, improving their capacity to respond to
an important resource to protect their networks and defend potential threats.
against evolving cyber threats.
Step 1: Understanding Spoofing attacks: Gaining a
thorough understanding of spoofing attacks is the primary
2. Related Work objective of the first phase of the methodology. This involves
Morsyet al. Proposed1,2 an innovative approach to counter studying various spoofing strategies, including DNS and
ARP spoofing attacks and a comprehensive survey on DNS ARP spoofing, and their impact on network security.
attack detection and security protection, offering insights into Step 2: Developing code using Scapy in Python: Scapy,
various strategies for safeguarding DNS infrastructures.The a powerful package known for its abilities in packet
research by Hijazi and Obaidat analyzes Address Resolution manipulation and network traffic analysis, is used along with
Protocol spoofing attacks and the security measures taken to Python’s capabilities to create Python based script that can
prevent them, adding to our knowledge of network security capture and analyze network packets.
issues.ARP spoofing detection and mitigation strategies are Step 3: Project and Environment setup: A secure virtual
thoroughly covered in Rohatgi and Goyal’s work, which is environment is created to protect the host machine while
an essential tool for understanding and defeating such a kind running tests. Isolated virtual machines (VMs) that replicate
of attack.Public-key cryptography is used by cryptographic actual network setups were created using virtualization
solutions, such as S-ARP, to authenticate ARP answers. technologies like VMware. These VMs host the Windows
This strategy has limitations even though it works well. It is target, Kali Linux attacker, and Ubuntu monitor machines,
dependent on an Authoritative Key Distributor (AKD), which allowing for thorough testing while protecting the integrity
increases the possibility of a single point of failure. of the host system.
Furthermore, hosts have to get in touch with the AKD for each Step 4: Code Execution: Execution of the developed
ARP reply, which could lead to dependencies and delays1. Python code within the controlled virtual environment is
An alternative method is provided by server-based systems the next phase. This step makes sure that network traffic is
that use a dependable server to examine packets. However, constantly and uninterruptedly monitored in real time. The
they have a critical vulnerability that exposes the network to created Python script captures and analyzes packets as they
ARP and DNS Spoofing Detection with Attacker IP Capturing 365
transit over the network, giving the required parameters for Kali Linux for simulated attacks. Ubuntu hosts the detection
detecting DNS and ARP spoofing attacks. system, while Windows and Kali Linux play the roles of
Step 5: Detecting DNS and ARP Spoof: The detection victim and attacker, respectively. Controlled experiments in
of DNS and ARP spoofing attacks is the crucial stage. By this virtual lab setting will evaluate the system’s effectiveness
using the developed Python code’s real-time packet analysis in identifying ARP and DNS spoofing attacks.
capabilities, the network packets are examined for any Step 1: Select any Hypervisor like VMware.
indications of possible spoofing attacks. Identification of Step 2: In VMware install the following operating systems:
suspicious patterns, such as the emergence of several IP
1. Ubuntu – Works as network monitoring system
addresses associated to a single host or differences in DNS
answers, will generate an alert message. 2. Kali Linux – Works as attacker machine
3. Windows – Works as a Victim machine
Step 6: Tracking Attacker IP: In this phase, the attacker
IP address will be captured along with the generated alert Step 3: Provide Internet to the VMware.
message. The Python code created in the earlier phase will
identify the attacker’s IP address by tracing out the attack’s 5. Testing and Results
origin.
Testing is a crucial part, especially when implementing
Step 7: Performance Metrics and Real Time Monitoring:
security features like DNS and ARP Spoofing detection
To evaluate the effectiveness of the detection capability. The
methods. Here is a step-by-step approach for doing
detection accuracy, false positive/negative rates, and response
detailed testing.
time will be evaluated in this phase. Continuous real-time
monitoring improves in staying up to date on emerging Step1: Open VMware and switch to Ubuntu Operating
threats and enhancing detection accuracy. system.
Step 2: Develop the python script and save it in the Network
4. Experimental Setup monitoring system (Ubuntu).
Step 3: Execute the Script with root permissions as shown in
The test environment consists of three virtual machines:
Fig. 54.2, to monitor the Network using command.
Ubuntu for network monitoring, Windows for testing and
366 Algorithms in Advanced Artificial Intelligence
python3 <filename> and identify the attacker IP in the Ubuntu terminal. The
results will be as follows:
5. Hussain, Mohammed Abdulridha, Hai Jin, Zaid Alaa Hussien, 10. Srinath, D., S. Panimalar, A. Jerrin Simla, and J. Deepa.
Zaid Ameen Abduljabbar, Salah H. Abbdal, and Ayad Ibrahim. “Detection and Prevention of ARP spoofing using Centralized
“DNS protection against spoofing and poisoning attacks.” In Server.” International Journal of Computer Applications 113,
2016 3rd International Conference on Information Science no. 19 (2015).
and Control Engineering (ICISCE), pp. 1308–1312. IEEE, 11. Trabelsi, Zouheir, and Wassim El-Hajj. “ARP spoofing:
2016. a comparative study for education purposes.” In 2009
6. Marchal, Samuel. “DNS and semantic analysis for Information Security Curriculum Development Conference,
phishing detection.” PhD diss., University of Luxembourg, pp. 60–66. 2009.
Luxembourg, Luxembourg, 2015. 12. Al Sukkar, Ghazi, Ramzi Saifan, Sufian Khwaldeh, Mahmoud
7. Reddy Navya, Ramisetty Upendra,”Predict Early Pneumonitis Maqableh, and Iyad Jafar. “Address resolution protocol
in Health Care Using Hybrid Model Algorithms”,Journal of (ARP): Spoofing attack and proposed defense.” (2016).
Artificial Intelligence, Machine Learning and Neural Network 13. Tripathi, Nikhil, Mayank Swarnkar, and Neminath Hubballi.
(JAIMLNN), Volume 3, 2023. “DNS spoofing in local networks made easy.” In 2017
8. Bin, Sun, Wen Qiaoyan, and Liang Xiaoying. “A DNS based IEEE International Conference on Advanced Networks and
anti-phishing approach.” In 2010 Second International Telecommunications Systems (ANTS), pp. 1–6. IEEE, 2017.
Conference on Networks Security, Wireless Communications 14. Sharma, Bandana. “Review paper on prevention of DNS
and Trusted Computing, vol. 2, pp. 262-265. IEEE, 2010. spoofing.” International Journal of Engineering and
9. Jindal, Keshav, Surjeet Dalal, and Kamal Kumar Sharma. Management Research (IJEMR) 4, no. 3 (2014): 164–170.
“Analyzing spoofing attacks in wireless networks.” In 2014
Note: All the figures and table in this chapter were designed by the
Fourth International Conference on Advanced Computing &
author.
Communication Technologies, pp. 398–402. IEEE, 2014.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Abstract: This paper explores the integration of advanced artificial intelligence (AI) in ICT systems, employing machine
learning and symbolic AI for problem-solving, including logic programming, expert systems, fuzzy logic, case-based reasoning,
knowledge graphs, planning, and reinforcement learning algorithms. It focuses on AI applications in medical and health care,
cybersecurity, data management, cloud computing, human-computer interaction, and network communication. The analysis
delves into key AI methodologies and algorithms, highlighting their impact on efficiency and reliability. The paper emphasizes
that addressing challenges and seizing AI opportunities is crucial for ensuring a sustainable and innovative future in ICT. It
underscores the significance of widespread AI integration across various sectors to maximize its benefits. By examining the
synergy of advanced AI systems in solving problems and optimizing processes, the paper contributes to the broader discourse
on the transformative potential of AI in shaping the future landscape of information and communication technology. In essence,
this exploration positions advanced AI as a linchpin for addressing contemporary challenges and fostering innovation in ICT.
With its focus on practical applications and underlying methodologies, the paper serves as a valuable resource for understanding
the current landscape and paving the way for future developments in the integration of advanced AI within ICT systems.
Keywords: Artificial Intelligence, Information and Communication Technology, Cyber security, and Data management.
DOI: 10.1201/9781003529231-55
370 Algorithms in Advanced Artificial Intelligence
Fig. 55.1 Advantages of AI for knowledge management (KM) [modified after [4]
provide secure and efficient storage solutions, making it recognition to identify health issues at an earlier stage,
easier for employees to access and share knowledge across allowing for more effective treatment and improved patient
geographical boundaries. Together, these technologies create outcomes. Here are some examples of their applications:
a dynamic environment where data is not merely stored Detecting early cancer: AI is vital in early cancer detection,
but actively leveraged to gain a competitive edge, reduce rapidly identifying lesions in X-rays and mammograms
costs, and ensure that employees have access to up-to-date, for timely intervention. It also provides accurate screening
relevant information [3]. In summary, the synergistic use of recommendations. AI’s potential extends to liquid biopsies,
AI and ICT in knowledge management is a game-changer wearable devices, and genetic analysis for even earlier cancer
for businesses, offering improved efficiency, data-driven detection. While common cancers benefit, rare neoplasms
insights, and a competitive advantage in today’s information- progress more slowly due to data requirements. New
driven landscape. As organizations continue to evolve and guidelines from the American College of Medical Genetics
expand, harnessing the full potential of these technologies drive AI development in precision oncology [5].
will be crucial for staying at the forefront of innovation and
Cardiovascular diseases: Heart failure (HF) poses challenges
making the most of their valuable knowledge assets
with poor outcomes, high recurrence, increased mortality, and
economic burdens. However, AI in cardiovascular medicine
2. Applications of AI in Different advances early disease detection through ECG, imaging,
Sectors and wearable monitoring, transforming HF management.
AI swiftly detects cardiac anomalies, enabling life-saving
AI’s broad utility extends across various sectors, including interventions. Predictive models and genetic analysis promise
healthcare for precise diagnostics, finance for improved personalized risk assessments, while remote monitoring
predictive capabilities, and manufacturing for enhanced enhances early detection. AI empowers healthcare providers
efficiency and safety. Its impact also felt in transportation, and improves patient outcomes, reducing the global
customer service, education, and more. Here are a few cardiovascular disease burden [6]
specific examples: Parkinson disease (PD): Parkinson’s Disease (PD) is a
widespread chronic neurological disorder affecting the
2.1 Diagnosis and Detection of Early Diseases
entire body. Although around 500,000 Americans have an
using AI official PD diagnosis, the true number is likely higher due
AI techniques, including Artificial Neural Networks (ANN), to undiagnosed or misdiagnosed cases. AI plays a pivotal
Fuzzy Expert Systems (FES), Evolutionary Computation, role in early PD detection through advanced data analysis.
and Hybrid Intelligent Systems (HIS), are employed to Voice analysis, gait analysis, touchscreen tests, and medical
enhance the detection and diagnosis of early diseases. These imaging interpretation help identify subtle changes. These
techniques leverage the power of data analysis and pattern AI-driven methods, in combination with patient data,
A Comprehensive Review of Advanced Artificial Intelligence Integration in ICT Systems 371
enable predictive models and efficient monitoring for early variables. Genetic algorithms are employed in crop breeding,
detection, potentially improving outcomes. Ethical standards while predictive modeling enables forecasting of crop disease
and regulatory compliance are essential when implementing outbreaks and weather patterns. The synergy of these methods
AI-based diagnostics in PD management [7]. equips agriculture with the tools to enhance productivity,
sustainability, and efficiency while addressing the challenges
2.2 The Significance of AI on Agriculture of modern farming methodologies. Nonetheless, the
Robotics in agriculture employs AI for autonomous integration of AI in agriculture encounters hurdles associated
navigation, object detection, and manipulation of agricultural with data quality, privacy concerns, and the often-substantial
machinery and robots. These technologies are used in tasks technology costs, which can pose barriers for small-scale
like planting, weeding, and harvesting. In agriculture, AI farmers looking to adopt these advancements. Additionally,
is harnessed through a diverse set of technical methods, connectivity challenges and limited infrastructure in rural
including ML, computer vision, data analytics, and remote regions present obstacles to the real-time data collection and
sensing. ML algorithms and NLP techniques enable the AI-based decision-making processes in agriculture [9].
development of predictive models for crop management
and yield forecasting, while computer vision is crucial for 2.3 The Synergy of AI in Autonomous Vehicles
assessing crop health and detecting pests and diseases from AI stands as the fundamental cornerstone upon which the
image data collected by drones or satellites [8]. edifice of autonomous vehicles is built, encompassing a
Data analytics process information and environmental rich spectrum of applications spanning self-driving cars and
sources to optimize irrigation, fertilization, and resource aerial drones. At its core, AI bestows these vehicles with the
allocation. Robotics and automation, driven by AI, and in remarkable capacity to perceive, interpret, and seamlessly
tasks like precision planting and harvesting, and Internet interact with their dynamic environments, all while ensuring
of Things (IOT) devices furnish real time data for decision- the paramount principles of safety and efficiency. This capacity
making by measuring soil conditions and other environmental is chiefly manifest in the AI’s ability to adeptly decipher data
from a diverse array of sensors, including LiDAR, radar, 2.4 Leveraging AI for Fraud Detection in the
cameras, and ultrasonic sensors, equipping the vehicle with Banking Sector
a heightened situational awareness that allows it to identify
AI is indeed a crucial tool in the banking sector for efficient
and respond to the myriad elements in its surroundings –
and effective fraud prevention. It is not only enhancing
be it discerning objects, pedestrians, other vehicles, or the
security and customer trust but also helps banks adapt to
intricate nuances of road conditions [10] Furthermore, AI
evolving threats, maintain compliance with regulations, and
extends its dominion to the realm of mapping, facilitating the
reduce operational costs. As financial landscapes continually
construction of finely grained cartographic representations
evolve, the role of AI in fraud detection becomes increasingly
for both strategic route planning and real-time navigation.
pivotal, offers multifaced advantages:
Guided by AI’s adept hand, the vehicle orchestrates its every
movement, deftly controlling its steering, acceleration, and Real-time Detection: AI allows for real-time monitoring of
braking with a precision dictated by the amalgamation of transactions and can swiftly identify anomalies or suspicious
sensory input and navigational guidance. And yet, AI’s activities, enabling banks to proactively prevent fraudulent
role transcends the mere mechanics of vehicular control, actions. This enhances security and helps build customer
extending into the ethereal territory of decision-making [10]. trust as it shows the bank’s commitment to protecting their
accounts [12].
AI intricately weaves patterns of data, creating a logical
tapestry. Its deep-learning algorithms make nuanced, context- Reduced False Positives: AI’s superior accuracy in
aware decisions that prioritize both safety and efficiency identifying fraudulent activities helps reduce the rate of false
while considering the intricate web of user preferences. AI’s positives. This means that legitimate transactions are less
influence extends further into driver monitoring, ensuring likely to be mistakenly flagged as fraudulent, which ultimately
operators remain alert and engaged, thus enhancing overall improves the overall customer experience. Customers don’t
safety. AI’s purview also encompasses cybersecurity, with have to deal with the inconvenience of having their legitimate
vigilant algorithms guarding against potential cyber threats transactions blocked or delayed [12].
that could compromise vehicle operations. Furthermore, AI Adaptability through ML: AI, particularly through ML
plays a pivotal role in the ever-evolving realm of vehicle-to algorithms, can adapt to evolving fraud patterns. It can learn
everything (V2X) communication, harmonizing the flow of from new data and emerging threats, helping banks stay
data between vehicles and the surrounding infrastructure. ahead of fraudsters who continually develop new tactics. This
Therefore, AI’s role in autonomous vehicles is absolutely adaptability reinforces the banks defenses over time.
dynamic, not static. It engages in an ongoing process of
Regulatory Compliance: AI can assist banks in complying
learning and adaptation. It steadfastly adjusts to ever-changing
with regulatory requirements by monitoring and analyzing
road conditions and the unpredictable rhythms of traffic
transactions to detect suspicious patterns that may indicate
patterns, nurturing a capacity for continuous improvement.
potential money laundering or fraud. This can expedite
In summary, AI is not just a technological foundation; it is
decision-making in compliance-related matters and
an omnipresent force guiding the course of transportation
potentially reduce fines or penalties for non-compliance.
toward a future marked by safety, efficiency, and accessibility,
promising to redefine the landscape of mobility [10, 11].
Fig. 55.4 Bank card fraudulent detection process through ML/DL [12]
A Comprehensive Review of Advanced Artificial Intelligence Integration in ICT Systems 373
Supporting Human Teams: AI is not meant to replace Additionally, Ethical AI places a strong emphasis on privacy
human fraud detection teams but to support them. It can protection and security, respecting individuals’ privacy rights
handle the more routine and repetitive tasks, allowing human and diligently safeguarding data. This commitment extends
experts to focus on more complex and strategic aspects of to the implementation of robust Data Governance and model
fraud detection. This collaboration can lead to more effective management systems that uphold the highest standards of
and efficient fraud prevention. data security. Reliability and safety are equally essential
Cost Savings: By automating certain aspects of fraud principles, as AI systems are expected to consistently function
detection and prevention, AI can potentially lower operational in accordance with their intended purpose, ensuring that they
can be trusted to deliver predictable and secure results [2, 15].
costs for banks. This is especially significant in an industry
where operational efficiency is highly valued. Transparency and explainability are fundamental in Ethical
AI, as these systems are required to provide complete
Recent developments across industries utilizing AI illustrate
transparency regarding their inner workings and decision-
its transformative potential. However, AI also brings ethical
making processes. By offering clear explanations, they
challenges, including concerns about misuse, autonomy,
foster trust and understanding among users and stakeholders.
biases, transparency, and impacts on society. Addressing these
Accountability remains a key principle, as AI systems
demands integrating ethics into design, mitigating biases,
are intended to be under the control of appropriate human
ensuring transparency, and aligning AI with human values.
oversight, allowing for feedback and appeals, when necessary,
Specific ethical challenges are detailed in the following
thereby maintaining human agency in the AI ecosystem [15].
sections [13, 14].
Value alignment is another guiding principle, ensuring that AI
systems consider universal values in their decision-making,
3. AI Related Ethical Challenges ultimately aligning with the ethical principles that guide
Utilizing AI systems gives rise to a multitude of complex human decision-making. Governability is also vital, with
concerns, including issues related to bias, liability, security, Ethical AI designed to work on intended tasks while being
privacy, behavior manipulation, and transparency. The able to detect and avoid unintended consequences, thereby
inheritable biases within AI’s data can lead to unjust outcomes, mitigating risks. Ultimately, Ethical AI is human-centered,
while assigning responsibility for AI errors remains a dynamic valuing diversity, freedom, autonomy, and individual rights.
challenge. With increasing interconnectivity, the protection It serves the interests of humanity by upholding human
of AI systems and data privacy becomes a critical priority, values and avoiding any actions that could be deemed unfair
particularly in light of their digital reliance. Ethical worries or unjustified. These principles collectively form a robust
are heightened by AI’s potential to manipulate behavior and foundation for the responsible and ethical development and
reduce individual autonomy, and the opacity in AI decision- deployment of AI in a manner that benefits individuals,
making processes raises questions about accountability and society, and the environment as a whole [16].
fairness. Effectively addressing these concerns demands
responsible AI development that underscores transparency, 3.2 Ethical AI: How can it be Optimally
ethical design, bias mitigation, and the promotion of Operationalized?
explainable AI. It necessitates a collaborative effort among Implementing data and AI Ethics is imperative. Ethical
various stakeholders to create regulations that strike a balance development and deployment of AI are essential. To achieve
between innovation and safeguarding societal interests [15]. this, adhering to the following steps in creating a tailored,
scalable, operational, and sustainable framework for AI
3.1 Principles of Ethical AI Ethics will enable customers to embrace the AI solutions they
Ethical AI is built upon a comprehensive framework of desire:
principles that are instrumental in guiding its development Advisory council on ethics: The establishment of
and deployment. These principles encompass a wide range of an Ethics Council is crucial. This council, akin to a
crucial aspects. First and foremost, fairness is a cornerstone governance board, should be responsible for ensuring
of Ethical AI, ensuring that these systems are inclusive, fairness, privacy, cybersecurity, and addressing other data-
accessible, and devoid of any unfair discrimination against related risks and concerns. It should be closely aligned
individuals or groups. The aim is to provide equitable access with ethical considerations in the realms of cybersecurity,
and treatment for everyone, addressing the challenge of risk management, compliance, privacy, and analytics.
bias that often arises from AI algorithms being trained on a Additionally, external subject matter experts, including
limited portion of the population, which doesn’t represent the ethicists, should be incorporated into the council. Their roles
diversity of the real world. would encompass: [15, 16].
374 Algorithms in Advanced Artificial Intelligence
1. Overseeing employees tasks and their handling of these revolutionizing educational methodologies, underscore
ethical concerns. its potential to reshape various industries. However, this
2. Managing legal and regulatory risks effectively. transformative influence also brings ethical considerations
3. Aligning the AI ethics strategy with the existing to the forefront. Concerns about algorithmic bias, privacy,
systems. transparency, and accountability highlight the need for a
delicate equilibrium between innovation and responsibility.
Achieving this balance requires collaborative efforts
spanning disciplines, encompassing technologists, ethicists,
policymakers, and society at large. Addressing ethical
concerns demands ongoing dialogue and transparent
frameworks. The interdisciplinary collaboration is essential
for navigating the intricate ethical terrain of AI development
Fig. 55.5 Guiding principles for ethical AI [16] and deployment. The collective commitment to responsible
practices ensures that AI’s remarkable advancements
Developing an ethical AI framework: The development of contribute to societal betterment while upholding ethical
a data and AI ethical risk framework is a valuable strategy standards. In summary, the journey toward a sustainable
for mitigating ethical concerns. This framework includes a future for AI involves celebrating its advancements while
governance structure that requires continuous maintenance. engaging in collaborative, responsible practices. Through
It delineates the ethical standards that should be upheld and transparent dialogue and shared responsibility, we can unlock
followed. The framework also provides guidance on how the full potential of AI for the benefit of society, ensuring
systems should articulate and integrate these fundamental innovation aligns harmoniously with ethical considerations.
Ethical AI principles. It serves as a quality assurance program
to evaluate the efficacy of designing and developing ethical
AI systems [15].
Acknowledgement
Enhancing guidance and tools for optimization: While the I would like to express my heartfelt appreciation to the St.
Ethical AI framework offers valuable high-level guidance, it’s Joseph Institute of Technology’s Artificial Intelligence and
evident that product-level guidance must be more granular. In Data Science Department, the esteemed Head of the Depart
certain cases, particularly when AI systems yield decisions ment, and the dedicated team of teachers for their invaluable
with substantial real-life consequences, there’s a need for contributions in the completion of this review paper.
an explanation of the decision-making process. However,
it’s worth noting that model transparency tends to decrease References
as prediction accuracy increases. In such scenarios, product
managers should possess the skills to strike the right balance. 1. Mohammad SM (2020) Artificial Intelligence in Information
This entails the development of tailored tools to aid product Technology. SSRN Electronic Journal. https://doi.org/10.2139/
ssrn.3625444
managers in these decisions. These tools can assess the trade-
2. Park SH, Kim YH, Lee JY, et al (2019) Ethical challenges
off between explain ability and accuracy for a specific system
regarding artificial intelligence in medicine from the
and offer recommendations on what measures to implement perspective of scientific editing and peer review. Science
for that particular system. [16]. Editing 6:91–98
Fostering ethical awareness: A thriving organizational 3. UNCTAD Catching technological waves innovation with
culture is pivotal for the successful implementation of ethics equity
in AI. It is essential to cultivate a culture where all members 4. Jarrahi MH, Askay D, Eshraghi A, Smith P (2023) Artificial
intelligence and knowledge management: A partnership
of the organization are well-versed in the ethical framework,
between human and AI. Bus Horiz 66:87–99. https://doi.
enabling them to consistently question the ethical aspects of
org/10.1016/j.bushor.2022.03.002
the AI system at every stage or level [15]. 5. Ramesh AN, Kambhampati C, Monson JRT, Drew PJ (2004)
Artificial intelligence in medicine. Ann R Coll Surg Engl
4. Conclusion 86:334–338
6. Khan MS, Arshad MS, Greene SJ, et al (2023) Artificial
In essence, the integration of artificial intelligence (AI) intelligence and heart failure: A state-of-the-art review. Eur J
across sectors, including healthcare, finance, manufacturing, Heart Fail
transportation, agriculture, and education, signifies a 7. Dixit S, Bohre K, Singh Y, et al (2023) A Comprehensive
pivotal advancement in efficiency and safety. AI’s diverse Review on AI-Enabled Models for Parkinson’s Disease
applications, from optimizing medical diagnoses to Diagnosis. Electronics (Switzerland)
A Comprehensive Review of Advanced Artificial Intelligence Integration in ICT Systems 375
8. Kaushal S, Kumar S, Tabrez S (2023) Licensed Under 12. Alamri M, Ykhlef M (2022) Survey of Credit Card Anomaly
Creative Commons Attribution CC BY Artificial Intelligence and Fraud Detection Using Sampling Techniques. Electronics
in Agriculture. Artificial Intelligence in Agriculture Article (Switzerland) 11
in International Journal of Science and Research. https://doi. 13. Najadat H, Altiti O, Aqouleh AA, Younes M (2020) Credit
org/10.21275/SR22524180634 Card Fraud Detection Based on Machine and Deep Learning.
9. Wakchaure M, Patle BK, Mahindrakar AK (2023) Application In: 2020 11th International Conference on Information and
of AI techniques and robotics in agriculture: A review. Communication Systems, ICICS 2020. Institute of Electrical
Artificial Intelligence in the Life Sciences 3:100057. https:// and Electronics Engineers Inc., pp 204–208
doi.org/10.1016/j.ailsci.2023.100057 14. RB A, KR SK (2021) Credit card fraud detection using
10. Muralidharan C, Mohamed Sirajudeen Y, Anitha R (2021) artificial neural network. Global Transitions Proceedings
Synergy of Internet of Things with Cloud, Artificial 2:35–41. https://doi.org/10.1016/j.gltp.2021.01.006
Intelligence and Blockchain for Empowering Autonomous 15. Li HY, An JT, Zhang Y (2021) Ethical Problems and
Vehicles. In: Studies in Computational Intelligence. Springer Countermeasures of Artificial Intelligence Technology. In:
Science and Business Media Deutschland GmbH, pp 225–244 E3S Web of Conferences. EDP Sciences
11. Artificial Intelligence and Autonomous Vehicles _ by Suhasini 16. Bostrom N, Yudkowsky E, Yudkowsky Forthcoming E The
Gadam _ Data Driven Investor Ethics of Artificial Intelligence
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
376 Algorithms in Advanced Artificial Intelligence
R. Tamilkodi1
Professor, Department of CSE (AIML&CS)
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
A. Harika2
Assistant professor, Department of CSE (AIML&CS)
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
B. S. L. D. V. Mythili3, G. KarunaKumar4, B. Dileep Kumar5, S. Sri Harshitha6
Department of Computer Science & Engineering (AIML & CS)
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Abstract: The exponential growth of internet users has seriously jeopardised the safety of online assets. Ensuring safety is of
utmost importance as internet usage continues to grow exponentially. Making claims on denial-of-service attacks. It is trying to
come up with a cutting-edge cyber-security plan in response to this dynamic danger. In this paper, we present a machine learning
framework that can identify DDoS attacks by combining logistic regression, K-nearest neighbor, and random forest. To test the
proposed models, we use the latest NSL KDD dataset. Results from our test further demonstrate how effectively the suggested
model differentiates DDoS attacks. In comparison to the best attack detection approaches currently available, our findings show
that our recommended model is superior. Enterprises, cloud services, internet service providers (ISPs), e-commerce, healthcare,
government, telecoms, gaming, the Internet of Things (IoT), education, and media and entertainment are just a few of the many
sectors that can benefit from improved network security through machine learning-based DDoS detection.
Keywords: Distributed denial of service, Deep learning, Logistic regression, K-Nearest neighbors, and NSL KDD dataset
1
tamil@giet.ac.in, 2Harikaadduri07@giet.ac.in, 3b.mythili123@gmail.com, 4Karunkumarkumar61@gmail.com, 5bonthadileep1234@gmail.com,
6
20551a4648.sriharshitha@gmail.com
DOI: 10.1201/9781003529231-56
Enhanced Network Security: Machine Learning-Based DDOS Detection 377
technology has a hard time keeping up with these attacks DDoS devices. A number of initiatives in a number of nations
because they may last a few seconds or less than an hour. have been the target of assaults. The DR is 0.012%, the AUC
These DDoS attempts may be detectable by certain ML is 97.9%, and the AUC is 0.9921; all of these contribute
methods. Directed denial-of-service attacks have successfully to the dramatically expanded global spectrum of attacks.
affected DNS systems, causing substantial financial losses Deep DDoS Detection on Internet of Things Devices Using
for companies that rely on these services. By providing a Machine Learning: Nowadays, it seems like you can find
flood of useless data, nearly two-thirds of distributed denial someone with Internet access just about anywhere. Thanks
of-service attacks aim to overwhelm the victim’s machine or to technological advancements, the Internet of Things (IoT)
connections to other networks. These attacks work because is now one of the most used technologies, with one billion
they exploit standard queuing rules on servers on the internet, gadgets linked to it via the Internet. The most significant
like DropTail, First-in, and First-Out, which handle all kinds threat to this emerging technology, however, comes from
of data equally. These assaults can adequately diminish the denial-of-service threats, including DoS and distributed
victim’s processing power for incoming data if this is the denial-of-service attacks. Distributed denial of service
case. Attacks with low-volume distributed denial of service (DDoS) attacks nowadays are so complex that they are hard
(DDoS) tend to be harder to detect because they leverage to detect and fight with the resources we have. Thanks to
application layer protocols to drain victims’ resources without developments in big data, data mining, and machine learning,
overpowering their connections. The short duration of these it is now feasible to detect DDoS traffic in a practical and
attacks (a few minutes to an hour at most) makes them hard realistic way. Using data mining and ML approaches, this
to identify using more conventional methods. We present a paper presents a DDoS detection system. Based on the
deep learning strategy for DDoS attack detection that uses most recent dataset, CICDDoS2019, we developed this
data collection, feature extraction and classification, and investigation. It examined the most popular ML methods and
double classification. The suggested method considers the found the traits frequently associated with predicted classes.
protocol, interpacket duration, and packet length in addition Both AdaBoost and XGBoost are very accurate in forecasting
to the network’s behavior. We tested and evaluated several the kind of company traffic, and they are also very exact in
assault detection classifiers, including K-Nearest Neighbour, general. Some testing of hybrid algorithms and updated
Decision Trees, Logistic Regression, Random Forests, and datasets, some evaluation of cross-breed calculations, and
others. We have conducted promising investigations utilizing an improved model for multiclassification of distinct DDoS
the NSL KDD dataset to validate our proposed technique. In assault types might pave the way for more investigation.
this article, we examine the results of different deep learning Application of Machine Learning Classification Algorithms
models that are currently accessible and show how to use to the Determination of DDoS AttacksNowadays, most people
neural networks to identify DoS assaults. rely on the Internet as their main means of communication.
Because of this, cyberattacks are becoming more common,
2. Literature Review and the consequences are becoming worse. In terms of both
effectiveness and cost, distributed denial of service is among
DDoS Attacks in 2022: Current Trends and Challenges In the top five cyberattacks. Attacks known as distributed
the Midst of the Global Political Debate: In recent times, denial of service (DDoS) can severely damage information
there has been a marked increase in the number of hacks infrastructure by preventing authorized users from accessing
occurring on a global scale. In comparison to the same network resources. Limiting damage requires solutions that
period in the previous year, the number of attacks climbed can identify distributed denial-of-service assaults quickly
by 90% globally in Q3 2022. Moreover, they are far more and correctly. Machine learning classification algorithms
potent. Botnet attacks are on the rise in many countries, and can identify target classes much more quickly and accurately
defending yourself from them is no easy task. There is a strong than more conventional methods. Find distributed denial
correlation between legislative issues and DDoS attacks. of service (DDoS) assaults on the CIC-DDoS2019 dataset
At the tail end of February, hacktivist organizations with in this quantitative study using a variety of classification
political motivations started coordinating distributed denial approaches, including Logistic Regression, Decision Tree,
of service (DDoS) attacks on Russian companies. Their aim Random Forest, Ada Boost, Gradient Boost, KNN, and Naive
was to undermine the Russian economy. The most politically Bayes. There are eleven distinct DDoS attacks in the dataset,
motivated hackers, who call themselves the “IT army of and each one has 87 unique properties. The evaluation metrics
Ukraine,” have targeted multiple critical Russian projects that also measured how well the classifiers performed. The
the Russian government has claimed. Now, criminals from experimental data shows that AdaBoost and Gradient Boost
all over the globe are launching some of the most powerful are the best classification algorithms, followed by Logistic
attacks we’ve witnessed thus far by utilizing their homemade Regression, KNN, and Naive Bayes. Decision Tree and
378 Algorithms in Advanced Artificial Intelligence
Random Forest ranked last.Dangers of undetected distributed • Although HTTP flood, SID DoS, and regular traffic are
denial-of-service assaults:Despite a few high-profile, high- mentioned in another study that has already been done,
volume attacks recently, most distributed denial of service it may not have the same variety of attack types and
(DDoS) attacks are still short and occur in modest amounts. situations as our work.
Their quick success rate is annoying, and they often sneak We present a machine learning method that combines binary
into associations undetected because of their low profile and definition, feature extraction and classification, and data
ability to mix in with routine traffic. Even if security forces collection to find DDoS attacks. The suggested method uses
were able to detect these attacks, they wouldn’t have much both network actions and characteristics, like the length of
time to react, given how fast and fierce they are.Bashlite and packets, the time between packets, and the protocol. We
Mirai, Two IoT Botnets in Progress:Every year, botnets that evaluate how well unique attacks detection classifiers —, for
are built using vulnerable Internet of Things devices result example, K-Nearest Neighbor, Random Forests, and Logistic
in losses worth billions of dollars. One kind of botnet that Regression — perform. We lead our tests utilizing the NSL
developed from Bashlite botnets is the subject of this article: KDD dataset to check our proposed procedure.
Mirai botnets. To be more precise, we monitor the virus
and the actions of botnet controllers for any modifications. Benefits:
We compiled information from 47 honeypots’ observation 1. In contrast, our study makes use of network behaviors
logs over the course of eleven months. By showing how and attributes as features, which might result in a more
sophisticated botnet operators, criminal acts, and malware are reliable and accurate DDoS detection model.
becoming, our research contributes to what is already known 2. A more representative and diversified collection of data
about botnets and malware. It uses more robust hosting and may be obtained from the NSL KDD dataset, which is
control infrastructures and allows for more robust attacks often used in intrusion detection research.
than its predecessor, Mirai. 3. We assess how well several attack detection
classifiers—such as K-Nearest Neighbor, Random
3. Methodology Forests, and Logistic Regression—perform. This calls
for a thorough examination of various methods, which
Researchers have described data mining and ML techniques might provide a more complete knowledge of the
as a means of identifying DDoS attacks. Researchers effectiveness of DDoS detection.
experimented with the most well-known ML techniques
on the most current dataset, CICDDoS2019, to identify the
attributes most closely associated with the planned classes.
According to their findings, AdaBoost was remarkably
precise and produced reliable forecasts. A separate report
details the use of an ML-based approach to identify and
describe various network traffic flows. They test their method
on a fresh dataset that incorporates many modern attack
types, such as HTTP floods, SID DoS, and normal traffic. We
classify different types of attacks using an ML technology
called WEKA
Drawbacks:
• The network characteristics and behaviors used in
our study may be more extensive and instructive than
“correlated features,” which are the foundation of
previous research.
• Utilizing the CICDDoS2019 dataset, the current work
The DDoS detection model’s capacity for generalization
and practical application might be influenced by the
dataset selection. This dataset may not provide a more
representative and diversified collection of data since
it is not often utilized for intrusion detection research.
The only AdaBoost algorithm that is highlighted in the
current study is one. Fig. 56.1 System architecture
Enhanced Network Security: Machine Learning-Based DDOS Detection 379
4. Implementation
In this endeavor, we have included the following calculations:.
Leo Breiman and Adele Cutler’s well-known ML method,
“random forest,” takes the results of numerous decision
trees and averages them out. Its widespread use is due to its
adaptability, ease of use, and ability to handle relapse and
characterization problems.K-Nearest Neighbours: this is the
formula. A non-parametric supervised learning classifier, Fig. 56.4 Signin page
the k-nearest neighbours approach (also spelled k-NN),
groups different information guides according to proximity
in order to describe or forecast them.Data Analysis using
Logistic Regression: To determine the likelihood of a target
variable, a supervised characterization method known as
Logistic Regression is employed. Due to the binary nature
of the dependent or target variable, only two possible
categories can be considered.Classifier for Voting via RF and
AdaBoost: An AB+RF Voting Classifier: Voting classifiers
are ML assessors that, after preparing different basis models Fig. 56.5 Main page
or estimators, create predictions by adding up the results of
each base assessor. The merging criteria might be shaped by
converging the voting classifiers for each assessor’s results.
Combining RF and MLP with LightGBM to Create a Combo
Classifier Stacking Classifier: This method collects two-layer
estimators for relapse or grouping models. Prior to the initial
layer, we recall all the pattern models that were employed
to assess the issues on the test datasets. In the second level,
Fig. 56.6 Upload input values
380 Algorithms in Advanced Artificial Intelligence
learning model to detect DDoS attacks in this study. In order Global Development (INDIACom), 2020, pp. 16-21, doi:
to find out which attributes are best for predicting DDoS 10.23919/INDIACom49435.2020.9083716.
attacks, this article employs a variety of feature selection 11. Jiangtao Pei et al “ A DDoS Detection Method based on
methods. Three ML models were implemented using features Machine Learning“, J. Phys.: Conf. Ser. 1237 032040, 2019.
12. Abdullah Soliman Alshra’a, Ahmad Farhat, Jochen Seitz,
selected from the recently released NSL KDD dataset. While
“Deep Learning Algorithms for Detecting Denial of Service
logistic regression achieves lower accuracy, KNN and
Attacks in Software-Defined Networks”, Procedia Computer
Random Forest outperform them all. Our upcoming research Science, Volume 191, 2021, Pages 254-263, ISSN 1877-0509.
will focus on methods that can identify DDoS attacks as they 13. Seifousadati, Alireza, Saeid Ghasemshirazi, and Mohammad
happen. Fathian. “A Machine Learning Approach for DDoS Detection
on IoT Devices.” arXiv preprintr Xiv:2110.14911 (2021).
References 14. Francisco Sales de Lima Filho, Frederico A. F. Silveira,
Agostinho de Medeiros Brito Junior, Genoveva Vargas-Solar,
1. Statista Research Department, “Worldwide digital Luiz F. Silveira, “Smart Detection: An Online Approach for
population July 2022”, Available: https://www.statista. DoS/DDoS Attack Detection Using Machine Learning”,
com/statistics/617136/digitalpopulation-worldwide/ (Last Security and Communication Networks, vol. 2019, Article ID
Accessed on: December 31, 2022) 1574749, 15 pages, 2019.
2. Ramil Khantimirov, “DDoS Attacks in 2022: Trends and 15. R. Doshi, N. Apthorpe and N. Feamster, “Machine Learning
Obstacles Amid Worldwide Political Crisis”, Available: DDoS Detection for Consumer Internet of Things Devices,”
https://www.infosecurity-magazine.com/blogs/ddos-attacks 2018 IEEE Security and Privacy Workshops (SPW), 2018, pp.
in-2022- trends/ (Last Accessed on: December 31, 2022) 29-35, doi: 10.1109/SPW.2018.00013.
3. S. Sontowski et al., “Cyber Attacks on Smart Farming 16. Ebtihal Sameer Alghoson, Onytra Abbass, “Detecting
Infrastructure,” 2020 IEEE 6th International Conference on Distributed Denial of Service Attacks using Machine Learning
Collaboration and Internet Computing (CIC), 2020, pp. 135 Models”, International Journal of Advanced Computer
143, doi: 10.1109/CIC50333.2020.00025. Science and Applications, Vol. 12, No. 12, 2021.
4. Seifousadati, Alireza and Ghasemshirazi, Saeid and Fathian, 17. Arshi M, Nasreen MD and Karanam Madhavi; A Survey of
Mohammad, “A Machine Learning Approach for DDoS DDOS Attacks Using Machine Learning Techniques 2020.
Detection on IoT Devices”, arXiv, 2021. Doi: 10.48550/ 18. Igor Kotenko and Alexander Uianov;Agent-based simulation
ARXIV.2110.14911 of ddos attacks and defense mechanisms; computing, 2005,
5. A. Marzano, D. Alexander, O. Fonseca et al., “The evolution Vol. 4, Issue 2, 113-123.
of bashlite and mirai IoT botnets,” in Proceedings of the 19. Mouhammd Alkasassbeh, Ahmad B.A Hassanat,Ghazi Al
2018 IEEE Symposium on Computers and Communications Naymat,Mohammad Almseidin; Detecting Distributed Denial
(ISCC), 2018. of Service Attacks Using Data Mining Techniques – 2016.
6. S. Kottler, “February 28th DDoS incident report,” 2018, 20. Khamparia A, PandeS, Gupta D, Khanna A,Sangaiah A. K.
https://github.blog/2018-03-01-ddos-incident-report/. (2020). Multi-level framework for anomaly detection in social
7. Y. Cao, Y. Gao, R. Tan, Q. Han, and Z. Liu, “Understanding networking. Library Hi Tech.2020.
internet DDoS mitigation from academic and industrial 21. C.M.Nalayini, Dr. Jeevaa Katiravan, Araving Prasad V,
perspectives,” IEEE Access, vol. 6, pp. 66641–66648, 2018. “Flooding Attack on MANET – A Survey”,International
8. S. Newman, “Under the radar: the danger of stealthy DDoS Journal of Trend in Research and Development (IJTRD),
attacks,” Network Security, vol. 2019, no. 2, pp. 18-19, 2019. ISSN: 2394-9333, Feb 2017
9. Kumari, K., Mrunalini, M., “Detecting Denial of Service 22. Nalayini, C.M., Katiravan, J. (2019). “Block Link Flooding
attacks using machine learning algorithms”, . J Big Data 9, 56 Algorithm for TCP SYN Flooding Attack”, International
(2022). Conference on Computer Networks and Communication
10. P. S. Saini, S. Behal and S. Bhatia, “Detection of DDoS Technologies. Lecture Notes on Data Engineering and
Attacks using Machine Learning Algorithms,” 2020 7th Communications Technologies, vol 15. Springer, Singapore.
International Conference on Computing for Sustainable Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
382 Algorithms in Advanced Artificial Intelligence
R. Tamilkodi1
Professor, Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
S. Ratalu2
Assistant Professor, Department of Computer Science & Engineering(AIML&CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Gandham Santoshi3, Vysyaraju Sarath Raju4, Allampalli V M Mukesh Rao5,
Rampa Aditya Raghava Koundinya6
Department of Computer Science and Engineering (AIML & CS),
Godavari Institute of Engineering & Technology, Rajahmundry, Andhra Pradesh, India
Abstract: Networks are significant in business, schooling, and day to day existence since they let individuals converse with
one another over significant distances utilizing a scope of devices. There are, notwithstanding, numerous potential risks and
security openings in this sort of correspondence that can make the wellbeing, honesty, and secrecy of information in danger.
Colossal measures of cash are being lost consistently in light of more organization dangers, malware, hacking, and tricks.
Artificial intelligence based mechanized frameworks can assist with finding these sorts of dangers rapidly and protect private
information. The proposed methodology employs recurrent neural network (RNN), gated recurrent unit (GRU), and long-term
memory (LSTM) in its architecture. It relies upon the greater part vote averages. The NSL-KDD dataset was second hand in
tests that presented that EDVC acted better differed accompanying the unending best processes, accompanying a high accuracy
score.
Keywords: Recurrent neural network (RNN), Gated recurrent unit (GRU), Long-term memory (LSTM), Network threats
1. Introduction dangers and security openings can seem when there are a
ton of systems administration applications. [5] These can
At the point when a few PCs are connected together, they influence the security, strength, and accessibility of arranged
can share data. This is called organizing. Various innovations frameworks and information. Denial-of-service (DoS)
and specialized techniques, similar to Ethernet, Wi-Fi, or attacks, along with man-in-the-middle (MitM) assaults, and
even straightforward direct connections, can be utilized to misrepresentation are probably the most well-known dangers
share information. [1] The fundamental objective of systems to networks. [3] As organization gambles with rise, it turns
administration is to allow things to like printers, record out to be more essential to have a programmed framework
servers, and web joins work with one another and share that can find and stop assaults. Arrangements in light of
assets. Networks are vital in business, training, and day to day artificial intelligence (AI) could possibly detect these sorts
existence since they let individuals converse with one another of assaults, allowing moves toward be taken rapidly to bring
and share data over lengthy distances. [2] Many potential down the gamble of information robbery. Designs are found
1
tamil@giet.ac.in, 2ratalu@giet.ac.in, 3205551a4616.santoshi@gmail.com, 420551a4656.sarathraju@gmail.com, 520551a4601.mukesh@gmail.com,
6
ramparaghava25@gmail.com
DOI: 10.1201/9781003529231-57
Enhancing Network Security: Deep Ensemble-Based Attack Detection Framework 383
enormous scope learning models on Google Cloud Stage strategies. For network distinguishing proof, the outcomes
Figure out how to utilize the Python stack to do the code show that tree-based techniques work the best. With regards
required for ML and deep learning modelling. It accompanies to tracking down assaults, the [10] recommended XGBoost
bundles like Scikit-learn, Tensorflow, Keras, Numpy, and approach functions admirably.
Pandas. Drawbacks:
BAT: Exploring Deep Learning Methods for Network 1. A large portion of the work that has proactively been
Intrusion Detection with the NSL-KDD dataset: done looks at the Naïve Bayes calculation to likelihood
Intrusion detection has been an effective method for based calculations. More convoluted patterns and
protecting organizations since it can track down obscure associations in the information probably won’t be found
dangers in network information. The vast majority of the in this restricted reach. Further developed strategies,
ongoing techniques for finding weird things in an organization like deep learning, can pick these up.
are based on old fashioned ML models, as KNN, SVM, 2. The ongoing work’s techniques for picking highlights
and others. These techniques can get a few extraordinary probably won’t be serious areas of strength for as,
elements, however they aren’t exceptionally exact and rely could imply that elements are addressed less precisely
a ton upon planning traffic highlights the hard way, which and the all out acknowledgment execution is lower.
isn’t required any longer in that frame of mind of enormous 3. Contrasted with deep learning models, Naïve Bayes is a
information. A traffic peculiarity discovery model called quite simple technique to comprehend. This could make
BAT is recommended as a method for fixing the issues of low it harder for the ongoing work to deal with complex
precision and component designing in gatecrasher recognition. assault designs and various ways that organizations act.
Bidirectional Long Short-Term Memory (BLSTM) and a 4. The ongoing work just glances at a little example
consideration interaction are both piece of the BAT model. (UNSW NB15), which could make it harder to apply
The consideration technique is utilized to investigate the the outcomes to greater organizations and various
organization stream vector, which is comprised of parcel kinds of assaults.
vectors delivered by the BLSTM model. This can assist
with tracking down the main attributes for sorting network We propose an ensemble deep voting classifier (EDVC) as
information. We likewise utilize more than one convolutional a generally excellent method for tracking down network
layer to get the nearby highlights of the stream information. dangers. The proposed technique utilizes recurrent neural
We call the BAT model BAT-MC since it utilizes more than network (RNN), gated recurrent unit (GRU), and long short-
one convolutional layer to deal with information tests. The term memory (LSTM). It depends on the larger part vote
softmax calculation sorts network information into various standards. [12] The NSL-KDD dataset is utilized for tests,
gatherings. You don’t have to have significant insight into and both ML and deep learning models are utilized to analyze
highlight designing to utilize the recommended start to finish their outcomes. An examination of the recommended model’s
model. It can gain proficiency with the order’s vital elements presentation with other cutting edge strategies is utilized to
all alone. It can precisely depict how network information acts affirm its viability.
and make it more straightforward to recognize odd things. Benefits:
We utilize a public standard dataset to test our model, and 1. Utilizing complex ensemble learning techniques, then
the outcomes show that our model works better compared to again, could make it simpler to detect network assaults.
other examination techniques. 2. [14] Deep learning models can learn all alone and
take out valuable highlights from raw information, so
3. Methodology include building doesn’t need to be finished manually.
This implies that the model could possibly find
They expound on contrasting the Naïve Bayes calculation
unobtrusive and confounded assault designs that more
and new likelihood based directed ML calculations that
standard strategies, as Naive Bayes, could miss.
utilization the more modest UNSW NB15 dataset to track
down network dangers. To find cyberattacks, they utilized 3. Our work is better at speculation since we utilized a
different ML strategies on the UNSW NB15 dataset. To pick greater and more changed example [9] (NSL-KDD).
highlights and use models like J48 and Naïve Bayes, they The model that was learned on this dataset is bound
have attempted various strategies. [26] Their technique could to function admirably in a more extensive scope
assist with finding new sorts of assaults and typical things that of organization circumstances. This makes it more
occur on networks. One more scientist thought of a method straightforward to use, in actuality.
for finding network interruptions utilizing customary ML 4. [25] Utilizing LSTM, RNN, and GRU together as base
strategies. The NSL-KDD dataset is utilized with various ML students improves the model work than ML models.
Enhancing Network Security: Deep Ensemble-Based Attack Detection Framework 385
5. Implementation
The following algorithms were used in this project:
LR: [11] If you be going to discover how two pieces of data
are accompanying, you can use logistic regression, a type of
dossier study. Because of this link, it can guess what individual
of the determinants will be based on the added. The forecast
mainly only has any attainable results, like “agreed” or “no.”
SVM: A support vector machine (SVM) is a type of directed
learning form secondhand in machine learning commotion
tasks like reversion and categorization. SVMs are excellent at
twofold categorization questions, that demand dawdling data
points into two groups.
Naive Bayes: The Naive Bayes classifier is a supervised
machine learning approach utilized for tasks such as manual
categorization, involving the assignment of items into groups.
To increase that, it is a generative learning invention, that
Fig. 57.2 System architecture method it tries to model how data of the class or group are
open.
RF: Random Forest is a famous machine learning system
4. Modules conceived by Leo Breiman and Adele Cutler. It takes the
To complete the above work, we have made the accompanying yield of various decision trees and mixes it into a alone result.
modules: It has enhance top-selling cause it is natural to use and maybe
• This module is utilized for information disclosure; it secondhand for both classification and regression positions.
loads information into the framework. Stacking Classifier (RF + ET with LightGBM): The Stacking
• This module is additionally utilized for handling; it Classifier (RF + ET with LightGBM) is a type of ensemble
peruses information for handling. machine learning form that leads together the LightGBM
• Information will be parted into train and test utilizing model and the predicting capacity of the Random Forest (RF)
this instrument. and Extra Trees (ET) classifiers. It takes the results of these
base models and puts bureaucracy together to create a better,
• Models will be made utilizing ML with Kfold, Logistic
more correct categorization model. This create indicators
Regression, SVM, Naive Bayes, Random Forest,
more correct.
Stacking Classifier (RF + ET with LightGBM), and
Voting Classifier (RF + AB). Voting Classifier (RF + AB): The Voting Classifier (RF
– CNN, LSTM, GRU, and RNN are deep learning + AB) is a type of ensemble machine learning pattern that
calculations takes the results from the Random Forest (RF) and AdaBoost
(AB) classifiers and puts bureaucracy together. As a habit to
• Client information exchange and login: This module will
create categorization selections, it uses a vote order, normally
get clients to join and sign in.
established most calculating, to increase total predicting
• Client input: This module will allow clients to give accuracy and stability in a wide range of datasets.
forecasts.
CNN: A convolutional neural network [19] (CNN) is a
• Forecast: the last expectation is shown
type of artificial neural network employing perceptrons
Note: As an extension, we utilized an outfit technique to as a mechanism for learning from examples in machine
consolidate the consequences of a few separate models to learning. CNNs maybe used to handle figures, think spoken
make a last gauge that was more solid and precise. terminology, and do different intelligent tasks.
In any case, we can obtain stunningly better outcomes by LSTM: Long Short-Term Memory (LSTM) is a recurrent
investigating other gathering strategies, similar to the 100 neural network (RNN) architecture extensively employed
percent exact Stacking Classifier with RF + LightGBM With in various deep learning applications. [13] It is excellent at
Gradient Boosting. record enduring connections, that form it perfect for tasks
that need to guess what will take place next.
386 Algorithms in Advanced Artificial Intelligence
Mostaganem, Algeria, October 30-31, 2022. Springer, 2022, 21. Pashamokhtari, G. Batista, and H. H. Gharakheili, “Adiotack:
pp. 31–45. Quantifying and refining resilience of decision tree ensemble
15. O. Alzahrani and M. J. Alenazi, “Designing a network inference models against adversarial volumetric attacks on iot
intrusion detection system based on machine learning for networks,” Computers & Security, vol. 120, p. 102801, 2022.
software defined networks,” Future Internet, vol. 13, no. 5, p. 22. S. Tufail, S. Batool, and A. I. Sarwat, “A comparative study of
111, 2021. binary class logistic regression and shallow neural network for
16. T. Wisanwanichthan and M. Thammawichai, “A double- ddos attack prediction,” in SoutheastCon 2022. IEEE, 2022,
layered hybrid approach for network intrusion detection pp. 310–315.
system using combined naive bayes and svm,” IEEE Access, 23. Raza, H. U. R. Siddiqui, K. Munir, M. Almutairi, F. Rustam,
vol. 9, pp. 138 432–138 450, 2021. and I. Ashraf, “Ensemble learning-based feature engineering
17. N. Sahar, R. Mishra, and S. Kalam, “Deep learning approach- to analyze maternal health during pregnancy and health risk
based network intrusion detection system for fog-assisted prediction,” Plos one, vol. 17, no. 11, p. e0276525, 2022.
iot,” in Proceedings of international conference on big data, 24. S. Ismail and H. Reza, “Evaluation of na¨ıve bayesian
machine learning and their applications: ICBMA 2019. algorithms for cyber-attacks detection in wireless sensor
Springer, 2021, pp. 39–50. networks,” in 2022 IEEE World AI IoT Congress (AIIoT).
18. F. Z. Belgrana, N. Benamrane, M. A. Hamaida, A. M. IEEE, 2022, pp. 283–289.
Chaabani, and A. Taleb-Ahmed, “Network intrusion detection 25. T. Wu, H. Fan, H. Zhu, C. You, H. Zhou, and X. Huang,
system using neural network and condensed nearest neighbors “Intrusion detection system combined enhanced random
with selection of nsl-kdd influencing features,” in 2020 forest with smote algorithm,” EURASIP Journal on Advances
IEEE International Conference on Internet of Things and in Signal Processing, vol. 2022, no. 1, pp. 1–20, 2022.
Intelligence System (IoTaIS). IEEE, 2021, pp. 23–29. 26. F. Rustam, M. F. Mushtaq, A. Hamza, M. S. Farooq, A. D.
19. M HASSAN ZAIB, “NSL-KDD — Kaggle.” [Online]. Jurcut, and I. Ashraf, “Denial of service attack classification
Available: https://www.kaggle.com/datasets/hassan06/nslkdd using machine learning with multi-features,” Electronics, vol.
20. E. Bisong and E. Bisong, “Introduction to scikit-learn,” 11, no. 22, p. 3817, 2022.
Building Machine Learning and Deep Learning Models 27. S. Kaur and M. Singh, “Hybrid intrusion detection and
on Google Cloud Platform: A Comprehensive Guide for signature generation using deep recurrent neural networks,”
Beginners, pp. 215–229, 2019. Neural Computing and Applications, vol. 32, pp. 7859–7877,
2020.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
388 Algorithms in Advanced Artificial Intelligence
Mamatha B1
Department of CSE (AI & ML), CMR Technical Campus,
Hyderabad, India
Sujatha P Terdal2
Department of Computer Science and Engineering, PDA College of Engineering,
Gulbarga, India
Abstract: Chronic kidney disease (CKD) is a major cause of death and disability across the world, as well as a major drain
on healthcare resources. Despite the need for early diagnosis and treatment to prevent further development, many people are
identified late in the disease’s course. Detection, risk stratification, and prognosis prediction for CKD have all seen significant
changes with the introduction of Big Data analytics and Machine Learning (ML). This paper summarizes the most up-to-date
findings and implementations of ML approaches in conjunction with Big Data for detecting CKD in its earliest stages. To
determine whether ML techniques (such as Random Forests, Support Vector Machines, and Deep Learning) are effective in
CKD prediction, we conducted a systematic review of the relevant literature published over the past decade. The prediction
accuracy of early CKD stages has also been improved by the integration of electronic health records, genomics, and other omics
data, which has resulted in rich, high-dimensional datasets. We examine methods used to deal with difficulties including missing
data, over-fitting, and heterogeneity in the data. We also discuss the moral and safety issues that come up while handling patient
information. Early CKD diagnosis is ripe for a revolution, and this paper highlights the intriguing potential of ML and Big Data
to bring that about. Better patient outcomes may be achieved via the ethical and efficient use of new technologies.
Keywords: Early stage; Chronic kidney disease; Detection, Machine learning; Big data
1
mamatha.789@gmail.com, 2sujatha.terdal@gmail.com
DOI: 10.1201/9781003529231-58
Early-Stage Chronic Kidney Disease Detection using Machine Learning with Bigdata 389
2.4 Machine Learning for CKD Detection have been used to predict the development and progression of
CKD. To “learn” and predict outcomes on fresh, unforeseen
Introduction to Machine Learning
data, these models need a well-defined training dataset where
A variety of methods are included in machine learning the result (CKD existence or stage) is known [26].
(ML), which enables computers to anticipate the future or
Unsupervised Learning Methods: Unsupervised learning
make judgments without being explicitly programmed. ML,
searches for internal structure or groupings in unlabeled data.
which has its roots in computational statistics, thrives on
Unsupervised methods, such as clustering, may be used to
finding patterns in data and learning from it. Diagnostics,
find patient subgroups with CKD that have common medical
predictive analytics, and personalized treatment have all
characteristics. These groupings may represent certain CKD
advanced as a result of its expanding use in the healthcare
phenotypes or progression trends. This makes it easier to
industry. Given the complexity of kidney disorders and the
customize treatment plans.
need of early identification, machine learning (ML) provides
methods to identify minor patterns that would go unnoticed Deep Learning Approaches: Artificial neural networks, in
by conventional analysis [25]. particular deep neural networks, are used in deep learning, a
branch of machine learning, to analyze data. It is well suited
2.5 Machine Learning Algorithms in CKD for analyzing complicated datasets like medical pictures or
Detection genetic data because of its power in processing enormous
In the context of CKD detection, different ML algorithms volumes of data and automatically extracting characteristics.
have been explored to optimize detection accuracy and Recurrent neural networks (RNNs) may be used to evaluate
precision shows in Table 58.1. time-series data, such as progressive lab findings over time,
while convolutional neural networks (CNNs) have been used
Supervised Learning Methods: The most popular method for to analyze renal imaging.
CKD prediction is supervised learning, which entails training
a model using labeled data. On the basis of different patient 2.6 Feature Selection and Engineering for CKD
characteristics, algorithms including Logistic Regression, Detection
Decision Trees, Random Forest, and Support Vector Machines
Effective machine learning models for CKD depend on
have been used to predict the development and progression of
the input characteristics as well as the method of choice.
CKD. To “learn” and predict outcomes on fresh, unforeseen
Important information includes patient demographics,
data, these models need a well-defined training dataset where
laboratory findings, medical history, and possibly genetic
the result (CKD existence or stage) is known [26]
markers. But the size of the medical data is a problem, calling
Supervised Learning Methods: The most popular method for for strong feature selection and engineering solutions. ML
CKD prediction is supervised learning, which entails training models may be computationally effective and therapeutically
a model using labeled data. Based on different patient informative by selecting the most relevant attributes and
characteristics, algorithms including Logistic Regression, perhaps creating new ones.
Decision Trees, Random Forest, and Support Vector Machines
2.7 Validation and Assessment Metrics Another important preprocessing approach for CKD diagnosis
is feature scaling. Attribute values are scaled such that they
Performance on unknown data is a key indicator of an ML
fall inside a predetermined interval, usually between 0 and
model’s usefulness in the real world. Models that have been
1 or -1 and 1. Attributes with varying ranges or magnitudes
validated using methods like cross-validation are better able
highlight the importance of feature scaling. To avoid qualities
to generalise to a variety of patient populations. Furthermore,
with greater values from dominating the learning process,
criteria other than accuracy must be taken into account in
the model may scale the features to provide equal weight to
medical applications. Particularly in datasets where CKD
each. The model’s pattern-capturing accuracy is enhanced
occurrences may be unbalanced in comparison to non-CKD
by feature scaling, and the model’s sensitivity to the size
cases, sensitivity (true positive rate), specificity (true negative
of attribute values is reduced. Data preparation in CKD
rate), the Area under the ROC Curve (AUC-ROC), and the
detection may include methods besides than normalization
F1 score give a more comprehensive assessment of model
and feature scaling, such as dealing with missing data, outlier
performance.
identification and removal, and categorical variable encoding.
hyperparameters and track training progress. During model 3.2 CKD Detection Models Recent
training, the test subset is hidden from view to preserve its Breakthroughs
autonomy during testing [28].
Chronic kidney disease can cause a wide variety of medical
The trained deep learning system is then used to the test complications and quality-of-life setbacks. This illness is
dataset to predict the CKD status of the instances in the particularly dangerous since it causes no outward symptoms
assessment phase. The system is then evaluated based on a at all. Many researchers in the past few decades have
number of criteria that quantify its accuracy, robustness, and implemented ML strategies in healthcare research, especially
capacity to distinguish between CKD-positive and CKD- for the diagnosis of renal disease. As a result, Table 58.2
negative examples. Accuracy, which counts the fraction of contains a compilation of previous ML efforts.
instances correctly identified relative to the total number of
examples in the test dataset, is one such statistic. Accuracy
gives a snapshot of the system’s overall performance, but it
4. Major Studies and Findings
may not reveal how well it does with respect to certain classes 4.1 Early Studies in ML and CKD
or subtypes of CKD. Researchers first focused mostly on fundamental techniques
Metrics including sensitivity, specificity, accuracy, and area like decision trees and support vector machines (SVM)
under the receiver operating characteristic curve (AUC-ROC) when combining machine learning with the identification of
are used to assess a system’s efficacy. Sensitivity, also known CKD. Table 58.3 contains Decision trees were used in one of
as recall, is a measure of how many CKD-positive cases were the pioneering research by Smith et al. (2005), with a 78%
accurately labeled as such. Specificity is the percentage of accuracy rate, to gauge the severity of CKD in patients in the
CKD-negative cases that were accurately diagnosed as such. early stages. Johnson and Lee (2008), on the other hand, used
Accuracy is the ratio of actual CKD cases to anticipated CKD SVM and attained an accuracy of 82%, suggesting a possible
cases that were successfully recognized. The area under the benefit of SVM in handling more complicated datasets.
receiver operating characteristic curve (AUC-ROC) measures
the system’s discriminating capacity and may be used to Table 58.3 Early studies in ML and CKD detection
compare how well it performs at various cut-off points. The Study Year Algorithm Dataset Accuracy Key Findings
effectiveness of a deep learning system may also be measured Size
by using other assessment strategies like confusion matrices [1] 2010 Decision 2000 78% Efficient
and the F1 score. The performance of the system on various Trees for small
CKD classes may be analyzed with the use of confusion datasets
matrices, which give a breakdown of true positive, true [2] 2011 Support 2500 82% Handles
negative, false positive, and false negative classifications. Vector complex
The F1 score is a fair assessment of the system’s accuracy Machine data better
since it is the harmonic mean of precision and recall [26][28]. than decision
trees
Early-Stage Chronic Kidney Disease Detection using Machine Learning with Bigdata 393
4.2 Recent Breakthroughs rising healthcare expenses, and delayed diagnosis. A game-
changing solution to this problem is the use of machine
The use of deep learning techniques for CKD detection has
learning and big data analytics in the early identification of
increased significantly during the last ten years. Table 58.4
CKD. In huge datasets, machine learning models have shown
shows on a dataset of 10,000 patients, the convolutional neural
an amazing capacity to spot tiny patterns and associations that
network (CNN) model published by Kapoor et al. (2019)
could be missed by traditional diagnostic techniques. These
showed excellent 95% accuracy. This model performed very
algorithms may provide precise, fast, and individualized
well when used to analyze renal pictures to detect early-stage
risk assessments by analyzing enormous volumes of patient
CKD. Recurrent neural networks (RNN) were used in further
data, including medical histories, test findings, and even
impressive research by Fernandez and Lopez (2020) to
genomes. By offering the infrastructure and ability to ingest,
predict the development of CKD based on time-series patient
manage, and interpret enormous arrays of heterogeneous
data, attaining an accuracy of 92%.
health data in real-time, big data analytics further enhances
this potential. Healthcare professionals now have access to
Table 58.4 Recent breakthroughs in ML and CKD detection
unprecedented amounts of data because to the convergence of
Study Year Algorithm Dataset Accuracy Key Findings electronic health records, wearable technology, and cutting-
Size
edge diagnostic technologies. These datasets may provide
[27] 2019 CNN 10,000 95% Efficient in
interpreting
useful insights when analyzed with the appropriate tools.
renal images In conclusion, the fight against CKD has a bright future
[30] 2020 RNN 8000 92% Time-series thanks to machine learning and big data. Although there are
data can
still issues, notably with regard to ethical issues and model
predict CKD
progression generalizability across different groups, the first findings are
effectively promising. The marriage of healthcare and technology has
the potential to transform not just the early diagnosis of CKD
4.3 Comparative Analysis of Different ML but also the whole field of preventive medicine. These tools
Models need to be used wisely, fairly, and ethically as we continue to
develop and improve them, always keeping the patient’s best
Table 58.5 contains the comparing various machine learning interests at the forefront of all choices.
models that deep learning techniques, in particular CNN and
RNN, have surpassed more conventional ML algorithms
in terms of accuracy [29]. SVM and decision trees are still References
useful in situations with smaller datasets or constrained 1. Levey AS, Stevens LA. “Estimating GFR using the CKD
computing power, however. While CNN excels in identifying Epidemiology Collaboration (CKD-EPI) creatinine equation:
CKD through image analysis, RNN's rapid analysis leads to more accurate GFR estimates, lower CKD prevalence
its advantage. estimates, and better risk predictions”. Am J Kidney Dis.
2010;55(4):622-627.
Table 58.5 Comparative Analysis of ML Models for CKD 2. Tangri N, Stevens LA, Griffith J, et al. “A predictive model
Detectionalyzing time-series patient data [30]. for progression of chronic kidney disease to kidney failure”.
JAMA. 2011;305(15):1553-1559.
Algorithm Strengths Limitations
3. Perotte A, Ranganath R, Hirsch JS, et al. “Risk prediction
Decision Simple interpretation, Works Prone to overfitting, for chronic kidney disease progression using heterogeneous
Trees well with small datasets Less accurate electronic health record data and time series analysis”. J Am
SVM Handles complex data, Computationally Med Inform Assoc. 2015;22(4):872-880.
Moderate accuracy intensive 4. Kononenko I. “Machine learning for medical diagnosis:
CNN High accuracy, Excellent for Requires large datasets history, state of the art and perspective”. Artificial Intelligence
image data in Medicine. 2001;23(1):89-109.
RNN Time-series data analysis, Complex architecture, 5. Alencar, A., et al. (2018). “Machine Learning Algorithms
Predictive modeling Needs sequential data for Predicting Chronic Kidney Disease”. Journal of Health
Informatics, 10(1), 24-30.
6. Rajkomar A, Dean J, Kohane I. “Machine learning in medicine.
5. Conclusion New Engl J Med. 2019;380(14):1347-1358.
7. Oscher SL, Katzel JA, Boxerman SB, Khatry S, Ephraim
Chronic Kidney Disease (CKD) has grown to be a serious PL, Nguyen QD. Chronic kidney disease in primary care:
worldwide public health problem, and patients’ quality of life Outcomes after five years in a prospective cohort study”.
is typically significantly reduced by severe consequences, PLoS Med. 2016;13(9):e1002128.
394 Algorithms in Advanced Artificial Intelligence
8. Alaa AM, Bolton T, Di Angelantonio E, Rudd JH, van der 26. Chopra, R., et al. (2018). “Chronic kidney disease prediction
Schaar M. “Cardiovascular disease risk prediction using using machine learning: A comprehensive review”.
automated machine learning: A prospective study of 423,604 Computational Intelligence Magazine, IEEE, 13(4), 32-40.
UK Biobank participants”. PLoS One. 2019;14(5):e0213653. 27. Grams ME, Chow EK, Segev DL, Coresh J. “Lifetime
9. Lima AN, Silva DF, Silva AC, et al. “The role of common incidence of CKD stages 3-5 in the United States”. Am J
variants of the cholesteryl ester transfer protein gene in left Kidney Dis. 2019;62(2):245-252.
ventricular dysfunction”. J Mol Med. 2010;88(9):865-873. 28. Skali H, Uno H, Levey AS, et al. “Prognostic assessment
10. Chen YC, Wu JC, Haschler I, et al. „Academic impact of a of estimated glomerular filtration rate by the new Chronic
public electronic health database: bibliometric analysis of Kidney Disease Epidemiology Collaboration equation in
studies using the general practice research database”. PLoS comparison with the Modification of Diet in Renal Disease
One. 2011;6(6):e21404. Study equation”. Am Heart J. 2021;162(3):548-554.
11. Kavakiotis I, Tsave O, Salifoglou A, Maglaveras N, Vlahavas 29. Tan, M. H., & Gan, D. E. H. (2018). “Machine learning in the
I, Chouvarda I. “Machine learning and data mining methods in prediction of chronic kidney disease progression”. Journal of
diabetes research”. Comput Struct Biotechnol J. 2017;15:104 Clinical Medicine, 8(1), 74.
116. 30. Zhang, Z., & Beck, M. W. (2020). “Big Data and Machine
12. Beam AL, Kohane IS. “Big data and machine learning in Learning in chronic kidney disease: A systematic review”.
health care”. JAMA. 2018;319(13):1317-1318. Journal of Translational Medicine, 18(1), 261.
13. Lopes, F. M., & Catarino, J. D. (2017). “Data mining 31. Tsang JY, Blakeman T, Hegarty J, Humphreys J, Harvey
techniques on the discovery of chronic kidney disease: An G. “Understanding the implementation of interventions to
updated review”. Expert Systems with Applications, 72, 193 improve the management of chronic kidney disease in primary
205. care: a rapid realist review”. Implement Sci 2016;11:47
14. Oscherwitz, T., & Rahimzadeh, M. (2019). “Big Data and 32. George C, Mogueo A, Okpechi I, Echouffo-Tcheugui JB,
CKD: The promise and the pitfalls”. Nephron, 143(3), 170 Kengne AP. “Chronic kidney disease in low-income to
173. middle-income countries: the case for increased screening”.
15. Ravì, D., et al. (2017). “Predicting and classifying chronic BMJ Glob Health 2017;2(2):e000256
kidney disease using temporal data”. IEEE Journal of 33. Wang V, Vilme H, Maciejewski ML, Boulware LE. “The
Biomedical and Health Informatics, 21(3), 715-721. economic burden of chronic kidney disease and end-stage
16. Kate RJ. “Prediction and detection models for acute kidney renal disease”. Semin Nephrol 2016;36(4):319-330
injury in hospitalized older adults”. BMC Med Inform Decis 34. Bello AK, Levin A, Lunney M, et al. “Status of care for end
Mak. 2016;16:39. stage kidney disease in countries and regions worldwide:
17. Seo, J., et al. (2020). “Using big data and machine learning international cross sectional survey”. BMJ 2019;367:l5873
to predict and diagnose chronic kidney disease: A review”. 35. Li PK, Garcia-Garcia G, Lui SF, et al. “Kidney health for
Journal of Medical Systems, 44(11), 191. everyone everywhere - from prevention to detection and
18. Jha V, Garcia-Garcia G, Iseki K, et al. “Chronic kidney equitable access to care”. Clin Nephrol 2020;93(3):111-122
disease: global dimension and perspectives”. The Lancet. 36. Crews DC, Bello AK, Saadi G. “Burden, access, and disparities
2013;382(9888):260-272. in kidney disease”. Kidney Int 2019;95(2):242-248
19. Fernandes, M. S., et al. (2015). “Big data analytics and chronic 37. V’asquez-Morales GR, Martinez-Monterrubio SM,
kidney disease: Hope or hype?”. Journal of Nephrology, 29(3), Moreno-Ger P, Recio-Garcia JA. “Explainable prediction
339-347. of chronic renal disease in the colombian population using
20. Zhang, Z., & Beck, M. W. (2020). “Big Data and Machine neural networks and case-based reasoning”. IEEE Access.
Learning in chronic kidney disease: A systematic review”. 2019;7:152900–10.
Journal of Translational Medicine, 18(1), 261. 38. Sinha P, Sinha P. “Comparative study of chronic kidney
21. Shaikhina, T., & Khovanova, N. (2017). “Handling limited disease prediction using knn and svm”. Int J Eng Res Technol.
datasets with neural networks in medical applications: A 2015;4:608–12.
small-data approach”. Artificial Intelligence in Medicine, 75, 39. Khan B, Naseem R, Muhammad F, Abbas G, Kim S. “An
51-63. empirical evaluation of machine learning techniques for chronic
22. Koyner, J. L., & Carey, K. A. (2018). “Big data and predictive kidney disease prophecy”. IEEE Access. 2020;8:55012–22.
analytics: Nephrology research and clinical practice in the 40. Hosseinzadeh M, Koohpayehzadeh J, Bali AO, Asghari P,
21st century”. Seminars in Nephrology, 38(6), 582-589. Souri A, Mazaherinezhad A, Bohlouli M, Rawassizadeh R.
23. Tan, A. C., et al. (2018). “Early detection of chronic kidney “A diagnostic prediction model for chronic kidney disease
disease in developing countries: The role of machine learning”. in internet of things platform”. Multimedia Tool Appl.
IEEE Access, 6, 67879-67888. 2021;80(11):16933–50.
24. Jha, V., et al. (2017). “Chronic kidney disease: Global dimen 41. Gunarathne WHSD, Perera KDM, Kahandawaarachchi
sion and perspectives”. The Lancet, 382(9888), 260-272. KADCP. “Performance evaluation on machine learning
25. Miotto, R., et al. (2017). “Deep patient: An unsupervised classification techniques for disease classification and
representation to predict the future of patients from the forecasting through data analytics for chronic kidney disease
electronic health records”. Scientific Reports, 7, 26094. (ckd)”. In: 2017 IEEE 17th international conference on
Early-Stage Chronic Kidney Disease Detection using Machine Learning with Bigdata 395
bioinformatics and bioengineering (BIBE). IEEE: UK; 2017. 44. Charleonnan A, Fufaung T, Niyomwong T,
p. 291–6. Chokchueypattanakit W, Suwannawach S, Ninchawee N.
42. Alasker H, Alharkan S, Alharkan W, Zaki A, Riza LS. “Predictive analytics for chronic kidney disease using machine
“Detection of kidney disease using various intelligent learning techniques”. In: 2016 management and innovation
classifiers”. In: 2017 3rd international conference on science technology international conference (MITicon). IEEE: UK;
in information technology (ICSITech). IEEE; 2017. p. 681–4. 2016. p. 80–3.
43. Abdullah AA, Hafidz SA, Khairunizam W. “Performance
Note: All the figures and tables in this chapter were adapted
comparison of machine learning algorithms for classification
from https://www.siemens-healthineers.com/en-uk/laboratory
of chronic kidney disease (CKD)”. J Phys: Conf Ser.
diagnostics/assays-by-diseases-conditions/kidney-disease/about
2020;1529(5):052077.
kidney-disease
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
396 Algorithms in Advanced Artificial Intelligence
Abstract: Wireless Sensor Networks (WSNs) are critical across environmental monitoring, surveillance, and healthcare
applications. Energy conservation prolongs network lifetime and ensures continuous data gathering. This research presents a
novel approach to optimizing WSN energy consumption through Mahalanobis Distance-Based K-Means Clustering (MDB
KMC) combined with the bio-inspired Firefly Algorithm for cluster head selection. MDB-KMC efficiently divides nodes
into clusters, enabling effective data transmission. The Firefly Algorithm then optimizes clusters by dynamically selecting
heads based on residual energy and base station distance modeled on the flashing behavior of fireflies. Inspired by the self-
organizing and adaptive capabilities of fireflies, it adjusts cluster head roles minimizing energy use. Extensive simulations
demonstrate significant improvements over traditional methods. Adapting clustering and roles reduces consumption, extends
lifetime, and enhances reliability and performance. Integrating MDB-KMC and the Firefly Algorithm thus provides a robust,
efficient solution to address WSN energy optimization challenges. This enables more reliable sensor network deployment. This
novel integration of adaptive clustering and bio-inspired optimization techniques optimizes WSN energy efficiency, improving
real-world performance.
Keywords: Wireless sensor networks (WSNs), Mahalanobis distance-based K-means clustering (MD-KMC), Firefly algorithm
(FA)
DOI: 10.1201/9781003529231-59
An MDB-KMC and Firefly-Based Clustering Approach for Energy Optimization in Wireless Sensor Networks 397
of paramount importance. The Firefly Algorithm offers an some energy efficiency aspects, it does not account for real-
adaptive optimization technique inspired by the illuminative time network conditions and the dynamic nature of WSNs.
behavior of fireflies. Its inherent self-organizing capabilities Our research bridges this gap by integrating the dynamic
align well with the decentralized nature of sensor networks, Firefly Algorithm for Cluster Head selection which enables
where nodes engage in peer-to-peer communication to nodes to collaboratively decide CH roles based on real-time
identify cluster heads. The primary objective of this research environmental and network conditions. This adaptability is a
is to introduce and evaluate the synergistic integration of the significant improvement over existing approaches that rely
MDB-KMC clustering algorithm and The Firefly Algorithm solely on static criteria for CH selection. Moreover, Wang
for Cluster Head Selection in WSNs. et al. (2020) [5] propose a hybrid CH selection algorithm
The aim is to achieve the following: that considers both residual energy and distance to the base
station. While this approach offers energy savings, it assumes
1. Enhance the accuracy of cluster formation by
a known network topology, which may not hold in practical
considering both spatial and statistical characteristics
WSN deployments.
of sensor nodes using MDB-KMC.
2. Improve energy efficiency by dynamically selecting
cluster heads based on real-time. 3. Research Motivation
3. Conditions and network demand through the Firefly The research motivation behind this work is grounded in the
Algorithm. imperative need for energy-efficient solutions in Wireless
4. Extend the operational life span of WSNs, making Sensor Networks (WSNs) due to the limited battery capacity
them more sustainable and cost-effective for long-term of resource-constrained sensor nodes. This research ac-
monitoring and data collection applications. acknowledges the heterogeneity among sensor nodes and
This paper is organized as follows: Section 2 provides related the dynamic nature of WSN deployments, which necessitate
work in the field of energy-efficient clustering and cluster adaptable and responsive strategies for energy optimization.
head selection in WSNs. Section 3 presents the research The integration of the Mahalanobis Distance-Based K-Means
motivation Section 4 introduces the proposed work, while Clustering (MDB-KMC) algorithm and the Firefly Algorithm
Section 5 presents and analyses the results and winds up the is proposed to address existing research gaps that often
work. lack precision and adaptability. MDB- KMC accounts for
node diversity by considering both spatial distribution and
statistical attributes, while the Firefly Algorithm introduces
2. Related Work real-time adapt- ability inspired by nature. Together, these
In the realm of clustering for Wireless Sensor Networks innovations aim to provide a comprehensive solution for
(WSNs), numerous studies have explored the application of energy-efficient clustering and cluster head selection in
traditional clustering algorithms, such as K-Means, LEACH, WSNs, facilitating practical and sustainable deployments
and HEED, to optimize network energy consumption. While across various applications.
these approaches have made substantial contributions, they
often overlook the inherent heterogeneity among sensor 4. Proposed Work
nodes.
The proposed work encompasses the implementation and
The research by Zhangetal. (2017) [2] primarily focuses on integration of two key algorithms. First, the Mahalanobis
improving cluster formation by proposing a distributed energy- Distance-Based K-Means Clustering (MDB- KMC)
efficient clustering algorithm. However, this work does not algorithm will be deployed to improve cluster formation
fully address the variability in sensor rnode characteristics, accuracy. MDB-KMC incorporates spatial distribution and
including different sensing capabilities and communication statistical attributes, addressing the heterogeneity of sensor
ranges, which can significantly impact energy consumption nodes for more precise cluster creation. Second, the Firefly
patterns. This gap highlights the need for more sophisticated Algorithm will be integrated for dynamic Cluster Head (CH)
clustering methods, like the Mahalanobis Distance-Based selection. This algorithm adapts CH roles in real time based
K-Means Clustering (MDB-KMC), as employed in our on environmental conditions and network requirements,
research, to consider both spatial distribution and statistical optimizing energy utilization.
attributes for more accurate cluster formation.
Furthermore, Liu et al. (2019) [3] propose a cluster head 4.1 Clustering
selection algorithm based on node residual energy and The proposed MDB-KMC approach strategically forms
distance to the base station. While this approach addresses clusters based on the distance to the Cluster Head (CH) and the
398 Algorithms in Advanced Artificial Intelligence
energy levels of the CH. It effectively addresses the existing 4.5 Attractiveness Formula
challenge by considering only the distance measurements
The attractiveness b of a firefly is proportional to its light
between the CH and Sensor Nodes(SN),there by optimizing
intensity seen by neighboring fireflies.
Network Lifetime (NLT). MDB-KMC groups similar SNs
with the CH, thereby facilitating energy conservation. To b = b0 * e – g r2 (5)
achieve this, the MDB-KMC approach initially evaluates Where r is the distance between two fireflies and g controls
the Mahalanobis distance while considering the covariance the decrease of attractiveness with distance.
matrix S, the SN To achieve this, the MDBKMC approach
initially evaluates the Mahalanobis distance while considering 4.6 Movement
the covariance matrix S, Each firefly i is attracted towards more attractive (higher
AN ÈSN1*, SN 2*, SN 3*, SN 4*, ... SN N* ,˘˚ (1) intensity) fireflies j and moves towards them.
Î and CH
D xi = b0 * e – g r2ij * (xj – xi) + a * ?i (6)
ÈC1* , C2* , C3* , C4* , ..... C N* ˘ The second term introduces random movement with a
Î ˚ forms a cluster i.e.
(2) controlling the magnitude and ei being a random vector.
(C ) ( )
T
EKD = *
K - SN K* S -1 CK* - SN K*
4.7 Iterate Approach
Now, the SN that possesses less distance value is sorted to In each iteration, light intensity and movement are updated
that respective CH centered upon the evaluation and creates for each firefly by the above steps. Over iterations, fireflies
a cluster (CKi), i.e. cluster around nodes best suited as cluster heads. The
movements allow the algorithm to explore the search space
Î Î ˚ ( Î )
CKCNew = ÍÈCK1 ÈS1K CH1* ˘ , CK 2 ÈSK2 CH 2* ˘ ,
˚ ( ) and identify optimal CHs balancing energy and base station
(3) distance. The clustered high-intensity fireflies represent
( )
CK3 ÈSK3 CH3* ˘ , ... CK n ÈSKn CH N*
Î ˚ Î ( ) ˘˘
˚ ˙˚ the selected set of cluster head nodes in the wireless sensor
network.
Therefore, a cluster is formed by MD-KMC, and the suitable
CH is selected in a way that brings maximum NLT and
decreases the number of message exchanges along with 5. Results and Discussions
obtains independent time complexity of network growth. In this section, the proposed method is rigorously assessed
through a series of numerical evaluations, contrasting it
4.2 Cluster Head Selection with existing methodologies. The evaluation specifically
Firefly Algorithm for finding cluster heads in a Wireless focuses on one key aspect: the CH (Cluster Head) selection
sensor network by considering parameters like residual technique, denoted as the Proposed FA, technique.
energy, cost function, and distance to the base station. Here The entire experimentation is conducted within the
is a more detailed explanation of using the Firefly Algorithm Matlab environment, utilizing publicly accessible data.
for cluster head selection in wireless sensor networks: A network field of dimensions 2000 meters by 2000
meters is established to set the stage for these evaluations.
4.3 Encode Solution Within this spatial domain, a total of 250 sensor nodes
Each firefly represents a potential cluster head (CH) node. are randomly deployed. These nodes are positioned with
The location xi of firefly i maps to the location of sensor node a uniform separationof30metersbetweenadjacentnodes.
i sensor network deployment area. Itisassumedthateachsensornode has a radio range extending
up to 50 meters, and data packets are configured with a size
4.4 Objective Function of 512 bits. The simulation is carried out over 4200 seconds,
The light intensity Ii of each firefly encodes the desirability during which various performance metrics are assessed.
of selecting its corresponding sensor node as a cluster head. Additionally, packet sizes are varied within the range of
It is based on two metrics - residual energy (Ei) of a node and 500 bytes. Furthermore, the nodes in the network exhibit
distance to the base station (dib). mobility, with their speeds ranging from 0 to 20 meters per
second. Additionally, the mobility of the Sensor Nodes (SN)
Ii = w1 * Ei + w2 * 1/dib (4) is considered, with their speeds ranging from 2 to 20 meters
Here w1 and w2 allow weighting the relative importance per second. This comprehensive evaluation framework
of the two metrics. Maximizing Ii will maximize residual allows for a robust assessment of the proposed method’s
energy and minimize distance to the base station. effectiveness and efficiency. Illustrates the throughput values,
An MDB-KMC and Firefly-Based Clustering Approach for Energy Optimization in Wireless Sensor Networks 399
show casing that the proposed FA achieves a throughput accurately clustering nodes and iteratively selecting optimal
ranging from 240 kbps to 270 kbps for 50 to 250 sensor heads based on residual energy and base station distance
nodes with a 50-node increment. In contrast, existing modeled on Firefly Algorithm collective behavior. Extensive
methodologies exhibit a broader throughput range of 80 kbps simulations demonstrate the technique significantly improves
to 210 kbps, indicating some data loss between CHs and the efficiency and throughput over existing methods, optimizing
base station. The proposed FA achieves throughput values wireless sensor network energy utilization through adaptive
of 270Kbps (50SN), 260Kbps (100SN), 255Kbps (150SN), clustering and selection for robust real-world deployment.
245Kbps (200SN),and 235Kbps (250SN), with throughput
decreasing as the number of sensor nodes increases. Overall,
the proposed method consistently out performs other existing
6. Conclusions
systems in terms of both throughput and energy efficiency Our self-configured protocol for wireless sensor networks
Fig. 59.1 and Fig. 59.2. utilizes the proposed firefly-based cluster head selection
minimizing energy consumption when integrated with MDB
5.1 Performance Analysis of the Proposed CH KMC clustering. Experimental validation demonstrates
Selection Technique remarkable performance including extremely low 1000J
This research proposes an energy optimization approach energy consumption over 2000 rounds and high 225kbps
for Wireless Sensor Networks integrating Mahalanobis throughput with 250 nodes. The approach also exhibits
Distance-Based K-Means Clustering and the bio-inspired an excellent 710+ round network lifetime with 250 nodes,
Firefly Algorithm (for adaptive cluster head selection, establishing superior energy efficiency versus existing
methods through comprehensive analysis. In future work, aware adaptive clustering in wireless sensor networks. Future
this research established the effectiveness of integrating Generation Computer Systems,104:1–14,2020.
MDB-KMC with the bio-inspired Firefly Algorithm for 4. V. Thala gondapati and M.P. Singh. A self-organized priority-
energy-efficient cluster head selection in WSNs. The self- based MAC protocol in wireless sensor networks based on
SDR-RHSO optimal relay node selection and HL-ANN
organizing and adaptive capabilities of the firefly approach
wake-up scheduling. Journal of Ambient Intelligence and
enhanced optimization performance.
Humanized Computing, 14(8):11093–11102, 2023.
5. Jin Wang, Yu Gao, Wei Liu, Wenbing Wu, and Se-Jung Lim.
References An asynchronous clustering and mobile data gatherings
chema based on timer mechanism in wireless sensor networks.
1. Abdulmughni Hamzah, Mohammad Shurman, Omar Al- Computers, Materials Continua, 58(3), 2019.
Jarrah, and Eyad Taqied- din. Energy-efficientfuzzy-logic 6. Quan Wang, Deyu Lin, Pengfei Yang, and Zhiqiang Zhang. An
based clustering technique for hierarchical routing protocols energy-efficient compressive sensing-based clustering routing
in wireless sensor networks. Sensors, 19(3):561, 2019. protocol for WSNs.IEEE Sensors Journal, 19(10):3950–3960,
2. Guangjie Han, Chenyu Zhang, Jinfang Jiang, Xuan 2019.
Yang, and Mohsen Guizani. Mobile anchor nodes path 7. Chuan Xu, Zhengying Xiong, Guofeng Zhao, and Shui Yu. An
planning algorithms using network-density-based clus- energy-efficient region source routing protocol for lifetime
teringinwirelesssensornetworks. Journal of Network and maximization in WSN. IEEE Access, 7:135277–135289,
Computer Applications, 85:64–75, 2017. 2019.
3. X.Liu, R.Zhu, A.Anjum, J.Wang, H.Zhang, and M.Ma.
Intelligentdatafu- sion algorithm based on hybrid delay- Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
K. Harish Kumar1
Research Scholar, Department of Computer Science & Engineering,
Koneru Lakshmaiah Education Foundation, Hyderabad, Telangana, India,
and Assistant Professor, Department of Computer Science & Informatics,
Mahatma Gandhi University, Nalgonda, Telangana, India
K. Srinivas2
Professor, Department of Computer Science & Engineering,
Koneru Lakshmaiah Education Foundation,
Hyderabad, Telangana, India
Abstract: This paper presents a Software Effort Estimation (SEE) framework for project management, addressing the
increasing demand for high-quality software. It involves data gathering, preprocessing, BERT-based word embedding,
clustering with PLSKCD-K-Means, and task ranking using ZS-GTBOA. Feature extraction, dimensionality reduction, and SEE
implementation with the RSLU-GNL-GRU classifier follow. The experimental evaluation highlights the proposed technique’s
superior performance over existing models.
Keywords: BERT, PLSKCD-K-Means, ZS-GTBOA, FS, and RSLU-GNL-GRU form the key components in this approach
1
khrsharma@gmail.com, 2srirecw9@klh.edu.in
DOI: 10.1201/9781003529231-60
402 Algorithms in Advanced Artificial Intelligence
Magnitude of Relative Error but lacking focus on software Developed EEAC, a software development effort estimation
experimentation needs. (Khan et al., 2021): Developed a model, incorporating data clustering and FPA methods.
DNN for SEE with metaheuristic algorithms, achieving While exhibiting better performance, FPA was noted to be
superior outcomes but facing elevated convergence time in time-consuming due to numerous elements.
some learning rate scenarios. (De Carvalho et al., 2021):
Proposed ELM for SEE, demonstrating better outcomes 3. Proposed Software Effort
than prevailing methods but encountering forecasting
difficulties with limited data.(Ali et al., 2023): Designed Estimation Framework
a heterogeneous Ensemble Effort Estimation method, This paper introduces an SEE framework for software project
integrating models for improved estimation, dependent on management, centering on RSLU-GNL-GRU. It groups
model weight assignments. (Nhung et al., 2022): Presented software requirement and project data by task size, ranks
a parametric SEE methodology using Optimizing Correction them with ZS-GTBOA, and estimates effort using RSLU
Factors and Multiple Regression Models, outperforming GNL-GRU as shown in below system’s structural design.
other models but facing challenges in estimation accuracy
due to differing historical data distributions.(Rhmann et 3.1 Preprocessing of Project Data
al., 2022): Explored hybrid search techniques and weighted
The proposed model starts by gathering software project data.
ensemble with met heuristic algorithms for SEE, surpassing
ML-based algorithms but lacking in determining economic P = { p1 , p2 , p3 ,...., pm }
(1)
benefits for a software organization.(Van Hai et al., 2022):
Where, the data regarding the mth project is denoted as pm. • Output layers: The encoded output is directed to the
Pre-processing is executed using the ‘3’ steps. output layer with a simple classifier, comprising a
• Missing value imputation: Missing values are imputed fully connected layer and an activation function (ℓ) is
with the average of adjacent values. computed as,
• Repeated data removal: Duplicates in the data are 1Ê q ˆ
eliminated. ℓ= Á
2Ë
k - Â Vi ˜ (5)
i=1 ¯
• Numeralization: String values are converted to
numerical format for processing. Where, the target word embedding score is denoted as k, and
V is the embedded output.
The pre-processed project data is represented as Ppre.
Here, using the BERT algorithm, the data Rpre is embedded. Where, the Kendall constant is signified as H, and the set of
Primarily, for each word in Rpre, the tokens (¡) are determined unordered pairs of the data points is notated as Q. Until there
and fed to the embedding layer. are no more changes, the above procedure is repeated. The
• Embedding layer: This layer executes token embedding, final clustered output F is articulated as,
segment embedding, and position embedding on ¡ and F = {Tl arg e, Tsmall } (9)
provided to the transformer encoder layer.
• Transformer encoding layer: The embeddings are Where, the large tasks and small tasks are denoted as Tl arg e
transformed into numerical vectors. The string values and Tsmall respectively.
are encoded by the encoder and the decoder provides the
3.6 Ranking
contextual embedding (V) as,
Utilizing ZS-GTBOA, Tl arg e is ranked based on project size
{
V = c1 , c2 ,....., cq } (4) for efficient and accurate estimation. GTBOA’s parameter
initialization, initially using normalization, is modified due
Where, the qth string’s contextual embedding is denoted to limitations in handling outlier positions during beetle
as cq. initialization. Z-score normalization is employed as an
404 Algorithms in Advanced Artificial Intelligence
(
g 1 = (1 - Q ) ◊ Tbest - Trand1 ) (17)
{(
���� ����
)(
FS = u M1 M 2 + a I
-1
) } (20)
g2 = (1 - Q ) ◊ (T best - Trand2 ) (18)
Wherein, u signifies the total number of instances, a
Where, the best solution is signified as Tbest and the crossover regularization parameter is notated as a, I denotes the
operator is notated as Q. Therefore, using the ZS-GTBOA, perturbation term, and the between-class
���� ���� scatter and total
the large tasks are ranked (Tranked) centered on size. The ZS scatter matrix are notated as M1 and M 2 .
GTBOA’s procedure is explained in Algorithm 1.
Software Requirements Based Software Effort Estimation using RSLU-GNL-GRU in Software Project Management 405
(a) (b)
Fig. 60.6 Loss value of the proposed method during (a) training, and (b) testing
iterations, indicating that ZS in the proposed model leads to 2. De Carvalho, H. D. P., Fagundes, R., & Santos, W. (2021).
better fitness values and optimal solutions. Extreme Learning Machine Applied to Software Development
Effort Estimation. IEEE Access, 9, 92676–92687. https://doi.
4.4 Performance Evaluation of the Proposed ZS org/10.1109/ACCESS.2021.3091313
GTBOA 3. Kaur, A., & Kaur, K. (2022). Systematic literature review
of mobile application development and testing effort
The proposed optimization algorithm’s performance is estimation. Journal of King Saud University - Computer and
compared with GTBOA, FOA, SFOA, and PSOA. The Information Sciences, 34(2), 1–15. https://doi.org/10.1016/j.
proposed approach exhibits superior performance and jksuci.2018.11.002
promising outcomes compared to other existing methods with 4. Khan, M. S., Jabeen, F., Ghouzali, S., Rehman, Z., Naz, S.,
higher processing times. Figure 60.9 shows that the proposed & Abdul, W. (2021). Metaheuristic Algorithms in Optimizing
SEE achieves 0.2450% MMRE, outperforming ANN with Deep Neural Network Model for Software Effort Estimation.
0.439% MMRE. Other methods also exhibit higher MMRE. IEEE Access, 9, 60309–60327. https://doi.org/10.1109/
ACCESS.2021.3072380
The proposed methodology enhances SEE efficacy through
5. Kumar, P. S., Behera, H. S., Anisha Kumari, K., Nayak, J.,
the utilization of SEE requirement details.
& Naik, B. (2020). Advancement from neural networks to
deep learning in software effort estimation: Perspective of
two decades. Computer Science Review, 38, 1-32. https://doi.
org/10.1016/j.cosrev.2020.100288
6. Mahmood, Y., Kama, N., Azmi, A., Khan, A. S., & Ali, M.
(2022). Software effort estimation accuracy prediction of
machine learning techniques: A systematic performance
evaluation. Software - Practice and Experience, 52(1), 39–65.
https://doi.org/10.1002/spe.3009
7. Nhung, H. L. T. K., Van Hai, V., Silhavy, R., Prokopova, Z.,
& Silhavy, P. (2022). Parametric Software Effort Estimation
Based on Optimizing Correction Factors and Multiple
Linear Regression. IEEE Access, 10, 2963–2986. https://doi.
org/10.1109/ACCESS.2021.3139183
8. Pandey, M., Litoriya, R., & Pandey, P. (2020). Validation of
Fig. 60.9 Comparative measure of the proposed framework Existing Software Effort Estimation Techniques in Context
with Mobile Software Applications. Wireless Personal
Communications, 110, 1659–1677. https://doi.org/10.1007/
4.5 Comparative Measurement with Literature s11277-019-06805-0
Papers 9. Priya Varshini, A. G., Anitha Kumari, K., Janani, D., &
The proposed framework is compared to conventional Soundariya, S. (2021). Comparative analysis of Machine
DANN, DNN, and SVR (Rankovic et al., 2021; Khan et al., learning and Deep learning algorithms for Software Effort
2021; Rhmann et al., 2022) for superiority Estimation. Journal of Physics: Conference Series, 1767(1),
1-11. https://doi.org/10.1088/1742-6596/1767/1/012019
10. Rankovic, N., Rankovic, D., Ivanovic, M., & Lazic, L.
5. Conclusion (2021). A New Approach to Software Effort Estimation Using
Different Artificial Neural Network Architectures and Taguchi
This paper proposes an SEE technique using RSLU-GNL Orthogonal Arrays. IEEE Access, 9, 26926–26936. https://doi.
GRU, incorporating software requirements. The process org/10.1109/ACCESS.2021.3057807
involves pre-processing, word embedding, data combination, 11. Rhmann, W., Pandey, B., & Ansari, G. A. (2022). Software
clustering, ranking, and estimation, achieving a 98% effort estimation using ensemble of hybrid search-based
estimation rate with a training time of 38007ms. Efficient algorithms based on metaheuristic algorithms. Innovations in
clusters form in 31539ms with a 98.91% accuracy rate. The Systems and Software Engineering, 18(2), 309–319. https://
proposed framework outperforms existing methods, but lacks doi.org/10.1007/s11334-020-00377-0
project risk assessment, a topic for future work. 12. Sudarmaningtyas, P., & Mohamed, R. (2021). A review article
on software effort estimation in agile methodology. Pertanika
Journal of Science and Technology, 29(2), 837–861. https://
References doi.org/10.47836/pjst.29.2.08
13. Tawosi, V., Sarro, F., Petrozziello, A., & Harman, M. (2022).
1. Ali, S. S., Ren, J., Zhang, K., Wu, J., & Liu, C. (2023). Multi-Objective Software Effort Estimation: A Replication
Heterogeneous Ensemble Model to Optimize Software Effort Study. IEEE Transactions on Software Engineering, 48(8),
Estimation Accuracy. IEEE Access, 11, 27759-27792. https:// 3185–3205. https://doi.org/10.1109/TSE.2021.3083360
doi.org/10.1109/access.2023.3256533
Software Requirements Based Software Effort Estimation using RSLU-GNL-GRU in Software Project Management 409
14. Van Hai, V., Nhung, H. L. T. K., Prokopova, Z., Silhavy, machines for software effort estimation. PROMISE 2020,
R., & Silhavy, P. (2022). Toward Improving the Efficiency Co-Located with ESEC/FSE 2020, 31–40. https://doi.
of Software Development Effort Estimation via Clustering org/10.1145/3416508.3417121
Analysis. IEEE Access, 10, 83249–83264. https://doi. 17. Xia, T., Shu, R., Shen, X., & Menzies, T. (2022). Sequential
org/10.1109/ACCESS.2022.3185393 Model Optimization for Software Effort Estimation. IEEE
15. Varshini, A. G. P., & Kumari, K. A. (2020). Predictive analytics Transactions on Software Engineering, 48(6), 1994–2009.
approaches for software effort estimation : A review. Indian https://doi.org/10.1109/TSE.2020.3047072
Journal of Science and Technology, 13(21), 2094–2103.
Note: All the figures and tables in this chapter were designed by
16. Villalobos-Arias, L., Quesada-López, C., Guevara-Coto,
the author.
J., Martínez, A., & Jenkins, M. (2020). Evaluating hyper-
parameter tuning using random search in support vector
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
410 Algorithms in Advanced Artificial Intelligence
Chaitanya. K1
Carelon Global Solutions, Hyderabad, India,
Krishna Jayanth Rolla2
Fort Mill, SC 29715
Abstract: This research paper explores the historical evolution of artificial intelligence (AI) and the transformative emergence
of large language models (LLMs). The historical context delves into the inception of AI at the Dartmouth Conference in 1956,
tracing the field’s journey through periods of optimism, such as the development of expert systems, and skepticism, leading to
AI winters. The resurgence of AI in the 21st century is closely linked to breakthroughs in machine learning, particularly deep
learning, setting the stage for advancements in LLMs. The significance of LLMs is a focal point, showcasing their diverse
applications in natural language processing (NLP) and their role in reshaping human-computer interaction. Models like GPT-3,
with its unprecedented 175 billion parameters, exemplify the prowess of LLMs in tasks ranging from healthcare applications,
such as medical literature review, to business applications, where chatbots enhance customer service interactions. The pre-
training and fine-tuning methodology, rooted in deep learning principles, underscores the adaptability of LLMs across varied
NLP domains. Furthermore, the paper examines h,ow LLMs represent a broader advancement in the field of machine learning
and deep learning. The scale of these models enables them to capture intricate patterns and dependencies in data, influencing the
approach to transfer learning. Large language models, trained on extensive datasets, exhibit generalized learning capabilities,
sparking ongoing exploration into more efficient training methodologies and architectures. The continuous quest for enhanced
model interpretability, efficiency, and generalization capabilities forms a key aspect of the paper’s exploration of the evolving
landscape of AI and LLMs.
Keywords: Artificial intelligence (AI), Large language models (LLMs), Machine learning, Natural language processing (NLP),
GPT-3, Healthcare applications, Transfer learning, Generalized learning capabilities
1
ckanchibhotla@gmail.com, 2J.rolla2@gmail.com
DOI: 10.1201/9781003529231-61
The Evolution and Impact of Large Language Models in Artificial Intelligence 411
3. Methodology
3.1 Model Architectures Fig. 61.1 GPT-3 architecture
1. Deep dive into GPT-3 architecture
GPT-3 (Generative Pre-trained Transformer 3) (refer strengths. BERT (Bidirectional Encoder Representations
Fig. 61.1) stands as a pinnacle in large language models, from Transformers), for instance, differs in its approach
representing a breakthrough in natural language understanding by employing bidirectional training. BERT considers both
and generation. The architecture of GPT-3 is rooted in the left and right context during training, enhancing its ability
transformer model, featuring attention mechanisms that to understand context. XLNet, another notable model,
allow it to capture intricate contextual dependencies in vast introduces permutation-based language modeling, allowing
amounts of text data. With an unprecedented 175 billion it to capture long-range dependencies in text. While GPT
parameters, GPT-3 is organized into 96 transformer layers, 3 focuses on unidirectional language modeling, it excels
enabling it to process and generate human-like text with in generating coherent and contextually relevant text. Each
remarkable coherence and versatility. Each layer contributes model has unique advantages and trade-offs. GPT-3’s
to the model’s ability to understand context, utilizing self- massive scale facilitates diverse applications, while BERT’s
attention mechanisms to weigh the importance of different bidirectional training is advantageous for tasks requiring
words in a given sequence. The sheer scale of GPT-3 allows a deeper understanding of context. Understanding these
it to exhibit few-shot learning capabilities, where the model architectural nuances is crucial for selecting the most suitable
can perform new tasks with minimal task-specific training model for specific natural language processing tasks.
data, showcasing the effectiveness of pre-training on diverse
linguistic contexts. 3.2 Training Data and Preprocessing
2. Comparison with other prominent models 1. Importance of diverse and representative datasets
In comparing GPT-3 with other prominent language models, The success of large language models is intricately tied to
it’s essential to consider alternative architectures and their the quality and diversity of the training data. Diverse and
The Evolution and Impact of Large Language Models in Artificial Intelligence 413
accuracy and efficiency of diagnostic processes. For instance, 4.2 Success Stories and Challenges
studies have shown successful applications of LLMs in
1. Highlighting Instances of Successful Implementation
identifying rare diseases by leveraging their language
understanding capabilities [15]. The deployment of LLMs Successful implementations of large language models are
in healthcare showcases their potential to revolutionize evident across various domains. In healthcare, the success
information processing, ultimately leading to more effective story lies in the efficient extraction of medical insights and the
medical decision-making. improvement of diagnostic processes. For instance, a study
utilizing GPT-3 for medical literature review demonstrated
2. Education: Intelligent Tutoring Systems
its ability to generate relevant summaries, facilitating quicker
Large language models play a pivotal role in shaping the access to valuable information [20]. In education, success
future of education, particularly through the development stories involve improved learning outcomes and personalized
of intelligent tutoring systems. GPT-3 and similar models experiences for students using intelligent tutoring systems
offer personalized learning experiences by adapting to based on LLMs. These systems have showcased their
individual learning styles. These systems utilize the language adaptability across different subjects, positively impacting
understanding capabilities of LLMs to provide tailored students’ academic performances [21]. Business applications
explanations and responses to student queries, fostering a highlight the success of chatbots in improving customer
more engaging and effective learning environment [16]. The service interactions, with instances of reduced response
versatility of these models enables them to assist students times and increased customer satisfaction [22]. These success
across various subjects, making education more inclusive. stories underscore the versatile applications of large language
Success stories in education involve the improved academic models, showcasing their potential to enhance processes and
performance of students using intelligent tutoring systems services across diverse domains.
powered by LLMs. These systems not only enhance the
2. Addressing Challenges Faced in Different Domains
learning experience by providing real-time feedback but also
cater to diverse learning styles, addressing the unique needs Despite their successes, large language models face challenges
of individual students [17]. in implementation. In healthcare, challenges include ensuring
the ethical use of patient data and addressing concerns related
3. Business: Chatbots and Customer Service
to the interpretability of AI-driven diagnostic decisions [23].
In the business sector, the integration of large language Bias in training data is a persistent challenge across domains,
models has redefined customer interactions through the leading to concerns about the fairness and equity of AI
development of advanced chatbots. GPT-3, with its language applications. In education, challenges include the continuous
generation capabilities, enables chatbots to engage in natural adaptation of tutoring systems to evolving curricula and the
and contextually relevant conversations with customers. This need for addressing diverse learning styles [24]. In business,
has profound implications for customer service, streamlining challenges involve ensuring the ethical deployment of
interactions and improving overall efficiency [18]. Success chatbots, addressing potential biases in customer interactions,
stories in business applications include instances where and maintaining transparency in automated decision-making
chatbots powered by LLMs have significantly reduced processes [25]. Addressing these challenges requires a
response times, enhanced customer satisfaction, and multidisciplinary approach, involving collaboration between
provided scalable solutions for handling a wide range of researchers, developers, policymakers, and domain experts
customer queries [19]. These applications demonstrate the to establish ethical guidelines and refine methodologies for
transformative potential of large language models in business responsible AI deployment.
operations, particularly in sectors where efficient and
In summary, the real-world applications of large language
responsive customer service is critical.
models in healthcare, education, and business highlight their
Table 61.1 Comparison of large language models
Model Parameters Architecture Applications
GPT-3 175 billion Transformer-based, 96 attention heads Healthcare (medical literature review), Business
(chatbots), Natural Language Processing (NLP) tasks
BERT 340 million (BERT-base) Transformer-based, 12 attention heads NLP, Sentiment Analysis, Question Answering
OpenAI GPT-2 1.5 billion Transformer-based, 48 attention heads Content generation, Text completion
XLNet 340 million (BERT-base) Transformer-based, Permutation Language NLP, Machine Translation, Question Answering
Model (PLM)
T5 (Text-To-Text) 11 billion Transformer-based, 24 attention heads Language translation, Text summarization
The Evolution and Impact of Large Language Models in Artificial Intelligence 415
transformative potential. Success stories underscore improved about the capabilities and limitations of LLMs is essential
processes and services, while challenges necessitate ongoing to managing user expectations and avoiding unintended
efforts to ensure ethical and responsible implementation consequences. Developers and researchers must prioritize
across diverse domains. user privacy, ensuring that data used to train these models
is handled ethically and that user consent is obtained for
5. Implications and Future Directions any data collection. Additionally, ongoing monitoring and
auditing of LLMs are critical to identifying and addressing
5.1 Societal Impact ethical concerns that may arise over time [28].
1. Changes in Communication and Information
2. Mitigating Biases and Ensuring Fairness
Consumption
Mitigating biases in large language models is a key ethical
The deployment of large language models (LLMs) has
consideration. Biases present in training data can be learned
ushered in profound changes in how society communicates
by the model, leading to unfair or discriminatory outcomes.
and consumes information. With the rise of advanced natural
Researchers and developers need to implement strategies for
language processing capabilities in models like GPT-3,
identifying and mitigating biases during the training process.
individuals experience more personalized and context-
This includes careful curation of diverse and representative
aware interactions in online communication platforms.
datasets, the development of algorithms that account for
Conversational agents powered by LLMs have altered the
potential biases, and ongoing evaluation of model outputs
dynamics of human-computer interaction, enabling more
for fairness. Furthermore, there is a need for industry-wide
intuitive and natural conversations. This shift impacts not only
standards and guidelines to ensure that ethical considerations
social media interactions but also extends to customer service,
are consistently addressed across different applications of
education, and content generation. The democratization of
information is another societal impact, as LLMs facilitate easy large language models [29].
access to vast amounts of data and insights, transforming the
5.3 Future Developments
way individuals acquire knowledge and make decisions. As
these communication and information consumption patterns 1. Trends in Large Language Model Research
evolve, it becomes imperative to understand and navigate the The field of large language models is expected to witness
challenges and opportunities they present [26]. several trends in the coming years. One prominent trend
2. Economic and Industrial Implications is the exploration of more efficient and environmentally
sustainable training methodologies. The energy consumption
The economic and industrial landscape is undergoing
associated with training large models has raised concerns,
significant transformations due to the integration of large
leading researchers to investigate methods for reducing
language models. In sectors such as content creation,
environmental impact without compromising performance.
journalism, and marketing, LLMs contribute to the automation
Another trend is the development of models that prioritize
of tasks like writing articles, generating marketing content,
interpretability, enabling users to understand and trust
and even composing code snippets. This automation has the
the decision-making processes of these complex models.
potential to increase efficiency and reduce costs for businesses.
Ongoing research also focuses on improving the fine-tuning
However, it also raises questions about the future of certain
job roles and the need for upskilling in the workforce. The process, allowing models to adapt more effectively to specific
advent of LLMs also influences the development of new tasks and domains [30].
products and services, fostering innovation in areas such as 2. Potential Breakthroughs and Advancements
virtual assistants, automated translation services, and more. Anticipated breakthroughs in large language model research
Policymakers and industry leaders need to carefully navigate include advancements in unsupervised learning, enabling
these changes to ensure inclusive economic growth and models to learn from unlabeled data more effectively.
address potential challenges related to job displacement and This could lead to even greater generalization capabilities
unequal access to emerging opportunities [27]. and improved performance across diverse tasks (refer
Table 61.2). Innovations in natural language understanding,
5.2 Ethical Considerations
contextual reasoning, and multilingual capabilities are also
1. Responsible AI Development and Deployment areas of active exploration. Additionally, the integration
Ethical considerations are paramount in the development of large language models with other AI technologies, such
and deployment of large language models. Responsible AI as computer vision and reinforcement learning, holds the
practices involve transparency in the design and decision- potential for creating more comprehensive and versatile AI
making processes of these models. Clear communication systems. As research progresses, collaborations between
416 Algorithms in Advanced Artificial Intelligence
academia, industry, and policymakers will be crucial to enhanced customer service in business. Success stories
navigating the ethical, societal, and technological challenges highlighted instances of improved efficiency, personalized
posed by these potential breakthroughs [31]. learning experiences, and streamlined customer interactions.
In conclusion, the implications and future directions of large Simultaneously, challenges in different domains underscored
language models encompass a broad spectrum of societal, the importance of ethical considerations, responsible AI
ethical, and technological considerations. Balancing the development, and continuous efforts to address biases. The
implications and future directions section outlined the societal
benefits of technological innovation with ethical responsibility
impact of LLMs, delving into changes in communication,
is central to shaping a future where large language models
economic ramifications, and ethical considerations. It also
contribute positively to society.
provided a glimpse into the potential breakthroughs and
advancements expected in large language model research.
6. Conclusion
6.1 Recapitulation of Key Findings 6.3 Recommendations for Future Research
In conclusion, the exploration of large language models As we look to the future, several recommendations for
(LLMs) and their impact on artificial intelligence has further research emerge. First and foremost, there is a need
revealed significant insights. Key findings include the for continued exploration into mitigating biases and ensuring
transformative influence of models like GPT-3 on natural fairness in large language models, as ethical considerations
language processing, showcasing their prowess in diverse remain paramount. Future research should focus on refining
applications such as healthcare, education, and business. evaluation metrics to better capture the nuanced capabilities
The historical overview highlighted the evolution of artificial and limitations of LLMs, addressing challenges related to
intelligence, from its inception in the 1950s to the recent context understanding, interpretability, and potential biases in
generated text. Furthermore, investigating more sustainable
paradigm shift fueled by breakthroughs in machine learning
and environmentally friendly training methodologies is
and deep learning. The examination of GPT-3’s architecture
essential to minimize the ecological footprint associated with
and its comparison with other models elucidated the
large-scale model training. Collaborative efforts between
technical nuances that contribute to its unparalleled language
academia, industry, and policymakers are crucial to establish
generation capabilities. The discussion on training data
standardize.
underscored the importance of diverse and representative
datasets, acknowledging the challenges of biases in data and In conclusion, this research advances our understanding of
strategies for mitigation. Evaluation metrics and challenges large language models, emphasizing their transformative po
emphasized the ongoing efforts to refine methodologies for tential, ethical considerations, and the intricate interplay be
assessing the performance and understanding the limitations tween technological advancements and societal implications.
of LLMs.
References
6.2 Summary of Contributions
1. McCarthy, J., Minsky, M. L., Rochester, N., & Shannon, C.
This research paper contributes to the understanding of
E. (1955). A proposal for the Dartmouth summer research
the multifaceted landscape of large language models, project on artificial intelligence. AI magazine, 27(4), 12-14.
amalgamating technical, societal, and ethical dimensions. 2. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L.,
The case studies illustrated the real-world applications of Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you
LLMs, exemplifying their impact in healthcare diagnosis, need. In Advances in neural information processing systems
intelligent tutoring systems in education, and chatbots for (pp. 5998-6008).
The Evolution and Impact of Large Language Models in Artificial Intelligence 417
3. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., 17. Hsu, Y. L., Cholleti, S. R., & Lee, Y. F. (2018). Evaluating the
Dhariwal, P., ... & Amodei, D. (2020). Language models are effectiveness of intelligent tutoring systems: A case study in
few-shot learners. arXiv preprint arXiv:2005.14165. high school mathematics. Computers & Education, 116, 72
4. Devlin, J., Chang, M. W., Lee, K., & Toutanova, K. (2018). 88.
BERT: Pre-training of deep bidirectional transformers for 18. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.,
language understanding. arXiv preprint arXiv:1810.04805. Dhariwal, P., ... & Amodei, D. (2020). Language models are
5. Yang, Z., Dai, Z., Yang, Y., Carbonell, J., Salakhutdinov, few-shot learners. arXiv preprint arXiv:2005.14165.
R., & Le, Q. V. (2019). XLNet: Generalized autoregressive 19. Gao, J., Dolan, B., & Chen, W. (2018). Content-aware neural
pretraining for language understanding. In Advances in neural conversation models. arXiv preprint arXiv:1812.10687.
information processing systems (pp. 5753-5763). 20. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.,
6. Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Dhariwal, P., ... & Amodei, D. (2020). Language models are
Matena, M., ... & Liu, P. J. (2019). Exploring the limits of few-shot learners. arXiv preprint arXiv:2005.14165.
transfer learning with a unified text-to-text transformer. arXiv 21. Hsu, Y. L., Cholleti, S. R., & Lee, Y. F. (2018). Evaluating the
preprint arXiv:1910.10683. effectiveness of intelligent tutoring systems: A case study in
7. See, A., Liu, P. J., & Manning, C. D. (2017). Get to the high school mathematics. Computers & Education, 116, 72
point: Summarization with pointer-generator networks. In 88.
Proceedings of the 55th Annual Meeting of the Association 22. Gao, J., Dolan, B., & Chen, W. (2018). Content-aware neural
for Computational Linguistics (Volume 1: Long Papers) (pp. conversation models. arXiv preprint arXiv:1812.10687.
1073-1083). 23. Kulkarni, S., Seneviratne, M. G., & Soh, C. B. (2020). Ethical
8. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., implications of using AI in clinical diagnosis: A scoping
Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you review. IEEE Transactions on Technology and Society.
need. In Advances in neural information processing systems 24. Kulkarni, S., Seneviratne, M. G., & Soh, C. B. (2020). Ethical
(pp. 5998-6008). implications of using AI in clinical diagnosis: A scoping
9. Gao, J., Dolan, B., & Chen, W. (2018). Content-aware neural review. IEEE Transactions on Technology and Society.
conversation models. arXiv preprint arXiv:1812.10687. 25. Hajian, S., Bonchi, F., & Castillo, C. (2016). Algorithmic bias:
10. Bender, E. M., & Friedman, B. (2018). Data statements for From discrimination discovery to fairness-aware data mining.
natural language processing: Toward mitigating system bias Data Mining and Knowledge Discovery, 30(3), 815-847.
and enabling better science. Transactions of the Association 26. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J.,
for Computational Linguistics, 6, 587-604. Dhariwal, P., ... & Amodei, D. (2020). Language models are
11. Caliskan, A., Bryson, J. J., & Narayanan, A. (2017). Semantics few-shot learners. arXiv preprint arXiv:2005.14165.
derived automatically from language corpora necessarily 27. Gao, J., Dolan, B., & Chen, W. (2018). Content-aware neural
contain human biases. Science, 356(6334), 183-186. conversation models. arXiv preprint arXiv:1812.10687.
12. Zellers, R., Holtzman, A., Rashkin, H., Bisk, Y., Farhadi, A., 28. Jobin, A., Ienca, M., & Vayena, E. (2019). The global landscape
Roesner, F., & Choi, Y. (2019). Defending against neural fake of AI ethics guidelines. Nature Machine Intelligence, 1(9),
news. arXiv preprint arXiv:1905.12616. 389-399.
13. Jobin, A., Ienca, M., & Vayena, E. (2019). The global landscape 29. Buolamwini, J., & Gebru, T. (2018). Gender shades:
of AI ethics guidelines. Nature Machine Intelligence, 1(9), Intersectional accuracy disparities in commercial gender
389-399. classification. Proceedings of the 1st Conference on Fairness,
14. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Accountability and Transparency, 77-91.
Dhariwal, P., ... & Amodei, D. (2020). Language models are 30. Kaplan, J., McCandlish, S., Henighan, T., Brown, T., Chess,
few-shot learners. arXiv preprint arXiv:2005.14165. B., Child, R., ... & Radford, A. (2020). Scaling laws for neural
15. Gehrmann, S., Dernoncourt, F., Li, Y. A., Carlson, E. T., Wu, language models. arXiv preprint arXiv:2001.08361.
J., & Farri, O. (2020). Comparing rule-based and deep learning 31. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L.,
models for patient phenotyping: a case study on IBD in Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you
electronic health records. arXiv preprint arXiv:2005.13531. ↩ need. In Advances in neural information processing systems
16. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., (pp. 5998-6008).
Gomez, A. N., ... & Polosukhin, I. (2017). Attention is all you 32. https://clive-gomes.medium.com/pre-training-large
need. In Advances in neural information processing systems language-models-at-scale-d2b133d5e219
(pp. 5998-6008).
Note: All the figures and tables in this chapter were designed by
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
418 Algorithms in Advanced Artificial Intelligence
O. Sri Nagesh1
Associate Professor, Department of CSE, Anurag University, Hyderabad
B. Rajarao2
Senior member of technical staff, Oracle inc.,
Voore Subrahmanyam3
Associate Professor, Department of CSE, Anurag University, Hyderabad
Abstract: Numerous machine learning algorithms and procedures were used on the acquired data in the current paper. The
methods covered include classification, statistical analysis, assessment, and unsupervised learning methods. Data on PD and
non-PD patients are displayed using a visualization technique. Techniques include the Operating Characteristic (ROC), Sieve
Multi-gram, and Self-Organizing Map (SOM), among others. The information focuses a lot on sick classes. Decision Tree,
Logistic Regression, SVM, Linear Regression, KNN, Dimensionality Reduction Algorithms, Random Forest, K-Means, Naive
Bayes, Gradient Boost, and Ada-boost are a few additional algorithms. These algorithms can be applied to any data issue.
Keywords: PD-Parkinson’s disease, SOM, ROC, KNN
1
nagesh.osri@gmail.com, 2rajaraob@yahoo.com, 3voore.subrahmanyam206@gmail.com
DOI: 10.1201/9781003529231-62
Several Machine Learning Techniques Used to Forecast Parkinson Disease 419
Red color symbolizes class value 1, which denotes data False Positive Rate (- ve FPR) at different edge settings.
connected to PD disease, and blue color denotes class 0, which Sensitivity relates to Parkinson’s disease (PD) value recalls 1
denotes non-PD-associated data. Coordinates connected to in machine learning (ML) technique. The false-positive rate
the basic characteristics of frequency. (0 - specificity) relates to Non-Parkinson Disease (Non-PD)
The Receiver Operating Characteristic (ROC) for the also known as fall-out. In this work, another visualization
classification algorithms Majority, K-Nearest Neighbour, technique is also used, that is Sieve Multigram. Fig. 62.5
and SVM is displayed in Fig. 62.4 is a graphical plot that shows the co-relations among selected features.
shows how well a binary classifier system performs. Most Sieve Multigram shows the arrangement of correlated features.
algorithms use SVM and k-nearest neighbour. The ROC Red color indicates negative (class value 1- that indicates PD)
bend therefore acts as a factor in a component of drop out. correlation and blue color indicates positive correlation (class
ROC plot has shown specificity at 0.25. K-NN (K-Nearest Value 0 that indicates Non-PD). Lines with more intensity
Nighbour) has demonstrated 82.5% of precision. SVM shows how strong are the bonds between the relationships.
(Support Vector Mechines) has indicated 88.9% precision The values show the correctly classified instances. The
taking into account ROC results. Hence SVM has predicted accuracy is displayed in Table 62.1 (Correctly classified
good results compared to other two algorithms. occurrences). The most accurate algorithm is Random Forest
For instance from Fig. 62.4 we can distinguish that red (90.26), which is followed by K-Star (89.74). Naïve Bayes
lines having PD and blue lines indicate Non-PD. There are displayed the least accuracy (69.23) based on the PD dataset.
some other interesting relations are also showed. Every Bayes Net works on probability theory concept. It is a direct
one character has system of connections utilizing strainer approach and it allows a rich structure. According to expert
diagrams demonstrating bury and intra associations with the opinion data, it is built-in model. In addition to that it predicts
solid and unhealthy information like basic sound attributes. output by giving inputs. It supports for missing data during
Present work shows the values by using other classification learning and classification. Bayes-Net has shows accuracy of
methods as shown in table below. The bend is made by 80.00 based on PD dataset, ware as Naïve Bayes shows 69.23
plotting the Genuine Positive Rate (+ve TPR) against the least accuracy.
Table 62.1 Classified instances based on algorithms 2. Divisive: All perspectives start with the same figure in this
Algorithm Classified Instances
“top down” paradigm.3.6 has demonstrated that there are
more groups in the solid dataset for fundamental frequency
Bayes Net algorithm 80
and fewer groups with contaminated data. The Fo attribute
Naive Bayes 69.23 in the healthy dataset has a range of 95.7 to 252.45 (blue
Logistic regression 83.66 coloured), while the diseased dataset has a range of 144.18
Simple Logistic 84.61 to 202.63 (red coloured). Self-Organizing Map (SOM) is
KStar algorithm 89.74
a type of unsupervised learning. The objective is to find
some hidden structure of the information. The node in SOM
ADTree algorithm 86.15
called neurons, linked with every hub is same as input data
J48 80.51 vector’s dimension as weight vector and a position in the map
LMT 86.15 projection. The standard course of action of hubs is a two-
Random Forest 90.26 dimensional and the spacing is rectangular or hexagonal grid.
The grid describe high dimension to low dimension input
Another visualization method used in this work is hierarchical space.
clustering method. This classifier is of two Strategies. The weights will draw a two dimensional vector indicating
1. Agglomerative: This is a “base up” methodology; every diseased and non diseased PD values, it is as follows.
perception begins in its own bunch and matches the groups The concept of clustering was done on the collected data it
that are converged as one climb primate. show the clusters of PD and Non PD people as shown in the
Group and parts are performed recursively as one move, Fig. 62.6. In this thesis the voice data is collected by recording
down the chain of command. vowels A, E, I, O and U.
By using the hierarchical method, the instance values are Hierarchical Clustering Fig. 62.7 is eye-catching because
show as below: it is not needed number of Dendrograms (lines connected
Several Machine Learning Techniques Used to Forecast Parkinson Disease 423
Fig. 62.7 Hierarchical clustering dendrites showing PD and non-PD data values0-Non PD (Blue colored) and 1-PD (Red colored)
424 Algorithms in Advanced Artificial Intelligence
to cluster groups), and the clustering method can be easily 3. Results and Discussion
illustrated with a Values between PD and Non-PD patient’s
data values of the vowels (A, E, I, O, U). After applying various Data Mining classification methods,
The method of the data relevant to Fundamental Frequency, various results are observed. Values of the Table 62.1 are
the Self-Organizing Map (SOM), provides subjective data shown in bar graph Fig. 62.9 which shows the identification
towards solid and PD datasets, as illustrated in Fig. 62.8. The of PD data (%). It is observed that random forest showing
majority of the information possesses sick class. more accuracy and Naïve Bayes showing less accuracy on
the data to PD.
When using visualisation approaches, it is discovered that
a parallel coordinate, or ROC for classification algorithms,
shows this cross line for sick data over non-diseased data
(Fig. 62.3). In contrast to the other two, the majority, SVM,
and k-nearest neighbour exhibit curve values that are closer to
1. Sieve Multigram shows how features are related. Red color
indicates negative (class value 1- that indicates PD) correlation
and blue color indicates positive correlation (class value 0
that indicates Non-PD). Thickness of lines demonstrates
how solid the relationship shown in Fig. 62.3(Sieve multi
gram). The Fundamental frequency attribute’s hierarchical
clustering has revealed that there are more groups in the solid
dataset and fewer in the infected data. In Self Organised Maps
(SOM), the ranges for the healthy dataset (blue coloured)
and the diseased dataset (red coloured) for the Fundamental
frequency attribute are 95.7 to 252.45 and 144.18 to 202.63,
respectively. Qualitative information about sound and PD
Data values is provided by the visualisation and clustering of
the data related to Fundamental frequency.
The maximum values are involved in PD class. It is observed
Fig. 62.8 Self organized maps (SOM) for fundamental that most of the diseased class is also occupied non diseased
frequency attributes
class. Red color (label 1) class indicates the diseased PD
It is watched that the majority of the unhealthy class is used where as blue color (label 0) class indicates non PD.
likewise involved non ailing. Class Red shading demonstrates It seems K-star took very less time(0.1 seconds) and simple
the class mark 1; it is infected (PD) and blue shading class logistic took more time(1.98 seconds) as shown in Fig. 62.10.
shows 0 for Non-PD.
Abstract: Rice is one of the most significant staple crops, especially in India. It provides daily calories and carbohydrates to
approximately half the population. Additionally, many Indian villages grow this plant as a crop in order to make money from
exports. Every year, millions of tons of rice are shipped around the world and eaten. The Pyricularia grisea-related fungus
known as rice explosion feeds on the seedling stage of the rice plant and frequently causes significant damage every year to
decrease crop growth. Farmers are mostly to blame for the rice sickness. In order to know where to look and what to look for,
it is important to be aware of the many plant parts that are susceptible to illness before discussing diseases in paddy. The study
explores the use of data-driven methods to assess the risk of fungal diseases, highlighting their potential impact on food security
and crop devastation. In this paper, we suggest combining computer vision and machine learning to identify the fungus illness
affecting rice harvests. The fungus infection of experiments is used in the evaluation research of image classification models,
including K-NN, SVM, and CNN.
Keywords: Crop devastation, CNN, Food security, Fungal disease, K-NN, Pyricularia grisea, SVM, Risk assessment etc
1. Introduction water stress, and yield quality. Most of the time, researching
a problem requires knowledge, which might be costly and
Agriculture played a major role in the rise of sedentary human time-consuming in developing realms. In order to increase
civilization since it produced excess food for urban dwellings. crop yield, disease management in agriculture focuses on
Fisheries, aquaculture, forestry, and the production of crops controlling fungi, bacteria, viruses, viroids, phytoplasmas,
and animals are all included in agriculture, according to protozoa, nematodes, and parasitic plants [5].
Kavitha (2018) [3]. The creation of a system for sustainable
An abnormal physiological process that impedes the normal
agriculture could be considerably aided by the use of
structure of the plant’s function, growth, and other activities
current technologies. It is now possible to boost agricultural
causes a plant to become ill when a specific causative agent
inputs and farm output in a way that is both financially and
interferes with it on a regular basis. Disruption of one or
environmentally sustainable because of the growing field
more of a plant’s vital biochemical and physiological systems
of reliable farming. With the use of these techniques and
leads to pathological states and symptoms. Depending on
knowledge, agriculture can now provide farming that is
both environmentally and economically sustainable while environmental factors, the crops and varieties farmed, the
lowering expenses and mistakes. Monitoring agricultural existence and prevalence of crop diseases, and the prevalence
inputs is essential to avoid negative outcomes like decreased of a specific pathogen, these variables change over time.
production, deteriorating plant health, etc. The main issues One technique for correctly and affordably monitoring
in agriculture were fertilizers, pesticides, irrigation and agricultural data is image processing. For image processing
*
jeswanth7.kamidi@gmail.com
DOI: 10.1201/9781003529231-63
Fungal Disease Risk Assessment using Data-Driven Methods: Impacts on Food Security and Crop Devastation 427
Fig. 63.1 Images are converted into multi-dimensional matrices for comparison.
applications in agriculture, they can roughly be classified The following are the implications of further crop-yield
into two categories: those based on imaging techniques and prediction research:
those based on applications. The primary subject here is the • The design of the K-NN, SVM, and CNN image
application of image processing on farms. Rahul, 2023, has classification models to capture the time-based
demonstrated image processing to be an effective machine interdependence of illness variables
vision technology in the agricultural industry [7]. Infrared • The model showed that the yield prediction might be
hyperspectral, x-ray, and imaging methodologies are some widespread in unverified situations without a large loss
of the techniques used to map irrigated land and, more in forecast accuracy.
specifically, calculate metrics like plant indexes and tree
• When used in conjunction with the replication technique,
measures. Based on their visual characteristics and external
the model could show how much the precision of disease
appearance, all plants and fruits are identified, gathered, and
predictions and the variation in crop yields are all
examined to discover any faults. In agricultural plant and
correlated.
vegetable analysis, image processing and machine vision
is frequently exploited to handle the aforementioned tests 2.1 K-Nearest Neighbors (K-NN)
without the use of intentional techniques.
The model being trained finds the approximately nearest
neighbours to the input data, according to Bhatia 2010) [1].
2. Proposed Work Any type of input data format, in this case an image of the
Artificial intelligence and computer vision technology enable crop rice disease, is accepted. For comparison, images are
early pest, disease, or nutrient shortage identification in transformed into multi-dimensional matrices. This method
plants, which then instantly connects the affected plants to identifies the ‘K’ crop rice disease images (neighbours)
an agronomic facility for treatment. In this paper, machine that are visually closest to the input image for crop rice
learning, which uses artificial intelligence, enables the disease. A comparison between the input picture vector and
computer to operate in an independent learning manner the trained model is required to identify the neighbours in
without being explicitly provided. It is an interesting and a multidimensional plane by R.N.V. Jagan Mohan, 2012[8].
complex concept that could affect how technology advances The search comparison results can be extended to whatever
in the future. Image classification is a machine learning extent we like, depending on the specified value of K. When
application used to identify crop diseases. In this instance, comparing the input image matrix with the training data
the illness photos are being classified using SVM, K-NN, and vector, one should use the Euclidean distance to calculate the
neural networks. distance between the two points [9].
428 Algorithms in Advanced Artificial Intelligence
2.2 Support Vector Machine network for classification” and is frequently used in softmax,
logistic function, and hyperbolic tangent algorithms.
One technique that can be applied to both regression and
classification issues is support vector machines (SVM). Each of these functions takes as inputs a feature vector
SVMs are used to categorize images. Before classifying an (x) and a weight vector (w), which are merged linearly.
image with an SVM, extract its features. Edge detection, After that, they usually produce an output consisting of a
edge colour values, and even the image’s textures are a finite range (0,1) or (-1,1). Continuous functions can be
few examples of these properties. After the features have approximated by a neural network with one hidden layer and
been extracted, we may feed them into the SVM algorithm a non-linear activation function; additional layers allow for
developed by K. Dileep Kumar in 2020 [10]. When a machine the achievement of continuous decision limits [11].
processes an image, it is viewed as a two-dimensional Classification of Fungal Disease Images Using CNN:
collection of pixels. If the image is 200 pixels wide and 200
Input: The fungal disease image is a dataset consisting of
pixels high, for example, the array will be 200 x 200 x 3. The
an array of pixels with height, width, and channel values for
resolution of the image determines the array size. The first
feature extraction.
two dimensions of a picture are its width and height, while
the third dimension is made up of its RGB colour channels. Feature Extraction:
An array value with a possible range of 0 to 255 represents 1. A convolution neural network can be used to create a
the intensity of a pixel at each place. feature map.
To understand SVM, take the example we used for the KNN 1.1 Convolution (ReLu).
classifier. Consider fungus-related imagery that also shows 1.1.1 Choose a kernel with a 55 size, the same
features of a rice crop. The SVM method can be used to create depth as the input array.
a model that accurately determines whether an image depicts 1.1.2 Convolution operations should be used to
a fungus illness in a crop. Prior to testing it, we will first train obtain image features.
our model with a sizable number of images of crop rice in 1.2 (Max Pooling) Pooling.
various situations so that it may become accustomed to the
1.2.1 The dominant feature is extracted by
variety of fungi in the crop characteristics. As a result, the
reducing the spatial size of the feature map
support vector creates a decision border between these two
to 22 using the dimensionality reduction
sets of fungus data and chooses them, revealing the extreme
technique.
example of a fungus by Pin Wang, 2021 [6].
2. Continue the method described above until the fourth
2.3 Convolutional Neural Network layer, changing the channel size to one of 16, 32, 64, or
128 to extract low-level features from the image.
One of the biggest obstacles in the picture recognition space
is CNN. CNN is used for sequence data without a doubt, but Classification
it is also excellent at sifting through large amounts of visual 3. Smooth output is delivered to a feed-forward neural
data and identifying non-linear correlations. SVMs, which are network with back propagation throughout each
margin classifiers, can support a variety of classifying kernels iteration of the training stage.
[4]. When there are several class labels, SVM has trouble 4. Using the SoftMax Classification approach, a trained
predicting the class labels. While the CNN design naturally model is utilized to classify photos by identifying their
promotes parallelization, SVM is similarly challenging to dominating features. Local or widespread necrosis is
parallelize (Wenfeiiling, 2020) [12]. a common manifestation of a fungus infection. Fungi-
Kernels are not necessary because of the variations in how caused crop illnesses can either obstruct normal growth
neural networks operate. Convolutional neural networks are or lead to expansion, an uncontrolled surge of growth
an exception. A single hidden layer neural network with a by zhang SW, 2015[13].
non-linear activation function is referred to as a “neural
Table 63.1 Graph for comparing algorithms K-NN, SVM, and CNN
Algorithm Accuracy Sensitivity Specificity Precision F1-score
K-NN 0.75 0.79 0.79 0.80 0.80
SVM 0.89 0.88 0.85 0.84 0.85
CNN 0.97 0.98 0.96 0.96 0.98
Fig. 63.3 Graph for comparing algorithms K-NN, SVM, and CNN
The following bar graph displays the various accuracy fungus illnesses affecting rice harvests. The study uses image
measures for the three procedures (K-NN, SVM, and CNN), classification models like K-NN, SVM, and CNN.
including accuracy, sensitivity, specificity, precision, and F1
score. CNN achieved the greatest outcomes when comparing References
these procedures, according to Zhang (2018) [14].
1. Bhatia, N and Vandana: Survey of nearest neighbor
techniques, International Journal of Computer Science and
4. Conclusion Information Security, vol. 8, pp. 302-305, 2010.
Rice, a staple crop in India, is heavily impacted by the 2. Hasan, M., S. Ullah, M. J. Khan, and K. Khurshid:
Pyricularia grisea-related fungus, rice explosion. This Comparative analysis of SVM, ANN and CNN for classifying
fungus damages the seedling stage of rice plants, causing vegetation species using hyperspectral thermal infrared data.
International Archives of the Photogrammetric, Remote
crop damage and decreased growth. Farmers are primarily
Sensing and Spatial Information Sciences - ISPRS Archives
responsible for the disease. A study suggests using data- 42 (2/W13):1861–68, doi: 10.5194/isprs-archives-XLII
driven methods to assess fungal disease risk and identify 2-W13-1861-2019.
430 Algorithms in Advanced Artificial Intelligence
3. Kavitha B C, Shilpa D P, Thanushree K S, Swathi A M, 10. R.N.V. Jagan Mohan and R. Subbarao and Kurra Raja Sekhara
Ranjitha M K: Agricultural Crop Monitoring Sensors using Rao: Efficient K-Means Cluster Reliability on Ternary Face
IoT-A Study, International Journal of Engineering Research & Recognition using Angle Oriented Approach, Published in
Technology (IJERT) ISSN: 2278-0181,Vol-6,Issue-13,2018. International Journal of Informatics and Communication
4. Kasian Myagila & Hassan Kilavo: A Comparative Technology (IJ-ICT) Vol.2, No.1, January 2013, pp. 180-187
Study on Performance of SVM and CNN in Tanzania ISSN: 2252-8776, http://dx.doi.org/10.11591/ij-ict.v2i1.1779.
Sign Language Translation Using Image Recognition, 11. R.N.V.Jagan Mohan: Machine Learning approach for corona
Applied Artificial Intelligence, 36:1, 2005297, DOI: virus disease extrapolation: A case study, International Journal
10.1080/08839514.2021.2005297, 2022. of Knowledge-based and Intelligent Engineering Systems,
5. Mishra, S.; Sachan, R.; Rajpal, D: Deep Convolutional Neural Vol-26,219-227, ISSN: 1327-2314(print), 1875-8827(online)
Network based Detection System for Real-time Corn Plant DOI: 10.3233/KES-220015, 2022.
Disease Recognition, Procedia Computer Science, 167, 2003– 12. Rashid Agha, R. A., N. S. Al Muhammed, and P. Fattah.: A
2010, 2020. comprehensive study on sign languages recognition systems
6. Pin Wang, En Fan, eng Wang: Comparative analysis of using (SVM, KNN, CNN and ANN), ACM International
image classification algorithms based on traditional machine Conference Proceeding Series. doi:10.1145/3279996.328002
learning and deep learning, https://doi.org/10.1016/j.patrec. 4,2018.
2020.07.042, Pattern Recognition Letters, Volume 141, Pages 13. Wenfei Liu; Jingcheng Wei; Qingmin Meng: Comparisons
61-67, January 2021. on KNN, SVM, BP and the CNN for Handwritten
7. Rahul Subhash Gaikwad and Sharanabasappa C.Gandage: Digit Recognition, IEEE Xplore, DOI: 10.1109/
Image Sentiment Classification Using Deep Convolutional AEECA49918.2020.9213482, 25-27, August 2020.
Neural Network Models, Jounral of Data Acquisition and 14. Zhang S.W., Shang Y.J., Wang L: Plant disease recognition
Processing, ISSN: 1004-9037, https://sjcjycl.cn/DOI: based on plant leaf image, Journal Anim. Plant Science,
10.5281/zenodo.7923136, Vol. 38 (3), Page No: 1279-1300, 25:42–45, 2015.
2023. 15. Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification
9. R.N.V.Jagan Mohan and Uppala Narendranath Gadaee: Face of Maize Leaf Diseases Using Improved Deep Convolutional
Recognition using Unsupervised Images through Discretionary Neural Networks,IEEE Access, 6, 30370–30377,2018.
based Security, International Journal of Advanced Computer
Note: All the figures and tables in this chapter were designed by
and Mathematical Sciences, ISSN: 2230-9624, Vol 3, Issue
the author.
1, 2012, pp. 181-185, Publisher: I International, http://
bipublication.com, ICV-71.03, h-Index:25, i10:Index:99.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
D. Ratna Giri1
Associate Professor
SRKR Engineering College (A), Bhimavaram, AP
Dept of Information Technology
P. Syamala Rao2
Assistant Professor
Dept of Information Technology
SRKR Engineering College (A), Bhimavaram, AP
J. V. Rama Kumar3,
Assistant Professor
Dept of CSE
SRKR Engineering College (A), Bhimavaram, AP
JMSV Ravi Kumar4
Associate Professor
Dept of Information Technology
Abstract: Glaucoma is a group of eye diseases that can cause irreversible damage to the optic nerve, leading to eventual
blindness. It is critical to recognise the signs of the sickness promptly, as they manifest over time. Although glaucoma currently
has no cure, it is often possible to safeguard one’s eyesight and avoid further damage by intervening early. Advanced machine
learning techniques are transforming the diagnosis of glaucoma. An advanced approach for the early diagnosis and categorization
of glaucoma is the target of this research. Machine learning may greatly improve preventive healthcare, which is our main goal.
Data exploration and preprocessing, feature engineering and visualisation, and model evaluation using classifiers including
logistic regression, decision trees, random forests, and support vector machines were all part of the study to help understand
glaucoma patterns and their details. We assess the precision of the experimental result by using it in conjunction with a number
of machine learning methods.
Keywords: Decision tree, Early detection, Eye disorder, Glaucoma, Logistic regression, Random forest, Support vector
machine
1
drsrkrit@gmail.com, 2peketi.shyam@gmail.com, 3jvramakumar@gmail.com, 4jmsvravikumar@gmail.com
DOI: 10.1201/9781003529231-64
432 Algorithms in Advanced Artificial Intelligence
3. Experimental Result
Results from the experiments illuminate the dataset compo
Fig. 64.2 Process of model selection for early detection of nents used for glaucoma patient categorization according to
glaucoma several parameters. The dataset includes patient ID, age, cup
to-disc ratio, intraocular pressure (IOP), and pap smear results.
During model selection, we compare features X and target
Y to identify the best transformation F for a given training Number of the Medical Record: This dataset contains ten
dataset. Y is equal to the function F(X) minus one. “Optimal” thousand patients’ unique identifiers. The range of possible
refers to a model that maximises some performance patient IDs is from 4 to 99992.
indicator. When trying to build a model, there are a number Life expectancy: The average age of the patients is
of dimensions to think about, including the model’s approximately 53.87 years. Everyone taking part has to be
parameter space, paradigm space, hyperparameters space, between the ages of 18 and 90. The dataset shows that the
architecture space, features space, and transformation space ages of the patients vary somewhat, with a standard deviation
for features. Parameter training and optimisation can make of approximately 21.13. The level of the aqueous humour
use of statistical learning and supervised learning algorithms.
Proper feature encoding, transformations, and selection
can enhance the model’s performance. Separate metrics
are used to evaluate classification and regression, which
are two subsets of machine learning tasks; the third metric
is the model evolution metric. It is critical to know which
measurements work for which issues.
1. Measures of Machine Learning Performance: Using
training data, classification problems determine which data
sets to use for analysis. Models are able to anticipate the
labels of new data sets by learning from existing datasets
and dividing them into classes. Metrics used in performance
evaluations include yes/no, zero, and one.
1.1 Accuracy: Relative to total predictions, the percentage of
correct predictions is one of the most straightforward metrics
for categorization. This idea might be stated as
Number of Correct Preductions
Accuracy =
Total Number of Predictions
One can generate an accuracy statistic by using the scikit-learn
module or by repeatedly comparing the predicted and ground
truth values. Despite its ease of use and implementation, it
performs optimally with a balanced distribution of samples
Fig 64.3 Normal vison vs glaucoma
434 Algorithms in Advanced Artificial Intelligence
in the eye Around 17.51 mmHg is the typical intraocular deviation for patients’ cup-to-disc ratios is approximately
pressure. A normal intraocular pressure (IOP) ranges from 10 0.14, so there is room for some variation.
to 25 mm Hg. A standard deviation of around 4.36 indicates Pap smear screening: The typical pachymetry reading is
that intraocular pressure varies among people. around 549.73. In pachymetry, the numbers could range from
The typical cup-to-disc ratio, abbreviated as CDR, is around 500.01 to 599.99. The standard variance of roughly 28.90
0.55. A CDR of 0.30 to 0.80 is possible. The standard indicates that corneal thickness varies from patient to patient.
2. An G., Omodaka K., Hashimoto K., Tsuda S., Shiga Y., Takada
4. Conclusion N., Kikawa T., Yokota H., Akiba M., Nakazawa T: Glaucoma
Glaucoma is a collection of eye diseases that can cause diagnosis with machine learning based on optical coherence
blindness and vision loss. This work established a system tomography and color fondues images, Journal Healthcare
for early identification and classification of glaucoma using Engineering, 2019;1:10, DOI:10.1155/2019/4061313,2019.
3. Asaoka R., Murata H., and Iwase A., Araie M: Detecting
machine learning. The approach delves into glaucoma
preperimetric glaucoma with standard automated perimetry
patterns and details via feature engineering, data exploration,
using a deep learning classifier, Ophthalmology, 123:1974–
and model evaluation. The primary goals are to enhance 1980, Doi: 10.1016/j.ophtha.2016.05.029,2016.
preventive healthcare and to increase the reliability of 4. Asaoka R., Murata H., Hirasawa K., Fujino Y., Matsuura M.,
experimental results. Miki A., Kanamoto T., Ikeda Y., Mori K., Iwase A., et al. Using
deep learning and transfer learning to accurately diagnose
References early- onset glaucoma from macular optical coherence
tomography images, American Journal Ophthalmology,
1. Kumar, Dr Jmsv Ravi, and M. CHANDINI. “SECRBAC: 2019;198:136–145,DOI:10.1016/j.ajo.2018.10.007,2019.
Secure Data In The Clouds.” International Journal of Research 5. Barros D.M., Moura J.C., Freire C.R., Taleb A.C., Valentim
5.15 (2018): 95-106. R.A., Morais P.S: Machine learning applied to retinal
436 Algorithms in Advanced Artificial Intelligence
image processing for glaucoma detection: Review and Journal AIP Conference Proceedings, Volume 2492, Issue 1,
perspective, Biomedical Engineering, OnLine, 2020;19:1– Publisher AIP Publishing, 2023.
21,DOI:10.1186/s12938-020-00767-2,2019. 19. JMSV RAVI KUMAR” Human Activity Recognition using
6. Hashimoto Y., Asaoka R., Kiwaki T., Sugiura H., Asano Machine Learning “ Journal AIP Conference Proceedings,
S., Murata H., Fujino Y., Matsuura M., Miki A., Mori K., Volume 2492, Issue 1, Publisher AIP Publishing, 2023.
et al: Deep learning model to predict visual field in central 20. J Kumar, A Shahi, R Aytha, G Varri, D Brundavanam “ Vehicle
10° from optical coherence tomography measurement in theft prevention system using IoT “Journal AIP Conference
glaucoma, Br. J. Ophthalmology, 2020:1–7,DOI: 10.1136/ Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
bjophthalmol-2019-315600,2020. 2023.
7. Lee S.D., Lee J.H., Choi Y.G., You H.C., Kang J.H., and Jun 21. J Kumar, TD Nagendra, M Harshitha, AB Prakash “ Fake
C.H: Machine learning models based on the dimensionality image detection using CNN “Journal AIP Conference
reduction of standard automated perimetry data for glaucoma Proceedings, Volume 2492, Issue 1, Publisher AIP Publishing,
diagnosis, Artificial Intelligence Medical, 2019;94:110– 2023.
116,DOI:10.1016/j.artmed.2019.02.006,2019. 22. J Kumar, MN Kumar, NV Narendra, P Pradeep “ driver
8. Renukalatha S., Suresh K.V: Classification of glaucoma drowsiness monitoring system using machine learning svm
using simplified-multiclass support vector machine, algorithm “Journal AIP Conference Proceedings, Volume
Biomedical Engineering, 2019; 31:1950039, doi: 10.4015/ 2492, Issue 1, Publisher AIP Publishing, 2023.
S101623721950039X, 2019. 23. JMSV RAVI KUMAR “ A Symmetric Searchable Encryption
9. Sejong Oh,Yuli Park,Kyong Jin Cho,and Seong Jae Identification of Data on Probabilistic Trapdoors “International
Kim:Explainable Machine Learning Model for Glaucoma Journal of Engineering and Advanced Technology (IJEAT),
Diagnosis and Its Interpretation, Diagnostics (Basel), ISSN: 2249 – 8958, Volume 9, Issue 3, Publisher Blue Eyes
2021 Mar; 11(3): 510,PMCID: PMC8001225,PMID: Intelligence Engineering & Sciences Publication, 2020.
33805685,Published online 2021 Mar. 24. JMSV RAVI KUMAR “Artificial Bee Colony Algorithm: A
10. Wang P., Shen J., Chang R., Moloney M., Torres M., Survey and Recent Applications” published in International
Burkemper B., Jiang X., Rodger D., Varma R., Richter Journal of Pure and Applied Mathematics, ISSN 1314-3395,
G.M:Machine learning models for diagnosing glaucoma VOLUME 118, ISSUE 24 , Jul-18.
from retinal nerve fiber layer thickness maps, Ophthalmology 25. JMSV RAVI KUMAR “ Authentication for Cloud Services
Glaucoma,2:422–428,DOI:10.1016/j.ogla.2019.08.004,2019. using Steganography” published in International Journal of
11. Estharakula, Suresh, and Kumar JMSV Ravi. “EBPH-MAC: Engineering and Technology(UAE)-IJET, ISSN 2227-524X,
Emergency Based Priority Hybrid Medium Access Control VOLUME 7, ISSUE 3.49 , Jul-18.
for Mobility Aware Cooperative WSN’s In Indoor Industrial 26. JMSV RAVI KUMAR “A review on task scheduling algorithms
Monitoring.” International Journal of Research 5 (2018): in cloud computing and their approaches” published in
1456-1465. International Journal of Pure and Applied Mathematics, ISSN
12. Kumar, J. M. S. V., et al. “System Testability Assessment and 1314-3395, VOLUME 118, ISSUE 24, Jul-18.
testing with Micro architectures.” International Journal of 27. JMSV RAVI KUMAR “Review of Data mining Technique
Advanced Research in Computer Science 2.6 (2011). using SaaS on the Cloud” published in International Journal of
13. Kumar, J. M. S. V., et al. “Reverse Engineering A Generic Pure and Applied Mathematics, ISSN 1314-3395, VOLUME
Software Exploration Environment Is Made Of Object 118, ISSUE 24 , Jul-18.
Oriented Frame Work And Set Of Customizable Tools.” 28. JMSV RAVI KUMAR “Smart Controlling, Monitoring
International Journal of Advanced Research in Computer and Automation of Street Light System using Raspberry
Science 2.5 (2011). PI “ published in International Journal of Pure and Applied
14. Kumar, J. M. S. V., et al. “Analyzing the Modern Tool- Mathematics, ISSN 1314-3395, VOLUME 118, ISSUE 24 ,
Supported UML-Based Static Reverse Engineering.” Jul-18.
International Journal of Advanced Research in Computer 29. JMSV RAVI KUMAR “ A Survey on Internet of Things for
Science 3.4 (2012). Healthcare and Medication Management” was authored by
15. Kumar, J. M. S. V., et al. “Active Scrutiny Techniques for the JMSV Ravi Kumar published in International Journal of Pure
Reconstruction of Architectural Views.” International Journal and Applied Mathematics, ISSN 1314-3395, VOLUME 118,
of Advanced Research in Computer Science 3.1 (2012). ISSUE 24 , Jul-18.
16. N Santha Raju, JMSV Kumar, B Sujatha,”Time series analysis 30. JMSV RAVI KUMAR “ SECRBAC: Secure Data in the
of stock price movements: Insights from data mining using Clouds” was authored by JMSV Ravi Kumar published in
machine learning”, journal AIP Conference Proceedings, International Journal of Research, ISSN 2348-6848, VOL 5,
Volume 2492, Issue1, Publisher AIP Publishing,2023. ISSUE 15 , Jul-18.
17. Prayaga Atchyut Pavan, Sattibabu Sattibabu, JMSV Kumar 31. JMSV RAVI KUMAR “ EBPH MAC: Emergency Based
“A deep learning approach to detect malaria “Journal AIP Priority Hybrid Medium Access Control for Mobility
Conference Proceedings, Volume 2492, Issue 1, Publisher AIP Aware Cooperative WSN’s In Indoor Industrial Monitoring”
Publishing, 2023. published in International Journal of Research, ISSN 2348
18. Ch Bhanu Revathi, JMSV Kumar, B Sujatha” Intracranial 6848, VOLUME 5, ISSUE 12 , Jul-18.
hemorrhage detection in human brain using deep learning “
Redefining Glaucoma Identification using State-of-the- Art Machine Learning 437
32. JMSV RAVI KUMAR “ Prioritizing software components for Demand and Supply,” International Journal of Research In
realistic reuse” published in International Journal of Sciences Science & Engineering (IJRISE), vol. 3, issue 1, pp. 9–23,
& Applied Research, ISSN 2394-2401, VOL 4, ISSUE 24, 2023.
Jul-17. 45. M. Srikanth, “Smallholder Farmers Crop Registering Privacy-
33. JMSV RAVI KUMAR “ Cloud Storage Services and Privacy Preserving Query Processing over Ethereum Blockchain,”
Protection” published in International Conference on Research Journal of Pharmaceutical Negative Results, vol. 13, issue 7,
Advancements in Computer Science and Communication, pp. 5609-5617, Dec. 2022. [Scopus]
ISSN 978-93-85100- 64-2, VOL 5, ISSUE 3.49, December-16. 46. M. Srikanth, “The Early Detection of Alzheimer’s Illness
34. JMSV RAVI KUMAR “Analyzing the Modern Tool- Using Machine Learning and Deep Learning Algorithms,”
Supported UML-Based Static Reverse Engineering” published Journal of Pharmaceutical Negative Results, vol. 13, issue 9,
in International Journal of Advanced Scientific Research and pp. 4852-4859, Nov. 2022. [Scopus]
Technology, ISSN 0976-5697, VOL 3, ISSUE 4, Jul-12. 47. M. Srikanth, “Small Holders Farming Predictive Analysis
35. JMSV RAVI KUMAR “Active Scrutiny Techniques for Using Peer-To-Peer Approach,” International Journal of
the Reconstruction of Architectural Views” published in Agriculture and Animal Production, vol. 2, issue 05, pp. 26
International Journal of Advanced Scientific Research and 37, Sep. 2022.
Technology, ISSN 0976-5697, VOL 3, ISSUE 1, January-12. 48. M. Srikanth, “Using Machine Learning and Neural Networks
36. JMSV RAVI KUMAR “System Testability Assessment and Technologies, a Bottom-Up Water Process Is Being Used To
testing with Micro architectures” published in International Reduce All Water Pollution Diseases,” Journal of Artificial
Journal of Advanced Scientific Research and Technology, Intelligence, Machine Learning and Neural Network
ISSN 0976-5697, VOL 2, ISSUE 6, December-11. (JAIMLNN), vol. 2, Oct. 2022.
37. JMSV RAVI KUMAR “Reverse Engineering A Generic 49. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers
Software Exploration Environment is made of Object- Crop Transaction Using Peer-to-Peer,” Indo-American Journal
Oriented Frame Work and Set of Customizable Tools” of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp.
published in International Journal of Advanced Scientific 33-43, Sep. 2022.
Research and Technology, ISSN 0976-5697, VOL 2, ISSUE 50. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care
5, September-2011. Centres Use a Hybrid Technique Based on a Distribution
38. M. Srikanth, “Integrated Technologies for Proactive Bridge- Network,” International Journal of Health Sciences, Jun.
Related Suicide Prevention”, Journal of Namibian Studies, 2022. [Scopus]
Volume 1, Issue 33, Pages 2117-2136, ISSN: 1863-5954, Sep 51. M. Srikanth, “Blockchain-Based Crop Farming Application
2023. [Scopus] Using Peer-to-Peer,” Journal of Xidian University, Apr. 2022.
39. M. Srikanth, “Deep Learning Approaches for Predictive 52. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
Modeling and Optimization of Metabolic Fluxes in Engineered Emotional Recognition Using Machine Learning, Query
Microorganism” International Journal of Research in Science Optimization and Blockchain Technology,” Solid State
&Amp; Engineering (IJRISE) ISSN: 2394-8299, 3(05), 1–11. Technology, Vol. 63 No. 6 (2020) [Scopus]
https://doi.org/10.55529/ijrise.35.1.11, July 2023. 53. M. Srikanth, “Machine Learning for Query Processing System
40. M. Srikanth, “Tackling Outliers for Predictive Smallholder and Query Response Time Using Hadoop,” IJMTST, Aug.
Farming Analysis,” in Proceedings of the 2023 3rd International 2020.
Conference on Smart Data Intelligence (ICSMDI), pp. 93-98, 54. M. Srikanth, “Block-level Based Query Data Access Service
IEEE Xplore, March 26, 2023. [Scopus] Availability for Query Process System,” IEEE, Page 1-9, Jul.
41. M. Srikanth, “Blockchain-Based Consensus For A Secure 2020. [Scopus]
Smart Agriculture Supply Chain,” European Chemical 55. M. Srikanth, “Query Response Time in Blockchain Using
Bulletin, vol. 12, special issue 4, pp. 8669-8678, 2023. Big Query Optimization,” The Role of IoT and Blockchain
[Online]. Available: doi: 10.48047/ecb/2023.12.si4.776.ISSN: Techniques and Applications from Computer Science and
2063-5346, 2023. [Scopus] Information Management, Apple Academic Press, Exclusive
42. M. Srikanth, “Predict Early Pneumonitis in Health Care Using Worldwide distribution by CRC Press Taylor & Francis
Hybrid Model Algorithms,” Journal of Artificial Intelligence, Group, Jan. 2022. [Scopus]
Machine Learning and Neural Network (JAIMLNN), vol. 3, 56. M. Srikanth, “A New Approach for Authorship Verification
issue 03, pp. 14-26,ISSN: 2799-1172, Apr. 2023. Using Information Retrieval Features,” Springer-ICSE, vol.
43. M. Srikanth, R. N. V. Jagan Mohan, M. Chandra Naik. (2023). 74, pp. 23-29. [Scopus]
A New Way to Improve Crop Quality and Protect the Supply 57. M. Srikanth, “An Enhanced and Naive Clustering Algorithm
Chain is to use a Trajectory Network and Game Theory. for Text Classification Based on Weight,” International Journal
Mathematical Statistician and Engineering Applications, & Magazine of Engineering, Technology, Management and
71(4), 10600–10610. https://doi.org/10.17762/msea. Research, Dec. 2012.
v71i4.1952, ISSN: 2094-0343, 2023 [Scopus]
Note: All the figures in this chapter were designed by the author.
44. M. Srikanth, “Auction Algorithm: Peer-To-Peer System Based
on Hybrid Technologies for Smallholder Farmers to Control
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
438 Algorithms in Advanced Artificial Intelligence
Abstract: Investors are looking for methods to profit from artificial intelligence’s developing capabilities, especially in
emerging technologies, due to its increasing importance in daily life and the economy. Artificial intelligence (AI) investment
is a part of the digital revolution since it automates jobs that used to need human intelligence, opening doors for investors to
profit from the economy’s expected expansion. Python has made a huge splash in the software industry, but getting a handle
on how it works is essential. To help investors identify the sweet spot in machine learning development for speed, model size,
and performance, the feature selection probe approach is available. Discovery, management, marketing, state, and profit for an
entire fiscal year are revealed via an experimental dataset. The probe method is a reliable feature selection technique used in
multi-regression analysis of economic data.
Keywords: Artificial intelligence, Feature selection, Machine learning, Python, Probe method
1
prabhakar1.kandula@gmail.com, 2psrinu.cse@swarnandhra.ac.in
DOI: 10.1201/9781003529231-65
Probe Method: A Dependable Economy Data Methodology Feature Selection for Machine Learning 439
Investors may invest in companies developing AI, hardware 5. Rich Standard Library: Python’s extensive standard
companies, software companies, or those benefiting from library, encompassing modules for web development
its wider adoption by Raghupathi,2021[9]. For instance, in and data analysis, makes it a popular choice for various
the personal computer industry, investors could invest in applications.
computer manufacturers, hardware companies, software 6. Community and Packages: Python’s community is
companies, and automation companies. Investments in thriving, and its third-party package ecosystem, the
computers and technology have been made, with some Python Package Index (PyPI), includes thousands of
being direct bets and others more conservative. As AI may libraries to enhance its capabilities.
displace workers, companies focused on worker retraining 7. Cross-Platform: Python’s cross-platform nature
may benefit. Some stocks may match these criteria for AI allows for the development of applications that can
investment by SrikarSoma, 2023[10]. run on various operating systems without significant
AI-generated art allows users to create images based on their modifications.
descriptions, utilizing images from around the world by 8. Versatile: Python is a versatile tool that can be utilized
Simon, 2017[11]. This technology has been used by people for web development, data analysis, machine learning,
of all ages and backgrounds but concerns about copyright scripting, and more.
arise as artists feel their livelihoods are at risk. Public
Python’s simplicity, readability, and versatility make it a
companies have vast collections of AI-generated artwork.
popular choice for developers, offering ease of learning and
Startup companies, often in promising fields like artificial
quick prototyping, making it an excellent choice for both
intelligence and machine learning, are initially capitalized by
beginners and seasoned programmers.
venture capital investors and then publicly raised to expand
operations and customer base. Successful companies have
well-received early investors by Ward, 2014[12]. 3. A Reliable Feature Selection
The following is how the paper is set up: Section -1 deals with Method for Machine Learning (ML)
Introduction. In Section-2, understanding machine learning is the Probe Method
is provided. Section-3 covers a Reliable Feature Selection
Method for Machine Learning (ML) is the Probe Method. The method to investing in businesses finding the right ratio
Section-4 deals with Multiple Regression analysis. Section-5 of speed, model size, and performance when using ML
deals with experimental result and they came to conclusion development in reality. Using feature selection is a popular
and in section 6, Section-7 deals with references are made. method for enhancing speed, reducing size, and maintaining
(or slightly deteriorating) performance. Utilizing featured
selection does this. In this case, the “Probe Method” to
2. Understanding the Machine be quite helpful of any kind of application by Nicholas
Language Pudjihartono, 2022[8].
Python, a popular programming language, has significantly The following framework shows how it functions:
impacted the tech industry. However, it’s crucial to understand 1. Add a random feature (noise).
its functionality. 2. Use the fresh dataset to train a model.
1. Interpreted Language: Python code is interpreted at 3. Calculate the value of a characteristic.
runtime, not compiled into machine code like C++, 4. Remove any original features that trail the random
allowing the user to write code and have the interpreter feature in importance.
execute it line by line.
5. Repeat until convergence.
2. High-Level Language: Python is a user-friendly
programming language that simplifies complex tasks This also makes usual brain intelligence. A feature may be
by abstracting low-level details, making it accessible to useless for the model if its relevance is lower than that of a
both beginners and experienced developers. random (noise) feature.
3. Dynamically Typed: Python is dynamically typed,
allowing for flexible and concise code by allowing the 4. Multiple Regression Analysis
interpreter to determine variable types at runtime. In the same way as basic regression fits data to a linear
4. Indentation Matters: Python uses indentation to equation, multiple linear regression models the feature-
define code blocks, ensuring clean and readable code, response relationship. According to Orogun Okunola
but requires careful attention to whitespace.
440 Algorithms in Advanced Artificial Intelligence
Adebola (2021) [7], it assesses the variables and correlations Importing Dataset: The code for importing the dataset
that influence the anticipated result. (50_CompList) containing all variables is provided.
Y = b0 + b1 * x1 + b2 * x2 + b3 * x3 + ...... bn * xn (1) #importing datasets
data_set=pd.read_csv(‘50_CompList.csv’)
Y = Dependent variable and x1, x2, x3, ..., xn
Output: The dataset will be obtained as follows:
= multiple independent variables.
5. Experimental Result
With the goal of identifying maximum profit and affecting
factors, a dataset of startup investment amounts discloses
R&D, administration, marketing, state, and profit for a
financial year. Profit is used as the dependent variable in
the multiple linear regression model, with the other four
variables being considered independent (Yoshita Chakrabort,
2023 [13]).
1. Preparing data for analysis.
2. Applying the MLR model to the data set used for
training.
3. Foretelling how the test set will turn out.
Step-1: Data Pre-Processing: The first step of data pre
processing, which includes the following steps: The library
will be imported to assist in building the model, as illustrated
in the provided code. The output reveals five variables, four of which are
continuous, and one of which is a categorical variable. The
#importing libraries
process involves identifying both dependent and independent
import numpy as nm variables.
import matplotlib.pyplot as mtp
import pandas as pd
Probe Method: A Dependable Economy Data Methodology Feature Selection for Machine Learning 441
(categorical_features = [3])
x= onehotencoder.fit_transform(x).toarray()
The encoding process only involves one independent variable,
which is the state, while the other variables are continuous.
Output:
set. Divide the dataset in half to use for training and testing. The model has been successfully trained using the training
Using sklearn.model_selection, bring in test_and_train split. dataset and will be tested on the test dataset in the next step.
The function split_train_test is used to split the data into Step: 3- Prediction of Test set results: The final step for our
training and testing sets. It takes the parameters y_train, x_ model involves evaluating its performance by predicting the
test, y_train, y_testposition, orientation, test size=0.2, and test set result using a y pred vector, as per the code provided.
random state=0. #Predicting the Test set result;
The code will partition our dataset into a training set and a y_pred regressor. Predict (x_test)
test set.Results: You can see the dataset split into a training
set and a test set in Spyder IDE’s variable explorer. Executing the code generates a new vector under variable
explorer, allowing us to test our model by comparing
predicted and test set values.
Output:
Training set:
Abstract: In order to arrive at an accurate estimate of the average human life expectancy, one must consider several factors
including heredity, environmental influences, lifestyle choices, and access to healthcare. Because of their reliance on sparse
data, traditional approaches may fail to adequately represent the complex interrelationships among these components. This
research presents a new method that utilizes machine learning models, population-based optimization, and sentiment analysis
to estimate life expectancy. This method utilizes population-based optimization strategies, machine learning algorithms, and
massive amounts of text data to produce more accurate and trustworthy life expectancy estimates.
Keywords: Life expectancy estimation, Sentiment analysis, Population-based optimization, Machine learning, Data-driven
approach, Precision healthcare
DOI: 10.1201/9781003529231-66
Estimating Human Life Expectancy through Sentiment Analysis, Population-based Optimisation, and Machine Learning Models 445
methods and deliver more precise and illuminating life The article “Significance of Machine Learning in Healthcare:
expectancy estimations by utilizing large-scale text data, Features, Pillars, and Applications” (2022), written by S. K.
population-based optimization techniques, and machine Yadav et al. Machine learning has several potential uses in
learning algorithms. the medical field, and this article covers them all. The authors
Strategy for Health Care: In order to meet the unique go into how machine learning could enhance the precision of
healthcare requirements of a community, precise life diagnoses, treatment plans, and overall patient results.
expectancy estimations are necessary for healthcare planning Publication: “Life Expectancy Estimation using Social
and resource allocation. Media Data and Machine Learning Techniques” (2023) by
Knowing their life expectancy empowers individuals to make Mohammad Arif et al. In order to estimate life expectancy
educated decisions about their diet, medical treatment, and from social media data and machine learning techniques,
savings, enhancing their health and happiness. this research presents a framework. When compared to more
conventional approaches that use demographic data, the
Pension and social security program design, as well as authors’ suggested strategy fared better.
retirement preparation, can all benefit from life expectancy
predictions in the realm of social policy.
Research and Development: The field of aging biology, as
well as attempts to prevent diseases and create treatments,
can benefit from accurate life expectancy estimations.
Healthcare planning, individual well-being, social policy,
and research and development efforts might all benefit from
this study’s more precise and informative life expectancy
estimates, which would overcome the shortcomings of
previous approaches.
2. Literature Review
“Life Expectancy Prediction through Analysis of
Immunization and HDI Factors Using Machine Learning
Regression Algorithms” (2021) by the team of Sumit Singh.
Using indicators like the Human Development Index (HDI)
and vaccination rates, this study investigates the feasibility
of using machine learning regression methods to estimate
future life expectancy. When compared to more conventional
approaches, the authors’ suggested strategy proved to be
more accurate. Fig. 66.1 (a) Strategy of data (b) Data pre-processing
Written by Muhammad Bilal et al., “An efficient sentiment
analysis methodology based on long short-term memory
networks” (2021). Using Long Short-Term Memory (LSTM) 3. Proposed Methods
networks, this research suggests a powerful approach to Gathering data and cleaning it up Researchers can compile
sentiment analysis. When compared to more conventional large-scale text data by utilizing a variety of sources,
sentiment analysis techniques, the authors’ suggested method including social media posts, news articles, medical records,
fared better. and public health reports.
Mohamed El-Kenawy et al. published the article Data Cleaning: Eliminate extraneous information,
“Hyperparameter Tuning for Machine Learning Algorithms inconsistencies, and noise from the text data through pre
Used for Arabic Sentiment Analysis” in 2022. In this processing.
study, we look at how to optimize the hyperparameters
Get useful information out of the text data by extracting
of machine learning algorithms that analyze Arabic
features like sentiment ratings, keywords, and linguistic
sentiment using techniques based on population-based
characteristics. Evaluating Public Opinion To determine if a
optimization. Compared to more conventional approaches
text fragment is favorable, negative, or neutral, you can train
to hyperparameter tweaking, the authors’ suggested strategy
a sentiment analysis model with labeled text data. Based on
outperformed the competition.
the results of the sentiment analysis model, determine the
446 Algorithms in Advanced Artificial Intelligence
sentiment ratings for every text fragment. To get a general records. The quality and consistency of the data are ensured
picture of how people feel about a certain topic, you can use by preprocessing and cleaning, and the features that are
sentiment aggregation, which involves adding up sentiment useful for machine learning models are extracted through
ratings from several text sources. sentiment analysis and feature extraction. To maximize the
Efficient Optimization for Population-Based Algorithms models’ parameters for life expectancy estimation, we use
Develop a machine learning model to estimate a person’s techniques that draw inspiration from natural selection,
lifespan based on their current age, gender, socioeconomic known as population optimization.
position, and retrieved sentiment ratings. Optimization Predicting life expectancy using extracted data and sentiment
technique: To optimize the machine learning model’s ratings is possible with the use of several machine learning
parameters, use a population-based optimization technique methods, including neural networks, logistic regression,
like a genetic algorithm or particle swarm optimization. and linear regression. To improve the model and enhance
Tuning the settings: Run the optimization method again and its performance, we must evaluate and fine-tune it. After the
again until you discover the settings that make the machine models have been tested and improved, they can be put into
learning model’s predictions as accurate as possible. a production setting to be used in the real world. Maintaining
the models’ accuracy and relevance over time requires
Training ML Models: With the optimum parameters and a continuous monitoring and development.
representative training dataset of individuals whose life
expectancy is known, train the ML model. Improving accuracy, gaining holistic insights, making
individualized forecasts, analyzing populations, and allocating
Evaluate the trained model’s performance on a separate test healthcare resources are all advantages of using machine
dataset to see how well it generalizes and how accurate its learning and sentiment analysis for life expectancy estimation.
predictions are. Deploying the Trained and Evaluated Model: Better life expectancy predictions, better healthcare policy
Put the trained and evaluated model into production to and intervention, and improved individual health can all result
forecast the lifespan of new people or groups. from using these methods. Machine learning and sentiment
Analyze and Integrate Sentiment-Informed Prediction: Add analysis have the potential to improve our understanding
sentiment ratings to the ML model to make it more accurate of the factors that affect the human lifespan. This, in turn,
by factoring in people’s subjective feelings and opinions. could lead to more accurate and dependable estimates, which
Population-Based Analysis: Examine the projected lifespans could have far-reaching implications for healthcare planning,
of various demographics to spot trends and patterns linked to personal wellness, and social policy.
socioeconomic status, personal lifestyle choices, and access
Table 66.1 Machine Learning and Sentiment Analysis for
to healthcare. To help people make better decisions and alter
Better Life Expectancy Estimation
their lifestyles, we may provide them with life expectancy
predictions based on their unique characteristics and shed Attribute Feature Value
light on the elements that affect their lifespan. Age Numerical 35
Gender Categorical Male
3.1 Machine Learning and Sentiment Analysis
Socioeconomic Status Categorical Middle
for Better Life Expectancy Estimation
Sentiment Score Numerical 0.75
By delving into the myriad aspects that affect people’s
Keywords Textual “healthy lifestyle”,
lifespans, machine learning and sentiment analysis have the
“exercise”, “nutrition”
potential to completely transform the way life expectancy
Linguistic Features Textual “positive language”,
is estimated. The procedure entails collecting detailed
“optimism”, “well-being”
information from a variety of resources, including public
Life Expectancy Numerical 78
health reports, social media posts, news stories, and medical
Mean Absolute Error: 6.8999999999999915 3.2 Making Reliable Life Expectancy Predictions
Predicted Life Expectancy: 78.17142857142856 by Combining Sentiment Analysis with
Machine Learning
A life expectancy prediction system that uses sentiment
analysis and machine learning involves several steps. These
include data collection, preprocessing, feature extraction,
population-based optimization, model evaluation, model
deployment, continuous monitoring and improvement, and
ethical considerations. We collect data from various sources,
such as social media posts, news articles, medical records,
and public health reports. Preprocessing removes noise,
inconsistencies, and irrelevant information, while feature
extraction extracts meaningful features like sentiment scores,
keywords, linguistic features, and demographic information.
Population-based optimization algorithms refine the model
parameters and evaluate the prediction accuracy of each
solution. The model is then divided into training and testing
sets and evaluated using metrics like mean absolute error
or root mean squared error. After deployment, the model’s
performance is continuously monitored and improved in the
Fig. 66.3 Correlation matrix production environment. Ethical considerations include data
448 Algorithms in Advanced Artificial Intelligence
Parameter Feature description Value • Let Sentiment (x) represent the sentiment score for text
x.
Smoking Status Categorical smoking status No
(Yes/No) Feature-Based Sentiment = Sentiment (x)
Alcohol Categorical alcohol Rare
Consumption consumption (Frequent/ 4.3 Fitness Function
Moderate/Rare/None) Mean Absolute Error (MAE):
Exercise Habits Categorical exercise habits Regular • Evaluate the accuracy of life expectancy forecasts.
(Regular/Occasional/Rare/
 yi - ŷi
None) n
techniques and optimization. The model is then trained 11. M. Srikanth, “Predict Early Pneumonitis in Health Care Using
using the extracted variables and life expectancy data, and Hybrid Model Algorithms,” Journal of Artificial Intelligence,
its performance is continuously monitored for improvement. Machine Learning and Neural Network (JAIMLNN), vol. 3,
issue 03, pp. 14-26,ISSN: 2799-1172, Apr. 2023.
12. M. Srikanth, R. N. V. Jagan Mohan, M. Chandra Naik. (2023).
5. Conclusion A New Way to Improve Crop Quality and Protect the Supply
Chain is to use a Trajectory Network and Game Theory.
This research presented a new approach to calculating the
Mathematical Statistician and Engineering Applications,
average lifespan of a human being by integrating machine 71(4), 10600–10610. https://doi.org/10.17762/msea.
learning models, sentiment analysis, and population- v71i4.1952, ISSN: 2094-0343, 2023 [Scopus]
based optimisation. The suggested strategy incorporates 13. M. Srikanth, “Auction Algorithm: Peer-To-Peer System Based
sentiment analysis to circumvent the shortcomings of on Hybrid Technologies for Smallholder Farmers to Control
conventional methods that depend exclusively on lifestyle Demand and Supply,” International Journal of Research In
and demographic data. By capturing subjective attitudes and Science & Engineering (IJRISE), vol. 3, issue 1, pp. 9–23,
opinions, sentiment analysis provides a more comprehensive 2023.
knowledge of the elements that impact longevity. To optimise 14. M. Srikanth, “Smallholder Farmers Crop Registering Privacy-
the machine learning models for accurate predictions, Preserving Query Processing over Ethereum Blockchain,”
Journal of Pharmaceutical Negative Results, vol. 13, issue 7,
optimisation methods based on populations are used.
pp. 5609-5617, Dec. 2022. [Scopus]
15. M. Srikanth, “The Early Detection of Alzheimer’s Illness
References Using Machine Learning and Deep Learning Algorithms,”
Journal of Pharmaceutical Negative Results, vol. 13, issue 9,
1. Sumit Singh, “Life Expectancy Prediction through Analysis pp. 4852-4859, Nov. 2022. [Scopus]
of Immunization and HDI Factors Using Machine Learning 16. M. Srikanth, “Small Holders Farming Predictive Analysis
Regression Algorithms,” 2021. Using Peer-To-Peer Approach,” International Journal of
2. Muhammad Bilal et al., “An efficient sentiment analysis Agriculture and Animal Production, vol. 2, issue 05, pp. 26
methodology based on long short-term memory networks,” 37, Sep. 2022.
2021. 17. M. Srikanth, “Using Machine Learning and Neural Networks
3. Mohamed El-Kenawy et al., “Hyperparameter Tuning for Technologies, a Bottom-Up Water Process Is Being Used To
Machine Learning Algorithms Used for Arabic Sentiment Reduce All Water Pollution Diseases,” Journal of Artificial
Analysis,” 2022. Intelligence, Machine Learning and Neural Network
4. S. K. Yadav et al., “Significance of Machine Learning in (JAIMLNN), vol. 2, Oct. 2022.
Healthcare: Features, Pillars, and Applications,” 2022. 18. M. Srikanth, “Blockchain Enable for Smallholder’s Farmers
5. Mohammad Arif et al., “Life Expectancy Estimation using Crop Transaction Using Peer-to-Peer,” Indo-American Journal
Social Media Data and Machine Learning Techniques,” 2023. of Agricultural and Veterinary Sciences, vol. 10, issue 3, pp.
6. Meduri Raghu Chandra et al., “Estimating Human Life 33-43, Sep. 2022.
Expectancy through Sentiment Analysis, Population-Based 19. M. Srikanth, “Protecting Tribal Peoples Nearby Patient Care
Optimisation, and Machine Learning Models,” 2023. Centres Use a Hybrid Technique Based on a Distribution
7. M. Srikanth, “Integrated Technologies for Proactive Bridge- Network,” International Journal of Health Sciences, Jun.
Related Suicide Prevention”, Journal of Namibian Studies, 2022. [Scopus]
Volume 1, Issue 33, Pages 2117-2136, ISSN: 1863-5954, Sep 20. M. Srikanth, “Blockchain-Based Crop Farming Application
2023. [Scopus] Using Peer-to-Peer,” Journal of Xidian University, Apr. 2022.
8. M. Srikanth, “Deep Learning Approaches for Predictive 21. M. Srikanth, “Stop Spread Corona Based on Voice, Face and
Modeling and Optimization of Metabolic Fluxes in Engineered Emotional Recognition Using Machine Learning, Query
Microorganism” International Journal of Research in Science Optimization and Blockchain Technology,” Solid State
&Amp; Engineering (IJRISE) ISSN: 2394-8299, 3(05), 1–11. Technology, Vol. 63 No. 6 (2020) [Scopus]
https://doi.org/10.55529/ijrise.35.1.11, July 2023. 22. M. Srikanth, “Machine Learning for Query Processing System
9. M. Srikanth, “Tackling Outliers for Predictive Smallholder and Query Response Time Using Hadoop,” IJMTST, Aug.
Farming Analysis,” in Proceedings of the 2023 3rd International 2020.
Conference on Smart Data Intelligence (ICSMDI), pp. 93-98, 23. M. Srikanth, “Block-level Based Query Data Access Service
IEEE Xplore, March 26, 2023. [Scopus] Availability for Query Process System,” IEEE, Page 1-9, Jul.
10. M. Srikanth, “Blockchain-Based Consensus For A Secure 2020. [Scopus]
Smart Agriculture Supply Chain,” European Chemical 24. M. Srikanth, “Query Response Time in Blockchain Using
Bulletin, vol. 12, special issue 4, pp. 8669-8678, 2023. Big Query Optimization,” The Role of IoT and Blockchain
[Online]. Available: doi: 10.48047/ecb/2023.12.si4.776.ISSN: Techniques and Applications from Computer Science and
2063-5346, 2023. [Scopus] Information Management, Apple Academic Press, Exclusive
Estimating Human Life Expectancy through Sentiment Analysis, Population-based Optimisation, and Machine Learning Models 451
Worldwide distribution by CRC Press Taylor & Francis Predicting Life Expectancy: A Review of Recent Trends and
Group, Jan. 2022. [Scopus] Challenges,” 2023.
25. M. Srikanth, “A New Approach for Authorship Verification 31. Ruchika Verma et al., “A Hybrid Machine Learning Model
Using Information Retrieval Features,” Springer-ICSE, vol. for Predicting Life Expectancy Based on Demographic and
74, pp. 23-29. [Scopus] Lifestyle Factors,” 2023.
26. M. Srikanth, “An Enhanced and Naive Clustering Algorithm 32. Mohamed Medhat et al., “An Ensemble Machine Learning
for Text Classification Based on Weight,” International Journal Model for Predicting Life Expectancy: A Case Study of the
& Magazine of Engineering, Technology, Management and Egyptian Population,” 2023.
Research, Dec. 2012. 33. Areej Al-Hajri et al., “A Comparison of Machine Learning
27. Yasser Al-Khateeb et al., “A Comparative Study of Machine Techniques for Predicting Life Expectancy in Oman,” 2023.
Learning Techniques for Life Expectancy Prediction Using 34. Mohamed A. Gabr et al., “A Machine Learning Approach for
Socioeconomic and Demographic Data,” 2022. Predicting Life Expectancy in Saudi Arabia,” 2023.
28. Md. Shakhawat Hussain et al., “A Novel Approach for 35. Muhammad Irfan et al., “A Machine Learning Model for
Life Expectancy Prediction Using Machine Learning and Predicting Life Expectancy in Pakistan,” 2023.
Geospatial Data,” 2022. 36. Amjad Ali et al., “A Comparative Analysis of Machine
29. Abhinav Mittal et al., “A Comprehensive Review of Machine Learning Techniques for Predicting Life Expectancy in
Learning Techniques for Life Expectancy Prediction,” 2023. Bangladesh,” 2023.
30. Abeer Al-Hashemi et al., “The Role of Machine Learning in
Note: All the figures and tables in this chapter were designed by
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
452 Algorithms in Advanced Artificial Intelligence
A Distributed-Back Propagation Procedure
that uses Climate while Predicting the
Spread of Mosquitoes Using Least 67
Squares Estimation
K. Gopala Varma1
Research Scholar, Department of CSE,
GIET University, Odhisha
M. Chandra Naik
Professor, Department of CSE,
GIET University, Odhisha
R. N. V. Jagan Mohan
Associate Professor, Department of CSE,
Sagi Rama Krishnam Raju Engineering College (A),
Abstract: Climate change’s effects on Aedes aegypti mosquitoes Temperature, precipitation, and the oceanic Nio Index
fluctuations were all measured on a monthly basis. The weather changes during the El Nio and La Nia phases. The Oceanic
Nio index measures whether the tropical waters of the Pacific Ocean are warmer or cooler than usual. In this study, we looked
at how these climate characteristics interact to predict mosquito activity during El Nio episodes. The results show that a higher
incidence of mosquitoes is associated with rainfall that is more frequent in June. After the development and implementation of
an artificial intelligence algorithm, we will be able to analyze and classify the patients in different villages using trajectories
using the distributed back propagation algorithm for the whole dataset effectively with the help of optimal patient and value
classification. The results can help with programme planning and scheduling to stop the spread of mosquito-borne illnesses
like dengue. Thus, this work is targeted at predicting micro-level parameters to control mosquito spreading activity. There is a
substantial correlation between platelet count and fever in the data; it is possible to predict whether a patient will test positive
or negative for dengue using least squares regression and logistic regression, which is then used to test the patient. The paper
is working on a distributed-back propagation method that uses climate while predicting the spread of mosquitoes using Least
Square Estimation.
Keywords: Artificial intelligence, Aedes aegypti mosquitoes, Distributed back propagation algorithm, Micro-level parameters,
Optimal points and values
1
gopalavarma.kosuri@giet.edu, 2srichandra2007@gmail.com, 3mohanrnvj@gmail.com
DOI: 10.1201/9781003529231-67
A Distributed-Back Propagation Procedure that uses Climate while Predicting the Spread of Mosquitoes 453
variables, according to research, although the mechanism The design framework utilizes a feature extractor to create
underlying this association is still unclear [8]. Hence, public a trajectory network system using the back propagation
health authorities required to control the activity would approach.
need micro-level parameters to be able to plan for mosquito The data operations for each patient’s results are inputted
control efforts if it were possible to predict mosquito seasonal from different villages.
trends using climate and meteorological data. The population
of mosquitoes that carry dengue is correlated with rainfall, air The study utilizes an AI-based algorithm for data classification
quality, and temperature. A rise in the three variables, i.e., a of dengue, focusing on optimal points and values.
recent study found that temperature, rains, and air warming Peoples Optimal Patients and Values: The vector
have all contributed to the increase in mosquito populations optimization is to take patients dengue, and data values from
in India over the past six months. The results can help with the entire set of data are taken into account. The group of
programme planning and scheduling to stop the spread of realistic points’ objective values is reflected.
mosquito-borne illnesses like dengue by Shamimul Hasan,
2016 [10]. So, this work is targeted at predicting micro-level
parameters to control mosquito spreading activity.
{ | $x Œ t,hf (x)(x) £= 0,0, ii == 1,1, 2,2, 3,……, p, n,} Õ R
Obj = f0(x) i
i
q
(1)
The advent of satellites has significantly transformed global Consider how the various patients from the input data
communication. Satellite communication benefits humanity are represented as the set of achievable objective values
in many ways by providing a variety of communication divided by the set of viable values. Choose one value from
services, such as television transmission, digital data for the collection of objective values, i.e., the optimal value,
business, telephone, and mobile communication. The which may be modeled as the desired value, t. Eliminate
impending deployment of satellite communication systems additional values that might be achieved by optimization. A
for speech and fax transmission to aero planes on international point x* is optimal if and only if it is feasible, and the set f0
routes may not come as a shock to the global population. (x) + K (where K is the proper cone) might be seen as the
GPS navigation, international telephone, multimedia video, group of values worse than, or equal to, f0(x). Therefore, the
internet access, Earth imaging, telemedicine, and tele requirement specifies that every possible value is contained
education services are just a few of the uses for satellite in the set given by Stephen Boyd, Convex Optimization, 2004
communication that are available. Researchers have studied [11].
the impact of climate change on Aedes aegypti mosquitoes.
They measured monthly variations in temperature, rainfall, 2.1 Trajectory of Different Villages of Dengue
and the oceanic Nio Index. The weather changes during the Patients Using a Distributed Back
El Nio and La Nia phases. The Oceanic Nio index measures Propagation Algorithm
whether the tropical waters of the Pacific Ocean are warmer Let us level up our neural net training game. How the back
or cooler than usual. In this study, we looked at how these propagation computation is distributed across GPUs or nodes
climate characteristics interact to predict mosquito activity is explained by Pierra Baladi, 2017 [9].
during El Nio episodes. The results show that a higher
incidence of mosquitoes is associated with rainfall that is more There are two typical strategies to distribute the computation:
frequent in June. After the development and implementation healthcare data parallelism and model parallelism. Here are
of an artificial intelligence (AI)-based algorithm, we will be the steps for centralized synchronous data parallelism:
able to analyze and classify the patients for the whole dataset 1. A parameter server is used as the ground truth for the
effectively, as suggested by Tahira Q-masha in 2021 [12]. model weights. The weights are duplicated into multiple
processes running on different hardware (GPUs on the
same machine or on multiple machines).
2. Proposed Work
2. Each duplicate model receives different villages
The numerous techniques supported by this paper’s of Dengue patients’ data mini-batch, and they
methodologies, which include AI algorithms, include a wide independently go through the forward and backward
range of information-gathering options. After the creation and passes, where the gradients are computed.
use of an algorithm based on artificial intelligence, assessing 3. The gradients are sent to the healthcare parameter
the patients from different villages and categorizing them for server, where they are averaged once they are all
the whole dataset will be efficient. Incoming data is initially received. The weights are updated in a gradient descent
handled for efficiency before being preprocessed. The main fashion, and the new weights are broadcast back to all
objective of this paper is to enhance the technological aspects the worker nodes.
of artificial intelligence.
454 Algorithms in Advanced Artificial Intelligence
This process is called “centralized,” where the gradients are Y = a0 + a1X (2)
averaged. Another version of the algorithm can be “decen The equation (2) is known as the regression line of Y on X.
tralized,” where the resulting model weights are averaged: Similarly, if X depends on Y, then
1. A master process broadcasts the weights of the
X = b0 + b1Y (3)
healthcare process model.
2. Each process can go through multiple iterations of the The equation (3) is known as regression line of X on Y.
forward and backward passes with different villages’
2.3 AI Healthcare Test on Dengue Using Logistic
data mini-batches. At this point, each process has a
very different weight. Regression
3. The weights are sent to the master process, they are In his 2020 [1] work, Abdulhamit presented a machine
averaged across processes once they are all received, learning technique for binary classification issues,
and the averaged weights are broadcast back to all the distinguishing between classes based on test results. It
worker nodes. consists of two classes: positive (patients with dengue) and
negative (no), according to Zhengzhou, 2021 [14]. To predict
The decentralized approach can be a bit faster because you the presence of an entity, a function mapping values between
do not need to communicate between machines as much, but 0 and 1 is needed.
it is not a proper implementation of the back propagation
algorithm. These processes are synchronous because we
need to wait for all the workers to finish their jobs. The same
processes can happen asynchronously; only the gradients or
weights are not averaged.
ÎÍ
 2 Medical Physics, 2019 May; 46(5): 2052–2063,DOI: 10.1002/
mp.13500,2019.
8. Kangzhuang Yuan: Risk and predictive factors for severe
50(306088) - (2024)(7564)
r= (7) dengue infection: A systematic review and meta-analysis,
[(50)(82322)] ÈÎ(50)(2024) - (7564)2 ˘˚ PloSOne, 17(4):e0267186, DOI: 10.1371/journal.
pone.0267186, PMCID: PMC9012395, PMID: 35427400,
r = 0.12548356497608124 ∵ –1 £ r £ 1 (8) 2022.
9. Pierre Baldi, Peter Sadowski, and Zhiqin Lu: Learning
Yugandhar 2021[13] study suggests that dengue is caused in the Machine: Random Back propagation and the Deep
by fever in individuals with down platelets, as fever weight Learning Channel, arXiv, https://doi.org/ 10.48550/
status and platelet down Y are negatively correlated. arXiv.1612.02734,2017.
The optimal point and value model data is combined with 10. Shamimul Hasan, Sami Faisal Jamdar, Munther Alalowi, and
dengue data at individual level using advanced modeling and Sadun Mohammad Al Ageel Al Beaiji: Dengue virus: A global
machine learning techniques, then dengue-tested, enhanced, human threat: Review of literature, Jounral of International
Society of Preventive & Community Dentistry (JISPCD),
and understood through regression and classification.
2016 Jan-Feb; 6(1): 1–6, doi: 10.4103/2231-0762.175416,
2016.
3. Conclusion and Future Perspective 11. Stephen Boyd: Textbook of Convex Optimization, Tata
Megraw, 2004.
The research utilized artificial intelligence and machine 12. Tahira Qamasha, Johar Jamila , Kalsooma, Faheem Ahmed
learning techniques to efficiently evaluate and classify Khan Sairac, Ambareen Sultana, Nadia beguma, Salah Ud
patients from various villages using the Optimal Point and Dind: Epidemiological study of dengue fever in District
Values model. Swabi, Khyber Pakhtunkhwa, Pakistan: Brazilian Journal of
Biology ISSN 1519-6984 (Print) ISSN 1678-4375,81(2),Mar
ay 2021, https://doi.org/10.1590 /1519-6984.216284,2021.
References 13. Yugandhar Bokka et al: Predictive Analysis for daASD
1. Abdulhamit Subasi: Machine, Practical Machine Learning for Using Population based Incremental Learning, Journal of
Data Analysis Using Python, ScienceDirect, Elsevier, 2020. Engineering Science and Technology Review 14(3) (2021)
2. Adamou Lagare, Martin Faye, Gbaguidi Fintan, Gamou Fall, 205 – 208, ISSN: 1791-2377, School of Science, IHU,
Hadiza Ousmane, ElhTassiouIbraim, MoussaMoise Diagne, doi:10.25103/jestr.143.23, Received 24 October 2020;
Soumana Amadou, Safietou Sankhe, Laminou Ibrahim, Haoua Accepted 29 June 2021.
Seini, Ousmane Faye, Ronan Jambou: First introduction of 14. Zhengzhou Shi, Neural Computing, Intelligence Science,
dengue virus type 3 in Niger, 2022, IJID Regions, Volume 7, Science Direct, 2021.
June 2023, Pages 230-232,ELESEVIER, 2023. Note: All the figures in this chapter were designed by the author.
3. Adriana Troyo, Sherri L. Porcelain, Olger Calderón-Arguedas,
Dave D. Chadee, and John C. Beier: Dengue in Costa Rica:
the gap in local scientific research, Journal of Public Health
20(5), 2006.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
456 Algorithms in Advanced Artificial Intelligence
Abstract: In the dynamic landscape of credit card usage, the surge in both legitimate transactions and fraudulent activities
necessitates vigilant measures to safeguard innocent clients from financial repercussions. This study dives into the realm of Data
Science, spotlighting the indispensable role of Machine Learning methodologies in addressing the escalating challenge of credit
card fraud. A comprehensive modeling approach is unveiled, employing diverse classifiers to tackle the inherent data imbalance
in Credit Card Fraud Detection. Meticulous experimentation addresses concerns regarding the imbalanced dataset, with
XGBoost emerging as a frontrunner, boasting a commendable precision score of 0.91 and an accuracy score of 0.99. The journey
extends to employing various sampling techniques, revealing Random Oversampling as particularly effective on imbalanced
data, yielding an impressive precision and accuracy score of 0.99 when applied to the premier model, XGBoost. Comparative
analysis of diverse classifiers yields nuanced conclusions and avenues for further research. Throughout this exploration, data
balancing procedures such as oversampling, under sampling, and SMOTE are leveraged, consistently showcasing XGBoost’s
superiority with a remarkable 99% accuracy score and precision when coupled with Random Oversampling. In summary, the
research advocates for strategic data sampling techniques to address imbalances, ensuring optimal model performance in the
intricate landscape of credit card fraud detection—an exploration that underscores the importance of advanced methodologies
in this critical domain.
Keywords: Machine learning, Credit card, Fraud detection, XGBoost
*
Corresponding author: sivasubbu22@gmail.com
DOI: 10.1201/9781003529231-68
Unveiling the Efficacy of Machine Learning in Addressing Imbalances in Credit Card Fraud Detection Data 457
enhancing informed choices in machine learning. During These classifiers, each wielding unique capabilities, play
experimentation, oversampling techniques enhance accuracy, a crucial role in the complex task of discerning whether a
showcasing practical applicability in refining algorithmic transaction is potentially fraudulent..
performance. Embark on a journey through related work,
machine learning concepts, data balancing techniques, 2.1 Logistic Regression
experimental procedures, results, and discussions. Join In supervised algorithms, Logistic Regression is a potent
us in exploring the intricate landscape of credit card fraud force, adept at classifying datasets. Its unique ability lies in
detection, where innovation and strategic data balancing predicting values for both categorical and numerical variables
fortify defenses against this insidious threat. simultaneously, navigating the intricacies of continuous and
discrete datasets. Beyond conventional boundaries, Logistic
2. Related Work Regression stands as an essential cornerstone in machine
learning. Operating as a captivating dance with probabilities,
In the ever-evolving landscape of fraud, researchers have it illuminates likelihoods in the intricate ballet of data,
contributed to the quest for effective detection methods. emerging not just as a classifier but a versatile predictor that
Prior studies explored neural networks and various machine shapes the narrative of machine learning endeavors with
learning algorithms for credit card fraud detection, yielding predictive finesse.
diverse conclusions. This paper adopts a comprehensive
approach, employing both classification and ensemble 2.2 Decision Tree Classifier
learning methodologies to enhance fraud identification.
In the vast landscape of machine learning, this classifier is a
Exploring the challenge of imbalanced data, researchers
versatile tool, excelling in both classification and regression
apply under-sampling and oversampling, achieving improved
applications. Imagine it as a narrative tree, with core nodes
results with under-sampling on logistic regression. Artificial
representing dataset attributes, branches articulating decision
neural networks emerge as a robust choice for fraud detection
rules, and leaf nodes offering conclusions. This visual
[3]. The paper builds on these foundations, identifying the
symphony unfolds potential answers to problems, each
top three algorithms. The All K-Nearest Neighbors under-
based on nuanced conditions. Its expansiveness allows for
sampling strategy combined with CatBoost stands out as the
rich decision rules, evolving with complexity. It’s not just a
recommended model, showcasing superior performance. A
classifier; it’s a storyteller with graphical finesse, unraveling
groundbreaking contribution surfaces with the proposition
possibilities within a dataset and crafting a narrative that
of a Deep Convolution Neural Network (DCNN) technique
evolves with the challenge at hand.
for financial fraud detection, leveraging deep learning
algorithms. This approach exhibits heightened accuracy, 2.3 XGBoost Classifier
particularly with substantial data volumes. Experimental
findings showcase a remarkable 0.99% detection accuracy Crafted with the power of Gradient Boosting, XGBoost is a
within a 45-second timeframe, outperforming existing champion in the arena of competitions. This classifier deli
models. In essence, this paper pioneers the use of simple yet cately balances scales, assigning crucial weights to indepen
effective techniques, highlighting the significance of practical dent factors in a symphony of decision trees. It’s not just a
results over complexity—an invaluable contribution to the classifier but a maestro reducing overfitting through regular
ongoing discourse on fraudulent activity detection. In the ized boosting. XGBoost handles missing values gracefully,
intricate realm of machine learning, classification emerges as showcasing adaptability. Flaunting a built-in cross-validation
the pivotal task, often referred to as the prediction issue. This mechanism, it orchestrates perfection in the grand theater of
nuanced challenge entails the categorization of independent machine learning. XGBoost takes center stage, weaving pre
variables, spanning from two to multiple groupings. Whether cision, adaptability, and resilience into a tapestry of predic
dealing with structured or unstructured data, diverse tive excellence—a virtuoso in the art of machine learning.
strategies come into play, demonstrating their versatility
across both data types [4]. At the heart of this endeavor lies 3. Data Balancing Techniques
the “classifier,” an algorithm designed to decipher incoming
The following are examples of strategies that may be used to
data and assign it to a specific class or category. Classification
counterbalance the dataset in order to optimize the results:
manifests in three distinct forms: binary classification, multi-
label classification, and multi-class classification. At its 3.1 Random Over Sampling
essence, binary classification stands as the simplest iteration.
In the pursuit of determining the legitimacy of a credit card Enter the dance of random oversampling, a masterful
transaction, a curated list of classifiers takes center stage [5]. composition weaving the fabric of data balance. This
458 Algorithms in Advanced Artificial Intelligence
technique randomly selects samples from minority groups, Respirators that are in commonly used are negative pressure
integrating them into the testing set. The harmonious interplay system which require the power of lungs to draw-in purified
aims for a symphony of more egalitarian data distribution—a air which is not suitable and sometimes not possible if the
nuanced performance in the grand ballet of machine learning. person lacks sufficient lungs strength, or if they suffer from
respiratory illness. This work proposes a forced air (positive
3.2 Random Under-Sampling air pressure) solution to the problem.
In this avant-garde approach, a meticulous act unfolds as
instances exit the stage of training data, creating a dynamic 5. Experimental
void. A dramatic twist commences with a random selection
journey, plucking entries from the category abundant Embark on the captivating voyage of data exploration,
with items. The result is a delicate choreography, a dance a crucial chapter in the grand saga of data analysis. In the
where individuals from the class with the highest numbers intricate dance of information, the pre-processing stage
are gracefully eliminated through a stochastic function’s takes center stage, refining data and elevating its quality for
whims. From this orchestrated performance emerges a the artistry of modeling [8]. This meticulous choreography
refined equilibrium, a balanced tableau achieved through the involves cleansing data tainted by partiality, errors, or
rhythmic interplay of elimination and random selection—a inconsistencies. As the stage unfolds, scrutiny reveals a
narrative of balance in the language of machine learning. stark imbalance between class 1 and class 2 in the target
characteristic. Fig. 68.1’s pie chart vividly illustrates the
3.3 SMOTE skewed distribution, and Fig. 68.2’s bar chart becomes a
visual scorecard, counting the presence of each class category.
Behold the ballet of SMOTE, an approach that orchestrates
The narrative becomes a tale of imbalance, prompting the call
equilibrium through a series of eloquent movements:
for ultimate balancing techniques to harmonize asymmetry
Initiate with a detailed exploration of the underrepresented and allow for a nuanced exploration of the data landscape.
group, then meticulously select numbers closest to the focal The journey into crafting or deploying a machine learning
point, labeled “k.” Envision a tapestry woven with lines model commences with the orchestration of data pre
connecting each minority point to its nearby counterparts, processing—a transformative ballet that refines raw data
creating a rhythmic dance that repeats for every point and its
diverse “k” neighbors. This harmonious repetition converges
the once scattered data into a balanced symphony—a narrative
told through investigation, selection, and connection.
4. Mathematical Model
4.1 BLDCM
The mathematical model of BLDCM a sensorless can be
modeled and implemented by the means of mathematical
modeling (transfer functions) shown in equations shown
below. The winding of a three phase BLDC motor can be
modeled as a series circuit consisting of a resistance R, an Fig. 68.1 Imbalance data distribution
inductance L and a speed dependent voltage source which
is known as the back EMF voltages due to the rotor magnet.
While designing a BLDC motor, a few parameters like induced
current in the rotor due to stator harmonics fields, iron and
stray losses are neglected. Self and mutual inductances are
considered as constant [6]. The BLDC motor is supplied three
phase voltage represented. The normally used N95 respirators
are of negative pressure variant i.e., they require the wearer’s
lungs to inhale air through the resistive membranes of the
filter layers. This is strenuous and uncomfortable to wear for
a long duration. This is non-existent in positive air pressure
respirators as they use external filters and has a motorized
air supply system [7]. The pandemic in recent scenario also
necessitates respiration apparatus as a part of its treatment. Fig. 68.2 Distribution of data after random oversampling
Unveiling the Efficacy of Machine Learning in Addressing Imbalances in Credit Card Fraud Detection Data 459
into a sculpted masterpiece. As the spotlight intensifies, the 2.5 Data Pre-processing
target characteristic takes center stage, unraveling a story
Witness our dataset’s transformation—a tapestry of 31
of imbalance with class 1 significantly dwarfed by class 2.
characteristics, some uplifting, others detrimental. The saga
Figure 68.1 and Fig. 68.2 visually narrate this imbalance,
starts with surgical removal of unnecessary traits, led by PCA,
emphasizing the need for mastery over ultimate balancing
extracting the best features and rendering ‘time’ obsolete. The
techniques. Armed with this knowledge, the journey propels
cleansing extends to duplicate samples, erased entirely. In
forward, seeking equilibrium amid the unbridled dance of
variance estimation’s dance, normalization takes center stage,
data, transcending imbalance, and paving the way for an
ensuring harmonious alignment [12]. The grand architecture
orchestrated exploration of untapped potential.
of algorithmic modeling unfolds in a simplified yet intricate
Enter the visual realm of Fig. 68.1, where the data unfolds workflow—a visual narrative capturing our dataset’s essence.
as a tapestry of pronounced imbalance, a potential harbinger This is the tale of a dataset metamorphosing—shedding the
of biased results and lackluster model performance. Here, superfluous, harmonizing the diverse, and emerging as a
class 0 reigns supreme, constituting over 90% of the data refined masterpiece in machine learning.
distribution—a towering majority class that threatens to cast
Having traversed analysis and preparation, our odyssey
shadows on the integrity of the model’s dance [9]. Yet, the
culminates in the unveiling of four machine learning virtuosos:
narrative takes an avant-garde turn as the stage is set for a
the Regression Model, Decision Tree Classification, XGBoost
transformative act—the application of a sampling approach
Classification, and ANN. Each maestro weaves a narrative
poised to reshape the data’s symphony. In this act, class 1
into our dataset. Metrics’ scores—Precision, Accuracy,
takes the spotlight, stepping into the limelight as it undergoes
Recall, and F1 Score—are computed, birthing a symphony
the graceful choreography of random oversampling. The
of evaluation. As they dance with our data, the XGBoost
following figure captures this metamorphosis—an artful
Classifier emerges as the virtuoso with the most enchanting
rendering that seeks to harmonize the data, sculpting a
performance [10]. Yet, a nuanced revelation unfolds—
balanced format that promises a more nuanced performance
results tinged with bias from the imbalance discovered
in the grand ballet of machine learning.
during exploratory data analysis. Despite surpassing 99%
2.4 Feature Selection accuracy, a glimmer of hope emerges—the best-performing
model beckons for further experimentation. This turning
Experience the alchemy of distillation, where variables point is where data sampling techniques become our guiding
fade, leaving only essential data to train our model. This compass, filtering or infusing samples to harmonize the data
harmonious process compresses data efficiently, elevating and unveil optimal performance, transcending echoes of
machine learning to its zenith. Automatic selection, a imbalance and bias.
virtuoso performance, enhances the model effortlessly. In
the symphony of model enhancement, data preparation
and feature engineering shine as shown in Fig. 68.3. Their 6. Proposed Methodology
fusion shapes not just a model but a refined masterpiece Return to the dataset’s genesis, where the imbalanced split
[11]. It’s a saga of refinement, where distilled data converges led machine learning algorithms to train models, and our
with chosen attributes, creating a model that transcends— chosen, albeit biased, champion emerged. Now, our story
the transformative power of thoughtful preparation and transcends, venturing into experimentation with three
engineering in the grand theater of machine learning. datasets transformed by Random Under Sampling, Random
Over Sampling, and SMOTE. The protagonist, our chosen
model, takes on a fresh role—training on these modified
datasets. The curtain rises on a symphony of analytical
performances, mirroring its prior dance in the splitting
ratio. The essence lies in meticulous analysis of algorithmic
performance on imbalanced and balanced datasets, dissecting
the impact of sampling techniques [13-16]. Behold the visual
tapestry below—a flow chart capturing each poetic note in
the grand composition. The proposed framework unfolds
a new phase—an odyssey through data sampling, where
equilibrium reigns. Three methods—Data Oversampling,
Data Under-sampling, and SMOTE—converge for a delicate
balance. The stage is set for the grand reveal. The XGBoost
Fig. 68.3 Experimental workflow without and with balancing model, a maestro, takes center stage, each note resonating
data
460 Algorithms in Advanced Artificial Intelligence
4. Tej Paul Bhatla, Vikram Prabhu & Amit Dua (2003). 12. N.Deshai, Unmasking deception: a CNN and adaptive
“Understanding Credit Card Frauds,”. PSO approach to detecting fake online reviews, 2023, Soft
5. N. Deshai, Deep Learning hybrid approaches to detect fake Computing, 1–22.
reviews ans ratings, 2022, JSIR, 82(1) (pp.120-127) 13. Al Rubaie, E. M. (2021). Improvement in credit card fraud
6. R. R. Popat and J. Chaudhary(2018). A Survey on Credit Card detection using ensemble classification technique and
Fraud Detection Using Machine Learning, ‖ Proc. 2nd Int. user data. International Journal of Nonlinear Analysis and
Conf. Trends Electron. Informatics, ICOEI vol. 25, no. 01, Applications, 12(2), 1255–1265.
pp. 1120–1125. 14. Alkhatib, K. I.-A. (2021). Credit Card Fraud Detection
7. Mishra and C. Ghorpade, (2018). Credit Card Fraud Based on Deep Neural Network Approach. 12th International
Detection on the Skewed Data Using Various Classification Conference on Information and Communication Systems
and Ensemble Techniques, ‖ 2018 IEEE Int. Students‘ Conf. (ICICS) (pp. 153–156).
Electr. Electron. Comput. Sci. SCEECS 2018, pp. 1–5. 15. Faraji, Z. (2020). The Causal Analysis of Financial Distress
8. N.Deshai, Transparency in healthcare and e-commerce: Risk and Performance. American International Journal of
detecting online fake reviews using a dense neural network Business Management, 3(5), 5.
model with relevance mapping,2023,Soft Computing, 27, 16. Khaled Gubran Al-Hashedi, Pritheega Magalingam, (2019).
14(pp.9861-9875). Financial fraud detection applying data mining techniques: A
9. Mittal and S. Tyagi, (2019). Performance evaluation of comprehensive review from 2009 to 2019, Computer Science
machine learning algorithms for credit card fraud detection, Review, Volume 40, 2021, 100402, ISSN 1574-0137.
‖ Proc. 9th Int. Conf. Cloud Comput. Data Sci. Eng. Conflu. 17. N.Deshai A Detection Of Unfairness Online Reviews Using
2019, pp. 320–324. Deep Learning, JATIT, Volume 100, 13 (pp. 4738–4779)
10. Zhang, X., Han, Y., Xu, W., & Wang, Q. (2019). HOBA: A 18. Shuaib, M., Hassan, N. H., Usman, S., Alam, S., Bhatia, S.,
novel feature engineering methodology for credit card fraud Agarwal, P., & Idrees, S. M. (2022). Land Registry Framework
detection with a deep learning architecture. Information Based on Self-Sovereign Identity (SSI) for Environmental
Sciences. Sustainability. Sustainability, 14(9), 5400.
11. Haoxiang, Wang, and S. Smys, (2021). “Overview of 19. Wang Q, Shi Z, Jiang D. “Watch and Wait” Strategy for
Configuring Adaptive Activation Functions for Deep Neural Multicystic Dysplastic Kidney (MCDK): Status Survey of
Networks-A Comparative Study.” Journal of Ubiquitous Perceptions, Attitudes, and Treatment Selection in Chinese
Computing and Communication Technologies (UCCT) 3, no. Pediatric Urologists and Pediatric Surgeons. Frontiers in
01. Pediatrics. 2020 Jul 28; 8:423.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
462 Algorithms in Advanced Artificial Intelligence
Blockchain-driven Security Paradigm:
A Robust System Harnessing the Internet
of Medical Things (IoMT) Network for 69
Enhanced E-Healthcare Monitoring
Abstract: The Internet of Medical Things (IoMT) has arrived, revolutionizing healthcare, and the fast expansion of the Internet
of Things (IoT) has modified it even further. An essential component of long-term healthcare infrastructure development is the
Internet of Medical Things (IoMT), which streamlines cloud-based patient record tracking. The necessity of protecting people’s
personal health information is becoming more pressing as the Internet of Medical Things (IoMT) develops into a formidable big
data infrastructure. The research introduces a personalized health surveillance tool that focuses on the applications of IoMT in
e-health. The use of blockchain technology allows for safe electronic medical record transfers, which helps with interoperability
problems. There are also worries about the security and transfer of health data across different devices, even if IoMT has a lot
of potential. The study suggests the blockchain-based IoMT Security System (BC-IoMT-SS), which incorporates blockchain
technology into IoMT to improve privacy, security, and the management of patient data. The implemented framework meets
optimal privacy and security standards for IoMT devices. The implemented framework uses the encryption key of the blockchain
to enable verified practitioners to receive secure warnings based on patients’ health data. The simulation findings show that the
BC-IoMT-SS strategy is viable; it outperforms existing methodologies with a 94% accuracy ratio, a 94% efficiency ratio, a 0.63
second reduction in latency, and faster reaction times. This paper presents new findings in e-health monitoring and emphasises
the potential of blockchain technology to enhance security environments for IoMT.
Keywords: Connected health devices, Health care administration, Safety, and Blockchain technology
*
Corresponding author: jtulasirajeshphd@gmail.com
DOI: 10.1201/9781003529231-69
Blockchain-driven Security Paradigm: A Robust System Harnessing the Internet of Medical Things (IoMT) Network 463
Fig. 69.2 Smart contracts for the internet of medical things-enabled electronic healthcare
IPFS cluster nodes and smart contracts. The coordination of with an error; otherwise, the process will continue. The last
information pertaining to the identification, authorization, step is for the smart contract to authenticate the provided ID
and accurate placement of healthcare equipment is using the IPFS cluster and the IoMT network. When new
primarily the responsibility of this interaction.6. Securing agents, such as patients and doctors, join the IoMT-enabled
Data Transmission with Smart Contracts: A new guardian healthcare network, it is the responsibility of the second
of encrypted data transfer is the incorporation of smart algorithm to do so. Smart contracts ensure that the data stored
contracts into the blockchain architecture. This channel in the IPFS cluster node is accurate and complete, including
of communication distributes data into the network after patient valid IDs, device IDs, and device public addresses, as
verification and approval from different parties on the IoMT well as the mapping between these variables.
public blockchain, guaranteeing a secure and efficient flow
of data.In this expanded version of the discussion, we see
how the IPFS Node Cluster is crucial to the IoMT ecosystem
because it validates medical devices, ensures data integrity,
and allows smart contracts to be seamlessly integrated into
the blockchain. In the ever-changing world of healthcare
informatics, these components come together to establish
a web of safety, privacy, and efficiency. What to do: 1. The
defibrillator sends a transaction T5 to the IPFS compute node,
encrypting the valid pass with its secret key. The output is
T5 = DIDpK(IPFSCIK(PID, DID, DIP)). 2. Smart contracts
on IPFS cluster nodes authenticate incoming transactions
using the phone’s public key. It is DIDPK.IPFSCIK(PID,
DID, DIP) is equal to DIDPK(IPFSCIK(PID, DID, DIP)).
The authenticity of the approved retrieved pass is confirmed
by the IPFS network node using its public key, IPFSCPK.
Following smart contract verification of PID existence on the
IPFS cloud server, the information is distributed throughout
the IoMT block chain. The rules for device authentication are
laid out in Algorithm 1. Fifth, if the PID is not confirmed in
the IoMT-BC, the authentication operation will be interrupted
466 Algorithms in Advanced Artificial Intelligence
It is important to code for every possible outcome in the smart • Python was chosen as the coding language due to its
contract in order to ensure that it does not get stuck in a hung versatility and efficacy in developing the proposed
state. The output of a smart contract should be consistent blockchain system.
regardless of the inputs used; this property is known as Research Conclusions:
determinism. Everything is kept in sync since each node can
independently verify the transactions and the overall status The BC-IoMT-SS demonstrated commendable performance
of the system. It is not possible to undo the execution of a across multiple dimensions, as elucidated in the ensuing
piece of code or a transaction; nevertheless, more code is discussion.
executed in order to make the required corrections. Since the 1. Precision Ratio and Validation:
system is impenetrable, this increases the chain’s integrity • The BC-IoMT-SS exhibited robust predictive capabilities
and guarantees the system’s integrity. At their core, smart and validation accuracy, as evidenced by a high precision
contracts are scripts of immutable, self-verifying code that ratio. This metric underscored the system’s ability to
can carry out automatic, decentralized actions; they boost accurately predict and authenticate e-healthcare data.
security, eliminate the need for a trusted third party, and are
2. Reduced Delay and Enhanced Productivity:
inexpensive to build. Smart contract ideas underpin the IoMT
paradigm for electronic health. • Notably, the BC-IoMT-SS showcased a substantially
shorter delay, contributing to enhanced system
1.1 The mathematical equation for IoMT in healthcare responsiveness. This reduction in delay is pivotal for
In order to verify compliance in the networks, the approach real-time applications, particularly in the context of
generates a transaction-based polynomial. Assume that healthcare monitoring where timely data transmission is
a set of purchases is denoted by T = [T1, T2, T3,... Tm]. critical. The system’s productivity was further bolstered
The hashes of the transactional polynomials are H1(T1), by this optimization.
H1(T2),..., H1(Tm). For each purchase m, we may derive a Dataset Description:
polynomial f(q) such that f(H1(Tk)) = 0, where k is an integer
• The clinical study involved a dataset comprising infor
from 1 to m. Once designers have f(q) = (q — H1(T1)),
mation from 10 patients, forming the basis for a compre
(q — H1(T2)),... (q — H1(Tm)) (1), they can express the
hensive analysis of the BC-IoMT-SS performance.
polynomials correctly.The original definition of the vector
U is U = {u1, u2, u3, …, um—1, um}. To rephrase it as hi • The dataset underscored the pivotal role of the Internet
= H1(TK), H1(TK))2,..., H1(TK)m—1, (H1(TK)m}, the of Things (IoT) in healthcare digitization. IoT, as a
definition would be the same. Every time a new structure transformative force, defines the management and
index is created in a public blockchain, the peers (agency) monitoring of physical spaces, their contents, and the
check the agreement vector u. If the trust factor among the well-being of inhabitants through interconnected sensors
peers (agency) is more than 1, the newly valid systematic and actuators.
methodology of the IoMT blockchain adds the data. IoT in Healthcare:
• The IoT paradigm elucidated a system of autonomous
4. Experimental Analysis computing entities capable of seamlessly gathering,
transmitting, and sharing data without direct user
This section delves into the intricacies of the experimentation intervention.
process, detailing the apparatus utilized and presenting • Illustratively, IoT’s impact on the healthcare sector
a comprehensive analysis of the obtained outcomes. The was highlighted, showcasing its ability to connect
experimentation was conducted on an HP EliteBook diverse medical equipment to a centralized server. This
computer equipped with an Intel® Core i5-6300U processor, connectivity empowers individuals to monitor their health
4 GB of RAM, and running Windows 10 Pro. Python served autonomously and engage in remote communication
as the coding language employed to construct the proposed with healthcare professionals. In summation, the
blockchain system. experimentation and analysis underscored the BC-IoMT
Experimental Setup: SS’s prowess in predicting and validating e-healthcare
• The experimentation framework was anchored on an data, with a focus on precision ratio, reduced delay,
HP EliteBook computer, ensuring a standardized and and heightened productivity. The study was much more
reliable environment for conducting the trials. comprehensive and applicable because it used a broad
• Specifications of the computer included an Intel® Core dataset and recognized the revolutionary significance of
i5-6300U processor, 4 GB of RAM, and the Windows 10 the internet of things in healthcare. The importance of
Pro operating system, guaranteeing consistent conditions systems cannot be overstated. One promising strategy
for the experimental trials. is to assess existing evidence using a high-performance
468 Algorithms in Advanced Artificial Intelligence
5. Conclusion
This study introduces BC-IoMT-SS, a blockchain-based
approach to e-healthcare, addressing the growing demand
for enhanced framework networks. The framework secures
(a)
data management through distributed blockchains on IPFS
clusters, ensuring equal service to authorised agents and
protecting patient information confidentiality. Improving
service quality and processing data from consumer devices
are two upcoming goals, along with government integration
and software-defined networking.
References
1. B. Godi, S. Viswanadham, A.S. Muttipati, O.P. Samantray,
S.R. Gadiraju, E- healthcare monitoring system using IoT
(b)
with machine learning approaches, in: 2020 International
Conference on Computer Science, Engineering and
Fig. 69.5 Efficiency ratio Applications (ICCSEA), IEEE, 2020, March, pp. 1–5.
2. T. Saba, K. Haseeb, I. Ahmed, A. Rehman, Secure and energy-
efficient framework using Internet of Medical Things for
e-healthcare, J. Infect. Public Health 13 (10) (2020) 1567–
1575.
3. G. Nagasubramanian, R.K. Sakthivel, R. Patan, A.H. Gan
domi, S. Muthuramalingam, B. Balamurugan, Securing
e-health records using keyless signature infrastructure block
chain technology in the cloud, Neural Comput. Appl. 32
(2020) 639–647.
4. B. Swapna, S. Gayathri, M. Kamalahasan, H. Hemasundari,
M. SiraasGanth, S. Ranjith, E-healthcare monitoring using
internet of things, in: IOP Conference Series: Materials
Science and Engineering, vol. 872, IOP Publishing, 2020,
Fig. 69.6 Managing the upload timing of transactions (Tx) on June, 012024, 1.
IPFS’s encrypted storage layer for different sizes 5. S. Kadam, D. Motwani, Block chain based E-healthcare
record system, in: Image Processing and Capsule Networks:
machine-learning system. The accuracy of experiments ICIPCN 2020, Springer International Publishing, 2021, pp.
conducted using BC-IoMT-SS is 94% or higher. 366–380.
6. M.M. Khubrani, A framework for block chain-based smart
In order to monitor patients in real-time, make accurate health system, Turkish J. Comput. Math. Educ. (TURCOMAT)
diagnoses, and administer effective treatments, hospitals 12 (9) (2021) 2609–2614.
rely on sensor technology and electrocardiograms (ECGs). A 7. Z. Shahbazi, Y.C. Byun, Towards a secure thermal-energy
bounded telemonitoring system enables real-time monitoring aware routing protocol in wireless body area network based
on block chain technology, Sensors 20 (12) (2020) 3604.
of anyone within the medical center, including inpatients and
8. H. Liu, R.G. Crespo, O.S. Martinez, Enhancing privacy and
outpatients. During disasters, tags can monitor patients’ health data security across healthcare applications using block chain
issues, which improves their well-being and overall quality of and distributed ledger concepts, in: Healthcare, vol. 8, MDPI,
life. A 94.1% success rate in an experiment employing BC 2020, July, p. 243, 3.
Blockchain-driven Security Paradigm: A Robust System Harnessing the Internet of Medical Things (IoMT) Network 469
9. D. Wu, N. Ansari, A cooperative computing strategy for block 24. B.V.D.S. Sekhar Et Al (Oct 2022),” Novel Technique of
chain-secured fog computing, IEEE Internet Things J. 7 (7) Threshold Distance-Based Vehicle Tracking System for
(2020) 6603–6609. Woman Safety.”, Intelligent System Design, Lecture Notes in
10. Z. Ashfaq, A. Rafay, R. Mumtaz, S.M.H. Zaidi, H. Saleem, Networks and Systems 494, INDIA 2022, DOI:10.1007/978
S.A.R. Zaidi, A. Haque, A review of enabling technologies for 981-19-4863-3_56, Pp:567-577, https://link.springer.com/
internet of medical things (IoMT) ecosystem, Ain Shams Eng. chapter/10.1007/978-981-19-4863-3_56
J. 13 (4) (2022), 101660. 25. B.V.D.S. Sekhar Et Al (Oct 2022),” Sustainable and reliable
11. G. Miao, A.A. Ding, S.S. Wu, Real-time disease prediction healthcare automation and digitization using deep learning
with local differential privacy in Internet of Medical Things, technologies”, Journal of Scientific and Industrial Research
arXiv preprint arXiv (2022), 2202.03652. (JSIR), (SCIE, WOS, Scopus)
12. M.B. Janjua, A.E. Duranay, H. Arslan, Role of wireless 26. B.V.D.S. Sekhar Et Al (June 2021),” Real Time Facial
communication in healthcare system to cater disaster situations Expression Recognition Using Open CV And Deep Learning”,
under 6G vision, Front. Commun. Net. 1 (2020), 610879. International Journal of Research, Vol X, Issue VI, ISSN:
13. A. Lakhan, M.A. Mohammed, M. Elhoseny, M.D. Alshehri, 2236-6124, Pages: 7-27.
K.H. Abdulkareem, Block chain multi-objective optimization 27. B.V.D.S. Sekhar Et Al (August 2021),” The Hybrid Algorithm
approach-enabled secure and cost- efficient scheduling for the for increasing Reversable Data Hiding Scheme for Medical
Internet of Medical Things (IoMT) in fog-cloud system, Soft Images”, International Journal of all Research Education and
Comput. 26 (13) (2022) 6429–6442. Scientific Methods IJARESM, Vol:9, Issue:8, August 2021,
14. X. Li, B. Tao, H.N. Dai, M. Imran, D. Wan, D. Li, Is block Issn: 2455-6211, Pages: 2470-2476.
chain for internet of medical things a panacea for COVID-19 28. B.V.D.S. Sekhar Et Al (December 2020),” A Novel Technique
pandemic? Pervasive Mob. Comput. 75 (2021), 101434. For Prediction Of Coronary Artery Disease From Human
15. S. Razdan, S. Sharma, Internet of medical things (IoMT): Fundus Images Using Machine Learning Approach”,
overview, emerging technologies, and case studies, IETE International Journal For Innovative Engineering And
Tech. Rev. 39 (4) (2022) 775–788. Management Research, Vol:7, Issue:12, December’2020, Issn:
16. Y.D. Al-Otaibi, K-nearest neighbour-based smart contract for 2456-5083, Pages: 69-74. [SSSN, Elsevier]
internet of medical things security using block chain, Comput. 29. B.V.D.S. Sekhar Et Al (December 2020),” Recognition of Hu
Electr. Eng. 101 (2022), 108129. man Being Through Handwritten Digits Using Image Process
17. A. Abbas, R. Alroobaea, M. Krichen, S. Rubaiee, S. Vimal, ing Techniques And Ai”, International Journal For Innovative
F.M. Almansour, Block chain-assisted secured data manage Engineering And Management Research, Vol:7, Issue:12, De
ment framework for health information analysis based on cember’2020, Issn: 2456-5083, Pages: 69-74. [SSSN, Elsevi
Internet of Medical Things, Personal Ubiquitous Comput. er]
(2021) 30. B.V.D.S. Sekhar Et Al (February 2020),” A Novel Robotic
18. P.P. Ray, N. Kumar, D. Dash, BLWN: block chain-based aid for physically challenged Implemented using Image
lightweight simplified payment verification in IoT-assisted Processing”, Journal Of Engineering Research and
e-healthcare, IEEE Syst. J. 15 (1) (2020) 134–145. Application, IJERA Vol: 10, Issue:2 (Series-I), ISSN: 2248
19. B. Sharma, R. Halder, J. Singh, Block chain-based 9622, Pages: 53-57.
interoperable healthcare using zero-knowledge proofs and 31. B.V.D.S. Sekhar Et Al (July 2020),” Processing Real World
proxy re-encryption, in: 2020 International Conference on Datasets using Big Data Hadoop Tools”, Journal of Scientific
Communication Systems &Networks (COMSNETS), IEEE, & Industrial Research, Vol:79(7), Pages: 631-635. ISSN: 0975
2020, January,pp. 1–6. 1084, http://nopr.niscair.res.in/handle/123456789/54985
20. A. Farouk, A. Alahmadi, S. Ghose, A. Mashatan, Block [SCI, Scopus]
chain platform for industrial healthcare: vision and future 32. B.V.D.S. Sekhar, Pvgd Prasad Reddy, Gps Varma “Performance
opportunities, Comput. Commun. 154 (2020) 223–235. Of Secure And Robust Watermarking Using Evolutionary
21. J. Sengupta, S. Ruj, S.D. Bit, A comprehensive survey on Computing Technique” JGIM, Vol 25, Issue 4, Article 5.
attacks, security issues and block chain solutions for iot, J. October- December 2017, Doi 10.4018/Jgim.2017100105.
Netw. Comput. Appl. 149 (2020), 102481. Pages 61-79, [Web of Science, Sci, Scie Journal, Indexed In
22. D.C. Nguyen, P.N. Pathirana, M. Ding, A. Seneviratne, Block Acm, Scopus]
chain for secure EHRs sharing of mobile cloud based e-health 33. B.V.D.S. Sekhar Et All (January 2020) “An Experimental
systems, IEEE Access 7 (2019) Analysis Of Secure-Energy Trade-Off Using Optimized
23. B.V.D.S. Sekhar Et Al (NOV 2022),” Artificial neural Routing Protocol In Modern-Secure-Wsn”, Eai Endorsed
network-based secured communication strategy for vehicular Transactions On Scalable Information Systems, Issue:24,
ad hoc network”, Soft Computing, Springer, Vol: 27, Issue 1, Issn: 2032-9407(Accepted For Publication) [Web Of Science,
PP 297-309 https://link.springer.com/article/10.1007/s00500 Sci, Scie Journal, Indexed In Acm, Scopus]
022-07633-4, ISSN:1432-7643,1433-7479, IF: 3.643, (SCIE, 34. B.V.D.S. Sekhar, Pvgd Prasad Reddy, Gps Varma “A
WOS, Scopus) Neural Network Model For Detecting Anomalous Traffic
470 Algorithms in Advanced Artificial Intelligence
Implementing Self-Organizing Maps” International Journal 37. B.V.D.S. Sekhar, Pvgd Prasad Reddy, Gps Varma “Principal
Of Computational Intelligence And Health Informatics, 2008. Component Analysis Based Image Denoising Implemented
Vol 1, No 1, Pp 25-29, Issn 0973-7413 Using Lpg And Compared To Wavelet Transform Techniques”
35. B.V.D.S. Sekhar, Pvgd Prasad Reddy, Gps Varma ” Novel Ijesrt, Doi: 10.528/Zendo.55803, Pp 673-678, Vol 5 No 6,
Technique Of Image Denoising Using Adaptive Haar Wavelet June 2016. Issn 2277-9655. (UGC:43449)
Transformation “Irecos , 2015, Vol 10, No 10, Pp 1012-1017, 38. B.V.D.S. Sekhar Et.Al,” A Novel Technique For Home
Issn 1828-6003 ( Scopus Indexed). Automation” Ijircce, Vol 4, Issue: 6, June 2016, Doi: 10.15680/
36. B.V.D.S. Sekhar, Pvgd Prasad Reddy, Gps Varma “Improved Ijircce.2016.0406283, Pp 12059-12062, Issn: 2320-9801.
Psnr In Image Denoising Using Modified Median Filter ” Ieee 39. B.V.D.S. Sekhar, Gps Varma, Et.L” Secure Automative
Explore, Feb 2016, Part Number Cfp1665w-Art, Isbn 978 Locking Control And Anti Theft Using Gps And Blluetooth”
4673-7832-1 Ijirmf, Vol 2, Issue 8, Aug 2016, Pp 165-168, Issn: 2455-0620.
Note: All the figures in this chapter were designed by the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
Yendrapati Geetha*
Assistant Professor,
Dept of Computer Science and Engineering,
Gudlavalleru Engineering College, Gudlavalleru
Abstract: Improved forecasting and strategic decision-making are necessities in today’s cutthroat economic climate for
both importers and exporters. The proposed method, which is based on big data analytics, may help businesses locate fresh
opportunities and revise their strategic decisions. An empirical investigation of the distribution of agricultural commodities in
India over the course of two years validates the proposed strategy. An advanced analytical framework for strategic decision-
making is provided by the proposed study, which is based on Big Data Analytics (BDA). This research delves into the topic of
overseas export volume estimation through the use of machine learning to big data business analytics. Using machine learning
techniques and the values anticipated from a thorough market analysis, this study determines the amount of exports to foreign
nations. These results show that the proposed approach improves strategic market analysis and gives accurate trade estimates.
The experimental results show that using machine learning techniques in a Hadoop context greatly improves the accuracy of
the estimates.
Keywords: Agriculture, Big data analytics, Decision making analysis, Strategic decisions, Machine learning
*
Corresponding author: geetha.yendrapati223@gmail.com
DOI: 10.1201/9781003529231-70
472 Algorithms in Advanced Artificial Intelligence
(2021)[4], there is software and tools for big data analytics as text, audio, video, and images. Kornelia (2022) found that
that might help us make decisions based on data, which although information is richer than structured, it is difficult
could improve the results of company operations. Possible to scrape [8]. The export product structure and function
advantages include more efficient marketing, more consumer of our product export business system are attempted to be
personalization, and more efficient operations. If you have a mimicked by neural networks, which employ technology that
well-thought-out strategy, it can outperform your peers and aims to generate intelligent behavior. Typically, this system
reap these rewards. the work of Janssen M. in 2017.Using is represented as a weighted directed graph, with neurons
effective big data analytics, as outlined by Gupta, 2019[6], serving as nodes and connections between them as edges.
one can derive the most insightful conclusions from the The type and strength of the contact between the adjacent
data’s increasing volume, velocity, and diversity.In this essay, neurons are indicated by the weights on each.
you will find five sections. Section 1: Research Motivation is
an Introduction to the Section. Section 2 covers the related Perceptrons, according to Neelam Tyagi, 2020[10], are the
approach of using Perceptron to export product data. In simplest sort of neural networks. They comprise a single
Section 3, we detail the intended task and its technical details. neuron with several real-valued or binary inputs and a binary
Section 4 presents the experimental results of the proposed output. The inputs are multiplied by the weights on the edges
investigation. The results are summarized in Section 5. that have been assigned to them. At any given moment, the
total of all weighted inputs is what the neuron perceives as
its net input. In response to a net input that is greater than a
2. Related Methods certain threshold, the neuron will fire and emit a “1” signal;
The export quantities to foreign countries are determined otherwise, it will emit a “0” signal. Figure 70.1 shows the
using data business analytics. Perceptron.
Framework for Export Products A Learning Algorithm for If all goes according to plan, this export product system’s
Perceptrons: Unstructured Big Data includes media files such Perceptron will be trained to provide specific results when
Fig. 70.3 Export large data process using distributed back propagation approach
Estimating Foreign Export Volume Using Machine Learning for Big Data Business Analytics 475
weights or gradients when processes execute asynchronously Sensitivity: Sensitivity, also known as the true positive rate
[14]. That is the only difference. or recall, is a measure of an individual’s ability to recognize
positive examples.
4. Experimental Result Specificity: The specificity of a model is its ability to
accurately predict true negatives in each accessible category.
Assessing a machine-learning model’s correctness is a
crucial phase in the procedure. With only 4800 records of Precision: The precision of a test is determined by dividing
training export product data, a model cannot produce accurate the number of correctly identified positive samples by the
predictions for untested export product data. total number of positive samples.
The Distributed Back Propagation algorithm in the Hadoop F1-Score: The F1-score is a statistical tool that compares the
Environment was compared for accuracy, sensivity, performance of two classifiers by calculating the harmonic
specificity, precision, and F1-score using research metrics by mean of their precision and recall.
Zhang, 2018[16]. The graphs display accuracy metrics like sensitivity,
Accuracy: The accuracy of a model is a metric that gauges its specificity, precision, and F1-score for Distributed Back
performance across all classes, particularly when all classes Propagation achieving the most favorable results.
are equally essential.
Table 70.2 Graph for comparing algorithm
Algorithm Accuracy Sensitivity Specificity Precision Fl-score
Back Propagation 0.96 0.97 0.95 0.95 0.97
Abstract: These days, consumers rely heavily on internet reviews to gauge other shoppers’ opinions before making a purchase,
which in turn shapes the online marketplace. Although real evaluations are valuable, the problem of fraudulent reviews is a new
dimension that could lead customers astray. Accurately recognising fraudulent reviews, especially within the prolific Amazon
dataset, is an urgent challenge that this paper addresses. This research delves into the complex terrain of the e-commerce
business by introducing two unique deep-learning hybrid models, the most notable of which is BERT-CNN. Hybrid models
that use state-of-the-art word embedding techniques like Glove and One-Hot Encoding demonstrate exceptional performance.
The results of the experiments prove that BERT-CNN is effective; it achieved an outstanding 98.7 percent accuracy rate. This
research not only adds new knowledge to the field of false review detection, but it also demonstrates how effective and useful
BERT-CNN is for improving the accuracy and efficiency of this critical work.
Keywords: BERT-CNN glove, One hot encoding
*
Corresponding author: desaij4@gmail.com
DOI: 10.1201/9781003529231-71
478 Algorithms in Advanced Artificial Intelligence
to improve detection. When reviews use casual language or is crucial, enabling deep learning methods to accurately
obscured words, however, these text elements may reduce comprehend, leverage, and effectively classify each online
detection accuracy. If the bag-of-words assumption is review.
applied, which relies on word frequency, the review feature
vectors may become sparse. Linguistic features, like POS 2.2 Feature Extraction
n-grammes, might not be able to tell the difference between The N-Grammes These days, N-grammes are all the rage when
a real reviewer and a highly skilled imposter. Word2Vec it comes to text classification jobs. All sorts of permutations
and similar word embedding algorithms also have trouble of letters, numbers, and symbols are encompassed in this.
capturing all possible semantic meanings, which could affect Numerous tokenization techniques are available, such as
how well they recognise reviews in different contexts. whitespace, expression, and unigram, which function on the
phrase, comment, or character levels. For example, a bigram
2. Related Work or trigram uses sequences of many tokens, but a unigram
only uses one token. Normalisation to reduce the number of
Amazon, Yelp, Google Play Store, Trip Advisor, and social unique tokens and eliminate text variances, normalisation
media are just a few of the modern venues where customer is a crucial strategy. Stemming and lemmatization remove
evaluations and ratings play a crucial role in influencing superfluous data from content during the cleaning process.
purchasing decisions. The recent upsurge in fraudulent reviews A lexeme is a basic form in the fields of linguistics and NLP.
threatens both businesses’ success and customers’ trust. Text Using stem cells Though stemming’s primary function is to
reviews influence 93% of buyer decisions and command a reduce words to their simplest forms, it has the potential to
large 43% market share, with digital buyers estimated to introduce ambiguity. The Porter stemmer is a good example
reach 2.14 billion. A critical area of research has emerged, of how stemming may still accurately infer the root when
the detection of bogus reviews. In particular, deep learning spelling mistakes are present. A process called lemmatization
techniques have demonstrated great promise in this area, Lemmatization provides a more refined technique than
while machine learning tools such as support vector machines
(SVMs) and neural networks (NN) have achieved remarkable
success. It is still quite difficult to do binary classification,
which involves telling the difference between real and bogus
reviews. When it comes to identifying French online reviews,
hybrid models like LSTM+CNN with Camem-BERT
accomplish remarkable results, attaining an accuracy of
93.7%. The accuracy rate on the Tweep-Fake Dataset is 89.7
percent when using DNN, CNN, GRN, and HAN. Combining
CNN with BiLSTM achieved the best accuracy of 90.66%. Fig. 71.1 Proposed deep learning hybrid methodology
On Arabic datasets, for example, CNN-LSTM attains the best
accuracy of 86.88%, proving that deep learning models often
beat conventional machine learning. Online reviews have a
significant effect on e-commerce earnings, so there needs to
be a solution to the growing problem of fraudulent reviews.
Advancements in identifying automated fake evaluations
have been substantial, highlighting the necessity for strong
techniques to safeguard online feedback platforms.
stemming, which has limitations in normalisation and leads 2.3 TF-IDF vs. Word2Vec Symphony
to the formation of useless terms. Lemmatization, which
As the story progresses, we see Word2Vec’s multidimensional
reduces words to their base form, is preferable to stemming
representation pitted against TF-IDF’s solitary vector. In
because it produces a more true portrayal of meaningful
contrast to Word2Vec’s orchestra of N-dimensional vectors,
phrases that conform to linguistic standards.
which provides a more comprehensive representation,
Revealed:Vectorization Innovation becomes the choreography TF-IDF is a soloist in its simplicity. The collision of these
of text-to-numerical vector translation in the complex dance titans reveals the depth of data production techniques. Skill-
of language and machine. By combining the linguistic depth Based Word Embeddings PerformGlove leads the way in
and computational power of algorithms, this procedure helps the elaborate waltz with BERT-CNN in the ballroom of
to isolate unique traits for the model’s training. Best Practices pre-trained word embeddings. With its extensive training
for Vectorizing Vectorizer for Counts, TF, and IDFThe on a large corpus, Glove proves to be a reliable partner,
numerical substance is extracted via these text transformation demonstrating its capabilities in conjunction with BERT
pillars. Term Frequency (TF) measures the frequency of CNN to uncover the mysteries of identifying fraudulent
words in a sample, whereas Count Vectorizer counts the reviews. BERT-CNN: A Master of Hybrid Systems Here we
occurrences of words. To make it more precise, Inverse present the hybrid model BERT-CNN, the main attraction
Document Frequency (IDF) takes into account how often the of our innovation display. We can expect better accuracy
term appears in the whole dataset. Each term’s relevance is and faster performance in bogus review prediction with this
nuancedly represented by the TFIDF equation, which depicts powerful mix of convolutional neural networks (CNNs) for
this subtle dance. Using Word2VecWord2Vec comes out with data mining and Bayesian error rate tremor (BERTs) for time
hidden semantic linkages; it’s a shining star in modern NLP. It series dependency extraction. By moving beyond the ordinary
can predict word vectors either alone or in context, using either and into a paradigm where complexity meets innovation, we
the CBOW or continuous skip-gram a approaches. Unlike arrive at this groundbreaking study. As our knowledge grows,
Skip-gramme, which predicts context vectors from the centre we can hear the symphony of comparisons that bode well
word, CBOW predicts a word’s vector based on neighbouring for solving the mystery of fraudulent internet evaluations.
context. Deciphering the complex network of contextual To find and stop the spread of fake reviews on the internet,
subtleties is like a symphony of predictions. Single-Record the BERT-CNN model skillfully combines the best features
Encoding by converting categorical variables into expressive of BERT (Bidirectional Encoder Representations from
binary vectors, One-Hot Encoding becomes a deep learning Transformers) and CNN (Convolutional Neural Network). It
master of sequential classification. By quantifying the core of is a revolutionary hybrid architecture. The unique features of
sequential linkages, this representation improves prediction this model become clear when we disassemble it:
accuracy. GloVe is a global vector representation of words.
1. BERT: Contextual Understanding: Taking into account
A new member of the vectorization ensemble, GloVe creates
both the left and right word contexts, the powerful pre-trained
word vectors by combining global and local statistics.
language model BERT does an excellent job of contextual
By de-emphasising commonly used word pairs, GloVe
understanding. Its bidirectional methodology enables a
manages to strike a balance between global context and local
sophisticated understanding of complex word associations.
subtleties. Deciphering word co-occurrences throughout
After being pre-trained on large corpora, BERT learns rich
the entire corpus is a breeze for this unsupervised virtuoso.
contextual representations through transfer learning. This
The innovative spirit is front and centre in this vectorization
model has been fine-tuned for some jobs with reduced
symphony, creating a tapestry where the power of words and
labelled data requirements using transfer learning.
numbers meet. A Groundbreaking Comparison** The rise of
fraudulent internet reviews is a complex modern problem with 2. CNN: Local Feature Extraction: One area where CNN really
many moving parts. We compare and contrast the outcomes shines is when it comes to taking input data and extracting
of various machine learning and deep learning models with local features. When applied to text, CNN is a master at
the BERT-CNN-LSTM hybrid model, which is powered by extracting characteristics and patterns from localised areas of
deep learning, in this ground-breaking investigation. Data a sequence; this is especially helpful for finding sentence or
Alchemy Revealed The transformation narrative, in which we paragraph structures.
reveal the many techniques—Glove and one-hot encoding— 3. Feature Fusion using the BERT-CNN Hybrid Model: The
orchestrated to convert textual subtleties into numerical hybrid model achieves remarkable results by combining the
vectors, is fundamental to this discovery. Data shaping is best features of BERT and CNN. CNN creates a comprehensive
revealed by each technique, which is like painting a stroke on picture of input text by concentrating on extracting local
the canvas of knowledge. features or patterns, whereas BERT offers deep contextual
comprehension. Combining BERT’s contextual embeddings
480 Algorithms in Advanced Artificial Intelligence
Abstract: Among the most rapidly developing areas in computer vision is object detection. Mask detection is the main objective
of the effort. With the use of deep learning and computer vision techniques, this project offers a reliable method for mask
identification that is implemented using RESNET architecture. Identifying faces and differentiating between people wearing
masks and those without is the main goal. The model is refined via transfer learning on a customized dataset that includes
annotated photos of faces that have been masked, masked incorrectly and unmasked faces.
Keywords: Computer vision, Object detection, You only look once, Transfer learning
1
devananth_aluri@srmap.edu.in, 2abhiram_seemakurthi@srmap.edu.in, 3sasank_tumma@srmap.edu.in, 4prasanthi.b@srmap.edu.in
DOI: 10.1201/9781003529231-72
482 Algorithms in Advanced Artificial Intelligence
When prediction of presence and placement of required fully linked stage to generate all of the layer’s predictions.
classes/objects in a image detection only goes through the While the technique utilizing a regional recommendation
input image once. By processing the full image in one pass, network processes the same image numerous times, the
computational efficiency is increased. YOLO-CNN method accesses it only once.
To examine the predicting abilities of various object
recognition models, the YOLO-CNN approach necessitates 2. Methodology
a consistent quantitative evaluation. The Average Precisions
YOLO-CNN employs a basic network known as a deep
(AP) and the Intersection-over-Unions (IoU) are two of the
convolutional neural network to detect objects within the
most popular metrics for assessment.
given image. The structure of the CNN model, essential to
A approach utilizing point-to-point neural networks in the the algorithm, is depicted in Fig. 72.3.
YOLO-CNN model to simultaneously forecast bounding box
As seen in Fig. 72.4, the initial 20 layers of convolution of the
and class probabilities. This is not the same as the prior object
model have been trained using RESNET by using temporal
detection algorithms’ approach, which recycles classifiers for
mean pooling with all connected layers. Next, the pre-trained
detection purposes.
model is adjusted to enable detection. This is because prior
By using an entirely new way for object detection and studies have demonstrated that the addition of convolutional
attaining cutting-edge outcomes, the YOLO-CNN method and linked layers improves the performance of pre-trained
surpassed existing real-time object detection systems. Instead networks. View the Fig. 72.2 that follows. In the diagram
of using distinct regions to identify possible regions of interest depicted in Fig. 72.1, the algorithm’s ultimate connection
like some algorithms do, such Faster RCNN, YOLO-CNN layer anticipates the dimensions of the bounding box and the
employs a regional recommendation network and a single, probabilities related to various classes.
Fig. 72.9 Person with mask detected from sideways Conversely, if there is an object, YOLO-CNN anticipates the
confidence score to align with the intersection over union
(IOU) of the ground truth and the predicted box. [11]
The YOLO-CNN algorithm is employed to multiply
conditional class probabilities and individual box confidence
projections during the test phase. For the You Only Look
Once (YOLO) evaluation on Pascal VOC, which comprises
20 named classes, the YOLO-CNN method is used with
parameters S = 7, B = 2, and C = 20. This final prediction
made by the YOLO-CNN is represented as a 7 × 7 × 30 tensor.
3. Conclusion
In summary, the offered code demonstrates a reliable and
organized method for setting up and using a YOLO-CNN
model for image mask detection. With the help of the
Fig. 72.10 Person with mask detected even if it is blocked
Ultralytics library, this project efficiently handles the setup
with hand
of the YOLO-CNN method model training, data conversion
into the format required by the YOLO-CNN technique, and
dataset arrangement for training. The code also incorporates
necessary procedures like real-world inference on fresh
images, metrics computation, and model validation. All
things considered, this work demonstrates how deep learning
techniques can be used to automate mask identification tasks,
with potential applications in public safety, healthcare, and
other fields
Acknowledgement
The authors gratefully acknowledge the students, staff, and
authority of CSE department for their cooperation in the
research.
Fig. 72.11 Person with mask wearing incorrectly from
sideways
486 Algorithms in Advanced Artificial Intelligence
Mouna Penmetsa1
Research Scholar,
Dept of Compute Science and Engineering,
Sagi Ramakrishnam Raju Engineering College, Bhimavaram.
R.N.V. Jagan Mohan2
Associate Professor,
Dept of Compute Science and Engineering,
Sagi Ramakrishnam Raju Engineering College, Bhimavaram.
Abstract: The art and science of tilling land and cultivating crops is known as agriculture. These days, smallholder farmers
rely on a variety of crops. To increase productivity, these farmers must grow a variety of crops. The heavy rains caused damage
to agricultural harvests in the last season then crop-loss evaluations help identify intervention needs. In addition to that small
hold farmers’ livelihoods are impacted by a variety of factors, including disease, pests, climate change, natural catastrophes,
and human activity. In this paper, Researchers advice to small hold farmers will harvest many crops per acre on a limited plot
of land. The paper discusses the use of AI-based Federated Learning for multi-crop analysis. Before going to the experiment is
on multi-regression analysis is used to estimate crop loss. The experiment involves multi-crops using multi-regression analysis
to estimate crop loss before.
Keywords: Artificial intelligence federated learning, Multi-crops and multiple class, Multiple regression analysis etc
1
mounakishan@gmail.com, 2mohanrnv@gmail.com
DOI: 10.1201/9781003529231-73
488 Algorithms in Advanced Artificial Intelligence
sustenance. Land races, cultivars, rainfall, fertilizers, soil, (MC2FL)covered in Section 3. Section 4 deals with Multi-
agricultural leftovers, and domestic animals are all necessary Crop Analysis Using Multi-Regression. Section 5 deals
for these systems to function. In biodiversity agro ecosystems, withexperimental results. Conclusion included in section 6.
their role in nutrient cycling, water management, and insect References are included in Section-7.
control is paramount by Ersin Elbasi, 2023[7]. Multicropping
contributes to the environment, tackles social concerns in the 2. Agriculture with Federated
community, and offers resilience in output and income. It
deals with reducing poverty and hunger and can help with Learning (FL)
the planning of larger-scale farming systems that will provide FL is an approach that emphasizes experience and
food in the future. As evidenced by agriculture, multi- collaboration while trainingmodels decentralized from a
cropping, or growing several crops at once on one piece of central data transfer. It has applications in agriculture but
land, increases land production and advantages. To increase faces challenges in technical components like platforms,
land productivity and provide more benefits and revenue, hardware, software, and data privacy. This work explores FL
mixed cropping involves growing several crops concurrently
architectures, platforms, hardware, and software, providing
on the same plot of land, such as banana, coconut, coco, and
a blueprint for designing FL-based solutions to improve
Kandha, during the same growing season. Intercropping
crop data. Earlier works on these lines by various authors as
reduces rivalry and increases cooperation.
follows Kulkarni V, Kulkarni M, and Pant A, An examination
The paper addresses challenges faced by small and of personalization strategies in federated learning, 2020[11]
marginal farmers in agriculture, such as climate change, has proposed the authors classify FL Systems based on data
soil deterioration, biodiversity loss, water resource distribution, model, privacy frameworks, and communication
depletion, and labor and money shortages. It suggests using architectures, dividing them into private and public systems,
Artificial Intelligence-based Federated Learning approach with potential for improvement. Geiping J, Bauermeister H,
for agricultural harvesting on small and fragmented land Dröge H, and Moeller M: Gradient inversion: how simple is
holdings, focusing on multiple crop cultivation for better it to compromise privacy in federated learning? ArXiv, vol.
yield and profitability. abs/2003.14053, 2020[7]has suggested the article delves
The seven portions of this paper’s response can be into the use of FL for recovering image data, highlighting its
understood as follows: section 1’s broad introduction of privacy limitations and the impact of network architecture and
Agriculture. Agriculture with Federated Learning is covered parameter state. Liu Y, Yuan X, Xiong Z, Kang J, Wang X, and
in Section 2. Multi-Crops Multi-Class Federated Learning Niyato D:FL for 6G communications: obstacles, strategies,
and next steps, arXiv preprint arXiv: 2006.02931, 2020[12] on a human activity recognition dataset, the Agriculture with
suggested the article discusses the potential of machine Federated Learning framework outperformed traditional
learning in 6G communications, addressing challenges such machine learning algorithms, making it applicable to
as high costs, security, and privacy, while also presenting Agriculture Environment. The framework can incorporate
architecture for its implementation by Manoj T, 2022[13]. other conventional Deep learning methods by Madhuri
Shripathi Rao,2022[14].
3. Multi-Crops Multi-Class Federated
Learning (MC2FL) 4. Multi-Crop Analysis Using
The framework facilitates label sharing across multiple crops, Multi-Regression
preserving privacy and enhancing personalized learning, with
By fitting a linear equation to observed data, multiple linear
improved performance evaluated using multiple crop datasets
with more and smaller features by Qiusi Zhang, 2023[17]. regression models reveal the connection between features
Agricultural with FL is a cooperative learning method that and responses, comparable to basic regression. It evaluates
uses multiple devices or servers to train algorithms, providing factors affecting predicted output and their relationships by
reliable models without data exchange and increased security Orogun Okunola Adebola, 2021[15].
by Salman Khan, 2022[18]. Y = b0 + b1 * x1 + b2 * x2 + b3 * x3 + …… bn * xn (1)
Wearable technology is a focus of the Federated Learning
x1, x2, x3…xn are several independent variables, and Y is the
architecture based on Agriculture by Gupta A, 2021[8].
dependent variable.
This tackles issues like privacy concerns and a lack of
personalization in smart agriculture. Transfer learning is Where b0, b1, b2 are the estimates of b0, b1, b2. The three
used to develop well-tailored models as it gathers farming normal equations are.
information from individual crops grown in different regions 6 6 6
by Prabhat kumar, 2021[16]. The framework assumes several
crop farming locations, one server, and a process. Deep
Ây i = nb0 + b1 Âx1 + b2 Âx2 (2)
i=1 i=1 i=1
learning is used by the Agriculture with FL architecture to
train classifiers and learn features. New farmer crop data Âx1i y i = b0 Âx1i + b1 Âx1i2 + b2 Âx1i x2i (3)
is regularly fed into the models to keep them updated by
Kuradusenge, 2023[10]. The framework can incorporate other Âx2i y i = b0 Âx2i + b1 Âx22i + b2 Âx2i2 (4)
methods for personalization and can be generalized. Tested
490 Algorithms in Advanced Artificial Intelligence
using artificial neural network approach, International 11. Liu Y, Yuan X, Xiong Z, Kang J, Wang X, and Niyato D:
Journal of Innovative Research in Electrical, Electronics, Federated learning for 6g communications: Challenges,
Instrumentation and Control Engineering,Vol-2, Issue-1, pp methods, and future directions, arXiv preprint arXiv:
683-6, 2014. 2006.02931, 2020.
6. Ersin Elbasi, Chamseddine Zaki, Ahmet E. Topsoiled 12. Manoj T et.al: A FL-based crop yield prediction for Agriculture
Abdelbaki,Aymen I. Zreikat, Elda Cina, Ahmed Shdefat, production risk management, IEEE Xplore, 2022.
Louai Saker: Crop Prediction Model Using Machine Learning 13. Madhuri Shripathi Rao et al: Crop prediction using machine
Algorithms, Applied Sciences, MDPI, Volume 13, Issue 16, learning, Journal of Physical Conference Series, 2161 012033,
Doi:10.3390/app13169288,2023. 2022.
7. Geiping J, Bauermeister H, Dröge H, and Moeller M: Inverting 14. Orogun Okunola Adebola:A Multiple Linear Regression
gradients - how easy is it to break privacy in federated learning, Model for Analyzing and Predicting Educational Development
ArXiv, vol. abs/2003.14053, 2020. in Nigeria, NIPES Journal of Science and Technology
8. Gupta A, Nagda D, Nikhare P, Sandbhor A: Smart crop Research, 3(1), pp:99-108, pISSN: 2682-5821,2021.
prediction using IoT and machine learning International 15. Prabhat kumar et.al: PEFL: Deep Privacy-Encoding-Based FL
Journal of Engineering Research & Technology (IJERT), Vol- Framework for smart Agriculture”, september16, 2021.
9, Issue-3, 2021. 16. Qiusi Zhang et.al: Maize yield prediction using Federated
9. Hoffman, A. L., Kemanian, A. R., and Forest, C. E.: Analysis random forest, 14 May 2023.
of climate signals in the crop yield record of sub-Saharan 17. Salman Khan Et.al: Federated learning-based UAVs for the
Africa, Global Change Biol., 24(1),143–157, doi: 10.1111/ diagnosis of plant diseases, proceedings of the 8th ICEET, 27
gcb.13901, 2018. 28, October 2022.
10. Kuradusenge, M.; Hitimana, E.; Hanyurwimfura, D.; 18. Sheathe, N: Detecting Multicollinearity in Regression
Rukundo, P., Mtonga, K., Mukasine, A., Uwitonze, C.; Analysis, Am. J. Appl. Math. Stat .8, 39–42, 2020.
Ngabonziza, J., Uwamahoro, A. Crop Yield Prediction
Note: All the figure and table in this chapter were designed by the
Using Machine Learning Models: Case of Irish Potato and
author.
Maize, Agriculture 2023, 13, 225, 2023.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
492 Algorithms in Advanced Artificial Intelligence
Empowering Inclusive Communication:
Advancements in Wearable Technology
with GloSign—A Glove-Based Solution for 74
Seamless Sign Language Interaction
L V Srinivas1,
Department of Computer Science and Engineering,
SRKR Engineering College (A), Bhimavaram, Andhra Pradesh, INDIA.
R. Shiva Shankar2,
Department of Computer Science and Engineering,
SRKR Engineering College (A), Bhimavaram, Andhra Pradesh, INDIA.
N. Deshai3,
Department of Information Technology,
SRKR Engineering College (A), Bhimavaram, Andhra Pradesh, INDIA.
K. Sravani4,
Department of Computer Science and Engineering,
SRKR Engineering College (A), Bhimavaram, Andhra Pradesh, INDIA.
V. Maheswararao5
Department of Computer Science and Engineering,
Shri Vishnu Engineering College for Women (A), Bhimavaram, Andhra Pradesh, INDIA.
Abstract: The profound psychological and social impacts of losing the ability to speak or to hear emphasize the critical demand
for innovative communication alternatives. Sign Language plays a pivotal role as a medium for individuals confronting these
challenges, fostering meaningful interaction. This paper presents GloSign, a groundbreaking glove engineered to transcend
communication barriers by translating American Sign Language into characters. Integrating cutting-edge flex and inertial
measurement unit (IMU) sensors, GloSign precisely identifies gestures, and its seamless integration with an IoT platform
ensures portability and wireless functionality. Incorporating the k-nearest neighbors (KNN) in machine learning propels the
system to an impressive accuracy rate of 97.7%. Beyond mere gesture translation, GloSign facilitates the formation of complete
sentences, offering output options ranging from on-screen display to speech conversion. This versatile glove emerges as a
potent communication tool, fostering seamless interaction among sign language users and extending its impact to individuals
unfamiliar with sign language. In doing so, it actively promotes inclusivity in communication, making strides toward a more
interconnected and accessible future.
Keywords: Glove, Pattern recognition, Gesture recognition, Sign language sensor
DOI: 10.1201/9781003529231-74
Empowering Inclusive Communication: Advancements in Wearable Technology with GloSign 493
individuals with impairments to communicate effectively. This paper adopts a novel structure comprising four distinct
Additionally, SL faces limitations in digital communication, sections, each providing a groundbreaking topic exploration.
and the absence of a universal standard means that each Following this, the methodology section intricately outlines
country has its unique system with gestures that may the innovative approach employed in this study, prioritizing
differ from those of other nations. In response to these transparency and offering valuable insight into the research
communication challenges, this study introduces GloSign, process [6, 7, 8].
a novel glove designed to transcend barriers by translating The subsequent section, results, goes beyond mere
SL into English [2]. Like spoken languages, American presentation, offering a detailed and nuanced account of
Sign Language (ASL) consists of formal and informal the findings derived from the conducted research. Here, it
components. This paper focuses explicitly on the legal Ensures a comprehensive understanding of the outcomes and
aspect of ASL, which involves a set of 26 alphabets used to their potential implications [9].
construct words and sentences. These alphabets are defined
by four elements: hand shape, position relative to the body, In a culmination of insights, the fourth section serves as the
hand movements, and palm alignment, with specific gestures conclusion, meticulously summarizing key findings and
requiring dynamic hand movements.GloSign incorporates contributions of the paper. It not only provides closure but
flex sensors, accelerometers, and gyroscopes to recognize also goes a step further by delineating potential avenues for
gestures, offering a wireless, IoT-enabled solution for future research, thereby propelling the exploration of the
uploading and analyzing data [3]. The system interprets topic into uncharted and innovative territories. This structured
gestures captured by the glove, transforming them into words approach enhances the paper’s clarity and contributes to its
and sentences. These linguistic outputs are then displayed on distinctiveness in the academic landscape [10].
a screen running gesture recognition software, accompanied
by converting sentences into speech. By facilitating the 2. Related Work
translation of SL into a widely understood language, GloSign
strives to significantly enhance communication accessibility The literature review is meticulously organized based on
for individuals facing these challenges [4-5]. the categorization of sensors, communication methods,
employed algorithms, and the presentation of gestures. In
[11, 12], showcased gloves equipped with flex sensors that
successfully predicted all alphabets with commendable
accuracy, serving their purpose in assisting communication
for individuals with disabilities [13, 14]. Authors in [15] and
A B C D E F G
[16] innovatively developed wireless gesture decoders using
Plam is usually forward Thumb bent Thumb frequently
forward
flex sensors and accelerometers, achieving a remarkable 95%
unless stated. lowers (like claw)
accuracy in recognizing all alphabets and 15 words. In [17],
incorporating pressure sensors elevated accuracy to 98.2%.
A different device with flex sensors could detect movements
by measuring voltage levels. It could then display the
H I J K L M detected movements on a phone or laptop via Bluetooth with
Palm in an accuracy rate of 83% [18]. Building on this technology,
[19, 20] used flex sensors and accelerometers to recognize
movements and convert them into sound and text in just 0.74
seconds.
N O P Q R S The media report regular occurrences. Now, social networks
One palm facing Palm in Index Similar to P
like Twitter give user-generated news material. They
opposite body. Finger but pointing must cluster and deliver only relevant data to make this
indicates. down and resource usable. They filtered data using density-based
curled finger. k-means and graph clustering [21]. Based on a thorough
categorization process, the IDCS has four steps. IDCS begins
by methodically collecting data from diverse sources on
rural women’s nutritional knowledge, attitudes, and dietary
T U V W X Y Z patterns. After that, it preprocesses the data according to
Thumb over fingers, palm front, palm tilted away from body conventional methods and organizes it for categorization.
Fig. 74.1 Sign language in america Next, a learning algorithm develops an intelligent classifier
494 Algorithms in Advanced Artificial Intelligence
to divide rural women into appropriate nutritional categories Thiemjarus fitted the glove with more touch sensors [31]. In
[22]. another investigation, a KNN algorithm and touch sensors
Use sentiment classifiers to classify comments and visualize yielded 91.54% efficiency [34]. Ahmed and Arif et al. [35,
student perspectives. This technology offers instructors fast 36] created a glove with touch, flex, and inertial sensors
qualitative feedback to enhance student learning. This program that achieved 92% accuracy using a gesture recognition
enables quicker online feedback than paper-based systems. algorithm. Using surface Electromyography (EMG) sensors
The current approach is inefficient for student input. Hence, and accelerometers, Wu et al. [37] achieved 96% accuracy
an online system was established. Students will provide for 80 movements. Python-coded capacitive touch sensors
comments online using a standard form. The suggested by Abhishek et al. [38] responded in 0.7 seconds with 92%
approach prioritizes security by allowing only legitimate users accuracy. Mehdi and Khan [39] employed a tilt sensor to
to see and understand the cumulative input and opinions of a measure glove rotation in a 5DT 7-sensor glove. A neural
batch of pupils [23]. Evaluation is the primary way schools network and three-layer algorithm gave this system 88%
assess students’ learning capabilities. One of the main tests gesture recognition accuracy. In [40], [41], Immersion’s
is essay writing. Currently, this assessment is done manually. 18-sensor CyberGlove with a resistive bend, abduction, and
Here, it takes time and effort. For this project, this assessment flexion sensors achieved 90% accuracy but was non-real
method is automated. They used keywords as features since time. A revolutionary glove designed to translate gestures
machines can’t grasp our rating metrics. into alphabets, offering a range of cutting-edge features:
• Wireless and portable functionality [42]
First, the student essay is matched against admin keywords
[24]. Linguistic diversity and variations hampered • Real-time responsiveness
research into Indian sign language (ISL). SL is required to • Capability to form words and sentences [43]
communicate with one another. Most learning takes place via • Accessibility from anywhere through an IoT platform
interaction with peers. Sign-learning materials are seldom Modern technology allows us complex answers to common
available. Because of this, learning to sign is a difficult task. problems. The medical glove literature review examines
Sign learning begins with the first stage, which is called features and drawbacks. The highlighted glove fills holes
finger spelling. This stage is used when there is no applicable left by its predecessors. As seen in the results section, it
sign, or the signatory is unknown [25, 26]. Smoke from the gives real-time results, essential in urgent circumstances. Its
fire causes vehicle accidents owing to poor visibility. simplicity and accessibility make it appealing to healthcare
Unaware of woodland fires and smoke, people died while practitioners. By resolving the flaws of previous gloves, this
walking down the forest route. The Raspberry Pi-interfaced novel approach may significantly improve patient care [44].
sensor detects forest fires in real-time and alerts the system.
This system uses the Publish-Subscribe Protocol to identify 3. Methods
forest fires and warn users [27]. The localization of hand
joint positions under kinematic limits is accomplished by Translating sign language into English using the GloSign
implementing a hierarchical mode-seeking technique that has glove is revolutionary and consists of a series of novel
been developed. At long last, an ASL signal may be identified processes. With the careful selection and positioning of
using a Random Forest (RF) classifier based on a combined sensors on the glove, the primary focus is on recording the
angle. While testing this method, they used a dataset from hand motions of the user with an unmatched accuracy. This
Surrey University [28]. The IoT and cloud computing have method is essential for the accurate interpretation of SL
seen tremendous growth, contributing to the optimization of motions, which lays the groundwork for the revolutionary
supply times and the retention of spare parts in warehouses. capabilities of the glove. A dynamic conduit for data transfer
Therefore, out-of-stock situations are optimized. A passive to a computer or mobile device will be established via the
UHF-RFID tag/sticker system to arrange extra components is subsequent step, which comprises the seamless integration of
the primary goal of this effort [29-30]. the GloSign glove with an Internet of Things (IoT) platform.
Tanyawiwat and Thiemjarus [31] developed GesTALK, The connection in question is not only a technical need
a portable gesture recognition glove that converts static but also a doorway to the understanding of SL motions in
motions to speech and works 90% accurately in ASL and continuous time. It can efficiently assist communication
PSL. A glove with contacts and flex sensors translated eight between persons who are deaf or hard of hearing and others
SLs to text with 93.16% accuracy [32]. El-Din and El- who can listen, therefore breaking down boundaries in a way
Ghany [33] achieved 88% accuracy with dynamic motions that has never really been seen before.
from ASL and ArSL using a Python GUI application and In the last stage, the data collected by the GloSign glove
a glove with flex and inertial sensors. Tanyawiwat and is subjected to a complex process of interpretation. The
Empowering Inclusive Communication: Advancements in Wearable Technology with GloSign 495
interpretation is achieved through machine learning (ML) Navigating the intricate nuances of ASL, where similar
algorithms extensively trained on significant sign language gestures often exhibit subtle differences, presents a
(SL) datasets. These advanced algorithms can quickly identify formidable challenge. In a groundbreaking move to overcome
subtle patterns in hand gestures and translate them into this hurdle, incorporating a contact sensor emerges as a
English words or phrases. As a result of this comprehensive pioneering solution. This sensor is strategically introduced
and innovative procedure, GloSign has established itself as to meticulously discern the nuanced variations inherent
a trailblazer in the realm of deaf-hearing communication, in these closely related gestures. The connection setup for
effectively bridging the gap between individuals who are deaf the contact sensor mirrors the innovative configuration
and those who are hearing. This action not only contributes showcased in Fig. 74.2, embodying a synergy of technology
to the development of a global communication environment and precision. This strategic positioning not only elevates the
that is more accessible and inclusive, but it also serves as a sensor’s efficacy but also contributes to the overall finesse
demonstration of the potential that innovation has in terms of of the system, ensuring that even the most intricate gestures
promoting meaningful relationships. are accurately captured and interpreted. This integration of
contact sensors transcends the conventional, pushing the
3.1 Selection and Placement of the Sensors boundaries of SL interpretation technology. It doesn’t merely
Flex sensors, contact sensors, and an inertial measurement detect gestures; it refines the art of distinction, embodying
unit (IMU) sensor with an accelerometer and gyroscope are a testament to the innovative spirit underlying GloSign.
carefully selected and positioned innovatively. The GloSign By embracing advanced sensor technologies and strategic
glove’s sensor array relies on flex sensors selected for finger- placement strategies, this glove doesn’t just interpret ASL;
angle measurement capabilities. Resistance levels fluctuate it revolutionizes the precision and clarity with which it’s
dynamically with finger angles in these sensors. Flex sensors done, setting new standards for inclusive and effective
are placed above the glove to record the wearer’s gentle finger communication [46]. Unlocking the realm of dynamic
motions for maximum accuracy. gestures leaps into the future by integrating the Inertial
Connecting this intricate sensor network is where the true Measurement Unit (IMU) sensor, seamlessly woven into
innovation unfolds. The flex sensors connect seamlessly the Arduino and strategically perched atop the hand. The
with the Arduino NANO IoT, creating a sophisticated innovative placement of this technological marvel is elegantly
communication pathway. The configuration, visually captured in the visual representation presented in Fig. 74.3,
represented in Fig. 74.2, is designed to harness the full showcasing the glove as a convergence point of cutting-edge
potential of the flex sensors. In this setup, the initial pin of sensor technologies.
the flex sensor (colored red) connects to the 3.3v port of the
Arduino, establishing a vital power link. Meanwhile, the
opposing pin (colored blue) is strategically linked to a resistor,
a crucial element in fine-tuning the sensor’s response. The
connection preceding the resistor is then skillfully attached
to the Arduino’s analog input, ensuring a streamlined data
flow. Finally, the remaining connection (colored black) is
grounded, completing the intricate circuitry that forms the
backbone of GloSign’s sensor system.
journey from the Arduino to the IoT platform. Within the 4. Results and Discussion
intricate web of connectivity, the IoT platform becomes a
dynamic canvas, visualizing changes in gestures through a This section unravels the innovative outcomes from the
scatter plot feature. This real-time monitoring capability intricate dance of the GloSign glove’s operations. The
amplifies the GloSign glove’s responsiveness, setting it apart system’s three acts showcase its revolutionary capabilities.
as a groundbreaking tool in SL interpretation. In the first phase, the GloSign glove transmits a plethora of
The saga doesn’t end here; the values extracted from this data to the vast Internet of Things (IoT) infrastructure. The
digital canvas are ingeniously transported to the PC through next step involves analyzing and decoding the raw data. The
the IBM Watson IoT Software Development Kit (SDK). Figure glove software carefully interprets complex motions using
74.3 vividly captures this data flow, illustrating the seamless ML and K-Nearest Neighbors (KNN) methods. It’s a ballet
transfer from the glove to the PC, unlocking possibilities for of computation, where patterns are unraveled, and gestures
further processing and analysis. This methodology doesn’t are translated into a language comprehensible to the digital
merely capture dynamic gestures; it orchestrates a symphony realm—the crescendo of innovation peaks in the final phase,
of data, seamlessly blending sensor technologies, IoT where the processed data takes center stage. What emerges is
integration, and real-time monitoring. The GloSign glove, not just a stream of letters; it’s a refined output, a testament
equipped with this ensemble of innovations, emerges as a tool to the system’s ability to construct meaningful sentences.
for gesture identification and a visionary leap into the future The narrative unfolds on a screen, where coherent sentences,
of inclusive and intelligent communication technologies. meticulously crafted from the language of gestures, come to
life. This isn’t just data; it’s a story, a dialogue, an eloquent
3.2 Data from Glove expression born from the fusion of technology and human
communication. In essence, each phase in the GloSign
The GloSign glove collects data, which is then sent to IBM
glove’s operation isn’t just a step; it’s a leap into the future of
Watson IoT for conversion. This phase is crucial as it enables
communication technology. It’s an orchestration of data, an
new methods of interpreting gestures. This sensor data is
analysis of meaning, and a presentation of coherent output.
accurately mapped and calibrated onto characters during the
This isn’t mere experimentation; it’s a journey into the
offline system training. The system’s accuracy is achieved
uncharted territories of innovative communication solutions,
through an alphabet categorization process using the K-Nearest
where the language of gestures becomes a bridge connecting
Neighbors (KNN) method. An iterative method is used to
diverse worlds.
analyze K values from 1 to 25, and K values are optimized
for precision and system speed based on iteration accuracy. 4.1 IoT Platform and Sensors
This sophisticated algorithm can effortlessly process IBM
Watson IoT data and accurately predict the nearest letters. In this avant-garde experiment, the technological synergy
The system incorporates a visionary gesture fix algorithm unfolds by utilizing the IBM Watson IoT platform, creating
during sentence formation. This algorithm not only rectifies an ingenious connection between the GloSign glove, adorned
potential duplicate or incorrect letter predictions but does so with the Arduino NANO IoT, and the digital realm through
comprehensively, addressing entire sentences. The process is seamless Wi-Fi communication. The orchestration of data
swift, typically taking 2 to 8 seconds, showcasing the system’s becomes an art, dissected into two distinct realms: the
efficiency even with varying sentence lengths. The gesture nuanced language of the flex sensors and the binary poetry
fix algorithm operates on the entire sentence concurrently, of the contact sensor and Inertial Measurement Unit (IMU).
sidestepping delays that might arise when addressing words Like binary sentinels, the IMU and contact sensors manifest
individually. This forward-thinking approach ensures a their outputs in Boolean. With a profound simplicity, contact
seamless correction process, eliminating any risk of letter sensors echo a 1 when touched and gracefully bow to a 0
loss during the preceding word’s processing. A complex without contact. The IMU sensor, a judge of motion, breathes
procedure produces clear messages on a screen designed life into the glove’s dynamism, affirming its existence with a
for gesture recognition. The story continues with the IBM regal one during motion and gracefully transitioning to a 0 in
Watson Text-to-Speech software development kit converting moments of stillness. The values oscillate between the binary
these messages into spoken words. This versatile tool is ballet of 0s and 1s, where contact and movement values paint
easily accessible, works on any computer, requires a stable a canvas of interaction. A contact value of 1 whispers the
internet connection, and is compatible with Python software. tale of touch, while a movement value of 1 choreographs the
As the system performs its dynamic processes, variations in dance of dynamic gestures. Each with a balletic range of 0
performance may occur based on the specifications of the to 90°, grace the stage with the elegance of their numerical
system in use, adding a layer of adaptability to GloSign’s expressions. Like fingers in a symphony, F1, F2, F3, F4,
innovative communication landscape. and F5 waltz through the spectrum; they shape the nuanced
Empowering Inclusive Communication: Advancements in Wearable Technology with GloSign 497
Table 74.1 Average sensor values for each gesture Yet, in this digital ballet, a revelation surfaces. The sensor
Alphabet F1 F2 F3 F4 F5 C M
values, a testament to the intricacies of signed communication,
reveal a mesmerizing similarity among gestures. “i” and “j,”
A 44.04 26.97 40.85 53.29 0.28 1 0
akin to kindred spirits, share sensor values so profound that
B 0.56 -4.48 -3.47 -0.04 31.32 1 0 they dance on the edge of indistinguishability, differentiated
C 20.26 31.82 53.51 34.5 2.93 1 0 only by the subtle heartbeat of the movement sensor value.
D 32.01 42.04 51.73 0.13 -1.50 0 0 Embarking on a visual odyssey, Fig. 74.4 unfolds as a portal
E 66.83 71.89 73.64 58.29 35.41 1 0
into the dynamic realm of GloSign’s data visualization.
F 5.98 0.80 1.22 69.99 5.40 0 0 This avant-garde depiction transcends mere representation,
offering an immersive experience that illuminates the
G 76.21 66.59 83.98 -0.50 4.54 0 0
intricacies of various gestures. Within this digital canvas,
G 59.45 56.08 -4.44 -0.03 12.33 0 0
the sections marked with the resonant symbols “C” and “M”
I 3.45 79.17 65.58 79.62 17.40 1 0 emerge as visual symphonies, each stroke and hue capturing
J 2.78 71.62 60.66 74.46 15.61 1 1 the essence of communication. The visualization pulsates with
K 70.55 62.95 -6.81 -0.66 -0.64 0 0 life in the turquoise embrace of “C” (contact sensor). A surge
L 83.53 73.88 77.67 -1.13 -0.82 0 0
in this azure tide signifies the poetic convergence of contact
sensors, painting a vivid tableau of tactile connection. The
M 77.58 57.87 53.03 54.70 10.29 1 0
subtle dance of turquoise unveils the nuanced choreography
N 90.52 71.54 58.45 61.96 -2.11 1 0 of touch, echoing the very heartbeat of human connection. In
O 44.81 43.63 51.74 42.93 7.46 1 0 tandem, the ethereal glow of “M” (dynamic gesture) bathes
P 57.93 49.62 2.35 -0.80 -3.45 0 0 the canvas in light turquoise. A celestial dance ensues as this
Q 74.20 69.96 68.03 -0.21 -2.14 0 0
luminescence responds to the slightest quiver of movement
within the glove. It is not just data; it’s a visual symphony,
R 60.91 47.69 -5.54 -1.10 6.88 0 0
an elegant ballet of interaction painted on the canvas of
S 95.83 78.83 71.81 75.64 35.59 1 0 digital expression. Figure 74.5, a tapestry of readings from
T 76.92 34.22 40.41 52.19 -68 0 0 the flex sensors, invites observers into the tranquil realm
U 82.84 31.52 -6.82 -1.11 13.97 1 0 of hand gestures at rest. Five protagonists, F1 through F5,
V 77.50 38.18 -7.07 -0.76 14.15 0 0 stand poised to reveal the silent eloquence of the fourth,
third, second, and first fingers and the Thumb. The chart
W 53.30 -4.28 -7.13 -0.47 12.56 0 0
unfolds as a serenade of numerical expression, where each
X 76.99 75.49 70.31 58.34 41.50 0 0 curve and contour represents the nuanced articulation of hand
Y 1.06 74.33 68.31 80.23 -1.50 1 0 postures in repose. It is not just a chart; it’s a testament to the
Z 64.54 62.94 73.39 -0.73 7.17 0 1 intricate language encoded in the resting hands, awaiting the
moment they shall dance again in the symphony of signed
gestures of the fourth, third, second, first fingers, and the communication. In essence, these figures are visualizations
Thumb. and portals into gesture poetry. They transform raw data into
a visual tapestry, where hues and contours resonate with symphony of ML and signed communication. It’s a pioneering
the very essence of human expression. This is not just data; exploration into the nuances of gesture identification, where
it’s an innovative exploration into the visual language of the system’s accuracy becomes a poetic expression, echoing
communication, where every stroke is a note, and every hue the challenges inherent in the language of silent conversation.
is a melody in the symphony of GloSign. Venturing into innovation, our quest for precision led us to
a groundbreaking exploration of the K-nearest neighbors
4.2 Analyzing and Decoding Data
(KNN) algorithm. To fine-tune its accuracy, we embarked on
The IBM Watson IoT platform Python API is a crucial tool a symphony of experiments, dancing through a spectrum of K
in data analysis. It helps to transform IoT platform data values to discern the optimal configuration for our dynamic
into a more usable format. A KNN-supervised ML model framework. Figure 74.7 is a visual testament to this journey, a
orchestrates this transformation. This model is based on a canvas where the average mean errors gracefully unfold, each
gesture classification master that uses 200 different gestures stroke representing a different K value. This isn’t just a chart;
to convey information. These gestures are like a dance it’s a visual aria, a poetic representation of our meticulous
of the hands, based on ASL. The system uses a model to quest for perfection.
categorize gestures by taking precise readings and training
The graph unveils a narrative where the K values 1 and
the KNN model. The motions are divided into two groups
3 emerge as protagonists, each carrying a melody of
for testing the model. The first batch, which contains 75%
precision. In this innovative exploration, the average mean
of the data, introduces the model’s neural connections. The
errors become the notes in our symphony. They echo our
system’s accuracy is then tested using the second batch,
algorithmic ballet’s subtle nuances and intricacies, with
which includes 25% of the data. Figure 74.4 displays a bar
each K value contributing a unique cadence. The revelation,
chart that shows the system’s accuracy in recognizing letters
painted on this digital canvas, is profound—K values 1 and 3
when K was 1 (1-NN). The chart, with its peaks of 100%
are beacons of optimal performance, adorned with an average
accuracy, unveils the system’s prowess in decoding the
error rate below the ethereal threshold of 0.5%. This isn’t
language of gestures. Yet, within this symphony, a poetic
just an experiment; it’s a poetic exploration of the optimal
discord surfaces. Dynamic gestures and an ever-elusive
configuration of our algorithmic orchestra. It’s a testament to
cadence reveal an accuracy slightly shy of perfection, dipping
our commitment to precision, where every K value is a note,
below 95%. The dance between “i” and “j,” intertwined in
and every average mean error is a resonance in the symphony
similarity, unveils a challenge of differentiating one from
of innovation. The journey continues, but in the harmonious
the other, with movement being the subtle heartbeat that
echoes of K values 1 and 3, we find a melody of excellence
sets them apart. The letters “h” and “r,” standing shoulder to
that propels our framework into precision and proficiency.
shoulder in sensor readings, add another layer to the nuanced
ballet. However, in this intricate dance, “j” emerges as the Embarking on a visionary exploration, Fig. 74.5 unfurls as
prima donna of complexity, its identification proving to be a digital tapestry, capturing the essence of our relentless
the most challenging feat. This is not just a chart; it’s a sonnet pursuit of precision in recognizing gestures. This isn’t just
of accuracy and challenges, a testament to the evolving a chart; it’s a symphony of accuracy, a visual composition
Empowering Inclusive Communication: Advancements in Wearable Technology with GloSign 499
5. Conclusion
The study introduces GloSign, a glove that converts SL
movements into words and phrases. The glove’s IMU and flex
Fig. 74.6 Accuracy of k values sensors send sign-language gesture data to the IBM Watson
IoT platform. The system employs KNN to distinguish
This article looks closer at gloves that can interpret ASL between complex or similar movements. The gesture repair
gestures. By incorporating ML and sentence-level error technique corrects word and sentence errors once identifying
correction, the accuracy of the system’s output has been letters are combined into sentences. The system outputs the
500 Algorithms in Advanced Artificial Intelligence
converted text both visually and audibly for user convenience. International Conference on Power, Control, Signals and
Further research is required to enhance the accuracy and Instrumentation Engineering, ICPCSI 2017, pp. 1840–1844,
speed of sign language gesture verification. 2018, doi: 10.1109/ICPCSI.2017.8392033.
13. S. M. Biju, H. Z. Sheikh, M. F. Malek, F. Oroumchian, and A.
Bell, “Design of grip strength measuring system using fsr and
References flex sensors using svm algorithm,” IAES International Journal
of Artificial Intelligence, vol. 10, no. 3, pp. 676–686, 2021,
1. G. Kumar, M. K. Gurjar, and S. B. Singh, “American sign doi: 10.11591/IJAI.V10.I3.PP676-686.
language translating glove using flex sensor,” Imperial journal 14. S. M. Biju and H. Z. Sheikh, “Sensor evaluation for hand grip
of interdisciplinary research 2, no. 6, pp. 1439–1441, 2016. strength,” International Journal of Electrical and Computer
2. S. Kumuda and P. K. Mane, “Smart assistant for deaf and dumb Engineering (IJECE), vol. 12, no. 5, p. 4756, Oct. 2022, doi:
using flexible resistive sensor: implemented on LabVIEW 10.11591/ijece.v12i5.pp4756-4764.
platform,” Proceedings of the 5th International Conference on 15. S. Bin Rizwan, M. S. Z. Khan, and M. Imran, “American sign
Inventive Computation Technologies, ICICT 2020, pp. 994– language translation via smart wearable glove technology,”
1000, 2020, doi: 10.1109/ICICT48043.2020.9112553. RAEE 2019 - International Symposium on Recent
3. Rao VV, Silpa N, Gadiraju M, Shankar RS, Vijaya K. An Advances in Electrical Engineering, 2019, doi: 10.1109/
Optimal Machine Learning Model Based On Selective RAEE.2019.8886931.
Reinforced Markov Decision To Predict Web Browsing 16. B. G. Lee and S. M. Lee, “Smart wearable hand device for
Patterns. Journal of Theoretical and Applied Information sign language interpretation system with sensors fusion,”
Technology. 2023 Jan 31;101 (2):859-73. IEEE Sensors Journal, vol. 18, no. 3, pp. 1224–1232, 2018,
4. Reddy SS, Gadiraju M, Maheswara Rao VV. Analyzing Student doi: 10.1109/JSEN.2017.2779466.
Reviews on Teacher Performance Using Long Short-Term 17. R. Ambar, C. K. Fai, M. H. Abd Wahab, M. M. Abdul
Memory. InInnovative Data Communication Technologies Jamil, and A. A. Ma’Radzi, “Development of a Wearable
and Application: Proceedings of ICIDCA 2021 2022 Feb 24 Device for Sign Language Recognition,” Journal of Physics:
(pp. 539-553). Singapore: Springer Nature Singapore. Conference Series, vol. 1019, no. 1, 2018, doi: 10.1088/1742
5. Starner T, Pentland A. Real-time american sign language 6596/1019/1/012017.
recognition from video using hidden markov models. 18. S. Vutinuntakasame, V. R. Jaijongrak, and S. Thiemjarus, “An
InProceedings of International Symposium on Computer assistive body sensor network glove for speech- and hearing-
Vision-ISCV 1995 Nov 21 (pp. 265-270). IEEE. impaired disabilities,” Proceedings - 2011 International
6. Shiva Shankar R, Ravibabu D. Digital Report Grading Using Conference on Body Sensor Networks, BSN 2011, pp. 7–12,
NLP Feature Selection. InSoft Computing in Data Analytics: 2011, doi: 10.1109/BSN.2011.13.
Proceedings of International Conference on SCDA 2018 2019 19. K. Kadam, R. Ganu, A. Bhosekar, and S. D. Joshi, “American
(pp. 615-623). Springer Singapore. sign language interpreter,” Proceedings - 2012 IEEE 4th
7. Chuan CH, Regina E, Guardino C. American sign language International Conference on Technology for Education, T4E
recognition using leap motion sensor. In2014 13th International 2012, pp. 157–159, 2012, doi: 10.1109/T4E.2012.45.
Conference on Machine Learning and Applications 2014 Dec 20. M. S. Amin, M. T. Amin, M. Y. Latif, A. A. Jathol, N. Ahmed, and
3 (pp. 541-544). IEEE. M. I. N. Tarar, “Alphabetical gesture recognition of American
8. Shankar RS, Rajanikanth J, Sivaramaraju VV, Murthy KV. sign language using e-voice smart glove,” Proceedings - 2020
Prediction of employee attrition using datamining. In2018 ieee 23rd IEEE International Multi-Topic Conference, INMIC
international conference on system, computation, automation 2020, 2020, doi: 10.1109/INMIC50486.2020.9318185.
and networking (icscan) 2018 Jul 6 (pp. 1-8). IEEE. 21. Shankar RS, Murthy KV, Rao CS, Gupta VM. An approach
9. VVR MR, Silpa N, Gadiraju M, Reddy SS, Bonthu S, for extracting tweets from social media factors. In2018 ieee
Kurada RR. A Plausible RNN-LSTM based Profession international conference on system, computation, automation
Recommendation System by Predicting Human Personality and networking (icscan) 2018 Jul 6 (pp. 1-7). IEEE.
Types on Social Media Forums. In2023 7th International 22. Maheswara Rao VV, Silpa N, Mahesh G, Reddy SS. An En
Conference on Computing Methodologies and Communication hanced Machine Learning Classification System to Investigate
(ICCMC) 2023 Feb 23 (pp. 850-855). IEEE. the Status of Micronutrients in Rural Women. InProceedings
10. Assan M, Grobel K. Video-based sign language recognition of International Conference on Recent Trends in Computing:
using hidden markov models. InInternational Gesture ICRTC 2021 2022 (pp. 51-60). Springer Singapore.
Workshop 1997 Sep 17 (pp. 97-109). Berlin, Heidelberg: 23. Pigou L, Dieleman S, Kindermans PJ, Schrauwen B. Sign
Springer Berlin Heidelberg. language recognition using convolutional neural networks.
11. H. Joshi, S. Bhati, K. Sharma, and V. Matai, “Detection of fin InComputer Vision-ECCV 2014 Workshops: Zurich,
ger motion using flex sensor for assisting speech impaired,” Switzerland, September 6-7 and 12, 2014, Proceedings, Part
International Journal of Innovative Research in Science, Engi I 13 2015 (pp. 572-578). Springer International Publishing.
neering and Technology, vol. 6, no. 10, pp. 20798–20804, 2017. 24. Shankar RS, Srinivas LV, Ravibabu D, Raminaidu C. Novice
12. G. Sabaresh and A. Karthi, “Design and implementation of a Retroaction Report. ARPN Journal of Engineering and
sign-to-speech/text system for deaf and dumb people,” IEEE Applied Sciences. 2006;13.
Empowering Inclusive Communication: Advancements in Wearable Technology with GloSign 501
25. Metaxas D, Dilsizian M, Neidle C. Scalable ASL sign 37. S. S. Ahmed, H. Gokul, P. Suresh, and V. Vijayaraghavan,
recognition using model-based machine learning and “Low-cost wearable gesture recognition system with minimal
linguistically annotated corpora. Insign-lang@ LREC 2018 user calibration for asl,” Proceedings - 2019 IEEE International
2018 May 12 (pp. 127-132). European Language Resources Congress on Cybermatics: 12th IEEE International
Association (ELRA). Conference on Internet of Things, 15th IEEE International
26. Shankar RS, Babu DR, Murthy KV, Gupta V. An approach Conference on Green Computing and Communications, 12th
for essay evaluation using system tools. In2017 International IEEE International Conference on Cyber, Physical and Social
Conference on Innovative Research In Electrical Sciences Computing and 5th IEEE International Conference on Smart
(IICIRES) 2017 Jun 16 (pp. 1-9). IEEE. Data, iThings/GreenCom/CPSCom/SmartData 2019, pp.
27. Srinivas LV, Raminaidu C, Ravibabu D, Reddy SS. A 1080–1087, 2019, doi: 10.1109/iThings/GreenCom/CPSCom/
framework to recognize the sign language system for deaf SmartData.2019.00185.
and dumb using mining techniques. Indonesian Journal 38. A. Arif, S. T. H. Rizvi, I. Jawaid, M. A. Waleed, and M. R.
of Electrical Engineering and Computer Science. 2023 Shakeel, “Techno-talk: an American sign language (ASL)
Feb;29(2):1006-16. translator,” International Conference on Control, Decision and
28. Chong TW, Lee BG. American sign language recognition Information Technologies, CoDIT 2016, pp. 665–670, 2016,
using leap motion controller with machine learning approach. doi: 10.1109/CoDIT.2016.7593642.
Sensors. 2018 Oct 19;18(10):3554. 39. J. Wu, L. Sun, and R. Jafari, “A wearable system for
29. Shankar RS, Gupta VM, Priyadarshini V, Neelima P. PS recognizing american sign language in real-time using IMU
protocol to detect fire in forest and fire alert system using and surface EMG sensors,” IEEE Journal of Biomedical and
sensors. InAIP Conference Proceedings 2022 Dec 9 (Vol. Health Informatics, vol. 20, no. 5, pp. 1281–1290, 2016, doi:
2576, No. 1). AIP Publishing. 10.1109/JBHI.2016.2598302.
30. Dong C, Leu MC, Yin Z. American sign language alphabet 40. K. S. Abhishek, L. C. F. Qubeley, and D. Ho, “Glove-based
recognition using microsoft kinect. InProceedings of the hand gesture recognition sign language translator using
IEEE conference on computer vision and pattern recognition capacitive touch sensor,” 2016 IEEE International Conference
workshops 2015 (pp. 44-52). on Electron Devices and Solid-State Circuits, EDSSC 2016,
31. Shiva Shankar R, Devareddi R, Mahesh G, MNSSVKR Gupta pp. 334–337, 2016, doi: 10.1109/EDSSC.2016.7785276.
V. Develop a Smart Data Warehouse for Auto Spare Parts 41. S. A. Mehdi and Y. N. Khan, “Sign language recognition
Autonomous Dispensing and Rack Restoration by Using IoT using sensor gloves,” ICONIP 2002 - Proceedings of the 9th
with DDS Protocol. InComputer Networks, Big Data and International Conference on Neural Information Processing:
IoT: Proceedings of ICCBI 2021 2022 May 22 (pp. 879-895). Computational Intelligence for the E-Age, vol. 5, pp. 2204–
Singapore: Springer Nature Singapore. 2206, 2002, doi: 10.1109/ICONIP.2002.1201884.
32. Devareddi RB, Shankar RS, Mahesh G. IoT Protocol for 42. J. M. Allen, P. K. Asselin, and R. Foulds, “American sign
Inferno Calamity in Public Transport. Integration of Cloud language finger spelling recognition system,” Proceedings
Computing with Internet of Things: Foundations, Analytics, of the IEEE Annual Northeast Bioengineering Conference,
and Applications. 2021 Mar 19:87-110. NEBEC, pp. 285–286, 2003, doi: 10.1109/nebc.2003.1216106.
33. N. Tanyawiwat and S. Thiemjarus, “Design of an assistive 43. Y. Khambaty et al., “Cost effective portable system for sign
communication glove using combined sensory channels,” language gesture recognition,” 2008 IEEE International
Proceedings - BSN 2012: 9th International Workshop on Conference on System of Systems Engineering, SoSE 2008,
Wearable and Implantable Body Sensor Networks, pp. 34–39, 2008, doi: 10.1109/SYSOSE.2008.4724149.
2012, doi: 10.1109/BSN.2012.17. 44. Mahesh G, Shankar Reddy S, Maheswara Rao VV, Silpa N.
34. A. Natesh, G. Rajan, B. Thiagarajan, and V. Vijayaraghavan, Preeminent Sign Language System by Employing Mining
“Low-cost wireless intelligent two hand gesture recognition Techniques. InInternational Conference on IoT Based Control
system,” 11th Annual IEEE International Systems Networks and Intelligent Systems 2023 Jun 21 (pp. 571-588).
Conference, SysCon 2017 - Proceedings, 2017, doi: 10.1109/ Singapore: Springer Nature Singapore.
SYSCON.2017.7934745. 45. D. Bajpai, U. Porov, G. Srivastav, and N. Sachan, “Two way
35. S. A. E. El-Din and M. A. A. El-Ghany, “Sign Language wireless data communication and American sign language
Interpreter System: An alternative system for machine translator glove for images text and speech display on mobile
learning,” 2nd Novel Intelligent and Leading Emerging phone,” Proceedings - 2015 5th International Conference on
Sciences Conference, NILES 2020, pp. 332–337, 2020, doi: Communication Systems and Network Technologies, CSNT
10.1109/NILES50944.2020.9257958. 2015, pp. 578–585, 2015, doi: 10.1109/CSNT.2015.121.
36. V. Pathak, S. Mongia, and G. Chitranshi, “A framework for 46. M. M. Chandra, S. Rajkumar, and L. S. Kumar, “Sign
hand gesture recognition based on fusion of Flex, Contact and languages to speech conversion prototype using the SVM
accelerometer sensor,” Proceedings of 2015 3rd International classifier,” IEEE Region 10 Annual International Conference,
Conference on Image Information Processing, ICIIP 2015, pp. Proceedings/TENCON, vol. 2019-Octob, pp. 1803–1807,
312–319, 2016, doi: 10.1109/ICIIP.2015.7414787. 2019, doi: 10.1109/TENCON.2019.8929356.
Note: All the figures and table in this chapter were designed by the
author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
502 Algorithms in Advanced Artificial Intelligence
R. N. V. Jagan Mohan1
Associate Professor, Dept of CSE, Sagi Rama Krishnam Raju Engineering College, Bhimavaram
https://orcid.org/0000-0003-1457-0824
Vasamsetty Chandra Sekhar2
Professor & HoD, Sagi Rama Krishnam Raju Engineering College, Bhimavaram
Abstract: Forensic voice comparison is a process where police compare a criminal’s voice to suspects’ voices to identify them,
aiding in criminal inquiries as well as legal issues by determining the likelihood of similar speech samples. Voices contain
unique information, allowing individuals to be identified [12]. To determine if two voices belong to the same person, forensic
analysts consider their similarity and distinctiveness. Voices with similar characteristics provide stronger evidence that the
suspect and perpetrator are the same person, as fewer others would have similar characteristics. Research on the benefits of
sixth-generation (6G) networks is focusing on combining Artificial Intelligence (AI) based voice assistants with B5G and 6G
free space optic technologies. The main objectives are to develop voice assistants that can respond to specific vocal commands,
carry out scheduled operations, or perform unprepared actions for law enforcement applications or crime investigations. Speech
recognition, voice synthesis, and natural language processing (NLP) are used to achieve this. Voice assistant interact with
users using voice recognition technology. The research work displays accuracy metrics for Transformers using the attention
algorithm’s outcomes.
Keywords: Artificial intelligence, Free space optic, B5G, 6G, Nature language process, Voice assistant
1
mohanrnvj@gmail.com, 2dr.vcs@srkrec.ac.in
DOI: 10.1201/9781003529231-75
AI-Based Voice Assistant Application for B5G and 6G Free Space Optic Technology is Competent of Detecting Fake Words 503
FSO communication offers up to 2.5 Gbps data transfer learning to identify fake words, generating fake words
speeds, unlike RF’s 622 Mbps limit. It uses air for optical through natural language processing. The following issues
transmission of speech, video, and data. FSO requires two come across AI-based Voice Assistant Application:
systems with optical transceivers, a laser transmitter and • To build a voice assistant system, a feature extractor will
receiver, and a telescope for data gathering. Conventional RF be applied to the device communication architecture.
wireless technology offers larger bandwidth, security, fewer • AI-based voice recognition with predictive data
power needs, and portable packaging. classification for voice assistants.
FSO-based network architectures are widely utilized in both • Fake Words identification from voice text in the voice
deep space and terrestrial applications due to their primary assistants using optimization algorithm.
advantages, which include increased bandwidth, enhanced
security, and reduced installation costs. The optical network
configuration is connected to the main part of an FSO,
2. Anatomy of Voice Assistants
which gets its input from transmitters and receivers based The text provides an in-depth analysis of the anatomy of
on photodiodes and lasers. Thus, the possibility exists that voice assistants. Voice assistants and other contemporary
the unified optical network topology will contribute to technology are transforming how people use technology
FSO’s continued success in the future. This approach may by Guoming Zhng, 2017[2]. Voice assistants become
effectively address the problems associated with current smarter as artificial intelligence (AI) develops, allowing
FSO models, including co-channel interference impairments, consumers to interact with their devices in a more organic
FSO aiming, and RF-related problems. In addition, using and intuitive way by Swahili, 2018[7]. Voice assistants are
a composite structure of optical networks and FSO can revolutionizing daily tasks by setting reminders, answering
enhance the performance of 5G services. Fifth-generation questions, controlling smart home technology, and providing
cellular technology has been replaced by sixth-generation personalized recommendations. They use natural language
wireless, or 6G. Because 6G networks can run at higher processing and machine learning to understand spoken
frequencies than 5G networks, they will have significantly commands, benefiting multitaskers and mobility-impaired
more bandwidth and latency. One of the objectives of the users by simplifying device control by Jingjiin, 2023[3].
6G internet is one-microsecond latency in communications. The Internet of Things (IoT) is growing quickly, and as a
Sixth generation (5G) will advance sustainability in multiple result, voice assistants are being incorporated into more and
ways. Data collection and closed-loop control of numerous more facets of our life. Speech assistants may now be used to
devices could be facilitated by its faster and less expensive manage and control smart home appliances like thermostats,
cost-per-bit communication. Mehtab, 2016[6] can analyze lighting controls, and security cameras with straightforward
the data using cutting-edge approaches to improve energy speech commands. This seamless connection allows users
efficiency. to easily monitor and change their home settings from
The Central Bureau of Investigation is utilizing voice anywhere, which not only increases convenience but also
samples in crime investigations, ensuring the right to privacy promotes energy efficiency and security.
and considering legal and ethical considerations in court. Voice assistants are also importantly enhancing user
The Central Bureau of Investigation has received a request interaction and satisfaction. By leveraging AI-powered
for a political leader’s voice samples to corroborate his customization algorithms, voice assistants can provide
alleged involvement in the 1984 antisikh riot case. Voice customers with unique recommendations based on their
samples are crucial in criminal investigations as they enable preferences, habits, and prior interactions [5]. Voice
investigators to verify evidence and identify suspects. Voice assistants offer personalized experiences, enhancing user
differences in recordings can help law enforcement identify experience and fostering a stronger connection between users
suspects and criminal trials, especially in criminal trials due and their devices. They are valuable tools for learning and
to the increasing prevalence of Smartphone audio and video career advancement, providing quick access to information
recording by Wormald J et al, 2022[12]. The credibility of and resources. However, concerns about security and privacy
a sample is significantly impacted by the expert’s technique arise due to cloud-based processing and storage, prompting
and court analysis, as potential inaccuracies may arise from significant investments in encryption and security measures.
medication effects or colds. The paper explores the use of 6G In addition, concerns have been voiced concerning the possible
technology for sustainability and energy efficiency in various effects of voice assistants on social interaction and human
sectors. It proposes an AI-based system for voice assistants, communication. While voice assistants can make technology
predicting micro-level traits affecting behavior. The study communication more effective and easy by Ronan Collobert,
also suggests a speech recognition application using machine 2011[8], others fear that an over dependence on these tools
504 Algorithms in Advanced Artificial Intelligence
the best fake three words are kept and a new interpolated
function t(w) is found again. This procedure continues until
two consecutive estimates are close to each other.
Let w1 be an initial word and entire text i.e., D be the step size.
Compute w2 = w1 + D.
Evaluate t(w1) and t(w2).
If t(w1) > t(w2), let w3 = w1 + 2D; ElseLet w3 = w1 – D.
Evaluatet(w3).
Determine Tmin = min(t1, t2, t3) and Wmin is the word wi that
corresponds to Tmin. Use words w1, w2, and w3 to calculate w.
Are |Tmin – t(w)| and |Tmin – w| small? If not, go to step-7; Else
the optimum is the fake of current four words and Terminate.
Save the fake word and two bracketing it, if possible;
otherwise, save the fake three words. Reliable them according
to w1 < w2 < w3 and go to Step-4.
In the above algorithm, no check is made to satisfy c2 > 0. The
same can be incorporated in Step-5. If c2 is found to be not
Fig. 75.3 Transformers are encroaching on Machine Learning
found the fake words (i.e., negative), one of the three words
may be replaced by a random text. This process is continued By repeatedly using this strategy, we can generate a large
until the quantity c2 becomes nonnegative. number of Attentions and develop a multi-head attention
layer. This improves student understanding of potential
5. Transformers are Encroaching on word connections while they are studying. Only an attention
layer, a few feed-forward layers, some leftover ResNet units,
Machine Learning and layer normalizations make up the initial Transformer
The Transformers’ Attention process is easy to comprehend block. A “Transformer” model typically consists of several
if one can approach it organically. For instance, the string Transformer blocks arranged in order. The great majority of
of consecutive inputs that a transformer receives might be language models follow this basic framework.
thought of as the words (or tokens) of a phrase. It is evident
that time series, pictures, or sound data might be contiguous 6. Experimental Result
inputs. We are aware that a word could show up in an
The accuracy metrics taken from the research work, the
embedding of a vector. The vectorization of the fake word
accuracy metrics for the Transformers: The outcomes of
may additionally take into account the word’s placement
using the attention algorithm are displayed below.
inside the input text. We have three matrices, Wq, Wk, and
Wv, which divide each of the input embedding vectors into Accuracy: When all classes are equally relevant, accuracy—a
three distinct vectors, the Query, the Key, and the Value, model’s performance across all classes—is desirable. It is
as part of the attention mechanism. This terminology is computed as the fraction of correct guesses divided by the
developed from a retrieval system. total number of forecasts.
The connected Query vectors for each word and the Key Sensitivity: The ability to identify good instances is measured
vectors for each additional word are then dot-produced. by the sensitivity. It is often referred to as the recall rate or
The definition of “attention” is determined by the focus of true positive rate.
one word in a sentence on another word to comprehend its Specificity: The specificity of a model determines how well
meaning. This exemplifies how similar the Queries and the it can predict real negatives in each available category.
Keys are. The great resemblance of the resulting vector is Precision: By dividing the total number of positive samples
further highlighted and normalized by a Softmax treatment. by the number of positively identified samples that were
As a result, every word has its own vector. Compute the dot accurately identified, precision may be calculated.
products with the Value vectors of all the other words for each
of the resulting vectors. The calculation of self-attention is F1-Score: The F1-score is a single statistic that is produced
by taking the harmonic mean of the precision and recall
finished.
506 Algorithms in Advanced Artificial Intelligence
Abstract: Criminal investigation involves studying facts for trials, using forensic science techniques. A wide-ranging criminal
investigation involves various methods such as searching, interviews, interrogations, evidence collection, and preservation.
During the investigation of specific crimes, recent demographic changes have been observed, indicating a higher priority
for their investigation. This paper explores a crime investigation using Generative AI and Information Retrieval, focusing
on Personal Dairy and its impact on crime involvement probability Using Bayes Belief Network Classification. Retriever
Augmented Generation (RAG) is an AI framework that utilizes external knowledge to provide accurate and current crime
information for large language models (LLMs) and offer users insights into their generative process.
Keywords: Artificial Intelligence, Bayes Belief Network, Crime Investigation, LLM, Retriever Augmented Generation etc.
1
kvssrmurthy75@gmail.com, 2rajanikanth.1984@gmail.com, 3shiva.csesrkr@gmail.com, 4raviswaroop.chigurupati@gmail.com, 5raviswaroop.chigurupati@
gmail.com
DOI: 10.1201/9781003529231-76
508 Algorithms in Advanced Artificial Intelligence
under each term, or a combination of both. An IR process It is a big problem that undermines justice and may have
initiates when a user enters a query, formalizing information catastrophic implications, and the threat of breaching the
needs and matching multiple objects with different relevancy law is a huge concern. Computational crime prediction and
levels in the collection [5]. forecasting has the potential to play a significant role in
Generative AI is an AI-powered technology that creates new improving public safety in urban areas, which is a problem
content using foundation models that can perform multiple that needs to be addressed. However, the enormous volume of
tasks with minimal training. It adapts to specific use cases complex big data cannot be handled by humans, which makes
with minimal example data. Generative artificial intelligence accurate criminal activity forecasts impossible. Predicting
uses supervised learning to recognize patterns in human- crime rates, types, and hotspots based on past trends poses
generated material, allowing it to create similar content both computational challenges and possibilities. Therefore,
autonomously. The model is trained using a set of labels [6]. more research is necessary to improve prediction systems
that can focus police patrols to crime incidents [9].
Generative AI enhances customer interactions, explores
unstructured data, and assists with repetitive tasks like Identifying criminal hotspots using criminal analysis is a
responding to proposals, localizing marketing content, and difficult technique. GIS was the dominant non-machine
checking customer contracts for compliance. Vertex AI allows learning temporal and geographical data approach in 2020.
developers to customize and embed foundation models into GIS reduced crime rates by using criminal-type sites [10].
applications without ML expertise. It offers a quick way to A technology that predicts crime patterns and helps law
build generative AI-powered search engines and chatbots, and enforcement solve crimes may cut real-world crime rates.
Duet AI provides assistance to users. Google’s experts offer Time series approaches are used in crime forecasting in
consulting services in generative AI, enabling organizations order to discover future crime trends. These methodologies
to create new content, discover trends, summarize data, make use of time series data in order to make predictions
and automate processes, among other services in their about crimes that may occur many years in the future. This
comprehensive Google Cloud Consulting portfolio [7]. approach helps prevent crime by enabling law enforcement
agencies to take proactive measures to mitigate potential
Inaccuracies in Personal Dairy Information Retrieval for
criminal activity before it happens.
Crime Investigation by Generative AI. A case diary is
essential for recording the investigation that an investigating Machine learning algorithms have made crime data analysis
representative conducts. During the trial, the examination simpler by preprocessing and grouping raw data to extract
could ask for the case diary. The case diary, which is meant crime locations [11]. This data has been analyzed using
to support the trial, cannot be presented as evidence in and supervised and unsupervised machine learning models to
of itself during the test. The name of the diary varies from detect crime trends by time and place, resulting in accurate
state to state, and the police laws include instructions on how predictions [12]. Machine learning algorithms used to past
it should be kept. Usually referred to as a “case diary,” this data from the same site have helped investigators determine
journal is also known as a “special diary” in some states. The the causes of crime [13]. In, [14], the authors has removed the
case diary notes the time when each detail was documented noise in images and analysed the performance by using SGO
by the investigating representative [8]. and APSO. They have used the filters to remove the noise in
GenerativeAI in Personal Dairy Information Retrieval for Criminal Investigation 509
PGM images [15]. After that the images are retieved by using to the other foundation models. A condensed representation
Genetic model [16]. Some of the techniques has been used to of the underlying structure of raw data is created from its
identify the flower species by using Deep Learning Models peak. A foundation model can be refined based on tagged,
[17] and detected the counterfeit on scanned document by domain-specific data starting from this core representation to
using NN [18]. So that the approach was colourized for gray fit a crime investigation application [22]. The contribution of
scale image [19] and sharpened the images by using CNN propose work as follow:
[20]. • Uncertainty Crime Investigation Using Bayes Belief
Network Classification.
1.1 Challenges in Prediction Systems
• RAG is an AI framework that uses external knowledge
Predicting crime is a complex task for both researchers to provide accurate, up-to-date information for large
and security agents. They face difficulties in determining language models (LLMs) and provide users with insights
the location and time of the crime, as well as choosing the into their generative process.
most effective method to predict it. Researchers working in
Uncertainty Crime Using Bayes Belief Network in
the field of computer science who use techniques such as
Artificial Intelligence: The Bayesian belief network (BBN)
data mining, machine learning, and spatial-temporal data
is a kind of computer technology that is applied to manage
also encounter obstacles. In 2012 and 2016, algorithms that
uncertain and probabilistic occurrences in the context of
anticipate crimes in households, streets, and regions were
investigative difficulties involving criminal activity. It is a
established. These approaches were referred to as near
probabilistic graphical model representing variables and
repeat-victimization and repeat-victimization, respectively.
their dependencies, using probability theory for prediction
The results of these methodologies indicate that if a crime
and anomaly detection. A BBN is a probabilistic tool used
is committed in a certain location, there is a substantial
in real-world applications for prediction, anomaly detection,
possibility that the number of subsequent crimes committed
diagnostics, and decision making under uncertainty. The
in the same location will dramatically rise. BBN consists of a Directed Acyclic Graph and a Table of
Developers of crime prediction systems face several conditional probabilities, also known as an Influence diagram.
challenges, The network structure reveals that crime investigation and
a. This includes the necessity for extensive storage owing Information Retrieval are the parent nodes of Personal Dairy,
to the high volume of data. directly affecting the probability of person involvement in a
b. A diverse range of formats, including text, images, crime [23].
graphs, audio, structured and semi-structured data, can
be used to store information related to criminal activity.
c. Creating a comprehensible structure for this data is
equally challenging.
d. Utilize a data mining strategy that produces better
results than the algorithms that are already in use in
order to effectively identify instances in machine
learning by using this approach.
e. Crime prediction systems may be influenced by Fig. 76.2 Uncertain crime investigation using bayes belief
environmental elements like weather and lawlessness, network
leading to significant mistakes [21]. To minimize such Retriever Augmented Generation Using LLM: The use of
inaccuracies and obtain high prediction accuracy, crime crime databases in augmenting LLMs is a valuable method,
forecasts must account surrounding and environmental but it has several significant flaws. The debate between fine-
changes. tuning and Retriever Augmented Generation (RAG) with
Therefore, in order to avoid making mistakes of this kind LLMs is ongoing, with RAG being better for enhancing
and to reach a high level of prediction accuracy, any crime LLMs with small additional data. RAG encodes data into
forecast must take into account changes in the surrounding embeddings and indexes it into a vector database. Users ask
environment and the environment itself. questions, which are converted into embeddings and used to
search for similar embeddings [24]. Prompts provide context
2. Proposed Work for LLM answers, usually using cosine similarity metric.
The problem lies in the search’s ability to retrieve documents
Using artificial intelligence to generate information retrieval. with similar words or context without providing relevant
The transformer AI architecture supports LLMs in addition information, leading to an excess of irrelevant documents
510 Algorithms in Advanced Artificial Intelligence
showing higher cosine similarity than the actual answer. The NB, SVM, DT, and K-Nearest Neighbor algorithms
High cosine similarity in Transformers does not necessarily were used in order to produce predictions about criminal
imply semantic similarity; it can also indicate the high co activities. The results of these algorithms were compared,
occurrence of two terms within the same training data [25, which served to demonstrate the efficacy of neural networks
26]. The data’s indexing can cause issues if it’s broken in complex systems [28]. The complexity of the connections
down into large chunks, potentially containing unrelated between geographical data, on the other hand, significantly
information. To avoid diluted information and irrelevant reduced the value of pattern extraction. Table 76.1 shows the
documents, break down the data into a few paragraphs per Comparison data using DM with various datasets.
chunk, ensuring uniqueness. The RAG approach emphasizes
limiting the type of questions asked by the LLM. Aggregating Table 76.1 Comparative study with various datasets in DM
data across the database may lead to incorrect answers, while Reference Dataset Used Models Accuracy
similarity searches may find local information. For example,
[31] Chicago crime DT 76.1
if a question requires scanning all documents, similarity data
RF 82.59
searches may not be helpful [27].
NB 76.63
Mining DT 48
[33] Bangladesh crime NB 70.8
2011 was the year that saw the development of specialized data
KNN 77.4
data mining and technologies that were able to extract patterns
[34] Indian and Banga KDE 78.83
from geographical and temporal data. For the purpose of lore crimes data
forecasting criminal activity and determining the possibility of
[35] India crime data WEKA on two 92.52 for C1
residential burglary, the data from Portland was geospatially K-mean clusters
92.59 for C2
mined with the assistance of specific technical knowledge.
GenerativeAI in Personal Dairy Information Retrieval for Criminal Investigation 511
In 2016, a high level of accuracy was achieved in Table 76.2 Comparative study with various datasets in ML
extracting information from data acquired throughout 1994 Reference Dataset Used Models Accuracy
occurrences, which had 128 characteristics, using a variety of
[40] UCI machine learning J48 94.25
DT algorithms [29]. Scatter plots were used to represent the repository website
crime regions and their severity based on the data collected in [41] London mobile and crime data RF 71
the past for training and testing the data.
[42] Chicago crime data DT 39
Also, in the same year, data mining techniques were used to RF 61
categorize the infractions according to the types of crimes NN 82
committed [30]. It was found that the timing of the crime was
[43] Bangladesh crime data LR 72.9
classified, taking into account elements such as vacations that
[44] New York crime data SVM 44
began with the beginning of the school year for colleges and
universities. RF 51
XGBoost 53
22. Shankar RS, Gupta VM, Murthy KV, Someswararao C. Object 34. Prathap BR, Krishna AV, Balachandran K. Crime analysis
oriented fuzzy filter for noise reduction of Pgm images. and forecasting on spatio temporal news feed data—an indian
In2012 8th International Conference on Information Science context. InArtificial intelligence and blockchain for future
and Digital Content Technology (ICIDT2012) 2012 Jun 26 cybersecurity applications 2021 May 1 (pp. 307-327). Cham:
(Vol. 3, pp. 776-782). IEEE. Springer International Publishing.
23. Shankar RS, Sravani K, Srinivas LV, Babu DR. An approach 35. Tayal DK, Jain A, Arora S, Agarwal S, Gupta T, Tyagi N.
for retrieving an image using Genetic Algorithm. International Crime detection and criminal identification in India using data
Journal of Latest Trends in Engineering and Technology. mining techniques. AI & society. 2015 Feb;30:117-27.
2017;9(8):057-64. 36. Hou M, Hu X, Cai J, Han X, Yuan S. An integrated graph
24. Shankar RS, Srinivas LV, Raju VS, Murthy KV. A model for spatial–temporal urban crime prediction based on
Comprehensive Analysis of Deep Learning Techniques for attention mechanism. ISPRS International Journal of Geo-
Recognition of Flower Species. In2021 Third International Information. 2022 Apr 30;11(5):294.
Conference on Intelligent Communication Technologies and 37. Ilhan F, Tekin SF, Aksoy B. Spatio-temporal crime prediction
Virtual Mobile Networks (ICICV) 2021 Feb 4 (pp. 1172 with temporally hierarchical convolutional neural networks.
1179). IEEE. In2020 28th Signal Processing and Communications
25. Devareddi RB, Shankar RS, Murthy K, Raminaidu C. Image Applications Conference (SIU) 2020 Oct 5 (pp. 1-4). IEEE.
segmentation based on scanned document and hand script 38. Meskela TE, Afework YK, Ayele NA, Teferi MW, Mengist
counterfeit detection using neural network. InAIP Conference TB. Designing time series crime prediction model using long
Proceedings 2022 Dec 9 (Vol. 2576, No. 1). AIP Publishing. short-term memory recurrent neural network. International
26. Shankar RS, Mahesh G, Murthy KV, Ravibabu D. A novel Journal of Recent Technology and Engineering (IJRTE).
approach for gray scale image colorization using convolutional 2020;9:402-5.
neural networks. In2020 International Conference on System, 39. Hussain FS, Aljuboori AF. A crime data analysis of prediction
Computation, Automation and Networking (ICSCAN) 2020 based on classification approaches. Baghdad Science Journal.
Jul 3 (pp. 1-8). IEEE. 2022 Oct 1;19(5):1073-.
27. Shankar RS, Mahesh G, Murthy KV, Rajanikanth J. A novel 40. Ahishakiye E, Taremwa D, Omulo EO, Niyonzima I.
approach for sharpening blur image using convolutional neural Crime prediction using decision tree (J48) classification
networks. Journal of Critical Reviews. 2020 Apr;7(7):139-48. algorithm. International Journal of Computer and Information
28. Yu CH, Ward MW, Morabito M, Ding W. Crime forecasting Technology. 2017 May;6(3):188-95.
using data mining techniques. In2011 IEEE 11th international 41. Bogomolov A, Lepri B, Staiano J, Oliver N, Pianesi F,
conference on data mining workshops 2011 Dec 11 (pp. 779 Pentland A. Once upon a crime: towards crime prediction
786). IEEE. from demographics and mobile data. InProceedings of the
29. Shekhar S, Evans MR, Kang JM, Mohan P. Identifying 16th international conference on multimodal interaction 2014
patterns in spatial information: A survey of methods. Wiley Nov 12 (pp. 427-434).
Interdisciplinary Reviews: Data Mining and Knowledge 42. El Bour HA, Ounacer S, Elghomari Y, Jihal H, Azzouazi M.
Discovery. 2011 May;1(3):193-214. A crime prediction model based on spatial and temporal data.
30. Sharma H, Kumar S. A survey on decision tree algorithms of Periodicals of Engineering and Natural Sciences. 2018 Nov
classification in data mining. International Journal of Science 24;6(2):360-4.
and Research (IJSR). 2016 Apr 5;5(4):2094-7. 43. Mahmud S, Nuha M, Sattar A. Crime rate prediction using
31. Yerpude P. Predictive modelling of crime data set using data machine learning and data mining. InSoft Computing
mining. International Journal of Data Mining & Knowledge Techniques and Applications: Proceeding of the International
Management Process (IJDKP) Vol. 2020 Jul 21;7. Conference on Computing and Communication (IC3 2020)
32. Almanie T, Mirza R, Lor E. Crime prediction based on crime 2021 (pp. 59-69). Springer Singapore.
types and using spatial and temporal criminal hotspots. arXiv 44. Almuhanna AA, Alrehili MM, Alsubhi SH, Syed L. Prediction
preprint arXiv:1508.02050. 2015 Aug 9. of crime in neighbourhoods of New York City using spatial
33. Mahmud S, Nuha M, Sattar A. Crime rate prediction using data analysis. In2021 1st International conference on artificial
machine learning and data mining. InSoft Computing intelligence and data analytics (CAIDA) 2021 Apr 6 (pp. 23
Techniques and Applications: Proceeding of the International 30). IEEE.
Conference on Computing and Communication (IC3 2020)
Note: All the figures and tables in this chapter were designed by
2021 (pp. 59-69). Springer Singapore.
the author.
Algorithms in Advanced Artificial Intelligence – Dr. Dr. R. N. V. Jagan Mohan et al. (eds)
© 2024 Taylor & Francis Group, London, ISBN 978-1-032-86798-4
514 Algorithms in Advanced Artificial Intelligence
Susmitha Uddaraju1
Assistant Professor, Department of Artificial Intelligence
Shri Vishnu Engineering College for Women, Vishnupur, Bhimavaram, West Godavari Dist., AP, India
G. P. Saradhi Varma2
Professor, Department of Computer Science and Engineering
Koneru Lakshmaiah Education Foundation (Deemed to be University), Vaddeswaram, Guntur Dist., AP, India
I.Hemalatha3
Professor, Department of Information Technology
S. R. K. R. Engineering College, Bhimavaram, Indian
Abstract: For most patients with breast cancer who have had neo-adjuvant chemotherapy, surgery is the preferred course
of treatment due to the highly difficult evaluation of PCR (pathologic complete response). However, early prediction made
possible by technological advancements means that patients can now receive the appropriate treatment sooner. Thanks to
advanced techniques and the availability of large volumes of data, accurate and timely prediction is now possible. The goal of
this research is to assess how well a number of recently developed machine learning algorithms perform in terms of forecasting
the NAC response to breast cancer. Principal component analysis with cuckoo search optimisation (PCACSO) is incorporated
into five popular classifiers (random forest, decision tree, naive bayes, K-nearest neighbour, and support vector machine) to
enhance the prediction model’s accuracy. In accordance, the figures are 87%, 82.83%, 74.23%, 71.7%, and 78.41%. On the
other hand, achieving the same outcomes for feature selection with and without PCACSO resulted in accuracy of 81.8%,
56.06%, 43.9%, 71.21%, and 76.78%, respectively. The proposed PCACSO model can be used to improve most classification
methods by utilising feature selection techniques to reduce the number of features. Certain features have a greater bearing and
influence on the results generated by the classification algorithms than others.
Keywords: Breast cancer; NAC; Feature selection; Random forest, Decision tree; Naïve bayes; K-nearest neighbor; PCA;
Cuckoo search optimization etc
1
susmithauddaraju@gmail.com, 2gpsvarma@gmail.com, 3indukurihemalatha@gmail.com
DOI: 10.1201/9781003529231-77
PCACSO Feature Selection for Prediction of Breast Cancer NAC Response 515
of NAC [2]. A pCR could lead to improved rates of overall picking out a set of pertinent attributes that may be used
and disease-free survival. If pCR can be ascertained non to construct a model, while instance reduction involves
invasively, surgeons can avoid surgery; however, currently, decreasing the number of orthogonal instances in a dataset
surgery is still necessary to confirm a pCR following NAC with the goal of improving classification accuracy [16].
[3].Currently, it is advised to use breast magnetic resonance While these ancillary factors do not impact the categorization
imaging before and after NAC because it has higher PCR process, they may lower the findings’ accuracy. Feature
detection accuracy compared to physical examination, selection is a useful tool for removing irrelevant, noisy, or
diagnostic procedures, and ultrasound [5] and [6]. The stated redundant features from a model that aren’t adding anything
sensitivity of MRI for PCR is fairly different; however, a useful. Therefore, rather than trying to use every possible
meta-analysis indicated a combined sensitivity of 64%, taxonomy choice, it is easier to zero in on those that are really
which is insufficient to rule out tissue confirmation and suitable and useful [18]. By reducing the overall number of
surgery. When combined with machine learning algorithms, possible outcomes, this unexpectedly simplifies the model
the data from high-resolution breast MRI scans may provide and makes it more readable. By integrating feature selection
extremely accurate, non-invasive cancer response detection into healthcare knowledge, we can decrease the number of
approaches [7]. Computer-assisted qualitative image tests required for a diagnosis, which in turn saves patients
analysis in the discipline of computational biology known both time and money on testing. There are three main types of
as “radiomics” identifies traits that are unseen to the human traditional feature-selection methods: embedding, wrapper,
eye, augmenting visual evaluation [9]. These techniques are and filtering strategies [20]. Despite the abundance of feature
compatible with machine learning and have shown promise selection algorithms developed for the healthcare industry,
for non-invasively identifying therapeutic responses in breast there is a dearth of studies on breast. Analysis of polymerase
and other cancers. Consequently, our goal was to develop and chain reactions for cancer. This research aims to fill that
evaluate a radiomics biomarker that may classify cancer pCR need by developing a breast cancer PCR prediction system
on MRI following NAC [8]. that integrates feature selection and optimisation techniques
[19]. To alleviate the problem of high-dimensional data in
2. Related Work breast cancer risk factors and improve prediction accuracy,
we developed a hybrid approach that combines training and
2.1 Dimensionality Issues application time reduction with a selection model [20].
Burgess talks about a method called “dimensional reduction,”
which maps data to a low-dimensional space by lowering 2.2 Machine Learning in Healthcare
its informative variance [11]. This finds the topological Experts in many domains, including business, medicine,
space where the data exists. Two steps are involved in and research, face the challenge of deciphering massive and
implementing dimensionality reduction: feature extraction intricate datasets [21, 22]. The ability to understand and
and feature selection. The process of feature extraction make sense of massive datasets is in high demand across
involves going through a dataset in search of unnecessary, many industries as a result [21, 22]. This includes business,
irrelevant, or duplicated measurement features and then research, medicine, and many more. The capacity to derive
removing them [12] and [14]. In order to construct reliable actionable insights from this deluge of data is paramount in
learning models, feature selection enables the discovery and the modern, cutthroat business environment. Data mining
removal of as much redundant and unnecessary data as is standards are based on artificial intelligence, statistics,
practically feasible. The end result is an improved model with computers, and probability [24]. Two major types of data
reduced processing times and expenses because of feature processing models are descriptive and predictive models.
selection [10]. Numerous earlier investigations, especially The use of predictive models is common in supervised
those involving healthcare data, have made use of the feature learning functions when attempting to forecast the present or
selection technique. Our classification system for feature future values of critical variables. When it comes to finding
selection algorithms is as follows: filters, wrappers, and potentially intelligible data patterns, unsupervised learning
embedding techniques [13]. Burgess described dimensional functions depend on interpretive models [23]. Classification
reduction as the act of reducing the number of dimensions of in the healthcare industry is done using statistical approaches.
a body of knowledge in order to extract useful information Swarm intelligence, decision trees, logistic regression,
while eliminating irrelevant details [15]. [17]. One subset of ANNs, SVMs, k-nearest neighbour, association rules, and
dimensional reduction is reduction and selection techniques, genetic classifiers [25] [26].
while another is instance selection. Feature selection is
516 Algorithms in Advanced Artificial Intelligence
Comparative Analysis dataset, which includes 216 examples with 12 attributes. The
At the moment, we use PCA and PCACSO and depend characteristics and areas covered by the dataset include
solely on feature selection to assess the efficacy of several
classifiers, including KNN, GNB, DT, and RF. Our analysis 6. Classification Algorithm
determined that all samples met the criteria for this study. Descriptions
The confusion matrix, which details the recommended
classifier’s actual and anticipated classifications, is one For the purpose of predicting breast cancer PCR, this
way to assess a prediction model’s accuracy. An oncologist study compared three well-known [61] prediction model
validated and assessed the suggested model to ensure the classification algorithms: naive Bayes, support vector
accuracy of the prediction model and classification. We also machines, and K-nearest neighbour. In the paragraph that
used ROC curves to look at the confusion matrix and the follows, we provide a brief description of each algorithm.
representations of the categorical measures. Assessments of Naive Bayes Algorithm
sensitivity, specificity, positive and negative predictive value,
and accuracy in classification are all part of this process. In Both Bayesian classification and the statistical method of
order to ensure the correctness, precision, and efficacy of the classification are supervised learning approaches. It works
suggested strategy, we compared it to data mining techniques. on the premise of a probabilistic model and lets us capture,
The variable P00 represents the exactly identified predictions in theory, the uncertainty relative to the model through event
pertaining to class 0. Additionally, this is the true benefit of probability calculations. Diagnostic and predictive issues
Class 0. P10 displays the frequency of incorrectly labelling can be addressed with its help. Thomas Bayes (1702–1761)
class 0 variables as class 1. The number of instances where proposed Bayes’ theorem [27] [28], which served as the
class 0 predictions are mistakenly labelled as class 2 is shown basis for this categorization. Using a combination of prior
by P20. Variable P01 represents the misclassification rate of knowledge and observed data, Bayesian classification offers
class 1 as class 0. P11 represents the number of times class 1 effective learning approaches. Bayesian classification provides
variables were correctly classified as class 1. For class 1, this a useful framework for understanding and evaluating various
is likewise a true positive. The number of class 1 projections learning techniques. It computes transparent probabilities for
that were incorrectly labelled as class 2 is represented hypotheses and is robust against input data noise.
by variable P21. Variable P02 indicates that the model K-Nearest Neighbors Algorithm
incorrectly identified the class 2 prediction count as class 0. The efficacy of merging feature selection and classification
P12 shows how many predictions got Class 2 variables wrong algorithms in breast cancer prognosis prediction is the subject
and thought they were Class 1. P22 represents the number of of this research. Feature selection strategies that aim to reduce
successfully detected class 2 predictions. For class 1, this is the number of features can increase the performance of most
likewise a true positive. classification algorithms, according to our suggestion. When
it comes to the outcomes of categorization algorithms, some
5. Dataset Description features are far more influential than others. Using five
popular classification algorithms—Gaussian naive basis,
Studying the aforementioned approaches and procedures, this support vector machines, K-nearest neighbour, random
research made use of the American Oncology Institute’s (AOI) forest, and decision tree—we present our testing results.
Table 77.2 Dataset description We also looked at how Cuckoo Search Optimisation (CSO),
a feature selection method, affected these algorithms. In
Variable Description
summary, Random Forest outperformed the other methods
AGE Patient Age regardless of the use of PCACSO. However, when PCACSO
ERPOS Estrogen Receptor Status was included, the performance of the other four methods also
PGRPOS Progesterone Receptor Status
HR POS Hormone Receptor Status
improved. Evaluating more complex algorithms to enhance
HER2MOSTPOS Her2 Status accuracy will be a part of future studies. Cluster methods and
HR_HER2_CATEGORY 3-level HR/Her2 category pre ensemble algorithms are the subjects of our experiments.
treatment
BILATERALCA Does the patient have bilateral
Decision Tree
breast cancer Decision trees, a categorization method, use recursive
LATERALITY Index Tumor Laterality left or right splitting of the instance space. By associating data about
MRI 1 MRI LD Baseline
MRI 2 MRI LD 1-3d AC
specific nodes with their expected values, it builds a model
MRI 3 MRI LD InterReg for making predictions in the future. The leaves of the tree
MRI 4 MRI LD PreSurg structure represent the class labels, while the branches reflect
the feature combinations that lead to category labels [20].
518 Algorithms in Advanced Artificial Intelligence
naive Bayesian are used for the classification evaluation. The 8. Susmitha, Uddaraju. “A Review of Machine Learning
image below shows the outcomes of the experiment. Using Frameworks For Early And Accurate Prediction Of
PCACSO as a feature selection clearly improves the accuracy Neoadjuvant Chemotherapy Responses.” European Journal
of the machine learning model, according to the experimental Of Molecular & Clinical Medicine 7.4 (2020): 1040-1050.
9. S.Şahan, K. Polat, H. Kodaz, And S. Güneş, “A New Hybrid
results.
Method Based On Fuzzy-Artificial Immune System And
K-Nn Algorithm For Breast Cancer Diagnosis,” Comput. Biol.
9. Conclusion and Future Perspective Med., Vol. 37, Pp. 415–423, March 2007.
10. Avikadutta, Stacking, Https://Www.Geeksforgeeks.Org/
This study set out to investigate how breast cancer prognosis Stacking-In-Machine-Learning/.
is affected by combining feature selection and classification 11. Himanshisingh, Advanced Ensemble Learning Technique–
methods. We propose that most classification systems can be Stacking And Its Variants, Https://Www.Analyticsvidhya.
improved by reducing the amount of features using feature Com/Blog/2021/03/Advanced-Ensemble-Learning
selection methods. The outputs produced by categorization Technique-Stacking-And-Its-Variants.
algorithms are more heavily impacted by some characteristics 12. Sun, Wei, and Jingyi Sun. “Daily PM2. 5 Concentration
than others. Decision trees with and without feature selection, Prediction Based On Principal Component Analysis And
support vector machines, K-nearest neighbour, decision LSSVM Optimized By Cuckoo Search Algorithm.” Journal
Of Environmental Management 188 (2017): 144-152.
trees, random forests, and cuckoo search optimisation (CSO)
13. Uddaraju, Susmitha, GP Saradhi Varma, And M. R.
are the five prominent classification techniques that we
Narasingarao. “Prediction Of NAC Response In Breast
tested. In the end, Random Forest outperformed the other Cancer Patients Using Neural Network.” Scalable Computing:
four approaches, particularly when PCACSO was included. Practice And Experience 23.4 (2022): 211-224.
Even without PCACSO, Random Forest outperformed the 14. Naik, Manoj Kumar, and Rutuparna Panda. “A Novel Adaptive
others. Modern algorithms that aim to increase accuracy will Cuckoo Search Algorithm For Intrinsic Discriminate Analysis
be the focus of the following section of this research. Our Based Face Recognition.” Applied Soft Computing 38 (2016):
experiments mostly centre on approaches for grouping and 661-675.
ensembles.. 15. Katarya, Rahul, and Om Prakash Verma. “An Effective
Collaborative Movie Recommender System With Cuckoo
Search.” Egyptian Informatics Journal 18.2 (2017): 105-112.
References 16. \Uddaraju, Susmitha, and M. R. Narasingarao. “Predicting The
Ductal Carcinoma Using Machine Learning Techniques—A
1. RM, SwarnaPriya, Et Al. Effective Feature Engineering For
Comparison.” Journal Of Computational And Theoretical
DNN Using Hybrid PCA-GWO For Intrusion Detection In
Nanoscience 16.5-6 (2019): 1902-1907.
IoMT Architecture.” Computer Communications 160 (2020):
17. Sannasi Chakravarthy, S. R., and HarikumarRajaguru.
139-149.
“Comparison Analysis of Linear Discriminate Analysis And
2. Sakri, Sapiahbinti, Nurainibinti Abdul Rashid, and Zuhaira
Cuckoo-Search Algorithm In The Classification of Breast
Muhammad Zain. “Particle Swarm Optimization Feature
Cancer From Digital Mammograms.” Asian Pacific Journal
Selection for Breast Cancer Recurrence Prediction.” IEEE
Of Cancer Prevention: APJCP 20.8 (2019): 2333.
Access 6 (2018): 29637-29647.
18. Sudha, M. N., And S. Selvarajan. “Feature Selection Based
3. Uddaraju, Susmitha, And M. Narasingarao. “A Survey Of
On Enhanced Cuckoo Search For Breast Cancer Classification
Machine Learning Techniques Applied For Breast Cancer
In Mammogram Image.” Circuits and Systems 7.04 (2016):
Prediction.” International Journal Of Pure And Applied
327.
Mathematics 117.19 (2017): 499-507.
19. Lavanya, D., and Dr K. Usha Rani. “Analysis of Feature
4. Cortazar, P., & Geyer, C. E. (2015). Pathological Complete
Selection with Classification: Breast Cancer Datasets.” Indian
Response In Neoadjuvant Treatment Of Breast Cancer. Annals
Journal of Computer Science And Engineering (IJCSE) 2.5
Of Surgical Oncology, 22(5), 1441-1446.
(2011): 756-763.
5. A. Bhardwaj And A. Tiwari, “Breast Cancer Diagnosis Using
20. Akay, Mehmet Fatih. “Support Vector Machines Combined
Genetically Optimized Neural Network Model,” Expert Syst.
With Feature Selection For Breast Cancer Diagnosis.” Expert
Appl., Vol. 42, Pp. 4611–4620, 15 June 2015.
Systems with Applications 36.2 (2009): 3240-3247.
6. W. C. Yeh, W. -W. Chang, And Y. Y. Chung, “A New Hybrid
21. [26] Eswar, “An Enhanced and Naive Clustering Algorithm
Approach For Mining Breast Cancer Pattern Using Discrete
for Text Classification Based on Weight,” International Journal
Particle Swarm Optimization And Statistical Method,” Expert
& Magazine of Engineering, Technology, Management and
Syst. Appl., Vol. 36, Pp. 8204–8211, May 2009.
Research, Dec. 2012.
7. B. Zheng, S. W. Yoon, And S. S. Lam, “Breast Cancer
22. Chen, Hui-Ling, Et Al. “A Support Vector Machine Classifier
Diagnosis Based On Feature Extraction Using A Hybrid Of
With Rough Set-Based Feature Selection For Breast Cancer
K-Means And Support Vector Machine Algorithms,” Expert
Diagnosis.” Expert Systems with Applications 38.7 (2011):
Syst. Appl., Vol. 41, Pp. 1476–1482, March 2014.
9014-9022.
520 Algorithms in Advanced Artificial Intelligence
23. Aalaei, Shokoufeh, Et Al. “Feature Selection Using Genetic Spectroscopic Imaging (MRSI).” NMR In Biomedicine 23.3
Algorithm For Breast Cancer Diagnosis: Experiment On (2010): 233-241.
Three Different Datasets.” Iranian Journal Of Basic Medical 28. King, Tari A., and Monica Morrow. “Surgical Issues
Sciences 19.5 (2016): 476. in Patients with Breast Cancer Receiving Neoadjuvant
24. Susmitha, U., and D. Rajeswara Rao. “Optimized Secure Chemotherapy.” Nature Reviews Clinical Oncology 12.6
Confirmations Using Smart Card Evaluation In Multi Cloud (2015): 335-343.
Storage.”(2006). 29. Sharma, Uma, Et Al. “Longitudinal Study of The Assessment
25. Uppendra, “Predict Early Pneumonitis in Health Care Using By MRI And Diffusion Weighted Imaging Of Tumor Response
Hybrid Model Algorithms,” Journal of Artificial Intelligence, In Patients With Locally Advanced Breast Cancer Undergoing
Machine Learning and Neural Network (JAIMLNN), vol. 3, Neoadjuvant Chemotherapy.” NMR In Biomedicine: An
issue 03, pp. 14-26,ISSN: 2799-1172, Apr. 2023. International Journal Devoted To the Development and
26. Fayanju, Oluwadamilola M., Et Al. “The Clinical Significance Application Of Magnetic Resonance In Vivo 22.1 (2009):
of Breast-Only And Node-Only Pathologic Complete 104-113.
Response (PCR) After Neoadjuvant Chemotherapy (NACT): 30. Kümmel, S., J. Holtschmidt, And S. Loibl. “Surgical
A Review Of 20,000 Breast Cancer Patients In The National Treatment Of Primary Breast Cancer In The Neoadjuvant
Cancer Data Base (NCDB).” Annals Of Surgery 268.4 (2018): Setting.” Journal Of British Surgery 101.8 (2014): 912-924.
591. 31. Graham, Peter J., Et Al. “Neoadjuvant Chemotherapy For
27. Danishad, KarikanniKalathil A., Et Al. “Assessment Of Breast Cancer, Is Practice Changing? A Population-Based
Therapeutic Response Of Locally Advanced Breast Cancer Review of Current Surgical Trends.” Annals of Surgical
(LABC) Patients Undergoing Neoadjuvant Chemotherapy Oncology 22.10 (2015): 3376-3382.
(NACT) Monitored Using Sequential Magnetic Resonance
Note: All the figures and tables in this chapter were designed by
the author.