0% found this document useful (0 votes)
32 views49 pages

Fullyandfinal

The document discusses a project report on image analysis using convolutional neural networks. It includes sections on introduction, literature review, proposed work, methodology, results and discussion, and conclusion. The report is submitted by three students for their bachelor's degree and aims to explore CNN architectures, functionality, and applications in image analysis.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
32 views49 pages

Fullyandfinal

The document discusses a project report on image analysis using convolutional neural networks. It includes sections on introduction, literature review, proposed work, methodology, results and discussion, and conclusion. The report is submitted by three students for their bachelor's degree and aims to explore CNN architectures, functionality, and applications in image analysis.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 49

A PROJECT REPORT

ON
Image Analysis using Convolutional Neural Network

For the partial fulfillment for the award of the degree of


BACHELOR OF TECHNOLOGY
In
Computer Science (Artificial Intelligent and Machine Learning)
Submitted By
Prateek Joshi (2001921530037)
Aman Kumar (2001921530008)
Rahul Kumar (2001921530041)

Under the Supervision of


Ms. Hina Gupta

G.L. BAJAJ INSTITUTE OF TECHNOLOGY &


MANAGEMENT, GREATER NOIDA
Affiliated to

DR. APJ ABDUL KALAM TECHNICAL UNIVERSITY,


LUCKNOW

2023-2024
Declaration

We hereby declare that the project work presented in this report entitled “Image
Analysis using Convolutional Neural Network” in partial fulfillment of the
requirement for the award of the degree of Bachelor of Technology in
“Computer Science (AIML)” submitted to A.P.J. Abdul Kalam Technical
University, Lucknow, is based on our own work carried out at Department of
Applied Computational Science and Engineering (ACSE), G.L. Bajaj Institute of
Technology & Management, Greater Noida. The work contained in the report is
true and original to the best of our knowledge and project work reported in this
report has not been submitted by us for award of any other degree or diploma in
any other institute / college/university.

Signature:
Name: Prateek Joshi

Roll No: 2001921530037

Signature:
Name: Aman Kumar
Roll No:
2001921530008

Signature:
Name: Rahul Kumar

Roll No: 2001921530041

Date:22 may 2024


Place: Greater Noida

ii
Certificate

This is to certify that the Project report entitled “Image Analysis using
Convolutional Neural Network” done by Prateek Joshi (2001921530037), Aman
Kumar (2001921530008) and Rahul Kumar (2001921530041) is an original
work carried out by them in Department of Applied Computational Science and
Engineering (ACSE), G.L. Bajaj Institute of Technology & Management, Greater
Noida under my supervision. The matter embodied in this project work has not
been submitted earlier for the award of any degree or diploma to the best of my
knowledge and belief.

Date:22 may 2024

Ms. Hina Gupta


Signature of the

Supervisor

iii
Acknowledgement

The merciful guidance bestowed to us by the almighty made us stick out this
project to a successful end. We humbly pray with sincere heart for his guidance to
continue forever.

We pay thanks to our project guide Ms. Hina Gupta who has given guidance and
light to us during this project. His/her versatile knowledge has helped us in the
critical times during the span of this project.

We pay special thanks to our Dean of the Department Dr. Naresh Kumar and Head
of Department Dr. Mayank Singh who has been always present as a support and
help us in all possible way during this project.

We also take this opportunity to express our gratitude to all those people who
have been directly and indirectly with us during the completion of the project.

We want to thanks our friends who have always encouraged us during this project.

At the last but not least thanks to all the faculty of CSE department who provided
valuable suggestions during the period of project.

Rahul Kumar(2001921530041)
Prateek Joshi(2001921530037)
Aman Kumar(2001921530008)

iv
Abstract

Image analysis using Convolutional Neural Networks (CNNs) has revolutionized


the field of computer vision, offering unprecedented capabilities in object
detection, image classification, and semantic segmentation. This paper explores
the architecture, functionality, and applications of CNNs in image analysis. CNNs
leverage hierarchical layers of convolutions to automatically and adaptively learn
spatial hierarchies of features from input images. This ability to capture spatial
and temporal dependencies in an image with a minimal amount of pre-processing
has established CNNs as the cornerstone of modern image analysis tasks.

The discussion begins with a detailed examination of the fundamental components


of CNNs, including convolutional layers, pooling layers, and fully connected
layers. The role of activation functions and the backpropagation algorithm in
training these networks is also elucidated. Various architectures such as LeNet,
AlexNet, VGG, and ResNet are compared, highlighting their contributions to
advancing CNN capabilities.

The paper further delves into the practical applications of CNNs in diverse
domains such as medical imaging, autonomous vehicles, facial recognition, and
anomaly detection. Techniques for enhancing CNN performance, including data
augmentation, transfer learning, and the use of advanced optimization algorithms,
are reviewed. Additionally, challenges such as overfitting, the need for large
labeled datasets, and computational requirements are discussed, along with
potential solutions like regularization techniques and unsupervised learning.

The conclusion emphasizes the transformative impact of CNNs on image analysis


and anticipates future trends, including the integration of CNNs with other deep
learning models and the development of more efficient and interpretable
networks. Overall, this paper underscores the critical role of CNNs in the ongoing
evolution of image analysis technologies.

v
Table Of Content

Declaration...................................................................................................................... (ii)
Certificate ................................................................................................................. (iii)
Acknowledgement........................................................................................................................................ (iv)
Abstract ................................................................................................................. (v)
Table of Content................................................................................................................. (vi)
List of Figures ………………………………………………………………………… (vii)
List of Tables …………………………………………………………………………. (viii)
Detail of published/ communicated research paper /patent ……………………………. (ix)

Chapter 1. Introduction ..................................................................................... Pg.No


1.1 Project Definition……….......................................................................
1.2 Problem Statement &Objectives ………………....................................
1.3 Benefits of research……………………………………..……...............
1.4 Limitation of research ………………………………………………..
Chapter Literature Survey ................................................................................ Pg.No
2.
2.1 Introduction …………………………………………………………..
2.2 Literature Review....................................................................................
2.3 Inferences Drawn from Literature Review………………….................
Chapter 3. Proposed Work………………………………………………..………. Pg.No
3.1 Introduction ………………………………………………………….
3.2 Proposed Work ………………………………………………….……
Chapter 4. Methodology…................................................................................. Pg.No
4.1 Introduction …………………………………….…………………….
4.2 Implementation Strategy (Flowchart, Algorithm etc.) ………………
4.3 Tools/Hardware/Software Requirements..………………………….
Chapter 5. Result & Discussion ……………………….......................................... Pg.No
5.1 Introduction ……………………………………………………….
5.2 Performance Metrics with details
Chapter 6. Conclusion & Future Scope.………………………………………...... Pg.No
References

vi
LIST OF
FIGURES

Page No.

Figure 1.1 Pg.No


Figure 1.2 Pg.No
Figure 2.1 Pg.No
Figure 2.2 Pg.No
Figure 3.1 Pg.No

vii
Chapter 1
Introduction

1.1 Project Definition:

Maintaining a healthy body in modern society requires careful monitoring of calorie intake

to balance calorie expenditure. An optimal Body Mass Index (BMI) should range between

22 and 28; a value higher than this indicates overweight, and a BMI over 30 signifies

obesity. Achieving and maintaining a healthy weight necessitates vigilance regarding calorie

consumption.

Traditional methods for calorie estimation are manual and often cumbersome, making them

less practical for everyday use. This project presents a novel approach to calorie

measurement utilizing deep learning algorithms, specifically convolutional neural networks

(CNNs), to estimate the calorie content of food from images. This technique is particularly

valuable in the medical field, where accurate calorie counting is essential for dietary

management.

Our model leverages TensorFlow, a robust framework for implementing machine learning

methods, to classify and calculate the calorie content of various food items, including fruits

and vegetables. Users simply capture an image of their food, which serves as the input for

our trained CNN model to detect the food object and calculate its calorie value.

To enhance the user experience, we integrate Google's Generative AI (Gemini) and prompt

engineering techniques. After the image is analyzed by the machine learning model, these

technologies generate comprehensive nutritional information.

Additionally, the knowledge of fine-tuning, prompt engineering, persona pattern, and


template patterns are applied to control the output of the model. This ensures that the

generated nutritional information is accurate and relevant to the user's needs.

This system provides a seamless, automated solution for dietary monitoring, promoting

better health management through precise and convenient calorie counting while

empowering users to understand the health benefits of their food.

1.2 Problem Statement

 Traditional methods of estimating calorie intake and nutritional values, such as

manual tracking or database lookups, are prone to errors, time-consuming, and often

lack accuracy.

 These methods can lead to inaccuracies in dietary records, making it difficult for

users to manage their health effectively and track their nutritional intake

comprehensively.

1.3 Objectives

 Develop a robust food recognition system capable of accurately identifying food

items based on image inputs.

 Implement an intuitive user interface for users to capture food images, view

nutritional information, and track their dietary intake.

 Enhance the efficiency and reliability of nutritional tracking while minimizing the

need for manual input and calculations.

1.4 Benefits of research

 The project will focus on developing a standalone food recognition system that can

be used by individuals for personal dietary management.

 The system will support the identification of multiple food items, real-time image

processing, and nutritional information retrieval.

2
 Integration with existing health and fitness applications or databases may be

considered for future enhancements.

1.5Functional Requirements

1.5.1 User Registration: Allow users to create profiles and store personal dietary

preferences and goals.

1.5.2 Image Capture: Implement a feature for users to capture images of their meals

using a smartphone camera or other devices.

1.5.3 Food Detection: Develop a food detection algorithm to locate and identify

food items in the captured images.

1.5.4 Nutritional Analysis: Retrieve detailed nutritional information for identified

food items, including calories, vitamins, minerals, and other macronutrients.

1.5.5 Dietary Tracking: Automatically log the nutritional values of consumed foods

in a centralized database or personal dietary log.

1.5.6 User Management: Provide functionalities for users to update their profiles,

set dietary goals, and review their nutritional intake history.

1.6 Non-Functional Requirements

1.6.1 Accuracy: The system must achieve high accuracy in food detection and

nutritional analysis to minimize errors.

1.6.2 Speed: Ensure real-time processing of food images for seamless nutritional

tracking without significant delays.

1.6.3 Security: Implement measures to protect sensitive user data, such as personal

dietary information and nutritional logs, from unauthorized access or misuse.

1.6.4 Scalability: Design the system to handle a large number of users and

accommodate future growth in user enrollment and nutritional tracking.

2
1.6.5 Usability: Create an intuitive and user-friendly interface for users to interact

with the system easily.

1.6.6 Reliability: Ensure the system's stability and robustness under various

environmental conditions and input scenarios.

1.6 Limitation of Project

Image analysis using Convolutional Neural Networks (CNNs) has made


significant strides in recent years, but like any technology, it also has its
limitations. Here are some common limitations of image analysis using CNNs:

1.6.1 Data Dependency:


CNNs require large amounts of labeled data for training, which may be
challenging to obtain, especially for niche or specialized domains.

Limited or biased training data can lead to overfitting or poor generalization,


affecting the model's performance on unseen data.

1.6.2 Computational Resources:


Training CNN models, especially deep architectures with millions of parameters,
requires substantial computational resources, including high-performance GPUs
or TPUs.

Inference (i.e., making predictions) may also be computationally intensive,


limiting real-time applications on resource-constrained devices.

1.6.3 Interpretability:
CNNs are often referred to as "black-box" models because of their complex
architectures and high dimensionality, making it challenging to interpret how they
arrive at their predictions.
1.7 Problem Identification and Feasibility

1.7.1 Problem Identification

2
Traditional methods of estimating calorie intake often rely on manual tracking or generic

databases, which can be time-consuming, inaccurate, and inconvenient. These methods do

not provide detailed nutritional information and require significant effort from users to input

data and calculate their intake. This issue is particularly critical for individuals managing

specific dietary needs or medical conditions, where precise nutritional information is vital.

1.7.2 Technical Feasibility

 Machine Learning Models: Advances in machine learning and computer vision,

particularly convolutional neural networks (CNNs), have made it possible to

accurately detect and analyze food items from images.

 Google's Generative AI (Gemini): The use of Gemini and prompt engineering

techniques can generate detailed nutritional information for identified food items,

enhancing the accuracy and utility of the system.

 Hardware Requirements: The system requires standard hardware such as

smartphones or cameras, which are widely accessible and user-friendly.

 Software Development: Utilizing TensorFlow and other machine learning

frameworks, the project involves developing algorithms capable of recognizing food

items and calculating their nutritional values. These technologies are well-

documented and supported by a large developer community.

1.7.3 Operational Feasibility

 User Acceptance: The system aims to offer an intuitive and user-friendly interface,

encouraging user adoption and minimizing the learning curve.

2
 Integration with Existing Health Apps: The system can integrate with existing

health and fitness applications, allowing users to seamlessly incorporate the tool into

their daily routines.

 Training and Support: Comprehensive training materials and technical support will

be provided to users, ensuring smooth implementation and effective use of the

system.

1.7.4 Financial Feasibility

 Cost-Benefit Analysis: Initial investments in hardware and software development

are necessary. However, the benefits, such as accurate nutritional tracking, improved

health outcomes, and reduced effort for users, justify these costs.

 Return on Investment (ROI): The system's ROI will be measured user health,

reduced healthcare costs due to better dietary management, and increased user

engagement and satisfaction.

1.7.5 Legal and Ethical Feasibility

 Data Privacy and Security: The system will comply with data privacy regulations,

ensuring the secure handling and storage of user data. Encryption, access controls,

and secure data storage practices will be implemented.

 Ethical Considerations: Ethical considerations, such as the potential for bias in the

AI model and user privacy concerns, will be addressed through transparent policies

and ongoing stakeholder engagement.

1.8 Constraints

 Hardware Requirements: The system may require specific hardware components,

such as cameras with adequate resolution and processing power for real-time food

recognition.

2
 Lighting Conditions: The performance of the food recognition algorithm may be

affected by variations in lighting conditions, requiring sufficient illumination for

accurate results.

 Environmental Factors: External factors such as noise, occlusions, or changes in

food presentation (e.g., different plating styles) may impact the system's

performance.

1.9 Assumptions

 The system assumes cooperative users who willingly participate in the image capture

process and adhere to dietary tracking procedures.

 It assumes a stable internet connection for accessing remote databases or cloud

storage for storing nutritional data.

2
Chapter 2
Literature Survey

2.1 literature Review

Roth et al. [12] In their work on improving existing CAD systems, the researchers
applied Convolutional Neural Networks (CNNs) to enhance the detection accuracy of
colonic polyps on CT colonography, sclerotic spine metastases on body CT, and
enlarged lymph nodes on body CT. They utilized previously developed candidate
detectors and generated 2D patches from three orthogonal directions, supplemented by
up to 100 randomly rotated views. This method, referred to as "2.5D" views, serves as
a decompositional image representation of the original 3D data. The CNNs processed
these 2.5D views, and their predictions were aggregated to boost accuracy further. The
integration of CNNs led to a significant improvement in sensitivity for lesion detection,
ranging from 13% to 34% across all three CAD systems. This substantial enhancement
demonstrates the generalizability and scalability of the CNN-based approach.
Achieving such improvements had been nearly unattainable with traditional non-deep
learning classifiers, such as committees of support vector machines (SVMs). This
indicates that CNNs provide a more powerful tool for enhancing the performance of
CAD systems in medical imaging.

Dou et al. [13] The researchers focused on detecting cerebral microbleeds from
susceptibility-weighted MRI scans using a two-stage approach involving 3D
Convolutional Neural Networks (CNNs). Unlike traditional methods, they replaced the
candidate detection stage with a CNN, streamlining the detection process. Their results
indicated that the 3D CNN significantly outperformed various classical methods and
2D CNN approaches previously documented in the literature. The authors validated
this by re-implementing, training, and testing these alternative methods on the same
dataset, demonstrating the superior effectiveness of their 3D CNN model.

2
Sirinukunwattana et al. [14] The researchers worked on detecting and classifying nuclei
in histopathological images using a specialized Convolutional Neural Network (CNN).
Their CNN processes small image patches and, rather than simply determining if the
central pixel is part of a cell nucleus, it generates an output characterized by a high
peak near the center of each nucleus and flat responses elsewhere. This spatially
constrained approach, combined with the fusion of overlapping patches during the
testing phase, results in superior performance compared to previous techniques. These
earlier methods included both other CNN-based approaches and classical feature-based
methods.

Anthimopoulos et al. [13] The researchers concentrated on detecting patterns of


interstitial lung diseases using 2D patches from chest CT scans. They were one of three
groups in the study (along with Shin et al. and van Tulder et al.) utilizing a public
dataset. They trained a CNN to classify 32x32 pixel patches into one of seven classes.
Their CNN-based approach achieved higher accuracy compared to three previously
published methods that relied on hand-crafted features.

Van Grinsven et al. [17] The researchers aimed to improve and accelerate CNN
training for medical image analysis by dynamically selecting misclassified negative
samples during training. CNN training involves many iterations (epochs) to optimize
network parameters, with each epoch using a randomly selected subset of training data
for parameter updates via back-propagation. In medical classification tasks, which
often involve distinguishing between normal and pathological cases, the normal class is
usually over-represented. Additionally, many normal samples are highly correlated due
to repetitive patterns in the images, making only a small fraction of these samples
informative. Uniformly treating all data during training wastes time on non-informative
samples, prolonging the process. By identifying and focusing on informative normal
samples, the researchers increased the efficiency of the CNN learning process and
reduced training time.

Tajbakhsh et al. [18] Experiments consistently demonstrated that using pre-trained


CNNs with fine-tuning yielded the best results across various application domains
(Tajbakhsh et al.) and network architectures (Shin et al.). Tajbakhsh et al. found that
deep fine-tuning significantly improved performance compared to shallow fine-tuning,

2
with the benefits becoming more pronounced as the size of the training sets decreased.
Shin et al. reported that the GoogLeNet architecture achieved state-of-the-art detection
of mediastinal lymph nodes, outperforming other shallower architectures.

(Setio et al. [11]) A study by Setio et al. led to the creation of a challenge for
pulmonary nodule detection, organized alongside the IEEE ISBI conference. This
challenge utilizes the publicly available LIDC/IDRI dataset, allowing the system
described in their study to be directly compared with alternative approaches. More
information about the challenge can be found at http://luna16.grand-challenge.org/.

Kallenberg et al. [20] present an unsupervised feature learning method for


mammography risk scoring. This approach builds a feature hierarchy from unlabeled
data, which is then used as input for a simple classifier. The method addresses two
tasks: breast density segmentation and mammographic texture scoring, achieving state-
of-the-art results. To control the model's complexity, a sparsity regularizer that includes
both lifetime and population sparsity is used. The convolutional layers in the
unsupervised sections are trained as autoencoders. In the supervised part, the pre-
trained weights and bias terms are fine-tuned using softmax regression.

The literature features numerous network architectures, with significant variability


evident across current journal works. Options include selecting a known architecture,
designing a task-specific one, or combining different architectures. An interesting
future question is whether the very deep residual networks, which used 152 layers and
excelled in the ILSVRC 2015 classification task [36], will also achieve good results on
medical tasks.

Several challenges exist in this area. Firstly, securing funding for constructing data sets
is difficult. Secondly, high-quality annotation of medical imaging data requires scarce
and expensive medical expertise. Thirdly, privacy concerns make sharing medical data
more challenging compared to natural images. Fourthly, the wide range of applications
in medical imaging necessitates the collection of diverse data sets. Despite these
challenges, there is rapid progress in data collection and sharing. Numerous public data
sets have been released and are routinely used in experimental validation. Notable
2
examples include VISCERAL and The Cancer Imaging Archive
(http://www.visceral.eu/ and http://www.cancerimagingarchive.net/). Roth et al. [12]
and Shin et al. [14] have analyzed a dataset of enlarged lymph nodes on CT scans,
which they have made publicly available on The Cancer Imaging Archive [22]. The
same group has also released a pancreas dataset online [38].

Since 2007, it has become common practice to organize challenge workshops at


medical imaging conferences like MICCAI, ISBI, and SPIE Medical Imaging. This has
resulted in the accumulation of a large number of data sets and ongoing benchmark
studies, cataloged at the website http://www.grand-challenge.org/. Utilizing these
public benchmark data sets offers a distinct advantage over using public data sets
alone. Challenges provide a precise definition of the task to be addressed and establish
one or more evaluation metrics that enable fair and standardized comparisons between
proposed algorithms. Without such standardization, comparing different approaches to
the same problem, even if they use the same data set, can be challenging. This issue is
highlighted in this journal, where three studies (Anthimopoulos et al. [13], Shin et al.
[14], and van Tulder et al. [15]) utilize the same data set of chest CT scans with
annotations of interstitial lung disease patterns [16], yet each reports results in a
different manner.

In other words, are the very deep residual networks, like the one with 152 layers that
excelled in the ILSVRC 2015 classification task, also effective for medical tasks? This
question highlights the potential transferability of network architectures across
different domains, particularly from image classification tasks like ILSVRC to medical
image analysis tasks. While the success of a network architecture in one domain doesn't
guarantee its effectiveness in another, exploring whether these deep residual networks
can achieve good results in medical tasks could lead to valuable insights and
advancements in medical image analysis.

In other words, obtaining accurate ground-truth data for medical tasks presents a
significant challenge due to its scarcity and the complexity involved in its collection.
While crowdsourcing has shown potential in annotating large datasets for general
images, its application in the biomedical field demands a more nuanced approach due
to the need for precise annotations and a deeper understanding of the medical tasks at

2
hand. This challenge underscores the importance of developing specialized methods for
annotating medical data accurately and efficiently.

Albarqouni Al [19] have addressed these challenges by introducing a network


architecture that incorporates an aggregation layer within the convolutional neural
network (CNN). This layer allows the network to learn from annotations provided by
crowdsourcing as part of its training process. Their results offer valuable insights into
how deep CNNs can effectively learn from annotations generated by crowds.
Moreover, what's particularly remarkable in crowdsourcing studies within the medical
domain is the finding that a crowd of non-professionals and inexperienced individuals
can achieve performance comparable to that of medical experts. This observation has
been noted not only by Albarqouni et al. but also by Nguyen et al. and McKenna et al.
in studies involving radiology images.

Brosch et al.[21] and Dou et al. [12] much of the analysis conducted in current works is
carried out in two dimensions (2D), prompting discussions about whether transitioning
to three-dimensional (3D) analysis could lead to significant performance
improvements. Various approaches to data augmentation, including the use of 2.5D
techniques, have been explored. For instance, Roth et al. utilized axial, coronal, and
sagittal images centered on a voxel in a colonic polyp or lymph node candidate,
feeding them into the cudaconvnet CNN. This CNN incorporates three channels
typically used to represent the red, green, and blue color channels of a natural light
image. Explicitly employing 3D convolutional neural networks (CNNs) has been
demonstrated in the works of Brosch et al. and Dou et al., marking a departure from
traditional 2D approaches and potentially offering advantages in capturing spatial
information across multiple dimensions.

2
1.7 Literature Survey:

1.7.1 Automated plankton image analysis using convolutional neural


networks. [3]

The document discusses the development and application of an automated


classification algorithm for in situ plankton imaging systems using convolutional
neural networks (CNNs). The abstract highlights the need for fast and accurate
classification tools for identifying a wide range of organisms and nonliving
particles in plankton images. Previous methods had limitations in resolving
multiple groups accurately, but with CNNs, there was a significant improvement
in classification

accuracy. The document describes an image processing procedure involving


preprocessing, segmentation, classification, and postprocessing to identify 108
classes of plankton using spatially sparse CNNs. The results showed an average
precision of 84% and a recall of 40% for all groups, with higher accuracy
achieved after excluding the rarest taxa.

Fig.2.1

The conclusion emphasizes the effectiveness of the automated classification


scheme using deep learning methods for plankton imaging systems. The study
demonstrated the successful classification of a large dataset into 108 classes, with

2
a focus on biological groups. By applying filtering thresholds on classification
probabilities and grouping classes into taxonomically meaningful groups, the
precision rate for nonrare biological groups reached 90.7%. The document also
discusses the importance of the training set, the impact of filtering on
classification statistics, and the potential for future developments in automated
plankton image analysis using CNNs.

In summary, the document presents a detailed process of using CNNs for


automated plankton image classification, showcasing the significant improvement
in accuracy compared to previous methods. The study highlights the potential for
applying deep learning tools to enhance ecological monitoring and fisheries
management through efficient and accurate classification of plankton images.

They demonstrated the successful application of an image processing procedure,


using a deep learning CNN, to classify a ~ 40 h, 10 TB in situ plankton imaging
dataset containing 25 million image segments into 108 classes. After applying a
filtering threshold on the classification probabilities, and grouping the classes into
37 taxonomically and functionally meaningful groups, the average classifier
precision on nonrare biological groups (n = 23) was 90.7%, which is higher than
any previous attempt on high-sampling volume, in situ plankton images.

1.7.2 Medical Image Analysis using Deep Convolutional Neural


Networks: CNN Architectures and Transfer Learning. [4]

The document delves into the utilization of deep convolutional neural networks
(CNNs) in medical image analysis, particularly in computer-aided detection
(CAD) systems for diseases like breast cancer, lung nodules, and prostate cancer.
The abstract emphasizes the significance of CNNs in enhancing diagnostic
accuracy and efficiency in medical imaging. It discusses the training strategies for
CNNs, including transfer learning and fine-tuning, to address the challenges of
limited labeled data. The study showcases applications of CNNs in various
medical imaging tasks and highlights the potential of deep learning methods in
revolutionizing healthcare practices.

2
Fig.2.2

In conclusion, the document underscores the pivotal role of CNN-based methods


in advancing medical image analysis and CAD systems. It emphasizes the
promising

results achieved in disease detection and diagnosis through CNN architectures and
transfer learning techniques. The study suggests future directions for integrating
deep learning tools in precision medicine, radiomics, and clinical decision support
systems to improve patient outcomes and streamline healthcare processes.

Fig.2.3

The document also provides insights into the evolution of computer-aided


detection (CAD) systems in medical imaging, highlighting the transition from
traditional machine learning techniques to deep learning methods, particularly
CNNs. It discusses the workflow of a typical CAD system, involving image
preprocessing, segmentation, feature extraction, and lesion classification. The
2
paper also addresses the challenges associated with CAD systems in clinical
settings, such as limited annotated data, standardization of datasets, and
integration with existing hospital systems. It explores the potential applications of
CNN-based methods in lesion detection, and classification tasks across various
medical imaging modalities.

1.7.3 CNN-Based Image Analysis for Malaria Diagnosis. [5]

The document discusses the use of CNN-based image analysis for malaria
diagnosis, highlighting the inefficiency of traditional methods and the potential of
machine learning. The abstract introduces a new 16-layer CNN model with a
97.37% accuracy in classifying infected and uninfected red blood cells. Transfer
learning, with 91.99% accuracy, is also compared.

The CNN model outperforms in sensitivity, specificity, precision, F1 score, and

Matthew’s correlation coefficient. The introduction emphasizes the global health


threat of malaria and the need for accurate diagnostics.

The CNN architecture, data preprocessing, and model training process are
detailed. Results show the CNN model's superior performance, attributing it to
both architecture and training data volume. The conclusion suggests deep
learning's potential to enhance malaria diagnosis efficiency and accuracy.
Acknowledgments and references are included, acknowledging funding sources
and previous studies on deep learning for genomics.

The study's significance lies in the improved classification performance of the


CNN model for blood smear analysis, indicating a promising direction for health-
related applications.

2
Fig.2.4

1.7.4 Chest X-ray image analysis and classification for COVID-19


pneumonia detection using Deep CNN. [6]

To accelerate the understanding of COVID-19 mechanisms, this study devised a


novel diagnostic platform utilizing a deep convolutional neural network (CNN) to
aid radiologists in differentiating COVID-19 pneumonia from other types based
on chest X-ray analysis at Middlemore Hospital. This tool streamlines chest X-ray
interpretation, enhancing diagnostic accuracy and expediting COVID-19
detection. By training the CNN on a diverse set of X-ray lung images (normal,
bacterial infection, viral infections including COVID-19), the model learns to
discern relevant information, enabling it to identify patterns indicative of diseases
like coronavirus infection in new images. Employing supervised learning, akin to
a doctor overseeing the learning process, the CNN's accuracy improves with the

2
increasing number of analyzed images. This approach mimics the training process
for a physician but leverages the vast dataset to potentially achieve higher
accuracy levels than human counterparts. The model's ability to learn from a wide
array of images underscores its potential for heightened diagnostic precision in
medical imaging.

Fig.2.5

This research aimed to expedite the understanding of COVID-19 disease


mechanisms by developing a new diagnostic platform using a deep convolutional
neural network (CNN) to assist radiologists in distinguishing COVID-19
pneumonia from other types based on chest X-ray analysis at Middlemore
Hospital. By training the CNN on a diverse set of X-ray lung images, including
normal, bacterial infection, and viral infections such as COVID-19, the model
learned to differentiate between noise and relevant information to identify disease
patterns. Through supervised learning, the CNN's accuracy improved with an
increasing number of analyzed images, potentially surpassing human accuracy
levels.

2
Fig.2.6

2
Chapter 3
Proposed Work

3.1 Proposed Work

3.1.1. Model Optimization and Hyperparameter Tuning:

 Explore different architectures or variations of MobileNetV2 to optimize model

performance.

 Conduct hyperparameter tuning experiments to find the optimal learning rate, batch

size, and other parameters for improved training efficiency and accuracy.

3.1.2. Data Augmentation and Transfer Learning Techniques:

 Experiment with various data augmentation techniques such as rotation, flipping,

and scaling to further enhance model generalization.

 Investigate different transfer learning strategies, including fine-tuning additional

layers of the pre-trained model or using different pre-trained models for feature

extraction.

3.1.3. Class Imbalance Handling:

 Address class imbalance issues in the dataset by employing techniques such as

oversampling, undersampling, or class weighting to ensure fair representation of all

classes during training.

3.1.4. Interpretability and Visualization:

2
 Implement methods for interpreting model predictions and understanding the

underlying decision-making process.

 Visualize feature maps, activation patterns, and class activation maps to gain insights

into what parts of the input images are influencing the model's predictions.

3.1.5. Model Deployment and Integration:

 Develop a user-friendly interface or application for deploying the trained model,

allowing users to upload images and receive predictions in real-time.

 Integrate the model into existing platforms or systems relevant to fruit and vegetable

recognition, such as agricultural monitoring systems or nutrition tracking

applications.

3.1.6. Performance Evaluation and Benchmarking:

 Conduct comprehensive performance evaluation of the trained model on diverse

datasets, including datasets with variations in lighting conditions, backgrounds, and

image qualities.

 Benchmark the model against state-of-the-art approaches in fruit and vegetable

recognition to assess its competitiveness and identify areas for further improvement.

3.1.7. Continuous Monitoring and Model Maintenance:

 Establish a framework for continuous monitoring of model performance and stability

in production environments.

 Implement strategies for model retraining and updating to adapt to changes in data

distribution or user requirements over time.

3.1.8. Collaboration and Knowledge Sharing:

2
 Foster collaboration with domain experts, researchers, and stakeholders in relevant

fields such as agriculture, nutrition, and computer vision to leverage domain

knowledge and ensure the practical relevance of the project.

 Share insights, findings, and best practices through publications, presentations, and

open-source contributions to contribute to the broader research community and drive

collective progress in the field.

3.1.9. Ethical Considerations and Responsible AI Practices:

 Prioritize ethical considerations in model development, deployment, and usage,

ensuring fairness, transparency, and accountability in decision-making processes.

 Proactively address potential biases, privacy concerns, and societal impacts

associated with the application of AI technologies in fruit and vegetable recognition,

promoting responsible AI practices and social good.

3.1.10. Future Directions and Innovation:

 Explore emerging trends and technologies in AI, such as federated learning, meta-

learning, or self-supervised learning, for potential application in fruit and vegetable

recognition tasks.

 Foster a culture of innovation and continuous learning within the project team,

encouraging experimentation, creativity, and exploration of new ideas to push the

boundaries of what is possible in the field.

3.2 Tech Stack Used


3.2.1 Programming Language: Python

Python is the primary programming language used for developing the project

due to its simplicity, readability, and extensive support for data manipulation,

machine learning, and deep learning tasks.

3.2 Libraries and Frameworks:


2
3.2.2.1 NumPy: NumPy is a fundamental package for scientific computing

with Python, providing support for powerful numerical operations and array

manipulation.

3.2.2.2 Pandas: pandas is a versatile data manipulation library that offers

data structures and tools for working with structured data, such as data

frames, and facilitates data preprocessing and analysis.

3.2.2.3 Matplotlib: Matplotlib is a comprehensive library for creating

static, interactive, and animated visualizations in Python, making it

suitable for data visualization tasks.

3.2.2.4 TensorFlow: TensorFlow is an open-source deep learning framework

developed by Google, offering comprehensive support for building and

training deep neural networks.

3.2.2.5 Scikit-learn: scikit-learn is a widely-used machine learning library in

Python, providing simple and efficient tools for data mining and data analysis

tasks, including classification, regression, clustering, and dimensionality

reduction.

3.2.2.6 Keras: Keras is a high-level neural networks API that runs on top of

TensorFlow, allowing for easy and fast prototyping of deep learning models

with a focus on user-friendliness, modularity, and extensibility.

3.2.2.7 Google Colab: Google Colab is a cloud-based Jupyter notebook

environment provided by Google, offering free access to GPU and TPU

resources for running Python code, especially for machine learning and deep

learning tasks.

2
3.2.2.8 Google Generative AI: Google Generative AI is a collection of APIs

and models developed by Google for various generative tasks, such as text

generation, image generation, and content creation.

3.2.3 Development Environment:

3.2.3.1 Google Colab: The code is developed and executed in a Google

Colab environment, which provides a convenient platform for writing,

executing, and sharing Python code in a Jupyter notebook format, with

access to GPU and TPU resources for accelerating deep learning tasks.

3.2.4 Additional Tools and APIs:

3.2.4.1 Kaggle API: The Kaggle API is used to download datasets directly

from the Kaggle platform, simplifying the process of acquiring relevant

datasets for the project.

3.2.4.2 Google Generative AI API: The Google Generative AI API is

utilized for generating content, such as text descriptions and summaries,

based on provided prompts or input data.

3.3 Design
3.3.1 Project Structure:

 The project follows a structured directory layout, with separate directories for data,

code, and outputs.

 The fruit_vegetable_dataset directory contains the dataset used for training and

testing the machine learning model.

 Code files are organized into logical units, such as data preprocessing, model

development, and evaluation.

2
 Output files, such as trained model weights and predicted labels, are saved in

designated directories for easy access and reference.

3.3.2 Data Preprocessing:

 Data preprocessing steps include loading images, resizing them to a uniform size,

and converting them into numerical arrays for input to the machine learning model.

 Image augmentation techniques, such as rotation, zooming, and flipping, are

applied to increase the diversity of the training data and improve model

generalization.

3.3.3 Model Development:

 The project utilizes a pre-trained MobileNetV2 model as the base convolutional

neural network (CNN) architecture.

 Transfer learning is employed by fine-tuning the pre-trained model on the dataset

for the specific task of fruit and vegetable image recognition.

 Additional dense layers are added on top of the pre-trained model to capture high-

level features and perform classification.

 The model is compiled with appropriate loss function, optimizer, and evaluation

metrics for training.

3.3.4 Training and Evaluation:

 The dataset is split into training, validation, and testing sets to train and evaluate the

model.

 The model is trained using the training set, with performance monitored on the

validation set to prevent overfitting.

 Early stopping is employed as a regularization technique to halt training when the

validation loss does not improve.

2
 Model performance metrics, such as accuracy, are computed on the testing set to

evaluate its effectiveness in classifying fruit and vegetable images.

3.3.5 Content Generation:

 Predictions are generated for test images using the trained model, and the predicted

labels are mapped to their corresponding classes.

 Content generation APIs, such as Google Generative AI, are utilized to generate

additional information or descriptions based on the predicted labels or input

prompts.

 Generated content, such as nutritional information, descriptions, or summaries, is

saved and displayed for user reference or further analysis.

3.3.6 User Interface (Optional):

 A user interface may be developed to provide a user-friendly interaction for users

to upload images, view predictions, and access generated content.

 The user interface may also include features for visualizing model performance

metrics, exploring datasets, and customizing model settings.

2
Chapter 4
Methodology

3.1 Introduction:

Convolutional Neural Networks (CNNs) have revolutionized image analysis,


enabling us to extract meaningful information from visual data. This detailed
introduction delves into the core steps involved in implementing an image
analysis system using CNNs, equipping you to harness their power for various
tasks.

Fig.3.1

2
Fig.3.2

2
Fig.3.3

Fig.3.4

2
Fig.3.5

3.2 Hardware and Software Specification

 Hardware Requirements: Processor: Multi-core CPU with sufficient processing


power to efficiently handle image processing tasks.
 A quad-core processor or higher is recommended, such as Intel Core i5 or AMD
Ryzen 5.
 Memory (RAM): Sufficient RAM is essential to effectively handle large data sets
and model training processes.
 A minimum of 8 GB of RAM is recommended, but larger capacity (16 GB or
more) may be beneficial when working with larger data sets or complex models.
 Graphics Processing Unit (GPU): Although not required, using a GPU for model
training can significantly speed up the training process, especially for deep learning
models.
 NVIDIA GPUs such as the GeForce GTX and RTX series are often used to
accelerate deep learning tasks.
 Storage: Sufficient storage space is required to store datasets, model checkpoints,
and other project-related files.
 Solid-state drives (SSDs) are recommended due to faster data access and better
performance compared to traditional hard disk drives (HDDs).
2
 Software Requirements: Operating System: The code appears to be developed and
run in the Google Colab environment, which provides a cloud-based Jupyter
Notebook environment.
 Therefore, operating system requirements are minimal and the code must be
compatible with various operating systems such as Windows, macOS, and Linux.
 Python: code is written in the Python programming language, which is widely used
for machine learning and deep learning tasks.
 Verify that Python is installed on your system.
 Make sure you have the latest version installed (Python 3.
 7 or newer) if possible.
 Libraries and Frameworks: The code utilizes several Python libraries and
frameworks, including NumPy, Pandas, TensorFlow, Matplotlib, and scikit-learn.
 To ensure code compatibility and proper functionality, install these libraries
using a package manager such as pip or conda.
 Kaggle API: This code can also interact with the Kaggle API to download datasets.
 Make sure the Kaggle API is properly configured and authenticated with the
correct credentials.
 Google Generative AI API: The code integrates with the Google Generative AI API
for content generation tasks.
 Make sure your API key is properly configured and has access to the code that
uses the API functions.
 Development Environment: Integrated Development Environment (IDE): Although
your code can run in a Jupyter Notebook environment such as Google Colab, a
feature-rich IDE such as PyCharm or Visual Studio Code makes development easie
 It's more efficient and easier
 Productive debugging and code management.
 Version Control: Implement version control using platforms like Git and GitHub to
track code changes, collaborate with team members, and ensure code reproducibility
and maintainability.
 Documentation and Reports: Maintain comprehensive documentation for your
project, including code comments, README files, and project reports.

2
 Create well-structured and visually appealing documents using tools Jupyter

Chapter 5
Result and Discussion

5.1.1 Model Improvement:

 Enhanced Accuracy: Further improve the accuracy of the model by

experimenting with more advanced deep learning architectures, such as

EfficientNet or ResNet.

 Hyperparameter Tuning: Conduct extensive hyperparameter tuning to

optimize model performance.

 Transfer Learning: Explore transfer learning from other pre-trained models

specific to food or agricultural datasets to enhance recognition accuracy.

5.1.2 Dataset Expansion:

 Increased Diversity: Expand the dataset to include a wider variety of fruits

and vegetables, including different species, ripeness levels, and conditions

(e.g., fresh, rotten).

 Larger Dataset: Increase the number of images per class to provide the

model with more training data and improve generalization.

 Crowdsourced Data: Encourage contributions from users to upload images

and help in annotating them to continuously grow and diversify the dataset.

5.1.3 Real-Time Application:

2
 Mobile Application: Develop a mobile application that uses the trained

model for real-time fruit and vegetable recognition through the device’s

camera.

 Edge Computing: Implement edge computing solutions to perform

inference on-device, reducing latency and dependency on internet

connectivity.

5.1.4 Additional Features:

 Nutritional Information: Integrate a comprehensive database of nutritional

information that provides detailed insights into the health benefits, calorie

content, and other nutritional values of the recognized fruits and vegetables.

 Recipe Suggestions: Provide recipe suggestions based on the recognized

fruits and vegetables, helping users make healthy and creative meals.

 Shopping Assistance: Incorporate a shopping list feature that allows users to

add recognized items directly to a shopping list, with options for nutritional

goals or dietary restriction

5.1.5 Integration with Other Technologies:

 Augmented Reality (AR): Use AR to overlay information about fruits and

vegetables in real-time, enhancing the user experience in grocery stores or

educational settings.

 Voice Assistants: Integrate with voice assistants like Google Assistant or

Alexa to enable hands-free interaction and information retrieval.

 IoT Integration: Connect with IoT devices such as smart refrigerators to

keep track of available produce and suggest what to buy next or what recipes

to make.

2
5.1.6 Research and Development:

 Advanced Algorithms: Explore advanced computer vision techniques, such

as GANs (Generative Adversarial Networks) to generate synthetic data for

training or style transfer for better image augmentation.

 Explainability: Develop methods to provide explainable AI (XAI) for the

model, allowing users to understand how and why certain predictions are

made.

 Continuous Learning: Implement mechanisms for the model to

continuously learn from new data, improving its performance over time

without extensive retraining.

5.1.7 Commercial Applications:

 Retail Integration: Partner with grocery stores and supermarkets to integrate

the technology into self-checkout systems, enhancing customer experience

and reducing the need for manual barcode scanning.

 Agricultural Sector: Use the model to assist farmers in monitoring crop

health, identifying pests, and optimizing harvest times through accurate

recognition of fruits and vegetables.

2. 5.2 Snapshot of system:

2
Fig.5.1

5.3 Snapshots of database tables:

2
Fig.5.2

2
Chapter 6
Conclusion, Limitation and Future Scope

5.1 Conclusion

The project successfully developed a deep learning model capable of recognizing and

classifying a variety of fruits and vegetables from images. By leveraging the

MobileNetV2 architecture pre-trained on the ImageNet dataset, we achieved a

balance between model performance and computational efficiency. The use of data

augmentation techniques and the ImageDataGenerator for preprocessing ensured that

the model generalizes well to new, unseen data.

5.2 Key Achievements:

 Dataset Preparation: A comprehensive dataset of fruits and vegetables was prepared,

including training, validation, and test splits. This dataset was crucial for training a robust

model.

 Model Training: The MobileNetV2 architecture was fine-tuned to classify 36 different

categories of fruits and vegetables. The model demonstrated strong performance metrics,

indicating its reliability.

 Image Processing Pipeline: A systematic pipeline was implemented for image loading,

preprocessing, and augmentation, enhancing the model's ability to handle real-world

variability in input images.

 Prediction and Output Generation: The model was capable of making accurate predictions on

test images. An example workflow was provided to demonstrate the process of using the

trained model to predict and store the labels of new images.


2
 Integration with Google Gemini AI:

 Enhanced Nutritional Lookup: By integrating the model with Google Gemini AI, we added a

layer of intelligence to the application. Gemini AI provides detailed nutritional information

and descriptive content about the identified fruits and vegetables, enriching the user

experience.

 Natural Language Processing: Using Google Gemini AI’s generative capabilities, the

application generates comprehensive and user-friendly descriptions of each fruit or

vegetable, including calorie information. This integration bridges the gap between image

recognition and contextual understanding, making the app more informative and engaging.

5.3 Practical Applications

 Nutrition Lookup: By integrating the model with an application that provides nutritional

information, users can easily obtain detailed insights about various fruits and vegetables.

 Educational Tool: The project can serve as an educational tool for learning about different

fruits and vegetables, their nutritional content, and their health benefits.

 Retail and Agriculture: The technology can be extended to commercial applications in the

retail and agricultural sectors, enhancing inventory management and crop monitoring.

5.4 Future Prospects:

 The project lays a solid foundation for future enhancements, including the exploration of

more advanced architectures, expansion of the dataset, and integration with mobile and IoT

devices. By continually improving the model and expanding its application scope, this

project can evolve into a valuable tool for a wide range of users.

 In conclusion, this project showcases the potential of deep learning and generative AI in

practical image recognition tasks. It demonstrates the feasibility of creating an efficient and

accurate model for fruit and vegetable classification, enhanced by the contextual intelligence

provided by Google Gemini AI. This combination paves the way for further innovations and

applications in the field of computer vision and beyond.


2
REFRENCES

1. TensorFlow Documentation

TensorFlow Core Team. (2023). TensorFlow 2.0 Documentation. Retrieved from

https://www.tensorflow.org/

2. Keras Documentation

Chollet, F. et al. (2023). Keras: The Python Deep Learning library. Retrieved from

https://keras.io/

3. MobileNetV2: Inverted Residuals and Linear Bottlenecks

Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. (2018).

MobileNetV2: Inverted Residuals and Linear Bottlenecks. Proceedings of the IEEE

Conference on Computer Vision and Pattern Recognition (CVPR), 4510-4520.

Retrieved from https://arxiv.org/abs/1801.04381

4. ImageDataGenerator Documentation

Keras Team. (2023). Image Preprocessing. Retrieved from

https://keras.io/api/preprocessing/image/

5. Kaggle Dataset: Fruit and Vegetable Image Recognition

Seth, K. (2023). Fruit and Vegetable Image Recognition. Retrieved from

https://www.kaggle.com/datasets/kritikseth/fruit-and-vegetable-image-recognition

6. Google Gemini AI Documentation

Google Cloud AI Team. (2024). Google Gemini AI API Documentation. Retrieved

from https://cloud.google.com/gemini-ai

2
7. Python Data Analysis Library (Pandas)

McKinney, W. (2023). Pandas: Powerful Python data analysis toolkit. Retrieved

from https://pandas.pydata.org/

8. NumPy Documentation

Harris, C. R., Millman, K. J., van der Walt, S. J., et al. (2020). Array programming

with NumPy. Nature, 585(7825), 357-362. Retrieved from https://numpy.org/

9. Matplotlib: Visualization with Python

Hunter, J. D. (2007). Matplotlib: A 2D Graphics Environment. Computing in

Science & Engineering, 9(3), 90-95. Retrieved from https://matplotlib.org/

10. IPython and Jupyter Notebooks

Perez, F., & Granger, B. E. (2007). IPython: A System for Interactive Scientific

Computing. Computing in Science & Engineering, 9(3), 21-29. Retrieved from

https://ipython.org/

11. A. Setio et al., “Pulmonary nodule detection in CT images using multiview

convolutional networks,” IEEE Trans. Med. Imag., vol. 35, no. 5, pp. 1160–1169,

May 2016.

12. H. Roth et al., “Improving computer-aided detection using convolutional neural

networks and random view aggregation,” IEEE Trans. Med. Imag., vol. 35, no. 5,

pp. 1170–1181, May 2016.

13. M. Anthimopoulos, S. Christodoulidis, A. Christe, and S. Mougiakakou, “Lung

pattern classification for interstitial lung diseases using a deep convolutional neural

network,” IEEE Trans. Med. Imag., vol. 35, no. 5, pp. 1207–1216, May 2016.

2
14. ] H.-C. Shin et al., “Deep convolutional neural networks for computeraided

detection: CNN architectures, dataset characteristics and transfer learning,” IEEE

Trans. Med. Imag., vol. 35, no. 5, pp. 1285–1298, May 2016

15. G. van Tulder and M. de Bruijne, “Combining generative and discriminative

representation learning in convolutional restricted Boltzmann machines,” IEEE

Trans. Med. Imag., vol. 35, no. 5, pp. 1262–1272, May 2016.

16. ] A. Depeursinge et al., Comput. Med. Imag. Graph., vol. 36, no. 3, pp. 227–238,

2012.

17. ] M. van Grinsven, B. van Ginneken, C. Hoyng, T. Theelen, and C. Sánchez, “Fast

convolutional neural network training using selective data sampling: Application to

hemorrhage detection in color fundus images,” IEEE Trans. Med. Imag., vol. 35, no.

5, pp. 1273–1284, May 2016

18. ] N. Tajbakhsh et al., “Convolutional neural networks for medical image analysis:

Full training or fine tuning?,” IEEE Trans. Med. Imag., vol. 35, no. 5, pp. 1299–

1312, May 2016.

19. S. Albarqouni, C. Baur, F. Achilles, V. Belagiannis, S. Demirci, and N. Navab,

“Agg-Net: Deep learning from crowds for mitosis detection in breast cancer

histology images,” IEEE Trans. Med. Imag., vol. 35, no. 5, pp. 1313–1321, May

2016.

20. ] M. Kallenberg et al., “Unsupervised deep learning applied to breast density

segmentation and mammographic risk scoring,” IEEE Trans. Med. Imag., vol. 35,

no. 5, pp. 1322–1331, May 2016.

21. ] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition

ArXiv, 2015 [Online]. Available: arXiv:1512.03385, to be published

2
22. [37] H. R. Roth et al., “A new 2.5 d representation for lymph node detection in CT,”

Cancer Imag. Arch., 2015 [Online]. Available: http://dx.doi.

org/10.7937/K9/TCIA.2015.AQI

23. [38] H. R. Roth et al., “Data from pancreas-CT,” Cancer Imag. Arch., 2016

[Online]. Available: http://dx.doi.org/10.7937/K9/TCIA.2016. tNB1kqB

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy