Mobilenet For Image Classification
Mobilenet For Image Classification
1 Introduction
Deep learning has revolutionised computer vision, particularly in tasks such as image
classification, object detection, and semantic segmentation. However, traditional
convolutional neural networks like VGG and ResNet are computationally intensive, limiting
their deployment in mobile or edge environments.
Despite their success, CNNs such as AlexNet, VGG16, and ResNet are computationally
expensive, requiring powerful GPUs and significant memory, which makes them unsuitable
for real-time deployment on mobile and IoT devices.
3.2 MobileNetV1
Introduced in 2017 by Howard et al. (Google), MobileNetV1 is based entirely on depthwise
separable convolutions. It includes two hyperparameters:
3.3 MobileNetV2
MobileNetV2, released in 2018, introduces:
• Inverted Residual Blocks: A residual structure with narrow input/output and wide
intermediate layers.
• Linear Bottlenecks: Prevent information loss by using linear activation at the output of
residual blocks.
3.4 MobileNetV3
MobileNetV3, released in 2019, combines:
Using transfer learning to adapt the model to new domains with minimal training data.
Tools such as TensorFlow Lite, PyTorch Mobile, and ONNX allow developers to deploy and
optimise MobileNet models for inference on mobile and embedded systems.
7 Conclusion
MobileNet provides an effective solution for performing image classification on devices with
limited computational resources. Its design, rooted in efficient convolutional operations and
compact architecture, has proven essential for applications in mobile AI and edge
computing. With continuing improvements in model compression and neural architecture
search, MobileNet remains a key model in the ongoing development of lightweight deep
learning solutions.