Mini Project Report Format
Mini Project Report Format
BACHELOR OF ENGINEERING
in
SAHYADRI
College of Engineering & Management
An Autonomous Institution
MANGALURU
2024 - 25
SAHYADRI
College of Engineering & Management
An Autonomous Institution
MANGALURU
COMPUTER SCIENCE AND ENGINEERING
(ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING)
CERTIFICATE
This is to certify that the Mini Project entitled “Hand Gesture Recognition
Model” has been carried out by Afnan Abdul Rahiman (4SF22CI006), Ahmed
Luthfulla Kazi (4SF22CI007), Chiranthan Sukumar (4SF22CI025) and
Subhajit Sahoo (4SF22CI099), the bonafide students of Sahyadri College of
Engineering & Management in partial fulfillment of the requirements for the V semester
Mini Project (AM522P7A) of Bachelor of Engineering in Computer Science
and Engineering(AI&ML) of Visvesvaraya Technological University, Belagavi during
the year 2024 - 25. It is certified that all corrections/suggestions indicated for Internal
Assessment have been incorporated in the report deposited in the departmental library.
The mini project report has been approved as it satisfies the academic requirements in
respect of mini project work.
DECLARATION
We hereby declare that the entire work embodied in this Mini Project Report titled
“Hand Gesture Recognition Model” has been carried out by us at Sahyadri
College of Engineering and Management, Mangaluru under the supervision of Mrs.
Sadhana Rai as the part of the V semester Mini Project (AM522P7A) of
Bachelor of Engineering in Computer Science and Engineering(AI&ML). This
report has not been submitted to this or any other University.
This research introduces an innovative hand gesture recognition model developed for real-
time human-computer interaction. The model leverages the power of convolutional neural
networks (CNNs) combined with advanced image processing techniques to accurately
detect and interpret both static and dynamic hand gestures. By incorporating a diverse
range of datasets, the model is designed to ensure robustness and adaptability, making
it capable of performing well under various environmental and lighting conditions.
The versatility of the model is evident in its wide range of applications. It holds
significant promise in areas such as sign language interpretation, enabling seamless
communication for individuals with hearing and speech impairments. In the gaming
industry, it opens up possibilities for immersive and interactive gaming experiences
through touchless controls. Additionally, the model has potential in touchless interfaces,
contributing to more hygienic and user-friendly interaction systems in healthcare, public
spaces, and smart homes.
The system demonstrates exceptional accuracy and computational efficiency,
ensuring real-time performance without compromising precision. This makes it suitable
for deployment on edge devices, smartphones, or integrated systems. By bridging the
gap between human gestures and machine understanding, this model represents a
significant step forward in creating intuitive, accessible, and interactive technology
solutions for a wide range of users and applications.
i
Acknowledgement
It is with great satisfaction and euphoria that we are submitting the Mini Project
Report on “Hand Gesture Recognition Model”. We have completed it as a part of
the V semester Mini Project (AM522P7A) of Bachelor of Engineering in
Computer Science and Engineering(AI&ML) of Visvesvaraya Technological
University, Belagavi.
We are profoundly indebted to our guide, Mrs. Sadhana Rai , Assistant Professor,
Department of Computer Science and Engineering(AI&ML) for innumerable acts of
timely advice, encouragement and We sincerely express our gratitude.
We express our sincere gratitude to Dr. Pushpalatha K, Professor & Head, Department
of CSE(AI&ML) for her invaluable support and guidance.
Finally, yet importantly, We express our heartfelt thanks to our family & friends for
their wishes and encouragement throughout the work.
ii
Table of Contents
Abstract i
Acknowledgement ii
Table of Contents iv
List of Figures v
List of Tables v
1 Introduction 1
1.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Purpose . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3 Scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
2 Literature Review 3
2.1 Literature Survey . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
3 Problem Formulation 8
3.1 Problem Statement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.2 Problem Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.3 Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
4 Requirements Specification 10
4.1 Hardware Specification . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
4.2 Software Specification . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
5 System Design 11
5.1 System Architecture Diagram . . . . . . . . . . . . . . . . . . . . . . . . 11
6 Implementation 14
iii
8 Conclusion and Future work 20
References 22
iv
List of Figures
v
Chapter 1
Introduction
1.1 Overview
Hand gesture recognition is a cutting-edge technology that bridges the gap between
humans and machines, enabling intuitive interaction without the need for traditional
input devices. By leveraging advancements in computer vision, machine learning, and
1
Hand Gesture Recognition Model Chapter 1
1.2 Purpose
The purpose of hand gesture recognition is to create an intuitive and natural interface
for human-computer interaction, eliminating the need for traditional input devices like
keyboards, mice, or touchscreens. By interpreting human gestures as commands, this
technology enhances accessibility, improves user experience, and enables seamless control
in various applications. It aims to provide solutions in fields such as virtual reality,
robotics, gaming, assistive technologies, and smart home automation. Additionally, hand
gesture recognition seeks to make interactions more inclusive, catering to individuals with
disabilities, while advancing the development of touchless systems for improved hygiene
and convenience in healthcare, public spaces, and beyond.
1.3 Scope
The scope of hand gesture recognition spans various fields, including virtual and
augmented reality, robotics, assistive devices, gaming, and smart home systems. It
enhances natural and touchless interaction, empowering individuals and transforming
industries like healthcare, education, and public safety. With advancements in AI and
sensor technologies, its applications are rapidly expanding, offering intuitive and
immersive solutions.
Literature Review
The literature survey helps in understanding the existing research done on refinement of
Unstructured Electronic Medical records. The current state of research demonstrates the
knowledge of scholarly debates around a topic. The information gathered here will help
identify the gaps in the current work. This chapter is the most important part of the
report as it shows the broader picture of Hand Gesture Recognition Model.
Hand gesture recognition (HGR) has become a key area of research, enabling a natural
and intuitive interface for human-computer interaction (HCI). By interpreting hand
movements and postures, HGR systems allow users to control devices and applications
without the need for physical input devices. This technology is widely applied in
gaming, virtual reality, sign language interpretation, and assistive technologies. The rise
of deep learning has transformed HGR, particularly with the use of Convolutional
Neural Networks (CNNs). CNNs excel at extracting intricate features from hand images
and videos, achieving state-of-the-art performance by leveraging large datasets and
modern computing power. These models have significantly improved accuracy and
robustness, making HGR systems practical and effective across diverse environments.
Beyond traditional applications, HGR has the potential to revolutionize areas such as
healthcare, education, and remote collaboration. For instance, it can facilitate touchless
interaction in sterile environments, enhance accessibility for individuals with physical
disabilities, and support immersive learning experiences in virtual classrooms. With the
integration of advanced sensors and multimodal data, the scope of HGR is continually
expanding. With continued advancements in deep learning and hardware technologies,
HGR is poised to play a pivotal role in shaping the future of human-computer interaction.
3
Hand Gesture Recognition Model Chapter 2
sequential data. It emphasizes the need for efficient computation to handle large
vocabularies in real-time applications.
Problem Formulation
3.3 Objectives
• To develop a system that can recognize and interpret common hand gestures in
real-time for practical applications.
8
Hand Gesture Recognition Model Chapter 3
• To provide an easy-to-use solution for interacting with devices using simple hand
movements.
Requirements Specification
• RAM : 8GB
10
Chapter 5
System Design
The architecture diagram depicts the workflow of a hand gesture recognition system. It
highlights both the training and testing phases.
11
Hand Gesture Recognition Model Chapter 5
Training Phase:
3. Hand Region Detection The system identifies the hand region within each image,
separating it from the background. This step ensures that only the hand gesture is
processed further.
5. Features Database The extracted features are stored in a database. This database
serves as a reference for comparing and identifying gestures during the testing phase.
Testing Phase:
1. Input: Testing Image A single testing image containing a hand gesture is fed
into the system to test its recognition capabilities.
2. Pre-Processing Similar to the training phase, the testing image undergoes pre-
processing to standardize it for feature comparison.
3. Hand Region Detection The system identifies the hand region in the testing
image to isolate the gesture.
4. Feature Extraction Features from the detected hand region in the testing image
are extracted using the same process as in the training phase.
5. Feature Matching The extracted features of the testing image are compared
against the stored features in the database. The matching process determines the
gesture by finding the closest match.
Key Workflow:
• Testing Phase utilizes this database to recognize gestures from new images.
This architecture ensures that the system learns from training data and accurately
classifies gestures during testing.
Implementation
This chapter describes the key modules of the Hand Gesture Recognition System and
their functionality. Each module plays a specific role in ensuring accurate and efficient
recognition of hand gestures for human-computer interaction. Below are the main
modules, their explanations, and pseudocode.
The code is divided into several modules, each handling a specific aspect of the hand
gesture recognition system. Below are the main modules with their detailed functionality:
1. Input Module
14
Hand Gesture Recognition Model Chapter 6
This module detects the presence of a hand in the video frame using the HandDetector
class.
Functionality:
• Identifies hands in the frame and returns their bounding box and landmarks.
3. Preprocessing Module
This module prepares the detected hand region for classification by cropping, resizing,
and standardizing the input.
Functionality:
• Resizes the cropped image to fit a fixed input size (300x300 pixels). j
4. Classification Module
This module uses a pre-trained deep learning model to classify the hand gesture.
Functionality:
This module overlays the classification results on the video frame and displays it.
Functionality:
• Displays intermediate outputs (cropped hand, resized image) and the final output
frame.
The implementation of this hand gesture recognition system provides the following results:
• It identifies gestures from a predefined set of labels (Hello, I love you, No, Okay,
Please, Thank you, Yes) with accuracy depending on the training quality of the
model.
• The confidence level for each classification is generated, providing insight into the
reliability of the prediction.
3. Robust Preprocessing
• The system handles variable hand sizes and positions by maintaining the aspect
ratio during preprocessing.
• It ensures uniformity in input to the classifier by centering the resized hand image
on a white canvas.
17
Hand Gesture Recognition Model Chapter 7
4. Visual Output
• The bounding box and the recognized gesture label are displayed on the video frame.
• Intermediate outputs (cropped hand and resized hand on the white canvas) are
shown for debugging and verification.
6. Usability
• The system provides a natural and intuitive way for Human-Computer Interaction
(HCI) without physical input devices.
The hand gesture recognition system developed in this project successfully demonstrates
the potential of computer vision and deep learning for intuitive and natural human-
computer interaction. By leveraging advanced preprocessing techniques and a pre-trained
CNN model, the system is capable of accurately detecting and classifying predefined
gestures in real time. This implementation highlights the importance of effective image
preprocessing, such as aspect ratio preservation and input normalization, in ensuring
consistent performance. The system’s usability extends to applications in virtual reality,
assistive technologies, and touchless interfaces, providing a step toward more immersive
and accessible interaction mechanisms. However, the project also underscores the need
for addressing challenges like lighting variability, background noise, and gesture diversity
to improve adaptability in real-world scenarios. With further enhancements in dataset
diversity, model training, and environmental robustness, this system has the potential to
serve as a foundational tool for numerous innovative applications, including sign language
recognition, gaming, and beyond.
The current hand gesture recognition system provides a solid foundation for
real-time interaction, but several avenues exist for future improvement and expansion.
These include increasing the gesture set to handle complex interactions like sign
language, enhancing environmental robustness by addressing varying conditions like
lighting and occlusions, and extending the system to recognize multi-hand gestures.
Incorporating 3D gesture recognition with depth sensors, adapting the system for
seamless integration with AR/VR platforms, and optimizing the model for edge devices
are also key areas for development. Furthermore, implementing a real-time feedback
system, integrating context-aware gestures, supporting continuous gesture streams, and
enabling personalized gesture recognition would significantly enhance user experience
20
Hand Gesture Recognition Model Chapter 8
and interaction capabilities. Addressing these future directions will make the system
more versatile and applicable in various fields such as healthcare, education, and
entertainment.
[1] Mohammed, A. A., Lv, J., Islam, M. S., Sang, Y. (2023). Multi-model ensemble
gesture recognition network for high-accuracy dynamic hand gesture recognition.
Journal of Ambient Intelligence and Humanized Computing, 14 (6), 6829-6842.
[2] Smith, J., Johnson, R., Davis, L. (2022). A geometric model-based approach to hand
gesture recognition. Proceedings of the IEEE International Conference on Computer
Vision and Pattern Recognition (CVPR), 135–142.
[3] Martinez, R., Zhou, P. (2020). Hand gesture recognition based on computer
vision: A review of techniques. Proceedings of the IEEE International Symposium
on Advances in Artificial Intelligence (AIAI), 78–85.
[4] Kim, H., Park, J., Lee, T. (2021). Real-time hand gesture recognition based on deep
learning YOLOv3 model. IEEE International Conference on Artificial Intelligence
Applications (ICAI), 244–250.
[5] Roy, P., Bhattacharya, A., Sen, S. (2021). A review of the hand gesture recognition
system: Current progress and future directions. IEEE Transactions on Human-
Machine Systems, 51 (4), 379–390.
[6] Chen, L., Wang, H., Li, K. (2020). Real-time hand gesture recognition using
surface electromyography and machine learning. Journal of Neural Engineering,
17 (5), 056014.
[7] Agrawal, M., Ainapure, R., Agrawal, S., Bhosale, S., Desai, S. (2020). Models
for hand gesture recognition using deep learning. In 2020 IEEE 5th International
Conference on Computing Communication and Automation (ICCCA) (pp. 589–594).
IEEE.
[8] Smith, L., Wang, Y., Zhou, H. (2024). HaGRID—Hand Gesture Recognition Image
22
Dataset. Proceedings of the 18th International Conference on Pattern Recognition
(ICPR), 1620–1627.
[9] Brown, A., Johnson, D. (2024). Diverse hand gesture recognition dataset.
International Journal of Computer Vision, 132 (7), 1152–1163.
[10] Lee, C., Kim, J., Choi, Y. (2022). Real-time hand gesture recognition using fine-
tuned convolutional neural network. Applied Soft Computing, 123, 107895.
[11] Wu, T., Li, Y. (2021). Development of hand gesture recognition system using
machine learning. Expert Systems with Applications, 165, 113896.
[12] Smith, A., Johnson, M., Lee, P. (2012). A comparison of machine learning algorithms
applied to hand gesture recognition. Pattern Recognition Letters, 33 (9), 1233–1240.
[13] Gupta, P., Sharma, A., Jain, M. (2022). An experimental analysis of various machine
learning algorithms for hand gesture recognition. Journal of Artificial Intelligence
Research, 67 (1), 45–56.
[14] Park, D., Kim, H., Choi, S. (2017). Hand gesture recognition using machine learning
and the Myo armband. Sensors, 17 (9), 2043.
[15] Wang, Z., Zhang, Y. (2021). Hand gesture recognition using machine learning and
infrared information. IEEE Transactions on Instrumentation and Measurement, 70,
1–8.
[16] Zhou, X., Chen, L. (2017). Deep learning in vision-based static hand gesture
recognition. Pattern Recognition, 72, 285–292.
[17] Liu, M., Zhang, X. (2018). Research on the hand gesture recognition based on
deep learning. Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition Workshops (CVPRW), 67–75.
[18] Brown, K., Wilson, T. (2020). Hand gesture recognition using machine learning
algorithms. Proceedings of the 2020 International Conference on Machine Learning
and Applications (ICMLA), 123–130.
[19] Johnson, R., Smith, J. (2019). Impact of machine learning techniques on hand
gesture recognition. IEEE Transactions on Multimedia, 21 (5), 1137–1147.
23
[20] Singh, A., Patel, R. (2020). Deep learning-based approach for sign language gesture
recognition. Pattern Recognition, 108, 107561.
24