Talk N Move
Talk N Move
Project report in partial fulfillment of the requirement for the award of the degree of
Bachelor of Technology
In
Computer Science and Engineering
Submitted By
University Area, Plot No. III – B/5, New Town, Action Area – III, Kolkata – 700160.
1
2
ACKNOWLEDGEMENT
We would like to take this opportunity to thank everyone whose cooperation and
encouragement throughout the ongoing course of this project remains invaluable to
us.
We are sincerely grateful to our guide Prof. Nilanjan Chatterjee and Prof. Anay
Ghosh of the Department of Computer Science and Engineering, UEM, Kolkata, for
his wisdom, guidance and inspiration that helped us to go through with this project
and take it to where it stands now.
Last but not the least, we would like to extend our warm regards to our families and
peers who have kept supporting us and always had faith in our work.
Shrestha Paul
Debkanta Biswas
Dhurbojyoti Bhattacharjee
Ronik Mondal
Parthiv Sikdar
Kaustav Mukherjee
Soumili Ghosh
Shibaji Chakraborty
Shreyasi Hazari
Jeet Dutta
3
TABLE OF CONTENTS
ABSTRACT.............................................<<5-6>>
CHAPTER – 1: INTRODUCTION...........................<<7-8>>
.............................................<<9-10>>
.............................................<<11-13>>
.............................................<<14-15>>
.............................................<<16>>
.............................................<<17-18>>
.............................................<<19>>
.............................................<<20>>
BIBLIOGRAPHY................................................<<21>>
4
ABSTRACT
In this project involves the creation of a hand gesture recognition
system based on computer vision for mouse pointer control using real-
time hand movements. Using Python and computer vision
libraries such as OpenCV, the system monitors the hand gestures of the
user and converts them
into equivalent mouse activities like movement, left-click, right-click,
and scrolling. The project does away with the need
for conventional input devices, providing a hands-free and interactive
experience.
6
INTRODUCTION
Over the last few years, motion detection and gesture recognition
technology have picked up steam in various markets. As computer vision
capabilities evolve and touch-free interactions gain greater importance,
both technologies are changing the dynamics of how we communicate
with digital technology. From advanced gaming applications to seamless
accessibility technology, gesture recognition has become increasingly
sought after to boost user engagement and convenience.
Gesture recognition has received increased interest in the past few years
as it promises to offer a more immersive and convenient experience.
Through the elimination of physical input devices, users can naturally
interact with their computers via easy hand movements. This technology
is especially useful in applications such as healthcare, where contactless
interaction is paramount, and in gaming, where hand gestures can
increase the level of realism of the gaming experience.
8
Automotive technology also has an advantage in gesture
recognition. Advanced driver-assistance systems (ADAS) incorporate
gesture controls to make the driving experience more comfortable and
safer. Drivers are able to manage tasks such as volume adjustment, call
acceptance, or navigation systems using hand movements, reducing
distraction.
Apart from its functional uses, this project also solves issues of
accessibility. People with mobility impairments might have a hard time
using conventional input devices. With the implementation of gesture
control, the system offers an accessible computing experience that can
improve productivity and autonomy among users with physical
disabilities.
9
LITERATURE SURVEY
1. Voice Calling:
Hands-free communication is particularly beneficial for individuals
with physical disabilities, allowing them to make and manage calls
effortlessly.
Reduces distractions for professionals, such as drivers or workers
handling machinery, by enabling voice-activated dialling.
10
Enhances convenience by eliminating the need to manually search
contacts and dial numbers.
PROBLEM STATEMENT
Motion recognition combined with AI voice control has proven to be a
highly promising development in human-computer interaction. This
technology uses artificial intelligence, computer vision, and voice
recognition to provide natural and easy-to-use user interfaces. Through
integrating gesture recognition and voice commands, AI voice assistants
provide users with greater control over devices and applications. This
survey of literature tries to present a thorough review of motion
detection systems, their functions, voice calling, email management,
search operations, and mouse gesture control applications, especially in
helping disabled people.
AI voice assistants such as Siri, Google Assistant, and Alexa have already
proven themselves to be effective tools for voice calling and email
management. Recent developments have added motion detection to
further simplify these features. Research indicates that users can make,
receive, or reject calls through simple hand movements. For instance,
waving a hand can reject an incoming call, while a thumbs-up can accept
it.
Mouse gesture control systems have attracted much attention for their
use in personal computing. By monitoring hand movements through a
webcam or dedicated sensors, these systems map gestures onto cursor
movement. Different algorithms, including Kalman filtering and optical
flow analysis, have been used by researchers to provide smooth and
accurate cursor control.
Notwithstanding the progress, there are still some challenges to the use
of motion detection AI voice assistants. Varying lighting levels,
background noise, and hardware capabilities can affect system
performance. Experts recommend the development of strong algorithms
that can perform in different environments.
7. Conclusion
15
PROPOSED
SOLUTION
16
Support voice composition of emails with real-time suggestions based on
NLP.
Add a gesture-based system for rapid-action replies, forwards, or
deleting emails.
6. Accessibility Features:
Implement an accessible system for people with disabilities by providing
gesture customization.
Integrate AI-driven gesture prediction for persons with mobility
disabilities.
Offer voice-activated cues for visually impaired users.
17
EXPERIMENTAL SETUP AND RESULT
ANALYSIS
Hardware Requirements:
Computer: Ensure the hardware has sufficient computational
resources to run the assistant smoothly.
Microphone: Choose a quality microphone
for accurate speech input recognition.
Camera: If implementing camera functionality, select a
suitable camera compatible with the hardware and software
setup.
Software Requirements:
Python and Necessary Libraries: Install Python and required
libraries using package managers like pip.
Development Environment: Set up a development
environment such as Anaconda or a virtual environment for
managing dependencies.
VoIP Service: If incorporating calling functionality, sign up for a
VoIP service like Twilio and configure it for integration with the
assistant.
Libraries Required:
18
wikipedia – Retrieves summarized information from Wikipedia
based on user queries.
Installation of Dependencies.
Install all required Python libraries using pip install -r requirements.txt.
Ensure API keys for weather, news, and Twilio are properly configured in
the script.
19
Testing Environment:
Check if the assistant accurately tracks hand gestures for mouse control.
Evaluate click, right-click, and cursor movement accuracy.
Measure latency in gesture recognition and execution.
20
CONCLUSION
With the rapid evolution of AI, future AI-powered voice assistants will
become more intuitive, context-aware, and capable of handling
complex tasks with minimal user effort. The seamless integration of
motion detection with AI voice technology will open new avenues for
innovation, ultimately reshaping the way humans interact with digital
systems. As researchers and developers refine these systems, AI-driven
voice and motion control will play a crucial role in making technology
more accessible, efficient, and user-friendly in the coming years.
The motion detection system for converting hand gestures into mouse
activities successfully demonstrates its potential as a user-friendly and
accessible interface. Our project demonstrates the feasibility of using
computer vision and machine learning techniques to develop a robust
and intuitive hand gesture recognition system. By offering an
alternative to traditional input devices, the solution promotes
inclusivity, enhances user experience, and paves the way for more
interactive computing methodologies.
21
In conclusion, our proposed system successfully bridges the gap
between traditional input devices and modern, intuitive interaction
methods by implementing real-time gesture recognition for mouse
control. The project demonstrates significant potential in areas such as
gaming, accessibility, and medical applications, offering an innovative,
hands-free computing experience.
22
FUTURE SCOPE
23
BIBLIOGRAPHY
Bradski, G. (2000). The OpenCV Library. Dr. Dobb’s Journal of Software Tools.