Skip to content

Leveraging Intelligent Processing Tools and Algorithms to help the Visually Impaired see and navigate ๐Ÿ’ฅโœจ

License

Notifications You must be signed in to change notification settings

HarshCasper/HelpingHand

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

38 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

HelpingHand

Logo

All Contributors

forthebadge forthebadge forthebadge forthebadge

Introduction ๐Ÿ“Œ

Even with the rise of tools and technologies, mankind hasnโ€™t implemented applications that could help visually impaired people. With the rise of Data Modelling techniques that can be used to infuse โ€œintelligenceโ€ even in dumb computers and the ease of accessibility, this โ€œintelligenceโ€ can be extended to our Smartphone to help the visually impaired people cope up with their surroundings and get a helping hand in their daily activities. Our Application aims to bridge the gap between them and the visual world by leveraging the power of Deep Learning which can be made accessible even on low-ended devices with a lucid User-Interface that would exactly allow them to better understand the world around.

Technology Stack ๐Ÿ

Instructions and Information ๐Ÿ’ฅ

To know more about the Build Instructions for the App and to better understand the technicalities, check out our Wiki Page.

Application Preview ๐Ÿ‘€

image

๐Ÿƒโ€โ™‚๏ธ Why this Project?

Our primary purpose behind this project is to leverage and study how Deep Learning Architectures along with easy prototyping tools can help us develop applications that can be easily rendered even on low-end devices. With this Application, we will develop a one-stop solution to allow the Blind or Partially Blind People to better understand the surroundings around them and to be able to cope with the dynamic world ahead of them.

The Minimal Viable Product (MVP) would allow the Users to leverage Image Captioning Architecture to generate a real-time insight into their surroundings while using Natural Language Processing to speak out in a lucid manner. The cornerstone of the Application would be its User Interface which would infuse a lucid experience for the User with its ease of handling and use.

For this project, we will be collaborating on various domains like:

  • Data Modelling
  • RESTful API Development
  • Prototyping Mobile Application using Flutter
  • UI/UX Designing

This would be an enriching experience for all of us that are part of this team.

How our project is more accessible for visually impaired people

  • Every feature, from image labeling to currency detection uses a text-to-speech feature to speak out to the user whatever is detected ๐Ÿ—ฃ๏ธ
  • Each screen vibrates with different intensity on being opened helping to user navigate. The buttons also have unique vibrations for better accessibility ๐Ÿ“ณ
  • We have used a minimum number of buttons, but whichever buttons are there, are of a large size. For instance, the top half of the screen will be one button, and the bottom half, another button so that a user does not need to precisely click on a particular position ๐Ÿ”˜
  • All features, except the image captioning feature, work completely offline and do not require any internet connection๐Ÿ“ถ
  • All offline features work in real-time and do not need any pre-processing time for the models to make predictions so the user can get instant updates๐ŸŽ๏ธ

For contributors

  • To get a deeper understanding and build insturctions for the project please check out our project's Wiki Page.
  • Before contributing do go through the Code of Conduct and the Contributors Guidelines.
  • If you find any bug in the application, or a feature you think would be nice to have, please open an issue.
  • If you want to contribute code, please fork the project and submit a pull request.

LICENSE ๐Ÿ“œ

MIT

Contributors โœจ

Thanks goes to these wonderful people (emoji key):


Shambhavi Aggarwal

๐Ÿ’ป ๐Ÿค”

Yash Khare

๐Ÿ’ป ๐Ÿ“– ๐Ÿค”

Harsh Bardhan Mishra

๐Ÿ“– ๐Ÿ’ป

This project follows the all-contributors specification. Contributions of any kind welcome!

About

Leveraging Intelligent Processing Tools and Algorithms to help the Visually Impaired see and navigate ๐Ÿ’ฅโœจ

Topics

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •  
pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy