Skip to content

WeSpeak is primarily an Indian Sign Language interpretation tool that uses machine learning and image processing to translate sign language in real time.

Notifications You must be signed in to change notification settings

roshrizvi/WeSpeak

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

WeSpeak

WeSpeak is a project that aims to bridge the current communication gap faced by speech-impaired people by interpreting Indian Sign Language (ISL) in real-time using machine learning and image processing.

This project was the winner of the PCCE "IDEAS 2024" competition, where it secured first place from 30+ other projects. It is one of the few projects to be selected by CIBA (Centre for Incubation & Business Acceleration) to be taken to the next level and is now being further developed and expanded upon in CIBA.

Key Features

  • Indian Sign Language Interpretation For Letters: WeSpeak interprets individual letters in the Indian Sign Language using machine learning and image processing so that the user can spell out a word letter-by-letter.
  • Sign Language Interpretation For Words: WeSpeak offers real-time interpretation for words gestured out in sign language, bridging the communication gap that currently exists and aiding communication for the speech-impaired.
  • Sign Language Learning: Users can learn Indian Sign Language through interactive modules and exercises provided by WeSpeak.

Additional Features

  • Hands-Free Presentation Control: Utilizing image processing technology, WeSpeak offers a hands-free presentation control feature that can be utilized in schools, colleges, meeting rooms, and other institutions.
  • Volume Control: Using the same hand tracking technology used for the sign language interpretation, we can control many other features without the need for a physical interface. This is one of the examples as a proof-of-concept.
  • Snake Game with Gesture Control: To show how quick and responsive our hand tracking technology is, WeSpeak includes a fun snake game that uses hands-free image processing technology to control the snake on the screen, making it an interactive and enjoyable experience.

Team Members

  • Leander Fernandes: Primary back-end developer and Tech Lead. He made critical architectural decisions and offered insightful suggestions that shaped the direction of the project. His leadership and technical prowess were instrumental in driving the project forward.

  • Roshan Rizvi: Primary front-end developer and back-end infrastructure and integration architect. His focus on creating an intuitive frontend interface and his contributions to creating the backend infrastructure and integrating all the components together were essential to the project's success.

  • Maithily Naik: Maithily provided valuable contributions to the WeSpeak project, offering support in coding tasks, training the models during the development stages, and assisting in various aspects throughout the development of the project. Her contributions are irreplaceable!

About

WeSpeak is primarily an Indian Sign Language interpretation tool that uses machine learning and image processing to translate sign language in real time.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages