A Deep Learning Application made with python to detect and recognize Sign Language gestures using "ASL" Standard then translates it into words.
We made this Application as our graduation project also we hope it could be used as an extension for video conference applications so that deaf and dump people can interact online easily with others.
The App Consists of 3 main modules.
1- Deep Learning Model.
2- Hand Detection.
3- Auto Correction.
Note: you can run the command pip install -r requirments.txt
and all libraries will be downloaded directly
Used Libraries:
- MediaPipe
- Numpy
- Tensorflow
- OpenCv-python
- pyspellchecker
- sklearn
- matplotlib
- Camera (not less than 3 mega pixels)
- CPU (2.7 Ghz Multi-core Processor)
- Python version >= 3.8
- You Can run the Application by simply openning
hand_detector
module then press f5. - You can also open
hand_detector
using python IDLE then from run window chooserun module
- Model Trainning Process is Done seperately using Google Colab.
- To get the best efficiency from this application please make sure to run it in a bright place with plain background.