This project implements a deep learning model to detect and classify sign language gestures from various input sources, including images, videos, and live webcam streams. The model is deployed in an interactive Streamlit web application, providing users with an intuitive interface for real-time gesture recognition.
The application utilizes a pre-trained YOLO-NAS model tailored for sign language detection. The model was trained using a custom dataset following the approach outlined in this project by Muhammad Moin Faisal. Building on top of this foundation, the model has been integrated into a comprehensive web platform that supports multiple input types for gesture detection.
- Image Upload: Users can upload an image, and the system will detect sign language gestures within the image.
- Video Upload: The model processes uploaded videos to identify gestures throughout the duration of the video.
- Real-time Webcam Detection: Users can activate their webcam to detect and classify gestures in real-time.
- Homepage: The app includes a user-friendly homepage to guide users through the available functionalities.
- Python 3.7 - 3.9
- Streamlit
- SuperGradients
- OpenCV
- PyTorch
- Additional dependencies can be installed from the
requirements.txt
file.
-
Clone the repository:
git clone https://github.com/darshneesunderraj/Sign-Language_Detection.git cd Sign-Language_Detection
-
Install the required dependencies:
pip install -r requirements.txt
-
Download the trained YOLO-NAS model and place it in the
model_weights
folder:- Create a directory named
model_weights
in the project root if it doesn't already exist:mkdir model_weights
- Download the pre-trained model weights file and save it in the
model_weights
directory asckpt_best.pth
:wget <url_to_your_trained_model> -O model_weights/ckpt_best.pth
- Alternatively, manually download the file and move it into the
model_weights
folder.
- Create a directory named
-
Run the Streamlit application:
streamlit run main.py
The sign language detection model used in this project was originally trained using the approach outlined by Muhammad Moin Faisal. The model architecture is based on YOLO-NAS and has been fine-tuned to detect 26 unique sign language gestures. This project extends the model's functionality by deploying it in a web-based interface for broader accessibility and interaction.
This project is inspired by the work on YOLO-NAS Sign Language Detection by Muhammad Moin Faisal, with significant adaptations to integrate the model into a fully functional Streamlit web application, allowing for image, video, and webcam-based gesture recognition.
This project is licensed under the MIT License.