Skip to content

PoojaK97/IndianSignLanguage_To_Speech

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Indian-Sign-Language-Recognition-System

Indian Sign Language (ISL) is one of the most used sign languages in the Indian subcontinent. This project aims at developing a simple Indian sign language recognition system based on convolutional neural network (CNN). The proposed system needs webcam and laptop. CNN is used for image classification. Scale invariant feature transformation (SIFT) is hybridized with adaptive thresholding and Gaussian blur image smoothing for feature extraction. Due to unavailability of ISL dataset, a dataset of 5000 images, 100 images each for 50 gestures, has been created. The system is implemented and tested using python-based library Keras. The proposed CNN with hybrid SIFT implementation achieves 92.78% accuracy, whereas the accuracy of 91.84% was achieved for CNN with adaptive thresholding.

Indian Sign Language Dataset

Due to unavailability of Indian Sign language dataset (ISL), We have created dataset of 5000 images. Data set contains 100 images each for 50 Indian signs.

Hardware Requirements

• 4 GB Ram
• 1 GB Free Space
• Web Cam (5 MP preferable)

Software Installation Requirements

• Python 3.6.10
• OPENCV 3.3.1
• Keras 2.3.1
• Theano 1.0.4

How to Use System?

1. Creation of the Data-set:

Step 1: Open the terminal and move to the project folder.
Step 2: Run the command 'set KERAS_BACKEND=theano' followed by 'python cnnCreateDataSet.py'.
Step 3: Press key n to capture image for the given sign.
Step 4: When system says “Change gesture” then change the gesture.

2. Training of Model

Step 1: Open the terminal and move to the project folder.
Step 2: Run the command 'set KERAS_BACKEND=theano' followed by 'python cnnTrain.py'.
Step 3: As per menu shown on terminal select the data filter on which you want to train the CNN model.
Step 4: Once the training is completed, a message "Model trained successfully" will be shown on the terminal.

3. Prediction of Sign

Step 1: Open the terminal and move to the project folder.
Step 2: Run the command 'set KERAS_BACKEND=theano' followed by 'python main.py'.
Step 3: Follow the menu shown on the terminal to adjust the region of gesture (Green box).
Step 4: Once the Region of Gesture is fixed press key 'P' to start prediction.
Step 5: System will show predicted sign on display. For audio format, enable sound drivers.

About

Mini Project

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •