Journal article Open Access

Hand Gesture Recognition using Deep Learning

A. Geetha Devi; M. Aparna; N. Mounika; U. Pavan kalian; R. Meghna Nath

Sponsor(s)
Blue Eyes Intelligence Engineering & Sciences Publication(BEIESP)

Hearing impaired individuals use sign languages to communicate with others within the community. Because of the wide spread use of this language, hard-of-hearing individuals can easily understand it but it is not known by a lot of normal people. In this paper a hand gesture recognition system has been developed to overcome this problem, for those who don't recognize sign language to communicate simply with hard-of-hearing individuals. In this paper a computer vision-based system is designed to detect sign Language. Datasets used in this paper are binary images. These images are given to the convolution neural network (CNN). This model extracts the features of the image and classifies the images, and it recognises the gestures. The gestures used in this paper are of American Sign Language. In real time system the images are converted to binary images using Hue, Saturation, and Value (HSV) colour model. In this model 87.5% of data is used for training and 12.5% of data is used for testing and the accuracy obtained with this model is 97%. .

Files (445.3 kB)
Name Size
D6765049420.pdf
md5:be1c62fcbb441fb86f72600fde0771a4
445.3 kB Download
15
6
views
downloads
Views 15
Downloads 6
Data volume 2.7 MB
Unique views 15
Unique downloads 6

Share

Cite as