Published April 1, 2023 | Version v1
Journal article Open

Real-time recognition of American sign language using longshort term memory neural network and hand detection

  • 1. Department of Software Engineering, Faculty of Computers Science, October University for Modern Science and Arts, Cairo, Egypt
  • 2. Faculty of Computers and Artificial Intelligence, Helwan University, Cairo, Egypt

Description

Sign language recognition is very important for deaf and mute people because it has many facilities for them, it converts hand gestures into text or speech. It also helps deaf and mute people to communicate and express mutual feelings. This paper's goal is to estimate sign language using action detection by predicting what action is being demonstrated at any given time without forcing the user to wear any external devices. We captured user signs with a webcam. For example; if we signed “thank you”, it will take the entire set of frames for that action to determine what sign is being demonstrated. The long short-term memory (LSTM) model is used to produce a real-time sign language detection and prediction flow. We also applied dropout layers for both training and testing dataset to handle overfitting in deep learning models which made a good improvement for the final result accuracy. We achieved a 99.35% accuracy after training and implementing the model which allows the deaf and mute communicate more easily with society.

Files

27716-61497-1-PB.pdf

Files (814.9 kB)

Name Size Download all
md5:3b703619839fb05796b313464c2ec180
814.9 kB Preview Download