Computer vision-based Bengali sign language to text generation
MetadataShow full item record
In the whole world, around 7% of people have hearing and speech impairment problems. They use sign language as their communication method. People from various countries use a variety of sign languages. As an example, there are ASL, CSL, JSL, etc. Even in our country, there are lots of people born with hearing and speech impairment problems. So, our primary focus is to work for those people by converting Bangla sign language into text. There are already various projects on Bangla sign language done by other people. However, they focused more on the separate alphabets and numerical numbers. That is why we want to concentrate on Bangla word signs since these people prefer to communicate using words or phrases rather than alphabets. There is not any proper database for Bangla word sign language, so we are making a database for Bangla word sign language for our work. In recognition of sign language (SLR), there usually are two types of scenarios: isolated SLR, which takes words by word and completes recognize action, and the other one is continuous SLR, which completes action by translating the whole sentence at once. We are working on isolated SLR. We introduce a method where we are going to use PyTorch and YOLOv5 for a video classification model to convert Bangla sign language into the text from the video where each video has only one sign language word. Here,we have achieved an accuracy rate of 76.29% on the training dataset and 51.44% on the testing dataset. We are working to build a system that will make it easier for hearing and speech-disabled people to interact with the general public.