A Comparison of LSTM and GRU for Bengali Speech-to-Text Transformation

Publisher:
Springer Nature
Publication Type:
Conference Proceeding
Citation:
Lecture Notes in Networks and Systems, 2023, 700 LNNS, pp. 214-224
Issue Date:
2023-01-01
Filename Description Size
978-3-031-33743-7_18.pdfPublished version387.9 kB
Adobe PDF
Full metadata record
This paper represents an approach to speech-to-text conversion in the Bengali language. In this area, we have found most of the methodologies were focused on other languages rather than Bengali. We started with a novel dataset of 56 unique words from 160 individual subjects was prepared. Then in this paper, we illustrate the approach to increasing accuracy in a speech-to-text over the Bengali language where initially we started with Gated Recurrent Unit(GRU) and Long short-term memory (LSTM) algorithms. During further observation, we found that the output of the GRU failed to give any stable output. So, we moved completely to the LSTM algorithm where we achieved 90% accuracy on an unexplored dataset. Voices of several demographic populations and noises were used to validate the model. In the testing phase, we tried a variety of classes based on their length, complexity, noise, and gender variant. Moreover, we expect that this research will help to develop a real-time Bengali speak-to-text recognition model.
Please use this identifier to cite or link to this item: