
Improving Pronunciation for Non-Native Speakers Using Neural Networks
Author(s) -
R Santhoshi
Publication year - 2021
Publication title -
international journal for research in applied science and engineering technology
Language(s) - English
Resource type - Journals
ISSN - 2321-9653
DOI - 10.22214/ijraset.2021.37234
Subject(s) - pronunciation , computer science , focus (optics) , speech recognition , natural language processing , artificial neural network , artificial intelligence , linguistics , philosophy , physics , optics
While learning a new language through the internet or applications, a lot of them focus on teaching words and sentences and do not concentrate on the pronouncing ability of the users. Even though many speakers are proficient in a language, their pronunciation might be influenced by their native language. For people who are interested in improving their pronunciation capabilities this proposed system was introduced. This system is primarily focused on improving the pronunciation of English words and sentences for non-native speakers i.e., for whom English is a second language. For a given audio clip, we scale the audio and extract the features, input the features to the model developed and the output of the model gives the phonemes that are spoken in the clip. Many models detect phonemes and various methods have been proposed but the main reason for choosing deep learning is that the learning and the features that we tend to oversee or overlook are picked up by the model provided the dataset is balanced and the model is built properly. The features to be considered varies for every speech processing project and through previous research work and through trial and error we choose the features that work best for us. Comparing the phonemes with the actual phonemes present, we can give the speaker which part of their speech they need to work on. Based on the phoneme, feedback is given on how to improve their pronunciation.