14
Views
1
CrossRef citations to date
0
Altmetric
Research Article

Code-switched end-to-end Marathi speech recognition for especially abled people

&

References

  • Delić, V., Sečujski, M., Bojanić, M., Knežević, D., Vujnović Sedlar, N., Mak, R.: Aids for the Disabled Based on Speech Technologies - Case Study for the Serbian Language. In: 11th Int. Conf. ETAI, pp. E2-1.1-4, Ohrid, Macedonia (2013).
  • N. Morgan and H. Bourlard, “Continuous speech recognition usingmultilayer perceptrons with hidden markov models,” inProceedings ofInternational Conference on Acoustics, Speech and Signal Processing(ICASSP), 1990, pp. 413–416.G.
  • H. Bourlard and C. J. Wellekens, “Links between markov models andmultilayer perceptrons,”IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 12, no. 12, pp. 1167–1178, 1990. doi: https://doi.org/10.1109/34.62605
  • N. Morgan and H. Bourlard, “Neural networks for statistical recognitionof continuous speech,” Proceedings of the IEEE, vol. 83, no. 5, pp.742–772, 1995. doi: https://doi.org/10.1109/5.381844
  • H. Sak, A. Senior, and F. Beaufays, “Long short-term memory recurrentneural network architectures for large scale acoustic modeling,” in Proceedings of Annual Conference of the International Speech Commu-nication Association (INTERSPEECH), 2014, pp. 338–342.
  • Y. M. Qian, M. X. Bi, T. Tan, and K. Yu, “Very deep convolutional neuralnetworks for noise robust speech recognition,” IEEE/ACM Transactionson Audio, Speech, and Language Processing, vol. 24, no. 12, pp. 2263–2276, 2016.
  • A. Graves, S. Ferńandez, F. Gomez, and J. Schmidhuber, “Connectionisttemporal classification: Labelling unsegmented sequence data withrecurrent neural networks,” in Proceedings of International Conferenceon Machine Learning (ICML), 2006, pp. 369–376.
  • Q. Liu, L. J. Wang, and Q. Huo, “A study on effects of implicitand explicit language model information for DBLSTM-CTC basedhandwriting recognition,” inProceedings of International Conference on Document Analysis and Recognition (ICDAR), 2015, pp. 461–465.
  • A. Graves and N. Jaitly, “Towards end-to-end speech recognition withrecurrent neural networks,” in Proceedings of International Conferenceon Machine Learning (ICML), 2014, pp. 1764–1772.
  • Gaikwad, S., Gawali, B., Yannawar, P. and Mehrotra, S. (2011): Feature extraction using fusion MFCC for continuous marathi speech recognition. In 2011 Annual IEEE India Conference(pp. 1-5). IEEE
  • Gaikwad, S., Gawali, B. and Mehrotra, S., (2013): November. Creation of Marathi speech corpus for automatic speech recognition. International Conference Oriental COCOSDA held jointly with 2013 Conference on Asian Spoken Language Research and valuation (O-COCOSDA/CASLRE) (pp. 1-5). IEEE
  • Kurzekar, P.K., Deshmukh, R.R., Waghmare, V.B. and Shrishrimal, P.P.,(2014): Continuous Speech Recognition System: A Review. Asian Journal of Computer Science and Information Technology, 4(6), pp.62-66.
  • Khetri, G.P., Padme, S.L., Jain, D.C., Fadewar, H.S., Sontakke, B.R. and Pawar, V.P. (2012): Automatic Speech Recognition for Marathi Isolated Words. International Journal of Application or Innovation in Engineering & Management (IJAIEM), 1(3)
  • Shrishirmal, P.P., Deshmukh, R.R., Waghmare, V.B., Borade, S., Janse, P.V. and Janvale, G.B. (2015): Development of Marathi Language Speech Database from Marathwada Region. IEEE, 978-1-4673-8279-3/15.
  • Shrishrimal, P.P., Deshmukh, R.R. and Waghmare, V.B., (2012): Indian language speech database: A review. International journal of Computer applications, 47(5), pp.17-21. doi: https://doi.org/10.5120/7184-9893
  • Shrishrimal, P.P., Deshmukh, R.R., Janwale, G.B. and Kulkarni, D.S. (2017): Marathi Digit Recognition System based on MFCC and LPC. reason, 2(67), pp.17-9.
  • Waghmare, S. D., Deshmukh R. R., Shrishrimal P.P., Waghmare V.B., Janwale G.B, “ Stuttered Isolated Spoken Marathi Speech Recognition by using MFCC and LPC. International Journal of Innovations in Engineering and Technology, Vol-8, Issue-3, ISSN:2319-1058
  • O. V. William Chan, Navdeep Jaitly, Quoc Le, (2016): Listen , Attend And Spell : A Neural Network For Large Vocabulary Conversational Speech Recognition. IEEE Int. Conf. Acoust. Speech, Signal Process. 2016, pp. 4960–4964, 2016.
  • N. Palia, S. Kant, and A. Dev, (2019): Performance evaluation of speaker recognition system, J. Discret. Math. Sci. Cryptogr., vol. 22, no. 2, pp. 203–218, 2019, doi: https://doi.org/10.1080/09720529.2019.1582868.
  • D. Zhao, (2017): Design of continuous recognition algorithm for online interactive english speech segment, J. Discret. Math. Sci. Cryptogr., vol. 20, no. 6–7, pp. 1513–1517, 2017, doi: https://doi.org/10.1080/09720529.2017.1392472.
  • K E M., Rao, K.S., Jayagopi, D.B., Ramasubramanian, V. (2018): Indian Languages ASR: A Multilingual Phone Recognition Framework with IPA Based Common Phone-set, Predicted Articulatory Features and Feature fusion. Proc. Interspeech 2018, 1016-1020, doi: https://doi.org/10.21437/Interspeech.2018-2529.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.