End-to-end automated speech recognition using a character based small scale transformer architecture

dc.contributor.authorLoubser, Alexander
dc.contributor.authorDe Villiers, Johan Pieter
dc.contributor.authorDe Freitas, Allan
dc.contributor.emaila.loubser@tuks.co.zaen_US
dc.date.accessioned2024-05-21T07:46:44Z
dc.date.available2024-05-21T07:46:44Z
dc.date.issued2024-10
dc.descriptionDATA AVAILABILITY: Data will be made available on request.en_US
dc.description.abstractThis study explores the feasibility of constructing a small-scale speech recognition system capable of competing with larger, modern automated speech recognition (ASR) systems in both performance and word error rate (WER). Our central hypothesis posits that a compact transformer-based ASR model can yield comparable results, specifically in terms of WER, to traditional ASR models while challenging contemporary ASR systems that boast significantly larger computational sizes. The aim is to extend ASR capabilities to under-resourced languages with limited corpora, catering to scenarios where practitioners face constraints in both data availability and computational resources. The model, comprising a compact convolutional neural network (CNN) and transformer architecture with 2.214 million parameters, challenges the conventional wisdom that large-scale transformer-based ASR systems are essential for achieving high accuracy. In comparison, contemporary ASR systems often deploy over 300 million parameters. Trained on a modest dataset of approximately 3000 h – significantly less than the 50,000 h used in larger systems – the proposed model leverages the Common Voice and LibriSpeech datasets. Evaluation on the LibriSpeech test-clean and test-other datasets produced character error rates (CERs) of 6.40% and 16.73% and WERs of 16.03% and 35.51% respectively. Comparisons with existing architectures showcase the efficiency of our model. A gated recurrent unit (GRU) architecture, albeit achieving lower error rates, incurred a computational cost 24 times larger than our proposed model. Large-scale transformer architectures, while achieving marginally lower WERs (2%–4% on LibriSpeech test-clean), require 200 times more parameters and 53,000 additional hours of training data. Modern large language models are used to improve the WERs, but require large computational resources. To further enhance performance, a small 4-g language model was integrated into our end-to-end ASR model, resulting in improved WERs. The overarching goal of this work is to provide a practical solution for practitioners dealing with limited datasets and computational resources, particularly in the context of under-resourced languages.en_US
dc.description.departmentElectrical, Electronic and Computer Engineeringen_US
dc.description.librarianhj2024en_US
dc.description.sdgSDG-09: Industry, innovation and infrastructureen_US
dc.description.sponsorshipThe MultiChoice Chair of Machine Learning.en_US
dc.description.urihttps://www.elsevier.com/locate/eswaen_US
dc.identifier.citationLoubser, A., De Villiers, P. & De Freitas, A. 2024, 'End-to-end automated speech recognition using a character based small scale transformer architecture', Expert Systems with Applications, vol. 252, part A, art. 124119, pp. 1-11, doi : 10.1016/j.eswa.2024.124119.en_US
dc.identifier.issn0957-4174 (print)
dc.identifier.issn1873-6793 (online)
dc.identifier.other10.1016/j.eswa.2024.124119
dc.identifier.urihttp://hdl.handle.net/2263/96106
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.rights© 2024 The Author(s). Published by Elsevier Ltd. This is an open access article under the CC BY license.en_US
dc.subjectAutomated speech recognition (ASR)en_US
dc.subjectSpeech recognitionen_US
dc.subjectTransformeren_US
dc.subjectEnd-to-enden_US
dc.subjectCharacter baseden_US
dc.subjectConnectionist temporal classificationen_US
dc.subjectConvolutional neural network (CNN)en_US
dc.subjectWord error rate (WER)en_US
dc.subjectCharacter error rate (CER)en_US
dc.subjectSDG-09: Industry, innovation and infrastructureen_US
dc.titleEnd-to-end automated speech recognition using a character based small scale transformer architectureen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Loubser_EndToEnd_2024_2024.pdf
Size:
645.93 KB
Format:
Adobe Portable Document Format
Description:
Article

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: