Automatic speech recognition for low-resource languages and dialects
Hoffer-Sohn, Yda
This item is only available for download by members of the University of Illinois community. Students, faculty, and staff at the U of I may log in with your NetID and password to view the item. If you are trying to access an Illinois-restricted dissertation or thesis, you can request a copy through your library's Inter-Library Loan office or purchase a copy directly from ProQuest.
Permalink
https://hdl.handle.net/2142/104017
Description
Title
Automatic speech recognition for low-resource languages and dialects
Author(s)
Hoffer-Sohn, Yda
Contributor(s)
Hasegawa-Johnson, Mark
Issue Date
2019-05
Keyword(s)
Automatic Speech Recognition
Dialect Variation and Adaptation
Abstract
Most widely spoken languages have numerous dialects or accents which can vary in degree of mutual intelligibility between speakers and automatic speech recognition systems alike. Oftentimes it is difficult or costly to build a specified dataset for training an automatic speech recognizer, and in some cases, it is not feasible to do so. In order to increase the usability of speech recognition software trained on available data, this thesis research explored methods of training ASR models on dialectal speech.
Common methods for training ASR models on low-resource and dialectal speech often rely on the idea of “there’s no data like more data”, and employ various methods for data augmentation, multi-lingual and cross-lingual training, or gleaning data from less obvious sources. In contrast to these common approaches, this thesis discusses a proposed phonetic mapping layer added to a high-resource single language speech recognition system to branch between a standard dialect and a regional dialect with a set of common alternate phoneme pronunciations.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.