Speech Technology for Everyone: Automatic Speech Recognition for Non-Native English with Transfer Learning
To address the performance gap of English ASR models on L2 English speakers, we evaluate fine-tuning of pretrained wav2vec 2.0 models (Baevski et al., 2020; Xu et al., 2021) on L2-ARCTIC, a non-native English speech corpus (Zhao et al., 2018) under different training settings. We compare \textbf{(a)...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
15.10.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | To address the performance gap of English ASR models on L2 English speakers, we evaluate fine-tuning of pretrained wav2vec 2.0 models (Baevski et al., 2020; Xu et al., 2021) on L2-ARCTIC, a non-native English speech corpus (Zhao et al., 2018) under different training settings. We compare \textbf{(a)} models trained with a combination of diverse accents to ones trained with only specific accents and \textbf{(b)} results from different single-accent models. Our experiments demonstrate the promise of developing ASR models for non-native English speakers, even with small amounts of L2 training data and even without a language model. Our models also excel in the zero-shot setting where we train on multiple L2 datasets and test on a blind L2 test set. |
---|---|
AbstractList | To address the performance gap of English ASR models on L2 English speakers, we evaluate fine-tuning of pretrained wav2vec 2.0 models (Baevski et al., 2020; Xu et al., 2021) on L2-ARCTIC, a non-native English speech corpus (Zhao et al., 2018) under different training settings. We compare \textbf{(a)} models trained with a combination of diverse accents to ones trained with only specific accents and \textbf{(b)} results from different single-accent models. Our experiments demonstrate the promise of developing ASR models for non-native English speakers, even with small amounts of L2 training data and even without a language model. Our models also excel in the zero-shot setting where we train on multiple L2 datasets and test on a blind L2 test set. |
Author | Abdul-Mageed, Muhammad Zhang, Xinyi Sullivan, Peter Cho, Haejin Shibano, Toshiko Mia Taige Li |
Author_xml | – sequence: 1 givenname: Toshiko surname: Shibano fullname: Shibano, Toshiko – sequence: 2 givenname: Xinyi surname: Zhang fullname: Zhang, Xinyi – sequence: 3 fullname: Mia Taige Li – sequence: 4 givenname: Haejin surname: Cho fullname: Cho, Haejin – sequence: 5 givenname: Peter surname: Sullivan fullname: Sullivan, Peter – sequence: 6 givenname: Muhammad surname: Abdul-Mageed fullname: Abdul-Mageed, Muhammad |
BookMark | eNqNzE0KwjAQBeAgCv71DgHXhZgaa92JVFyIC3UvpUzTlDqjSap4e4N4ADfzFu97M2Z9JIQeG8kkmcerhZRDFjnXCCHkMpVKJSPWnO8AZc0v4SC1pN-8IsvzJ9h3GK_5pvN0K7wp-U-eoCSNxhvCLz0SxscAnsBz1K1xNX8ZHz7aAl0Flh-gsGhQT9mgKloH0S8nbLbLL9t9fLf06MD5a0OdxVBdpUozOVeZEsl_6gN9CUpK |
ContentType | Paper |
Copyright | 2021. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2021. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central UK/Ireland ProQuest Central Essentials ProQuest Central Technology Collection ProQuest One Community College ProQuest Central SciTech Premium Collection ProQuest Engineering Collection Engineering Database Publicly Available Content Database ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_25792159503 |
IEDL.DBID | 8FG |
IngestDate | Tue Sep 24 18:58:09 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_25792159503 |
OpenAccessLink | https://www.proquest.com/docview/2579215950/abstract/?pq-origsite=%requestingapplication% |
PQID | 2579215950 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_2579215950 |
PublicationCentury | 2000 |
PublicationDate | 20211015 |
PublicationDateYYYYMMDD | 2021-10-15 |
PublicationDate_xml | – month: 10 year: 2021 text: 20211015 day: 15 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2021 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.348693 |
SecondaryResourceType | preprint |
Snippet | To address the performance gap of English ASR models on L2 English speakers, we evaluate fine-tuning of pretrained wav2vec 2.0 models (Baevski et al., 2020; Xu... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Automatic speech recognition Training |
Title | Speech Technology for Everyone: Automatic Speech Recognition for Non-Native English with Transfer Learning |
URI | https://www.proquest.com/docview/2579215950/abstract/ |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1LSwMxEB5qF8GbT3zUEtBrWDf79iIqWxehS6kKvZUkO1U87G6324MXf7tJ3LUHoceQkBfDzPDlmy8A14gYuotcUCEDj3qO4FTlzYzmAZeSRdz_FdMZZ0H65j3P_FkP0q4WRtMqO59oHHVeSo2R28q0YhWeYv_G5kKjALKx76ol1f9H6XfW9jONHbAcrYmna8ZHT39oCwtClTu7_xyuiSKjfbAmvML6AHpYHMKuIV_K1RF8vlSI8oNsYG6iUkmSKCv7Kgu8JffrpjTSqqQdOe1oP2VhhmZlQTMj4U3aulyiAVZiItECa9LKqL4fw9UoeX1MabfBeWtMq_nm6O4J9Au18CmQmEUiFE4kXcm8KHC4yvsxj5ngzOV5JM5gsG2m8-3dF7DHNHlDUzf8AfSbeo2XKvo2YmgudgjWQ5JNpqo1_k5-AOhUkf0 |
link.rule.ids | 786,790,12792,21416,33408,33779,43635,43840 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1LS8QwEB50F9GbT3ysGtBr0CZ9ehGRrVV3i-gKeytJml3x0Na2e_Dfm8TUPQh7zpAXw3zDly8zAJdSyoDOco658F3sOpxhlTcTnPtMCBIy77eYzjj1k3f3aepNLeHWWFllFxNNoM5LoTnyK-VakYKnyLu-rb6w7hqlX1dtC4116LtUQaf-KR4__HEsxA9Uxkz_hVmDHfE29F9YJesdWJPFLmwYyaVo9uDzrZJSfKAluY1UAomGyre-y0LeoLtFW5qCqshavnZin7IwpmlZ4NQU7kb2Ny7StCoy-DOTNbLFU-f7cBEPJ_cJ7jaYWRdqsuWB6QH0CrXwIaCIhDzgTiioIG7oO0xl-zKPCGeEsjzkRzBYNdPx6uFz2Ewm41E2ekyfT2CLaPmGFm94A-i19UKeKvxt-Zm55B_7u45g |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Speech+Technology+for+Everyone%3A+Automatic+Speech+Recognition+for+Non-Native+English+with+Transfer+Learning&rft.jtitle=arXiv.org&rft.au=Shibano%2C+Toshiko&rft.au=Zhang%2C+Xinyi&rft.au=Mia+Taige+Li&rft.au=Cho%2C+Haejin&rft.date=2021-10-15&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |