ELF: Encoding Speaker-Specific Latent Speech Feature for Speech Synthesis
In this work, we propose a novel method for modeling numerous speakers, which enables expressing the overall characteristics of speakers in detail like a trained multi-speaker model without additional training on the target speaker's dataset. Although various works with similar purposes have be...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , , , , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
31.05.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | In this work, we propose a novel method for modeling numerous speakers, which enables expressing the overall characteristics of speakers in detail like a trained multi-speaker model without additional training on the target speaker's dataset. Although various works with similar purposes have been actively studied, their performance has not yet reached that of trained multi-speaker models due to their fundamental limitations. To overcome previous limitations, we propose effective methods for feature learning and representing target speakers' speech characteristics by discretizing the features and conditioning them to a speech synthesis model. Our method obtained a significantly higher similarity mean opinion score (SMOS) in subjective similarity evaluation than seen speakers of a high-performance multi-speaker model, even with unseen speakers. The proposed method also outperforms a zero-shot method by significant margins. Furthermore, our method shows remarkable performance in generating new artificial speakers. In addition, we demonstrate that the encoded latent features are sufficiently informative to reconstruct an original speaker's speech completely. It implies that our method can be used as a general methodology to encode and reconstruct speakers' characteristics in various tasks. |
---|---|
AbstractList | In this work, we propose a novel method for modeling numerous speakers, which enables expressing the overall characteristics of speakers in detail like a trained multi-speaker model without additional training on the target speaker's dataset. Although various works with similar purposes have been actively studied, their performance has not yet reached that of trained multi-speaker models due to their fundamental limitations. To overcome previous limitations, we propose effective methods for feature learning and representing target speakers' speech characteristics by discretizing the features and conditioning them to a speech synthesis model. Our method obtained a significantly higher similarity mean opinion score (SMOS) in subjective similarity evaluation than seen speakers of a high-performance multi-speaker model, even with unseen speakers. The proposed method also outperforms a zero-shot method by significant margins. Furthermore, our method shows remarkable performance in generating new artificial speakers. In addition, we demonstrate that the encoded latent features are sufficiently informative to reconstruct an original speaker's speech completely. It implies that our method can be used as a general methodology to encode and reconstruct speakers' characteristics in various tasks. |
Author | Kong, Dohee Kim, Beomjeong Kim, Sangjin Park, Jihoon Kong, Jungil Kim, Jeongmin Lee, Changheon Lee, Junmo |
Author_xml | – sequence: 1 givenname: Jungil surname: Kong fullname: Kong, Jungil – sequence: 2 givenname: Junmo surname: Lee fullname: Lee, Junmo – sequence: 3 givenname: Jeongmin surname: Kim fullname: Kim, Jeongmin – sequence: 4 givenname: Beomjeong surname: Kim fullname: Kim, Beomjeong – sequence: 5 givenname: Jihoon surname: Park fullname: Park, Jihoon – sequence: 6 givenname: Dohee surname: Kong fullname: Kong, Dohee – sequence: 7 givenname: Changheon surname: Lee fullname: Lee, Changheon – sequence: 8 givenname: Sangjin surname: Kim fullname: Kim, Sangjin |
BookMark | eNqNikELgjAYQEcUZOV_GHQW1jc17RpKgTe7y1ifOYvNtnno32dQ904P3nsrMtdG44wEwPkuymKAJQmd6xljkO4hSXhAzkVVHmihpbkqfaP1gOKONpooVaskrYRH7T8eZUdLFH60SFtjf6p-ad-hU25DFq14OAy_XJNtWVyOp2iw5jmi801vRqun1ECWxyzLU0j5f9cbStA9Qw |
ContentType | Paper |
Copyright | 2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central ProQuest Central Essentials ProQuest Central Technology Collection ProQuest One Community College ProQuest Central SciTech Premium Collection ProQuest Engineering Collection Engineering Database Publicly Available Content Database ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_28940896263 |
IEDL.DBID | BENPR |
IngestDate | Thu Oct 10 17:54:02 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_28940896263 |
OpenAccessLink | https://www.proquest.com/docview/2894089626?pq-origsite=%requestingapplication% |
PQID | 2894089626 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_2894089626 |
PublicationCentury | 2000 |
PublicationDate | 20240531 |
PublicationDateYYYYMMDD | 2024-05-31 |
PublicationDate_xml | – month: 05 year: 2024 text: 20240531 day: 31 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2024 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.5375953 |
SecondaryResourceType | preprint |
Snippet | In this work, we propose a novel method for modeling numerous speakers, which enables expressing the overall characteristics of speakers in detail like a... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Similarity Speech recognition |
Title | ELF: Encoding Speaker-Specific Latent Speech Feature for Speech Synthesis |
URI | https://www.proquest.com/docview/2894089626 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfZ3fS8MwEMcP1yL45k_8MUdAX4MxTX9sL4LSOmWO4RT2NtI0xTHoats9-OLfvktt9UHYYy6QkBDu873LwQFc37I4CFw_oUyknIpUSBp4rE8TB-nkC2moY6otxt7wXTzP3FmTcCubssrWJ9aOOlkpkyO_wcBAsKCP-vsu_6Sma5T5XW1aaHTA5hgpMAvs-3A8ef3NsnDPR83s_HO0NT2ifbAnMtfFAezo7BB266JLVR7BUziKBiTM1MoAhExzLZe6oHVL-HShyAh1YFYZu1YfxIi1daEJqszWNP3KUL-Vi_IYrqLw7WFI2_3nzRsp538nck7AwmBfnwLhQinBWZKmyHlX-VKK2HMSZE3sCVfqM-huW-l8-_QF7HGE8s_vdxesqljrS4RqFfegE0SPveb-cPTyHW4A49qAlg |
link.rule.ids | 783,787,12777,21400,33385,33756,43612,43817 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfZ3PS8MwFMcfuiHu5k_8MTWg12BtX3_Miwdp7bQOYRN2K2ma4hC62nYH__u91FYPwq4JJCSE9_m-H-EB3NwZiefZbsoNzEyOGQruOcaIpxbRyUWhqaOrLSZO-I7Pc3veBtyqtqyys4mNoU6XUsfIb8kxQMMbkf5-KL647hqls6ttC41t6KNFrNY_xYOn3xiL6bikmK1_ZrZhR7AH_TdRqHIftlR-ADtNyaWsDmHsR8E983O51Phg00KJT1XypiF8tpAsIhWY13pcyQ-mpdqqVIw0Zjc0_c5JvVWL6giuA3_2GPJu_7h9IVX8dx7rGHrk6qsTYCZKiaaRZhlR3pauEJg4VkqkSRy0hTqF4aaVzjZPX8FuOHuN4mg8eTmHgUl4_smDD6FXlyt1QXitk8vmDtfXNYAK |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=ELF%3A+Encoding+Speaker-Specific+Latent+Speech+Feature+for+Speech+Synthesis&rft.jtitle=arXiv.org&rft.au=Kong%2C+Jungil&rft.au=Lee%2C+Junmo&rft.au=Kim%2C+Jeongmin&rft.au=Kim%2C+Beomjeong&rft.date=2024-05-31&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |