Language Scaling for Universal Suggested Replies Model
We consider the problem of scaling automated suggested replies for Outlook email system to multiple languages. Faced with increased compute requirements and low resources for language expansion, we build a single universal model for improving the quality and reducing run-time costs of our production...
Saved in:
Main Authors | , , , , , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
03.06.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | We consider the problem of scaling automated suggested replies for Outlook
email system to multiple languages. Faced with increased compute requirements
and low resources for language expansion, we build a single universal model for
improving the quality and reducing run-time costs of our production system.
However, restricted data movement across regional centers prevents joint
training across languages. To this end, we propose a multi-task continual
learning framework, with auxiliary tasks and language adapters to learn
universal language representation across regions. The experimental results show
positive cross-lingual transfer across languages while reducing catastrophic
forgetting across regions. Our online results on real user traffic show
significant gains in CTR and characters saved, as well as 65% training cost
reduction compared with per-language models. As a consequence, we have scaled
the feature in multiple languages including low-resource markets. |
---|---|
AbstractList | We consider the problem of scaling automated suggested replies for Outlook
email system to multiple languages. Faced with increased compute requirements
and low resources for language expansion, we build a single universal model for
improving the quality and reducing run-time costs of our production system.
However, restricted data movement across regional centers prevents joint
training across languages. To this end, we propose a multi-task continual
learning framework, with auxiliary tasks and language adapters to learn
universal language representation across regions. The experimental results show
positive cross-lingual transfer across languages while reducing catastrophic
forgetting across regions. Our online results on real user traffic show
significant gains in CTR and characters saved, as well as 65% training cost
reduction compared with per-language models. As a consequence, we have scaled
the feature in multiple languages including low-resource markets. |
Author | Bajaj, Payal Ying, Qianlan Lin, Bojia Wang, Wei Jiang, Daxin Song, Xia Yang, Yu Shokouhi, Milad Yang, Yang Deb, Budhaditya |
Author_xml | – sequence: 1 givenname: Qianlan surname: Ying fullname: Ying, Qianlan – sequence: 2 givenname: Payal surname: Bajaj fullname: Bajaj, Payal – sequence: 3 givenname: Budhaditya surname: Deb fullname: Deb, Budhaditya – sequence: 4 givenname: Yu surname: Yang fullname: Yang, Yu – sequence: 5 givenname: Wei surname: Wang fullname: Wang, Wei – sequence: 6 givenname: Bojia surname: Lin fullname: Lin, Bojia – sequence: 7 givenname: Milad surname: Shokouhi fullname: Shokouhi, Milad – sequence: 8 givenname: Xia surname: Song fullname: Song, Xia – sequence: 9 givenname: Yang surname: Yang fullname: Yang, Yang – sequence: 10 givenname: Daxin surname: Jiang fullname: Jiang, Daxin |
BackLink | https://doi.org/10.48550/arXiv.2106.02232$$DView paper in arXiv |
BookMark | eNotj8tKAzEYRrPQhdY-gCvzAjMmfya3pRRvMCLYdj1kmj8hEDMlY4u-vVpdHfgWh-9ckrMyFSTkmrO2M1KyW1c_07EFzlTLAARcENW7Eg8uIl3vXE4l0jBVui3piHV2ma4PMeL8gZ6-4T4nnOnL5DFfkfPg8ozLfy7I5uF-s3pq-tfH59Vd3ziloZGdGYP0EpThodsZzgxj1unAjfkZPWptRw2dUGCRBx00l8GPwmoEb00QC3Lzpz0dH_Y1vbv6NfwGDKcA8Q0Rd0C1 |
ContentType | Journal Article |
Copyright | http://creativecommons.org/licenses/by/4.0 |
Copyright_xml | – notice: http://creativecommons.org/licenses/by/4.0 |
DBID | AKY GOX |
DOI | 10.48550/arxiv.2106.02232 |
DatabaseName | arXiv Computer Science arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2106_02232 |
GroupedDBID | AKY GOX |
ID | FETCH-LOGICAL-a672-548bf5d52681f4c8108009a7f188526de779b7243629e1f7f715fdb397e2d98f3 |
IEDL.DBID | GOX |
IngestDate | Mon Jan 08 05:41:52 EST 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a672-548bf5d52681f4c8108009a7f188526de779b7243629e1f7f715fdb397e2d98f3 |
OpenAccessLink | https://arxiv.org/abs/2106.02232 |
ParticipantIDs | arxiv_primary_2106_02232 |
PublicationCentury | 2000 |
PublicationDate | 2021-06-03 |
PublicationDateYYYYMMDD | 2021-06-03 |
PublicationDate_xml | – month: 06 year: 2021 text: 2021-06-03 day: 03 |
PublicationDecade | 2020 |
PublicationYear | 2021 |
Score | 1.8059456 |
SecondaryResourceType | preprint |
Snippet | We consider the problem of scaling automated suggested replies for Outlook
email system to multiple languages. Faced with increased compute requirements
and... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Artificial Intelligence Computer Science - Computation and Language |
Title | Language Scaling for Universal Suggested Replies Model |
URI | https://arxiv.org/abs/2106.02232 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV07SwQxEB7Oq2xEUTmfpLCN3maTTbYU8TzER3EnbLfkNYcgIvcQf76T7Io2tsk0M4F830y-zABcSByXKKziNkTFpVKKO4uRhyQKsy4Yg6mg__hUTV_kfaOaAbCfvzB2-fX62fUHdqsrykeqS0KZki7ZLSGSZOvuuekeJ3Mrrt7-1444Zl76AxKTXdjp2R277o5jDwbxfR-qh74myGYUEIIKRkSR9YoIsp5tFotcdWTEhokSrlgaUPZ2APPJ7fxmyvtxBdxWOg0UMA5VSO1TCpTeJPHeuLYaC2NoMUSta6eFJMSoY4EadaEwOOIDUYTaYHkIQ8r44wiYVcJb45UvainRodXpo4_wpY-UXaE-glF2sv3oOlK0yf82-3_8_9YJbIskyEglhPIUhuvlJp4Roq7deQ7rNwJedbw |
link.rule.ids | 228,230,783,888 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Language+Scaling+for+Universal+Suggested+Replies+Model&rft.au=Ying%2C+Qianlan&rft.au=Bajaj%2C+Payal&rft.au=Deb%2C+Budhaditya&rft.au=Yang%2C+Yu&rft.date=2021-06-03&rft_id=info:doi/10.48550%2Farxiv.2106.02232&rft.externalDocID=2106_02232 |