Language Scaling for Universal Suggested Replies Model

We consider the problem of scaling automated suggested replies for Outlook email system to multiple languages. Faced with increased compute requirements and low resources for language expansion, we build a single universal model for improving the quality and reducing run-time costs of our production...

Full description

Saved in:
Bibliographic Details
Main Authors Ying, Qianlan, Bajaj, Payal, Deb, Budhaditya, Yang, Yu, Wang, Wei, Lin, Bojia, Shokouhi, Milad, Song, Xia, Yang, Yang, Jiang, Daxin
Format Journal Article
LanguageEnglish
Published 03.06.2021
Subjects
Online AccessGet full text

Cover

Loading…
Abstract We consider the problem of scaling automated suggested replies for Outlook email system to multiple languages. Faced with increased compute requirements and low resources for language expansion, we build a single universal model for improving the quality and reducing run-time costs of our production system. However, restricted data movement across regional centers prevents joint training across languages. To this end, we propose a multi-task continual learning framework, with auxiliary tasks and language adapters to learn universal language representation across regions. The experimental results show positive cross-lingual transfer across languages while reducing catastrophic forgetting across regions. Our online results on real user traffic show significant gains in CTR and characters saved, as well as 65% training cost reduction compared with per-language models. As a consequence, we have scaled the feature in multiple languages including low-resource markets.
AbstractList We consider the problem of scaling automated suggested replies for Outlook email system to multiple languages. Faced with increased compute requirements and low resources for language expansion, we build a single universal model for improving the quality and reducing run-time costs of our production system. However, restricted data movement across regional centers prevents joint training across languages. To this end, we propose a multi-task continual learning framework, with auxiliary tasks and language adapters to learn universal language representation across regions. The experimental results show positive cross-lingual transfer across languages while reducing catastrophic forgetting across regions. Our online results on real user traffic show significant gains in CTR and characters saved, as well as 65% training cost reduction compared with per-language models. As a consequence, we have scaled the feature in multiple languages including low-resource markets.
Author Bajaj, Payal
Ying, Qianlan
Lin, Bojia
Wang, Wei
Jiang, Daxin
Song, Xia
Yang, Yu
Shokouhi, Milad
Yang, Yang
Deb, Budhaditya
Author_xml – sequence: 1
  givenname: Qianlan
  surname: Ying
  fullname: Ying, Qianlan
– sequence: 2
  givenname: Payal
  surname: Bajaj
  fullname: Bajaj, Payal
– sequence: 3
  givenname: Budhaditya
  surname: Deb
  fullname: Deb, Budhaditya
– sequence: 4
  givenname: Yu
  surname: Yang
  fullname: Yang, Yu
– sequence: 5
  givenname: Wei
  surname: Wang
  fullname: Wang, Wei
– sequence: 6
  givenname: Bojia
  surname: Lin
  fullname: Lin, Bojia
– sequence: 7
  givenname: Milad
  surname: Shokouhi
  fullname: Shokouhi, Milad
– sequence: 8
  givenname: Xia
  surname: Song
  fullname: Song, Xia
– sequence: 9
  givenname: Yang
  surname: Yang
  fullname: Yang, Yang
– sequence: 10
  givenname: Daxin
  surname: Jiang
  fullname: Jiang, Daxin
BackLink https://doi.org/10.48550/arXiv.2106.02232$$DView paper in arXiv
BookMark eNotj8tKAzEYRrPQhdY-gCvzAjMmfya3pRRvMCLYdj1kmj8hEDMlY4u-vVpdHfgWh-9ckrMyFSTkmrO2M1KyW1c_07EFzlTLAARcENW7Eg8uIl3vXE4l0jBVui3piHV2ma4PMeL8gZ6-4T4nnOnL5DFfkfPg8ozLfy7I5uF-s3pq-tfH59Vd3ziloZGdGYP0EpThodsZzgxj1unAjfkZPWptRw2dUGCRBx00l8GPwmoEb00QC3Lzpz0dH_Y1vbv6NfwGDKcA8Q0Rd0C1
ContentType Journal Article
Copyright http://creativecommons.org/licenses/by/4.0
Copyright_xml – notice: http://creativecommons.org/licenses/by/4.0
DBID AKY
GOX
DOI 10.48550/arxiv.2106.02232
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2106_02232
GroupedDBID AKY
GOX
ID FETCH-LOGICAL-a672-548bf5d52681f4c8108009a7f188526de779b7243629e1f7f715fdb397e2d98f3
IEDL.DBID GOX
IngestDate Mon Jan 08 05:41:52 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a672-548bf5d52681f4c8108009a7f188526de779b7243629e1f7f715fdb397e2d98f3
OpenAccessLink https://arxiv.org/abs/2106.02232
ParticipantIDs arxiv_primary_2106_02232
PublicationCentury 2000
PublicationDate 2021-06-03
PublicationDateYYYYMMDD 2021-06-03
PublicationDate_xml – month: 06
  year: 2021
  text: 2021-06-03
  day: 03
PublicationDecade 2020
PublicationYear 2021
Score 1.8059456
SecondaryResourceType preprint
Snippet We consider the problem of scaling automated suggested replies for Outlook email system to multiple languages. Faced with increased compute requirements and...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Artificial Intelligence
Computer Science - Computation and Language
Title Language Scaling for Universal Suggested Replies Model
URI https://arxiv.org/abs/2106.02232
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV07SwQxEB7Oq2xEUTmfpLCN3maTTbYU8TzER3EnbLfkNYcgIvcQf76T7Io2tsk0M4F830y-zABcSByXKKziNkTFpVKKO4uRhyQKsy4Yg6mg__hUTV_kfaOaAbCfvzB2-fX62fUHdqsrykeqS0KZki7ZLSGSZOvuuekeJ3Mrrt7-1444Zl76AxKTXdjp2R277o5jDwbxfR-qh74myGYUEIIKRkSR9YoIsp5tFotcdWTEhokSrlgaUPZ2APPJ7fxmyvtxBdxWOg0UMA5VSO1TCpTeJPHeuLYaC2NoMUSta6eFJMSoY4EadaEwOOIDUYTaYHkIQ8r44wiYVcJb45UvainRodXpo4_wpY-UXaE-glF2sv3oOlK0yf82-3_8_9YJbIskyEglhPIUhuvlJp4Roq7deQ7rNwJedbw
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Language+Scaling+for+Universal+Suggested+Replies+Model&rft.au=Ying%2C+Qianlan&rft.au=Bajaj%2C+Payal&rft.au=Deb%2C+Budhaditya&rft.au=Yang%2C+Yu&rft.date=2021-06-03&rft_id=info:doi/10.48550%2Farxiv.2106.02232&rft.externalDocID=2106_02232