Achieving User-Side Fairness in Contextual Bandits

Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the recommendation strategy based on feedback. However, unfairness could incur in personalized recommendation. In this paper, we study how to achieve us...

Full description

Saved in:
Bibliographic Details
Main Authors Huang, Wen, Labille, Kevin, Wu, Xintao, Lee, Dongwon, Heffernan, Neil
Format Journal Article
LanguageEnglish
Published 22.10.2020
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the recommendation strategy based on feedback. However, unfairness could incur in personalized recommendation. In this paper, we study how to achieve user-side fairness in personalized recommendation. We formulate our fair personalized recommendation as a modified contextual bandit and focus on achieving fairness on the individual whom is being recommended an item as opposed to achieving fairness on the items that are being recommended. We introduce and define a metric that captures the fairness in terms of rewards received for both the privileged and protected groups. We develop a fair contextual bandit algorithm, Fair-LinUCB, that improves upon the traditional LinUCB algorithm to achieve group-level fairness of users. Our algorithm detects and monitors unfairness while it learns to recommend personalized videos to students to achieve high efficiency. We provide a theoretical regret analysis and show that our algorithm has a slightly higher regret bound than LinUCB. We conduct numerous experimental evaluations to compare the performances of our fair contextual bandit to that of LinUCB and show that our approach achieves group-level fairness while maintaining a high utility.
AbstractList Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the recommendation strategy based on feedback. However, unfairness could incur in personalized recommendation. In this paper, we study how to achieve user-side fairness in personalized recommendation. We formulate our fair personalized recommendation as a modified contextual bandit and focus on achieving fairness on the individual whom is being recommended an item as opposed to achieving fairness on the items that are being recommended. We introduce and define a metric that captures the fairness in terms of rewards received for both the privileged and protected groups. We develop a fair contextual bandit algorithm, Fair-LinUCB, that improves upon the traditional LinUCB algorithm to achieve group-level fairness of users. Our algorithm detects and monitors unfairness while it learns to recommend personalized videos to students to achieve high efficiency. We provide a theoretical regret analysis and show that our algorithm has a slightly higher regret bound than LinUCB. We conduct numerous experimental evaluations to compare the performances of our fair contextual bandit to that of LinUCB and show that our approach achieves group-level fairness while maintaining a high utility.
Author Lee, Dongwon
Heffernan, Neil
Huang, Wen
Wu, Xintao
Labille, Kevin
Author_xml – sequence: 1
  givenname: Wen
  surname: Huang
  fullname: Huang, Wen
– sequence: 2
  givenname: Kevin
  surname: Labille
  fullname: Labille, Kevin
– sequence: 3
  givenname: Xintao
  surname: Wu
  fullname: Wu, Xintao
– sequence: 4
  givenname: Dongwon
  surname: Lee
  fullname: Lee, Dongwon
– sequence: 5
  givenname: Neil
  surname: Heffernan
  fullname: Heffernan, Neil
BackLink https://doi.org/10.48550/arXiv.2010.12102$$DView paper in arXiv
BookMark eNotzrFOwzAUhWEPMEDpAzDhF0i5voljZywRBaRKDLRzdB1fg6XiIjtU5e2BwnSkfzj6LsVZ2icW4lrBorFawy3lYzwsEH6CQgV4IXA5vkU-xPQqt4Vz9RI9yxXFnLgUGZPs92ni4_RJO3lHycepXInzQLvC8_-dic3qftM_Vuvnh6d-ua6oNVh5Z8fA0LIiAMOaPDoy6BoTjPIdW2qYXUfYOg2Eo9UBDJGpicZgXVfPxM3f7Qk9fOT4Tvlr-MUPJ3z9DVqgQTg
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
GOX
DOI 10.48550/arxiv.2010.12102
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2010_12102
GroupedDBID AKY
GOX
ID FETCH-LOGICAL-a672-db8cfe06e1a007e5ad2ba72b47f71d9e8a4eeb9a26b50a2c85f07aa73aacf8b93
IEDL.DBID GOX
IngestDate Mon Jan 08 05:46:56 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a672-db8cfe06e1a007e5ad2ba72b47f71d9e8a4eeb9a26b50a2c85f07aa73aacf8b93
OpenAccessLink https://arxiv.org/abs/2010.12102
ParticipantIDs arxiv_primary_2010_12102
PublicationCentury 2000
PublicationDate 2020-10-22
PublicationDateYYYYMMDD 2020-10-22
PublicationDate_xml – month: 10
  year: 2020
  text: 2020-10-22
  day: 22
PublicationDecade 2020
PublicationYear 2020
Score 1.7866527
SecondaryResourceType preprint
Snippet Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Artificial Intelligence
Computer Science - Learning
Title Achieving User-Side Fairness in Contextual Bandits
URI https://arxiv.org/abs/2010.12102
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NT8MwDLW2nbggEKDxqRy4RrRpm7THgRgTEnBgk3ar7MYRvUxoG2g_f0laBBeuiS-OJb9n2X4BuNVJk6mCS8nOocxzYoml07KyTnu6TT4hho7uy6ueLfLnZbEcgPjZhcH1rv3u9IFpcxcnr6LE1RCGSoWRrae3ZdecjFJcvf2vneeY8egPSEyP4LBnd2LSheMYBrw6ATVpPloOhbtY-IDL99aymGK7DllGtCsRFaJ2YZFD3Ictk-3mFObTx_nDTPZfFUjURklLZeM40Zyix1wu0CpCoyg3zqS24hJzZqpQaSoSVE1ZuMQgmgyxcSVV2RmMfLXPYxCeXqQWXcbaM33rDBaNp2xVSkROkUvOYRwdrD87NYo6-F5H3y_-v7qEAxUKRZ90lbqC0Xb9xdceTbd0E590D74BdYc
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Achieving+User-Side+Fairness+in+Contextual+Bandits&rft.au=Huang%2C+Wen&rft.au=Labille%2C+Kevin&rft.au=Wu%2C+Xintao&rft.au=Lee%2C+Dongwon&rft.date=2020-10-22&rft_id=info:doi/10.48550%2Farxiv.2010.12102&rft.externalDocID=2010_12102