Achieving User-Side Fairness in Contextual Bandits
Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the recommendation strategy based on feedback. However, unfairness could incur in personalized recommendation. In this paper, we study how to achieve us...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
22.10.2020
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Personalized recommendation based on multi-arm bandit (MAB) algorithms has
shown to lead to high utility and efficiency as it can dynamically adapt the
recommendation strategy based on feedback. However, unfairness could incur in
personalized recommendation. In this paper, we study how to achieve user-side
fairness in personalized recommendation. We formulate our fair personalized
recommendation as a modified contextual bandit and focus on achieving fairness
on the individual whom is being recommended an item as opposed to achieving
fairness on the items that are being recommended. We introduce and define a
metric that captures the fairness in terms of rewards received for both the
privileged and protected groups. We develop a fair contextual bandit algorithm,
Fair-LinUCB, that improves upon the traditional LinUCB algorithm to achieve
group-level fairness of users. Our algorithm detects and monitors unfairness
while it learns to recommend personalized videos to students to achieve high
efficiency. We provide a theoretical regret analysis and show that our
algorithm has a slightly higher regret bound than LinUCB. We conduct numerous
experimental evaluations to compare the performances of our fair contextual
bandit to that of LinUCB and show that our approach achieves group-level
fairness while maintaining a high utility. |
---|---|
AbstractList | Personalized recommendation based on multi-arm bandit (MAB) algorithms has
shown to lead to high utility and efficiency as it can dynamically adapt the
recommendation strategy based on feedback. However, unfairness could incur in
personalized recommendation. In this paper, we study how to achieve user-side
fairness in personalized recommendation. We formulate our fair personalized
recommendation as a modified contextual bandit and focus on achieving fairness
on the individual whom is being recommended an item as opposed to achieving
fairness on the items that are being recommended. We introduce and define a
metric that captures the fairness in terms of rewards received for both the
privileged and protected groups. We develop a fair contextual bandit algorithm,
Fair-LinUCB, that improves upon the traditional LinUCB algorithm to achieve
group-level fairness of users. Our algorithm detects and monitors unfairness
while it learns to recommend personalized videos to students to achieve high
efficiency. We provide a theoretical regret analysis and show that our
algorithm has a slightly higher regret bound than LinUCB. We conduct numerous
experimental evaluations to compare the performances of our fair contextual
bandit to that of LinUCB and show that our approach achieves group-level
fairness while maintaining a high utility. |
Author | Lee, Dongwon Heffernan, Neil Huang, Wen Wu, Xintao Labille, Kevin |
Author_xml | – sequence: 1 givenname: Wen surname: Huang fullname: Huang, Wen – sequence: 2 givenname: Kevin surname: Labille fullname: Labille, Kevin – sequence: 3 givenname: Xintao surname: Wu fullname: Wu, Xintao – sequence: 4 givenname: Dongwon surname: Lee fullname: Lee, Dongwon – sequence: 5 givenname: Neil surname: Heffernan fullname: Heffernan, Neil |
BackLink | https://doi.org/10.48550/arXiv.2010.12102$$DView paper in arXiv |
BookMark | eNotzrFOwzAUhWEPMEDpAzDhF0i5voljZywRBaRKDLRzdB1fg6XiIjtU5e2BwnSkfzj6LsVZ2icW4lrBorFawy3lYzwsEH6CQgV4IXA5vkU-xPQqt4Vz9RI9yxXFnLgUGZPs92ni4_RJO3lHycepXInzQLvC8_-dic3qftM_Vuvnh6d-ua6oNVh5Z8fA0LIiAMOaPDoy6BoTjPIdW2qYXUfYOg2Eo9UBDJGpicZgXVfPxM3f7Qk9fOT4Tvlr-MUPJ3z9DVqgQTg |
ContentType | Journal Article |
Copyright | http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
Copyright_xml | – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
DBID | AKY GOX |
DOI | 10.48550/arxiv.2010.12102 |
DatabaseName | arXiv Computer Science arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2010_12102 |
GroupedDBID | AKY GOX |
ID | FETCH-LOGICAL-a672-db8cfe06e1a007e5ad2ba72b47f71d9e8a4eeb9a26b50a2c85f07aa73aacf8b93 |
IEDL.DBID | GOX |
IngestDate | Mon Jan 08 05:46:56 EST 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a672-db8cfe06e1a007e5ad2ba72b47f71d9e8a4eeb9a26b50a2c85f07aa73aacf8b93 |
OpenAccessLink | https://arxiv.org/abs/2010.12102 |
ParticipantIDs | arxiv_primary_2010_12102 |
PublicationCentury | 2000 |
PublicationDate | 2020-10-22 |
PublicationDateYYYYMMDD | 2020-10-22 |
PublicationDate_xml | – month: 10 year: 2020 text: 2020-10-22 day: 22 |
PublicationDecade | 2020 |
PublicationYear | 2020 |
Score | 1.7866527 |
SecondaryResourceType | preprint |
Snippet | Personalized recommendation based on multi-arm bandit (MAB) algorithms has
shown to lead to high utility and efficiency as it can dynamically adapt the... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Artificial Intelligence Computer Science - Learning |
Title | Achieving User-Side Fairness in Contextual Bandits |
URI | https://arxiv.org/abs/2010.12102 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NT8MwDLW2nbggEKDxqRy4RrRpm7THgRgTEnBgk3ar7MYRvUxoG2g_f0laBBeuiS-OJb9n2X4BuNVJk6mCS8nOocxzYoml07KyTnu6TT4hho7uy6ueLfLnZbEcgPjZhcH1rv3u9IFpcxcnr6LE1RCGSoWRrae3ZdecjFJcvf2vneeY8egPSEyP4LBnd2LSheMYBrw6ATVpPloOhbtY-IDL99aymGK7DllGtCsRFaJ2YZFD3Ictk-3mFObTx_nDTPZfFUjURklLZeM40Zyix1wu0CpCoyg3zqS24hJzZqpQaSoSVE1ZuMQgmgyxcSVV2RmMfLXPYxCeXqQWXcbaM33rDBaNp2xVSkROkUvOYRwdrD87NYo6-F5H3y_-v7qEAxUKRZ90lbqC0Xb9xdceTbd0E590D74BdYc |
link.rule.ids | 228,230,783,888 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Achieving+User-Side+Fairness+in+Contextual+Bandits&rft.au=Huang%2C+Wen&rft.au=Labille%2C+Kevin&rft.au=Wu%2C+Xintao&rft.au=Lee%2C+Dongwon&rft.date=2020-10-22&rft_id=info:doi/10.48550%2Farxiv.2010.12102&rft.externalDocID=2010_12102 |