Personality Alignment of Large Language Models

Aligning large language models (LLMs) typically aim to reflect general human values and behaviors, but they often fail to capture the unique characteristics and preferences of individual users. To address this gap, we introduce the concept of Personality Alignment. This approach tailors LLMs' r...

Full description

Saved in:
Bibliographic Details
Main Authors Zhu, Minjun, Weng, Yixuan, Yang, Linyi, Zhang, Yue
Format Journal Article
LanguageEnglish
Published 21.08.2024
Subjects
Online AccessGet full text
DOI10.48550/arxiv.2408.11779

Cover

Loading…
Abstract Aligning large language models (LLMs) typically aim to reflect general human values and behaviors, but they often fail to capture the unique characteristics and preferences of individual users. To address this gap, we introduce the concept of Personality Alignment. This approach tailors LLMs' responses and decisions to match the specific preferences of individual users or closely related groups. Inspired by psychometrics, we created the Personality Alignment with Personality Inventories (PAPI) dataset, which includes data from over 320,000 real subjects across multiple personality assessments, including both the Big Five Personality Factors and Dark Triad traits. This comprehensive dataset enables quantitative evaluation of LLMs' alignment capabilities across both positive and potentially problematic personality dimensions. Recognizing the challenges of personality alignments, such as limited personal data, diverse preferences, and scalability requirements, we developed an activation intervention optimization method. This method enhances LLMs' ability to efficiently align with individual behavioral preferences using minimal data and computational resources. Remarkably, our method, PAS, achieves superior performance while requiring only 1/5 of the optimization time compared to DPO, offering practical value for personality alignment. Our work paves the way for future AI systems to make decisions and reason in truly personality ways, enhancing the relevance and meaning of AI interactions for each user and advancing human-centered artificial intelligence. The dataset and code are released at https://github.com/zhu-minjun/PAlign.
AbstractList Aligning large language models (LLMs) typically aim to reflect general human values and behaviors, but they often fail to capture the unique characteristics and preferences of individual users. To address this gap, we introduce the concept of Personality Alignment. This approach tailors LLMs' responses and decisions to match the specific preferences of individual users or closely related groups. Inspired by psychometrics, we created the Personality Alignment with Personality Inventories (PAPI) dataset, which includes data from over 320,000 real subjects across multiple personality assessments, including both the Big Five Personality Factors and Dark Triad traits. This comprehensive dataset enables quantitative evaluation of LLMs' alignment capabilities across both positive and potentially problematic personality dimensions. Recognizing the challenges of personality alignments, such as limited personal data, diverse preferences, and scalability requirements, we developed an activation intervention optimization method. This method enhances LLMs' ability to efficiently align with individual behavioral preferences using minimal data and computational resources. Remarkably, our method, PAS, achieves superior performance while requiring only 1/5 of the optimization time compared to DPO, offering practical value for personality alignment. Our work paves the way for future AI systems to make decisions and reason in truly personality ways, enhancing the relevance and meaning of AI interactions for each user and advancing human-centered artificial intelligence. The dataset and code are released at https://github.com/zhu-minjun/PAlign.
Author Zhu, Minjun
Zhang, Yue
Yang, Linyi
Weng, Yixuan
Author_xml – sequence: 1
  givenname: Minjun
  surname: Zhu
  fullname: Zhu, Minjun
– sequence: 2
  givenname: Yixuan
  surname: Weng
  fullname: Weng, Yixuan
– sequence: 3
  givenname: Linyi
  surname: Yang
  fullname: Yang, Linyi
– sequence: 4
  givenname: Yue
  surname: Zhang
  fullname: Zhang, Yue
BackLink https://doi.org/10.48550/arXiv.2408.11779$$DView paper in arXiv
BookMark eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTMzIxsNAzNDQ3t-Rk0AtILSrOz0vMySypVHDMyUzPy03NK1HIT1PwSSxKTwWSeemliUCGb35Kak4xDwNrWmJOcSovlOZmkHdzDXH20AWbHF9QlJmbWFQZD7IhHmyDMWEVAPMPMbY
ContentType Journal Article
Copyright http://creativecommons.org/licenses/by/4.0
Copyright_xml – notice: http://creativecommons.org/licenses/by/4.0
DBID AKY
GOX
DOI 10.48550/arxiv.2408.11779
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2408_11779
GroupedDBID AKY
GOX
ID FETCH-arxiv_primary_2408_117793
IEDL.DBID GOX
IngestDate Tue Jul 22 23:12:57 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_2408_117793
OpenAccessLink https://arxiv.org/abs/2408.11779
ParticipantIDs arxiv_primary_2408_11779
PublicationCentury 2000
PublicationDate 2024-08-21
PublicationDateYYYYMMDD 2024-08-21
PublicationDate_xml – month: 08
  year: 2024
  text: 2024-08-21
  day: 21
PublicationDecade 2020
PublicationYear 2024
Score 3.765956
SecondaryResourceType preprint
Snippet Aligning large language models (LLMs) typically aim to reflect general human values and behaviors, but they often fail to capture the unique characteristics...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Computation and Language
Title Personality Alignment of Large Language Models
URI https://arxiv.org/abs/2408.11779
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQMTU3Skw2NjfWtTBKM9I1STW01LU0MjfSNUw2SDNIATaQzMH7uH39zDxCTbwiTCOYGBRge2ESiyoyyyDnAycV64PO3wJPK1oyMzAbGYGWbLn7R0AmJ8FHcUHVI9QB25hgIaRKwk2QgR_aulNwhESHEANTap4Ig14AosGr4JiTmQ6egFfIT1PwAS3DBpKQIUMF0L1kOcWiDPJuriHOHrpgG-ILIMdBxIMsB5-ZbWksxsAC7LSnSjAomFmkpICm0AySzBJBJ9QnAatakxQTS2AHxdg8JSlZkkEClylSuKWkGbiMgJUqaEzTyFCGgaWkqDRVFlgpliTJgUMGALjeZGg
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Personality+Alignment+of+Large+Language+Models&rft.au=Zhu%2C+Minjun&rft.au=Weng%2C+Yixuan&rft.au=Yang%2C+Linyi&rft.au=Zhang%2C+Yue&rft.date=2024-08-21&rft_id=info:doi/10.48550%2Farxiv.2408.11779&rft.externalDocID=2408_11779