Learning Graph Representation of Person-specific Cognitive Processes from Audio-visual Behaviours for Automatic Personality Recognition

This approach builds on two following findings in cognitive science: (i) human cognition partially determines expressed behaviour and is directly linked to true personality traits; and (ii) in dyadic interactions individuals' nonverbal behaviours are influenced by their conversational partner b...

Full description

Saved in:
Bibliographic Details
Main Authors Song, Siyang, Shao, Zilong, Jaiswal, Shashank, Shen, Linlin, Valstar, Michel, Gunes, Hatice
Format Journal Article
LanguageEnglish
Published 26.10.2021
Subjects
Online AccessGet full text
DOI10.48550/arxiv.2110.13570

Cover

Abstract This approach builds on two following findings in cognitive science: (i) human cognition partially determines expressed behaviour and is directly linked to true personality traits; and (ii) in dyadic interactions individuals' nonverbal behaviours are influenced by their conversational partner behaviours. In this context, we hypothesise that during a dyadic interaction, a target subject's facial reactions are driven by two main factors, i.e. their internal (person-specific) cognitive process, and the externalised nonverbal behaviours of their conversational partner. Consequently, we propose to represent the target subjects (defined as the listener) person-specific cognition in the form of a person-specific CNN architecture that has unique architectural parameters and depth, which takes audio-visual non-verbal cues displayed by the conversational partner (defined as the speaker) as input, and is able to reproduce the target subject's facial reactions. Each person-specific CNN is explored by the Neural Architecture Search (NAS) and a novel adaptive loss function, which is then represented as a graph representation for recognising the target subject's true personality. Experimental results not only show that the produced graph representations are well associated with target subjects' personality traits in both human-human and human-machine interaction scenarios, and outperform the existing approaches with significant advantages, but also demonstrate that the proposed novel strategies such as adaptive loss, and the end-to-end vertices/edges feature learning, help the proposed approach in learning more reliable personality representations.
AbstractList This approach builds on two following findings in cognitive science: (i) human cognition partially determines expressed behaviour and is directly linked to true personality traits; and (ii) in dyadic interactions individuals' nonverbal behaviours are influenced by their conversational partner behaviours. In this context, we hypothesise that during a dyadic interaction, a target subject's facial reactions are driven by two main factors, i.e. their internal (person-specific) cognitive process, and the externalised nonverbal behaviours of their conversational partner. Consequently, we propose to represent the target subjects (defined as the listener) person-specific cognition in the form of a person-specific CNN architecture that has unique architectural parameters and depth, which takes audio-visual non-verbal cues displayed by the conversational partner (defined as the speaker) as input, and is able to reproduce the target subject's facial reactions. Each person-specific CNN is explored by the Neural Architecture Search (NAS) and a novel adaptive loss function, which is then represented as a graph representation for recognising the target subject's true personality. Experimental results not only show that the produced graph representations are well associated with target subjects' personality traits in both human-human and human-machine interaction scenarios, and outperform the existing approaches with significant advantages, but also demonstrate that the proposed novel strategies such as adaptive loss, and the end-to-end vertices/edges feature learning, help the proposed approach in learning more reliable personality representations.
Author Song, Siyang
Jaiswal, Shashank
Shao, Zilong
Valstar, Michel
Gunes, Hatice
Shen, Linlin
Author_xml – sequence: 1
  givenname: Siyang
  surname: Song
  fullname: Song, Siyang
– sequence: 2
  givenname: Zilong
  surname: Shao
  fullname: Shao, Zilong
– sequence: 3
  givenname: Shashank
  surname: Jaiswal
  fullname: Jaiswal, Shashank
– sequence: 4
  givenname: Linlin
  surname: Shen
  fullname: Shen, Linlin
– sequence: 5
  givenname: Michel
  surname: Valstar
  fullname: Valstar, Michel
– sequence: 6
  givenname: Hatice
  surname: Gunes
  fullname: Gunes, Hatice
BackLink https://doi.org/10.48550/arXiv.2110.13570$$DView paper in arXiv
BookMark eNqFjkFuwkAMRWfRLijlAKzwBUITIKLbgkq76AJV3UfW4IClxI48kwhO0Gt3Stl39SV_6_334O5EhZybFvl89VyW-RPamYf5okiHYlmu85H7_iA0YTnCm2F3gk_qjAJJxMgqoDXsyYJKFjryXLOHrR6FIw8Ee1NPIVCA2rSFl_7Amg0cemxgQyccWHtLpVrqorYJ6W84bDhe0pj_g6k8uvsam0CTW47dbPf6tX3PrspVZ9yiXapf9eqqvvz_4wcb-1Pw
ContentType Journal Article
Copyright http://creativecommons.org/publicdomain/zero/1.0
Copyright_xml – notice: http://creativecommons.org/publicdomain/zero/1.0
DBID AKY
GOX
DOI 10.48550/arxiv.2110.13570
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2110_13570
GroupedDBID AKY
GOX
ID FETCH-arxiv_primary_2110_135703
IEDL.DBID GOX
IngestDate Wed Jul 23 01:26:36 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_2110_135703
OpenAccessLink https://arxiv.org/abs/2110.13570
ParticipantIDs arxiv_primary_2110_13570
PublicationCentury 2000
PublicationDate 2021-10-26
PublicationDateYYYYMMDD 2021-10-26
PublicationDate_xml – month: 10
  year: 2021
  text: 2021-10-26
  day: 26
PublicationDecade 2020
PublicationYear 2021
Score 3.5518627
SecondaryResourceType preprint
Snippet This approach builds on two following findings in cognitive science: (i) human cognition partially determines expressed behaviour and is directly linked to...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Computer Vision and Pattern Recognition
Title Learning Graph Representation of Person-specific Cognitive Processes from Audio-visual Behaviours for Automatic Personality Recognition
URI https://arxiv.org/abs/2110.13570
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV27TsMwFL0qnVgQFaDyvgOrReQal45VRVsh8VAFUrYobmyUxUZNUvEH_DbXjiO6dI0tO36fI597DHDnFYsmMYIlWkgmCNAymsqaqUlOE0wrsw4mSS-vcvkpntOHtAfYxcLkm59y2_oDq-resxP_MsOYSPkB555cLd7S9nIyWHHF_P_5CGOGTzuHxPwYjiK6w2k7HAPoaXsCv9HD9AsX3h0aV0F8GmN-LDqD7wH2Mh_16JU7OOs0PRh1_LpCHweC06YoHduWVUO1RGtDovtIyJPSahf8V2NxAWDjqhMIOXsKt_Onj9mShV_Pvluficy3KgutGp1B3zqrh4BarB-JUuWJkUbIxND6G6lizIvcEHNSxTkM95VysT_pEg65F2vQpszlFfTrTaOv6bSt1U3o8j85XYeq
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Learning+Graph+Representation+of+Person-specific+Cognitive+Processes+from+Audio-visual+Behaviours+for+Automatic+Personality+Recognition&rft.au=Song%2C+Siyang&rft.au=Shao%2C+Zilong&rft.au=Jaiswal%2C+Shashank&rft.au=Shen%2C+Linlin&rft.date=2021-10-26&rft_id=info:doi/10.48550%2Farxiv.2110.13570&rft.externalDocID=2110_13570