Learning Graph Representation of Person-specific Cognitive Processes from Audio-visual Behaviours for Automatic Personality Recognition
This approach builds on two following findings in cognitive science: (i) human cognition partially determines expressed behaviour and is directly linked to true personality traits; and (ii) in dyadic interactions individuals' nonverbal behaviours are influenced by their conversational partner b...
Saved in:
Main Authors | , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
26.10.2021
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.2110.13570 |
Cover
Abstract | This approach builds on two following findings in cognitive science: (i)
human cognition partially determines expressed behaviour and is directly linked
to true personality traits; and (ii) in dyadic interactions individuals'
nonverbal behaviours are influenced by their conversational partner behaviours.
In this context, we hypothesise that during a dyadic interaction, a target
subject's facial reactions are driven by two main factors, i.e. their internal
(person-specific) cognitive process, and the externalised nonverbal behaviours
of their conversational partner. Consequently, we propose to represent the
target subjects (defined as the listener) person-specific cognition in the form
of a person-specific CNN architecture that has unique architectural parameters
and depth, which takes audio-visual non-verbal cues displayed by the
conversational partner (defined as the speaker) as input, and is able to
reproduce the target subject's facial reactions. Each person-specific CNN is
explored by the Neural Architecture Search (NAS) and a novel adaptive loss
function, which is then represented as a graph representation for recognising
the target subject's true personality. Experimental results not only show that
the produced graph representations are well associated with target subjects'
personality traits in both human-human and human-machine interaction scenarios,
and outperform the existing approaches with significant advantages, but also
demonstrate that the proposed novel strategies such as adaptive loss, and the
end-to-end vertices/edges feature learning, help the proposed approach in
learning more reliable personality representations. |
---|---|
AbstractList | This approach builds on two following findings in cognitive science: (i)
human cognition partially determines expressed behaviour and is directly linked
to true personality traits; and (ii) in dyadic interactions individuals'
nonverbal behaviours are influenced by their conversational partner behaviours.
In this context, we hypothesise that during a dyadic interaction, a target
subject's facial reactions are driven by two main factors, i.e. their internal
(person-specific) cognitive process, and the externalised nonverbal behaviours
of their conversational partner. Consequently, we propose to represent the
target subjects (defined as the listener) person-specific cognition in the form
of a person-specific CNN architecture that has unique architectural parameters
and depth, which takes audio-visual non-verbal cues displayed by the
conversational partner (defined as the speaker) as input, and is able to
reproduce the target subject's facial reactions. Each person-specific CNN is
explored by the Neural Architecture Search (NAS) and a novel adaptive loss
function, which is then represented as a graph representation for recognising
the target subject's true personality. Experimental results not only show that
the produced graph representations are well associated with target subjects'
personality traits in both human-human and human-machine interaction scenarios,
and outperform the existing approaches with significant advantages, but also
demonstrate that the proposed novel strategies such as adaptive loss, and the
end-to-end vertices/edges feature learning, help the proposed approach in
learning more reliable personality representations. |
Author | Song, Siyang Jaiswal, Shashank Shao, Zilong Valstar, Michel Gunes, Hatice Shen, Linlin |
Author_xml | – sequence: 1 givenname: Siyang surname: Song fullname: Song, Siyang – sequence: 2 givenname: Zilong surname: Shao fullname: Shao, Zilong – sequence: 3 givenname: Shashank surname: Jaiswal fullname: Jaiswal, Shashank – sequence: 4 givenname: Linlin surname: Shen fullname: Shen, Linlin – sequence: 5 givenname: Michel surname: Valstar fullname: Valstar, Michel – sequence: 6 givenname: Hatice surname: Gunes fullname: Gunes, Hatice |
BackLink | https://doi.org/10.48550/arXiv.2110.13570$$DView paper in arXiv |
BookMark | eNqFjkFuwkAMRWfRLijlAKzwBUITIKLbgkq76AJV3UfW4IClxI48kwhO0Gt3Stl39SV_6_334O5EhZybFvl89VyW-RPamYf5okiHYlmu85H7_iA0YTnCm2F3gk_qjAJJxMgqoDXsyYJKFjryXLOHrR6FIw8Ee1NPIVCA2rSFl_7Amg0cemxgQyccWHtLpVrqorYJ6W84bDhe0pj_g6k8uvsam0CTW47dbPf6tX3PrspVZ9yiXapf9eqqvvz_4wcb-1Pw |
ContentType | Journal Article |
Copyright | http://creativecommons.org/publicdomain/zero/1.0 |
Copyright_xml | – notice: http://creativecommons.org/publicdomain/zero/1.0 |
DBID | AKY GOX |
DOI | 10.48550/arxiv.2110.13570 |
DatabaseName | arXiv Computer Science arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2110_13570 |
GroupedDBID | AKY GOX |
ID | FETCH-arxiv_primary_2110_135703 |
IEDL.DBID | GOX |
IngestDate | Wed Jul 23 01:26:36 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-arxiv_primary_2110_135703 |
OpenAccessLink | https://arxiv.org/abs/2110.13570 |
ParticipantIDs | arxiv_primary_2110_13570 |
PublicationCentury | 2000 |
PublicationDate | 2021-10-26 |
PublicationDateYYYYMMDD | 2021-10-26 |
PublicationDate_xml | – month: 10 year: 2021 text: 2021-10-26 day: 26 |
PublicationDecade | 2020 |
PublicationYear | 2021 |
Score | 3.5518627 |
SecondaryResourceType | preprint |
Snippet | This approach builds on two following findings in cognitive science: (i)
human cognition partially determines expressed behaviour and is directly linked
to... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Computer Vision and Pattern Recognition |
Title | Learning Graph Representation of Person-specific Cognitive Processes from Audio-visual Behaviours for Automatic Personality Recognition |
URI | https://arxiv.org/abs/2110.13570 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV27TsMwFL0qnVgQFaDyvgOrReQal45VRVsh8VAFUrYobmyUxUZNUvEH_DbXjiO6dI0tO36fI597DHDnFYsmMYIlWkgmCNAymsqaqUlOE0wrsw4mSS-vcvkpntOHtAfYxcLkm59y2_oDq-resxP_MsOYSPkB555cLd7S9nIyWHHF_P_5CGOGTzuHxPwYjiK6w2k7HAPoaXsCv9HD9AsX3h0aV0F8GmN-LDqD7wH2Mh_16JU7OOs0PRh1_LpCHweC06YoHduWVUO1RGtDovtIyJPSahf8V2NxAWDjqhMIOXsKt_Onj9mShV_Pvluficy3KgutGp1B3zqrh4BarB-JUuWJkUbIxND6G6lizIvcEHNSxTkM95VysT_pEg65F2vQpszlFfTrTaOv6bSt1U3o8j85XYeq |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Learning+Graph+Representation+of+Person-specific+Cognitive+Processes+from+Audio-visual+Behaviours+for+Automatic+Personality+Recognition&rft.au=Song%2C+Siyang&rft.au=Shao%2C+Zilong&rft.au=Jaiswal%2C+Shashank&rft.au=Shen%2C+Linlin&rft.date=2021-10-26&rft_id=info:doi/10.48550%2Farxiv.2110.13570&rft.externalDocID=2110_13570 |