Do large language models resemble humans in language use?

Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable capacities in comprehending and producing language. However, their internal workings remain a black box, and it is unclear whether LLMs and chatbots can develop humanlike characteristics in language use. Cognitive scientis...

Full description

Saved in:
Bibliographic Details
Main Authors Cai, Zhenguang G, Duan, Xufeng, Haslett, David A, Wang, Shuqi, Pickering, Martin J
Format Journal Article
LanguageEnglish
Published 10.03.2023
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable capacities in comprehending and producing language. However, their internal workings remain a black box, and it is unclear whether LLMs and chatbots can develop humanlike characteristics in language use. Cognitive scientists have devised many experiments that probe, and have made great progress in explaining, how people comprehend and produce language. We subjected ChatGPT and Vicuna to 12 of these experiments ranging from sounds to dialogue, preregistered and with 1000 runs (i.e., iterations) per experiment. ChatGPT and Vicuna replicated the human pattern of language use in 10 and 7 out of the 12 experiments, respectively. The models associated unfamiliar words with different meanings depending on their forms, continued to access recently encountered meanings of ambiguous words, reused recent sentence structures, attributed causality as a function of verb semantics, and accessed different meanings and retrieved different words depending on an interlocutor's identity. In addition, ChatGPT, but not Vicuna, nonliterally interpreted implausible sentences that were likely to have been corrupted by noise, drew reasonable inferences, and overlooked semantic fallacies in a sentence. Finally, unlike humans, neither model preferred using shorter words to convey less informative content, nor did they use context to resolve syntactic ambiguities. We discuss how these convergences and divergences may result from the transformer architecture. Overall, these experiments demonstrate that LLMs such as ChatGPT (and Vicuna to a lesser extent) are humanlike in many aspects of human language processing.
AbstractList Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable capacities in comprehending and producing language. However, their internal workings remain a black box, and it is unclear whether LLMs and chatbots can develop humanlike characteristics in language use. Cognitive scientists have devised many experiments that probe, and have made great progress in explaining, how people comprehend and produce language. We subjected ChatGPT and Vicuna to 12 of these experiments ranging from sounds to dialogue, preregistered and with 1000 runs (i.e., iterations) per experiment. ChatGPT and Vicuna replicated the human pattern of language use in 10 and 7 out of the 12 experiments, respectively. The models associated unfamiliar words with different meanings depending on their forms, continued to access recently encountered meanings of ambiguous words, reused recent sentence structures, attributed causality as a function of verb semantics, and accessed different meanings and retrieved different words depending on an interlocutor's identity. In addition, ChatGPT, but not Vicuna, nonliterally interpreted implausible sentences that were likely to have been corrupted by noise, drew reasonable inferences, and overlooked semantic fallacies in a sentence. Finally, unlike humans, neither model preferred using shorter words to convey less informative content, nor did they use context to resolve syntactic ambiguities. We discuss how these convergences and divergences may result from the transformer architecture. Overall, these experiments demonstrate that LLMs such as ChatGPT (and Vicuna to a lesser extent) are humanlike in many aspects of human language processing.
Author Duan, Xufeng
Pickering, Martin J
Haslett, David A
Wang, Shuqi
Cai, Zhenguang G
Author_xml – sequence: 1
  givenname: Zhenguang G
  surname: Cai
  fullname: Cai, Zhenguang G
– sequence: 2
  givenname: Xufeng
  surname: Duan
  fullname: Duan, Xufeng
– sequence: 3
  givenname: David A
  surname: Haslett
  fullname: Haslett, David A
– sequence: 4
  givenname: Shuqi
  surname: Wang
  fullname: Wang, Shuqi
– sequence: 5
  givenname: Martin J
  surname: Pickering
  fullname: Pickering, Martin J
BackLink https://doi.org/10.48550/arXiv.2303.08014$$DView paper in arXiv
BookMark eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTMzI2MNYzsDAwNOFksHTJV8hJLEpPBZJ56aWJQEZufkpqTrFCUWpxam5STqpCRmluYl6xQmYeQklpcao9DwNrWmJOcSovlOZmkHdzDXH20AVbEl9QlJmbWFQZD7IsHmyZMWEVAFT4NXI
ContentType Journal Article
Copyright http://creativecommons.org/licenses/by/4.0
Copyright_xml – notice: http://creativecommons.org/licenses/by/4.0
DBID AKY
GOX
DOI 10.48550/arxiv.2303.08014
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2303_08014
GroupedDBID AKY
GOX
ID FETCH-arxiv_primary_2303_080143
IEDL.DBID GOX
IngestDate Thu Mar 28 12:15:21 EDT 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-arxiv_primary_2303_080143
OpenAccessLink https://arxiv.org/abs/2303.08014
ParticipantIDs arxiv_primary_2303_08014
PublicationCentury 2000
PublicationDate 2023-03-10
PublicationDateYYYYMMDD 2023-03-10
PublicationDate_xml – month: 03
  year: 2023
  text: 2023-03-10
  day: 10
PublicationDecade 2020
PublicationYear 2023
Score 3.7457256
SecondaryResourceType preprint
Snippet Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable capacities in comprehending and producing language. However, their internal...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Computation and Language
Title Do large language models resemble humans in language use?
URI https://arxiv.org/abs/2303.08014
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQAaZZQ2Ctk6SbapyYrGtiYm4JOvLWSDcRVB0bglK0IWiDs6-fmUeoiVeEaQQTgwJsL0xiUUVmGeR84KRifWD72FjPAHTACTMDs5ERaMmWu38EZHISfBQXVD1CHbCNCRZCqiTcBBn4oa07BUdIdAgxMKXmiTBYuuQr5IAWXCvABgcVwPfPFCuAtv7kJuWkKoBvyitWyMxDKCktTrUXZZB3cw1x9tAFWxZfADkZIh7kjniwO4zFGFiA_fdUCQYFAyPLlORES6PUFGDXxzDJyCIlzdLIyNIoKckg1TA1NUmSQQKXKVK4paQZuEA3n-uCl5bJMLCUFJWmygLrx5IkOXAgAQAyTmhZ
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Do+large+language+models+resemble+humans+in+language+use%3F&rft.au=Cai%2C+Zhenguang+G&rft.au=Duan%2C+Xufeng&rft.au=Haslett%2C+David+A&rft.au=Wang%2C+Shuqi&rft.date=2023-03-10&rft_id=info:doi/10.48550%2Farxiv.2303.08014&rft.externalDocID=2303_08014