Do large language models resemble humans in language use?
Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable capacities in comprehending and producing language. However, their internal workings remain a black box, and it is unclear whether LLMs and chatbots can develop humanlike characteristics in language use. Cognitive scientis...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
10.03.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable
capacities in comprehending and producing language. However, their internal
workings remain a black box, and it is unclear whether LLMs and chatbots can
develop humanlike characteristics in language use. Cognitive scientists have
devised many experiments that probe, and have made great progress in
explaining, how people comprehend and produce language. We subjected ChatGPT
and Vicuna to 12 of these experiments ranging from sounds to dialogue,
preregistered and with 1000 runs (i.e., iterations) per experiment. ChatGPT and
Vicuna replicated the human pattern of language use in 10 and 7 out of the 12
experiments, respectively. The models associated unfamiliar words with
different meanings depending on their forms, continued to access recently
encountered meanings of ambiguous words, reused recent sentence structures,
attributed causality as a function of verb semantics, and accessed different
meanings and retrieved different words depending on an interlocutor's identity.
In addition, ChatGPT, but not Vicuna, nonliterally interpreted implausible
sentences that were likely to have been corrupted by noise, drew reasonable
inferences, and overlooked semantic fallacies in a sentence. Finally, unlike
humans, neither model preferred using shorter words to convey less informative
content, nor did they use context to resolve syntactic ambiguities. We discuss
how these convergences and divergences may result from the transformer
architecture. Overall, these experiments demonstrate that LLMs such as ChatGPT
(and Vicuna to a lesser extent) are humanlike in many aspects of human language
processing. |
---|---|
AbstractList | Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable
capacities in comprehending and producing language. However, their internal
workings remain a black box, and it is unclear whether LLMs and chatbots can
develop humanlike characteristics in language use. Cognitive scientists have
devised many experiments that probe, and have made great progress in
explaining, how people comprehend and produce language. We subjected ChatGPT
and Vicuna to 12 of these experiments ranging from sounds to dialogue,
preregistered and with 1000 runs (i.e., iterations) per experiment. ChatGPT and
Vicuna replicated the human pattern of language use in 10 and 7 out of the 12
experiments, respectively. The models associated unfamiliar words with
different meanings depending on their forms, continued to access recently
encountered meanings of ambiguous words, reused recent sentence structures,
attributed causality as a function of verb semantics, and accessed different
meanings and retrieved different words depending on an interlocutor's identity.
In addition, ChatGPT, but not Vicuna, nonliterally interpreted implausible
sentences that were likely to have been corrupted by noise, drew reasonable
inferences, and overlooked semantic fallacies in a sentence. Finally, unlike
humans, neither model preferred using shorter words to convey less informative
content, nor did they use context to resolve syntactic ambiguities. We discuss
how these convergences and divergences may result from the transformer
architecture. Overall, these experiments demonstrate that LLMs such as ChatGPT
(and Vicuna to a lesser extent) are humanlike in many aspects of human language
processing. |
Author | Duan, Xufeng Pickering, Martin J Haslett, David A Wang, Shuqi Cai, Zhenguang G |
Author_xml | – sequence: 1 givenname: Zhenguang G surname: Cai fullname: Cai, Zhenguang G – sequence: 2 givenname: Xufeng surname: Duan fullname: Duan, Xufeng – sequence: 3 givenname: David A surname: Haslett fullname: Haslett, David A – sequence: 4 givenname: Shuqi surname: Wang fullname: Wang, Shuqi – sequence: 5 givenname: Martin J surname: Pickering fullname: Pickering, Martin J |
BackLink | https://doi.org/10.48550/arXiv.2303.08014$$DView paper in arXiv |
BookMark | eNrjYmDJy89LZWCQNDTQM7EwNTXQTyyqyCzTMzI2MNYzsDAwNOFksHTJV8hJLEpPBZJ56aWJQEZufkpqTrFCUWpxam5STqpCRmluYl6xQmYeQklpcao9DwNrWmJOcSovlOZmkHdzDXH20AVbEl9QlJmbWFQZD7IsHmyZMWEVAFT4NXI |
ContentType | Journal Article |
Copyright | http://creativecommons.org/licenses/by/4.0 |
Copyright_xml | – notice: http://creativecommons.org/licenses/by/4.0 |
DBID | AKY GOX |
DOI | 10.48550/arxiv.2303.08014 |
DatabaseName | arXiv Computer Science arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2303_08014 |
GroupedDBID | AKY GOX |
ID | FETCH-arxiv_primary_2303_080143 |
IEDL.DBID | GOX |
IngestDate | Thu Mar 28 12:15:21 EDT 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-arxiv_primary_2303_080143 |
OpenAccessLink | https://arxiv.org/abs/2303.08014 |
ParticipantIDs | arxiv_primary_2303_08014 |
PublicationCentury | 2000 |
PublicationDate | 2023-03-10 |
PublicationDateYYYYMMDD | 2023-03-10 |
PublicationDate_xml | – month: 03 year: 2023 text: 2023-03-10 day: 10 |
PublicationDecade | 2020 |
PublicationYear | 2023 |
Score | 3.7457256 |
SecondaryResourceType | preprint |
Snippet | Large language models (LLMs) such as ChatGPT and Vicuna have shown remarkable
capacities in comprehending and producing language. However, their internal... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Computation and Language |
Title | Do large language models resemble humans in language use? |
URI | https://arxiv.org/abs/2303.08014 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQAaZZQ2Ctk6SbapyYrGtiYm4JOvLWSDcRVB0bglK0IWiDs6-fmUeoiVeEaQQTgwJsL0xiUUVmGeR84KRifWD72FjPAHTACTMDs5ERaMmWu38EZHISfBQXVD1CHbCNCRZCqiTcBBn4oa07BUdIdAgxMKXmiTBYuuQr5IAWXCvABgcVwPfPFCuAtv7kJuWkKoBvyitWyMxDKCktTrUXZZB3cw1x9tAFWxZfADkZIh7kjniwO4zFGFiA_fdUCQYFAyPLlORES6PUFGDXxzDJyCIlzdLIyNIoKckg1TA1NUmSQQKXKVK4paQZuEA3n-uCl5bJMLCUFJWmygLrx5IkOXAgAQAyTmhZ |
link.rule.ids | 228,230,783,888 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Do+large+language+models+resemble+humans+in+language+use%3F&rft.au=Cai%2C+Zhenguang+G&rft.au=Duan%2C+Xufeng&rft.au=Haslett%2C+David+A&rft.au=Wang%2C+Shuqi&rft.date=2023-03-10&rft_id=info:doi/10.48550%2Farxiv.2303.08014&rft.externalDocID=2303_08014 |