Topics in Contextualised Attention Embeddings
Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications. Complementary to these language models are probabilistic topic models that learn thematic patterns from the text. Recent work has demonstrated that cond...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
11.01.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications. Complementary to these language models are probabilistic topic models that learn thematic patterns from the text. Recent work has demonstrated that conducting clustering on the word-level contextual representations from a language model emulates word clusters that are discovered in latent topics of words from Latent Dirichlet Allocation. The important question is how such topical word clusters are automatically formed, through clustering, in the language model when it has not been explicitly designed to model latent topics. To address this question, we design different probe experiments. Using BERT and DistilBERT, we find that the attention framework plays a key role in modelling such word topic clusters. We strongly believe that our work paves way for further research into the relationships between probabilistic topic models and pre-trained language models. |
---|---|
AbstractList | Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications. Complementary to these language models are probabilistic topic models that learn thematic patterns from the text. Recent work has demonstrated that conducting clustering on the word-level contextual representations from a language model emulates word clusters that are discovered in latent topics of words from Latent Dirichlet Allocation. The important question is how such topical word clusters are automatically formed, through clustering, in the language model when it has not been explicitly designed to model latent topics. To address this question, we design different probe experiments. Using BERT and DistilBERT, we find that the attention framework plays a key role in modelling such word topic clusters. We strongly believe that our work paves way for further research into the relationships between probabilistic topic models and pre-trained language models. |
Author | Alba Garcia Seco de Herrera Talebpour, Mozhgan Shoaib Jameel |
Author_xml | – sequence: 1 givenname: Mozhgan surname: Talebpour fullname: Talebpour, Mozhgan – sequence: 2 fullname: Alba Garcia Seco de Herrera – sequence: 3 fullname: Shoaib Jameel |
BookMark | eNqNykEOgjAQQNHGaCIqd2jiukmZFqpLQzAegD1BGU0JTpEpicfXhQdw9Rfvb8SSAuFCJGBMpg4WYC1S5l5rDYWDPDeJUHUY_Y2lJ1kGiviOczt4xk6eYkSKPpCsnlfsOk8P3onVvR0Y01-3Yn-u6vKixim8ZuTY9GGe6EsNuMK6o80AzH_XBwYPNAU |
ContentType | Paper |
Copyright | 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central UK/Ireland ProQuest Central Essentials AUTh Library subscriptions: ProQuest Central Technology Collection ProQuest One Community College ProQuest Central SciTech Premium Collection (Proquest) (PQ_SDU_P3) ProQuest Engineering Collection ProQuest Engineering Database Access via ProQuest (Open Access) ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_27647941223 |
IEDL.DBID | 8FG |
IngestDate | Thu Oct 10 17:59:26 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_27647941223 |
OpenAccessLink | https://www.proquest.com/docview/2764794122?pq-origsite=%requestingapplication% |
PQID | 2764794122 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_2764794122 |
PublicationCentury | 2000 |
PublicationDate | 20230111 |
PublicationDateYYYYMMDD | 2023-01-11 |
PublicationDate_xml | – month: 01 year: 2023 text: 20230111 day: 11 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2023 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.4501863 |
SecondaryResourceType | preprint |
Snippet | Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications.... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Clustering Dirichlet problem Questions |
Title | Topics in Contextualised Attention Embeddings |
URI | https://www.proquest.com/docview/2764794122 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1LSwMxEB60i-CtvvBRS0CvQZPNa09FZdcitBSp0FvZZLPQg-22u4Inf7tJ3OpB6C1hYBLCMJP55gVwKxQjJGYM-56nmAle4iRnBisiNY9tQgrlC4VHYzF8Yy8zPmsBt7pNq9zqxKCoi5XxGPkdlcI3QyeUDqo19lOjfHS1HaGxDxGhUnqpVtnzL8ZChXQ_5vifmg22I-tCNMkruzmCPbs8hoOQcmnqE8DTVeUWaLFEoUXUp6_kWNS2QA9N85OEiNJ3bYsQHTqFmyydPg3x9ox5KwX1_O_O8Rl0nDtvzwFJYqh2lJzKkpWaK6ETyRPLrbo3zvu5gN4uTpe7yVdw6Aeie5CAkB50ms2HvXZms9H98DZ9iB7T8eTV7UZf6Tc6BnZP |
link.rule.ids | 783,787,12779,21402,33387,33758,43614,43819 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3dS8MwED90Rdybn_gxNaCvQZPmo30SlY6qWxlSYW-lSVPYg1tdK-zPN6mdPgh7CxwkIRx3ud_d_Q7gRgSMEJ8x7DhPMRO8xGHONA6IVNw3ISkC1yg8TkT8zl6mfNoBbnVXVrm2ia2hLhbaYeS3VApHhk4ova8-sZsa5bKr3QiNbfAcVZXVau8xSiZvvygLFdL-mf1_hrb1HsM98CZ5ZZb7sGXmB7DTFl3q-hBwuqjsAs3mqCWJWrlejlltCvTQND9liCj6UKZo80NHcD2M0qcYr8_IOj2os79b-8fQswG9OQEkiabKSnIqS1YqHggVSh4aboI7beOfUxhs2ulss_gKduN0PMpGz8nrOfTdeHQHGRAygF6z_DIX1ok26rJ7qW9rDXfV |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Topics+in+Contextualised+Attention+Embeddings&rft.jtitle=arXiv.org&rft.au=Talebpour%2C+Mozhgan&rft.au=Alba+Garcia+Seco+de+Herrera&rft.au=Shoaib+Jameel&rft.date=2023-01-11&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |