MAQA: A Multimodal QA Benchmark for Negation
Multimodal learning can benefit from the representation power of pretrained Large Language Models (LLMs). However, state-of-the-art transformer based LLMs often ignore negations in natural language and there is no existing benchmark to quantitatively evaluate whether multimodal transformers inherit...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
09.01.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Multimodal learning can benefit from the representation power of pretrained Large Language Models (LLMs). However, state-of-the-art transformer based LLMs often ignore negations in natural language and there is no existing benchmark to quantitatively evaluate whether multimodal transformers inherit this weakness. In this study, we present a new multimodal question answering (QA) benchmark adapted from labeled music videos in AudioSet (Gemmeke et al., 2017) with the goal of systematically evaluating if multimodal transformers can perform complex reasoning to recognize new concepts as negation of previously learned concepts. We show that with standard fine-tuning approach multimodal transformers are still incapable of correctly interpreting negation irrespective of model size. However, our experiments demonstrate that augmenting the original training task distributions with negated QA examples allow the model to reliably reason with negation. To do this, we describe a novel data generation procedure that prompts the 540B-parameter PaLM model to automatically generate negated QA examples as compositions of easily accessible video tags. The generated examples contain more natural linguistic patterns and the gains compared to template-based task augmentation approach are significant. |
---|---|
AbstractList | Multimodal learning can benefit from the representation power of pretrained Large Language Models (LLMs). However, state-of-the-art transformer based LLMs often ignore negations in natural language and there is no existing benchmark to quantitatively evaluate whether multimodal transformers inherit this weakness. In this study, we present a new multimodal question answering (QA) benchmark adapted from labeled music videos in AudioSet (Gemmeke et al., 2017) with the goal of systematically evaluating if multimodal transformers can perform complex reasoning to recognize new concepts as negation of previously learned concepts. We show that with standard fine-tuning approach multimodal transformers are still incapable of correctly interpreting negation irrespective of model size. However, our experiments demonstrate that augmenting the original training task distributions with negated QA examples allow the model to reliably reason with negation. To do this, we describe a novel data generation procedure that prompts the 540B-parameter PaLM model to automatically generate negated QA examples as compositions of easily accessible video tags. The generated examples contain more natural linguistic patterns and the gains compared to template-based task augmentation approach are significant. |
Author | Ganti, Ravi Huang, Qingqing Judith Yue Li Jansen, Aren Kuzmin, Dima Lee, Joonseok |
Author_xml | – sequence: 1 fullname: Judith Yue Li – sequence: 2 givenname: Aren surname: Jansen fullname: Jansen, Aren – sequence: 3 givenname: Qingqing surname: Huang fullname: Huang, Qingqing – sequence: 4 givenname: Joonseok surname: Lee fullname: Lee, Joonseok – sequence: 5 givenname: Ravi surname: Ganti fullname: Ganti, Ravi – sequence: 6 givenname: Dima surname: Kuzmin fullname: Kuzmin, Dima |
BookMark | eNrjYmDJy89LZWLgNDI2NtS1MDEy4mDgLS7OMjAwMDIzNzI1NeZk0PF1DHS0UnBU8C3NKcnMzU9JzFEIdFRwSs1LzshNLMpWSMsvUvBLTU8syczP42FgTUvMKU7lhdLcDMpuriHOHroFRfmFpanFJfFZ-aVFeUCpeCNzM2NDUwtDEwNj4lQBAOjwMcU |
ContentType | Paper |
Copyright | 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central UK/Ireland ProQuest Central Essentials AUTh Library subscriptions: ProQuest Central Technology Collection ProQuest One Community College ProQuest Central Korea SciTech Premium Collection (Proquest) (PQ_SDU_P3) ProQuest Engineering Collection ProQuest Engineering Database Publicly Available Content Database ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_27631581403 |
IEDL.DBID | BENPR |
IngestDate | Thu Oct 10 16:23:07 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_27631581403 |
OpenAccessLink | https://www.proquest.com/docview/2763158140?pq-origsite=%requestingapplication% |
PQID | 2763158140 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_2763158140 |
PublicationCentury | 2000 |
PublicationDate | 20230109 |
PublicationDateYYYYMMDD | 2023-01-09 |
PublicationDate_xml | – month: 01 year: 2023 text: 20230109 day: 09 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2023 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.4416032 |
SecondaryResourceType | preprint |
Snippet | Multimodal learning can benefit from the representation power of pretrained Large Language Models (LLMs). However, state-of-the-art transformer based LLMs... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Benchmarks Transformers |
Title | MAQA: A Multimodal QA Benchmark for Negation |
URI | https://www.proquest.com/docview/2763158140 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQsTAzSEu0NDfWTQH2rHVNLJMsdC1SDYx1LSxTTQwNk81SjY1A-519_cw8Qk28IkwjoANuxdBllbAyEVxQp-Qng8bI9Y2AGcHQFHQ-k31BoS7o1ijQ7Cr0Cg1mBlYjoIwBCwOrk6tfQBB8lMXIzBzYZjbGKGjBtYebIANrQGJBapEQA1NqnjADO3jRZXKxCIOOr2Ogo5WCowJ4F2xufkpijkKgo4ITMOFk5CYWZSsAG5QKfqnp4MATZVB2cw1x9tCFWREPTQbF8QhHG4sxsAD786kSDAoWliYmBqlpqcnGSUnA_gboiDxzg9Rkk6RUE1PTZKNESQYZfCZJ4ZeWZuAC3YgOHiWwlGFgKSkqTZUF1pslSXIMzBZu7nLQIALyfOtcAcXydUc |
link.rule.ids | 783,787,12777,21400,33385,33756,43612,43817 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwY2BQsTAzSEu0NDfWTQH2rHVNLJMsdC1SDYx1LSxTTQwNk81SjY1A-519_cw8Qk28IkwjoANuxdBllbAyEVxQp-Qng8bI9Y2AGcHQFHQ-k31BoS7o1ijQ7Cr0Cg1mBlbQUVXAzherk6tfQBB8lMXIzBzYZjbGKGjBtYebIANrQGJBapEQA1NqnjADO3jRZXKxCIOOr2Ogo5WCowJ4F2xufkpijkKgo4ITMOFk5CYWZSsAG5QKfqnp4MATZVB2cw1x9tCFWREPTQbF8QhHG4sxsAD786kSDAoWliYmBqlpqcnGSUnA_gbI3eYGqckmSakmpqbJRomSDDL4TJLCLy3PwOkR4usT7-Pp5y3NwAW6HR08YmApw8BSUlSaKgusQ0uS5KABBQDwTHYq |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=MAQA%3A+A+Multimodal+QA+Benchmark+for+Negation&rft.jtitle=arXiv.org&rft.au=Judith+Yue+Li&rft.au=Jansen%2C+Aren&rft.au=Huang%2C+Qingqing&rft.au=Lee%2C+Joonseok&rft.date=2023-01-09&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |