DALD: Improving Logits-based Detector without Logits from Black-box LLMs
The advent of Large Language Models (LLMs) has revolutionized text generation, producing outputs that closely mimic human writing. This blurring of lines between machine- and human-written text presents new challenges in distinguishing one from the other a task further complicated by the frequent up...
Saved in:
Published in | arXiv.org |
---|---|
Main Authors | , , , , , , , , , |
Format | Paper |
Language | English |
Published |
Ithaca
Cornell University Library, arXiv.org
27.10.2024
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | The advent of Large Language Models (LLMs) has revolutionized text generation, producing outputs that closely mimic human writing. This blurring of lines between machine- and human-written text presents new challenges in distinguishing one from the other a task further complicated by the frequent updates and closed nature of leading proprietary LLMs. Traditional logits-based detection methods leverage surrogate models for identifying LLM-generated content when the exact logits are unavailable from black-box LLMs. However, these methods grapple with the misalignment between the distributions of the surrogate and the often undisclosed target models, leading to performance degradation, particularly with the introduction of new, closed-source models. Furthermore, while current methodologies are generally effective when the source model is identified, they falter in scenarios where the model version remains unknown, or the test set comprises outputs from various source models. To address these limitations, we present Distribution-Aligned LLMs Detection (DALD), an innovative framework that redefines the state-of-the-art performance in black-box text detection even without logits from source LLMs. DALD is designed to align the surrogate model's distribution with that of unknown target LLMs, ensuring enhanced detection capability and resilience against rapid model iterations with minimal training investment. By leveraging corpus samples from publicly accessible outputs of advanced models such as ChatGPT, GPT-4 and Claude-3, DALD fine-tunes surrogate models to synchronize with unknown source model distributions effectively. |
---|---|
AbstractList | The advent of Large Language Models (LLMs) has revolutionized text generation, producing outputs that closely mimic human writing. This blurring of lines between machine- and human-written text presents new challenges in distinguishing one from the other a task further complicated by the frequent updates and closed nature of leading proprietary LLMs. Traditional logits-based detection methods leverage surrogate models for identifying LLM-generated content when the exact logits are unavailable from black-box LLMs. However, these methods grapple with the misalignment between the distributions of the surrogate and the often undisclosed target models, leading to performance degradation, particularly with the introduction of new, closed-source models. Furthermore, while current methodologies are generally effective when the source model is identified, they falter in scenarios where the model version remains unknown, or the test set comprises outputs from various source models. To address these limitations, we present Distribution-Aligned LLMs Detection (DALD), an innovative framework that redefines the state-of-the-art performance in black-box text detection even without logits from source LLMs. DALD is designed to align the surrogate model's distribution with that of unknown target LLMs, ensuring enhanced detection capability and resilience against rapid model iterations with minimal training investment. By leveraging corpus samples from publicly accessible outputs of advanced models such as ChatGPT, GPT-4 and Claude-3, DALD fine-tunes surrogate models to synchronize with unknown source model distributions effectively. |
Author | Zeng, Cong Yang, Xianjun Chen, Yuanzhou Sun, Yiyou Chen, Haifeng Xu, Zhiqiang Tang, Shengkun Cheng, Wei Xu, Dongkuan Yao, Li |
Author_xml | – sequence: 1 givenname: Cong surname: Zeng fullname: Zeng, Cong – sequence: 2 givenname: Shengkun surname: Tang fullname: Tang, Shengkun – sequence: 3 givenname: Xianjun surname: Yang fullname: Yang, Xianjun – sequence: 4 givenname: Yuanzhou surname: Chen fullname: Chen, Yuanzhou – sequence: 5 givenname: Yiyou surname: Sun fullname: Sun, Yiyou – sequence: 6 givenname: Zhiqiang surname: Xu fullname: Xu, Zhiqiang – sequence: 7 givenname: Li surname: Yao fullname: Yao, Li – sequence: 8 givenname: Haifeng surname: Chen fullname: Chen, Haifeng – sequence: 9 givenname: Wei surname: Cheng fullname: Cheng, Wei – sequence: 10 givenname: Dongkuan surname: Xu fullname: Xu, Dongkuan |
BookMark | eNqNjLEOgjAUABujiaj8QxPnJrWVYtxUNJjUzZ0AFgShT9uifr4MfIDTDXe5GRpr0GqEPMb5imzWjE2Rb21NKWUiZEHAPRRHOxlt8bl9GnhXusQSyspZkqVW3XCknModGPyp3B06N1hcGGjxvknzB8ngi6W82AWaFGljlT9wjpan4_UQk3786pR1SQ2d0b1KOBUhZVRwwf-rfgU0PJQ |
ContentType | Paper |
Copyright | 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
Copyright_xml | – notice: 2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. |
DBID | 8FE 8FG ABJCF ABUWG AFKRA AZQEC BENPR BGLVJ CCPQU DWQXO HCIFZ L6V M7S PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
DatabaseName | ProQuest SciTech Collection ProQuest Technology Collection Materials Science & Engineering Collection ProQuest Central (Alumni) ProQuest Central ProQuest Central Essentials ProQuest Central Technology Collection ProQuest One Community College ProQuest Central Korea SciTech Premium Collection ProQuest Engineering Collection Engineering Database Publicly Available Content Database ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China Engineering Collection |
DatabaseTitle | Publicly Available Content Database Engineering Database Technology Collection ProQuest Central Essentials ProQuest One Academic Eastern Edition ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Technology Collection ProQuest SciTech Collection ProQuest Central China ProQuest Central ProQuest Engineering Collection ProQuest One Academic UKI Edition ProQuest Central Korea Materials Science & Engineering Collection ProQuest One Academic Engineering Collection |
DatabaseTitleList | Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: 8FG name: ProQuest Technology Collection url: https://search.proquest.com/technologycollection1 sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Physics |
EISSN | 2331-8422 |
Genre | Working Paper/Pre-Print |
GroupedDBID | 8FE 8FG ABJCF ABUWG AFKRA ALMA_UNASSIGNED_HOLDINGS AZQEC BENPR BGLVJ CCPQU DWQXO FRJ HCIFZ L6V M7S M~E PIMPY PQEST PQQKQ PQUKI PRINS PTHSS |
ID | FETCH-proquest_journals_30670206363 |
IEDL.DBID | BENPR |
IngestDate | Wed Oct 30 05:17:50 EDT 2024 |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-proquest_journals_30670206363 |
OpenAccessLink | https://www.proquest.com/docview/3067020636?pq-origsite=%requestingapplication% |
PQID | 3067020636 |
PQPubID | 2050157 |
ParticipantIDs | proquest_journals_3067020636 |
PublicationCentury | 2000 |
PublicationDate | 20241027 |
PublicationDateYYYYMMDD | 2024-10-27 |
PublicationDate_xml | – month: 10 year: 2024 text: 20241027 day: 27 |
PublicationDecade | 2020 |
PublicationPlace | Ithaca |
PublicationPlace_xml | – name: Ithaca |
PublicationTitle | arXiv.org |
PublicationYear | 2024 |
Publisher | Cornell University Library, arXiv.org |
Publisher_xml | – name: Cornell University Library, arXiv.org |
SSID | ssj0002672553 |
Score | 3.5679984 |
SecondaryResourceType | preprint |
Snippet | The advent of Large Language Models (LLMs) has revolutionized text generation, producing outputs that closely mimic human writing. This blurring of lines... |
SourceID | proquest |
SourceType | Aggregation Database |
SubjectTerms | Black boxes Blurring Large language models Misalignment Performance degradation Target detection |
Title | DALD: Improving Logits-based Detector without Logits from Black-box LLMs |
URI | https://www.proquest.com/docview/3067020636 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV3PS8MwFH64FsGbP_HHHAG9BkuStZ0XUdtapB1DFHYbTft6tHPtwJN_u3klmwdhx_AgISF83_u-l_AAbnUha68ge6nSmhu-ljxUJXIdKBSFqo0iIGsgn_rph3qdj-fWcGvts8oNJvZAXTUleeR3lNqa1MaX_sPyi1PXKKqu2hYaA3CFUQqeA-5TPJ29bV0W4QcmZ5b_gLZnj-QQ3FmxxNUR7OHnMez3jy7L9gTS6DGL7tlW1zNqfNy1nJilYhF2vaPOyCpt1p2NMvoQwnrfjevmm2VZ3p7CTRK_P6d8s_zCXpF28bcheQaO0fp4Dkx5WtSqQBz7lQpLb4I1IlJ10gjGYiIvYLhrpsvd4Ss4EIaTCXpFMASnW63x2nBqp0cwCJOXkT0-M8p_4l8GDoDs |
link.rule.ids | 783,787,12777,21400,33385,33756,43612,43817 |
linkProvider | ProQuest |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1NS8QwEB10F9Gbn_ixakCvwZJk210vItZaNV08rLC3krTTo123XfDnmwnd9SDseSAhIbw3702GAbi1RlaBIXuptJY7vpZ8pArkNlIojKqcIiBrIJuE6ad6mw1nneHWdN8qV5jogbqsC_LI7yi1dalNKMOH-TenqVFUXe1GaGxDX0nH1dQpnrysPRYRRi5jlv9g1nNHsg_9DzPHxQFs4dch7Pgvl0VzBGn8qON7tlb1jMYetw0nXilZjK330xkZpfWy7aKM2kGYd924rX-Y1llzDDfJ8_Qp5avt8-6BNPnfceQJ9JzSx1NgKrCiUgZxGJZqVARjrBCRapNOLpqxPIPBppXON4evYTedZjrXr5P3C9gTjp0JhEU0gF67WOKlY9fWXvkr_AX3PYBg |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=DALD%3A+Improving+Logits-based+Detector+without+Logits+from+Black-box+LLMs&rft.jtitle=arXiv.org&rft.au=Zeng%2C+Cong&rft.au=Tang%2C+Shengkun&rft.au=Yang%2C+Xianjun&rft.au=Chen%2C+Yuanzhou&rft.date=2024-10-27&rft.pub=Cornell+University+Library%2C+arXiv.org&rft.eissn=2331-8422 |