Detecting images generated by diffusers

In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines betwe...

Full description

Saved in:
Bibliographic Details
Published inPeerJ. Computer science Vol. 10; p. e2127
Main Authors Coccomini, Davide Alessandro, Esuli, Andrea, Falchi, Fabrizio, Gennaro, Claudio, Amato, Giuseppe
Format Journal Article
LanguageEnglish
Published United States PeerJ. Ltd 10.07.2024
PeerJ Inc
Subjects
Online AccessGet full text
ISSN2376-5992
2376-5992
DOI10.7717/peerj-cs.2127

Cover

Loading…
Abstract In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers .
AbstractList In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers.
In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at:
In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers.In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers.
In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers .
ArticleNumber e2127
Audience Academic
Author Gennaro, Claudio
Amato, Giuseppe
Coccomini, Davide Alessandro
Falchi, Fabrizio
Esuli, Andrea
Author_xml – sequence: 1
  givenname: Davide Alessandro
  surname: Coccomini
  fullname: Coccomini, Davide Alessandro
  organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy, Information Engineering, University of Pisa, Pisa, Tuscany, Italy
– sequence: 2
  givenname: Andrea
  orcidid: 0000-0002-5725-4322
  surname: Esuli
  fullname: Esuli, Andrea
  organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy
– sequence: 3
  givenname: Fabrizio
  orcidid: 0000-0001-6258-5313
  surname: Falchi
  fullname: Falchi, Fabrizio
  organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy
– sequence: 4
  givenname: Claudio
  surname: Gennaro
  fullname: Gennaro, Claudio
  organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy
– sequence: 5
  givenname: Giuseppe
  surname: Amato
  fullname: Amato, Giuseppe
  organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy
BackLink https://www.ncbi.nlm.nih.gov/pubmed/39145210$$D View this record in MEDLINE/PubMed
BookMark eNp1kktv1DAQgC1UREvpkStaiQNwyOKx4zg5oarlsVIlJB5ny3HGwatsvNgOov--TrdUDQJfbI2_-TRjz1NyNPoRCXkOdC0lyLd7xLAtTFwzYPIROWFcVoVoGnb04HxMzmLcUkpBQF7NE3LMGygFA3pCXl1iQpPc2K_cTvcYVz2OGHTCbtVerzpn7RQxxGfksdVDxLO7_ZR8__D-28Wn4urzx83F-VVhhBCpENBVjNclGGZrFC01rbRQ1VboyqJtwXQVSETeakqFsbWstOyorQWWkjLgp2Rz8HZeb9U-5KLCtfLaqduAD73SITkzoDK6Q9mBZqXOyVa30DRCWN2xKptFlV3vDq791O6wMzimoIeFdHkzuh-q978UAGesqetseH1nCP7nhDGpnYsGh0GP6KeoOG04SAGiyejLA9rrXJsbrc9KM-PqvAYuaVkKkan1Pyg9t7JzJn-udTm-SHizSMhMwt-p11OMavP1y5J98bDf-0b__HYGigNggo8xoL1HgKp5oNTtQCkT1TxQmed_8cYlnZyfn8sN_8m6ATh8zoM
CitedBy_id crossref_primary_10_1111_exsy_13829
crossref_primary_10_7717_peerj_cs_2127
crossref_primary_10_1007_s00530_024_01626_z
Cites_doi 10.1109/CVPR.2017.19
10.48550/arXiv.2307.06272
10.3390/jimaging9050089
10.48550/arXiv.2010.11929
10.1016/j.patrec.2021.03.005
10.7717/peerj-cs.2127
10.1109/ICCV.2017.244
10.1007/978-3-031-06433-3_19
10.48550/arXiv.2303.05275
10.48550/arXiv.2210.06998
10.1109/CVPR.2019.00160
10.1109/OJSP.2023.3337714
10.1109/ICCV48922.2021.00632
10.1145/3665497
10.48550/arXiv.2211.10996
10.1109/CVPR42600.2020.00813
10.1145/3512732.3533587
10.3390/jimaging8100263
10.48550/arXiv.1907.11692
ContentType Journal Article
Copyright 2024 Coccomini et al.
COPYRIGHT 2024 PeerJ. Ltd.
2024 Coccomini et al. 2024 Coccomini et al.
Copyright_xml – notice: 2024 Coccomini et al.
– notice: COPYRIGHT 2024 PeerJ. Ltd.
– notice: 2024 Coccomini et al. 2024 Coccomini et al.
DBID AAYXX
CITATION
NPM
ISR
7X8
5PM
DOA
DOI 10.7717/peerj-cs.2127
DatabaseName CrossRef
PubMed
Gale In Context: Science
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle CrossRef
PubMed
MEDLINE - Academic
DatabaseTitleList

MEDLINE - Academic
PubMed

CrossRef
Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 2
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
DeliveryMethod fulltext_linktorsrc
Discipline Computer Science
EISSN 2376-5992
ExternalDocumentID oai_doaj_org_article_cade7d1a24ae47fab19955fad265cf56
PMC11322988
A813704455
39145210
10_7717_peerj_cs_2127
Genre Journal Article
GrantInformation_xml – fundername: MUR National Recovery and Resilience Plan funded by the European Union—NextGenerationEU
– fundername: AI4Media
  grantid: EC H2020 n. 951911
– fundername: SERICS
  grantid: (PE00000014) and FAIR (PE00000013)
GroupedDBID 53G
5VS
8FE
8FG
AAFWJ
AAYXX
ABUWG
ADBBV
AFKRA
AFPKN
ALMA_UNASSIGNED_HOLDINGS
ARAPS
ARCSS
AZQEC
BCNDV
BENPR
BGLVJ
BPHCQ
CCPQU
CITATION
DWQXO
FRP
GNUQQ
GROUPED_DOAJ
HCIFZ
IAO
ICD
IEA
ISR
ITC
K6V
K7-
M~E
OK1
P62
PHGZM
PHGZT
PIMPY
PQQKQ
PROAC
RPM
H13
NPM
PQGLB
PMFND
7X8
5PM
PUEGO
ID FETCH-LOGICAL-c555t-51d623841c2f8e5b0cb7f168f5a6fefb1cd617ee3ba005cf876a7d0f85e470213
IEDL.DBID DOA
ISSN 2376-5992
IngestDate Wed Aug 27 01:18:14 EDT 2025
Thu Aug 21 18:34:38 EDT 2025
Fri Jul 11 10:32:42 EDT 2025
Tue Jun 17 22:02:21 EDT 2025
Tue Jun 10 21:01:07 EDT 2025
Fri Jun 27 05:26:37 EDT 2025
Mon Jul 21 05:40:50 EDT 2025
Thu Apr 24 22:52:00 EDT 2025
Tue Jul 01 04:11:46 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Keywords Deep learning
Computer vision
Deepfake detection
Transformers
Synthetic image detection
Convolutional neural networks
Multimodal machine learning
CLIP
Language English
License https://creativecommons.org/licenses/by/4.0
2024 Coccomini et al.
This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited.
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-c555t-51d623841c2f8e5b0cb7f168f5a6fefb1cd617ee3ba005cf876a7d0f85e470213
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ORCID 0000-0002-5725-4322
0000-0001-6258-5313
OpenAccessLink https://doaj.org/article/cade7d1a24ae47fab19955fad265cf56
PMID 39145210
PQID 3093175159
PQPubID 23479
PageCount e2127
ParticipantIDs doaj_primary_oai_doaj_org_article_cade7d1a24ae47fab19955fad265cf56
pubmedcentral_primary_oai_pubmedcentral_nih_gov_11322988
proquest_miscellaneous_3093175159
gale_infotracmisc_A813704455
gale_infotracacademiconefile_A813704455
gale_incontextgauss_ISR_A813704455
pubmed_primary_39145210
crossref_primary_10_7717_peerj_cs_2127
crossref_citationtrail_10_7717_peerj_cs_2127
ProviderPackageCode CITATION
AAYXX
PublicationCentury 2000
PublicationDate 2024-07-10
PublicationDateYYYYMMDD 2024-07-10
PublicationDate_xml – month: 07
  year: 2024
  text: 2024-07-10
  day: 10
PublicationDecade 2020
PublicationPlace United States
PublicationPlace_xml – name: United States
– name: San Diego, USA
PublicationTitle PeerJ. Computer science
PublicationTitleAlternate PeerJ Comput Sci
PublicationYear 2024
Publisher PeerJ. Ltd
PeerJ Inc
Publisher_xml – name: PeerJ. Ltd
– name: PeerJ Inc
References Qiao (10.7717/peerj-cs.2127/ref-27) 2019
Rombach (10.7717/peerj-cs.2127/ref-30) 2022
Li (10.7717/peerj-cs.2127/ref-21) 2020
Bammey (10.7717/peerj-cs.2127/ref-2) 2024; 5
Baxevanakis (10.7717/peerj-cs.2127/ref-3) 2022
Karras (10.7717/peerj-cs.2127/ref-19) 2020
Zheng (10.7717/peerj-cs.2127/ref-36) 2021
Ledig (10.7717/peerj-cs.2127/ref-20) 2017
Liu (10.7717/peerj-cs.2127/ref-23) 2019
Mokhayeri (10.7717/peerj-cs.2127/ref-25) 2019
Sha (10.7717/peerj-cs.2127/ref-32) 2022
Caldelli (10.7717/peerj-cs.2127/ref-4) 2021; 146
Coccomini (10.7717/peerj-cs.2127/ref-9) 2024
Coccomini (10.7717/peerj-cs.2127/ref-7) 2022a
Ma (10.7717/peerj-cs.2127/ref-24) 2023
Coccomini (10.7717/peerj-cs.2127/ref-6) 2023a; 9
Dosovitskiy (10.7717/peerj-cs.2127/ref-15) 2020
Cozzolino (10.7717/peerj-cs.2127/ref-13) 2020
Deng (10.7717/peerj-cs.2127/ref-14) 2009
Goodfellow (10.7717/peerj-cs.2127/ref-16) 2014
Radford (10.7717/peerj-cs.2127/ref-28) 2021
Zhang (10.7717/peerj-cs.2127/ref-35) 2022
Amoroso (10.7717/peerj-cs.2127/ref-1) 2023
Zhu (10.7717/peerj-cs.2127/ref-37) 2017
Xu (10.7717/peerj-cs.2127/ref-34) 2021
Chollet (10.7717/peerj-cs.2127/ref-5) 2017
Srinivasan (10.7717/peerj-cs.2127/ref-33) 2021
Coccomini (10.7717/peerj-cs.2127/ref-10) 2022b
Ramesh (10.7717/peerj-cs.2127/ref-29) 2021
Guarnera (10.7717/peerj-cs.2127/ref-17) 2022; 8
Ruiz (10.7717/peerj-cs.2127/ref-31) 2020
He (10.7717/peerj-cs.2127/ref-18) 2016
Nichol (10.7717/peerj-cs.2127/ref-26) 2022
Lin (10.7717/peerj-cs.2127/ref-22) 2014
Coccomini (10.7717/peerj-cs.2127/ref-8) 2023b
Coccomini (10.7717/peerj-cs.2127/ref-11) 2022c
Corvi (10.7717/peerj-cs.2127/ref-12) 2022
References_xml – year: 2017
  ident: 10.7717/peerj-cs.2127/ref-20
  article-title: Photo-realistic single image super-resolution using a generative adversarial network
  doi: 10.1109/CVPR.2017.19
– volume-title: The Second Workshop on New Frontiers in Adversarial Machine Learning
  year: 2023
  ident: 10.7717/peerj-cs.2127/ref-24
  article-title: Exposing the fake: effective diffusion-generated images detection
  doi: 10.48550/arXiv.2307.06272
– volume: 9
  start-page: 89
  issue: 5
  year: 2023a
  ident: 10.7717/peerj-cs.2127/ref-6
  article-title: On the generalization of deep learning models in video deepfake detection
  publication-title: Journal of Imaging
  doi: 10.3390/jimaging9050089
– year: 2020
  ident: 10.7717/peerj-cs.2127/ref-15
  article-title: An image is worth 16x16 words: transformers for image recognition at scale
  doi: 10.48550/arXiv.2010.11929
– start-page: 5073
  year: 2020
  ident: 10.7717/peerj-cs.2127/ref-21
  article-title: Advancing high fidelity identity swapping for forgery detection
– volume: 146
  start-page: 31
  issue: 10
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-4
  article-title: Optical flow based cnn for detection of unlearnt deepfake manipulations
  publication-title: Pattern Recognition Letters
  doi: 10.1016/j.patrec.2021.03.005
– start-page: 52
  year: 2022a
  ident: 10.7717/peerj-cs.2127/ref-7
  article-title: Cross-forgery analysis of vision transformers and cnns for deepfake image detection
– start-page: 8748
  volume-title: Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-28
  article-title: Learning transferable visual models from natural language supervision
– year: 2024
  ident: 10.7717/peerj-cs.2127/ref-9
  article-title: Detecting images generated by diffusers
  doi: 10.7717/peerj-cs.2127
– year: 2020
  ident: 10.7717/peerj-cs.2127/ref-31
  article-title: Morphgan: one-shot face synthesis gan for detecting recognition bias
– year: 2017
  ident: 10.7717/peerj-cs.2127/ref-37
  article-title: Unpaired image-to-image translation using cycle-consistent adversarial networks
  doi: 10.1109/ICCV.2017.244
– year: 2019
  ident: 10.7717/peerj-cs.2127/ref-25
  article-title: Cross-domain face synthesis using a controllable gan
– year: 2022
  ident: 10.7717/peerj-cs.2127/ref-26
  article-title: Glide: towards photorealistic image generation and editing with text-guided diffusion models
– start-page: 219
  volume-title: Image Analysis and Processing—ICIAP 2022
  year: 2022b
  ident: 10.7717/peerj-cs.2127/ref-10
  article-title: Combining efficientnet and vision transformers for video deepfake detection
  doi: 10.1007/978-3-031-06433-3_19
– year: 2022
  ident: 10.7717/peerj-cs.2127/ref-12
  article-title: On the detection of synthetic images generated by diffusion models
  publication-title: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
– start-page: 740
  year: 2014
  ident: 10.7717/peerj-cs.2127/ref-22
  article-title: Microsoft coco: common objects in context
– year: 2023b
  ident: 10.7717/peerj-cs.2127/ref-8
  article-title: Detecting images generated by diffusers
  doi: 10.48550/arXiv.2303.05275
– start-page: 10684
  year: 2022
  ident: 10.7717/peerj-cs.2127/ref-30
  article-title: High-resolution image synthesis with latent diffusion models
– year: 2022
  ident: 10.7717/peerj-cs.2127/ref-32
  article-title: De-fake: detection and attribution of fake images generated by text-to-image diffusion models
  doi: 10.48550/arXiv.2210.06998
– start-page: 15088
  year: 2020
  ident: 10.7717/peerj-cs.2127/ref-13
  article-title: Id-reveal: identity-aware deepfake video detection
– year: 2019
  ident: 10.7717/peerj-cs.2127/ref-27
  article-title: Mirrorgan: learning text-to-image generation by redescription
  doi: 10.1109/CVPR.2019.00160
– volume: 5
  start-page: 1
  year: 2024
  ident: 10.7717/peerj-cs.2127/ref-2
  article-title: Synthbuster: towards detection of diffusion model generated images
  publication-title: IEEE Open Journal of Signal Processing
  doi: 10.1109/OJSP.2023.3337714
– start-page: 2672
  volume-title: Advances in Neural Information Processing Systems
  year: 2014
  ident: 10.7717/peerj-cs.2127/ref-16
  article-title: Generative adversarial nets
– start-page: 248
  year: 2009
  ident: 10.7717/peerj-cs.2127/ref-14
  article-title: Imagenet: a large-scale hierarchical image database
– start-page: 6363
  volume-title: 2021 IEEE/CVF International Conference on Computer Vision (ICCV)
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-34
  article-title: Drb-gan: a dynamic resblock generative adversarial network for artistic style transfer
  doi: 10.1109/ICCV48922.2021.00632
– volume-title: ACM Transactions on Multimedia Computing, Communications, and Applications
  year: 2023
  ident: 10.7717/peerj-cs.2127/ref-1
  article-title: Parents and children: distinguishing multimodal deepfakes from natural images
  doi: 10.1145/3665497
– year: 2022c
  ident: 10.7717/peerj-cs.2127/ref-11
  article-title: Mintime: multi-identity size-invariant video deepfake detection
  doi: 10.48550/arXiv.2211.10996
– start-page: 8107
  volume-title: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
  year: 2020
  ident: 10.7717/peerj-cs.2127/ref-19
  article-title: Analyzing and improving the image quality of stylegan
  doi: 10.1109/CVPR42600.2020.00813
– year: 2022
  ident: 10.7717/peerj-cs.2127/ref-3
  article-title: The mever deepfake detection service: lessons learnt from developing and deploying in the wild
  doi: 10.1145/3512732.3533587
– start-page: 770
  year: 2016
  ident: 10.7717/peerj-cs.2127/ref-18
  article-title: Deep residual learning for image recognition
– start-page: 8821
  volume-title: Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-29
  article-title: Zero-shot text-to-image generation
– start-page: 2443
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-33
  article-title: Wit: Wikipedia-based image text dataset for multimodal multilingual machine learning
– start-page: 15024
  volume-title: ICCV
  year: 2021
  ident: 10.7717/peerj-cs.2127/ref-36
  article-title: Exploring temporal coherence for more general video face forgery detection
– volume: 8
  start-page: 263
  issue: 10
  year: 2022
  ident: 10.7717/peerj-cs.2127/ref-17
  article-title: The face deepfake detection challenge
  publication-title: Journal of Imaging
  doi: 10.3390/jimaging8100263
– year: 2019
  ident: 10.7717/peerj-cs.2127/ref-23
  article-title: Roberta: a robustly optimized bert pretraining approach
  doi: 10.48550/arXiv.1907.11692
– start-page: 211
  volume-title: SPM7’22
  year: 2022
  ident: 10.7717/peerj-cs.2127/ref-35
  article-title: Face forgery detection of deepfake based on multiple convolutional neural networks
– start-page: 1800
  year: 2017
  ident: 10.7717/peerj-cs.2127/ref-5
  article-title: Xception: deep learning with depthwise separable convolutions
SSID ssj0001511119
Score 2.2753427
Snippet In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep...
SourceID doaj
pubmedcentral
proquest
gale
pubmed
crossref
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
Enrichment Source
StartPage e2127
SubjectTerms Artificial Intelligence
CLIP
Computational linguistics
Computer Vision
Data Mining and Machine Learning
Deep learning
Deepfake detection
Language processing
Multimodal machine learning
Natural language interfaces
Neural Networks
Synthetic image detection
Title Detecting images generated by diffusers
URI https://www.ncbi.nlm.nih.gov/pubmed/39145210
https://www.proquest.com/docview/3093175159
https://pubmed.ncbi.nlm.nih.gov/PMC11322988
https://doaj.org/article/cade7d1a24ae47fab19955fad265cf56
Volume 10
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1Lb9QwEB6VcuEClFdT2lVAFb0QGid27Bxb6Lb0UKFCpd4s27H7EM1Wze6Bf89Mkl0lQogL13gO9tgz840z_gZglxhESuaLhDuLCYpIeaJsmieZU1Z4IYlghaotzoqTC356KS4Hrb6oJqyjB-4Ut09V4rJiJuPGcxmMpTfFIpgqK4QLoiXbxpg3SKa698HkCsqOVFNiyrJ_7_3DbeKaT1nbQWYQhFqu_j898iAkjcslB_Fn-hye9sAxPugmvAFrvn4Bz5ZNGeLeRl_C3hdP_wUwIsU3d-gsmviqZZZGZBnbXzE1RKGLieYVXEyPfnw-SfpuCIkTQswTwSqEKoozlwXlhU2dlYEVKghTBB8scxWiEe9za9CyXEA3Z2SVBiVQbxjJ89ewXs9qvwlxQMNFZF1am-Xc-sxWQhplDMEpmXMewcelerTrqcKpY8VPjSkDaVO32tSu0aTNCD6sxO87joy_CR6SrldCRG3dfsAN1_2G639teATvaac0kVfUVB1zZRZNo79-P9cHiuUy5VyICPZ6oTDDmTvTPzbA9RPf1UhyeySJ1uVGw--WB0LTEJWk1X62aDT9QkbshXAwgjfdAVktLC8ZR1yURqBGR2e08vFIfXPdknszuh4oldr6H7p6C08yBGF0F83SbVifPyz8DoKouZ3AIzU9nsDjw6Ozb-eT1np-AyY3HsM
linkProvider Directory of Open Access Journals
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Detecting+images+generated+by+diffusers&rft.jtitle=PeerJ.+Computer+science&rft.au=Coccomini%2C+Davide+Alessandro&rft.au=Esuli%2C+Andrea&rft.au=Falchi%2C+Fabrizio&rft.au=Gennaro%2C+Claudio&rft.date=2024-07-10&rft.pub=PeerJ.+Ltd&rft.issn=2376-5992&rft.eissn=2376-5992&rft.volume=10&rft.spage=e2127&rft_id=info:doi/10.7717%2Fpeerj-cs.2127&rft.externalDBID=ISR&rft.externalDocID=A813704455
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2376-5992&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2376-5992&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2376-5992&client=summon