Detecting images generated by diffusers
In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines betwe...
Saved in:
Published in | PeerJ. Computer science Vol. 10; p. e2127 |
---|---|
Main Authors | , , , , |
Format | Journal Article |
Language | English |
Published |
United States
PeerJ. Ltd
10.07.2024
PeerJ Inc |
Subjects | |
Online Access | Get full text |
ISSN | 2376-5992 2376-5992 |
DOI | 10.7717/peerj-cs.2127 |
Cover
Loading…
Abstract | In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at:
https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers
. |
---|---|
AbstractList | In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers. In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers.In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers. In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep learning techniques. These synthetic images, often created through complex algorithms, closely mimic real photographs, blurring the lines between reality and artificiality. This proliferation of synthetic visuals presents a pressing challenge: how to accurately and reliably distinguish between genuine and generated images. This article, in particular, explores the task of detecting images generated by text-to-image diffusion models, highlighting the challenges and peculiarities of this field. To evaluate this, we consider images generated from captions in the MSCOCO and Wikimedia datasets using two state-of-the-art models: Stable Diffusion and GLIDE. Our experiments show that it is possible to detect the generated images using simple multi-layer perceptrons (MLPs), starting from features extracted by CLIP or RoBERTa, or using traditional convolutional neural networks (CNNs). These latter models achieve remarkable performances in particular when pretrained on large datasets. We also observe that models trained on images generated by Stable Diffusion can occasionally detect images generated by GLIDE, but only on the MSCOCO dataset. However, the reverse is not true. Lastly, we find that incorporating the associated textual information with the images in some cases can lead to a better generalization capability, especially if textual features are closely related to visual ones. We also discovered that the type of subject depicted in the image can significantly impact performance. This work provides insights into the feasibility of detecting generated images and has implications for security and privacy concerns in real-world applications. The code to reproduce our results is available at: https://github.com/davide-coccomini/Detecting-Images-Generated-by-Diffusers . |
ArticleNumber | e2127 |
Audience | Academic |
Author | Gennaro, Claudio Amato, Giuseppe Coccomini, Davide Alessandro Falchi, Fabrizio Esuli, Andrea |
Author_xml | – sequence: 1 givenname: Davide Alessandro surname: Coccomini fullname: Coccomini, Davide Alessandro organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy, Information Engineering, University of Pisa, Pisa, Tuscany, Italy – sequence: 2 givenname: Andrea orcidid: 0000-0002-5725-4322 surname: Esuli fullname: Esuli, Andrea organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy – sequence: 3 givenname: Fabrizio orcidid: 0000-0001-6258-5313 surname: Falchi fullname: Falchi, Fabrizio organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy – sequence: 4 givenname: Claudio surname: Gennaro fullname: Gennaro, Claudio organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy – sequence: 5 givenname: Giuseppe surname: Amato fullname: Amato, Giuseppe organization: Institute of Information Science and Technologies “Alessandro Faedo”, Italian National Research Council, Pisa, Tuscany, Italy |
BackLink | https://www.ncbi.nlm.nih.gov/pubmed/39145210$$D View this record in MEDLINE/PubMed |
BookMark | eNp1kktv1DAQgC1UREvpkStaiQNwyOKx4zg5oarlsVIlJB5ny3HGwatsvNgOov--TrdUDQJfbI2_-TRjz1NyNPoRCXkOdC0lyLd7xLAtTFwzYPIROWFcVoVoGnb04HxMzmLcUkpBQF7NE3LMGygFA3pCXl1iQpPc2K_cTvcYVz2OGHTCbtVerzpn7RQxxGfksdVDxLO7_ZR8__D-28Wn4urzx83F-VVhhBCpENBVjNclGGZrFC01rbRQ1VboyqJtwXQVSETeakqFsbWstOyorQWWkjLgp2Rz8HZeb9U-5KLCtfLaqduAD73SITkzoDK6Q9mBZqXOyVa30DRCWN2xKptFlV3vDq791O6wMzimoIeFdHkzuh-q978UAGesqetseH1nCP7nhDGpnYsGh0GP6KeoOG04SAGiyejLA9rrXJsbrc9KM-PqvAYuaVkKkan1Pyg9t7JzJn-udTm-SHizSMhMwt-p11OMavP1y5J98bDf-0b__HYGigNggo8xoL1HgKp5oNTtQCkT1TxQmed_8cYlnZyfn8sN_8m6ATh8zoM |
CitedBy_id | crossref_primary_10_1111_exsy_13829 crossref_primary_10_7717_peerj_cs_2127 crossref_primary_10_1007_s00530_024_01626_z |
Cites_doi | 10.1109/CVPR.2017.19 10.48550/arXiv.2307.06272 10.3390/jimaging9050089 10.48550/arXiv.2010.11929 10.1016/j.patrec.2021.03.005 10.7717/peerj-cs.2127 10.1109/ICCV.2017.244 10.1007/978-3-031-06433-3_19 10.48550/arXiv.2303.05275 10.48550/arXiv.2210.06998 10.1109/CVPR.2019.00160 10.1109/OJSP.2023.3337714 10.1109/ICCV48922.2021.00632 10.1145/3665497 10.48550/arXiv.2211.10996 10.1109/CVPR42600.2020.00813 10.1145/3512732.3533587 10.3390/jimaging8100263 10.48550/arXiv.1907.11692 |
ContentType | Journal Article |
Copyright | 2024 Coccomini et al. COPYRIGHT 2024 PeerJ. Ltd. 2024 Coccomini et al. 2024 Coccomini et al. |
Copyright_xml | – notice: 2024 Coccomini et al. – notice: COPYRIGHT 2024 PeerJ. Ltd. – notice: 2024 Coccomini et al. 2024 Coccomini et al. |
DBID | AAYXX CITATION NPM ISR 7X8 5PM DOA |
DOI | 10.7717/peerj-cs.2127 |
DatabaseName | CrossRef PubMed Gale In Context: Science MEDLINE - Academic PubMed Central (Full Participant titles) DOAJ Directory of Open Access Journals |
DatabaseTitle | CrossRef PubMed MEDLINE - Academic |
DatabaseTitleList | MEDLINE - Academic PubMed CrossRef |
Database_xml | – sequence: 1 dbid: DOA name: DOAJ Directory of Open Access Journals url: https://www.doaj.org/ sourceTypes: Open Website – sequence: 2 dbid: NPM name: PubMed url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed sourceTypes: Index Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Computer Science |
EISSN | 2376-5992 |
ExternalDocumentID | oai_doaj_org_article_cade7d1a24ae47fab19955fad265cf56 PMC11322988 A813704455 39145210 10_7717_peerj_cs_2127 |
Genre | Journal Article |
GrantInformation_xml | – fundername: MUR National Recovery and Resilience Plan funded by the European Union—NextGenerationEU – fundername: AI4Media grantid: EC H2020 n. 951911 – fundername: SERICS grantid: (PE00000014) and FAIR (PE00000013) |
GroupedDBID | 53G 5VS 8FE 8FG AAFWJ AAYXX ABUWG ADBBV AFKRA AFPKN ALMA_UNASSIGNED_HOLDINGS ARAPS ARCSS AZQEC BCNDV BENPR BGLVJ BPHCQ CCPQU CITATION DWQXO FRP GNUQQ GROUPED_DOAJ HCIFZ IAO ICD IEA ISR ITC K6V K7- M~E OK1 P62 PHGZM PHGZT PIMPY PQQKQ PROAC RPM H13 NPM PQGLB PMFND 7X8 5PM PUEGO |
ID | FETCH-LOGICAL-c555t-51d623841c2f8e5b0cb7f168f5a6fefb1cd617ee3ba005cf876a7d0f85e470213 |
IEDL.DBID | DOA |
ISSN | 2376-5992 |
IngestDate | Wed Aug 27 01:18:14 EDT 2025 Thu Aug 21 18:34:38 EDT 2025 Fri Jul 11 10:32:42 EDT 2025 Tue Jun 17 22:02:21 EDT 2025 Tue Jun 10 21:01:07 EDT 2025 Fri Jun 27 05:26:37 EDT 2025 Mon Jul 21 05:40:50 EDT 2025 Thu Apr 24 22:52:00 EDT 2025 Tue Jul 01 04:11:46 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Keywords | Deep learning Computer vision Deepfake detection Transformers Synthetic image detection Convolutional neural networks Multimodal machine learning CLIP |
Language | English |
License | https://creativecommons.org/licenses/by/4.0 2024 Coccomini et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c555t-51d623841c2f8e5b0cb7f168f5a6fefb1cd617ee3ba005cf876a7d0f85e470213 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 |
ORCID | 0000-0002-5725-4322 0000-0001-6258-5313 |
OpenAccessLink | https://doaj.org/article/cade7d1a24ae47fab19955fad265cf56 |
PMID | 39145210 |
PQID | 3093175159 |
PQPubID | 23479 |
PageCount | e2127 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_cade7d1a24ae47fab19955fad265cf56 pubmedcentral_primary_oai_pubmedcentral_nih_gov_11322988 proquest_miscellaneous_3093175159 gale_infotracmisc_A813704455 gale_infotracacademiconefile_A813704455 gale_incontextgauss_ISR_A813704455 pubmed_primary_39145210 crossref_primary_10_7717_peerj_cs_2127 crossref_citationtrail_10_7717_peerj_cs_2127 |
ProviderPackageCode | CITATION AAYXX |
PublicationCentury | 2000 |
PublicationDate | 2024-07-10 |
PublicationDateYYYYMMDD | 2024-07-10 |
PublicationDate_xml | – month: 07 year: 2024 text: 2024-07-10 day: 10 |
PublicationDecade | 2020 |
PublicationPlace | United States |
PublicationPlace_xml | – name: United States – name: San Diego, USA |
PublicationTitle | PeerJ. Computer science |
PublicationTitleAlternate | PeerJ Comput Sci |
PublicationYear | 2024 |
Publisher | PeerJ. Ltd PeerJ Inc |
Publisher_xml | – name: PeerJ. Ltd – name: PeerJ Inc |
References | Qiao (10.7717/peerj-cs.2127/ref-27) 2019 Rombach (10.7717/peerj-cs.2127/ref-30) 2022 Li (10.7717/peerj-cs.2127/ref-21) 2020 Bammey (10.7717/peerj-cs.2127/ref-2) 2024; 5 Baxevanakis (10.7717/peerj-cs.2127/ref-3) 2022 Karras (10.7717/peerj-cs.2127/ref-19) 2020 Zheng (10.7717/peerj-cs.2127/ref-36) 2021 Ledig (10.7717/peerj-cs.2127/ref-20) 2017 Liu (10.7717/peerj-cs.2127/ref-23) 2019 Mokhayeri (10.7717/peerj-cs.2127/ref-25) 2019 Sha (10.7717/peerj-cs.2127/ref-32) 2022 Caldelli (10.7717/peerj-cs.2127/ref-4) 2021; 146 Coccomini (10.7717/peerj-cs.2127/ref-9) 2024 Coccomini (10.7717/peerj-cs.2127/ref-7) 2022a Ma (10.7717/peerj-cs.2127/ref-24) 2023 Coccomini (10.7717/peerj-cs.2127/ref-6) 2023a; 9 Dosovitskiy (10.7717/peerj-cs.2127/ref-15) 2020 Cozzolino (10.7717/peerj-cs.2127/ref-13) 2020 Deng (10.7717/peerj-cs.2127/ref-14) 2009 Goodfellow (10.7717/peerj-cs.2127/ref-16) 2014 Radford (10.7717/peerj-cs.2127/ref-28) 2021 Zhang (10.7717/peerj-cs.2127/ref-35) 2022 Amoroso (10.7717/peerj-cs.2127/ref-1) 2023 Zhu (10.7717/peerj-cs.2127/ref-37) 2017 Xu (10.7717/peerj-cs.2127/ref-34) 2021 Chollet (10.7717/peerj-cs.2127/ref-5) 2017 Srinivasan (10.7717/peerj-cs.2127/ref-33) 2021 Coccomini (10.7717/peerj-cs.2127/ref-10) 2022b Ramesh (10.7717/peerj-cs.2127/ref-29) 2021 Guarnera (10.7717/peerj-cs.2127/ref-17) 2022; 8 Ruiz (10.7717/peerj-cs.2127/ref-31) 2020 He (10.7717/peerj-cs.2127/ref-18) 2016 Nichol (10.7717/peerj-cs.2127/ref-26) 2022 Lin (10.7717/peerj-cs.2127/ref-22) 2014 Coccomini (10.7717/peerj-cs.2127/ref-8) 2023b Coccomini (10.7717/peerj-cs.2127/ref-11) 2022c Corvi (10.7717/peerj-cs.2127/ref-12) 2022 |
References_xml | – year: 2017 ident: 10.7717/peerj-cs.2127/ref-20 article-title: Photo-realistic single image super-resolution using a generative adversarial network doi: 10.1109/CVPR.2017.19 – volume-title: The Second Workshop on New Frontiers in Adversarial Machine Learning year: 2023 ident: 10.7717/peerj-cs.2127/ref-24 article-title: Exposing the fake: effective diffusion-generated images detection doi: 10.48550/arXiv.2307.06272 – volume: 9 start-page: 89 issue: 5 year: 2023a ident: 10.7717/peerj-cs.2127/ref-6 article-title: On the generalization of deep learning models in video deepfake detection publication-title: Journal of Imaging doi: 10.3390/jimaging9050089 – year: 2020 ident: 10.7717/peerj-cs.2127/ref-15 article-title: An image is worth 16x16 words: transformers for image recognition at scale doi: 10.48550/arXiv.2010.11929 – start-page: 5073 year: 2020 ident: 10.7717/peerj-cs.2127/ref-21 article-title: Advancing high fidelity identity swapping for forgery detection – volume: 146 start-page: 31 issue: 10 year: 2021 ident: 10.7717/peerj-cs.2127/ref-4 article-title: Optical flow based cnn for detection of unlearnt deepfake manipulations publication-title: Pattern Recognition Letters doi: 10.1016/j.patrec.2021.03.005 – start-page: 52 year: 2022a ident: 10.7717/peerj-cs.2127/ref-7 article-title: Cross-forgery analysis of vision transformers and cnns for deepfake image detection – start-page: 8748 volume-title: Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research year: 2021 ident: 10.7717/peerj-cs.2127/ref-28 article-title: Learning transferable visual models from natural language supervision – year: 2024 ident: 10.7717/peerj-cs.2127/ref-9 article-title: Detecting images generated by diffusers doi: 10.7717/peerj-cs.2127 – year: 2020 ident: 10.7717/peerj-cs.2127/ref-31 article-title: Morphgan: one-shot face synthesis gan for detecting recognition bias – year: 2017 ident: 10.7717/peerj-cs.2127/ref-37 article-title: Unpaired image-to-image translation using cycle-consistent adversarial networks doi: 10.1109/ICCV.2017.244 – year: 2019 ident: 10.7717/peerj-cs.2127/ref-25 article-title: Cross-domain face synthesis using a controllable gan – year: 2022 ident: 10.7717/peerj-cs.2127/ref-26 article-title: Glide: towards photorealistic image generation and editing with text-guided diffusion models – start-page: 219 volume-title: Image Analysis and Processing—ICIAP 2022 year: 2022b ident: 10.7717/peerj-cs.2127/ref-10 article-title: Combining efficientnet and vision transformers for video deepfake detection doi: 10.1007/978-3-031-06433-3_19 – year: 2022 ident: 10.7717/peerj-cs.2127/ref-12 article-title: On the detection of synthetic images generated by diffusion models publication-title: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) – start-page: 740 year: 2014 ident: 10.7717/peerj-cs.2127/ref-22 article-title: Microsoft coco: common objects in context – year: 2023b ident: 10.7717/peerj-cs.2127/ref-8 article-title: Detecting images generated by diffusers doi: 10.48550/arXiv.2303.05275 – start-page: 10684 year: 2022 ident: 10.7717/peerj-cs.2127/ref-30 article-title: High-resolution image synthesis with latent diffusion models – year: 2022 ident: 10.7717/peerj-cs.2127/ref-32 article-title: De-fake: detection and attribution of fake images generated by text-to-image diffusion models doi: 10.48550/arXiv.2210.06998 – start-page: 15088 year: 2020 ident: 10.7717/peerj-cs.2127/ref-13 article-title: Id-reveal: identity-aware deepfake video detection – year: 2019 ident: 10.7717/peerj-cs.2127/ref-27 article-title: Mirrorgan: learning text-to-image generation by redescription doi: 10.1109/CVPR.2019.00160 – volume: 5 start-page: 1 year: 2024 ident: 10.7717/peerj-cs.2127/ref-2 article-title: Synthbuster: towards detection of diffusion model generated images publication-title: IEEE Open Journal of Signal Processing doi: 10.1109/OJSP.2023.3337714 – start-page: 2672 volume-title: Advances in Neural Information Processing Systems year: 2014 ident: 10.7717/peerj-cs.2127/ref-16 article-title: Generative adversarial nets – start-page: 248 year: 2009 ident: 10.7717/peerj-cs.2127/ref-14 article-title: Imagenet: a large-scale hierarchical image database – start-page: 6363 volume-title: 2021 IEEE/CVF International Conference on Computer Vision (ICCV) year: 2021 ident: 10.7717/peerj-cs.2127/ref-34 article-title: Drb-gan: a dynamic resblock generative adversarial network for artistic style transfer doi: 10.1109/ICCV48922.2021.00632 – volume-title: ACM Transactions on Multimedia Computing, Communications, and Applications year: 2023 ident: 10.7717/peerj-cs.2127/ref-1 article-title: Parents and children: distinguishing multimodal deepfakes from natural images doi: 10.1145/3665497 – year: 2022c ident: 10.7717/peerj-cs.2127/ref-11 article-title: Mintime: multi-identity size-invariant video deepfake detection doi: 10.48550/arXiv.2211.10996 – start-page: 8107 volume-title: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) year: 2020 ident: 10.7717/peerj-cs.2127/ref-19 article-title: Analyzing and improving the image quality of stylegan doi: 10.1109/CVPR42600.2020.00813 – year: 2022 ident: 10.7717/peerj-cs.2127/ref-3 article-title: The mever deepfake detection service: lessons learnt from developing and deploying in the wild doi: 10.1145/3512732.3533587 – start-page: 770 year: 2016 ident: 10.7717/peerj-cs.2127/ref-18 article-title: Deep residual learning for image recognition – start-page: 8821 volume-title: Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research year: 2021 ident: 10.7717/peerj-cs.2127/ref-29 article-title: Zero-shot text-to-image generation – start-page: 2443 year: 2021 ident: 10.7717/peerj-cs.2127/ref-33 article-title: Wit: Wikipedia-based image text dataset for multimodal multilingual machine learning – start-page: 15024 volume-title: ICCV year: 2021 ident: 10.7717/peerj-cs.2127/ref-36 article-title: Exploring temporal coherence for more general video face forgery detection – volume: 8 start-page: 263 issue: 10 year: 2022 ident: 10.7717/peerj-cs.2127/ref-17 article-title: The face deepfake detection challenge publication-title: Journal of Imaging doi: 10.3390/jimaging8100263 – year: 2019 ident: 10.7717/peerj-cs.2127/ref-23 article-title: Roberta: a robustly optimized bert pretraining approach doi: 10.48550/arXiv.1907.11692 – start-page: 211 volume-title: SPM7’22 year: 2022 ident: 10.7717/peerj-cs.2127/ref-35 article-title: Face forgery detection of deepfake based on multiple convolutional neural networks – start-page: 1800 year: 2017 ident: 10.7717/peerj-cs.2127/ref-5 article-title: Xception: deep learning with depthwise separable convolutions |
SSID | ssj0001511119 |
Score | 2.2753427 |
Snippet | In recent years, the field of artificial intelligence has witnessed a remarkable surge in the generation of synthetic images, driven by advancements in deep... |
SourceID | doaj pubmedcentral proquest gale pubmed crossref |
SourceType | Open Website Open Access Repository Aggregation Database Index Database Enrichment Source |
StartPage | e2127 |
SubjectTerms | Artificial Intelligence CLIP Computational linguistics Computer Vision Data Mining and Machine Learning Deep learning Deepfake detection Language processing Multimodal machine learning Natural language interfaces Neural Networks Synthetic image detection |
Title | Detecting images generated by diffusers |
URI | https://www.ncbi.nlm.nih.gov/pubmed/39145210 https://www.proquest.com/docview/3093175159 https://pubmed.ncbi.nlm.nih.gov/PMC11322988 https://doaj.org/article/cade7d1a24ae47fab19955fad265cf56 |
Volume | 10 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1Lb9QwEB6VcuEClFdT2lVAFb0QGid27Bxb6Lb0UKFCpd4s27H7EM1Wze6Bf89Mkl0lQogL13gO9tgz840z_gZglxhESuaLhDuLCYpIeaJsmieZU1Z4IYlghaotzoqTC356KS4Hrb6oJqyjB-4Ut09V4rJiJuPGcxmMpTfFIpgqK4QLoiXbxpg3SKa698HkCsqOVFNiyrJ_7_3DbeKaT1nbQWYQhFqu_j898iAkjcslB_Fn-hye9sAxPugmvAFrvn4Bz5ZNGeLeRl_C3hdP_wUwIsU3d-gsmviqZZZGZBnbXzE1RKGLieYVXEyPfnw-SfpuCIkTQswTwSqEKoozlwXlhU2dlYEVKghTBB8scxWiEe9za9CyXEA3Z2SVBiVQbxjJ89ewXs9qvwlxQMNFZF1am-Xc-sxWQhplDMEpmXMewcelerTrqcKpY8VPjSkDaVO32tSu0aTNCD6sxO87joy_CR6SrldCRG3dfsAN1_2G639teATvaac0kVfUVB1zZRZNo79-P9cHiuUy5VyICPZ6oTDDmTvTPzbA9RPf1UhyeySJ1uVGw--WB0LTEJWk1X62aDT9QkbshXAwgjfdAVktLC8ZR1yURqBGR2e08vFIfXPdknszuh4oldr6H7p6C08yBGF0F83SbVifPyz8DoKouZ3AIzU9nsDjw6Ozb-eT1np-AyY3HsM |
linkProvider | Directory of Open Access Journals |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Detecting+images+generated+by+diffusers&rft.jtitle=PeerJ.+Computer+science&rft.au=Coccomini%2C+Davide+Alessandro&rft.au=Esuli%2C+Andrea&rft.au=Falchi%2C+Fabrizio&rft.au=Gennaro%2C+Claudio&rft.date=2024-07-10&rft.pub=PeerJ.+Ltd&rft.issn=2376-5992&rft.eissn=2376-5992&rft.volume=10&rft.spage=e2127&rft_id=info:doi/10.7717%2Fpeerj-cs.2127&rft.externalDBID=ISR&rft.externalDocID=A813704455 |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=2376-5992&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=2376-5992&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=2376-5992&client=summon |