<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Robustness and Generalization of Synthetic Images Detectors</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Davide</forename><forename type="middle">Alessandro</forename><surname>Coccomini</surname></persName>
							<email>davidealessandro.coccomini@isti.cnr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">ISTI-CNR</orgName>
								<address>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Pisa</orgName>
								<address>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Roberto</forename><surname>Caldelli</surname></persName>
							<email>roberto.caldelli@unifi.it</email>
							<affiliation key="aff2">
								<orgName type="institution">CNIT</orgName>
								<address>
									<settlement>Florence</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
							<affiliation key="aff3">
								<orgName type="institution">Universitas Mercatorum</orgName>
								<address>
									<settlement>Rome</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Claudio</forename><surname>Gennaro</surname></persName>
							<email>claudio.gennaro@isti.cnr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">ISTI-CNR</orgName>
								<address>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giuseppe</forename><surname>Fiameni</surname></persName>
							<email>gfiameni@nvidia.com</email>
							<affiliation key="aff4">
								<orgName type="institution">NVIDIA AI Technology Center</orgName>
								<address>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Giuseppe</forename><surname>Amato</surname></persName>
							<email>giuseppe.amato@isti.cnr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">ISTI-CNR</orgName>
								<address>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Fabrizio</forename><surname>Falchi</surname></persName>
							<email>fabrizio.falchi@isti.cnr.it</email>
							<affiliation key="aff0">
								<orgName type="institution">ISTI-CNR</orgName>
								<address>
									<settlement>Pisa</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff5">
								<orgName type="department">Davide Alessandro Coccomini</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Robustness and Generalization of Synthetic Images Detectors</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">0B313AEC48F0C35F33C367B1CCBA2513</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Deepfake Detection, Deep Learning, Super Resolution F. Falchi) 0000-0002-0755-6154 (D. A. Coccomini)</term>
					<term>0000-0003-3471-1196 (R. Caldelli)</term>
					<term>0000-0002-3715-149X (C. Gennaro)</term>
					<term>0000-0001-8687-6609 (G. Fiameni)</term>
					<term>0000-0003-0171-4315 (G. Amato)</term>
					<term>0000-0001-6258-5313 (F. Falchi)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In recent times, the increasing spread of synthetic media, known as deepfakes has been made possible by the rapid progress in artificial intelligence technologies, especially deep learning algorithms. Growing worries about the increasing availability and believability of deepfakes have spurred researchers to concentrate on developing methods to detect them. In this field researchers at ISTI CNR's AIMH Lab, in collaboration with researchers from other organizations, have conducted research, investigations, and projects to contribute to combating this trend, exploring new solutions and threats. This article summarizes the most recent efforts made in this area by our researchers and in collaboration with other institutions and experts.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Deepfakes and synthetic media are becoming more prevalent and realistic day by day, presenting society with an increasingly urgent challenge, learning to distinguish reality from fiction effectively. These fake content can and are continually being used to spread disinformation, create smear campaigns, and manipulate reality with potentially devastating impacts for anyone who may end up a victim. To contrast this phenomenon, research has been conducted in recent years creating detectors, often based on deep learning techniques, that can classify a piece of content (such as an image) as realistic or fake. Despite many efforts, this discrimination capability is still insufficient today with many open problems in the field of deepfake detection. One example above all is that of generalization <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. In fact, deepfake detectors, although particularly effective in detecting images generated or manipulated by the same methods they are trained on, fail when using different and novel techniques. An-other problem is that of adversarial attacks, strategies of camouflaging traces, enhancing fake content or ad-hoc manipulations designed to fool the detector, which can be used to make detection even more complex. Deepfake detection models must therefore be designed so that they provide a high degree of robustness to possible adversarial attacks and also be able to effectively distinguish deepfakes without raising false alarms. For this reason, AIMH Lab at ISTI CNR has carried out numerous research attempts to explore new innovative techniques to advance this field but also to highlight possible hidden dangers that may damage the efforts made in previous research, representing dangers that detection systems may encounter in the real world. In particular, this paper summarizes the efforts made in <ref type="bibr" target="#b2">[3]</ref> and <ref type="bibr" target="#b3">[4]</ref>. process, some aspects of the image may change. For example, some previously visible details may become more blurred, totally disappear, or, conversely, be emphasized and brought to light. Deepfake generation algorithms commonly tend to introduce some more or less visible artifacts. In the case of human faces, these artifacts can be, for example, anomalies in pupils, contours of lips, eyes or ears, or accessories. Typically, deepfake detection models learn to recognize the specific anomalies introduced in manipulated images and, because of them, can discriminate between pristine and fake content. In <ref type="bibr" target="#b3">[4]</ref>, we explored whether Super Resolution techniques can be used as an adversarial attack to camouflage artifacts introduced by deepfake generations approaches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model</head><p>To do this, we proposed an SR-attack pipeline, whose purpose is to disguise artifacts present in deepfake images while still trying to keep the appearance as unaltered as possible. The pipeline begins with the detection of a face from the deepfake image that is subsequently downscaled by a factor 𝐾 using interpolation techniques. The resulting image is restored to its initial resolution through a super-resolution approach. The resulting face can eventually be reinserted into the original image, resulting in the camouflaged image.</p><p>An example of the impact of the proposed SR-attack is shown in Figure <ref type="figure" target="#fig_0">1</ref>. As the figure shows, the attack leads to the removal of artifacts introduced by deepfake generation techniques, such as noise around the mouth, and thus makes their detection extremely complex. In fact, to distinguish a counterfeit image from a pristine On the left, an example of manipulated face with a zoom on the artifacts around the mouth. On the right the same face but after the application of the SR-attack. From the zoom on the second one it can be seen how the artifacts are drastically smoothed.</p><p>one we often rely on observing these artifacts and oddities that can be introduced by the manipulation process. The fact that super-resolution leads to their removal or attenuation qualifies it as a potentially effective attack against deepfake detectors but also against the human eye itself.</p><p>The usage of the SR-attack conduct to a blurring effect on the artifacts introduced in the fake images and this makes them more difficult to detect. This is pretty evident in terms of performance; in fact, the use of the attack drastically degrades the performance of classifiers trained to do deepfake detection.</p><p>Table <ref type="table" target="#tab_0">1</ref> shows the accuracies of a Swin Transformer[6] and a Resnet50 <ref type="bibr" target="#b6">[7]</ref> on images manipulated with different techniques, considering them before and after SR-attack. The dataset used is FaceForensics++ <ref type="bibr" target="#b4">[5]</ref> and the deepfake generation methods considered are Deepfakes <ref type="bibr" target="#b7">[8]</ref>, Face2Face <ref type="bibr" target="#b8">[9]</ref>, FaceSwap <ref type="bibr" target="#b9">[10]</ref>, FaceShifter <ref type="bibr" target="#b10">[11]</ref> and NeuralTextures <ref type="bibr" target="#b11">[12]</ref>. This allows us to evaluate the effect of the super-resolution attack on different types of manipulations, thus highlighting on which it is more or less effective. Both the models are trained on the FaceForensics++ training set considering for the construction of the fake class, the same deepfake generation method used for the test.</p><p>According to our experiments, for both the considered models, when images are attacked with the proposed approach, there is an increase in False Negative Rate, particularly on some methods namely Face2Face, FaceSwap and FaceShifter. Others, however, are found to be more robust to attack, namely Deepfakes and NeuralTextures on which, however, there is an increase in False Positives. This behavior indicates that in some cases, the use of super-resolution techniques could lead to the elevation of false alarms, leading models to identify legitimately enhanced images through these approaches as deepfakes.</p><p>The latter result highlights a problem that could prove crippling to traditional deepfake detectors and could prevent their deployment in the real world. Indeed, it is plausible to think that on social networks it will become increasingly common to improve the quality of one's photos through Super-Resolution techniques. If this were to happen and in parallel the deepfake detectors were unable to understand that these are legitimate images but instead end up mistaking them for deepfakes, the number of false alarms would be such as to prevent their effective deployment on a large scale. It is therefore necessary on the one hand to defend against the malicious use of super-resolution to disguise artifacts introduced by manipulation techniques but also to make detectors robust so that they are able to recognize legitimately augmented images.</p><p>In these experiments we used only EDSR <ref type="bibr" target="#b12">[13]</ref> as the basis of our attack but the proposed attack can be conducted using different types of SR algorithms (such as BSRGAN <ref type="bibr" target="#b13">[14]</ref>), and depending on the peculiarities of each, greater or lesser effectiveness can be achieved on each specific deepfake generation method. The choice of the 𝐾 factor also has an impact in that as it increases, the detectors' errors increase but the quality of the image itself also deteriorates. Therefore, it is crucial to choose the SR algorithm and the value of the 𝐾 factor appropriately in order to achieve maximum effectiveness from the attack.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Future Works</head><p>In this section we expose some of the future works we are working on either as extensions of previously presented works or as new applications and solutions for effective deepfake detection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Robustness of Deepfake Detectors</head><p>The fact that Deepfake Detectors are susceptible to the use of SR techniques on images, whether fake or pristine, exposes a serious problem in their use in the real world and therefore requires that more studies be conducted to make them robust to this kind of content. It is necessary to find an effective way to make the models robust to this attack for example by introducing super-resolution as data augmentation during training.</p><p>The attack itself can also be further explored and improved by going to identify the optimal 𝐾 value and corresponding SR method as well as experimenting with different strategies for applying the attack, such as focusing on a frame rather than a detected face.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Deepfake Detection without Deepfakes</head><p>As stated before, one of the most stringent problems in the field of deepfake detection is that of generalization. Indeed, there is ample evidence that detectors tend to learn to effectively recognize deepfake content obtained through methods used to construct their training set, but fail when they need to classify content obtained through novel techniques. This leads to a total inadequacy of conventional deepfake detectors in being used in the real world. In fact, new deepfake generation techniques are continually being created, and it would be impractical to retrain the model each time to introduce every single possible method. In the context of synthetic images, this tendency of deepfake detectors stems from the fact that each generator introduces a specific fingerprint into the image <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b16">17]</ref>. It tends to be invisible to the human eye but involves the presence of structured patterns in the frequency domain (grids, symmetric peaks, halos, etc.).</p><p>From the observation of this phenomenon, as a future work we are exploring a new training technique for deepfake detectors that tries to stimulate the model to recognize the presence of structured patterns in the frequency domain and not to learn a specific fingerprint. The preliminary results of this approach can be found in <ref type="bibr" target="#b2">[3]</ref>.</p><p>We propose to reproduce prototype structured patterns inspired by what we observed from the fingerprints actually introduced by generative patterns of various types. These patterns, in frequency, are injected in pristine images and considered as "fake" in the training phase. During the training, we show the model pristine images and others on which a pattern has been applied, indicating them to the model as fakes.</p><p>From our preliminary experiments, we demonstrated that models trained on this proto-task, are extremely effective at identifying synthetic images despite never really seeing one in the training phase.</p><p>The use of synthetic patterns may also be used to support traditional training of deepfake detectors, introducing them only occasionally and still maintaining deepfakes in the training set. In addition, it will be possible to experiment with a virtually infinite number of patterns by searching for the most effective ones and studying their impact.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Conclusions</head><p>Carrying out research in the field of deepfake detection is an increasingly pressing need because of the multitude of techniques that now make it possible to produce synthetic or manipulated content with an increasing degree of credibility. As shown in our recent research, it is critical to explore both innovative methods to try to improve the capability of deepfake detectors looking for new training approaches and techniques. This could be needed to overcome the pressing problem of generalization. On the other hand, it is important to find new solution to face the possible risks and unexpected situations that these models might encounter in the real-world that could invalidate their potential, such as the adversarial attacks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Example of the impact of SR attack on fake images.On the left, an example of manipulated face with a zoom on the artifacts around the mouth. On the right the same face but after the application of the SR-attack. From the zoom on the second one it can be seen how the artifacts are drastically smoothed.</figDesc><graphic coords="2,302.62,348.08,203.36,168.58" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Evaluation on Faceforensics++<ref type="bibr" target="#b4">[5]</ref> test set. Each set is composed of half pristine and half fake images. The Attacked column indicates if the SR-attack has been applied to the images. The attack is applied to fake and pristine images.</figDesc><table><row><cell></cell><cell>Forgery</cell><cell>Attacked</cell><cell>Accuracy ↑</cell><cell>FNR ↓</cell><cell>FPR ↓</cell><cell>AUC ↑</cell></row><row><cell></cell><cell>Deepfakes</cell><cell>✖ ✓</cell><cell>95.3 90.7</cell><cell>5.9 6.1</cell><cell>3.6 12.4</cell><cell>99.1 97.4</cell></row><row><cell></cell><cell>NeuralTextures</cell><cell>✖ ✓</cell><cell>87.1 81.5</cell><cell>12.9 13.2</cell><cell>12.8 23.9</cell><cell>94.9 90.4</cell></row><row><cell>Swin</cell><cell>Face2Face</cell><cell>✖ ✓</cell><cell>95.2 87.0</cell><cell>6.3 24.4</cell><cell>3.3 1.7</cell><cell>98.9 96.1</cell></row><row><cell></cell><cell>FaceSwap</cell><cell>✖ ✓</cell><cell>95.2 86.4</cell><cell>4.9 21.9</cell><cell>4.6 5.3</cell><cell>98.6 93.9</cell></row><row><cell></cell><cell>FaceShifter</cell><cell>✖ ✓</cell><cell>94.4 89.0</cell><cell>7.2 18.9</cell><cell>4.1 3.1</cell><cell>98.7 97.4</cell></row><row><cell></cell><cell>Deepfakes</cell><cell>✖ ✓</cell><cell>95.6 95.0</cell><cell>5.5 6.9</cell><cell>3.2 10.1</cell><cell>99.2 98.9</cell></row><row><cell></cell><cell>NeuralTextures</cell><cell>✖ ✓</cell><cell>87.1 81.5</cell><cell>12.9 13.2</cell><cell>12.8 23.9</cell><cell>94.9 90.4</cell></row><row><cell>Resnet</cell><cell>Face2Face</cell><cell>✖ ✓</cell><cell>95.9 91.2</cell><cell>5.0 14.4</cell><cell>3.2 4.7</cell><cell>98.9 97.6</cell></row><row><cell></cell><cell>FaceSwap</cell><cell>✖ ✓</cell><cell>95.9 88.5</cell><cell>6.4 21.1</cell><cell>1.9 2.6</cell><cell>99.1 95.6</cell></row><row><cell></cell><cell>FaceShifter</cell><cell>✖ ✓</cell><cell>95.3 85.9</cell><cell>6.1 24.8</cell><cell>3.4 3.3</cell><cell>98.8 95.2</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.0.1.">Acknowledgments</head><p>This work was partially supported by the following projects: Tuscany Health Ecosystem (THE) (CUP B83C22003930001) and SERICS (PE00000014, MUR PNRR -NextGenerationEU), AI4Media (EC H2020 -n. 951911) and AI4Debunk (Horizon EU n. 101135757), FOSTERER (Italian MUR PRIN 2022). We acknowledge the CINECA award under the ISCRA initiative, for the availability of high-performance computing resources and support.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Cross-forgery analysis of vision transformers and CNNs for deepfake image detection</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Coccomini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caldelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Falchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gennaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Amato</surname></persName>
		</author>
		<idno type="DOI">10.1145/3512732.3533582</idno>
	</analytic>
	<monogr>
		<title level="m">International Conference on Multimedia Retrieval Workshop</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">On the generalization of deep learning models in video deepfake detection</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Coccomini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caldelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Falchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gennaro</surname></persName>
		</author>
		<idno type="DOI">10.3390/jimaging9050089</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Imaging</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Coccomini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caldelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gennaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Fiameni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Amato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Falchi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2403.13479</idno>
		<title level="m">Deepfake detection without deepfakes: Generalization via synthetic frequency patterns injection</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Adversarial magnification to deceive deepfake detection through super resolution (to appear)</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Coccomini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caldelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Falchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Amato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Falchi</surname></persName>
		</author>
		<author>
			<persName><surname>Gennaro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<meeting>European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Faceforensics++: Learning to detect manipulated facial images</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rossler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cozzolino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Verdoliva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Riess</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Thies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Niessner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF International Conference on Computer Vision</title>
				<meeting>the IEEE/CVF International Conference on Computer Vision</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Swin transformer: Hierarchical vision transformer using shifted windows</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Guo</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICCV48922.2021.00986</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE/CVF International Conference on Computer Vision (ICCV)</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Deep residual learning for image recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR.2016.90</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<imprint>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title/>
		<author>
			<persName><surname>Deepfakes</surname></persName>
		</author>
		<ptr target="https://github.com/deepfakes/faceswap" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Face2face: Real-time face capture and reenactment of rgb videos</title>
		<author>
			<persName><forename type="first">J</forename><surname>Thies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zollhöfer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stamminger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Theobalt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nießner</surname></persName>
		</author>
		<idno type="DOI">10.1145/3292039</idno>
	</analytic>
	<monogr>
		<title level="j">Commun. ACM</title>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">M</forename><surname>Faceswap</surname></persName>
		</author>
		<ptr target="https://github.com/MarekKowalski/FaceSwap/" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1912.13457</idno>
		<title level="m">Faceshifter: Towards high fidelity and occlusion aware face swapping</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Deferred neural rendering: Image synthesis using neural textures</title>
		<author>
			<persName><forename type="first">J</forename><surname>Thies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zollhöfer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nießner</surname></persName>
		</author>
		<idno type="DOI">10.1145/3306346.3323035</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Enhanced deep residual networks for single image super-resolution</title>
		<author>
			<persName><forename type="first">B</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Son</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Mu</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops</title>
				<meeting>the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Designing a practical degradation model for deep blind image super-resolution</title>
		<author>
			<persName><forename type="first">K</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Van Gool</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Timofte</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Computer Vision</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Detecting images generated by diffusers</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Coccomini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Esuli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Falchi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gennaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Amato</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2303.05275</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">CNN-generated images are surprisingly easy to spot</title>
		<author>
			<persName><forename type="first">S.-Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Owens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Efros</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR42600.2020.00872</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE Conf. Comput. Vis. Pattern Recog</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="8692" to="8701" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">On the detection of synthetic images generated by diffusion models</title>
		<author>
			<persName><forename type="first">R</forename><surname>Corvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cozzolino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zingarini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Poggi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Nagano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Verdoliva</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICASSP49357.2023.10095167</idno>
	</analytic>
	<monogr>
		<title level="m">Int. Conf. on Acoustics, Speech and Signal Processing</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="5" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
