<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Video Forgery Detection by Bitstream Analysis</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Hugo</forename><surname>Jean</surname></persName>
							<affiliation key="aff0">
								<orgName type="laboratory" key="lab1">UNICAEN</orgName>
								<orgName type="laboratory" key="lab2">ENSICAEN</orgName>
								<orgName type="institution" key="instit1">Normandie Univ</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">GREYC</orgName>
								<address>
									<postCode>14000</postCode>
									<settlement>Caen</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Emmanuel</forename><surname>Giguet</surname></persName>
							<email>emmanuel.giguet@unicaen.fr</email>
							<affiliation key="aff0">
								<orgName type="laboratory" key="lab1">UNICAEN</orgName>
								<orgName type="laboratory" key="lab2">ENSICAEN</orgName>
								<orgName type="institution" key="instit1">Normandie Univ</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">GREYC</orgName>
								<address>
									<postCode>14000</postCode>
									<settlement>Caen</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Christophe</forename><surname>Charrier</surname></persName>
							<email>christophe.charrier@unicaen.fr</email>
							<affiliation key="aff0">
								<orgName type="laboratory" key="lab1">UNICAEN</orgName>
								<orgName type="laboratory" key="lab2">ENSICAEN</orgName>
								<orgName type="institution" key="instit1">Normandie Univ</orgName>
								<orgName type="institution" key="instit2">CNRS</orgName>
								<orgName type="institution" key="instit3">GREYC</orgName>
								<address>
									<postCode>14000</postCode>
									<settlement>Caen</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Video Forgery Detection by Bitstream Analysis</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">9125482A9E97467C0E178D2BD682FDE9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T19:50+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Digital investigation</term>
					<term>Video forensics</term>
					<term>Video forgery</term>
					<term>Forgery detection</term>
					<term>Machine learning</term>
					<term>Bitstream analysis</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we propose a video tampering detection method based on bitstream analysis for videos in H.264 or MPEG-4 AVC format. This method aims at detecting inter-frame alterations: insertion, deletion, permutation, duplication. Features are extracted from the original bitstream. This method therefore does not require the decoding of the video, which improves the speed of analysis. The detection quality remains very significant in terms of binary detection, tampered / pristine video, with a F1 measure equal to 94.89. Concerning multiclass classification, F1 measure reaches 70.33 due to the difficulty to separate swap and duplication forgeries.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Nowadays, video content is transmitted in exponentially increasing volumes. Most of them are intended to be shared on social networks which have become so popular. This growth has been facilitated by the creation of powerful, easy-to-use video editing tools. Video editing has never been so easy: videos can be combined, unwanted part can be deleted, amazing scenes can be duplicated, frames can be altered to make objects or people disappear or appear, and this, according to one's desires or motivations. Technological advances in image and video editing have unleashed creativity, for humorous or artistic purposes, but also for misinformation, propaganda, and conspiracy. Therefore, the legitimacy, the reliability and the authenticity of the videos which broadcasted and relayed on Internet have become a major concern, in particular to detect disinformation attempts. Legally speaking, videos can now be used as evidence in court. The intentional modification of a video for the purpose of falsification, called video forgery, must be detectable. The challenge is to determine whether the video has been altered and, if possible, to qualify the nature of the alterations.</p><p>Many forgery detection methods have been proposed, but they are generally unable to detect simultaneously the different types of existing forgeries. Moreover, they require the entire video to be decoded beforehand in order to perform these detections.</p><p>In this work, we propose an original method for detecting inter-frame forgery in H.264 (or MPEG-4 AVC) videos, using the bitstream approach. This method detects insertion, deletion, permutation and duplication of frames. It is based on feature extraction directly from the bistream, i.e., from the compressed domain. Anomalies detection into a video are performed analyzing the variation of statistics computed on video fragments, taking into account the variation of the forward and backward motion vectors into the B and P frames by minimizing the false positives.</p><p>The paper is organized as follows. In Section 2, we introduce the current state-of-the-art for detecting inter-image falsification in video forgery. Our proposed methodology is then described in Section 3, including feature extraction and selected classification methods. In section 4, we provide a detailed description of the evaluation environment we set up for this study, including dataset construction, performance metrics, and two evaluation scenarii: a binary classification task and a multi-class classification task. In Section 5, we present the results we obtained for each scenario. Section 6 offers concluding remarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">State-of-the-art</head><p>In the literature, many methods for detecting inter-image falsifications are present. Whether these techniques are applied at the local level by LBP <ref type="bibr" target="#b0">[1]</ref>, by similarity measure computation <ref type="bibr" target="#b1">[2]</ref> like, for example, the MS-SSIM quality measure <ref type="bibr" target="#b2">[3]</ref>, by computing the Zernike opposite chromaticity moments <ref type="bibr" target="#b3">[4]</ref>, or even the histograms of oriented gradients and motion energy images <ref type="bibr" target="#b4">[5]</ref>, the announced performances are of high level. However, they decrease rapidly when the training conditions are more or less respected (dynamic video background, static video, and son on).</p><p>In 2014, Zhang et al <ref type="bibr" target="#b0">[1]</ref> calculated the correlation between each adjacent frame encoded with the LBP approach, to decipher the frame insertion and frame deletion fakes in a video. If the number of frame deletions is small, the performance of this technique degrades. Li et al. <ref type="bibr" target="#b1">[2]</ref> used the consistency of the quotient-of-mean structural similarity measure (QoMSSIM) to detect frame insertions and deletions. QoMSSIM is used as a feature and feeds an SVM classifier to detect the types of falsifications. However, the performance degrades when the videos are static, as is the case in video surveillance. Liu et al. <ref type="bibr" target="#b3">[4]</ref> proposed an approach based on coarse-to-fine investigation to detect tampering types by inserting, deleting, duplicating, and replacing frames in videos. In coarse detection, abnormal frame locations are detected using Zernike Opponent Chromaticity Moments (ZOCM-Zernike Opponent Chromaticity Moments). All images are transformed into color opposition space, and the Zernike moment correlation is calculated over the color space to obtain the ZOCM value. The coarse Tamura feature is extracted from the detected anomalous images, and the fine detection algorithm is run to reduce false positives. However, this approach fails when the background of the videos is dynamic. Recently, Fadl et al. <ref type="bibr" target="#b4">[5]</ref> used histograms of oriented gradients (HOG) and motion energy images (MEI) to design a passive detection technique to detect tampering by deleting, inserting, and reshuffling images. However, the performance of the proposed method quickly degrades when deleting images in a static scene video.</p><p>Concerning methods based on deep learning features, Long et al. <ref type="bibr" target="#b5">[6]</ref> used a 3DCNN network to detect frame deletion in a single 16-frame video shot and checked the center of the shot (between frames 8 and 9). They refined the confidence scores using peak detection and temporal scaling to reduce false alarms. They also proposed another method <ref type="bibr" target="#b6">[7]</ref> for image duplication using an I3D network (Two-Stream Inflated 3D ConvNet). The test video was divided into overlapping shots and the features of each shot were extracted using a pre-trained I3D network, and then the features of all the shots in the video were contacted to calculate the distance between them and detect similarity. Bakas et al. <ref type="bibr" target="#b7">[8]</ref> used three pre-trained 3DCNN models to detect deletion, insertion, and duplication of frames in a single video shot. In the proposed model, a difference layer is added in the CNN, which is mainly aimed at extracting temporal information from videos. The authors claim significant performance rates.</p><p>In recent years, techniques based on the use of CNNs (3DCNN, 2DCNN, etc.) <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8]</ref> have been widely used, showing significant performance rates.</p><p>All the previous methods rely on accessing the pixels of the video frames and then working in a transformed domain. They therefore require a complete and successful decoding of the encoded video files, which necessarily leads to a significant overall computation time, especially when processing several hours long videos.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The proposed approach</head><p>In order to be broadcast on the Internet, a video is encoded as a sequence of bits, commonly called a bitstream, using a compression algorithm, or codec. Among the most widely used are the H.264 codec and its successor the H.265 codec. Although the latter is more powerful, the H.264 codec is still widely used on the Internet today because of its better compatibility.</p><p>The video forgery detection method proposed here is illustrated in figure <ref type="figure" target="#fig_0">1</ref>. From the bitstream of the video, a feature extraction is performed using a stream analyzer. This set of features is then used to train learning models to classify the different types of tampering sought.</p><p>In the field of compression, a video is represented as a sequence of images. These images, of the intra or inter type, are organized into groups of images (GOP). Each GOP is composed of an intra (I) image, known as the key image, encoded in JPEG. This algorithm takes into account spatial redundancies in order to reduce the amount of data to be encoded. An intra image is followed by several inter images (B, P) represented by a set of motion vectors. These vectors symbolize the displacement of a pixel of the current image with respect to the reference images. This representation abstracts from temporal redundancies while encoding the motion content.</p><p>P-frames consider only the previous frames as reference frames while B-frames consider the following frames as additional references. The H.264 codec defines a frame as a set of slices, which are composed of macroblocks. An H.264 bitstream is structured in three layers. The Network Ability Layer (NAL) contains the video data blocks, called Video Coding Layers (VCLs). Each VCL describes a slice of image, named the Slice Layer. This layer is reused as the set of macroblocks that compose it. Each macroblock is finally described by its own characteristics at the level of the Macroblock Layer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Features extraction</head><p>In order to extract characteristics on the different layers of the bitstream, each VLC is inspected by the flow analyzer. The extracted parameters 𝑓 𝑙 are the following: The 𝑓 1 parameter is extracted directly at the Slice Layer while the rest is extracted at the Macroblock Layer. The features 𝑓 1 , 𝑓 2 and 𝑓 3 represent the distortion of the video while the moving content is symbolized by the features 𝑓 4 to 𝑓 7 . The encoder choices are finally transcribed from 𝑓 8 to 𝑓 27 . Each parameter is extracted for each image slice and then averaged over the current GOP size. The feature vector 𝑉 𝐺𝑂𝑃 𝑘 for each GOP 𝑘 is finally computed:</p><formula xml:id="formula_0">𝑉 𝐺𝑂𝑃 𝑘 = ( 1 𝑀𝑁 𝑀 ∑ 𝑗=1 𝑁 ∑ 𝑖=1 𝑓 𝑙,𝑖,𝑗 ) , ∀𝑙 ∈ [1, … , 27]<label>(1)</label></formula><p>where 𝑓 𝑙,𝑖,𝑗 represents the l-th feature of the i-th frame slice of the j-th frame of the k-th GOP of the video.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Selected classification methods</head><p>There are many binary and multiclass learning techniques in the literature. Their performance varies according to the problem to be solved. However, they are all implemented in software libraries so that it is now possible and easy to test several of them to compare their performance on different data sets. One of the best known and most robust libraries for machine learning is Scikit-learn, also called sklearn.</p><p>In order to study the adaptability of existing classification schemes to the bistream data, we compare, among the best performing strategies, the following approaches : <ref type="bibr" target="#b8">[9]</ref>: Gradient Boosting Classifier (GBC), Light-Gradient Boosting Machine (L-GBM), Logistic Regression (RL), Decision Tree Classifier (DTC), Random Forest Classifier (RFC), Support-Vector Machine (SVM) and K Nearest Neighbors (KNN).</p><p>We also tested the following methods: Ada Boost Classifier (ADA), Extra Trees Classifier (ETC), Linear Discriminant Analysis (LDA), Ridge Classifier (RC), Quadratic Discriminant Analysis (QDA), Dummy Classifier (DC) and Naive Bayes (NB).</p><p>In the end, fourteen methods are compared according to two scenarios: binary or multiclass classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experimental Setup</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Video Dataset Design</head><p>In order to evaluate our tampering detection method, we had to create our own dedicated video database, a dedicated database has been created, as we could not identify a free database containing the four types of inter-frame alterations (insertion, deletion, duplication, and frame swapping).</p><p>To build our artificial database, we proceeded by deriving videos from the LIVE Video Quality Challenge (VQC) database <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref> created by the University of California, Berkeley. This database was created by the University of Texas at Austin as part of the LIVE Video Quality Challenge (VQC). This original database consists of 585 unaltered videos featuring a wide variety of scenes, captured from 101 cameras representing 43 models, shot by 80 users, and with varying recording qualities. These videos have an average duration of 10.03 seconds, with variable formats, portrait or landscape, and variable resolutions.</p><p>For our evaluation campaign, we automated the creation of the altered video database from the VQC database. We had to define a falsification process covering the 4 types of alterations targeted, with sufficiently varied positions and alteration durations. From 82 videos randomly selected in the VQC database, a database of 410 videos is created by altering each original video according to one of four types. We have created a database of 410 videos by altering each original video in one of four ways: insertion, deletion, duplication and permutation.</p><p>To produce a video with insertion, a fragment to be inserted is extracted from a randomly selected video. The duration of this fragment is between 1 second and the total duration. The fragment is then inserted into the target video, at a position between the beginning of the target video and the end of the target video minus the insertion time.</p><p>To produce a video with delete, we randomly select a fragment to be deleted with a start position between the beginning and 75% of the video, and a random duration between 20 and 100% of the remaining duration.</p><p>To produce a video with duplication, we randomly select a fragment to duplicate of maximum 33% of the video, and starting between the beginning of the video and the end decreased the duration of the copy. The fragment is then inserted at a random point in the video.</p><p>To produce a video with permutation/swap, we randomly choose two fragments to permute, without overlapping range. To guarantee the non-overlap of the excerpts, we randomly choose a maximum duration of 33% of the video, and two distant starting points: the beginning of extract1 starts between the beginning and 33% of the video, that of extract2 between 35% and 65% of the video. The two extracts are then swapped.</p><p>All such tampered videos are then re-encoded using the H.264 codec using the default Constant Rate Factor (CRF) value equal to 23 in order to get quite good quality videos. Actually, since CRF is a "constant quality" encoding mode, as opposed to constant bitrate (CBR), it will compress different frames by different amounts, thus varying the Quantization Parameter (QP) as necessary to maintain a certain level of perceived quality.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Performance metrics</head><p>The performance of the 14 trail classification strategies selected was compared according to five criteria:</p><p>1. the accuracy which is the fraction of correct predictions of the model, 2. the precision which is the proportion of positive identification that is really correct, 3. the recall which is the proportion of real positives to have been correctly identified, 4. The F1 score which allows to evaluate the capacity of a classification model to predict efficiently the positive individuals, by making a compromise between precision and recall. It is defined by the harmonic mean of precision and recall, 5. The AUC (Area under the ROC Curve) provides an aggregate measure of performance for all possible classification thresholds. One way to interpret the AUC is the probability that the model ranks a random positive example higher than a random negative example. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Evaluation scenarii 4.3.1. Binary Classification Models</head><p>In this scenario, the goal is to classify the video into two classes: forged video, un-forged video.</p><p>The fourteen patterns presented in were tested to measure their ability to predict the class of the video.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.2.">Multi-class classification models</head><p>In this second scenario, we tested the ability of different classification models to predict the type of forgery (insertion, deletion, permutation and duplication), or the absence of forgery, using multiclass approaches. In this approach, the 6 models considered are: GBC, L-GBM, LR, DTC, SVM and KNN.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Performance Evaluation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Optimization and model training</head><p>Whether for binary or multiclass classification, the best combination of hyperparameters was performed using the Grid Search technique.</p><p>During the learning phase of the various schemes, 70% of the randomly drawn examples of the database constitute the learning database and the remaining 30% feed the test database. The 10-sub-sample cross-validation (𝑘 = 10) was used to evaluate the machine learning models.</p><p>The feature selection technique, or Features Selection, was not chosen as it was not appropriate. This technique is commonly used to select the features contributing to the performance of the model and to discard the less relevant ones. However, this process is not compatible with the </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Results</head><p>Table <ref type="table" target="#tab_0">1</ref> shows the results obtained for the binary classification. The Light Gradient Boosting Machine (L-GBM) obtains the best accuracy (91.63) and the best F1 measure (94.89). For multiclass classification, the table 2 presents the obtained results. The Gradient Boosting Classifier (GBC) obtains the best accuracy (70.63) and the best F1 measure (70.33).</p><p>Figure <ref type="figure" target="#fig_1">2</ref> presents the confusion matrix for the best classifier: LGBM. As we can observe, the classifier makes confusion for two kind of forgery: 1) swap and duplication. This is not really surprising since both swap and duplication are performed uusing an extract of the same video and thus, it, in general, is difficult to detect the difference between the a swap of two extracts of a duplication of an extract if a long term memory strategy is not used. One solution would to add such a strategy to be able to distinguish those two kinds of forgery. Except for this case, the obtained results clearly show that the LGBM classifier performs well to identify the type of forgery, and un-forged video.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper, we have proposed a video tampering detection method based on bitstream analysis for videos in H.264 or MPEG-4 AVC format. This forgery detection method aims at identifying inter-frame alterations: insertion, deletion, permutation, duplication. In our approach, the features taken into account during classification are directly derived from the file's bit sequence.</p><p>This video forgery detection method has the advantage to prevent decoding the video. Thus, it permits very fast and memory efficient analysis of the files. The binary classification, forged / un-forged video, remains very qualitative with an F1 measure equal to 94.89. It is obtained with the Light-Gradient Boosting Machine classification model. The multi-class classification task leads to promising results, with an F1 measure value equal to 70.33. It is obtained with the Gradient Boosting Classifier classification method.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Synopsis of the proposed model.</figDesc><graphic coords="4,151.80,84.19,291.68,292.93" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Confusion matrix for results obtained from the LGBM Classifier</figDesc><graphic coords="8,130.96,84.19,333.36,223.16" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Binary classifiers performance evaluation.</figDesc><table><row><cell cols="5">Model Accuracy AUC Recall Precision</cell><cell>F1</cell></row><row><cell>L-GBM</cell><cell>91.63</cell><cell cols="2">93.55 97.39</cell><cell>92.6</cell><cell>94.89</cell></row><row><cell>ADA</cell><cell>91.60</cell><cell>93.4</cell><cell>95.61</cell><cell>94.16</cell><cell>94.81</cell></row><row><cell>GBC</cell><cell>90.21</cell><cell>93.18</cell><cell>96.96</cell><cell>91.56</cell><cell>94.13</cell></row><row><cell>ETC</cell><cell>89.51</cell><cell cols="2">89.99 99.57</cell><cell>88.87</cell><cell>93.87</cell></row><row><cell>RFC</cell><cell>89.16</cell><cell>90.43</cell><cell>98.26</cell><cell>89.44</cell><cell>93.58</cell></row><row><cell>LR</cell><cell>87.77</cell><cell>87.85</cell><cell>94.33</cell><cell>91.23</cell><cell>92.51</cell></row><row><cell>LDA</cell><cell>87.44</cell><cell>88.49</cell><cell>93.87</cell><cell>91.19</cell><cell>92.3</cell></row><row><cell>RC</cell><cell>87.06</cell><cell>0.0</cell><cell>96.5</cell><cell>88.76</cell><cell>92.31</cell></row><row><cell>DTC</cell><cell>82.88</cell><cell>71.22</cell><cell>90.43</cell><cell>88.49</cell><cell>89.36</cell></row><row><cell>QDA</cell><cell>80.09</cell><cell>75.19</cell><cell>87.39</cell><cell>87.72</cell><cell>87.34</cell></row><row><cell>DC</cell><cell>80.09</cell><cell>0.5</cell><cell>1.0</cell><cell>80.09</cell><cell>88.94</cell></row><row><cell>KNN</cell><cell>78.04</cell><cell>75.31</cell><cell>89.51</cell><cell>84.14</cell><cell>86.64</cell></row><row><cell>SVM</cell><cell>76.56</cell><cell>0.0</cell><cell>89.35</cell><cell>82.81</cell><cell>85.</cell></row><row><cell>NB</cell><cell>72.04</cell><cell>84.85</cell><cell>68.99</cell><cell>94.68</cell><cell>79.16</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Multiclass classifiers performance evaluation</figDesc><table><row><cell cols="4">Model Accuracy Precision Recall</cell><cell>F1</cell><cell>AUC</cell></row><row><cell>GBC</cell><cell>70.63</cell><cell>71.50</cell><cell cols="3">68.70 70.33 90.53</cell></row><row><cell>LGB</cell><cell>68.19</cell><cell>69.16</cell><cell>66.38</cell><cell cols="2">67.83 90.55</cell></row><row><cell>LR</cell><cell>69.93</cell><cell>70.84</cell><cell cols="3">68.74 69.02 91.34</cell></row><row><cell>DTC</cell><cell>66.44</cell><cell>68.14</cell><cell>64.68</cell><cell cols="2">66.48 79.27</cell></row><row><cell>RFC</cell><cell>64.41</cell><cell>63.50</cell><cell>62.06</cell><cell></cell><cell>85.81</cell></row><row><cell>SVM</cell><cell>39.51</cell><cell>42.50</cell><cell>38.27</cell><cell>32.54</cell><cell>0.00</cell></row><row><cell>KNN</cell><cell>36.32</cell><cell>39.58</cell><cell>34.92</cell><cell cols="2">36.16 67.26</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Efficient video frame insertion and deletion detection based on inconsistency of correlations between local binary pattern coded frames</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Security and Communication Networks</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="311" to="320" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Video inter-frame forgery identification based on the consistency of quotient of mssim</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Security and Communication Networks</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="4548" to="4556" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Multi-scale structural similarity for image quality assessment</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">P</forename><surname>Simoncelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Asilomar Conference on Signals, Systems, and Computers</title>
				<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="1398" to="1402" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Exposing video inter-frame forgery by zernike opponent chromaticity moments and coarseness analysis</title>
	</analytic>
	<monogr>
		<title level="j">Multimedia Systems</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="223" to="238" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Surveillance video authentication using universal image quality index of temporal average</title>
		<author>
			<persName><forename type="first">S</forename><surname>Fadl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Digital Forensics and Watermarking</title>
				<editor>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Yoo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y.-Q</forename><surname>Shi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Kim</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Piva</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Kim</surname></persName>
		</editor>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="337" to="350" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A c3d-based convolutional neural network for frame dropping detection in a single video shot</title>
		<author>
			<persName><forename type="first">C</forename><surname>Long</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Basharat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hoogs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</title>
				<imprint>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="page" from="1898" to="1906" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">A coarse-to-fine deep convolutional neural network framework for frame duplication detection and localization in video forgery</title>
		<author>
			<persName><forename type="first">C</forename><surname>Long</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Basharat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hoogs</surname></persName>
		</author>
		<idno>ArXiv abs/1811.10762</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A digital forensic technique for inter-frame video forgery detection based on 3d cnn</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Naskar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Information Systems Security</title>
				<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="304" to="317" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Pattern recognition and machine learning</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Bishop</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">M</forename><surname>Nasrabadi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Springer</publisher>
			<biblScope unit="volume">4</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Large-scale study of perceptual video quality</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Sinno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
		<idno type="DOI">10.1109/TIP.2018.2869673</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Image Processing</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="612" to="627" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Large scale subjective video quality study</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Sinno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Bovik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">25th IEEE International Conference on Image Processing (ICIP)</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="276" to="280" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
