<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Detection by Bitstream Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hugo Jean</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emmanuel Giguet</string-name>
          <email>emmanuel.giguet@unicaen.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christophe Charrier</string-name>
          <email>christophe.charrier@unicaen.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Normandie Univ.</institution>
          ,
          <addr-line>UNICAEN, ENSICAEN, CNRS, GREYC, 14000 Caen</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <abstract>
        <p>In this paper, we propose a video tampering detection method based on bitstream analysis for videos in H.264 or MPEG-4 AVC format. This method aims at detecting inter-frame alterations: insertion, deletion, permutation, duplication. Features are extracted from the original bitstream. This method therefore does not require the decoding of the video, which improves the speed of analysis. The detection quality remains very significant in terms of binary detection, tampered / pristine video, with a F1 measure equal to 94.89. Concerning multiclass classification, F1 measure reaches 70.33 due to the dificulty to separate swap and duplication forgeries.</p>
      </abstract>
      <kwd-group>
        <kwd>Digital investigation</kwd>
        <kwd>Video forensics</kwd>
        <kwd>Video forgery</kwd>
        <kwd>Forgery detection</kwd>
        <kwd>Machine learning</kwd>
        <kwd>Bitstream</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>(C. Charrier) GLOBE</p>
      <p>https://giguete.users.greyc.fr/ (E. Giguet); https://
charrierc.users.greyc.fr/ (C. Charrier) Orcid</p>
      <p>0000-0002-3836-9829 (C. Charrier)
permutation and duplication of frames. It is based on feature extraction directly from the
bistream, i.e., from the compressed domain. Anomalies detection into a video are performed
analyzing the variation of statistics computed on video fragments, taking into account the
variation of the forward and backward motion vectors into the B and P frames by minimizing
the false positives.</p>
      <p>The paper is organized as follows. In Section 2, we introduce the current state-of-the-art
for detecting inter-image falsification in video forgery. Our proposed methodology is then
described in Section 3, including feature extraction and selected classification methods. In
section 4, we provide a detailed description of the evaluation environment we set up for this
study, including dataset construction, performance metrics, and two evaluation scenarii: a
binary classification task and a multi-class classification task. In Section 5, we present the
results we obtained for each scenario. Section 6 ofers concluding remarks.</p>
    </sec>
    <sec id="sec-2">
      <title>2. State-of-the-art</title>
      <p>
        In the literature, many methods for detecting inter-image falsifications are present. Whether
these techniques are applied at the local level by LBP [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], by similarity measure computation
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] like, for example, the MS-SSIM quality measure [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], by computing the Zernike opposite
chromaticity moments [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], or even the histograms of oriented gradients and motion energy
images [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the announced performances are of high level. However, they decrease rapidly
when the training conditions are more or less respected (dynamic video background, static
video, and son on).
      </p>
      <p>
        In 2014, Zhang et al [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] calculated the correlation between each adjacent frame encoded with
the LBP approach, to decipher the frame insertion and frame deletion fakes in a video. If the
number of frame deletions is small, the performance of this technique degrades. Li et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
used the consistency of the quotient-of-mean structural similarity measure (QoMSSIM) to detect
frame insertions and deletions. QoMSSIM is used as a feature and feeds an SVM classifier to
detect the types of falsifications. However, the performance degrades when the videos are static,
as is the case in video surveillance. Liu et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] proposed an approach based on coarse-to-fine
investigation to detect tampering types by inserting, deleting, duplicating, and replacing frames
in videos. In coarse detection, abnormal frame locations are detected using Zernike Opponent
Chromaticity Moments (ZOCM–Zernike Opponent Chromaticity Moments). All images are
transformed into color opposition space, and the Zernike moment correlation is calculated over
the color space to obtain the ZOCM value. The coarse Tamura feature is extracted from the
detected anomalous images, and the fine detection algorithm is run to reduce false positives.
However, this approach fails when the background of the videos is dynamic. Recently, Fadl et
al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] used histograms of oriented gradients (HOG) and motion energy images (MEI) to design
a passive detection technique to detect tampering by deleting, inserting, and reshufling images.
However, the performance of the proposed method quickly degrades when deleting images in a
static scene video.
      </p>
      <p>
        Concerning methods based on deep learning features, Long et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] used a 3DCNN network
to detect frame deletion in a single 16-frame video shot and checked the center of the shot
(between frames 8 and 9). They refined the confidence scores using peak detection and temporal
scaling to reduce false alarms. They also proposed another method [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] for image duplication
using an I3D network (Two-Stream Inflated 3D ConvNet ). The test video was divided into
overlapping shots and the features of each shot were extracted using a pre-trained I3D network,
and then the features of all the shots in the video were contacted to calculate the distance
between them and detect similarity. Bakas et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] used three pre-trained 3DCNN models to
detect deletion, insertion, and duplication of frames in a single video shot. In the proposed
model, a diference layer is added in the CNN, which is mainly aimed at extracting temporal
information from videos. The authors claim significant performance rates.
      </p>
      <p>
        In recent years, techniques based on the use of CNNs (3DCNN, 2DCNN, etc.) [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 7, 8</xref>
        ] have
been widely used, showing significant performance rates.
      </p>
      <p>All the previous methods rely on accessing the pixels of the video frames and then working
in a transformed domain. They therefore require a complete and successful decoding of the
encoded video files, which necessarily leads to a significant overall computation time, especially
when processing several hours long videos.</p>
    </sec>
    <sec id="sec-3">
      <title>3. The proposed approach</title>
      <p>In order to be broadcast on the Internet, a video is encoded as a sequence of bits, commonly
called a bitstream, using a compression algorithm, or codec. Among the most widely used are
the H.264 codec and its successor the H.265 codec. Although the latter is more powerful, the
H.264 codec is still widely used on the Internet today because of its better compatibility.</p>
      <p>The video forgery detection method proposed here is illustrated in figure 1. From the bitstream
of the video, a feature extraction is performed using a stream analyzer. This set of features is
then used to train learning models to classify the diferent types of tampering sought.</p>
      <p>In the field of compression, a video is represented as a sequence of images. These images, of
the intra or inter type, are organized into groups of images (GOP). Each GOP is composed of an
intra (I) image, known as the key image, encoded in JPEG. This algorithm takes into account
spatial redundancies in order to reduce the amount of data to be encoded. An intra image is
followed by several inter images (B, P) represented by a set of motion vectors. These vectors
symbolize the displacement of a pixel of the current image with respect to the reference images.
This representation abstracts from temporal redundancies while encoding the motion content.</p>
      <p>P-frames consider only the previous frames as reference frames while B-frames consider the
following frames as additional references. The H.264 codec defines a frame as a set of slices,
which are composed of macroblocks. An H.264 bitstream is structured in three layers. The
Network Ability Layer (NAL) contains the video data blocks, called Video Coding Layers (VCLs).
Each VCL describes a slice of image, named the Slice Layer. This layer is reused as the set of
macroblocks that compose it. Each macroblock is finally described by its own characteristics at
the level of the Macroblock Layer.</p>
      <sec id="sec-3-1">
        <title>3.1. Features extraction</title>
        <p>In order to extract characteristics on the diferent layers of the bitstream, each VLC is inspected
by the flow analyzer. The extracted parameters   are the following:</p>
        <p>-  1 : the bitrate
-  2 : the average Quantization Parameter (QP)
-  3 : the QP delta (ΔQP)
-  4,  5 : the average and maximum length of the motion vectors
-  6,  7 the average and maximum length of the prediction error on the motion vectors
-  8, ⋯ ,  10 : the percentage of intra (I), inter (B, P) and uncoded (skip) macroblocks
-  11, ⋯ ,  13 : the percentage of macroblocks of type I having a size of 16x16, 8x8 and 4x4.
-  14, ⋯ ,  17 : the percentage of macroblocks of type P having a size of 16x16, 16x8, 8x16
and 8x8.
-  18, ⋯ ,  20 : the percentage of sub-macroblocks of type P having a size of 8x4, 4x8 and
4x4.
-  21, ⋯ ,  24 : the percentage of type B macroblocks having a size of 16x16, 16x8, 8x16 and
8x8.
-  25, ⋯ ,  27 : the percentage of uncoded type B and P macroblocks and direct-coded type</p>
        <p>B macroblocks</p>
        <p>The  1 parameter is extracted directly at the Slice Layer while the rest is extracted at the
Macroblock Layer. The features  1,  2 and  3 represent the distortion of the video while the
moving content is symbolized by the features  4 to  7. The encoder choices are finally transcribed
from  8 to  27. Each parameter is extracted for each image slice and then averaged over the
current GOP size. The feature vector    for each GOP  is finally computed:
   = (  
1 ∑ ∑  ,, ) , ∀ ∈ [1, … , 27]
=1 =1
(1)
where  ,, represents the l-th feature of the i-th frame slice of the j-th frame of the k-th GOP of
the video.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Selected classification methods</title>
        <p>There are many binary and multiclass learning techniques in the literature. Their performance
varies according to the problem to be solved. However, they are all implemented in software
libraries so that it is now possible and easy to test several of them to compare their performance
on diferent data sets. One of the best known and most robust libraries for machine learning is
Scikit-learn, also called sklearn.</p>
        <p>In order to study the adaptability of existing classification schemes to the bistream data,
we compare, among the best performing strategies, the following approaches : [9]: Gradient
Boosting Classifier (GBC), Light-Gradient Boosting Machine (L-GBM), Logistic Regression (RL),
Decision Tree Classifier (DTC), Random Forest Classifier (RFC), Support-Vector Machine (SVM)
and K Nearest Neighbors (KNN).</p>
        <p>We also tested the following methods: Ada Boost Classifier (ADA), Extra Trees Classifier
(ETC), Linear Discriminant Analysis (LDA), Ridge Classifier (RC), Quadratic Discriminant Analysis
(QDA), Dummy Classifier (DC) and Naive Bayes (NB).</p>
        <p>In the end, fourteen methods are compared according to two scenarios: binary or multiclass
classification.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <sec id="sec-4-1">
        <title>4.1. Video Dataset Design</title>
        <p>In order to evaluate our tampering detection method, we had to create our own dedicated
video database, a dedicated database has been created, as we could not identify a free database
containing the four types of inter-frame alterations (insertion, deletion, duplication, and frame
swapping).</p>
        <p>To build our artificial database, we proceeded by deriving videos from the LIVE Video Quality
Challenge (VQC) database [10, 11] created by the University of California, Berkeley. This
database was created by the University of Texas at Austin as part of the LIVE Video Quality
Challenge (VQC). This original database consists of 585 unaltered videos featuring a wide
variety of scenes, captured from 101 cameras representing 43 models, shot by 80 users, and
with varying recording qualities. These videos have an average duration of 10.03 seconds, with
variable formats, portrait or landscape, and variable resolutions.</p>
        <p>For our evaluation campaign, we automated the creation of the altered video database from
the VQC database. We had to define a falsification process covering the 4 types of alterations
targeted, with suficiently varied positions and alteration durations. From 82 videos randomly
selected in the VQC database, a database of 410 videos is created by altering each original video
according to one of four types. We have created a database of 410 videos by altering each
original video in one of four ways: insertion, deletion, duplication and permutation.</p>
        <p>To produce a video with insertion, a fragment to be inserted is extracted from a randomly
selected video. The duration of this fragment is between 1 second and the total duration. The
fragment is then inserted into the target video, at a position between the beginning of the target
video and the end of the target video minus the insertion time.</p>
        <p>To produce a video with delete, we randomly select a fragment to be deleted with a start
position between the beginning and 75% of the video, and a random duration between 20 and
100% of the remaining duration.</p>
        <p>To produce a video with duplication, we randomly select a fragment to duplicate of maximum
33% of the video, and starting between the beginning of the video and the end decreased the
duration of the copy. The fragment is then inserted at a random point in the video.</p>
        <p>To produce a video with permutation/swap, we randomly choose two fragments to permute,
without overlapping range. To guarantee the non-overlap of the excerpts, we randomly choose
a maximum duration of 33% of the video, and two distant starting points: the beginning of
extract1 starts between the beginning and 33% of the video, that of extract2 between 35% and
65% of the video. The two extracts are then swapped.</p>
        <p>All such tampered videos are then re-encoded using the H.264 codec using the default
Constant Rate Factor (CRF) value equal to 23 in order to get quite good quality videos. Actually,
since CRF is a ”constant quality” encoding mode, as opposed to constant bitrate (CBR), it will
compress diferent frames by diferent amounts, thus varying the Quantization Parameter (QP)
as necessary to maintain a certain level of perceived quality.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Performance metrics</title>
        <p>The performance of the 14 trail classification strategies selected was compared according to five
criteria:
1. the accuracy which is the fraction of correct predictions of the model,
2. the precision which is the proportion of positive identification that is really correct,
3. the recall which is the proportion of real positives to have been correctly identified,
4. The F1 score which allows to evaluate the capacity of a classification model to predict
eficiently the positive individuals, by making a compromise between precision and recall.</p>
        <p>It is defined by the harmonic mean of precision and recall,
5. The AUC (Area under the ROC Curve) provides an aggregate measure of performance
for all possible classification thresholds. One way to interpret the AUC is the probability
that the model ranks a random positive example higher than a random negative example.</p>
        <p>Model
L-GBM</p>
        <p>ADA
GBC
ETC
RFC
LR
LDA
RC
DTC
QDA
DC
KNN
SVM
NB</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Evaluation scenarii</title>
        <sec id="sec-4-3-1">
          <title>4.3.1. Binary Classification Models</title>
          <p>In this scenario, the goal is to classify the video into two classes: forged video, un-forged video.
The fourteen patterns presented in were tested to measure their ability to predict the class of
the video.</p>
        </sec>
        <sec id="sec-4-3-2">
          <title>4.3.2. Multi-class classification models</title>
          <p>In this second scenario, we tested the ability of diferent classification models to predict the
type of forgery (insertion, deletion, permutation and duplication), or the absence of forgery,
using multiclass approaches. In this approach, the 6 models considered are: GBC, L-GBM, LR,
DTC, SVM and KNN.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Performance Evaluation</title>
      <sec id="sec-5-1">
        <title>5.1. Optimization and model training</title>
        <p>Whether for binary or multiclass classification, the best combination of hyperparameters was
performed using the Grid Search technique.</p>
        <p>During the learning phase of the various schemes, 70% of the randomly drawn examples of
the database constitute the learning database and the remaining 30% feed the test database. The
10-sub-sample cross-validation ( = 10 ) was used to evaluate the machine learning models.</p>
        <p>The feature selection technique, or Features Selection, was not chosen as it was not appropriate.
This technique is commonly used to select the features contributing to the performance of the
model and to discard the less relevant ones. However, this process is not compatible with the
fact that the detection of diferent types of alterations requires considering diferent subsets of
features.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Results</title>
        <p>For multiclass classification, the table 2 presents the obtained results. The Gradient Boosting
Classifier (GBC) obtains the best accuracy (70.63) and the best F1 measure (70.33).</p>
        <p>Figure 2 presents the confusion matrix for the best classifier: LGBM. As we can observe, the
classifier makes confusion for two kind of forgery: 1) swap and duplication. This is not really
surprising since both swap and duplication are performed uusing an extract of the same video
and thus, it, in general, is dificult to detect the diference between the a swap of two extracts of
a duplication of an extract if a long term memory strategy is not used. One solution would to
add such a strategy to be able to distinguish those two kinds of forgery. Except for this case,
the obtained results clearly show that the LGBM classifier performs well to identify the type of
forgery, and un-forged video.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this paper, we have proposed a video tampering detection method based on bitstream analysis
for videos in H.264 or MPEG-4 AVC format. This forgery detection method aims at identifying
inter-frame alterations: insertion, deletion, permutation, duplication. In our approach, the
features taken into account during classification are directly derived from the file’s bit sequence.</p>
      <p>This video forgery detection method has the advantage to prevent decoding the video. Thus,
it permits very fast and memory eficient analysis of the files. The binary classification, forged
/ un-forged video, remains very qualitative with an F1 measure equal to 94.89. It is obtained
with the Light-Gradient Boosting Machine classification model. The multi-class classification
task leads to promising results, with an F1 measure value equal to 70.33. It is obtained with the
Gradient Boosting Classifier classification method.
[9] C. M. Bishop, N. M. Nasrabadi, Pattern recognition and machine learning, volume 4,</p>
      <p>Springer, 2006.
[10] Z. Sinno, A. C. Bovik, Large-scale study of perceptual video quality, IEEE Transactions on</p>
      <p>Image Processing 28 (2019) 612–627. doi:1 0 . 1 1 0 9 / T I P . 2 0 1 8 . 2 8 6 9 6 7 3 .
[11] Z. Sinno, A. C. Bovik, Large scale subjective video quality study, in: 2018 25th IEEE
International Conference on Image Processing (ICIP), 2018, pp. 276–280.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Eficient video frame insertion and deletion detection based on inconsistency of correlations between local binary pattern coded frames</article-title>
          ,
          <source>Security and Communication Networks</source>
          <volume>8</volume>
          (
          <year>2015</year>
          )
          <fpage>311</fpage>
          -
          <lpage>320</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Guo,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Video inter-frame forgery identification based on the consistency of quotient of mssim</article-title>
          ,
          <source>Security and Communication Networks</source>
          <volume>9</volume>
          (
          <year>2016</year>
          )
          <fpage>4548</fpage>
          -
          <lpage>4556</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. P.</given-names>
            <surname>Simoncelli</surname>
          </string-name>
          , A. C.
          <article-title>Bovik, Multi-scale structural similarity for image quality assessment</article-title>
          ,
          <source>in: IEEE Asilomar Conference on Signals, Systems, and Computers</source>
          ,
          <year>2003</year>
          , pp.
          <fpage>1398</fpage>
          -
          <lpage>1402</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>[4] Exposing video inter-frame forgery by zernike opponent chromaticity moments and coarseness analysis</article-title>
          ,
          <source>Multimedia Systems</source>
          <volume>23</volume>
          (
          <year>2017</year>
          )
          <fpage>223</fpage>
          -
          <lpage>238</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fadl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Surveillance video authentication using universal image quality index of temporal average</article-title>
          , in: C. D.
          <string-name>
            <surname>Yoo</surname>
            ,
            <given-names>Y.-Q.</given-names>
          </string-name>
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Piva</surname>
          </string-name>
          , G. Kim (Eds.),
          <source>Digital Forensics and Watermarking</source>
          , Springer International Publishing,
          <year>2019</year>
          , pp.
          <fpage>337</fpage>
          -
          <lpage>350</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Basharat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hoogs</surname>
          </string-name>
          ,
          <article-title>A c3d-based convolutional neural network for frame dropping detection in a single video shot</article-title>
          ,
          <source>in: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1898</fpage>
          -
          <lpage>1906</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Basharat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hoogs</surname>
          </string-name>
          ,
          <article-title>A coarse-to-fine deep convolutional neural network framework for frame duplication detection and localization in video forgery</article-title>
          , ArXiv abs/
          <year>1811</year>
          .10762 (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bakas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Naskar</surname>
          </string-name>
          ,
          <article-title>A digital forensic technique for inter-frame video forgery detection based on 3d cnn</article-title>
          ,
          <source>in: Information Systems Security</source>
          , Springer International Publishing,
          <year>2018</year>
          , pp.
          <fpage>304</fpage>
          -
          <lpage>317</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>