<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Comparison of Supervised Machine Learning Techniques for CERN CMS O ine Data Certi cation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mantas Stankevicius</string-name>
          <email>mantas.stankevicius@cern.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Virginijus Marcinkevicius</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Valdas Rapsevicius</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Vilnius University, Faculty of Mathematics and Informatics</institution>
          ,
          <country country="LT">Lithuania</country>
        </aff>
      </contrib-group>
      <fpage>170</fpage>
      <lpage>176</lpage>
      <abstract>
        <p>The Compact Muon Solenoid (CMS) is one of the experiments at the CERN Large Hadron Collider (LHC). CMS is a generalpurpose detector able to record particle collisions up to 40 million times each second. 40 MHz rate results in approx 40 Tb/s, however journey of recorded CMS data is long and only small fraction makes to the nal step - physics analysis. Chain of automated and semi-automated processes lter, reconstruct, calibrate and verify data before it can be used for physics analysis. Data certi cation process starts with data aggregation in histograms, plots and various statistical quantities, and nishes with manual data quality assessment and decision. During the last step multiple experts review and evaluate (certify) data as being \good" or \bad". This results in high manpower demand and occasional involuntary human errors. Data certi cation process consists of online (to determine possible problems during the data taking, only the fraction of data is used) and o ine (not constrained in time full data set analysis including physics data reconstruction). Main goal of this research is to investigate the applicability and provide the comparison of various supervised machine learning techniques for o ine data certi cation process automation. Removal or signi cant reduction of manual labor and exclusion of human errors from the CMS data certi cation process are the main motivations of this research.</p>
      </abstract>
      <kwd-group>
        <kwd>CERN CMS</kwd>
        <kwd>Data certi cation</kwd>
        <kwd>Supervised machine learning</kwd>
        <kwd>Neural network</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Data Quality Monitoring (DQM) is one of the central and the critically
important projects of the Compact Muon Solenoid (CMS) detector at the CERN
Large Hadron Collider (LHC). Main goal of the DQM is to provide and support
a single end-to-end process for reliable certi cation of the recorded data. The
certi cation process comprises of multiple parts, starting from lling and
handling multiple histograms and scalar monitor elements and nishing with the
list of \good" list of runs and lumisections. The goal of the DQM is to discover
and pinpoint errors, problems occurring in detector hardware or reconstruction
software, early, with su cient accuracy and clarity to maintain good detector
and operation e ciency [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        While online DQM process only the small fraction of sampled data for
immediate response and provides just technical Detector Performance Group (DPG)
histograms, the o ine DQM examines the full dataset and generates both DPG
and physics POG (Physics Objects Group) histograms. Output is being
accumulated into the ROOT [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] format le and displayed to shifters and experts by
using DQMGUI application [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Then the certi cation process involves shifters
for monitoring various sets of histograms in online and o ine and marking
collected data as "good" or "bad" and providing comments. Certi cation is made on
run (data taking session) and luminosity section (lumisection) levels.
Lumisection is a data sample of approximately 23 seconds of data-taking. Final decision
is made by the data certi cation group experts by cross-checking assessments
and generating the nal list of good data [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] also known as GoldenJSON. Data
quality ags, comments and other related information is stored in Run Registry
database [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>Since the start of the experiment in 2008 the CMS DQM processes and tools
have been worked out and are stable. A decade of data taking have accumulated
the su cient amount of labeled DQM data. This allows the scientists to examine
the possibility to automate the nal step of the data certi cation process by using
the recent advances in Machine Learning algorithms for binary classi cation. In
the current system, hundreds of di erent histograms come as an input and the
output is one of \good" or \bad" for each lumisection.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Currently, there are several initiatives trying di erent approaches for the data
certi cation and anomaly detection. One of the most advanced research is done
by Yandex School of Data Analysis [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. They investigate applicability of deep
neural networks for anomaly detection and classi cation. Predictive power of
neural network is quite high - ROC AUC score equals to 0.96. The research is
based on data collected in 2010 by CERN CMS experiment. In this paper we
use newer and bigger dataset and compare other supervised machine learning
algorithms.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methodology</title>
      <p>
        Data used in this research was collected by CERN CMS [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] experiment at
LHC during 2016 data-taking. CMS DQM group provides aggregated data of
JetHT2016 dataset and so-called GoldenJSON. JetHT2016 dataset is widely
used at CMS for machine learning research. Dataset contains around 160,000
lumisections. GoldenJSON is a JSON le which contains labels for lumisections.
Labels are \good" or \bad" meaning if lumisection is good for further physics
analysis or not. Each lumisection consists of 401 parameters (histograms) which
are vectors of 7 numbers (mean, RMS and 5 quantiles of the histogram). So
in total each lumisection has 2807 features. There are several other parameters
used for data preprocessing but not for training, for example run number and
lumisection number, etc.
      </p>
      <p>Data preprocessing consists of two steps: merging dataset with GoldenJSON
and feature normalization. We used standard score normalization to center and
scale data so that distribution of each feature has a mean value 0 and standard
deviation of 1.
3.1</p>
      <sec id="sec-3-1">
        <title>Class Imbalance Problem</title>
        <p>Dataset contains very small amount (1.8%) of so-called \bad" lumisections,
hence class distribution ratio is 49:1. Therefore, multiple tactics are used to
deal class imbalance: class weights penalty, strati ed fold cross validation and
rank based performance metrics (see Section 3.3).</p>
        <p>First of all, we introduced class weights during training of a model. Additional
penalty for classi cation mistakes can a ect model to pay more attention to
minority class.</p>
        <p>
          Secondly, we use strati ed cross validation to preserve percentage of samples
for each class during cross validation. Natural distribution of classes in train and
test splits helps model to ght class imbalance during training. However, classical
strati ed cross validation is limited to a certain number of splits. We decided to
use Strati edShu eSplit [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] as it gives randomized folds. Using the randomized
folds model can be validated multiple times while keeping same train-test split
ratio.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Classi cation Methods</title>
        <p>Five methods were selected for a comparison: Support Vector Machine, Random
Forest, Naive Bayes, Gradient Boosted Trees and Arti cial Neural Network.
We chose popular methods from these categories: probabilistic, ensemble and
hierarchical.</p>
        <p>
          Support Vector Machine (SVM) We use scikit-learn [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] implementation
of Support Vector classi er. However after series of tries SVM turned out to
be inappropriate for the task. Large number of high-dimensional training data
badly a ected performance, therefore this method was eliminated from further
development and evaluation.
        </p>
        <p>
          Random Forest (RF) We use scikit-learn implementation of Random Forest
classi er. The forest has 64 trees with depth of 7 layers, these are hand-picked
parameters.
Gradient Boosted trees (XGB) We use XGBoost [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] implementation for
gradient boosted decision trees. Same as with Random Forest, we hand-picked
several parameters to improve performance: number of trees is 64 with max
depth of 7 layers.
        </p>
        <p>Gaussian Naive Bayes (NB) We use scikit-learn implementation of Gaussian
Naive Bayes algorithm for classi cation. The method trains very fast, but its
predictive power is too poor for this task.</p>
        <p>
          Arti cial Neural Network (ANN) We use Keras [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] library and Tensor ow
[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] backend. Neural network consists of 3 hidden layers. Each regular
denselyconnected NN layer uses recti ed linear unit (ReLU) activation function and is
followed by Dropout layer. Output layer uses sigmoid activation function (see
Fig. 1). Early stopping is used to avoid over- tting and stop training when
validation accuracy begins to decrease.
To determine the performance of di erent methods we use three performance
measures: accuracy (ACC), F1 score and Receiver Operating Characteristic
(ROC) with Area Under Curve (AUC). For curiosity only, we track training
time as well.
        </p>
        <p>Threshold based. Accuracy (ACC) is a very simple and an intuitive method
as it measures actual predicted values. However, that makes ACC poor metric
for imbalanced data.</p>
        <p>F1 score is a weighted average of precision and recall. Both false positives
and false negatives are taken into account which makes F1 score usually more
useful than accuracy.</p>
        <p>
          Rank based. Receiver Operating Characteristic (ROC) with Area Under Curve
(AUC) measures how well positive cases are ordered before negative cases [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
That makes it a great metric to evaluate model performance having uneven class
distribution.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental Results</title>
      <p>Experimental Setup. Software used: Python (v3.6), Keras (v2.1.5),
Tensorow (v1.7), XGBoost (v0.71), scikit-learn (v0.19.1). Hardware used: PC with
NVIDIA GPU (GeForce GTX 1080 Ti) and virtual machine (8 cores 2.2 GHz,
16 GB RAM)
Primary task of this research is to compare supervised machine learning
models and nd the most applicable for data certi cation. Each model was
crossvalidated for 10 times. Support Vector Classi er was eliminated at early stage
due to excessive training time. Naive Bayes classi er was eliminated second,
because of poor performance for this task. Other 3 models show good performance
with ROC AUC score over 0.95. The best model which works almost
out-ofthe-box is Gradient Boosted Trees (XGB) with average ROC AUC score 0.987.
Average ROC AUC scores are shown in Fig. 2. Full details about each model
and each metric is in Table 1. Best scores are in bold.</p>
      <p>1.0
0.8
e
taR0.6
e
iitsvPo
eu0.4
r
T
0.2
0.0</p>
      <p>Receiver operating characteristic
0.0
0.2</p>
      <p>ANN (AUC = 0.954)
XGB (AUC = 0.987)
Random Forest (AUC = 0.971)</p>
      <p>Naive Bayes (AUC = 0.706)</p>
      <p>Fa0ls.4e Positive0R.a6te 0.8 1.0
Secondary task follows the hypothetical automatic classi cation process in a
way that model is trained on historical data and only then it is used to classify the
new incoming data. In order to mimic this use-case we sorted dataset timewise
and split into 80:20. All models were trained using same split. Fig. 3 shows ROC
curves with AUC value. All models perform quite well with 94+% AUC score.
1.0
0.8
e
taR0.6
e
v
iits
o
P
eu0.4
r
T
0.2
0.0
0.0
0.2</p>
      <p>ANN (AUC = 0.937)
Random Forest (AUC = 0.968)
XGB (AUC = 0.983)</p>
      <p>Naive Bayes (AUC = 0.942)
0.4 0.6 0.8 1.0</p>
      <p>False Positive Rate</p>
      <p>However due to lucky train-test set distribution Naive Bayes model performs
so well, nearly 20% better than average (see Fig. 2). This behavior once again
proves necessity of cross validation for a proper model validation.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Works</title>
      <p>In this paper we ran experiments with 5 classi cation methods and compared
their performance on CERN CMS JetHT2016 dataset. Best performing model
is Gradient Boosted Trees which ROC AUC score equals 0.987. It seems to be
a su cient score already, but due to imbalanced class distribution (98.2% to
1.8%) it is barely better than "most popular class" classi er. Since only
manual search of hyper-parameters was performed, therefore we believe that full
potential of deep neural network is not yet discovered. Further grid-search of
hyper-parameters (learning rate, dropout rate, batch size, number of neurons,
hidden layers, epochs, etc) should be done to improve performance and predictive
power.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brevdo</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Citro</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Devin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghemawat</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harp</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Irving</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Isard</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jozefowicz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaiser</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kudlur</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levenberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mane</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Monga</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murray</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Olah</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuster</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shlens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steiner</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Talwar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tucker</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanhoucke</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vasudevan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Viegas</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vinyals</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warden</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wattenberg</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wicke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>TensorFlow: Large-scale machine learning on heterogeneous systems (</article-title>
          <year>2015</year>
          ), https://www.tensor ow.org/, software available from tensor ow.org
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Azzolini</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , et al.:
          <article-title>Deep learning for inferring cause of data anomalies (</article-title>
          <year>2017</year>
          ), http://inspirehep.net/record/1637193/ les/arXiv:
          <fpage>1711</fpage>
          .07051.pdf
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Borrello</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>The Data Quality Monitoring Software for the CMS experiment at the LHC</article-title>
          .
          <source>Tech. Rep. CMS-CR-2014-431</source>
          , CERN,
          <source>Geneva (Nov</source>
          <year>2014</year>
          ), http://cds.cern.ch/record/2121269
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Brun</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rademakers</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>ROOT: An object oriented data analysis framework</article-title>
          .
          <source>Nucl. Instrum. Meth. A389</source>
          ,
          <volume>81</volume>
          {
          <fpage>86</fpage>
          (
          <year>1997</year>
          ). https://doi.org/10.1016/S0168-
          <volume>9002</volume>
          (
          <issue>97</issue>
          )
          <fpage>00048</fpage>
          -X
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Buitinck</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Louppe</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mueller</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Niculae</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grobler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Layton</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , VanderPlas, J.,
          <string-name>
            <surname>Joly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Holt</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
          </string-name>
          , G.:
          <article-title>API design for machine learning software: experiences from the scikit-learn project</article-title>
          .
          <source>In: ECML PKDD Workshop: Languages for Data Mining and Machine Learning</source>
          . pp.
          <volume>108</volume>
          {
          <issue>122</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Caruana</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Niculescu-Mizil</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>An empirical comparison of supervised learning algorithms</article-title>
          .
          <source>In: Proceedings of the 23rd International Conference on Machine Learning</source>
          . pp.
          <volume>161</volume>
          {
          <fpage>168</fpage>
          . ICML '06,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2006</year>
          ). https://doi.org/10.1145/1143844.1143865, http://doi.acm.
          <source>org/10</source>
          .1145/1143844.1143865
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Chatrchyan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , et al.:
          <source>The CMS Experiment at the CERN LHC. JINST 3</source>
          ,
          <issue>S08004</issue>
          (
          <year>2008</year>
          ). https://doi.org/10.1088/
          <fpage>1748</fpage>
          -0221/3/08/S08004
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guestrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Xgboost: A scalable tree boosting system</article-title>
          .
          <source>In: Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>785</volume>
          {
          <fpage>794</fpage>
          . KDD '16,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA (
          <year>2016</year>
          ). https://doi.org/10.1145/2939672.2939785, http://doi.acm.
          <source>org/10</source>
          .1145/2939672.2939785
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Chollet</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , et al.: Keras. https://keras.io (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>De Guio</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>The CMS Data Quality Monitoring software experience and future improvements</article-title>
          .
          <source>Tech. Rep. CMS-CR-2013-350</source>
          , CERN, Geneva (Oct
          <year>2013</year>
          ), http://cds.cern.ch/record/2194526
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Rapsevicius</surname>
          </string-name>
          , V.:
          <article-title>CMS Run Registry: Data Certi cation Bookkeeping and Publication System</article-title>
          .
          <source>Tech. Rep. CMS-CR-2011-020</source>
          , CERN,
          <source>Geneva (Jan</source>
          <year>2011</year>
          ), http://cds.cern.ch/record/1345306
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Rovere</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The Data Quality Monitoring Software for the CMS experiment at the LHC</article-title>
          .
          <source>J. Phys.: Conf. Ser</source>
          .
          <volume>664</volume>
          (
          <issue>7</issue>
          ),
          <year>072039</year>
          . 8 p (
          <year>2015</year>
          ), http://cds.cern.ch/record/2159196
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>