<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Concept for Automated Polarized Web Content Annotation based on Multimodal Active Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>University of Kassel</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Wilhelmsh¨oher Allee</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kassel</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Germany</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>marek.herde</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>dhuseljic</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>bsick}@uni-kassel.de</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Passau</institution>
          ,
          <addr-line>Innstrasse 43, 94032 Passau</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Active learning (AL) techniques hardly cope with complex annotations tasks, where, for example, annotations might express relationships across data modalities. As a use case, we consider the task of automatically detecting and reporting multimodal, polarized web content (PWC). Samples of this content type emerge dynamically, covering a broad spectrum of topics. Thus, training machine learning systems for detecting PWC is challenging, particularly if it needs to be done with minimum annotation cost. In this article, we propose the concept of multimodal AL for complex annotations in the context of PWC detection and formulate the resulting challenges as questions for future research.</p>
      </abstract>
      <kwd-group>
        <kwd>Active Learning</kwd>
        <kwd>Multimodal Data</kwd>
        <kwd>Semantic Annotation</kwd>
        <kwd>Polarized Web Content</kwd>
        <kwd>Hateful Memes</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Supervised machine learning (ML) relies on vast amounts of annotated data
often provided by human annotators in a labor-intensive process. Active learning
(AL) addresses this problem of costly data annotation by intelligently querying
annotators [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The goal is to maximize an ML system’s performance while
minimizing the annotation cost. Although AL techniques have shown their benefit
for classification and regression tasks [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], they hardly cope with more complex
annotation tasks, where annotations might
– express relationships across data modalities (A1),
– describe (semantic) relationships between concepts (A2),
– come along with a high level of error-proneness and potential disagreement
among annotators due to an ambiguous context (A3),
– or require modeling background knowledge and sociodemographic factors of
annotators to estimate the quality of annotations (A4).
      </p>
      <p>
        © 2022 for this paper by its authors. Use permitted under CC BY 4.0.
As a use case, we consider the task of automatically detecting and reporting
potential multimodal [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], abusive web content in political communication, which
is in most cases strongly polarized. We use polarized web content (PWC) instead
of related expressions such as hateful memes [
        <xref ref-type="bibr" rid="ref4 ref5">4,5</xref>
        ] to highlight this polarized
nature. Generally, PWC comes in many forms, is subjective, depends on the
context, and frequently requires background knowledge to be understood [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
In this article, we refer to PWC as multimodal online content, mainly text and
images, that can be found on social media and has, e.g., defamatory or abusive
characteristics (at least from the viewpoint of certain groups of persons). The left
side of Fig. 1 shows a PWC sample composed of an image of the burning World
Trade Center on 09/11 and an image of a Muslim congresswoman, Mrs. Ilhan
Abdullahi Omar. These two images are combined with a textual contradiction
of “never forget” and “you have forgotten”. The polarized context arises from
combining images and text (A1), which relates the concepts Twin Towers to
Muslims and terrorism (A2). Identifying this polarization requires knowledge
about American history and politics (A4) or otherwise may result in erroneous
annotations (A3). Such PWC samples emerge dynamically and unforeseeably,
covering a broad spectrum of concepts. Thus, training ML systems for detecting
PWC is challenging, particularly if it needs to be done annotation cost-eficiently.
      </p>
      <p>
        Within this article, we view PWC detection as a challenging sample
application with real-world impact [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] to initiate research on extending AL systems
toward complex annotations of multimodal data. Therefore, we propose our
concept of multimodal active learning for complex annotations (MALCOM) and
formulate the associated challenges as questions for future research.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Concept</title>
      <p>
        We envision MALCOM as an extension of traditional AL [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which assumes a
single omniscient annotator providing categorical labels as annotations, toward
(1) semantic annotation graphs (SAGs) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] as complex, multimodal annotations
and (2) an AL strategy selecting pairs of annotators and queries, e.g., samples.
The objective is to semi-automatically build models that can identify PWC and
analyze it by annotating a potential PWC sample with an SAG. Such an SAG
describes the PWC samples’ contents, explains why its contents can be seen as
polarized, and reflects the potential uncertainty in that analysis. Fig. 1 shows a
PWC sample and its SAG to illustrate this objective. In the following, we outline
our two envisioned extensions of AL and PWC detection in more detail.
      </p>
      <p>
        Extension 1 – Complex, Multimodal Annotations: Existing PWC
detection approaches focus on standard supervised learning settings with
categorical labels as annotations [
        <xref ref-type="bibr" rid="ref1 ref16 ref6">1,6,16</xref>
        ]. The outputs or embeddings of vision and
language models are typically combined as input for a final decision model. Our
proposed SAGs represent an alternative combination strategy for the two
modalities of images and text. SAGs allow decisions on a higher semantic level, which
fosters explainability and decouples objective annotation tasks such as concept
analysis of images and texts from more subjective decisions on polarization. We
argue that this is a more eficient way of generating precise automatic
classifications of PWC. Methodologically, we have to go far beyond annotating images or
text individually but considering their relationships. Annotations may describe
positions of objects in images (regions of interest), comparisons of two images or
texts, the importance of specific contexts for decisions, a degree of polarization,
confidence estimates regarding decisions, etc. We need to develop a proper
semantic model, e.g., ontologies [
        <xref ref-type="bibr" rid="ref12 ref8">8,12</xref>
        ], covering the diferent modalities and being
understandable for annotators. This also includes the ability to express very
different PWC concepts over diferent modalities that go beyond contradictions but
include more fuzzy concepts such as antitheses or correlations between concepts.
      </p>
      <p>
        Extension 2 – Query and Annotator Selection: Identifying PWC
requires contextual knowledge of (very recent) events, e.g., pandemics [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. So
instead of building one generic model, we aim at building specialized models for
diferent kinds of PWC, which use pre-trained models (per modality), and
finetune them in an AL cycle. Extending the AL cycle towards complex annotations
of multimodal data, as sketched in Fig. 2, starts with the question of integrating
diferent modalities. First, we consider a pool of annotated unimodal data, i.e.,
texts and images, which we use to create unimodal models that can annotate
1 Image above is a compilation of assets, including ©Getty Images/Spencer Platt and
©Getty Images/Adam Bettcher, used under the “Hateful Memes Dataset License
Agreement”. It is taken from “The Hateful Memes Challenge” [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for illustrative
purposes only and any person depicted in the content is a model.
      </p>
      <p>useful queries</p>
      <p>selects
query-annotator pairs</p>
      <p>annotated queries
extends training set</p>
      <p>and ontology
Error-prone Annotators</p>
      <p>Annotator</p>
      <p>Model
Active
Learning
Strategy</p>
      <p>ML Models
Query Set
Text Data
Text 1: We blame ...</p>
      <p>Text 2: We blame ...</p>
      <p>Text 3: We blame ...
...</p>
      <p>Text N: We blame ...</p>
      <p>(1)
(4)
Training Set</p>
      <p>Text Data
Text 49: We blame ...</p>
      <p>Text 67: We blame ...</p>
      <p>Annotations
(2)
(3)
Image Data</p>
      <p>SAGs</p>
      <p>Image Data</p>
      <p>
        SAGs
assesses utilities
of queries
controls training/
fine-tuning procedure
information through models
updated PWC corpus
the unimodal data semantically. This process in each case results in an SAG,
i.e., a typed, attributed graph defined through an ontology-based annotation
scheme. Later, the SAGs are merged into a joint, multimodal SAG. Similar to
traditional AL strategies, we need to identify promising candidates – initially
images and texts, later multimodal SAGs – to be annotated. To consider the
problem’s multimodal nature, the annotations’ semantic properties, and the
annotators’ diverse backgrounds, we must develop new AL selection strategies that
account not only for the respective data sample but also for the diferent kinds of
queries and the qualifications of certain annotators regarding the PWC sample
at hand. These qualifications (also referred to as annotator performance [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]) may
depend on various aspects such as the respective PWC category (e.g., politics)
or educational background (e.g., Master’s degree in political sciences). The
annotator model predicting such qualifications needs to be sensitive to annotator
minorities, e.g., by estimating similarities between annotators. Otherwise, we
risk ignoring annotator minorities’ opinions regarding PWC. Moreover, we must
consider that answers regarding the degree to which content is polarized may
be highly subjective, i.e., uncertain from an ML perspective [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Establishing an
objective definition of PWC, similar to hate speech research [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], is a possible
way of reducing the subjectivity of PWC annotation.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Research Questions</title>
      <p>We conclude this article with the following six research questions derived from
the above key research objective and the required extensions.</p>
      <p>
        – How can we define ontology-based annotation schemes to express a human’s
reasoning over classifying web content as (gradually) polarized or not?
– How can we extract image descriptions (part of the SAG) from potentially
polarized images (part of the PWC) considering diferent uncertainty types?
– How can we extend AL for object detection in potentially polarized images?
– How can we extend AL over text extracted from the images to identify
rhetorical figures and automatically analyze textual content to create
semantic annotations automatically?
– How can we merge unimodal SAGs and extend AL to train models, e.g.,
graph convolutional networks [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], assessing PWC via multimodal SAGs?
– How can we evaluate the above techniques and build or extend data
corpora [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for research?
      </p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgements</title>
      <p>SPONSORED BY THE</p>
      <p>The project on which this article is based was partly funded
by the German Federal Ministry of Education and Research
(BMBF) under the funding code 01|S20049. The authors are
responsible for the content of this publication. Furthermore,
the authors thank Chandana Priya Nivarthi, Stephan Vogt,
Mohammad Wazed Ali, and the anonymous reviewers for
their insightful comments to improve this article.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gibert</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karatzas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Exploring Hate Speech Detection in Multimodal Publications</article-title>
          . In: WACV. pp.
          <fpage>1470</fpage>
          -
          <lpage>1478</lpage>
          . Snowmass Village, CO (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Herde</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huseljic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A Survey on Cost Types, Interaction Schemes, and Annotator Performance Models in Selection Algorithms for Active Learning in Classification</article-title>
          .
          <source>IEEE Access 9</source>
          ,
          <fpage>166970</fpage>
          -
          <lpage>166989</lpage>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Huseljic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sick</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Herde</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kottke</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Separation of Aleatoric and Epistemic Uncertainty in Deterministic Deep Neural Networks</article-title>
          .
          <source>In: ICPR</source>
          . pp.
          <fpage>9172</fpage>
          -
          <lpage>9179</lpage>
          .
          <string-name>
            <surname>Virtual</surname>
          </string-name>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Kiela</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Firooz</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goswami</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fitzpatrick</surname>
            ,
            <given-names>C.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bull</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lipstein</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nelli</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , et al.:
          <source>The Hateful Memes Challenge: Competition Report. In: NeurIPS 2020 Competition and Demonstration Track</source>
          . pp.
          <fpage>344</fpage>
          -
          <lpage>360</lpage>
          .
          <string-name>
            <surname>Virtual</surname>
          </string-name>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Kiela</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Firooz</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goswami</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ringshia</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Testuggine</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The Hateful Memes Challenge: Detecting Hate Speech in Multimodal Memes</article-title>
          . In: NeurIPS. pp.
          <fpage>2611</fpage>
          -
          <lpage>2624</lpage>
          .
          <string-name>
            <surname>Virtual</surname>
          </string-name>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sachdeva</surname>
          </string-name>
          , N.:
          <article-title>Multimodal cyberbullying detection using capsule network with dynamic routing and deep convolutional neural network</article-title>
          .
          <source>Multimed. Syst</source>
          . (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Active Learning Query Strategies for Classification, Regression, and Clustering: A Survey</article-title>
          .
          <source>JCST</source>
          <volume>35</volume>
          (
          <issue>4</issue>
          ),
          <fpage>913</fpage>
          -
          <lpage>945</lpage>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8. Ku¨hn, R., Mitrovi´c, J.,
          <string-name>
            <surname>Granitzer</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>GRhOOT: Ontology of Rhetorical Figures in German</article-title>
          . In: LREC. Marseille, France (
          <year>2022</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Lahat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adali</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jutten</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Multimodal Data Fusion: An Overview of Methods, Challenges, and Prospects</article-title>
          .
          <source>Proceedings of the IEEE</source>
          <volume>103</volume>
          (
          <issue>9</issue>
          ),
          <fpage>1449</fpage>
          -
          <lpage>1477</lpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>MacAvaney</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yao</surname>
            ,
            <given-names>H.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Russell</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frieder</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Hate speech detection: Challenges and solutions</article-title>
          .
          <source>PLOS ONE 14(8)</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Mishra</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yannakoudakis</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shutova</surname>
          </string-name>
          , E.:
          <article-title>Tackling Online Abuse: A Survey of Automated Abuse Detection Methods</article-title>
          . arXiv:
          <year>1908</year>
          .
          <volume>06024</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. Mitrovi´c, J.,
          <string-name>
            <given-names>O</given-names>
            <surname>'Reilly</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          , Mladenovi´c,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Handschuh</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          :
          <article-title>Ontological representations of rhetorical figures for argument mining</article-title>
          .
          <source>Argument &amp; Computat. 8</source>
          (
          <issue>3</issue>
          ),
          <fpage>267</fpage>
          -
          <lpage>287</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Sood</surname>
            ,
            <given-names>S.O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Antin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Churchill</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Using Crowdsourcing to Improve Profanity Detection</article-title>
          . In: AAAI Spring Symposium 2012 -
          <article-title>Wisdom of the Crowd</article-title>
          . pp.
          <fpage>69</fpage>
          -
          <lpage>74</lpage>
          . Palo Alto, CA (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Uyheng</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carley</surname>
            ,
            <given-names>K.M.</given-names>
          </string-name>
          :
          <article-title>Bots and online hate during the COVID-19 pandemic: case studies in the United States and the Philippines</article-title>
          .
          <source>JCSS</source>
          <volume>3</volume>
          (
          <issue>2</issue>
          ),
          <fpage>445</fpage>
          -
          <lpage>468</lpage>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Vidal</surname>
            ,
            <given-names>J.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lama</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Otero-Garc´ıa, E., Bugar´ın, A.:
          <article-title>Graph-based semantic annotation for enriching educational content with linked data</article-title>
          .
          <source>KBS 55</source>
          ,
          <fpage>29</fpage>
          -
          <lpage>42</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shilon</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Ma, H.,
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Predovic</surname>
          </string-name>
          , G.:
          <article-title>Exploring Deep Multimodal Fusion of Text and Photo for Hate Speech Classification</article-title>
          . In: ALW. pp.
          <fpage>11</fpage>
          -
          <lpage>18</lpage>
          . Florence,
          <string-name>
            <surname>Italy</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tong</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maciejewski</surname>
          </string-name>
          , R.:
          <article-title>Graph convolutional networks: a comprehensive review</article-title>
          .
          <source>Comput. Soc. Netw</source>
          .
          <volume>6</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>