<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Supporting Journalism by Combining Neural Language Generation and Knowledge Graphs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marco Cremaschi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Federico Bianchi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Maurino</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Primo Pierotti</string-name>
          <email>a.pierotti1@campus.unimib.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Sciences, Systems and Communications University of Milan-Bicocca Viale Sarca</institution>
          ,
          <addr-line>336 - 20126, Milan</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Natural Language Generation is a field that is becoming relevant in several domains, including journalism. Natural Language Generation techniques can be of great help to journalists, allowing a substantial reduction in the time required to complete repetitive tasks. In this position paper, we enforce the idea that automated tools can reduce the effort required to journalist when writing articles; at the same time we introduce GazelLex (Gazette Lexicalization), a prototype that covers several steps of Natural Language Generation, in order to create soccer articles automatically, using data from Knowledge Graphs, leaving journalists the possibility of refining and editing articles with additional information. We shall present our first results and current limits of the approach, and we shall also describe some lessons learned that might be useful to readers that want to explore this field.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Although automation is a phenomenon that is
becoming more and more visible today, there are
specialised jobs that require human effort to be
completed. The job of a journalist is among these
        <xref ref-type="bibr" rid="ref11">( O¨rnebring, 2010)</xref>
        . However, recent technological
progress in the field of Natural Language
Generation (NLG) and the use of increasingly
sophisticated techniques of artificial intelligence allow the
use of software capable of writing newspaper
articles almost indistinguishable from human ones.
These techniques can help journalists reduce the
      </p>
      <p>
        Copyright 2019 for this paper by its authors. Use
permitted under Creative Commons License Attribution 4.0
International (CC BY 4.0).
effort needed for repetitive tasks, such as data
collection and drafting writing. The name given to
this phenomenon is Automated Journalism; this
new type of journalism uses algorithms to generate
news under human supervision. During the past
years, several newsrooms have begun to
experiment this technology: Associated Press, Forbes,
Los Angeles Times, and ProPublica are among the
first, but adoption could spread out soon
        <xref ref-type="bibr" rid="ref7">(Graefe,
2016)</xref>
        . Automated Journalism can bring a massive
change to the sector: writing news is a business
that endeavours to minimise costs while
maintaining maximum efficiency and full speed, and thanks
to this software the above-mentioned objectives
can be achieved, generating good-quality articles
        <xref ref-type="bibr" rid="ref19">(van Dalen, 2012)</xref>
        . This new technology provides
many advantages: the most evident are speed and
the scale of news coverage. Of course, there are
also problems and limitations. One of the most
relevant is the dependence from structured data
        <xref ref-type="bibr" rid="ref7">(Graefe, 2016)</xref>
        , that is the reason why sports
reports, financial articles, and forecasts are the most
covered topics by software: they are all domains
where the complexity of the topic can be managed
from software using structured data. Similar
structured data are not always available in other fields.
In order to generate valuable text, approaches
considering data contained in the Knowledge Graphs
(KGs) have recently been introduced in literature
        <xref ref-type="bibr" rid="ref10 ref15 ref20 ref5 ref9">(Gardent et al., 2017; Trisedya et al., 2018)</xref>
        .
      </p>
      <p>
        A Knowledge Graph (KG) describes real-world
entities and the relations between them. KGs are
an essential source of information, and their
features allow the use of this information in different
contexts, such as link prediction
        <xref ref-type="bibr" rid="ref12 ref18 ref21">(Trouillon et al.,
2016)</xref>
        and recommendation
        <xref ref-type="bibr" rid="ref12 ref18 ref21">(Zhang et al., 2016)</xref>
        .
Popular KGs are the Google Knowledge Graph,
Wikidata and DBpedia
        <xref ref-type="bibr" rid="ref1">(Auer et al., 2007)</xref>
        .
Entities are defined in an ontology and thus can be
classified using a series of types. The primary
element of a KG to store entities information is a
Resource Description Framework (RDF) triple in
the format hsubject; predicate; objecti. As RDF
triples open many possibilities in Web data
representation, utilising this data also in the NLG
context is valuable
        <xref ref-type="bibr" rid="ref12 ref18 ref21">(Perera et al., 2016)</xref>
        . Interlinked
KGs can be used to automatically extend the
information relating to a given entity in an article.
      </p>
      <p>In our solution, we use DBpedia, one of the
fastest growing Linked Data resource that is
available free of charge; it is characterised by a high
number of links from the Linked Data Cloud2.
DBpedia is thus a central interlinking hub, an
access point for retrieving information to be inserted
in an article, as specified below.</p>
      <p>
        Up to 2010, commercial providers in the NLG
field were not popular, but in the last years few
companies have started to provide this kind of
services. In 2016 there were 13 companies covering
this field
        <xref ref-type="bibr" rid="ref4">(Drr, 2016)</xref>
        (e.g., AutomatedInsights3,
NarrativeScience4). Approaches that try to
integrate deep networks and text generation are now
common in literature
        <xref ref-type="bibr" rid="ref20 ref5 ref9">(Gardent et al., 2017)</xref>
        . These
automated tools are going to become a standard
method to help journalist during the news writing
process.
      </p>
      <p>
        We shall concentrate on examples of related
work in the context of lexicalization from RDF
data, we shall refer to surveys from the state of
the art for a more detailed overview of the field
        <xref ref-type="bibr" rid="ref10 ref13 ref15 ref6 ref8">(Reiter and Dale, 1997; Gatt and Krahmer, 2018;
Moussallem et al., 2018)</xref>
        . Semantic web
technologies like RDF can be used to enhance the power of
current algorithms
        <xref ref-type="bibr" rid="ref3">(Bouayad-Agha et al., 2012)</xref>
        .
The WebNLG challenge
        <xref ref-type="bibr" rid="ref20 ref5 ref9">(Gardent et al., 2017)</xref>
        has
been introduced to study the possibilities given by
the combination of deep learning techniques and
semantic web technologies. In a similar context,
an approach based on Long Short-Term Memory
(LSTM) networks has been proposed to generate
text lexicalizations from RDF triples
        <xref ref-type="bibr" rid="ref10 ref15">(Trisedya et
al., 2018)</xref>
        .
      </p>
      <p>In this work, we aim to describe what is the
possible automation process that can be used to
help journalist in the news writing process. At the
same time we describe a new prototype we have
created to support journalistic activities, GazelLex
(Gazette Lexicalization). GazelLex, through the
use of deep learning techniques implements a
2https://wiki.dbpedia.org/
dbpedia-2016-04-statistics
3https://automatedinsights.com/
4https://narrativescience.com/
Neural Machine Translation (NMT) approach to
generate articles (sentences) starting from data
composed by RDF triples. GazelLex is also able
to generate videos containing the images and the
prominent information of the article, and to
generate audio using a speech synthesis module (Figure
1). To the best of our knowledge, our prototype
is the first to provide an all-in-one integrated
approach to NLG with RDF triples in the context of
helping journalist in writing articles.</p>
      <p>This paper is structured as follows: in
Section 2, we analyse the state-of-the-art on Natural
Language Generation, showing that these
methods to generate natural language are becoming
popular. In Section 3 we describe our
prototype, GazelLex, that combines neural methods and
knowledge graphs to create soccer articles and
describe how this kind of tools can be of help to
journalism. In Section 4 we show a preliminary
experimental analysis, while in Section 5 we provide
conclusions.</p>
      <p>CONTENT DETERMINATION</p>
      <p>ATTRIBUTES
DEFINITION</p>
      <p>SCHEMA</p>
      <p>DEFINITION
TRAINING
TEXT GENERATION</p>
      <p>INFO
EXTRACTION</p>
      <p>RDF
TRIPLES</p>
      <p>OLD ARTICLE</p>
      <p>RDF
TRIPLES</p>
      <p>TEXT
STRUCTURING</p>
      <p>MODEL FOR
TRAINING DATASET
TRAINING DATASET</p>
      <p>GENERATION
UX</p>
      <p>DEEP</p>
      <p>
        LEARNING
VIDEO AUDIO NEW ARTICLE
NLG is a “sub-field of artificial intelligence and
computational linguistics that is concerned with
the construction of computer systems that can
produce understandable texts in English or other
human languages from some underlying
nonlinguistic representation of information”
        <xref ref-type="bibr" rid="ref13 ref14 ref8">(Reiter
and Dale, 1997; Reiter and Dale, 2000)</xref>
        . In NLG
six “problems” must be addressed: Content
determination: input data that is always more
detailed and richer than what we want to cover in the
text
        <xref ref-type="bibr" rid="ref6">(Gatt and Krahmer, 2018)</xref>
        and so the aim is to
filter and choose what to say. Text structuring:
a clear text structure and the order of presentation
of information are critical for readers, for this
reason, pre-defining the templates is necessary.
Sentence aggregation: sentences must not be
disconnected. Text needs therefore to be grouped in such
a way that a “more fluid and readable” text
        <xref ref-type="bibr" rid="ref6">(Gatt
and Krahmer, 2018)</xref>
        is generated. Lexicalization:
one of the most critical phases of NLG process is
how to express message blocks through words and
phrases. This task is called lexicalization and
concerns the actual conversion from messages to
natural language. Reference expression generation:
to avoid repetitions, selecting ways to refer to
entities using different methods (such as pronouns,
proper nouns, or descriptions) is essential.
Linguistic realisation: it concerns the combination
of relevant words and phrases to form a sentence.
      </p>
      <p>
        As we stated above, lexicalization is one of the
most critical and complex tasks in the NLG
process. Natural language vagueness and choosing
the right words to express a concept are
intricate issues to manage. Looking at the
state-ofthe-art, we see that recent research on this topic
shows that an interesting solution in these cases is
based on Machine Learning (ML)
        <xref ref-type="bibr" rid="ref6">(Gatt and
Krahmer, 2018)</xref>
        . Moreover, a recent challenge in the
NLG field, launched and published in 2017, called
WebNLG
        <xref ref-type="bibr" rid="ref20 ref5 ref9">(Gardent et al., 2017)</xref>
        confirms the idea
that not only we need to combine ML methods to
generate language, but we can also use KGs to
enrich sentences with additional contextual
information (e.g., contextual information about a player).
3
      </p>
    </sec>
    <sec id="sec-2">
      <title>GazelLex</title>
      <p>
        In this section, we shall give an example of the
NLG process in a domain specific view. As
introduced previously, we developed a software, named
GazelLex, that can produce soccer articles. There
are two main reasons for this choice: first of all,
the project was partly commissioned by an
Italian newspaper publisher. Furthermore, soccer and
sports, in general, are good domains to develop
NLG, because they are complex enough to be
challenging, yet they are easy to manage and many
data exist
        <xref ref-type="bibr" rid="ref2">(Barzilay and Lapata, 2005)</xref>
        . In this
scenario we focused our attention on the final
output, using a solution that combines neural network
with some handcrafted processes. We would like
to underline that the data related to the games (e.g.
number of goals, training) are extracted
automatically from online services.
      </p>
      <p>
        Our approach is divided into five tasks, in
order to address the five classic NLG sub-problems
        <xref ref-type="bibr" rid="ref6">(Gatt and Krahmer, 2018)</xref>
        : in the following, for
each phase, implementation details will be
provided.
3.1
      </p>
      <sec id="sec-2-1">
        <title>Content Determination</title>
        <p>
          To select the most relevant information, a
handcrafted approach was chosen. To select the
information to bring in the final output, we traced the
most used data in soccer articles. One of the
primary references was PASS, a personalised
automated text system developed to write soccer
articles
          <xref ref-type="bibr" rid="ref20 ref5 ref9">(van der Lee et al., 2017)</xref>
          . We took the kind
of information PASS used to fill its templates and
enriched them with our data fields. So we have
some entities of type “TEAM”, “FORMATION”,
“COACH” and some predicates like “injuryAt”,
“yellowCardAt”, and “violentFoulAt”5. The
software used this data to create triples, that
algorithms used to write the article.
3.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Text Structuring</title>
        <p>
          Being a domain specific process, we developed a
handcrafted template, based on real articles.
Aiming to get a similar output we imitated the
journalist’s job in the division of text and about
information contained in each part. We also considered the
text structuring approach usually developed in this
domain, that uses more general information and
after that a chronological order
          <xref ref-type="bibr" rid="ref6">(Gatt and Krahmer,
2018)</xref>
          . In GazelLex, it is possible to find templates
(e.g., complete or short article) resulting from the
process described above, but it is also possible to
modify them or create new ones (Fig. 2).
In soccer data, many events could be redundant
when written in an article. If a player scores
a hat trick in a match writing the same
sentence about each goal would be unpleasant to read
while grouping them in a single sentence could be
more concise and coherent. This task “focused
on domain- and application-specific rules”
          <xref ref-type="bibr" rid="ref6">(Gatt
and Krahmer, 2018)</xref>
          . We aggregated the RDF
triples defined in the preceding section to
generate a group of triples that represents the content of
our news article.
Like we said above, we considered lexicalization
like a NMT process, converting RDF data into
natural language. To achieve this aim, we used a
specific kind of neural network: LSTM
          <xref ref-type="bibr" rid="ref13 ref8">(Hochreiter
and Schmidhuber, 1997)</xref>
          . Their recent success in
NLG field is related to many advantages they
provide. Compared to the traditional neural network,
LSTM do not have limitations in input and
output length. Furthermore, input and output are not
independent, that is a vital advantage in language
generation. To predict a word in a sentence it is
useful to know and consider the previous one, and
the hidden states of the network keep the
memory about what happened in previous timesteps. In
this way, LSTM can combine the previous state,
the memory collected and the input, allowing
dependencies to be maintained in the long term. We
experimented NMT using a now widly recognized
tool for neural machine translation6
          <xref ref-type="bibr" rid="ref20 ref5 ref9">(Klein et al.,
2017)</xref>
          . Our neural architecture is based on a
standard encoder-decoder structure with 4 LSTM
layers containing 200 hidden neurons on both the
encoder and the decoder. Input tokenization is based
on the space character (recall that our RDF triples’
elements are separated by spaces).
3.5
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Reference expression generation</title>
        <p>We used different databases to avoid redundancy
and give a fluent text to the reader. Some online
resources help us to create a list of possible
replacements for a team or players’ name. Using
DBpedia, we can find a nickname for an entity
(Real Madrid players are also called Blancos or
Merengues). Other resources we used are
Wikidata list of soccer teams nicknames and Topend
Sports database.</p>
        <sec id="sec-2-3-1">
          <title>6http://opennmt.net</title>
          <p>
            In the following section, we shall show some
insights into our tool and on how it works. We
shall present a use case, a recent soccer match, for
which the generation process and the resulting text
will be shown. The initial dataset for the training
was created manually and consists of 4387 pairs
of triples and lexicalizations. We drew inspiration
from the state-of-the-art to devise the architecture
of our network
            <xref ref-type="bibr" rid="ref10 ref15 ref20 ref5 ref9">(Gardent et al., 2017; Trisedya
et al., 2018)</xref>
            . From our primary experiments the
best performing model required two layers of
bidirectional LSTM, but still, the model suffers from
some limitations (outlined in the related sec.).
To show the valid output of GazelLex, we took
an example match and generated its lexicalization.
We considered the football match played by
Juventus F.C. and A.C Chievo on the 21st of
January. Our application gathered data from an
online provider and converted data in a triple format.
A journalist can edit settings using a form (Figure
3): the journalist is in charge of deciding what is
worth writing in the article and how it should
appear to the end-user; we recall that we can also
define templates for our articles (Figure 2). The
final output of this process looks like the one that is
shown in Figure 4. GazelLex, in order to improve
the quality of the sentences and to obtain results as
close to the style of the journalist as possible (i.e.
style transfer), cyclically re-executes the training
phase using the sentences validated by the
journalist. The following is an example of lexicalization
of triples relative to the use case (Table 1).
In this section we would like to outline the
current limitations of our project and also report a
few lessons learned that might be useful for other
researchers who are currently exploring this field.
One key part of the development process comes
from the definition or the selection of a good
Knowledge Graph that can support the
lexicalization; moreover, the definition of the new RDF
predicates is a difficult process that must be done
carefully to avoid errors in the next steps. Our
application currently supports the lexicalization of
a small set of triples (i.e., we focused on goals
and final result); we decided to concentrate on this
small set to generate a set of resulting sentences
that can be manually inspected for quality. Our
NLG model is based on a deep learning
architecture, and thus some of the generated sentences are
not well-formed owing to the structure of the net
itself. While this is a problem that has to be solved
in our settings, we have a journalist reviewing the
article before it is released to the public: this
allows us to have a model that is more flexible than
standard pattern-based NLG, while the precision
of the output can be controlled in a
human-in-theloop setting. Regarding the configuration of our
model, we have replicated the state-of-the-art
experiments (i.e. approaches explained in
            <xref ref-type="bibr" rid="ref20 ref5 ref9">(Gardent
et al., 2017)</xref>
            ) and we are currently experimenting
those architectures on our domain dataset. The
results are yet to be quantitatively validated and they
are preliminary, but they are promising as reported
by journalists. In the future, we are planning to
carefully explore various architecture and consider
the use of word embeddings to solve some of our
current issues.
5
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Conclusion</title>
      <p>In this position paper we have analysed the
future possibilities given by automated journalism.
We have summarised the current state of art on
this topic showing that there is an increasing
interest towards automated natural language generation
for the news sector. While hereby, we showed an
application related to the soccer domain, the
principles and the methodologies described are
general, and they can be used in other fields (e.g.,
finance, weather reporting). We strongly believe
that these tools can greatly help journalists in
working on what is really important (e.g.,
investigation, fact checking), leaving high effort, but
low value tasks to computers. The prototype we
have described is a first step towards this
automated process and its results are surely
promising.</p>
    </sec>
    <sec id="sec-4">
      <title>Acknowledgments</title>
      <p>This research has been supported in part by the
Robo-Journalism project 2018-comm25-0047 in
collaboration with Instal S.r.l.7. Special thanks to
Carlo Mattioli and Alessandra Siano for their
support during the development of the project.</p>
      <sec id="sec-4-1">
        <title>7instal.com</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Auer et al.2007]
          <article-title>So¨ren Auer, Christian Bizer</article-title>
          , Georgi Kobilarov, Jens Lehmann, Richard Cyganiak, and
          <string-name>
            <given-names>Zachary</given-names>
            <surname>Ives</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Dbpedia: A nucleus for a web of open data</article-title>
          .
          <source>In The Semantic Web</source>
          , pages
          <fpage>722</fpage>
          -
          <lpage>735</lpage>
          , Berlin, Heidelberg. Springer Berlin Heidelberg.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <source>[Barzilay and Lapata2005] Regina Barzilay and Mirella Lapata</source>
          .
          <year>2005</year>
          .
          <article-title>Collective content selection for concept-to-text generation</article-title>
          .
          <source>In Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing, HLT '05</source>
          , pages
          <fpage>331</fpage>
          -
          <lpage>338</lpage>
          , Stroudsburg, PA, USA. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [
          <string-name>
            <surname>Bouayad-Agha</surname>
          </string-name>
          et al.
          <year>2012</year>
          ]
          <string-name>
            <given-names>N</given-names>
            <surname>Bouayad-Agha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G</given-names>
            <surname>Casamayor</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L</given-names>
            <surname>Wanner</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Natural language generation and semantic web technologies</article-title>
          .
          <source>Semantic Web Journal.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Drr2016] Konstantin Nicholas Drr</source>
          .
          <year>2016</year>
          .
          <article-title>Mapping the field of algorithmic journalism</article-title>
          .
          <source>Digital Journalism</source>
          ,
          <volume>4</volume>
          (
          <issue>6</issue>
          ):
          <fpage>700</fpage>
          -
          <lpage>722</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [Gardent et al.2017]
          <string-name>
            <given-names>Claire</given-names>
            <surname>Gardent</surname>
          </string-name>
          , Anastasia Shimorina, Shashi Narayan, and
          <string-name>
            <surname>Laura</surname>
          </string-name>
          Perez-Beltrachini.
          <year>2017</year>
          .
          <article-title>The webnlg challenge: Generating text from rdf data</article-title>
          .
          <source>In Proceedings of the 10th International Conference on Natural Language Generation</source>
          , pages
          <fpage>124</fpage>
          -
          <lpage>133</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>[Gatt and Krahmer2018] Albert Gatt and Emiel Krahmer</source>
          .
          <year>2018</year>
          .
          <article-title>Survey of the state of the art in natural language generation: Core tasks, applications and evaluation</article-title>
          .
          <source>J. Artif. Int. Res.</source>
          ,
          <volume>61</volume>
          (
          <issue>1</issue>
          ):
          <fpage>65</fpage>
          -
          <lpage>170</lpage>
          , January.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [Graefe2016]
          <string-name>
            <given-names>Andreas</given-names>
            <surname>Graefe</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Guide to automated journalism</article-title>
          .
          <source>Technical report</source>
          , Tow Center for Digital Journalism, Columbia University.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>[Hochreiter and Schmidhuber1997] Sepp Hochreiter</article-title>
          and
          <string-name>
            <given-names>Jrgen</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>Long short-term memory</article-title>
          .
          <source>Neural Computation</source>
          ,
          <volume>9</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [Klein et al.2017]
          <string-name>
            <given-names>Guillaume</given-names>
            <surname>Klein</surname>
          </string-name>
          , Yoon Kim, Yuntian Deng, Jean Senellart, and
          <string-name>
            <given-names>Alexander</given-names>
            <surname>Rush</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>OpenNMT: Open-source toolkit for neural machine translation</article-title>
          .
          <source>In Proceedings of ACL</source>
          <year>2017</year>
          ,
          <string-name>
            <given-names>System</given-names>
            <surname>Demonstrations</surname>
          </string-name>
          , pages
          <fpage>67</fpage>
          -
          <lpage>72</lpage>
          , Vancouver, Canada, July. Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [Moussallem et al.
          <year>2018</year>
          ] Diego Moussallem, Matthias Wauer, and
          <string-name>
            <surname>Axel-Cyrille Ngonga Ngomo</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Machine translation using semantic web technologies: A survey</article-title>
          .
          <source>Journal of Web Semantics</source>
          ,
          <volume>51</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [O¨ rnebring2010]
          <string-name>
            <surname>Henrik</surname>
            <given-names>O¨ rnebring.</given-names>
          </string-name>
          <year>2010</year>
          .
          <article-title>Technology and journalism-as-labour: Historical perspectives</article-title>
          .
          <source>Journalism</source>
          ,
          <volume>11</volume>
          (
          <issue>1</issue>
          ):
          <fpage>57</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [Perera et al.2016]
          <string-name>
            <given-names>Rivindu</given-names>
            <surname>Perera</surname>
          </string-name>
          , Parma Nand, and
          <string-name>
            <given-names>Gisela</given-names>
            <surname>Klette</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Realtext-lex: A lexicalization framework for rdf triples</article-title>
          .
          <source>The Prague Bulletin of Mathematical Linguistics</source>
          ,
          <volume>106</volume>
          (
          <issue>1</issue>
          ):
          <fpage>45</fpage>
          -
          <lpage>68</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <source>[Reiter and Dale1997] Ehud Reiter and Robert Dale</source>
          .
          <year>1997</year>
          .
          <article-title>Building applied natural language generation systems</article-title>
          .
          <source>Nat. Lang</source>
          . Eng.,
          <volume>3</volume>
          (
          <issue>1</issue>
          ):
          <fpage>57</fpage>
          -
          <lpage>87</lpage>
          ,
          <year>March</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Reiter and Dale2000] Ehud Reiter and Robert Dale</source>
          .
          <year>2000</year>
          .
          <article-title>Building Natural Language Generation Systems</article-title>
          . Cambridge University Press, New York, NY, USA.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [Trisedya et al.2018]
          <string-name>
            <given-names>Bayu</given-names>
            <surname>Distiawan</surname>
          </string-name>
          <string-name>
            <surname>Trisedya</surname>
          </string-name>
          , Jianzhong Qi, Rui Zhang, and
          <string-name>
            <given-names>Wei</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Gtr-lstm: A triple encoder for sentence generation from rdf data</article-title>
          .
          <source>In Proceedings of the 56th Annual Meeting of the Association for Computational</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Linguistics</surname>
          </string-name>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          , volume
          <volume>1</volume>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          pages
          <fpage>1627</fpage>
          -
          <lpage>1637</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [Trouillon et al.
          <year>2016</year>
          ]
          <article-title>The´o Trouillon, Johannes Welbl</article-title>
          ,
          <string-name>
            <given-names>Sebastian</given-names>
            <surname>Riedel</surname>
          </string-name>
          , E´ ric Gaussier, and
          <string-name>
            <given-names>Guillaume</given-names>
            <surname>Bouchard</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Complex embeddings for simple link prediction</article-title>
          .
          <source>In International Conference on Machine Learning</source>
          , pages
          <fpage>2071</fpage>
          -
          <lpage>2080</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [van Dalen2012]
          <string-name>
            <surname>Arjen van Dalen</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>The algorithms behind the headlines</article-title>
          .
          <source>Journalism Practice</source>
          ,
          <volume>6</volume>
          (
          <issue>5</issue>
          -6):
          <fpage>648</fpage>
          -
          <lpage>658</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>[van der Lee</surname>
          </string-name>
          et al.2017]
          <article-title>Chris van der Lee, Emiel Krahmer</article-title>
          , and
          <string-name>
            <given-names>Sander</given-names>
            <surname>Wubben</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>PASS: A Dutch data-to-text system for soccer, targeted towards specific audiences</article-title>
          .
          <source>In Proceedings of the 10th International Conference on Natural Language Generation</source>
          , pages
          <fpage>95</fpage>
          -
          <lpage>104</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [Zhang et al.2016]
          <string-name>
            <given-names>Fuzheng</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Nicholas Jing Yuan, Defu Lian, Xing Xie, and
          <string-name>
            <surname>Wei-Ying Ma</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Collaborative knowledge base embedding for recommender systems</article-title>
          .
          <source>In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          , pages
          <fpage>353</fpage>
          -
          <lpage>362</lpage>
          . ACM.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>