<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging Knowledge Graphs with Large Language Models for Classification Tasks in the Tourism Domain</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Cadeddu</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alessandro Chessa</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vincenzo De Leo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gianni Fenu</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Motta</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Osborne</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Diego Reforgiato Recupero</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Angelo Salatino</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luca Secchi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Linkalab s.r.l.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cagliari</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Italy</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Business and Law, University of Milano Bicocca</institution>
          ,
          <addr-line>Milan</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Mathematics and Computer Science, University of Cagliari</institution>
          ,
          <addr-line>Cagliari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Knowledge Media Institute, The Open University</institution>
          ,
          <addr-line>Milton Keynes</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Online platforms, serving as the primary conduit for travelers to seek, compare, and secure travel accommodations, require a profound understanding of user dynamics to craft competitive and enticing oferings. Concurrently, recent advancements in Natural Language Processing, particularly large language models, have made substantial strides in capturing the complexity of human language. Simultaneously, knowledge graphs have become a formidable instrument for structuring and categorizing information. This paper introduces a cutting-edge deep learning methodology integrating large language models with domain-specific knowledge graphs to classify tourism ofers. It aims at aiding hospitality operators in understanding their accommodation oferings' market positioning, taking into account the visit propensity and user review ratings, with the goal of optimizing the ofers themselves and enhancing their appeal. Comparative analysis against alternative methods on two datasets of London accommodation ofers attests to our approach's efectiveness, demonstrating superior results.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Knowledge Graphs</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>BERT</kwd>
        <kwd>Classification Tasks</kwd>
        <kwd>Feature Engineering</kwd>
        <kwd>Tourism</kwd>
        <kwd>Hospitality</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the age of digital transition, online platforms are a vital tool for travelers to explore and
reserve travel accommodations. Yet, with the surge of information, users can find it challenging
to choose the optimal option. Moreover, individual traveler preferences, such as location,
amenities, and price, further complicate this task. Hence, understanding these dynamics is
crucial for online platforms and revenue managers in the hospitality sector to better position
their accommodations in an ever-competitive market [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Advancements in Natural Language
Processing (NLP), notably large language models based on transformers, have greatly improved
automatic human language comprehension [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Similarly, knowledge graphs (KG) have gained
recognition as valuable tools for structured information organization [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. They provide a
semantic representation of all significant entities within a domain, serving as a powerful
resource for conveying valuable information to intelligent services.
      </p>
      <p>
        However, the integration of these technologies presents challenges, primarily in blending
unstructured and structured data and correctly encoding knowledge graph information [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>This paper introduces a novel deep-learning methodology integrating large language models
and domain knowledge graphs for tourism ofers classification. Specifically, it augments a
transformer model with a knowledge graph generated from Airbnb data to accurately categorize
accommodation descriptions. We evaluated our approach against a BERT (Bidirectional Encoder
Representations from Transformers) classifier and a baseline logistic regression classifier on a
dataset of over 15,000 accommodation ofers, yielding excellent results.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Integrating structured knowledge into deep learning architectures has been the focal point
of numerous studies. Typically the source of knowledge is encoded as knowledge graphs, as
they facilitate reasoning and can be refined and enhanced by employing diverse techniques for
knowledge graph completion. Furthermore, KG about tourism can be generated from a broad
spectrum of data sources utilizing information extraction pipelines [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8">5, 6, 7, 8</xref>
        ].
      </p>
      <p>
        In the realm of NLP, a significant body of research has been dedicated to integrate specialized
knowledge with Language Models [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], such as BERT [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. For example, Liu et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] proposed
to extend BERT by injecting knowledge graphs triples directly into the input text, whereas
Ostendorf et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] combines knowledge graphs embeddings and metadata to complement
the information presented to BERT through text.
      </p>
      <p>
        Previous studies in peer-to-peer accommodation business (like Airbnb) were focused on the
pricing issues [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] or in detecting the booking likelihood [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. However, these methods do not
consider the textual description of each accommodation and focus on numeric features, missing
an important factor in the tourist’s choices. Consequently, we introduced a novel approach
that can also process textual descriptions leveraging the combination of knowledge graphs and
Language Models within the context of tourism, with the aim of addressing practical use cases,
such as accommodation classification.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Background</title>
      <p>
        Our solution was designed to support the optimization of an accommodation ofer by solving
two classification tasks derived from collaborative discussions with stakeholders and revenue
managers1. The first task is visit propensity classification which aims to predict if the
accommodation would be visited or not. This is also called the booking likelihood in other studies [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
Airbnb allows a user to write a review only after the visit. Therefore, if the accommodation
1Within the tourism industry, an individual responsible for maximizing the performance of an accommodation is
commonly known as a “Revenue Manager" or a “Revenue Optimization Manager."
received at least one review (visit) in the previous year, the label is set to high propensity;
otherwise to low propensity. The second task is review rating classification, which aims to predict if
an accommodation would have a high review rating (&gt; 4.5 over a 1-5 Likert scale). They are
encoded as binary classification tasks, serving as practical checklists for revenue managers 2.
      </p>
      <p>The resulting characterization of an accommodation equips users with the capability to
scrutinize the quality of the ofer and investigate an array of strategies for its enhancement.
Furthermore, it enables assessing how potential changes would impact the predicted dimensions.</p>
      <p>
        To support the methodology described in this paper, we adopted the following resources:
• Tourism Analytic Ontology (TAO) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], an ontology designed to describe the complex
dynamics of the tourism domain and support intelligent services in this space, that we
used to model accommodation’s data.
• London Tourism Knowledge Graph (TKG), a knowledge graph based on TAO and built
by following the methodology introduced in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and applied in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. It covers Airbnb’s
London accommodations and is based on open data from the Inside Airbnb project3. It
was used as a source for factual knowledge.
• DBpedia Spotlight [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], a cross-domain entity linking tool built upon DBpedia public
knowledge graph, we used it for extracting supplementary information from the
descriptions of Airbnb accommodations and extend the Knowledge Graph.
• BERT, a well-known language model renowned for its ability to capture extensive
contextual representations of both words and sentences [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. We leveraged BERT as the core
element for our text classification system.
      </p>
      <p>Since TKG is stored in a triple-store database, we leverage the SPARQL language to extract
all relevant data. Subsequently, a data pipeline generates three distinct datasets that are utilized
for feature engineering. Dataset 1 associates each accommodation with its description text and
various properties represented as numbers, dates, or Boolean flags. Dataset 2 associates each
accommodation with its included amenities4 expressed using TAO classes. Dataset 3 associates
each accommodation with related DBpedia entities, which are extracted from accommodation
descriptions using DBpedia Spotlight.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>
        We introduce a methodology that combines a Transformer model for text processing and a
Multi-Layer Perceptron (MLP), used for incorporating additional features. This solution allows
knowledge infusion from TKG and uses Datasets 1, 2 and 3 to produce the following features: (i)
textual features, serving as a natural input for BERT, are derived from accommodation
descriptions in Dataset 1 using a BERT tokenizer5, (ii) numerical features are produced from numerical
2The use of just two states high/low is a design choice that gives managers good confidence in pursuing simple and
concrete objectives thanks to high accuracy levels in the predictions.
3See http://insideairbnb.com/about.
4Amenities refer to additional services, features, or facilities provided to guests during their stay.
5The tokenizer is responsible for dividing input text into individual tokens and applying additional tokenization
techniques, such as splitting words into subwords or adding special tokens for tasks like sentence classification or
question answering.
data in Dataset 1, like the number of rooms and beds, and represented as a  dimensional vector6
whose elements are normalized to have all values within the range [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ], (iii) categorical features
i.e., the list of amenities for each accommodation, from Dataset 2, are converted into numeric
vectors (of size 145) through the process of one-hot encoding, and (iv) linked entities, which are
DBpedia entities extracted from the descriptions and stored in Dataset 3, are also transformed
into numeric vectors (of size 625) through a one-hot encoding process after those associated
with less than 100 accommodations are filtered out.
      </p>
      <p>The training phase employs an end-to-end optimization for each classification task. It involves
ifne-tuning the BERT transformer on the description set, and training the MLP from scratch on
the BERT output combined with all other additional features. More in detail, the tokenized text
is processed by BERT using the English uncased model7. The [CLS] hidden vector state of size
768 is used as the BERT output. The tanh output from the pooling layer tied to BERT is scaled
between 0 to 1 to match other non-textual feature vectors. Numeric features are normalized
real number vectors, and categorical and linked entities’ features are one-hot encoded. The
four vectors are concatenated and fed into the MLP, composed of two layers with 1024 units
each and a ReLu activation function. All MLP layers undergo dropout (with default probability
p=0.1) during training to counter overfitting. The output layer of the MLP is a Sigmoid layer,
providing the classification probability output.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation</title>
      <p>To evaluate the proposed classification methodology, we created a balanced dataset for each of
the two classification tasks, leveraging the full datasets denfied in Section 3. We split the balanced
data set of each task into three parts: train, validation, and test set. Subsequently, starting from
the complete training set, we produced four distinct training sets for each task, progressively
increasing in size to contain 3000, 6000, 9000, and 12000 accommodations. The objective of
this process is to evaluate the influence of varying data quantities on the performance of the
classification tasks. We used the same validation and test set for each task with a fixed size
of 1800 elements in order to obtain comparable results. We assessed the performance using
precision, recall, and F1 score.</p>
      <p>
        To ensure robustness, we performed the training and hyperparameter tuning process five
times, resulting in five model versions for each experiment. Every model variant was assessed
using the test set, and the average metric values were subsequently computed. This approach
was employed to account for the potential accuracy fluctuations observed in previous studies
when fine-tuning BERT-based models multiple times on the same dataset with varying random
seeds [
        <xref ref-type="bibr" rid="ref14 ref2">2, 14</xref>
        ].
      </p>
      <p>We evaluated three approaches:
1. Logistic Regression, a Logistic Regression Classifier used as a baseline 8.
6With  = 15 for review rating classification task and  = 12 for visit propensity classification task.
7See Hugging Face repository https://huggingface.co/bert-base-uncased
8It is a network composed of one hidden layer, with a single unit, ReLu activation, and a Sigmoid layer for binary
classification probabilities. Unable to process text, it is fed only numerical, categorical, and linked entity features.
max_train_size
experiment
LOGISTIC REGRESSION 72.94 76.30 78.05
BERT 64.44 67.29 66.97
BERT complemented with KG 85.16 85.67 85.63
max_train_size
experiment
LOGISTIC REGRESSION 56.64 61.99 64.75
BERT 61.97 63.90 64.16
BERT complemented with KG 69.76 69.43 69.38
2. BERT, a BERT-based uncased model, trained on textual features9.
3. BERT complemented with KG, the methodology introduced in Section 4.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>This paper introduces a novel methodology that integrates language models and knowledge
graphs to enhance two classification tasks about accommodation ofers in the tourism domain.
To improve classification accuracy, we combined BERT with a knowledge graph to provide
numeric data, categorical information, and linked entities. Our approach outperformed BERT
with a mean increase of 12.5% points in F1 score.</p>
      <p>Future research directions will focus on enhancing the methodology for multi-class
classification and regression and explore how efectively a classifier trained for a specific tourist location,
such as London, could be transferred and applied to a diferent destination.
9The pooled output from BERT [CLS] token is fed to a final inner layer with one unit followed by a Sigmoid layer.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>González-Serrano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Talón-Ballestero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Revenue</given-names>
            <surname>Management</surname>
          </string-name>
          and E-Tourism:
          <article-title>The Past</article-title>
          ,
          <source>Present and Future</source>
          , Springer International Publishing, Cham,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
          . URL: https: //doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -05324-6_
          <fpage>76</fpage>
          -
          <lpage>1</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -05324-6_
          <fpage>76</fpage>
          -
          <lpage>1</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. T.</given-names>
            <surname>Google</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. I. Language</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          ,
          <source>Naacl-Hlt</source>
          <year>2019</year>
          (
          <year>2018</year>
          )
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Naseriparsa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <article-title>Knowledge graphs: opportunities and challenges</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <article-title>A survey of knowledge enhanced pre-trained models</article-title>
          ,
          <source>CoRR abs/2110</source>
          .00269 (
          <year>2021</year>
          ). URL: https://arxiv.org/abs/ 2110.00269. arXiv:
          <volume>2110</volume>
          .
          <fpage>00269</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chessa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Fenu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Motta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Reforgiato</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Secchi</surname>
          </string-name>
          ,
          <article-title>Data-driven methodology for knowledge graph generation within the tourism domain</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>67567</fpage>
          -
          <lpage>67599</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2023</year>
          .
          <volume>3292153</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chessa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Fenu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Motta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          , L. Secchi,
          <article-title>Enriching Data Lakes with Knowledge Graphs</article-title>
          ,
          <source>CEUR Workshop Proceedings</source>
          <volume>3184</volume>
          (
          <year>2022</year>
          )
          <fpage>123</fpage>
          -
          <lpage>131</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Troncy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jameson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Corcho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Palumbo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Ballesteros Hermida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Spirescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Kuhn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Barbu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Celino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Valla</surname>
          </string-name>
          , T. Haaker, 3cixty:
          <article-title>Building comprehensive knowledge bases for city exploration</article-title>
          ,
          <source>Journal of Web Semantics 46-47</source>
          (
          <year>2017</year>
          )
          <fpage>2</fpage>
          -
          <lpage>13</lpage>
          . URL: http://dx.doi.org/10.1016/j.websem.
          <year>2017</year>
          .
          <volume>07</volume>
          .002. doi:
          <volume>10</volume>
          .1016/j.websem.
          <year>2017</year>
          .
          <volume>07</volume>
          .002.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Gazzè</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Duca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marchetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tesconi</surname>
          </string-name>
          ,
          <article-title>An overview of the tourpedia linked dataset with a focus on relations discovery among places</article-title>
          ,
          <source>ACM International Conference Proceeding Series 16-17-Sept</source>
          (
          <year>2015</year>
          )
          <fpage>157</fpage>
          -
          <lpage>160</lpage>
          . doi:
          <volume>10</volume>
          .1145/2814864.2814876.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K-</surname>
          </string-name>
          <article-title>BERT: enabling language representation with knowledge graph</article-title>
          , CoRR abs/
          <year>1909</year>
          .07606 (
          <year>2019</year>
          ). URL: http://arxiv.org/ abs/
          <year>1909</year>
          .07606. arXiv:
          <year>1909</year>
          .07606.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bourgonje</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Rehm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <article-title>Enriching BERT with knowledge graph embeddings for document classification</article-title>
          , CoRR abs/
          <year>1909</year>
          .08402 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1909</year>
          .08402. arXiv:
          <year>1909</year>
          .08402.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Estimating spatial efects on peer-to-peer accommodation prices: Towards an innovative hedonic model approach</article-title>
          ,
          <source>International Journal of Hospitality Management</source>
          <volume>81</volume>
          (
          <year>2019</year>
          )
          <fpage>43</fpage>
          -
          <lpage>53</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/ S0278431918306984. doi:https://doi.org/10.1016/j.ijhm.
          <year>2019</year>
          .
          <volume>03</volume>
          .012.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Afrianto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wasesa</surname>
          </string-name>
          ,
          <article-title>Booking prediction models for peer-to-peer accommodation listings using logistics regression, decision tree, k-nearest neighbor, and random forest classifiers</article-title>
          ,
          <source>JISEBI</source>
          <volume>6</volume>
          (
          <year>2020</year>
          )
          <fpage>123</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>P. N.</given-names>
            <surname>Mendes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jakob</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>García-Silva</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Bizer, DBpedia spotlight: Shedding Light on the Web of Documents</article-title>
          ,
          <source>in: Proc. of the 7th Int. Conf. on Semantic Systems - I-Semantics '11</source>
          , ACM Press, New York, New York, USA,
          <year>2011</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . doi:
          <volume>10</volume>
          .1145/2063518.2063519.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dodge</surname>
          </string-name>
          , G. Ilharco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Farhadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hajishirzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <surname>Fine-Tuning Pretrained Language Models: Weight Initializations</surname>
            ,
            <given-names>Data</given-names>
          </string-name>
          <string-name>
            <surname>Orders</surname>
          </string-name>
          , and Early Stopping, CoRR abs/
          <year>2002</year>
          .0 (
          <year>2020</year>
          ). arXiv:arXiv:
          <year>2002</year>
          .06305v1.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>