<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Neural factorization for Ofer Recommendation using Knowledge Graph Embeddings</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gourab Chowdhury</string-name>
          <email>gourab@iitkgp.ac.in</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mainak Chain</string-name>
          <email>mainakchain@iitkgp.ac.in</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Neural Factorization, Recommender Systems, Knowledge Graph</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Madiraju Srilakshmi</string-name>
          <email>sreelakshmi@iitkgp.ac.in</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sudeshna Sarkar</string-name>
          <email>sudeshna@cse.iitkgp.ernet.in</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Embedding, E-commerce ofers</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Indian Institute of Technology Kharagpur</institution>
          ,
          <addr-line>West Bengal</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <abstract>
        <p>Companies send a lot of promotional ofers and coupons to customers to attract them to buy more. Ofer recommendation systems can help to identify relevant ofers to users. In this paper, we present a Neural Factorization (NF) model for the task of Ofer recommendation. We represent users and ofers with Knowledge Graph Embeddings (KGE). Specifically, we model the available data in the form of a Knowledge Graph (KG) and learn embeddings for entities and relations using a standard KGE technique called TransE. We also incorporate the user temporal features in the NF model using Long Short Term Memory (LSTM) with attention framework. We experiment with Kaggle Acquire Valued Shoppers Challenge dataset and show that the performance of our model is significantly better than tree-based methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
      <p>• Information systems → Recommender systems; •
Information Systems → Information retrieval; Recommender Systems;
• Computing methodologies → Neural networks.</p>
    </sec>
    <sec id="sec-2">
      <title>INTRODUCTION</title>
      <p>Marketing and promotions are used to attract customers in the
retail domain. Companies spend a lot of money to send promotional
offers or discounts to customers. It is therefore important to identify
relevant offers that the users are likely to accept.</p>
      <p>We consider the case in which an ofer is a flat discount for a
given brand and category combination. An example of an ofer is
"Rs 200 of on Polo T-shirts". There can be other types of ofers
such as coupons/promo codes (Use Coupon GO15O to get Rs 150
cash-back), combo ofers (Buy 3 get 1 free, Buy 3 get Rs 300 of)
and loyalty points.</p>
      <p>
        Xia et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] proposed an approach for the task of ofer
recommendation based on the features of users and ofers. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] used
tree-based methods namely Random forest [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Gradient boosted
trees [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to handle this task. The features involve attributes of the
users such as their email domain and OS systems and attributes of
ofers such as text description, discount amount as well as shopping
history features such as recent discount information and number
of times the user visited the coupon before.
      </p>
      <p>We propose to use a Neural Factorization (NF) model to learn
user-ofer interactions. The user and ofer representations are given
as input to the NF model whose output is the probability of the
ofer being accepted by the user. We predict the probabilities of all
ofers available at the given time and recommend the top k probable
ofers to the customer. In this work, we explore diferent ways of
representing users and ofers.</p>
      <p>In our first model, we represent users and ofers with features
extracted from the dataset. The user features contain the normalized
count of items purchased in a month, the normalized count of items
purchased in each category, days since the last visit, the average
amount paid per visit etc. The ofer features include category, brand
on which ofer is given, amount of discount, minimum quantity of
purchase etc.</p>
      <p>
        In our second model, we explore representing user and ofers as
embedding. For this, we construct a Knowledge Graph (KG)
involving users, categories, brands, price values as nodes and belongs-to,
purchase and price as edges between them. We adopt a
knowledge graph embedding technique called TransE [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to generate
embeddings of users and ofers.
      </p>
      <p>In our third model, we capture the user sequential behaviour
using Long Short Term Memory (LSTM) with attention framework.
The input to the model is the sequence of baskets purchased by the
user and the output is the probabilities on all categories available.
We incorporate this information as an additional input to the NF
model.</p>
      <p>We experiment with Kaggle Acquire Valued Shoppers Challenge
dataset which contains user ofer interactions, user purchase history
and ofer content information. We apply our models on this data
and show that the NF based models achieves better performance
than tree-based methods.
2</p>
    </sec>
    <sec id="sec-3">
      <title>PRELIMINARIES</title>
      <p>In this section, we formally define the task of ofer recommendation
and present the details of the data available to handle this task.
2.1</p>
    </sec>
    <sec id="sec-4">
      <title>Problem Definition</title>
      <p>Ofer Recommendation is the task of predicting the best ofers
for a given user. Let U = {u1, u2, ..., um } be the set of users and
O = {o1, o2, o3...on } be the set of ofers. The task is to recommend
top k ofers to the users so that they are likely to include the next
ofer converted by the user.
2.2</p>
    </sec>
    <sec id="sec-5">
      <title>Dataset</title>
      <p>In our work, we experiment with Kaggle Acquire Valued Shoppers
Challenge dataset 1.</p>
      <p>The dataset contains the transaction history of users from March
2012 to July 2013. A transaction consists of user_id, item_id and
date. The set of items purchased by the same user in the same date
are termed as a basket.</p>
      <p>The user-ofer interactions are recorded from March 2013 to July
2013. A user-ofer interaction consists of user_id, ofer_id and date.
Each user has availed exactly one ofer in this period. Each ofer
is specified by its category, brand, discount amount and minimum
quantity. The overall data statistics are listed in Table 1.
3</p>
    </sec>
    <sec id="sec-6">
      <title>NEURAL FACTORIZATION FOR OFFER</title>
    </sec>
    <sec id="sec-7">
      <title>RECOMMENDATION</title>
      <p>We use Neural Factorization (NF) model for the task of ofer
recommendation. We have experimented with diferent methods of
representing users and ofers. We first explain the basic neural
factorization framework and then introduce our methods.
3.1</p>
    </sec>
    <sec id="sec-8">
      <title>Architecture of Neural Factorization</title>
      <p>
        Our framework is based on Neural Collaborative Filtering proposed
by He et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The system [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] used Multi Layer Perceptron (MLP)
for modelling user-item interactions which is able to capture the
non-linear relations between users and items.
      </p>
      <p>In our model, the user vector (vu ) and ofer vector (vo ) are given
as input to two input layers. Each input layer is followed by a
dense layer. The output of these dense layers are concatenated and
are given as input to a Neural network. We use past user-ofer
1https://www.kaggle.com/c/acquire-valued-shoppers-challenge/data
(u,o)∈Y−∪Y+</p>
      <p>L =
interactions as positive samples while the negative samples are
generated by random sampling. The final output layer predicts yuo ,
the probability of user u accepting an ofer o. The architecture is
illustrated in Figure 2.</p>
      <p>The model is trained by minimizing the loss between predicted
value (yuˆo ) and the target value yuo . The target value represents
the user action towards an ofer. It is 1 when the user availed the
ofer and 0 otherwise. The loss function is defined as follows:
Õ
−(yuo log (yuˆo ) + (1 − yuo ) log (1 − yuˆo ))
(1)
where Y+ denotes the set of positive interaction of users, ofers
and Y− denotes the negative instances (sampled from unobserved
data).</p>
      <p>The user_ids and ofer_ids are diferent in the train and test
datasets. Therefore, we can't use the learned factors for the
prediction. We represent the user and ofer by their content information
and input them to the trained neural model and predict the
probability of the user accepting the ofer. Similarly, we find the probabilities
of all ofers available at the given time. The ofers are ranked based
on the probability values and top k ofers are recommended to the
user.
– Quantity to be purchased to avail discount
– How cheap the product is compared to other products in
the same category (Amount).</p>
      <p>The numeric value is used as input for the numeric features while
the non-numeric features such as brand and category are one-hot
encoded. Since these features have multiple possible values, the
feature dimensions become large. These features are also unable to
capture indirect relationships between users and ofers.
3.3</p>
    </sec>
    <sec id="sec-9">
      <title>Neural Factorization with Knowledge</title>
    </sec>
    <sec id="sec-10">
      <title>Graph Embeddings (NF+KGE)</title>
      <p>The representation of users and ofers plays a significant role in
the efectiveness of a recommendation model. We wish to use a
representation that is able to capture relevant knowledge of users,
items, ofers, the attributes of the above entities and the interactions
between them.</p>
      <p>
        We propose to use Knowledge Graph Embedding (KGE)
techniques to learn embeddings for users and ofers. These techniques
have been found to be efective in capturing complex and indirect
relationships among entities in the Knowledge Graph (KG) and are
proven to be successful in many applications such as link prediction
and recommendation systems etc [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>We construct a Knowledge Graph (KG) based on user purchase
history and ofer content information. The nodes of the graph
are users, categories, brands and price range. We find the average
price of items in a brand and categorize them into high and low.
We have 3 types of relations in our graph. The user nodes are
connected to category nodes by relation purchased, category nodes
are connected to brand nodes by relation belongs_to and brand
nodes are connected to price range nodes by relation price. The
graph is formed as a set of triplets (h, r , t ) i.e., head node (h) is
connected to tail node (t ) by relationship (r ). An example graph
representation is shown in Figure 3.</p>
      <p>The triplets generated for the example graph are as follows: T =
{(U1, purchased, C1), (U2, purchased, C1), (C1, belonдs_to, B1),
(C2, belonдs_to, B1), (C3, belonдs_to, B2), (B1, price, low),
(B2, price, hiдh)}.</p>
      <p>
        We use a standard knowledge graph embedding method called
TransE [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] to learn the embeddings for all entities and relationships
in the graph. We have chosen TransE because this method is simple
and has been found to be eficient in modelling multi-relational
data [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>
        Given a triplet of the form &lt;h,r,t&gt; which indicates that the head
entity (h) is connected to the tail entity (t) by the relationship (r),
TransE [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] learns the embedding such that h + r ≈ t (Figure 3).
TransE uses the following scoring function:
fr (h, t ) = −||h + r − t ||1/2
(2)
TransE obtains positive triplets from the graph and negative triplets
The above Knowledge Graph Embeddings (KGE) is unable to
capture the sequential behaviour of users since the knowledge graph
does not represent the time stamp of the interactions or the
sequentiality of the transactions. Therefore, we try to enhance our model
by incorporating a temporal component by considering the user
sequential purchase behaviour using a Long Short Term Memory
(LSTM) with attention model.
      </p>
      <p>Since the ofers are given on specific categories, we formulate
the task of predicting the next category to be purchased by the user.
We hypothesize that this information may help to identify suitable
ofers for a user.</p>
      <p>Each user has purchased a set of items per visit. The set of items
can be termed as a basket. We consider the category of an item
(category_id) instead of the exact item (item_id) in each basket. We
give the sequence of baskets purchased by the user as input to the
LSTM and predict the probabilities of he categories purchased in the
next basket. The predicted category probabilities are incorporated
as an additional input to the NF model.</p>
      <p>
        Let u = {b1, b2, . . . , bt } be the basket sequence of the target user.
Each basket is the group of categories: bk = {c1, c2, . . . , cn }, where
n is the size of the basket. We represent each category with the
embedding learned from the Knowledge Graph discussed earlier.
We use average pooling to represent the basket. This approach is
similar to the basket prediction method proposed by Yu et al.[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>The user sequence is now denoted as u = {v1, v2, . . . , vt }, where
vk is the average of graph-based category embeddings in the basket
bk . We input the user sequence till the current time t into the LSTM
model. Let ht be the LSTM hidden unit and yt be the output at t -th
time step. The hidden state ht of each interaction is updated by the
previous hidden state ht −1 and the current basket embedding vt .</p>
      <p>ht = LSTM(ht −1, vt )
We apply attention on top of the LSTM layer to give weights to
the baskets at diferent time-steps. Let H = (h1, h2, . . . , ht ) be the
output vectors that are produced by LSTM layer. They are inputs to
the attention layer and the weights for each time-step are learned
i.e. A = (a1, a2, . . . , at ). The weighted sum of the hidden states
(M) is input into a dense layer (D) to find the scores of all categories.
We find their probabilities using the sigmoid activation function.
This architecture is illustrated in Figure 4.</p>
      <p>A = softmax(wT ∗ tanh(H ))</p>
      <p>M = AT ∗ H
P (yt +1 | v1i≤i ≤t ) = sigmoid(D[M])
(4)
(5)
(6)
(7)</p>
      <p>The learned category probabilities are given as an auxiliary input
to the NF model. The rest of the architecture is similar and is shown
in Figure 5.
The train and test split are considered as given in the dataset i.e., the
ifrst 2 months of user-ofer interaction data are used as train and
the rest of the 3 months as the test set. The train and test statistics
are listed in Table 2. To train the Knowledge Graph Embeddings, we
have used recent 5 months of user purchase history before ofers
are given to the users (01-October-2012 to 28-February-2013). This
has been done on the assumption that the recent purchase history
reflects the current preferences of the user. The same data is used
for predicting user category preferences as well as feature creation.
4.2.2 Experimental Setup. We use the following parameters to
learn knowledge graph embedding. The size of the embedding for
all entities (user, brand, category, price range) is 100. We use Adam
optimizer, and the learning rate is set to 0.001.</p>
      <p>The parameters for predicting user next preferred categories
using LSTM are as follows. We use one LSTM layer with 100 hidden
4.93
6.73
15.41
21.77
15.96
units. There is a dropout layer in between the LSTM layer and the
attention layer with 25% dropout. The learning rate is set to 0.001.</p>
      <p>The parameters for neural factorization models are as follows.
The activation function for dense layers is Leaky Relu and the batch
size is set at 512. The learning rate is set to 0.0001. We applied batch
normalization at every layer.
4.3</p>
    </sec>
    <sec id="sec-11">
      <title>Comparison of Models</title>
      <p>We compare the NF based models namely NF+features, NF+KGE
and NF+KGE+TF against XGBoost and Random Forest Classifier
(RFC) methods with features.</p>
      <p>The standard item KNN method can't be applied to this dataset
because each user in the dataset has availed exactly one ofer. There
is no possibility of finding similar ofers to the ofers that are
previously availed by the user and recommend them.
4.4</p>
    </sec>
    <sec id="sec-12">
      <title>Result and Analysis</title>
      <p>The results for the two baseline methods and the three NF based
models discussed in this paper are presented in Table 1. It is evident
from the results that the NF based models outperform the XGboost
and RFC methods with features.</p>
      <p>Neural factorization model with graph-based embedding (NF+KGE)
performs better than Neural factorization with features (NF+features).</p>
      <p>The final variation of our model with temporal features (NF+KGE+TF)
gives the significant improvement over all other models considered
in terms of Recall@3 and Recall@5 and MRR@5. Neural
factorization model with graph-based embedding (NF+KGE) performs best
in terms of Recall@1.</p>
      <p>We hypothesize that the knowledge graph based embedding
is efective as it is able to use the connections between diferent
entities and can therefore efectively capture indirect as well as
latent relationships.</p>
      <p>However the limitation of the above model is that it fails to
capture the temporal interactions of the user. Our third model
(NF+KGE+TF) addresses this by using an additional input from the
LSTM based on the user’s sequence of baskets.
5</p>
    </sec>
    <sec id="sec-13">
      <title>RELATED WORK</title>
      <p>Related work to our model is categorized into three parts.
Subsection 5.1 reviews the work reported on ofer recommendation
systems. Subsection 5.2 reviews the methods used in a related task
of repeat purchase prediction methods. In subsection 5.2, we discuss
about Knowledge Graph Embedding (KGE) based recommendation
systems.
5.1</p>
    </sec>
    <sec id="sec-14">
      <title>Ofer Recommender Systems</title>
      <p>There are very few works on ofers and coupons recommendation
systems.</p>
      <p>
        Xia et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] approached the problem of ofer recommendation
in e-commerce domain. They used a private dataset consists of
customer’s shopping trips, shopping trip counts, clicked coupons,
and retailers that issued the coupon. The coupons are characterized
based on their textual descriptions and validity period. The authors
curated a number of features from the data to represent users and
coupons and ranked the coupons based on scores generated by
XGBoost [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and Random Forest [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] algorithms.
      </p>
      <p>
        Similar work is proposed by Hu et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] in the telecom domain.
Hu et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] used Random Forest method [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] to provide telecom
ofers to mobile users. The authors extracted user features such as
age, gender, voice call duration, SMS count etc. from the customer
profile repository, and historical usage repository. These features
are given as input to the Random Forest algorithm [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
5.2
      </p>
    </sec>
    <sec id="sec-15">
      <title>Repeat Purchase Prediction Systems</title>
      <p>Repeat purchase prediction is the task given for Kaggle Acquire
Valued Shoppers Challenge. Since we have used the same dataset
for our task of ofer recommendation systems, we present a review
of the work done on repeat purchase prediction.</p>
      <p>
        Anand et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] proposed a prediction model based on a
combination of temporal and aggregate level models. They extracted
three types of features capturing diferent aspects of user behaviour
namely customer-based, product-based, customer-product
interactions based. customer-based features include total visits made, total
spend, the loyalty of the customer etc. product-based features
include the fraction of repeat customers for the ofer-product etc.
customer-product interactions based features include the number of
visits, quantity bought, the amount spent etc.
      </p>
      <p>To capture the aggregate level behaviour of the user, the above
features are computed over the entire transaction history. To
capture the temporal behaviour, the features are split and computed
over non-overlapping time windows. The authors used Long Short
Term Memory (LSTM) model as the classifier for temporal features
and quantile regression (QR) model as the classifier for aggregate
level features. The two models are combined using a mixture of
experts (ME).</p>
      <p>
        Nikulin et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] used Random forest [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Gradient boosted
trees [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to predict the repeat purchase behaviour of the users. The
authors curated a number of statistical features from the data and
applied the above methods.
5.3
      </p>
    </sec>
    <sec id="sec-16">
      <title>Knowledge Graph Embedding based</title>
    </sec>
    <sec id="sec-17">
      <title>Recommender Systems</title>
      <p>
        In recent times, knowledge graph embeddings have been shown
to be efective for recommendation systems. The basic idea is to
represent the available data in the form of a graph, learn embeddings
for entities using Knowledge graph embedding methods [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and
incorporate them into recommendation.
      </p>
      <p>
        [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] presents Collaborative Knowledge base Embedding (CKE)
which uses TransR [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] to learn the structural representations of
items which are combined with visual and textual embeddings.
Deep Knowledge-aware Network (DKN) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] learns entity
embeddings using TransD [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and and designs a CNN framework by
combining them with word embeddings for news recommendation.
Ai et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] learn embedding of users and items by the method of
TransE [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and the recommendation is based on user-item similarity
score in the projected space.
6
      </p>
    </sec>
    <sec id="sec-18">
      <title>CONCLUSION</title>
      <p>In this paper, we have presented a neural factorization method for
the task of ofer recommendation with diferent representations of
users and ofers. We have shown that our models perform better
than the tree-based methods. We have also shown that the learned
graph-based user and ofer embeddings capture deeper and indirect
connections between users and ofers, which helps to improve the
quality of recommendation. The incorporation of temporal features
involving transaction sequences improves the performance further
in some cases.</p>
    </sec>
    <sec id="sec-19">
      <title>ACKNOWLEDGMENTS</title>
      <p>This research was supported by Capillary Technologies
International Pte Limited. We thank Dr. Subrat Panda and Mr. Jyotiska
Bhattacharjee from Capillary Technologies International Pte
Limited who provided insight and expertise that greatly assisted the
research.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Qingyao</given-names>
            <surname>Ai</surname>
          </string-name>
          , Vahid Azizi, Xu Chen,
          <string-name>
            <given-names>and Yongfeng</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <article-title>Learning heterogeneous knowledge base embeddings for explainable recommendation</article-title>
          .
          <source>Algorithms</source>
          ,
          <volume>11</volume>
          (
          <issue>9</issue>
          ):
          <fpage>137</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Gaurangi</given-names>
            <surname>Anand</surname>
          </string-name>
          , Auon H Kazmi, Pankaj Malhotra, Lovekesh Vig, Puneet Agarwal, and
          <string-name>
            <given-names>Gautam</given-names>
            <surname>Shrof</surname>
          </string-name>
          .
          <article-title>Deep temporal features to predict repeat buyers</article-title>
          .
          <source>In NIPS 2015 Workshop: Machine Learning for eCommerce</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Antoine</given-names>
            <surname>Bordes</surname>
          </string-name>
          , Nicolas Usunier, Alberto Garcia-Duran,
          <string-name>
            <given-names>Jason</given-names>
            <surname>Weston</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Oksana</given-names>
            <surname>Yakhnenko</surname>
          </string-name>
          .
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          , pages
          <fpage>2787</fpage>
          -
          <lpage>2795</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Leo</given-names>
            <surname>Breiman</surname>
          </string-name>
          .
          <article-title>Random forests</article-title>
          .
          <source>Machine learning</source>
          ,
          <volume>45</volume>
          (
          <issue>1</issue>
          ):
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Jerome</surname>
            <given-names>H</given-names>
          </string-name>
          <string-name>
            <surname>Friedman</surname>
          </string-name>
          .
          <article-title>Greedy function approximation: a gradient boosting machine</article-title>
          .
          <source>Annals of statistics</source>
          , pages
          <fpage>1189</fpage>
          -
          <lpage>1232</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Xiangnan</given-names>
            <surname>He</surname>
          </string-name>
          , Lizi Liao, Hanwang Zhang, Liqiang Nie, Xia Hu, and
          <string-name>
            <surname>Tat-Seng Chua</surname>
          </string-name>
          .
          <article-title>Neural collaborative filtering</article-title>
          .
          <source>In Proceedings of the 26th International Conference on World Wide Web</source>
          , pages
          <fpage>173</fpage>
          -
          <lpage>182</lpage>
          . International World Wide Web Conferences Steering Committee,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Wan-Hsun</surname>
            <given-names>Hu</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shih-Hsien</surname>
            <given-names>Tang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yin-Che</surname>
            <given-names>Chen</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chia-Hsuan Yu</surname>
          </string-name>
          , and Wen-Cheng Hsu.
          <article-title>Promotion recommendation method and system based on random forest</article-title>
          .
          <source>In Proceedings of the 5th Multidisciplinary International Social Networks Conference, page 11. ACM</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Guoliang</given-names>
            <surname>Ji</surname>
          </string-name>
          , Shizhu He, Liheng Xu, Kang Liu, and
          <string-name>
            <given-names>Jun</given-names>
            <surname>Zhao</surname>
          </string-name>
          .
          <article-title>Knowledge graph embedding via dynamic mapping matrix</article-title>
          .
          <source>In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)</source>
          , pages
          <fpage>687</fpage>
          -
          <lpage>696</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Yankai</given-names>
            <surname>Lin</surname>
          </string-name>
          , Zhiyuan Liu, Maosong Sun, Yang Liu, and
          <string-name>
            <given-names>Xuan</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <article-title>Learning entity and relation embeddings for knowledge graph completion</article-title>
          .
          <source>In Twenty-ninth AAAI conference on artificial intelligence</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Vladimir</given-names>
            <surname>Nikulin</surname>
          </string-name>
          .
          <article-title>Prediction of the shoppers loyalty with aggregated data streams</article-title>
          .
          <source>Journal of Artificial Intelligence and Soft Computing Research</source>
          ,
          <volume>6</volume>
          (
          <issue>2</issue>
          ):
          <fpage>69</fpage>
          -
          <lpage>79</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Hongwei</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Fuzheng Zhang, Xing Xie, and
          <string-name>
            <given-names>Minyi</given-names>
            <surname>Guo</surname>
          </string-name>
          . Dkn:
          <article-title>Deep knowledgeaware network for news recommendation</article-title>
          .
          <source>arXiv preprint arXiv:1801.08284</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Quan</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Zhendong Mao,
          <string-name>
            <given-names>Bin</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Li</given-names>
            <surname>Guo</surname>
          </string-name>
          .
          <article-title>Knowledge graph embedding: A survey of approaches and applications</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          ,
          <volume>29</volume>
          (
          <issue>12</issue>
          ):
          <fpage>2724</fpage>
          -
          <lpage>2743</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Yandi</surname>
            <given-names>Xia</given-names>
          </string-name>
          , Giuseppe Di Fabbrizio, Shikhar Vaibhav, and
          <string-name>
            <given-names>Ankur</given-names>
            <surname>Datta</surname>
          </string-name>
          .
          <article-title>A contentbased recommender system for e-commerce ofers and coupons</article-title>
          .
          <source>In eCOM@SIGIR</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Feng</surname>
            <given-names>Yu</given-names>
          </string-name>
          , Qiang Liu, Shu Wu,
          <string-name>
            <given-names>Liang</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Tieniu</given-names>
            <surname>Tan</surname>
          </string-name>
          .
          <article-title>A dynamic recurrent model for next basket recommendation</article-title>
          .
          <source>In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval</source>
          , pages
          <fpage>729</fpage>
          -
          <lpage>732</lpage>
          . ACM,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Fuzheng</surname>
            <given-names>Zhang</given-names>
          </string-name>
          , Nicholas Jing Yuan, Defu Lian, Xing Xie, and
          <string-name>
            <surname>Wei-Ying Ma</surname>
          </string-name>
          .
          <article-title>Collaborative knowledge base embedding for recommender systems</article-title>
          .
          <source>In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , pages
          <fpage>353</fpage>
          -
          <lpage>362</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>