=Paper=
{{Paper
|id=Vol-2319/paper25
|storemode=property
|title=Towards Optimization of e-Commerce Search and Discovery
|pdfUrl=https://ceur-ws.org/Vol-2319/paper25.pdf
|volume=Vol-2319
|authors=Anjan Goswami,Chengxiang Zhai,Prasant Mohapatra
|dblpUrl=https://dblp.org/rec/conf/sigir/GoswamiZM18
}}
==Towards Optimization of e-Commerce Search and Discovery==
Towards Optimization of E-Commerce Search and
Discovery
Anjan Goswami ChengXiang Zhai Prasant Mohapatra
University of California, Davis University of Illinois University of California, Davis
agoswami@ucdavis.edu Urbana-Champaign, IL pmohapatra@ucdavis.edu
czhai@illinois.edu
ABSTRACT Keywords
E-Commerce (E-Com) search is an emerging problem with e-com search, retrieval models, learning to rank, discover-
multiple new challenges. One of the primary challenges ability, exploration-exploitation
constitutes optimizing multiple objectives involving business
metrics such as sales and revenue and maintaining a discov- 1. INTRODUCTION
ery strategy for the site. In this paper, we formalize the
One of the most critical components of an e-commerce (e-
e-com search problem for optimizing metrics based on sales,
com) marketplace is its search functionality. The goal of an
revenue, and relevance. We define a notion of item discover-
e-commerce search engine is to help the buyers to find and
ability in search and show that learning to rank (LTR) algo-
purchase their desirable products. The buyers’ intent repre-
rithms trained with behavioral features from e-com customer
sents the demand side and the products listed by the sellers
interactions (eg. clicks,cart-adds, orders etc.) do not by
represent the supply side of an e-com platform. The e-com
themselves address the discoverability problem. Instead, a
search engine’s function can be considered to provide an ef-
suitable explore-exploit framework must be integrated with
ficient matching platform for the supply and the demand
the ranking algorithm. We thus construct a practical dis-
side and to facilitate transactions to generate revenue. Like
covery strategy by keeping a few top positions for discovery
a web search engine, an e-com search also intends to ensure
and populating some of the items selected through explo-
user satisfaction by presenting the most relevant items that
ration. Then, we present a few exploration strategies with
a user is searching for. However, an e-com search engine
low regret bounds in terms of business metrics. We conduct
generates revenue directly based on the products presented
a simulation study with a synthetically generated dataset
to the users; this is in sharp contrast with how a web search
that represents items with different utility distribution and
engine generates revenue through advertising and raises sig-
compares these strategies using metrics based on sales, rev-
nificant new challenges for designing effective search algo-
enue, relevance, and discovery. We find that a strategy based
rithms. Thus besides the usual goal of maximizing relevance
on adaptive submodular function based discovery framework
of the search results, an e-com search engine also needs to
can provide a nice balance of business metrics and discov-
optimize two additional goals that generally do not exist
erability compared to other strategies based on random ex-
in web search scenarios, i.e., it has to 1) maximize the busi-
ploration or multi-armed bandit. However, another strategy,
ness metrics such as revenue or sales from the purchases and
based on monotonic submodular optimization function that
2) minimize the inventory cost by selling the items faster.
needs to be integrated with linear LTR models also works
Both the objectives are dependent on the performance of
well for discovery and has nice performances with respect to
the underlying ranking algorithm of the search engine other
sales, revenue, and relevance.
than the assortment selection and demand generation. The
first set of goals of maximizing the business metrics can be
Categories and Subject Descriptors achieved by constructing a rank function using learning to
rank framework with a suitable optimization objective that
H.3.3 [Information Search and Retrieval]: E-Com Search is in line with the business goals of the e-commerce com-
pany. The second goal however requires the e-commerce
sites to have a strategy of discovery so that it can expose
General Terms as many items to the customers faster and determine what
Algorithms products are selling. This allows the companies to readjust
their assortment strategy to optimize the inventory and as-
sociated costs. The discovery strategies however, can hurt
the business goals since it requires showing previously unex-
plored items to the customers. In e-commerce, learning to
Permission to make digital or hard copies of all or part of this work for rank functions are trained by constructing mainly two types
personal or classroom use is granted without fee provided that copies are of features: (a) features that arise from static contents of
not made or distributed for profit or commercial advantage and that copies the product description (b) features that come from the be-
bear this©notice
Copyright 2018 by and the full
the paper’s citation
authors. onpermitted
Copying the firstforpage.
privateTo
andcopy otherwise,
academic purposes. to havioral signals such as clicks, cart-adds, sales, and review
In: J. Degenhardt,
republish, G. Di
to post onFabbrizio,
servers or S. Kallumadi, M. Kumar,
to redistribute Y.-C. requires
to lists, Lin, A. Trotman, H. Zhao
prior specific
(eds.): Proceedings of the SIGIR 2018 eCom workshop, 12 July, 2018, Ann Arbor, Michigan, USA,
permission and/or a fee. ratings. It has been observed [16, 21] that the second group
published at http://ceur-ws.org
Copyright 20XX ACM X-XXXXX-XX-X/XX/XX ...$15.00. of features is more useful for finding a comparative relevance
among the items. However, the use of the machine learning literatures. The algorithms that are commonly used for ex-
algorithms in search engines tends to bias a search engine ploration include -greedy [23], and variants of upper con-
toward favoring the viewed items by users due to the use of fidence bound (UCB) algorithm [2]. Additionally, k-armed
features that are computed based on user behavior such as dueling bandit framework [25] has been developed for find-
clicks, rates of “add to cart”, etc. Since a learning algorithm ing out optimal bandit in a scenario where each bandit can
would rank items that have already attracted many clicks be considered to optimize a specific objective. It is possible
on the top, it might overfit to promote the items viewed by to frame a multiobjective problem with exploration compo-
users. As a result, some items might never have an oppor- nent using such paradigm. In this paper, we aim to construct
tunity to be shown to a user (i.e., “discovered” by a user), a framework that is simpler than K-armed bandit but more
thus also losing the opportunity to potentially gain clicks. practical.
Such “undiscovered” products would then have to stay in the
inventory for a long time incurring extra cost and hurting
satisfaction of the product providers. Hence, it is important 3. OPTIMIZATION OF E-COM SEARCH
to consider the integrating the aspect of discovery with the We consider the problem of optimization of an E-Com
ranking mechanism for e-commerce search. In this paper, search engine over a period of time {1, · · · , T }. Let us as-
we provide a practical yet theoretically sound framework for sume that there are N queries, denoted by Q = (q1 , q2 , · · · , qN ),
learning to rank where it is possible to have a regulated dis- that can be sent to the search engine during this time. Let
covery mechanism minimizing the loss in revenue, sales, or Z = {ζ1 , · · · , ζM } be the set of M items and let the corre-
relevance metrics. sponding prices of the items be given by {ρ1 , · · · , ρM }. Now,
The key contributions of this paper can be described as consider a rank function π and π(qi , t) outputs top K items
follows: which is a permutation from the set Z. Now, let us consider
(1) This is the first paper that formalizes an e-com search two binary random variables ΛS C
ijt , Λijt where the first one de-
problem as a multi-objective LTR problem with continuous notes if a sale happens or not and the second one denotes if
exploration that we call learning to rank and discover prob- there is a click or not for an item ζj given a query qt at time t.
lem. (2) We discuss how we can construct a solution to this Now, given that we have these two binary random variables,
problem using the existing framework of multi-armed ban- we can define the probability of sales given a query qi , item
dit and auction mechanism. (3) We also provide a novel ζj at time t as p(ΛS ijt = 1|qi , ζj , t) and similarly probability
paradigm combining algorithms from multi-objective opti- of click for the same can be defined as p(ΛC ijt = 1|qi , ζj , t).
mization and explore and exploit paradigms. Let’s also create a normalization constant Γ that denotes
the total number of search sessions during the time period
2. BACKGROUND we are considering to conduct this study.
Now, we can define several objective functions for the e-
2.1 LTR and Ecommerce Search commerce search such as the following:
Conversion rate per visit (CRV):
Learning to rank (LTR) algorithms [19] involve learning a
function for optimizing a relevance measure using an offline t=T i=N j=M
1 XX X
labeled training data set. This function can be a neural net, gCRV = p(ΛS
ijt = 1|qi , ζj , t)
boosted tree [3, 22], support vector machine [15] or linear Γ t=1 i=1 j=1
models [13] etc. The common ranking measures [14] include
normalized discounted cumulative gain (NDCG), mean re- Revenue per visit (RPV):
ciprocal ranking (MRR), mean average precision (MAP) t=T i=N j=M
etc. The loss functions based on these ranking measures 1 XX X
gRP V = p(ΛS
ijt = 1|qi , ζj , t)ρj
are mostly non-smooth, however researchers [5, 17] derive Γ t=1 i=1 j=1
LTR algorithms such as LambdaMart [4] to address such
problems. There is also a work on multi-objective learning We now also define a relevance based objective as follows:
to rank [22] that uses a click rate along with the human t=T i=N
judged ratings to augment the computation of gradient di- 1 XX
gREL (π) = RM (π(qi , t))
rection from pairwise training data in LambdaMart algo- Γ t=1 i=1
rithm. This algorithm can potentially be modified to learn
to maximize multiple graded objectives where the priorities where RM can be any relevance measure such as normal-
are known in advance. There are not too many papers on ized discounted cumulative gain (NDCG), which is gener-
LTR algorithms applied to e-com data. Recently, researchers ally defined based on how well the ranked list π(q) matches
from Walmart [16] have conducted some experiments with the ideal ranking produced based on human annotations of
multiple LTR algorithms with e-com data. In their obser- relevance of each item to the query. The aggregation func-
vation, LambdaMart algorithm with a sales and click rate tion does not have to be a sum (over all the queries); it
based objective worked well for their data set. In a sepa- can also be, e.g., the minimum of relevance measure over all
rate paper, researchers from Amazon [21] reported use of the queries, which would allow us to encode the preference
gradient boosted tree based regression methods [8] for LTR for avoiding any query with a very low relevance measure
application. They also reported using various normalization score. Note that it is possible to define the NDCG based on
techniques to avoid the biases generated due to the heavy ratings created from direct user feedback such as clicks, add-
reliance of historical behavioral data in regression. to-cart, or sales, since comparing items based on relevance
in e-commerce can be hard [21]. In this paper, we assume
2.2 Online learning and Multi-armed bandit that the relevance function uses a rating system without
Multi-armed bandit (MAB) algorithms [11, 20] have a rich any specific notion of how that rating system is constructed.
With this generality, we can define the following objective approach is that we can directly build on the existing LTR
for relevance: framework, though the new objective function would pose
new challenges in designing effective and efficient optimiza-
gREL (π) = mini∈[1,N ],j∈[1,T ] RM (π(qi , t)) tion algorithms to actually compute optimal rankings. One
Now,we define a notion of discovery for an e-com engine. disadvantage of this strategy is that we cannot easily control
To formalize the notion of discoverability, we say that the the tradeoff between different objectives (e.g., we sometimes
LTR function f is β-discoverable if all items are shown at may want to set a lower bound on one objective rather than
least β times. Now, we can further define a β-discoverability to maximize it). To address this limitation, we may allow
rate as the percentage of items that are impressed at least some objectives to be treated as constraints.
β times in a fixed period of time. Let us now define again
a binary variable γi for every item ζi and then assume that
4.2 Incremental Optimization of e-Com Search
γi = 1 if the item got shown in the search results for β times An alternative strategy is to take an existing LTR ranking
and γi = 0 in case the item is not shown in the search results function as a basis and seek to improve the ranking (e.g., by
more than β times. We can express this as follows: perturbation) so as to optimize multiple objectives as de-
Pi=|Z| scribed above; such an incremental optimization strategy is
i=1 γi more tractable as we will be searching for solutions to the op-
gβ−discoverability =
|Z| timization problem in the neighborhood of an existing rank-
ing function. Specifically, we instantiate the optimization
We can further define another objective sales through rate problem discussed in the previous section by assuming that
(STR) to denote the number of items sold in a period of time the engine incrementally perturbs a learning to rank func-
as percentage of the inventory. This objective represent the tion slightly in a way that the original relevance does not
efficiency of selling items faster. Let us define again a binary vary too much and then each function π = (f1 , ..., fT ) at
variable λs for every item ζs and then assume that λs = 1 time step t = 1 to t = T gradually improves upon multiple
if the item got sold and λs = 0 in case the item is not sold. objectives. Intuitively, we start with a function f and gen-
Thus, the STR objective function when using policy π can erate the next function in the “neighborhood” of an existing
be written as follows: ranking function f within a relevance bound and we take
i=|Z|
X each step in a way that it attempts to improve the sales and
gST R (π) = λi revenue with high probability and it also tries to achieve bet-
i=1 ter β-discoverability and STR for the items in its inventory.
Now, we define the objective of an e-com search engine as However, the search space for all such ranking functions is
the following: still intractable and we need a heuristic to obtain a “good”
Pareto optimal solution for the multi objective optimization
maximize gCRV , gRPV , gREL , gβ−discoverability , gSTR problem.
We thus need to construct an optimization framework that
This is a multi-objective optimization and has an explo-
integrates an exploration mechanism with a learning to rank
ration component in the form of discovery objective. The
paradigm.
priorities of these objectives can be business dependent. For
example, a business aiming to maximize the gross revenue 4.3 Discovery Framework
can prioritize maximizing revenue and on the other hand, a One way of constructing a framework that optimizes an
business aiming to improve sales growth may want to maxi- e-com site for multiple business objectives along with an
mize the conversions. Similarly, a business that depends on exploration option can be to use a fixed number of rank
periodically bringing new items in the inventory may want positions say x for exploration and use the rest of the top
to prioritize the discovery objective along with sales and rev- (K − x) positions for exploitation. In this set-up, we can
enue. In most real scenarios, such business priorities come construct an explore set of items E where all the items have
from the strategic directions of an e-commerce platform. been shown less than β times. We then select top x candi-
We now discuss the strategies for solving this optimization dates for a query qi from the intersection of the recall set of
problem. items for the query and the set E. Let’s call this set corre-
sponding to query qi as Ei . The items selected for the rest
4. STRATEGIES FOR SOLVING THE OP- of the (K − x) positions can be based on a suitable multi-
TIMIZATION PROBLEM objective LTR ranking function. This is a framework that
is practical and the cost of the exploration can be estimated
Since there are multiple objectives to optimize, it is impos-
by the loss of revenue for selecting not the most optimal
sible to directly apply an existing Learning to Rank (LTR)
items for x positions. Note, it is also possible to select any
method to optimize all the objectives. However, there are
rank positions to show the explored items. However, in this
multiple ways to extend an LTR method to solve the prob-
paper, let us assume that we keep those x items at top for
lem as we will discuss below.
keeping our framework simple to analyze.
4.1 Direct extension of LTR 4.4 LTR and Discoverability
One simple strategy is to combine the multiple objectives
Now that we have a formal model for the e-commerce
with appropriate weighting parameters so as to form one
search optimization problem, we want to establish the fol-
single objective function, which can then be used in a tra-
lowing point:
ditional LTR framework to find a ranking that would opti-
mize the consolidated objective function. The weights can • A LTR function is not optimal for discoverability cri-
be based on the business priorities. The advantage of this teria.
on sales over impression for the query. If an item ζj is in
the set E ∩ Ri and is shown bj times in T iterations, and is
In an e-com site, only top K items are shown to the
sold aj times inqbetween, then the UCB score of the item ζj
customers. Suppose there is an item ζ that is rele- a
vant to a set of queries Q. Now, assume that for each is ucbj = bjj + 2 logb2jT +1 . Note, this is for a specific query.
query q ∈ Q there are K other relevant old items for We then select x items based on top UCB scores.
which f (X2)) > 0. Assuming that the K items are Explore LTR (eLTR In this, we define a function that
equally relevant, we can say that the ζ has same value we call explore LTR (eLTR) to select the x items. The rest
of f (X1) compared to the k old items. Then, ζ will al- of the items for top K can be chosen using the traditional
ways rank below the top K items for the set of queries LTR. Then, we can either keep the x items on top or we can
in Q. Suppose the current number of impressions for rerank all K items based on eLTR.
ζ is less than β, then it has no chance of receiving im- The main motivation for the eLTR is the observation that
pression under the ranking framework using function f there is inherent overfitting in the regular ranking function
because it will always be ranked below the top K items used in an e-com search engine that hinders exploration,
unless there is new query that ranks this item higher. i.e., hinders improvement of β-discoverability and STR. The
This means that as long as we can find a percentage of overfitting is mainly caused by a subset of features derived
items that receives below β impressions, it is hard to from user interaction data. Such features are very important
improve the β-discoverability using a traditional LTR as they help inferring a user’s preferences, and the overfit-
mechanism without any integrated discovery mecha- ting is actually desirable for many repeated queries and for
nism. items that have sustained interests t users (since they are
“test cases” that have occurred in the training data), but
5. DISCOVERY STRATEGIES it gives old items a biased advantage over the new items,
limiting the increase of β-discoverability and STR. Thus the
In this section, we propose several strategies to select the x main idea behind e-LTR is thus to separate such features
items for discovery that aims to minimize the loss of business and introduce a parameter to restrict their influences on the
metrics. ranking function, indirectly promoting STR. Formally, we
note that in general, a ranking function can be written in
5.1 Auction based strategy the following form:
In this strategy, the items for top x positions can be as-
signed based on an auction strategy where sellers can bid y = f (X) = g(f1 X1), f2 (X2))
for the positions. This strategy has been used in sponsored where y ∈ R denotes a ranking score and X ∈ RN is a
search for search engine companies such as Google and in N dimensional feature vector, X1 ∈ RN 1 and X2 ∈ RN 2
sponsored product for e-commerce company such as Ama- are two different groups of features such that N 1 + N 2 =
zon. Design of such auction mechanism that balances the N, X1 ∪ X2 = X. The two groups of features are meant to
revenue and relevance is a complex subject and has been distinguish features that are unbiased (e.g., content match-
discussed in recent papers by Ghose et al. [10] and Athey at ing features) from those that are inherently biased (e.g.,
al. [1]. However, many e-commerce businesses may not want clickthrough-based features). Here g is an aggregation func-
to use auction strategy to aid discovery since it can hurt the tion which is monotonic with respect to both arguments. It
growth of seller participation on those sites. is easy to show that any linear model can be written as a
monotonic aggregation function. It is not possible to use
5.2 Exploration with LTR (eLTR) such representation for models such as additive trees. How-
This is a strategy discussed in a recent paper by Goswami ever, our previous techniques do not have such limitation
et al. [12]. In this strategy, we define the set from which since they are completely separated from the LTR. In this
the LTR function selects the items as Li ⊂ Ri for a given paper, we keep our discussion limited to linear models. We
query qi . We assume that all the items outside set Li are now define explore LTR (eLTR) function as follows:
not β-discoverable. Then, L = ∪i=N i=1 Li is the set of all β-
y e = fe (X) = g(f (X1), × f (X2))
discoverable items. Hence, the set E = R \ L can then be
consisting of all the items that require exploration. where y e ∈ R and 0 ≤ ≤ 1 is a variable in our algorithmic
Goswami et al. [12] discuss three strategies to incorporate framework. Since, g is monotonic, fe (X) <= f (X) when
discovery in an e-commerce search. ≤ 1. Since feature set X2 is a biased feature set favoring
Random selection based exploration from the re- old items, we can expect ranking based on f e would be more
call set (RSE): This is a baseline strategy for continuous in favor of new items in comparison with the original f ,
exploration with a LTR algorithm. In this, for every query achieving the goal of emphasizing exploration of new items.
qi , we randomly select x items from the set E ∩Ri . Then, we Note that controls the amount of exploration: the smaller
put these x items on top of the other (k − x) items that are is, the more exploration (at the cost of exploitation). Since
selected using LTR from the set Ri . The regret here will be the maximum exploration is achieved when =0, in which
linear with the number of search sessions with exploration. case, ranking is entirely relying on f1 , the only loss in the
Upper confidence bound (UCB) based exploration original objective function is incurred by the removal of f2 .
from the recall set (UCBE): This is another simple strat- By controlling what features to be included in f2 , we can
egy that uses a variant of UCB based algorithm for explo- control the upper bound of the loss. In this sense, eLTR
ration instead of random sampling. Here, we maintain a ensures a “safe” exploration strategy since f1 is always active.
MAB for each query. We consider each item in the set Note, this composed function gradually becomes the LTR
E ∩ Ri as an arm for the MAB corresponding to a query function when tends to 1. There can be various ways of
qi . We maintain an UCB score for each of those items based constructing the such as the following:
eLTR basic exploration (eLTRb): In this strategy, lenges. Here we discuss some ideas for evaluating the pro-
I
we keep = Tmax . Here, I is an iteration and Tmax is a posed e-LTR algorithm, which we hope will stimulate more
maximum number of iteration after which everything can work in this direction.
be reset. This is a very simple strategy where the eLTR just Given a set of queries and an initial ranking function, an
increases the importances of the behavioral features gradu- e-LTR method is expected to “discover” the ideal ranking
ally with every iteration. for each query, over a sequence of iterations. Ideal ranking
eLTR ucb weighted exploration (eLTRu): In this may be defined as one that maximizes business metrics like
ucb
strategy, we keep = Ujj . Here, Uj is a normalization sales, revenue etc. Since exploration is involved, one can
factor and in our experiment it is chosen to be the maximum expect the quality of ranking (N DCG) and consequently
UCB score in the set E ∩ Ri . sales/revenue to fluctuate for individual queries during the
eLTR ucb weighted exploration and reranking (eL- discovery process. However, the benefit of an e-LTR method
TRur): This strategy first selects the top x items using lies in its ability to deliver greater aggregate sales/revenue
eLTRu and it selects the remaining (k − x) items using the over some reasonable number of iterations N , compared to
classic LTR and then it reranks the k items using eLTRu. a non-discovery LTR method. We must thus compare our
Goswami et al. [12] show in their paper that eLTR can methods on the iterative growth in quality of ranking, and
be considered as a monotonic sub-modular function and can aggregate gain in business metrics.
thus be approximated by a greedy algorithm. The ideal approach for conducting such an evaluation would
require simultaneously deploying all candidate methods to
5.3 Adaptive submodular function based eLTR live user traffic, and computing various user engagement
(AeLTR) metrics such as click through rate,sales, revenue etc. How-
ever this strategy is difficult to implement in practice. Since
This is an extension of the eLTR mechanism presented by
user traffic received by each candidate method is different,
Goswami et al. [12]. However, it makes the strategy much
we need to direct substantial amount of traffic to each method
more general by allowing us to use any LTR rank function
to make observations comparable and conclusions statisti-
going beyond the limitation of linear LTR models. In this,
cally significant. Deployment of a large number of experi-
we use the following function:
mental and likely sub-optimal ranking functions, especially
y a = fa (X) = a f (X) when evaluating baselines, can result in significant business
losses for e-Commerce search engines. More importantly,
Here, the a is a function of UCB as follows: such an approach of using live user traffic would not allow
ucbj us to have a fair comparison of any future algorithm with
a = the ones that we test today, so it does not facilitate further
Uj
study of the problem.
This algorithm has been discussed in a paper by Gabillon Perhaps a good and feasible strategy is to design a simulation-
et al. [9] and is proven to be also a monotonic sub-modular based counterfactual evaluation method [18]. Here, unlike a
function. Hence, a simple greedy algorithm can give a nice traditional static test collection, we also have to introduce
approximation also for this. We however do not know how models to simulate user behavior and model the biases ob-
this strategy can be compared in terms of losses and the served in real system. This can be a topic itself and we omit
discovery metric with respect to the other approaches that further discussion on this as it is out of scope for this short
we have discussed in this paper. paper.
5.4 Regret of the eLTR strategies
The RSE function has linear regret since it is random. The
7. EXPERIMENTAL RESULTS
regret for this can be arbitrarily bad and can thus be given In this section, we first construct a synthetic historical
by O(Rl ). The regret for the UCBE algorithm can be given dataset with queries, items and their prices. We also gener-
by O(log (Rl )). Both eLTR and AeLTR have similar regret ate the true purchase probabilities and utility scores for the
because they both uses monotonic sub-modularity property. item and query pairs. Additionally, we use a specific rank
These algorithms pull items using a score that is a func- function to simulate the behavior of a trained LTR model.
tion of the LTR score and are expected to provide a better Then we conduct a simulation as described in section ??
loss compared to the approaches that are based on multi- with various exploration strategies. During the simulation
armed bandit algorithms. However, it is not immediately we use the observed purchase probabilities estimated from
clear how these algorithms can optimize the discoverability the purchase feedback as the most important feature for the
objective. The regret of the eLTR group of algorithms can rank function but we use the true probabilities generated
|E| during the initial data generation phase to simulate the user
be estimated by O(1+1/e− x ) times worst compared to the
behavior.
optimal which comes from their monotonic sub-modularity
The main goal of this experimental study is to evaluate
property. The eLTR algorithms can be considered to be sim-
the behavior of the exploration strategies with different dis-
ilar to the -greedy approaches where the regret bound is not
tributions of the utility scores representing different state of
better than the UCB algorithm. However, in our simulation,
the inventory in an e-com company.
we want to study the discoverability metrics for such eLTR
We evaluate our algorithms by running the simulation for
algorithms.
T times. We compute RPV and β-discoverability at the end
of T iterations. We also compute a purchase based mean re-
6. EVALUATION METHODOLOGY ciprocal ranking [6] metric (MRR). This metric is computed
Due to the involvement of multiple objectives, the eval- by summing the reciprocal ranks of all the items that are
uation of E-Com search algorithms also presents new chal- purchased in various user visits for all queries. Moreover,
we also discretize our gold utility score between 1 to 5 and Algorithms RPV NDCG MRR β−d
generate a rating for each item. This also allows us to com- LTR 0.12 0.96 0.54 0.10
pute a mean NDCG score at k-th position for all the search RSE 0.06 0.76 0.21 0.78
sessions as a relevance metric. UCBE 0.08 0.87 0.28 0.66
We expect to see that the RPV and NDCG of the LTR eLTRb 0.09 0.94 0.33 0.67
function will be the best. however the β-discoverability val- eLTRu 0.09 0.94 0.33 0.67
ues will be better in ranking policies that use an exploration eLTRur 0.09 0.94 0.33 0.67
strategy. The new ranking strategies will incur loss in RPV AeLTR 0.10 0.95 0.40 0.51
and NDCG and based on our theoretical analysis we ex-
pect the eLTR methods to have less loss compared to the
RSE and UCB based approaches in those measures. We Table 1: Simulation of eLTR framework, with |Q| = 100, |Z| =
also expect to see a loss in MRR for all exploration meth- 5000, |L| = 200, β − d = 10%, β = 50, K = 6, x = 3, T = 50000.
ods. However, we mainly interested in observing how these
algorithms perform in β-discoverability metric compared to
LTR. through the search results one after another from the top
and can purchase an item based on that item’s purchase
probability for that query. Note, in order to keep the simu-
7.1 Synthetic data generation lation simple, we consider an user only purchases one item
We first generate a set of N queries and M items. We in one visit and leaves the site after that. We also can ap-
then assign the prices of the items by randomly drawing a ply a discount to the probability of purchase logarithmically
number between a minimum and a maximum price from a 1
for each lower rank by multiplying log (r+1) , where r is the
2
multi-modal Gaussian distribution that can have up to 1 to ranking position of the item. This is to create an effect of
10 peaks for a query. We select the specific number of peaks the position bias [7].
for a query uniform randomly. We also assign a subset of
the peak prices to a query to be considered as the set of it’s 7.2 Description of the experimental study
preferred prices. This makes a situation where every query
We conduct two sets of experiments with this simulated
may have a few preferred price peak points where it may
data.
also have the sales or revenue peaks.
In the first set of experiments we use a Gaussian distri-
Now that we have the items and queries defined, we ran-
bution with mean 0.5 and the variance 0.1 for generating
domly generate an utility score, denoted by (uij ) for every
the utility scores, but everything else is same as the previ-
item ζj for a query qi . In our set up, we use uniform ran-
ous experiment. This means that the utility of the items
dom, Gaussian and a long tailed distribution for selecting
in the inventory follows a normal distribution. We generate
the utilities. These three different distributions represent
the scores for all the products from a Gaussian distribution
three scenarios for a typical e-com company’s inventory. Ad-
with mean 0.5 and the variance 0.1. The table 1 shows the
ditionally, we generate a purchase probability between 0.02
tables with final metrics for all the algorithms. We observe
to 0.40 for every item for every query. We generate these
that with a Gaussian distribution of utility scores the eLTR
probabilities such that they correlates with the utility score.
approaches have better MRR, and β-discoverability. We ob-
We generate these numbers in a way so that these are cor-
serve that the AeLTR approach works well by producing
related with the utility scores with a statistically significant
good MRR and yet the β-discoverability is better than LTR.
(p-value less than 0.10) Pearson correlation coefficient [24].
This is a powerful result since it shows that it is possible to
We also intend to correlate the purchase probability with
integrate this with more sophisticated LTR algorithms
the preferred peak prices for a query. Hence, we give an
In the second set of experiment, we use a power law to
additive boost between 0 to 0.1 to the purchase probability
generate the utility distribution. This means that only a
in proportion to the absolute difference of the price of the
small set of items here can be considered valuable in this
item from the closest preferred mean price for that query.
scenario. The table 2 shows the final metrics for this case.
By generating the purchase probabilities in this way, we en-
We notice that even with this distribution of utility scores
sure that the actual purchase probabilities are related to the
the eLTR variants have smaller loss in RPV, NDCG, and
preferred prices for the queries, and also it is related to the
in MRR. Note that in this distribution, the discoverability
utility scores of the items for a given query. Now, we de-
can be considered to be naturally not so useful since a large
fine a β-discoverability rate β = b and selects b × M items
number of items are not that valuable. We expect in such
randomly from the set of all items. In our simulation, we
situation, a nice discoverability algorithm can help to elimi-
assume that the estimated (observed) purchase probability
nate items that do not get sold after sufficient exposure and
for all the items in set E at the beginning can be zero. The
enable the e-com company to optimize it’s inventory. We
rest of the items purchase probability are assumed to be
observe a very similar trend for AeLTR algorithm as our
estimated correctly at the beginning. Now, we create a sim-
previous experiment. It gives better RPV and MRR close
ple rank function that is a weighted linear function of the
to the LTR algorithm but it has less β-discoverability com-
utility score (u), the observed purchase probability (po ), and
pared to eLTR approaches.
the normalized absolute difference between the product price
and the closest preferred mean price (m̂p for the query such
that l = 0.60po + 0.20u + 0.20m̂p . Here l denotes the score 8. CONCLUSIONS AND FUTURE WORK
of the ranker. This ranker simulates a trained LTR function This paper represents a first step toward formalizing the
in e-com search where usually the sales can be considered emerging new E-Com search problem as an optimization
the most valuable behavioral signal. problem with multiple objectives including the sales(CRV),
We now construct an user model. Here, an user browses revenue (RPV), and discoverability besides relevance. We
Algorithms RPV NDCG MRR β−d [10] A. Ghose and S. Yang. An empirical analysis of search
LTR 9.56 0.57 0.45 0.11 engine advertising: Sponsored search in electronic
RSE 7.55 0.27 0.25 0.76 markets. Management Science, 55(10):1605–1622,
UCBE 7.55 0.27 0.26 0.66 2009.
eLTRb 8.2 0.33 0.31 0.67 [11] J. Gittins, K. Glazebrook, and R. Weber. Multi-armed
eLTRu 8.3 0.33 0.31 0.67 bandit allocation indices. John Wiley & Sons, 2011.
eLTRur 8.4 0.33 0.32 0.67 [12] A. Goswami, C. Zhai, and P. Mohapatra. Learning to
AeLTR 9.4 0.41 0.40 0.40 rank and discover for e-commerce search. In
International Conference on Machine Learning and
Data Mining in Pattern Recognition.
Table 2: Simulation of eLTR framework, with |Q| = 100, |Z| = [13] L. Hang. A short introduction to learning to rank.
5000, |L| = 200, β − d = 10%, β = 50, K = 6, x = 3, T = 50000.
IEICE TRANSACTIONS on Information and
Systems, 94(10):1854–1862, 2011.
formally define these objectives and discuss multiple strate- [14] K. Järvelin and J. Kekäläinen. Ir evaluation methods
gies for conducting exploration with the learning to rank for retrieving highly relevant documents. In
algorithms. We hope that our work will open up many new Proceedings of the 23rd annual international ACM
directions in research for optimizing e-com search. The obvi- SIGIR conference on Research and development in
ous next step is to empirically validate the strategies based information retrieval, pages 41–48. ACM, 2000.
on log data from an e-com search engine. The theoreti- [15] T. Joachims. Optimizing search engines using
cal framework also enables many interesting ways to further clickthrough data. In Proceedings of the eighth ACM
formalize the e-com search problem and develop new effec- SIGKDD international conference on Knowledge
tive e-com search algorithms. Finally, the proposed discov- discovery and data mining, pages 133–142. ACM,
ery framework is just a small step toward solving the new 2002.
problem of optimizing discoverability in e-com search; it is [16] S. K. Karmaker Santu, P. Sondhi, and C. Zhai. On
important to further develop more effective algorithms by application of learning to rank for e-commerce search.
using these algorithms as baselines. In Proceedings of the 40th International ACM SIGIR
Conference on Research and Development in
Information Retrieval, SIGIR ’17, pages 475–484, New
9. REFERENCES York, NY, USA, 2017. ACM.
[1] S. Athey and D. Nekipelov. A structural model of [17] H. Li. A short introduction to learning to rank. IEICE
sponsored search advertising auctions. In Sixth ad TRANSACTIONS on Information and Systems,
auctions workshop, volume 15, 2010. 94(10):1854–1862, 2011.
[2] P. Auer, N. Cesa-Bianchi, and P. Fischer. Finite-time [18] L. Li, S. Chen, J. Kleban, and A. Gupta.
analysis of the multiarmed bandit problem. Machine Counterfactual estimation and optimization of click
learning, 47(2-3):235–256, 2002. metrics in search engines: A case study. In Proceedings
[3] C. Burges, T. Shaked, E. Renshaw, A. Lazier, of the 24th International Conference on World Wide
M. Deeds, N. Hamilton, and G. Hullender. Learning to Web, pages 929–934. ACM, 2015.
rank using gradient descent. In Proceedings of the [19] T.-Y. Liu. Learning to rank for information retrieval.
22nd international conference on Machine learning, Foundations and Trends in Information Retrieval,
pages 89–96. ACM, 2005. 3(3):225–331, 2009.
[4] C. J. Burges. From ranknet to lambdarank to [20] A. Mahajan and D. Teneketzis. Multi-armed bandit
lambdamart: An overview. Learning, 11(23-581):81, problems. In Foundations and Applications of Sensor
2010. Management, pages 121–151. Springer, 2008.
[5] C. J. Burges, R. Ragno, and Q. V. Le. Learning to [21] D. Sorokina and E. Cantu-Paz. Amazon search: The
rank with nonsmooth cost functions. In Advances in joy of ranking products. In Proceedings of the 39th
neural information processing systems, pages 193–200, International ACM SIGIR Conference on Research
2007. and Development in Information Retrieval, SIGIR ’16,
[6] N. Craswell. Mean reciprocal rank. In Encyclopedia of pages 459–460, New York, NY, USA, 2016. ACM.
Database Systems, pages 1703–1703. Springer, 2009. [22] K. M. Svore, M. N. Volkovs, and C. J. Burges.
[7] N. Craswell, O. Zoeter, M. Taylor, and B. Ramsey. An Learning to rank with multiple objective functions. In
experimental comparison of click position-bias models. Proceedings of the 20th international conference on
In Proceedings of the 2008 International Conference World wide web, pages 367–376. ACM, 2011.
on Web Search and Data Mining, WSDM ’08, pages [23] J. Vermorel and M. Mohri. Multi-armed bandit
87–94. ACM, 2008. algorithms and empirical evaluation. In European
[8] J. H. Friedman. Greedy function approximation: a conference on machine learning, pages 437–448.
gradient boosting machine. Annals of statistics, pages Springer, 2005.
1189–1232, 2001. [24] R. R. Wilcox. Introduction to robust estimation and
[9] V. Gabillon, B. Kveton, Z. Wen, B. Eriksson, and hypothesis testing. Academic press, 2011.
S. Muthukrishnan. Adaptive submodular [25] Y. Yue, J. Broder, R. Kleinberg, and T. Joachims.
maximization in bandit setting. In Advances in Neural The k-armed dueling bandits problem. Journal of
Information Processing Systems, pages 2697–2705, Computer and System Sciences, 78(5):1538–1556,
2013. 2012.