<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Multimodal Recommender System for Large-scale Assortment Generation in E-commerce</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Murium</forename><surname>Iqbal</surname></persName>
							<email>miqbal@overstock.com</email>
						</author>
						<author>
							<persName><forename type="first">Adair</forename><surname>Kovac</surname></persName>
							<email>akovac@overstock.com</email>
						</author>
						<author>
							<persName><forename type="first">Kamelia</forename><surname>Aryafar</surname></persName>
							<email>karyafar@overstock.com</email>
						</author>
						<author>
							<affiliation key="aff0">
								<address>
									<settlement>Midvale</settlement>
									<region>Utah</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff1">
								<address>
									<settlement>Midvale</settlement>
									<region>Utah</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<address>
									<settlement>Midvale</settlement>
									<region>Utah</region>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<settlement>Ann Arbor</settlement>
									<region>Michigan</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Multimodal Recommender System for Large-scale Assortment Generation in E-commerce</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">101AC59FFF32D13F1A0EAE01BF8E141F</idno>
					<idno type="DOI">10.1145/nnnnnnn.nnnnnnn</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T06:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>recommender system</term>
					<term>visual document</term>
					<term>topic modeling</term>
					<term>set recommendation</term>
					<term>quadratic knapsack problem</term>
					<term>product recommendation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>E-commerce platforms surface interesting products largely through product recommendations that capture users' styles and aesthetic preferences. Curating recommendations as a complete complementary set, or assortment, is critical for a successful e-commerce experience, especially for product categories such as furniture, where items are selected together with the overall theme, style or ambiance of a space in mind. In this paper, we propose two visually-aware recommender systems that can automatically curate an assortment of living room furniture around a couple of pre-selected seed pieces for the room. The first system aims to maximize the visual-based style compatibility of the entire selection by making use of transfer learning and topic modeling. The second system extends the first by incorporating text data and applying polylingual topic modeling to infer style over both modalities. We review the production pipeline for surfacing these visually-aware recommender systems and compare them through offline validations and large-scale online A/B tests on Overstock 1 . Our experimental results show that complimentary style is best discovered over product sets when both visual and textual data are incorporated.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Overstock 1 is an e-commerce platform with the goal of creating dream homes for all. Users browse Overstock's catalog to select pieces that complement one another while matching the stylistic settings and color palettes of their rooms. As furniture is not a disposable product, furniture purchases are subject to careful scrutiny of aesthetics and strict budgeting. Brick and mortar stores often inspire consumers by creating furniture showrooms, in some cases pushing consumers to walk through their carefully selected furniture displays. These assorted showrooms alleviate the creative and stylistic pressure on the consumer, since the set is already on display with all the necessary pieces.</p><p>By crafting an appropriate recommender system, this experience can be recreated on an e-commerce platform. Figure <ref type="figure" target="#fig_0">1</ref> illustrates an example of the proposed system's automatically generated showroom, built around two seed products selected from Overstock 1 . The goal of this recommender system is to provide set recommendations that adhere to a general theme or cohesive visual style while accounting for essential item constraints 2 . Set recommendations have become more prominent with the rise of subscription box services in various domains such as fashion (e.g. StitchFix 3 ), jewelry (e.g. Rocksbox 4 ) and beauty products (e.g. Birchbox 5 ). The criteria for selecting an assortment depends heavily on the product space which can inform the latent space of user preferences and product representations. Personal care product recommendations, for example, rely on user preferences, such as skin type, which can be inferred from product text-based attributes. Fashion, jewelry and furniture shopping, on the other hand, are predominantly visual experiences. This motivates a visually-aware representation for products and style-based preferences for users.</p><p>We propose two assortment recommender systems. The first takes advantage of visually semantic features transferred from a deep convolutional neural network to learn style. The second uses both these visual features and product text-based attributes to learn style across the multimodal dataset. Our hypothesis is that while the visual style can help find similar products, use of text data in conjunction with images will result in more complimentary, and cohesive stylized assortments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>As the prevalence of recommendation systems grows, e-commerce platforms are looking to increase the influence on customer purchases by not just recommending one item at a time but several in a bundle together <ref type="bibr" target="#b24">[25]</ref> [27] <ref type="bibr">[1] [6]</ref>. This is true across industries, from travel agencies to clothing <ref type="bibr" target="#b9">[10]</ref> [14] <ref type="bibr" target="#b14">[15]</ref>. Specifically Zhu et al. <ref type="bibr" target="#b26">[27]</ref> model the problem of bundle recommendations as the Quadratic Knapsack Problem (QKP), and find an approximation to it. They use only implicit feedback data, rather than content data to build their system, allowing for the bundles to be personalized. Using this approach, we also view our system as a QKP, but look to leverage primarily content data, incorporating implicit feedback as a minor signal. To capture a user's preference, the system we propose allows a user to present a seed product, around which we can create the bundle. This allows us to circumvent the cold start problem, as any user, regardless of having a history on our site, can still build assortment recommendations. New products are also seamlessly incorporated, as long as images or text data is made available.</p><p>Recommendation systems is a well studied field which has two large over arching types: collaborative filtering methods <ref type="bibr" target="#b17">[18]</ref> [12] <ref type="bibr" target="#b6">[7]</ref>, and content based methods <ref type="bibr" target="#b3">[4]</ref> [21] <ref type="bibr" target="#b21">[22]</ref> of which many rely on learning latent representation of the input data. Topic modeling, specifically Latent Dirichlet Allocation (LDA) <ref type="bibr" target="#b2">[3]</ref>, has also been applied to create topic based recommendations both off of text input as well as from implicit feedback <ref type="bibr" target="#b10">[11]</ref> [2] <ref type="bibr" target="#b25">[26]</ref>. These systems work by scoring individual products against one another or against users to see which are most similar. Here we look to extend this methodology by applying topic modeling over both image and text data. Rather than simply concatenating image and text features and linearly combining them, PolyLDA enables us to learn two distinct but coupled latent style representations. This allows for a versatile interpretation of style <ref type="bibr" target="#b12">[13]</ref>. We then use these learned styles to make bundle or assortment recommendations rather than the traditional single item recommendations.</p><p>To facilitate style discovery from the images, we process them via prevalent deep learning techniques. Deep residual networks have recently been shown as a powerful model to capture style based preferences for creating visual wardrobes from fashion images <ref type="bibr" target="#b9">[10]</ref>. With the goal of learning visually coherent styles, we apply transfer learning, by using a Resnet-50 <ref type="foot" target="#foot_0">6</ref> which was pre-trained on ImageNet<ref type="foot" target="#foot_1">7</ref>  <ref type="bibr" target="#b7">[8]</ref>. We explore interpreting the convolutional neural network by indexing the activations on channels within the convolutional layers <ref type="bibr" target="#b18">[19]</ref>. We use a neural network to learn filters for our data and simply index their responses to images to create visual documents, similar to the older bag of word methods <ref type="bibr" target="#b23">[24]</ref>. To discover visually-aware trends, we use these documents with Latent Dirichlet Allocation (LDA) <ref type="bibr" target="#b2">[3]</ref>.</p><p>While we use LDA for topic modeling on single-modality image data, we need an extension to interpret both visual semantic features and text-based attribute data as complementary modalities. Roller et al. <ref type="bibr" target="#b19">[20]</ref> create a multimodal topic model that assumes words and corresponding visual features occur in pairs and should be captured as tuples within the topic distribution. This model would work well for descriptions of images, but the underlying assumption could prove to be too stringent to generalize to our application. Mimno et al. <ref type="bibr" target="#b16">[17]</ref> offer a more flexible extension of LDA, Polylingual LDA (PolyLDA), which handles tuples of documents in different languages with assumed identical topic distributions. The documents within a tuple do not need to be direct translations of one another, and the topics themselves have distinct sets of words for each language. By treating our image data and our text data as two separate languages, we can use PolyLDA directly to create a more flexible multimodal topic model with which to infer style.</p><p>Hsiao and Grauman <ref type="bibr" target="#b9">[10]</ref> also apply PolyLDA to infer style from images, but they use complementary clothing types (e.g. pants and blouse) as languages in a compatibility model instead of using PolyLDA to handle different modalities of data for the same item. The documents they use for their topic models are generated using a Resnet trained to automatically detect their pre-defined text attributes within images and apply the appropriate labels.</p><p>Much of the work performed on traditional recommendation systems can be applied here, but additional constraints must be taken into consideration. Products can't simply be visually similar to be purchased together, but must be complimentary as well. For example, having two similar brightly colored upholstered chairs as a result in a traditional recommendation is perfectly acceptable, but if they are both taken into one assortment with the intention of being bought together, they may clash. Here we show that by incorporating text data, we are able to avoid assortments of incredibly visually similar, interchangeable products which are not at all complimentary. These nuances can be further addressed by incorporating purchase data to infer what types of products are not just similar, but also compatible, and likely to be bought together. To create the final assortment recommendations, we define a distance measure that combines stylistic similarity with implicit feedback from users, in the form of purchases, to develop a concept of affinity similar to approaches in composite recommendation systems such as <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b22">23,</ref><ref type="bibr" target="#b26">27]</ref>.</p><p>This rest of this paper is organized as follows: In Section 3, we present our production pipeline for assortment recommendations as deployed in Overstock 1 and explore our product embeddings as used with topic modeling techniques. We also review our formulation of assortment generation as a 0-1 quadratic knapsack problem and our greedy optimization for building assortments. Section 4 presents our offline and online experimental results. Finally, we conclude our paper in Section 5 and propose future directions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">METHODOLOGY</head><p>Here we provide an overview of the production system used to surface recommendations throughout Overstock 1 and describe our assortment recommender system in detail. We first explain our product embeddings as text-based bag-of-words and bag-of-visualwords (BoVW) documents. Then we explain the LDA approaches used to discover text-based and visual styles across our platform. We finally discuss our formulation of assortment generation as a 0-1 quadratic knapsack problem with budget constraints and our greedy assortment recommendations algorithm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">System Overview</head><p>Product recommendations on Overstock 1 are surfaced through multiple carousels across the platform. User interactions with products and recommendations is recorded along side purchase data. Each product on the platform is represented by corresponding title, text attribute tags, text-based descriptions and images. The first step in generating recommendations is processing the large volume of product and user data which are hosted on Hadoop Servers.</p><p>After initial processing on Hadoop, the data is transfered to a onebox CUDA-enabled GPU server. Here product images are passed through a publicly available Resnet-50 to create image-based documents. Both the image and text documents are then fed into Mallet<ref type="foot" target="#foot_2">8</ref> to generate topic representations of products. We create two topic variants, a visual variant, that uses the image data with LDA and a multimodal variant, that uses both text and image data with PolyLDA.</p><p>Once the topic distributions are created, the products reside within their defined feature space. Hereafter, the methodology for both variants is identical. We construct a distance measure over the topic space, using implicit feedback from our users, to model compatibility. The assortments are then generated by finding products which minimize this distance to predefined seed product pairs.</p><p>The generated assortments are pushed to our Cassandra database cluster. Our website servers can query Cassandra directly to see We threshold these responses to determine whether or not the "word" each channel represents is present within the image. This methodology seems to work particularly well for our upholstered furniture, showing distinct responses to the different patterns.</p><p>which products have an associated assortment to present to users. Assortment recommendations are featured on corresponding product pages, below the product description, providing an option for users to purchase an entire assortment, or just parts of it. A brief diagram of the systems involved is presented in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Product Embeddings</head><p>Products on Overstock 1 have several images of the item for sale and different forms of associated text data including title, description and attribute tags. Product attributes are descriptive tags associated with a product. Some example attribute categories are color, size, style, composing materials, finish, and brand. In this paper, we utilize these attributes as the primary text-based information since they often provide a rich text representation of the item that compliments provided images. All product attribute data is processed to remove stop words and are concatenated with each product title to form a bag-of-words document.</p><p>We use the methodology described in <ref type="bibr" target="#b12">[13]</ref> to create visual documents to be used with LDA via a Resnet. We then combine both visual and text data together via PolyLDA.</p><p>The convolutional layers within a Resnet are composed of a series of small learned filters. Each filter is convolved with the input in steps along it's height and width. The result of this process is a 2-dimensional grid of activations that represents the response from the filter at corresponding locations of the input. The learned filters within a Resnet respond to specific patterns. By viewing these filters as words, and indexing them, Iqbal et al. <ref type="bibr" target="#b12">[13]</ref> show that a BoVW document can be created which, when combined with LDA, is effective at uncovering style. To index a layer, we simply threshold the response from the filter to indicate whether or not it is sufficiently activated by the image to be included within the image's document. Given this methodology for creating image documents, our visual word vocabulary is defined entirely by the channels in the layers that we choose to index. We tested creating documents using various layers through the network. We found that combining several middle layers gave us the best results. After experimenting we empirically chose to use Resnet-50 layers 8, 18, 31, and 43. The combination of these layers provided significantly better results, per visual inspection, than any single layer or other combinations. This yielded a total vocabulary size of 2816 visual words over our image documents. Since multiple images can be provided for each product, we take the union of all present visual words within the corresponding images as the visual document for the product itself. The process is depicted in Figure <ref type="figure" target="#fig_3">4</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Topic Modeling</head><p>Once the visual documents are created, they can be used directly with LDA to create topics for our visual variant. LDA is an unsupervised generative model which infers probability distributions of words within topics and models documents as a mixture of those topics <ref type="bibr" target="#b2">[3]</ref>. Given a set of documents and a number of target topics the model assumes the following generative process was used to create the documents.</p><p>(1) For M documents, initialize the set of topic distributions, θ i ∼ Dir(α) for {i = 1 . . . M} (2) For K topics, initialize the set of word distributions, ϕ k ∼ Dir(β) for {k = 1 . . . K} (3) For the k t h word in the i t h document select a topic z i from θ i and a word, w i,k , from ϕ z i</p><p>where Dir is the Dirichlet distribution.</p><p>The model initializes the parameters randomly then iteratively updates these parameters via Gibbs Sampling and variational Bayes inference. After many iterations the topic distributions converge to a stable state and the resulting topics can be used as a low dimensional feature space that captures the salient content of the original documents.</p><p>For the multimodal variant, each visual document is paired with it's text equivalent in a tuple. These tuples are then used with Polylingual LDA (PolyLDA) <ref type="bibr" target="#b16">[17]</ref>. Products that are missing either a text document or a visual document can still be used with PolyLDA, as it is robust to missing documents within tuples. A diagram of the system for multimodal document creation is provided in Figure <ref type="figure" target="#fig_3">4</ref>.</p><p>PolyLDA is an extension of LDA meant to handle loosely equivalent documents in different languages. It takes as input tuples of documents, each in a different language but with the same exact topic distribution. Each resulting latent topic from this method contains a distinct associated set of word distributions for each language. Documents aren't required to be direct translations of one another, which allows for flexibility. A word in one language could be attributed to a given topic while its direct translation in another language may be attributed to a different topic. Words that only appear in one language can also be present in topics despite having no equivalent translations in the corpus. PolyLDA extends LDA by assuming the following generative procedure:</p><p>(1) For a given tuple of documents, {d 1 . . . d L }, initialize a single set of topic distributions, θ ∼ Dir(α) (2) For K topic sets with L languages, initialize the set of word distributions, ϕ k,l ∼ Dir(β) for {k = 1 . . . K } and {l = 1 . . . L} (3) For the k th word in the l th document in the i th tuple select a topic z i,l from θ i and a word, w i,l,k , from ϕ z i,l</p><p>For our multimodal variant, we make use of the flexibility afforded by PolyLDA by applying it to our visual and text documents, representing each modality as a distinct language attempting to describe the same product. This method should allow for topics that afford better generality. Intuitively for our corpus, we suspect that a topic could capture relationships between certain visual features, like thick wood, with textual attributes, like "quality", that don't have a direct visual representation but do share an underlying contextual relationship. By combining modalities the model also affords us the ability to infer a more complete style representation given incomplete data. For example, we can apply our multimodal topics to products with only text attributes available and still infer style that would normally be captured by the missing image data.</p><p>We provide visualizations of the resulting topics from both variants in Figure <ref type="figure" target="#fig_4">5</ref>. To create these visualizations we selected the products that are maximally aligned with each topic. In our case, we have empirically seen that the topics relate well to various styles while reducing our feature space from a large number of attribute tags, title words, and image-based channel activations to a succinct k-dimensional space (where k is the number of topics). These styles are not necessarily complimentary though, especially in the case of the visual topics. We can see that the items within the topics seem more substitutable, e.g. in the cases of brightly colored furniture, similarly patterned upholstery, and mirrored surfaces.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Knapsack Problem</head><p>Once we have all our products residing within the topics' feature space, we can begin to build out our assortments. The process for generating the assortments is the same for both the visual variant and the multimodal variant, with the only difference between the two being the topic representations of products. To build these assortments, we first define the necessary components of an assortment as verticals. The verticals we define here enable us to build living room assortments, but we would like to emphasize that verticals can be defined for any e-commerce platform. If desired, the verticals can be left undefined, and the assortment can be built with no vertical constraints. For our specific application verticals are critical pieces of living room furniture, e.g. coffee table, chair, or accent table. Our definitions are manually applied to the products and are listed below. (Here a couch set can be either a sectional sofa or a sofa and loveseat combination.)</p><formula xml:id="formula_0">Verticals: {Couch Set, Coffee Table, Accent Table, Entertainment Center, Bookshelf, Ottoman, Chair}</formula><p>We chose to build our assortments around seed pieces. For our current application, we generated seeds, but we can easily allow users to provide their own seeds around which we can curate an assortment. We selected seeds as preferred pairings of the most crucial verticals in any living room space, namely a couch set and a coffee table. Although all the other verticals can be seen as optional, these two verticals are defining members of a living room and therefore must be present in any assortment. We chose pairs of sofas and coffee tables which are most frequently co-clicked for our seeds.</p><p>We then assume a budget constraint for the entire assortment. As such we can formulate our assortment generation as a 0-1 quadratic knapsack problem (0-1 QKP) as defined by Gallo et al <ref type="bibr" target="#b4">[5]</ref>. The 0-1 QKP assumes that a knapsack with limited space needs to be filled with a set of items. These items all have an intrinsic value. Items also have an additional associated pair-wise profit when selected together for the knapsack.</p><p>For a given seed, we can find the optimal assortments by trying to find the products that are nearest the seed in the topic space. We can add the additional constraint that we want all products within the assortment to be nearest one another in the topic space as well. The total assortment must remain within the assumed budget constraint. Each product has an associated vertical, and we can add in some constraints on the minimum and maximum number of products each vertical can contribute to the final assortment.</p><p>To calculate proximity we use a Mahalanobis distance built on purchase data similar to <ref type="bibr" target="#b15">[16]</ref>. The Mahalanobis distance allows us to leverage the implicit feedback of our users to understand how our topics distribute and relate to one another across multiple products within a purchase.</p><p>We assume that purchases of living room furniture made by the same user over a 3 month period are being used in the same room, and as such can be said to represent a compatible assortment. After trimming to include only purchases with 3-10 pieces of living room furniture, so as to avoid noise and erroneous signals from bulk purchasers, we create a dataset of roughly 7.5k purchased assortments. We take the L 2 normalized sum of the topic distributions of the copurchased products as a topic representation of the purchase itself, and use this dataset to calculate the covariance matrix used for our Mahalanobis distance.</p><p>The Mahalanobis distance is defined as d M (x i ,x j ) = (x i -x j )M(x ix j ) T , Where M is the covariance matrix for the topics learned from our purchase data. For each seed pair of coffee table and couch set, we greedily swap products in and out of our assortment until we </p><formula xml:id="formula_1">for i ∈ {V−S} do t A ← ( t p ) ∀p j ∈ A j,pr e , ∀j ∈ V i 8: while size(A i ) &lt; i size do A i ← A i ∪ argmin(d M (t o , t A )) 10: δ ← δ + d M (t A i , t A i,pr e )</formula><p>converge to an ideal set of products. The only verticals which are not allowed to change are those in the seed.</p><p>To formulate the O-1 QKP we first define some terminology. Each product o will have a vertical label a o , an associated price c o , and associated topic distribution t o . All products will have a pairwise distance associated with them and every other product d i, j which is the Mahalanobis distance between products o i and o j built on the purchases and topic distributions. Additionally the products all have a score q o to represent their value to the seed, here we use the inverse of the distance from the product to the seed. Let M be the set of products comprising the assortment. Let B be the total available budget minus the cost of the seed. The 0-1 knapsack problem can then be formulated as:</p><formula xml:id="formula_2">maximize M i=1 q i + M j=1 1/d i, j , j i subject to: ∀ i : M i=1 c i ≤ B ∀ k : min k ≤ count(a o = k) ≤ max k<label>(1)</label></formula><p>A greedy approximation can now be formulated to follow the constraints of the 0-1 QKP. First a solution is initialized, then iteratively, items are swapped until the system converges to an optimal solution. To get the initial set, products are all scored by their total potential for compatibility (the sum of the score of this product with the sum total of all reciprocals of pairwise Mahalanobis distance) divided by their price. Products are then sorted by this ratio and the highest scored are added until vertical constraints or budget constraints are met. Then each product is considered for a swap with other products to improve the total compatibility of the assortment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5">Greedy Method</head><p>For our online evaluation, we relax the budget constraints, so that we can display the assortment on product pages. This allows us to select the best assortment per product. We can easily reapply the budget constraint for user assortments. As such, we use the following greedy method to generate the assortments. The assortment is initialized as first only containing the seed items, all other verticals are empty. Each vertical is then iteratively considered for new products while holding all other verticals constant. New products are chosen from the pool of products that are labeled as candidates for this vertical. Those which are closest to the total assortment (minus the current vertical) at the previous time-step are added in, until the size of the vertical is met. The sizes of the verticals in our experiment were set by us, but a user can dictate how many side tables, bookshelves, chairs, etc they wish to add to their assortment.</p><p>Let V be the set of all verticals. S is the set of verticals used in the seed. A i is the set of products selected for the assortment in vertical i for the current timestep. Each vertical i has an associated size, i size , which is the number of products we want for this vertical. A i,pr e is the set of products selected for the assortment in vertical i for the previous timestep. p i is an element of A i,pr e . o i is a candidate product for vertical i. t o is the topic distribution of a product. d M is the Mahalanobis distance as defined above. δ is the change in the distance of the assortment from iteration to iteration. The greedy approach is described using this notation in Algorithm 1. Figure <ref type="figure">6</ref> illustrates the generated assortments using this method for both the visual and multimodal variants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">RESULTS AND DISCUSSION</head><p>This section describes a large-scale experiment on Overstock 1 to determine whether simultaneously learning style from text and images provides better results than learning style from only images. All offline validations and online A/B test assume 2 variants: a visual-only assortment recommender system, which is built on top of BoVW representation of product images, and a multimodal one, which is built on text-based attributes and visual words. Our site does offer manually curated collections, but there is little overlap on the products selected in the manual process and those selected for our automated system. The methodology used to create these collections also involves selecting products from the same vendor from the same product line, which can often be viewed as substitutable products as well. These would thus serve as a poor comparison to our model, which attempts to find stylistically complimentary products of different categories regardless of vendor, rather than similar/identical products. As such we do not use them as a baseline to compare our model against.</p><p>Online Evaluation. We run an A/B test on Overstock 1 with both the visual variant and multimodal variant assortments on product pages. The users are provided an option to add an entire assortment or individual items from the recommendations module to their cart. User engagement with recommendations are measured via clickthrough rate, CTR, on assortment recommendations is a classic measure of user engagement with product recommendations in ecommerce platforms. Our findings show that the multimodal variant outperforms the visual variant, showing a statistically significant lift of 10.9% relative to a visual only baseline.</p><p>Offline Evaluations. As we are recommending sets of products together, we can't use traditional offline evaluations, such as AUC, MAP and Recall. For scoring, we have taken the assortments and scored them with the average click-based Jaccard coefficient calculated pairwise over all products contained within the assortment. We make use of logs of product clicks over the last 2 months to build this score. This tells us how compatible the various products within an assortment are to one another.</p><formula xml:id="formula_3">(a) (b) (c) (d) (e) (f) (g) (h)</formula><p>Figure <ref type="figure">6</ref>: Assortments from our visual variant (top) and multimodal variant (bottom) with the same seeds are depicted above. The visual assortments tend to provide a more-of-the-same assortment, while the multimodal is able to create a diverse set of products that still forms a cohesive style. When the visual assortment can't find products similar to either seed, it selects products which are similar to one another. In some cases it can match products to one seed, and ignores the other, which leaves the second seed item looking out of place. The multimodal variant, which can take advantage of text data as well, is able to find pieces to create a cohesive look incorporating elements of both seeds.</p><p>To calculate the assortment Jaccard, we look at S, the set of all user browsing sessions that included clicks on at least 2 items of living room furniture. A user browsing session only spans one visit, so we can assume that when items are co-clicked within the same session, the user still has the same intent, and is implicitly seeing both products as satisfying this intent. We take all user browsing sessions over a two month span for our dataset. We denote the subset of S that includes sessions that have clicked on product a as S a .</p><p>We calculate the Jaccard coefficient for all pairs of products (a, b) within our corpus as follows:</p><formula xml:id="formula_4">(a, b) = |S a ∩ S b | |S a ∪ S b | In contrast to the standard Jaccard formulation, if |S a ∪ S b | = 0, we consider (a, b) = 0.</formula><p>This modification is to prevent products that were never visited together from being counted as perfectly compatible with each other, since the standard formulation defaults these cases to 1.</p><p>For assortment A we calculate an assortment Jaccard score A as a simple average of the pairwise Jaccard coefficient for all items within the assortment, excluding the Jaccard coefficient between the two seeds. An example assortment with a good Jaccard score is shown in Figure <ref type="figure" target="#fig_0">1</ref>.</p><p>The multimodal maintains a much higher average assortment Jaccard than the visual-only variant. The dramatic difference between </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Modality</head><p>Avg Max visual 0.0015 0.0220 multimodal 0.0027 0.0362 the multimodal variant and the visual variant may reflect an underlying bias in the click data: The text modality looks at the same data used in site search and navigation to surface groups of products to users, so co-click data may reflect better discoverability of items with similar textual features as opposed to similar visual features.</p><p>We also examine the distribution of topics from each assortment. The topic representation of products in our PolyLDA generated space is more diverse than that of the LDA. The products in the LDA space usually are strongly attributed to one or two products, while the PolyLDA does a better job of representing a product as a mixture of topics. This affects style as the products within the topics that are learned are very visually similar, lending themselves to substitutability rather than complimentary behavior.</p><p>Figure <ref type="figure">7</ref> shows how many topics compose each assortment. The distribution for the visual variant is heavily skewed, with the majority of assortments only being composed of a few topics. The multimodal </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Number of Topics in Assortment</head><p>Total Assortments</p><p>Visual Multimodal Figure <ref type="figure">7</ref>: The above chart tabulates the number of topics that nontrivially contribute to an assortment. The vast majority of visual assortments have 1-4 topics associated with them, as depicted by the peak in the red line. On the other hand, the multimodal assortments can have anywhere from one to 20 topics associated with them, which allows for more diverse product representation. This is desirable as the uncovered topics show very similar items, not necessarily complimentary ones as depicted in Figure <ref type="figure" target="#fig_4">5</ref>. A user would likely not want an entire set of bright red furniture. Mixing among the topics to create a complimentary set is more desirable than an entire set of the exact same visual features.</p><p>variant has much more variety, with assortments composed of only a few topics, to assortments composed of nearly half the topics being equally likely. This shows us that the multimodal variant offers more diverse assortments. This is preferred since, as depicted in 5, the topics themselves often contain substitutable rather than complimentary items. Properly blending the topics together to create a cohesive look results in better assortments, as depicted in Figure <ref type="figure">6</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">CONCLUSION</head><p>Online shopping is a visual experience. Visually-aware recommendations are crucial to online shopping and e-commerce platforms. Yet, systems which rely solely on images suffer from a lack of diversity. By incorporating both image and text data we are able to create cohesive styles.</p><p>In this paper we introduced a deep visually-aware large-scale assortments recommender system for Overstock 1 . Our assortment recommender system takes advantage of product images to create visually coherent trends from Overstock products. We introduced two variants: a visual-only variant and a multimodal variant. Our visual-only variant creates a bag-of-visual-words representation of product images by thresholding the activations from specific layers of a pre-trained deep residual neural network, Resnet-50. It then applies topic modeling (LDA) on product image representations to create visual trends among Overstock products. We then proposed a greedy approach ( with and without budget constraints ) to create assortment recommendations based on seed items that maximize the visual compatibility of the set. Our multimodal variant takes advantage of text-based product attributes in addition to image representation. This variant utilizes Polylingual LDA (PolyLDA) to create trends that are based on two modalities, images and text.</p><p>We have featured multiple assortments generated from both models. We have evaluated our results through a set of offline validations and an online large-scale A/B test on Overstock. Our experimental results indicate that incorporating both image and text data provides more a more cohesive visual style than using only images and can enhance user engagement metrics with recommendations module. We also show that PolyLDA provides a meaningful way to simultaneously learn style across text and image data.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: An automatically generated assortment from the multimodal approach is shown.</figDesc><graphic coords="1,321.75,183.55,232.65,153.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: A system overview diagram is presented above. We use product data and user engagement data residing within our Hadoop cluster and our NFS/S3 store to train the model. Most of the training is done on a one box GPU server, which runs the Resnet-50, Mallet and our greedy algorithm. All results are served to Overstock 1 users from Cassandra.</figDesc><graphic coords="2,68.71,82.69,210.42,127.73" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Here we depict activations from the layers used to create our visual documents. The leftmost column is the raw image. The next column to the right is the cropped patch which is passed through the network. In the next 4 columns, we see the response to these patches from 9 channels each of convolutional layers 8, 18, 31 and 43. Lighter colors mean strong responses, while darker colors mean weak responses.We threshold these responses to determine whether or not the "word" each channel represents is present within the image. This methodology seems to work particularly well for our upholstered furniture, showing distinct responses to the different patterns.</figDesc><graphic coords="3,322.18,82.71,231.77,144.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Each product's corresponding text and images are collected. Our system then discards images with noisy scenes, keeping only those with white backgrounds and passing them through Resnet-50. Channel activations from layers 8,18,31, and 43 are thresholded and indexed to create visual documents. The union of these documents is taken as the visual representation of the product. The text data is stripped of stopwords and paired with the corresponding visual document to create a tuple for ingestion by PolyLDA</figDesc><graphic coords="4,53.95,82.69,239.94,115.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Representative items from each of six topics in the visual variant on the left (a) and multimodal variant on the right (b) are depicted above.</figDesc><graphic coords="5,53.80,92.69,265.53,250.61" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Algorithm 1 Generate Assortments δ ← ϵ + 1 2: A i,pr e ← {}∀i∈ {V−S} A i,pr e ←{p i }∀i∈ {S} 4: while δ ≥ ϵ do δ ← 0</figDesc><table><row><cell>6:</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Jaccard Score averages and max values for assortments of both variants.</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_0">https://github.com/facebook/fb.resnet.torch</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_1">http://www.image-net.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_2">http://mallet.cs.umass.edu/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Recommender systems for product bundling</title>
		<author>
			<persName><forename type="first">Moran</forename><surname>Beladev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lior</forename><surname>Rokach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bracha</forename><surname>Shapira</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge-Based Systems</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<biblScope unit="page" from="193" to="206" />
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Comparing Topic Models for a Movie Recommendation System</title>
		<author>
			<persName><forename type="first">Sonia</forename><surname>Bergamaschi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Po</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Serena</forename><surname>Sorrentino</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">WEBIST</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="172" to="183" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Latent dirichlet allocation</title>
		<author>
			<persName><forename type="first">Andrew</forename><forename type="middle">Y</forename><surname>David M Blei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><forename type="middle">I</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><surname>Jordan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of machine Learning research</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="993" to="1022" />
			<date type="published" when="2003-01">2003. Jan (2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Metalearning for Context-aware Filtering: Selection of Tensor Factorization Algorithms</title>
		<author>
			<persName><forename type="first">Tiago</forename><surname>Cunha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Carlos</forename><surname>Soares</surname></persName>
		</author>
		<author>
			<persName><forename type="first">André Cplf</forename><surname>Carvalho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eleventh ACM Conference on Recommender Systems</title>
				<meeting>the Eleventh ACM Conference on Recommender Systems</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="14" to="22" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Quadratic knapsack problems</title>
		<author>
			<persName><forename type="first">Giorgio</forename><surname>Gallo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><forename type="middle">L</forename><surname>Hammer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bruno</forename><surname>Simeone</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Combinatorial optimization</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1980">1980</date>
			<biblScope unit="page" from="132" to="149" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Design of a shopbot and recommender system for bundle purchases</title>
		<author>
			<persName><forename type="first">Robert</forename><surname>Garfinkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ram</forename><surname>Gopal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arvind</forename><surname>Tripathi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fang</forename><surname>Yin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Decision Support Systems</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="1974" to="1986" />
			<date type="published" when="2006">2006. 2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Scalable Recommendation with Hierarchical Poisson Factorization</title>
		<author>
			<persName><forename type="first">Jake</forename><forename type="middle">M</forename><surname>Prem Gopalan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><forename type="middle">M</forename><surname>Hofman</surname></persName>
		</author>
		<author>
			<persName><surname>Blei</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">UAI</title>
		<imprint>
			<biblScope unit="page" from="326" to="335" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Deep residual learning for image recognition</title>
		<author>
			<persName><forename type="first">Kaiming</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiangyu</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shaoqing</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE conference on computer vision and pattern recognition</title>
				<meeting>the IEEE conference on computer vision and pattern recognition</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="770" to="778" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Learning compatibility across categories for heterogeneous item recommendation</title>
		<author>
			<persName><forename type="first">Ruining</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Charles</forename><surname>Packer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julian</forename><surname>Mcauley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Mining (ICDM), 2016 IEEE 16th International Conference on. IEEE</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="937" to="942" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Creating Capsule Wardrobes from Fashion Images</title>
		<author>
			<persName><forename type="first">Wei-Lin</forename><surname>Hsiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristen</forename><surname>Grauman</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1712.02662</idno>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Style in the long tail: Discovering unique interests with latent variable models in large scale social e-commerce</title>
		<author>
			<persName><forename type="first">Diane</forename><forename type="middle">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rob</forename><surname>Hall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Josh</forename><surname>Attenberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining</title>
				<meeting>the 20th ACM SIGKDD international conference on Knowledge discovery and data mining</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1640" to="1649" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Collaborative filtering for implicit feedback datasets</title>
		<author>
			<persName><forename type="first">Yifan</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yehuda</forename><surname>Koren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chris</forename><surname>Volinsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICDM&apos;08. Eighth IEEE International Conference on. Ieee</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="263" to="272" />
		</imprint>
	</monogr>
	<note>Data Mining</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Discovering Style Trends through Deep Visually Aware Latent Item Embeddings</title>
		<author>
			<persName><forename type="first">Murium</forename><surname>Iqbal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adair</forename><surname>Kovac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kamelia</forename><surname>Aryafar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CVPR Workshops</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Personalized travel package recommendation</title>
		<author>
			<persName><forename type="first">Qi</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yong</forename><surname>Ge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhongmou</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Enhong</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hui</forename><surname>Xiong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Mining (ICDM), 2011 IEEE 11th International Conference on. IEEE</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="407" to="416" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Hi, magic closet, tell me what to wear!</title>
		<author>
			<persName><forename type="first">Si</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiashi</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zheng</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tianzhu</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hanqing</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Changsheng</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shuicheng</forename><surname>Yan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th ACM international conference on Multimedia</title>
				<meeting>the 20th ACM international conference on Multimedia</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="619" to="628" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Image-based recommendations on styles and substitutes</title>
		<author>
			<persName><forename type="first">Julian</forename><surname>Mcauley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Christopher</forename><surname>Targett</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Qinfeng</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anton</forename><surname>Van Den</surname></persName>
		</author>
		<author>
			<persName><surname>Hengel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="43" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Polylingual topic models</title>
		<author>
			<persName><forename type="first">David</forename><surname>Mimno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jason</forename><surname>Hanna M Wallach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">David</forename><forename type="middle">A</forename><surname>Naradowsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Andrew</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><surname>Mccallum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2009 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="880" to="889" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Probabilistic matrix factorization</title>
		<author>
			<persName><forename type="first">Andriy</forename><surname>Mnih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ruslan</forename><forename type="middle">R</forename><surname>Salakhutdinov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="1257" to="1264" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Understanding trained CNNs by indexing neuron selectivity</title>
		<author>
			<persName><forename type="first">Ivet</forename><surname>Rafegas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maria</forename><surname>Vanrell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luís</forename><forename type="middle">A</forename><surname>Alexandre</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1702.00382</idno>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A multimodal LDA model integrating textual, cognitive and visual modalities</title>
		<author>
			<persName><forename type="first">Stephen</forename><surname>Roller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sabine</forename><surname>Schulte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Im</forename><surname>Walde</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2013 Conference on Empirical Methods in Natural Language Processing</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="1146" to="1157" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Deep content-based music recommendation</title>
		<author>
			<persName><forename type="first">Aaron</forename><surname>Van Den Oord</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sander</forename><surname>Dieleman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Benjamin</forename><surname>Schrauwen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="2643" to="2651" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Using content-based filtering for recommendation</title>
		<author>
			<persName><forename type="first">Robin</forename><surname>Van Meteren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Maarten</forename><surname>Van Someren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Machine Learning in the New Information Age: MLnet/ECML2000 Workshop</title>
				<meeting>the Machine Learning in the New Information Age: MLnet/ECML2000 Workshop</meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="47" to="56" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Learning visual clothing style with heterogeneous dyadic cooccurrences</title>
		<author>
			<persName><forename type="first">Andreas</forename><surname>Veit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Balazs</forename><surname>Kovacs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sean</forename><surname>Bell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julian</forename><surname>Mcauley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kavita</forename><surname>Bala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Serge</forename><surname>Belongie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Vision (ICCV), 2015 IEEE International Conference on. IEEE</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="4642" to="4650" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Feature hashing for large scale multitask learning</title>
		<author>
			<persName><forename type="first">Kilian</forename><surname>Weinberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anirban</forename><surname>Dasgupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><surname>Langford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><surname>Smola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Josh</forename><surname>Attenberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 26th Annual International Conference on Machine Learning</title>
				<meeting>the 26th Annual International Conference on Machine Learning</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="1113" to="1120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Breaking out of the box of recommendations: from items to packages</title>
		<author>
			<persName><forename type="first">Min</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laks</forename><surname>Vs Lakshmanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><forename type="middle">T</forename><surname>Wood</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the fourth ACM conference on Recommender systems</title>
				<meeting>the fourth ACM conference on Recommender systems</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2010">2010</date>
			<biblScope unit="page" from="151" to="158" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">A recommendation system combining LDA and collaborative filtering method for Scenic Spot</title>
		<author>
			<persName><forename type="first">Shengli</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yifan</forename><surname>Feng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Information Science and Control Engineering (ICISCE), 2015 2nd International Conference on. IEEE</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="67" to="71" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Bundle recommendation in ecommerce</title>
		<author>
			<persName><forename type="first">Tao</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrick</forename><surname>Harrington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Junjun</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lei</forename><surname>Tang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 37th international ACM SIGIR conference on Research &amp; development in information retrieval</title>
				<meeting>the 37th international ACM SIGIR conference on Research &amp; development in information retrieval</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="657" to="666" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
