<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Scaling up Analogy with Crowdsourcing and Machine Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Joel Chan</string-name>
          <email>joelchuc@cs.cmu.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tom Hope</string-name>
          <email>tom.hope@mail.huji.ac.il</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dafna Shahaf</string-name>
          <email>dshahaf@cs.huji.ac.il</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aniket Kittur</string-name>
          <email>nkittur@cs.cmu.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Human-Computer Interaction Institute Carnegie Mellon University</institution>
          ,
          <addr-line>Pittsburgh PA 15213</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Computer Science and Engineering Hebrew University of Jerusalem</institution>
          ,
          <addr-line>Jerusalem</addr-line>
          ,
          <country country="IL">Israel</country>
        </aff>
      </contrib-group>
      <fpage>31</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>Despite tremendous advances in computational models of human analogy, a persistent challenge has been scaling up to find useful analogies in large, messy, real-world data. The availability of large idea repositories (e.g., the U.S. patent database) could significantly accelerate innovation and discovery in a way never previously possible. Previous approaches have been limited by relying on hand-created databases that have high relational structure but are very sparse (e.g., predicate calculus representations). Traditional machine-learning/information-retrieval similarity metrics (e.g., LSA) can scale to large, natural-language datasets; however, while these methods are good at detecting surface similarity, they struggle to account for structural similarity. In this paper, we propose to leverage crowdsourcing techniques to construct a dataset with rich “analogy-tuning” signals, used to guide machine learning models towards matches based on relations rather than surface features. We demonstrate our approach with a crowdsourced analogy identification task, whose results are used to train deep learning algorithms. Our initial results suggest that a deep learning model trained on positive/negative example analogies from the task can find more analogous matches than an LSA baseline, and that incorporating behavioral signals (such as queries used to retrieve an analogy) can further boost its performance.</p>
      </abstract>
      <kwd-group>
        <kwd>Analogy</kwd>
        <kwd>crowdsourcing</kwd>
        <kwd>machine learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Invention by analogy (i.e., transferring ideas from other domains that are
structurally similar to a target problem) is a powerful way to create new innovations.
For example, a car mechanic invented a new low-cost way to ease difficult
childbirth by drawing an analogy to a cork extraction method in wineries (inserting
and inflating a small plastic bag in the bottle) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This award-winning device
has the potential to change lives worldwide, particularly women in developing
countries with limited medical resources.
      </p>
      <p>The recent growth of online innovation repositories represents an
unparalleled opportunity for invention by analogy. These repositories contain hundreds
of thousands (Quirky, OpenIDEO) or millions (the U.S. patent database, the
Web) of ideas that have the potential to be applied to other structurally
similar domains. However, the scale of these repositories presents a challenge to a
person’s ability to find useful analogies.</p>
      <p>
        Computational systems could greatly accelerate innovation by mining
analogies from these vast repositories. Indeed, decades of research on computational
models of human analogy-making have yielded tremendous advances in the
ability of computational systems to explain and simulate human-like analogical
reasoning. Yet, a persistent challenge has been scaling up computational
analogy systems to reliably find useful analogies in large, messy,
realworld data. Existing approaches are limited by either relying on hand-created
databases that have high relational structure but are small, domain-specific, and
costly to keep updated [
        <xref ref-type="bibr" rid="ref14 ref17">14, 17</xref>
        ], or on machine learning approaches that can scale
to large datasets but have difficulty encoding and matching relations [
        <xref ref-type="bibr" rid="ref16 ref5">5, 16</xref>
        ].
      </p>
      <p>In this paper, we propose a hybrid approach which combines crowdsourcing
with machine learning to develop a scalable approach to finding analogies in
large idea repositories. A key insight is that we aim to externalize and capture
the mental processes that humans use to find and evaluate analogies to serve
as training data for a machine learning model. The intuition is that instead of
trying to build a complete human-generated dataset or a machine learning model
driven only from existing data, the rich behavioral traces of how people query for
analogies can “tune” a more scalable computational approach towards matches
based on relations rather than surface features.</p>
      <p>We illustrate our approach through a crowdsourced analogy identification
task where people query a repository and find analogical matches to a target.
These matches and queries are used as training data (and for feature selection)
for deep learning algorithms. Our initial results suggest that a deep learning
model trained on positive/negative example analogies from the task can find
more analogous matches than an LSA baseline, and that incorporating
behavioral signals (such as queries entered) can further boost its performance.
2</p>
      <p>
        Related Work
Computational Analogy Systems. We argue that a crucial reason behind
the difficulty in scaling up computational analogy lies in a trade-off between
accuracy and scale in existing approaches. On the one hand, models that have
been the most successful at approaching human-level performance in analogical
matching—such as Hummel and Holyoak’s LISA analogy engine [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], Klenk and
Forbus’s [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] Companion for AP Physics problems, and Vattam and colleagues’
[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] Design Analogy to Nature Engine (DANE)—rely heavily on hand-created
databases that have high relational structure. Creating such databases involves
extensive knowledge engineering efforts. Vattam and colleagues [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] estimate
that converting a single (complex) biological system into a formal
representation requires between forty and one hundred person-hours of work. Consequently,
models that rely on hand-coded relational representations have yet to be
successfully applied to large, open repositories like the U.S. Patent Database.
      </p>
      <p>
        Conversely, a number of machine learning approaches exist that make
minimal assumptions about the input data, and in particular do not require explicitly
coded relational representations. Examples include word embedding models like
Word2Vec [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], vector-space models like Latent Semantic Indexing [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and
probabilistic topic modeling approaches like Latent Dirichlet Allocation [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. While
these approaches scale well to large datasets, they have difficulty encoding
relational similarity. One possible reason is that these approaches tend to rely
on co-occurrence patterns between words that describe higher-level “concepts”;
however, relational categories have very sparse and diverse term distributions [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
      </p>
      <p>
        Note that some approaches [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] try to capture structure by focusing on
particular types of words (e.g., verbs). However, parts-of-speech alone are not enough
to capture structural relations, and these methods suffer from a lot of noise.
Consequently, these systems tend to have low precision of analogical matches,
shifting the burden onto the user to sift through large amounts of false positives.
To illustrate, Fu and colleagues [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] found that, despite their approach
producing structures that experts found sensible, their “far” analogies actually were
perceived as “too far”, and hurt instead of helped creative output.
Crowdsourcing and Machine Learning. We believe a possibly fruitful way
forward lies in hybrid approaches that combine crowdsourcing and machine
learning. We are inspired by related efforts in human-computer interaction that
combine these technologies to crowdsource complex cognition at scale, in
particular clustering related items from rich and messy text sources [
        <xref ref-type="bibr" rid="ref2 ref9">2, 9</xref>
        ]. However,
these methods are not aimed at finding analogical clusters, which requires
supporting deep relational similarity rather than surface similarity.
3
      </p>
      <p>
        Computational Analogy at Scale with Crowdsourcing
and Machine Learning
We frame the problem of finding analogies in a large dataset as a hybrid human
and machine-learning problem. We propose to use crowdsourcing to obtain rich
“analogy-tuning” signals for machine learning models by capturing the process
by which people query for and evaluate analogies. By doing so we aim to collect
not just positive/negative examples of analogies, but also implicit and explicit
behavioral traces ranging from the queries people use to look for analogies to the
keywords they believe are discriminative. We believe these behavioral traces are
vital for bridging the gap between scalable machine learning approaches and the
structured representations of prior approaches. This machine-learning approach
is related to other efforts in case-based-reasoning that use machine-learning to
reduce the need for knowledge engineering [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. To illustrate the potential of this
approach, we present a system that seeds a deep learning model with
analogyrelevant signals harvested from a crowdsourced analogy querying task.
3.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Crowdsourcing Task for Collecting Analogy-Tuning Signals</title>
      <p>
        The goal of the crowdsourcing component is to obtain rich behavioral data (e.g.,
positive/negative examples of analogies, query sequences, keywords) that signal
the core relational structures of documents. Guided by the psychological insight
that comparison is a powerful way to get people to attend to the core structural
features of a description [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], we decided to embed the task of providing
analogytuning signals within a realistic task of finding analogies.
      </p>
      <p>Workers use a simple search interface to find product descriptions that are
analogous to a seed product. Figure 1 depicts the interface and the four main
components of the task: 1) searching for matches (A), 2) screening/processing
matches (B), 3) describing the analogy (C), and 4) highlighting keywords (E).</p>
      <p>This approach yields a rich set of signals that we can use for our machine
learning models. For example:
– What queries are used, in what sequence?
– Which documents are tagged as possible matches?
– Which documents are tagged as best matches?
– Which documents are implicitly rejected (i.e., ignored in the search result
list, despite appearing before matches)?
– Which documents are directly rejected?
– How is the best match described as being analogous to the seed document?
– Which key terms are highlighted?
Importantly, this task context enables us to harvest all these rich signals from a
natural search task that is easy and familiar. Additionally, the task design could
guide the instrumentation of an interface in a user-facing computational analogy
system to yield similar signals for ongoing refinement of the underlying models.
Deployment. We test our approach with a corpus of product descriptions from
Quirky.com, an online crowdsourced product innovation website. Quirky is
representative of the kinds of datasets we are interested in, because it is large (at the
time of writing, it hosts upwards of 10,000 product ideas), unstructured (ideas
are described in natural language), and covers a variety of domains (invention
categories), which makes cross-domain analogies possible. The following example
is representative of the length and “messiness” of product ideas in this dataset:</p>
      <p>Proximity wristband: Control of childs in public places is very difficult and
stressing. Parents fear that their childs may go too far unexpectedly for any reason,
and that actually happens no matter how careful they are, especially in crowdy places.
Because childs always run and move. Parents can’t relax and are obliged to keep their
eyes continuously on their childs. Furthermore, the consequences of a child going too
far from his parents may be very dangerous. A wristband is put on the wrist of the child.
The wristband has a radio connection (bluetooth) with one of the parents’ smartphone,
that has been previously matched with the wristband. Parents may activate/disactivate
the alarm on the wristband by tapping on the App installed on their smartphone. When
the alarm is activated, the App detects the distance between the 2 radio connected
devices (the wristband and the smartphone). If the distance gets higher than the maximum
value (changeable in the settings of the App) than a speaker integrated in the bracelet
emits a loud alarm and the smartphone starts ringing. The inside of the wristband hosts
a circular conductive element that loses its metallic continuity if the wristband opens
for any reason, so if this circuit is opened the wristband is programmed to emit the
alarm and alert the parents.</p>
      <p>We crowdsourced analogy finding within a set of 400 randomly sampled
Quirky products. Three hundred and ninety-four workers from Amazon
Mechanical Turk collected analogies for 227 “seed” documents (median of 1 unique
analogy per seed, range of 1-10 unique analogies). Median completion time for
each seed was 10 minutes, and pay was $6/hr (or $1 per completed seed). Workers
could complete as many seeds as they wanted.</p>
      <p>An example behavioral trace sequence from our data illustrates how the act
of comparison pushes people to focus on structure (along with the rich data
we can mine from the behavioral traces). Worker X received the “proximity
wristband” as a seed. She initially started with the query “alarm”, and tagged
as a possible match a “voltage plug” product that automatically alerts the user
if there are voltage problems for a given power outlet. She also rejected
nonanalogous results like a “smart doorbell chime”. Dissatisfied with the results, she
entered a new query (“wristband”), but didn’t find any useful matches. Finally,
she entered “proximity” as a query, and tagged a product about a “digital dog
fence” as a best match, explaining that both products are about “Proximity,
keeping object within a set distance, the object has it attached.”
3.2</p>
    </sec>
    <sec id="sec-3">
      <title>Task 1: Semantic Similarity from Traces</title>
      <p>
        An important challenge of working with natural language is finding appropriate
semantic similarities. Many existing similarity metrics, like Word2vec [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] do
poorly on verb and adjective similarity, which are central for structural similarity.
      </p>
      <p>We now use behavioral traces to expose similarity. We constructed a graph
from queries entered in our study: Nodes correspond to queries; there is an edge
if the queries resulted in the same analogy. In other words, if one user found an
analogy using the query “cover”, and another used “protect”, we add an edge.</p>
      <p>Figure 2 shows fragments from this graph. By construction, the query graph
exposes a lot of the desired semantic similarity. Semantically similar verbs often
form dense clusters (e.g., protect/defend/shield), and related terms (gps/location/find)
also tend to be a short distance away. Traversing the query graph can also reveal
analogies: for example, an analogy between a dog gps and a child tracking device
(left), and between products that protect from rain, water, and sun (right). This
graph could be used to guide feature selection in a machine learning model.
3.3</p>
    </sec>
    <sec id="sec-4">
      <title>Task 2: Learning Analogies</title>
      <p>We now demonstrate how to frame analogy-finding as a machine learning
problem. In this context, we are given a training set D = {(xi1, xi2, yi)}. (xi1, xi )
2
are pairs of product descriptions. Label yi 2 { 0, 1} corresponds to whether the
pair was tagged as an analogy (“best match”) or as a non-analogy (ignored in
search result, directly rejected). Our goal is learn a decision function for new pairs
whose score reflects their “degree of analogy": f (✓, ( ✓ (xi1), ✓ (xi2))), where ✓ (·)
is function embedding xi into a shared feature space and ✓ are model parameters.</p>
      <p>
        The document model we use to demonstrate our ideas is based on
convolutional neural network (CNN) architecture that has shown state-of-the-art results
in many NLP tasks [
        <xref ref-type="bibr" rid="ref13 ref4">13, 4</xref>
        ]. This distributional model learns to map texts to
vector representations. The objective guides the model to learn a representation
where texts tagged as analogous are close, and non-analogous texts are far.
      </p>
      <p>
        We use a Siamese Network architecture [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], where two identical copies (same
weights) of a function are applied to two inputs. Another layer measures distance
between the two inputs, computing whether they are similar. Figure 3 shows
the main components of the architecture. We represent each word as a
lowdimensional, real-valued dense vector w 2 Rd. Each xi is thus a sequence of
vectors w, which together form a matrix Mi where column j represents the jth
word in the input sequence.
(Pwroodrducetmtebxetdmdiantgrisx) Confevaotluurteiosnal Pooling Fuhlidydceonnnlaeycetred Contrastive Loss Next, our model learns to
compose the w into higher-level semantic
representations by applying
transformations to Mi. Each vector sequence
Dog colar with built-in GPS ∥max( 01, − − ∥ 2  ∥122 +−   2 ∥22) ilsaypears,saepdpltyhirnoguaghbaankcoofn“vsloidluintgiownina-l
dow filters" to extract local features of
small subsequences of words.
      </p>
      <p>To learn non-linear patterns,
convolutional layers are followed by
eleChild wristband with built-in Bluetooth mentwise activation functions. We
Fig. 3. Siamese Network architecture. use the ReLU function, max(0, x).
Words are embedded into a low-dimension The output of the activation
funcrepresentation and combined into a matrix tion is passed through a pooling
(left). Convolutional “sliding window fil- layer, which reduces dimensionality
ters” (blue) are applied to the matrix. Mul- by aggregating information,
capturtiple filters are applied (different colors), ing pertinent patterns and filtering
forming a feature map pooled to form an ag- noise. The pooling layer performs
gregated signal. The pooled representation max-pooling, returning the largest
goes through a fully connected layer. The value for each column. Finally, values
same weights are applied for both inputs pass through a fully connected layer
iancproustss aisllcloamyeprus.teFdin(Calolyn,trdaissttiavneceLobsest)w.een which computes a linear
transformation followed by ReLU non-linearity,
combining local features into a global semantic view of the text.</p>
      <p>This composition of functions, from embedding words to the final layer, yields
our function ✓ (·), mapping a text input xi into a new vector representation.
Crucially, in our model ✓ (·) is shared for (xi1, xi2), enabling the model to learn
a symmetric representation that is invariant to the order in which the texts are
provided. Finally, our objective function is the Contrastive Loss, defined as:
L( ✓ (xi1), ✓ (xi2)) = yiL+( ✓ (xi1), ✓ (xi2)) + (1
yi)L ( ✓ (xi1), ✓ (xi2)),
where</p>
      <p>L+( ✓ (xi1), ✓ (xi2)) = || ✓ (xi1)
✓ (xi2))||22
L ( ✓ (xi1), ✓ (xi2)) = max(0, m ||
✓ (xi1)
✓ (xi2))||22.</p>
      <p>L+ penalizes positive pairs far apart, and L penalizes negative pairs
(direct/implicit rejected pairs, from the crowd) closer than margin m. Learning
is done with gradient descent, via backpropagation.</p>
      <p>Results. We applied the model to the crowd-annotated Quirky data. We split
our data into training and evaluation sets, each containing distinct sets of “seed"
texts (to test the model’s ability to generalize). The training set consists of about
12500 pairs. Positive labels were assigned directly by crowd workers, while
negative labels mean the pairs were implicitly rejected by not being tagged despite
being viewed by a worker. Our evaluation set comprised of roughly 3000 pairs.</p>
      <p>The data was imbalanced, with about 10 negatives for each positive (⇡ 1, 100
positive pairs). To counter the imbalance, we use a weighted loss function. For
the LSA baseline, we compute the Singular Value Decomposition (SVD) on the
document-term matrix, with term weights given by the TF-IDF score. Cosine
similarity is then computed for each pair. The table below shows proportion of
analogies among pairs with highest predicted scores (precision@K). For example,
looking at the top 2% predictions of our model (2% evaluation-set instances with
lowest predicted distance; about 60 pairs), 64% were tagged as positive (vs. 46%
for LSA). The overall proportion of positive labels in the test set was only 11%.
Importantly, negative labels are derived implicitly, so many pairs with negative
annotations could possibly be “mislabeled” and are actually positive.</p>
      <p>Method Top 2% Top 5% Top 10% Top 15% Top 25%</p>
      <p>Siamese Net 0.64 0.54 0.39 0.32 0.28
Latent Semantic Analysis 0.46 0.40 0.34 0.29 0.25
In Table 1 we show examples of seed documents and their best predicted
matches. Our model recovers both “purpose ” matches (e.g., gloves that hold
nails, and a hammer handle to store tools) as well as “mechanism” (child / dog
tracking devices). Comparing our model to LSA-based similarity, one can
qualitatively observe that overall, LSA seems to focus more on surface similarity.
For example, when starting from a wristband that monitors child proximity,
LSA returns a baby recliner with sensors, a lawnmower connected to a
wristband, and skateboard shoes; our model returns a pet tracker, a wallet finder and
vehicle finder. Our model seems to “err" and return smart window blinds that
detect when homeowners are away – possibly recognizing the semantic “analogy”
between a child or pet wandering off and a homeowner being away.
3.4</p>
    </sec>
    <sec id="sec-5">
      <title>Task 3: Incorporating the query</title>
      <p>We now extend our machine learning setting, adding the query as additional
input. Our training data is now D = {(xi1, xi2, qi, yi)}, where qi is the user-entered
query. We apply the same representation ✓ (·) we use for x1, xi2 to qi, embedding
i
the query into a shared feature space. We seek to measure similarity between
seed and target in this space. Importantly, similarity should be relative in some
sense to the user’s query. We do so by using the projection of the query vector
onto the seed and target vectors – essentially “aligning" the query “concept"
with each product text. We then compute the contrastive loss as before. More
formally, we re-define the contrastive loss to incorporate the query as follows.</p>
      <p>L+( ✓ (xi1), ✓ (xi2), ✓ (qi)) = || ✓✓ ((xqi1i))·· ✓✓ ((xxi1i1)) ✓ (xi1) ✓✓ ((xqi2i))·· ✓✓ ((xxi2i2)) ✓ (xi2)||22
and similarly for L . In the table below we report results for this new model
and compare it to the model without the query. Since many queries did not have
matching targets we filter the data to a smaller subset with full information
(6000 less rows). The smaller training set drops base model accuracy to .46
(from .64). Interestingly, the query information has an apparent positive effect,
compensating for lack of data and boosting accuracy from .46 to .54. These
preliminary results validate our intuition that machine learning models would
benefit from a variety of human-generated analogy-relevant signals.</p>
      <sec id="sec-5-1">
        <title>Seed Siamese Net top matches LSA top matches</title>
        <p>Proximity wrist- A dog and cat collar that has Baby recliner - placing sensors on a
band for children built-in GPS tracking. // A wal- baby being walked by parent, learn
with a bluetooth let phone case that can be located movement and incorporate into
reclinconnection to with a key-chain remote control. ing chair. // Clamp on a lawnmower
parents’ smart- // Programmable smart window safety bar with a pin connected to a
phone. App blinds that can also detect when wristband. When user is too far pin
detects the dis- you are away. // A vehicle finder is pulled out and the clamp opens.
tance, emits an with a small portable electronic // Magnetic skateboard shoes with RF
alarm if child is device transmitting a signal from controller for turning on and off. Could
too far. parked car to smartphone. be fabricated into a ring or wristband.
Window louvers Programmable window blinds. // A window-mounted fan with sound
that stop rain A window-mounted fan with sound dampening louvers. // A glare-stopper
by funneling the dampening louvers. // Lamps that device preventing energy-efficient
winrain out. work off batteries for additional dows from melting sides of buildings
lighting without depending on or vehicles // A car screen to help
reelectricity duce excessive heat buildup when
parking outside.</p>
        <p>A flat magnet A golf glove with a magnetic strip A glove with adhesive contours
allowsewn into the around the wrist. // A hammer ing the user to pick up hair and lint in
posterior hand of with a handle that stores nails hard-to-reach places. // A can opener
a glove to hold and hex screwdrivers. // Tear- container device, with a magnet on
nails or screws. proof gloves to work easily with both ends to hold the can in place. //
tools. A box that clamps onto a ladder and
holds screws, nails and small items.</p>
      </sec>
      <sec id="sec-5-2">
        <title>Method</title>
        <p>Model with query
Model without query
Overall, we believe this is a promising time to make traction on the problem
of finding analogies in complex, messy data. There is a confluence of new work
on crowdsourcing complex cognition, machine learning tools for unsupervised
learning of semantics, and hybrid approaches combining crowds and machine
judgments to get the best of both worlds. We believe there is promise in an
approach that positions the process of crowdsourced analogical knowledge base
creation as input to machine learning models, using explicit and implicit
signals to augment the written text. In this paper we describe a prototype of this
approach and early results describing its potential value.</p>
        <p>We note that the approach described here does not tackle a number of
important problems that we acknowledge as limitations. For example, we do not yet
deal with the rich relational structure inherent in the source and target analogs.
We also note that our intent is to find interesting and useful analogs in large data
repositories, and as such we make no claims that the processes we describe match
up with those that human cognition engages in during analogical retrieval and
reasoning. However, our capture of the process by which people engage in while
searching for analogies, including the queries and keywords they use and their
perceived relevance judgments of their resulting matches, may prove valuable for
further psychological research on the process of analogical retrieval.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>D. M. Blei</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>M. I.</given-names>
          </string-name>
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>and J.</given-names>
          </string-name>
          <string-name>
            <surname>Lafferty</surname>
          </string-name>
          .
          <article-title>Latent Dirichlet Allocation</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          , pages
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>J.</given-names>
            <surname>Bragg</surname>
          </string-name>
          and
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Weld</surname>
          </string-name>
          .
          <article-title>Crowdsourcing Multi-Label Classification for Taxonomy Creation</article-title>
          .
          <source>In HCOMP'13</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>S.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hadsell</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>LeCun</surname>
          </string-name>
          .
          <article-title>Learning a similarity metric discriminatively, with application to face verification</article-title>
          .
          <source>In CVPR'05</source>
          , volume
          <volume>1</volume>
          . IEEE,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>R.</given-names>
            <surname>Collobert</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          .
          <article-title>A unified architecture for natural language processing: Deep neural networks with multitask learning</article-title>
          .
          <source>In ICML'08. ACM</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>S.</given-names>
            <surname>Deerwester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Dumais</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. W.</given-names>
            <surname>Furnas</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T. K.</given-names>
            <surname>Landauer</surname>
          </string-name>
          .
          <article-title>Indexing by Latent Semantic Analysis</article-title>
          .
          <source>JASIST</source>
          ,
          <volume>41</volume>
          (
          <issue>6</issue>
          ):
          <year>1990</year>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>K.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kotovsky</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K. L.</given-names>
            <surname>Wood</surname>
          </string-name>
          .
          <article-title>Discovering Structure In Design Databases Through Functional And Surface Based Mapping</article-title>
          . JMD,
          <volume>135</volume>
          :
          <fpage>031006</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>K.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kotovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Schunn</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Wood</surname>
          </string-name>
          .
          <article-title>The Meaning of Near and Far: The Impact of Structuring Design Databases and the Effect of Distance of Analogy on Design Output</article-title>
          . JMD,
          <volume>135</volume>
          (
          <issue>2</issue>
          ):
          <fpage>021007</fpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>D.</given-names>
            <surname>Gentner</surname>
          </string-name>
          and
          <string-name>
            <given-names>K. J.</given-names>
            <surname>Kurtz</surname>
          </string-name>
          .
          <article-title>Relational Categories. In Categorization inside and outside the laboratory: Essays in honor of Douglas L. Medin, APA decade of behavior series</article-title>
          . American Psychological Association, Washington, DC, US,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>N.</given-names>
            <surname>Hahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Kittur</surname>
          </string-name>
          .
          <article-title>The Knowledge Accelerator: Big Picture Thinking in Small Pieces</article-title>
          .
          <source>In CHI'16</source>
          , New York, NY, USA,
          <year>2016</year>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>K.</given-names>
            <surname>Hanney</surname>
          </string-name>
          and
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Keane</surname>
          </string-name>
          .
          <article-title>The adaptation knowledge bottleneck: How to ease it by learning from cases</article-title>
          . In D. B.
          <string-name>
            <surname>Leake</surname>
          </string-name>
          and E. Plaza, editors,
          <source>Case-Based Reasoning Research and Development, number 1266 in Lecture Notes in Computer Science</source>
          , pages
          <fpage>359</fpage>
          -
          <lpage>370</lpage>
          . Springer Berlin Heidelberg,
          <year>July 1997</year>
          . DOI:
          <volume>10</volume>
          .1007/3-540-63233- 6_
          <fpage>506</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Hummel</surname>
          </string-name>
          and
          <string-name>
            <given-names>K. J.</given-names>
            <surname>Holyoak</surname>
          </string-name>
          .
          <article-title>A symbolic-connectionist theory of relational inference and generalization</article-title>
          .
          <source>Psychological Review</source>
          ,
          <volume>110</volume>
          (
          <issue>2</issue>
          ):
          <fpage>220</fpage>
          -
          <lpage>264</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>D. G. M.</surname>
          </string-name>
          <article-title>Jr. Car Mechanic Dreams Up a Tool to Ease Births</article-title>
          . The New York Times, Nov.
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          .
          <article-title>Convolutional neural networks for sentence classification</article-title>
          .
          <source>arXiv preprint arXiv:1408.5882</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>M.</given-names>
            <surname>Klenk</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>Forbus</surname>
          </string-name>
          .
          <article-title>Analogical model formulation for transfer learning in</article-title>
          <source>AP Physics. Artificial Intelligence</source>
          ,
          <volume>173</volume>
          (
          <issue>18</issue>
          ):
          <fpage>1615</fpage>
          -
          <lpage>1638</lpage>
          , Dec.
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>L.</given-names>
            <surname>Kotovsky</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Gentner</surname>
          </string-name>
          .
          <article-title>Comparison and Categorization in the Development of Relational Similarity</article-title>
          .
          <source>Child Development</source>
          ,
          <volume>67</volume>
          (
          <issue>6</issue>
          ):
          <fpage>2797</fpage>
          -
          <lpage>2822</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <article-title>Efficient Estimation of Word Representations in Vector Space</article-title>
          . arXiv:
          <volume>1301</volume>
          .3781 [cs],
          <source>Jan</source>
          .
          <year>2013</year>
          . arXiv:
          <volume>1301</volume>
          .
          <fpage>3781</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>S.</given-names>
            <surname>Vattam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wiltgen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Helms</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Goel</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Yen</surname>
          </string-name>
          . DANE:
          <article-title>Fostering Creativity in and through Biologically Inspired Design</article-title>
          .
          <source>In Design Creativity</source>
          <year>2010</year>
          .
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>