<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Query Algebra for Quantum Information Retrieval</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Annalina Caputo</string-name>
          <email>acaputo@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benjamin Piwowarski</string-name>
          <email>benjamin@bpiwowar.net</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mounia Lalmas</string-name>
          <email>mounia@acm.org</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Bari</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Glasgow</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Yahoo! Research</institution>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The formalism of quantum physics is said to provide a sound basis for building a principled information retrieval framework. Such a framework is based on the notion of information need vector spaces, where events, such as document relevance, correspond to subspaces, and user information needs are represented as weighted sets of vectors. In this paper, we look at possible ways to build, using an algebra de ned over weighted sets, sophisticated query representations and report on corresponding experiments on TREC.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Quantum physics o ers a formalism based on the mathematics of Hilbert spaces,
to describe the behaviour of matter at (sub-)atomic scales. As van Rijsbergen
claims [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the \language" of quantum physics can be used to express the
various geometrical, probabilistic and logical IR models within a uni ed framework.
Driven by this motivation, research in quantum-inspired IR is investigating how
the mathematical framework of quantum physics can address diverse challenges
in IR. These e orts led to the recent proposal of a quantum IR (QIR)
framework [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        In this framework, the system view on the user information need (IN) is a set
of weighted vectors, where each vector represents a possible atomic IN, i.e., an
IN that cannot be further re ned. Computing a query representation from the
terms forming the query is more complex in the QIR framework, than in standard
vector space models. In the latter, the query is represented as the sum of the
vectors representing each query term; in the former, each term corresponds to a
weighted set of vectors (the possible atomic INs corresponding to this term), and
thus it is not straightforward to extend this summation technique. In addition,
many di erent operators (beside an extension of the sum) can be used [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        In the QIR framework, a query is processed in two steps. First, the di erent
aspects of an IN are identi ed, where each aspect address a di erent requirement
on the topic of a document (see Section 2.2). The aspects representations are
computed using two possible operators, namely, the extension of the sum, which
corresponds to a mixture of superpositions, and another operator, mixture. In [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
the best strategy was equalling each term to a di erent aspect. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], the two
operators led to an increase in performance depending on how some or all the
terms forming a query were related to each other; more precisely, whether the
terms together formed a \concept" (e.g. \social network") or whether they were
(more) independent (e.g. \blue car").
      </p>
      <p>
        Based on the ndings in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and that it is possible to freely combine operators
when representing a query (hence, to de ne an algebra with these operators), this
paper explores how complex query representations can be built using an algebra
based on quantum operators. In order to do so, we exploit query segmentation
and computational linguistics techniques to identify relationships between terms
in a query, and extract phrases. We propose a number of strategies, making use
of the terms and phrases in the query, to build a structured query representation.
      </p>
      <p>
        Structured queries have been used in the past in IR and when properly
expressed, retrieval performance is often excellent [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The fact that structured
queries, in general, are not widely used (e.g. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]) stem from users di culties to
express them correctly, although this can be somehow tackled through (pseudo)
relevance feedback. In this paper, we study operators from quantum physics as
a means to express structured queries in the QIR framework.
      </p>
      <p>The outline of this paper is as follows. We rst introduce the QIR framework
(Section 2). Section 3 contains the two paper contributions: We rst de ne the
query algebra (Section 3.1) and then de ne how to transform a query into the
algebra (Section 3.2). These constructions are evaluated in an ad hoc retrieval
scenario (Section 4). Finally, we conclude.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>The quantum IR framework</title>
      <sec id="sec-2-1">
        <title>Information Need Spaces</title>
        <p>
          The basic assumption of the QIR framework [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is that there exists a Hilbert
space H of information needs (IN), called information need space, where each
vector corresponds to what we call an atomic IN. The concept of an \atomic"
IN4 is central to our framework. It can be compared to the notion of \nugget" [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ],
used in summarisation and question-answering to assess the amount of relevant
information a summary or an answer contains.
        </p>
        <p>An atomic user IN is a unit vector ' (aka a state) in H. States determine
statistically any measure made on the user's IN, as for example the relevance of
a document. Given a state space H and a state vector ', a probabilistic event
is represented as a subspace S of H. A state vector ' induces a probability
distribution on events (i.e., the subspaces). The probability of an event is given
by the square of the length of the projection of ' onto the corresponding event
2
subspace S, that is by computing the value Sb' where Sb is the projector onto
the subspace S.</p>
        <p>In the QIR framework, we suppose that we have a system view on the user
IN. We assume that the system usually does not know what is the atomic IN
4 In this paper, we use \atomic" information need (\atomic IN") to distinguish it from
information need (\IN") in its more usual sense in IR.
state of the user, but knows with a probability V (') that the user is in the state
'. It is convenient to view V as a function that associates a weight V (') 2 [0; 1]
to each of its state vectors '. We say that ' 2 V if its weight is greater than
zero { hence V can be thought of as a weighted set of INs. For notations, we use
standard operations on function spaces: Adding two weighted sets V1 + V2, or
scaling by a factor V. We denote ' 7! w the weighted set that associates w to
' and 0 to the other vectors.</p>
        <p>As states are mutually exclusive (a user IN cannot be in two di erent states
at the same time), we require that a weighted set has weights that sum up to 1.
Given the weighted set V of possible state vectors, we then de ne the probability
of an event S as</p>
        <p>Pr (SjV) = X V (') Pr (Sj') = X V (') Sb'
'2V
'2V
2</p>
        <p>
          (1)
2
Note that equation 1 reduces to Sb' if V consists of only one vector ', i.e.
when there is no uncertainty about the user's IN state. This probability can
be computed e ciently when using low rank approximations (see [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] for more
details).
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Aspect Spaces</title>
        <p>
          As discussed in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], user INs may consist of several \aspects" that relevant
documents should address. For example, in the TREC-8 topic 408, \What tropical
storms (hurricanes and typhoons) have caused signi cant property damage and
loss of life?", we can identify two (topical) IN aspects, namely tropical storms
and signi cant damage/loss of life. Each IN aspect can be de ned within an IN
aspect space, where the state vectors are now called atomic IN aspects. Examples
of atomic IN aspects are the vectors representing \hurricane" and \typhoon" for
the rst IN aspect (tropical storms). We use the terminology atomic IN aspect,
since one atomic IN aspect addresses one aspect of the IN (tropical storms) in
the same way that an atomic IN addresses an IN. In this paper, following [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] we
only consider one aspect space, the topical space T , and equal it to the standard
term space where each term corresponds to a dimension.
        </p>
        <p>
          In this topical space, we can represent the event Sd that a document is
relevant to a given weighted set of topical IN aspects as a subspace. We give
here a quick overview of the construction process, which is described in more
details in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The document representation is based on the assumption that a
typical document answers various (atomic) IN aspects. It also assumes that each
document can be split into (possibly overlapping and non-contiguous) fragments,
where each fragment addresses one atomic IN aspect. This follows from research
in focused retrieval, which states that answers to a query, and hence aspects of
it, usually correspond to document fragments (sentences or paragraphs) and not
full documents. In this work, following [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], a document is decomposed into a set
of overlapping fragments of a xed length of 10 terms.
        </p>
        <p>Each fragment can be represented in the topical space T by a vector
corresponding to the atomic IN this fragment is relevant to, which is approximated by
a binary vector where components corresponding to terms in the fragment are
set to 1. The subspace Sd is then de ned as the subspace spanned by the set of
vectors in Vd, i.e. Sd is the smallest subspace such that a document is always a
fully relevant answer to a pure IN aspect it contains, or more formally such that
Pr (Sdj') = 1 for any ' 2 Vd. The document will be relevant to an atomic IN
with a probability that depends on the length of the projection of the atomic IN
vector onto the subspace Sd.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3 Information Need and Aspect Spaces</title>
        <p>The IN space is then de ned as a tensor product of aspect spaces, i.e. of topical
spaces T . Tensor products are a generalisation of the product of probabilistic
spaces for Hilbert spaces. We consider the simple case where the events in each
space are independent from each other, i.e. where aspects are independent.</p>
        <p>If we identify n aspects in a query, each represented by the weighted set of
atomic aspects Vi, then the IN space H is a tensor product of topical spaces
Nin=1 T . In this space H a document is represented by Sd = Nin=1 Sd, where Sd
is the subspace that represents the document in the topical space T (see previous
section). The actual IN is a tensor product of weighted sets V = Nn
i=1 Vi. We
describe several constructions of V based on a query q in Section 3.2 using
the algebra de ned in Section 3.1. Given a document representation Sd and a
weighted set V, we then compute the probability of relevance for the document
d as:</p>
        <p>Pr</p>
        <p>O Si O Vi = Y Pr (SijVi)
i
(2)</p>
        <p>Users might give di erent importance to certain aspects. This motivates the
introduction of a weighting scheme for aspects that would enable us to express
the fact that some aspects are more important than others. In our Hilbert space
of aspects, we chose to introduce a new \don't care" dimension, de ned by the
state vector '&gt;. We then modify each document subspace Sd so that the
subspace contains the \don't care" dimension. This implies that any document is
2
relevant to the \don't care" aspect '&gt;, since Pr (Sdj'&gt;) = Sbd'&gt; = 1; with
respect to Equation 2, the probability of a document to be relevant to Vi is now
Pr (SijVi) = Vi ('&gt;) + (1 Vi ('&gt;)) Pr (SijVi n f'&gt;g), and hence Vi ('&gt;) de nes
how important the ith IN aspect is within the product in Equation 2.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Representing queries</title>
      <p>We formally de ne an algebra over weighted sets of atomic IN aspects (Section3.1).
A query (sequence of terms) q is then transformed into a a structured
representation Vq expressed in this algebra (Section 3.2). The algebra provides the mean
to express relationships between query terms, thus allowing for more complex
query representations.
3.1</p>
      <sec id="sec-3-1">
        <title>An algebra of weighted sets of atomic INs</title>
        <p>We now de ne our algebra over weighted sets, i.e., its constants and operators.
Constants are used to represent the set of atomic INs corresponding to a term
or a phrase. This constant can be computed for each term/phrase at indexing
time from the corpus (Section 3.1.1). We use two operators, each corresponding
to one way to combine atomic INs: mixture (Section 3.1.2), where the possible
atomic aspects are drawn from the weighted sets; and mixture of
superpositions (Section 3.1.3), where the possible atomic aspects are (linear) combinations
of the atomic IN aspects drawn from each weighted set. Other operators could
be de ned (e.g. projection), but we omit them due to the lack of space.
3.1.1 Constants The constants of the algebra correspond to weighted set
of aspects associated with phrases. Let p = t1 : : : tm be a phrase made of m
terms, where m 1. We assume that the documents fragments that possibly
answer a query/phrase p are all the fragments containing this exact phrase.
From Section 2.2 a fragment is mapped to an atomic aspect vector. Thus we can
de ne over all the documents in the collection (1) the number of fragments Np
associated with the phrase p, and (2) the number of occurrences np (') of the
atomic IN aspect ' in the fragments containing the phrase p.</p>
        <p>Each of the Np fragments are associated with a vector, and as we have a
priori no way to distinguish between these di erent vectors, we assume that
each is equally likely to be a atomic IN aspect of a user that typed the phrase
p as a query. Hence, we de ne a rst weighted set V (p) for a phrase p as
V (p) = P ' 7! np (') =Np. In practice, this representation means that the more
vectors ' from V (p) lie in the document subspace, the higher the relevance of
the document to p.</p>
        <p>A query can be composed with more than one phrase; thus to give more or less
importance to a given phrase while composing the weighted sets, we use the fake
state (Section 2.2) with a given weight. This weight will be useful to determine
the contribution of a weighted set when combining it with other weighted sets
in a structured query. Thus, given an aspect Vp, we include the state '&gt; as one
of its possible atomic IN aspect states with a probability f (p). We thus modify
V (p) as follows:</p>
        <p>V (p) = (1
f (p)) V (p) + ('&gt; 7! f (p))
(3)
The second part of the sum assigns the weight f (p) to the don't care atomic
IN aspect, whereas the rst part scales the previous weights so that the sum
of all weights is 1 again. Equations 3 and 1 imply that the probability of a
document to be relevant for the aspect de ned by the phrase p is Pr (SdjV (p)) =
f (p) + (1 f (p)) Pr (SdjV (p)).
3.1.2 Mixture Mixture (by analogy with a probabilistic mixture) is the
simplest operator combining n weighted sets of IN aspects V1 : : : Vn. The
underlying assumption in our context is that the atomic IN aspect can be an
aspect coming from any of the Vi. Unless there is a special reason to believe that
the user atomic aspect is more likely to be from one than the other weighted
sets, we can assume that each weighted set have an equal probability of being
the one (that is, a probability of n1 ). We de ne formally the mixture operator
}in=1Vi of all the atomic IN aspect vectors associated with the weighted sets
V1 : : : Vn:
n
n 1 X
}i=1Vi = n</p>
        <p>Vi
i=1</p>
        <p>The mixture can be used when each IN aspect (weighted set) is an equally
important alternative to describe the user IN aspect.
3.1.3 Mixture of superpositions In standard IR, a query is represented by
a vector corresponding to a linear combination of the vectors associated with the
query terms. In quantum physics, a normalised linear combination corresponds
to the principle of superposition (normalised linear combination), where the
description of a system state can be superposed to describe a new system state.</p>
        <p>In our case, a system state is a user atomic IN aspect, and we use the
superposition principle to build new atomic IN aspects from existing ones. For
example, let 'p, 'uk and 'usa be three pairwise orthogonal vectors in a
threedimensional IN space which, respectively, represent the atomic IN aspects \I want
a pizza", \I want it to be delivered in Cambridge (UK)" and \I want it to be
delivered in Cambridge (USA)". The atomic IN aspect vector \I want a pizza to be
delivered in Cambridge (UK)" would be represented by a superposition 'p=uk of
'p and 'uk. We can similarly build the atomic IN aspect for Cambridge (USA).
To represent the ambiguous query \pizza delivery in Cambridge" where we do
not know whether Cambridge is in the USA or the UK, and assuming there is no
other source of ambiguity, we would use a mixture of the two possible superposed
atomic IN aspects. This brings us to another variant of combination of atomic
INs, the mixture of superpositions, which forms our second algebra operator.</p>
        <p>Each aspect vector weighted set resulting from the mixture of superpositions
is constructed as follows. We rst randomly select a vector from each set Vi. In
our previous example, the set V1 would be just one vector ('p), whereas V2 would
contain two vectors ('uk and 'usa). We then superpose the selected vectors (one
for each term), and obtain a new vector '. The above process is repeated for all
the possible selections of vectors. The associated weighted set is thus formally
de ned using the mixture of superpositions operator:
(4)
n
si=1Vi =</p>
        <p>X : : : X
'12V1
'n2Vn</p>
        <p>P 'i
kP 'ik 7!</p>
        <p>Y Vi ('i)
i
!
(5)
In the weighted set, each atomic IN aspect is a linear combination Pn
i=1 'i of
the IN aspect from each of the terms composing the query which is normalised
to ensure it is a unit vector. Each of these linear combinations is associated with
a weight that is the product of the weights5 Qi Vi ('i).</p>
        <p>The mixture of superpositions di ers from the mixture in the way we combine
weighted sets. As an example, consider two terms t1 and t2 that always occur in
a document, but half of the time in the same fragments, the other half in distinct
fragments. A mixture V (t1) } V (t2) retrieves all the documents containing the
terms, but a mixture of superpositions V (t1) sV (t2) gives a higher probability
to the documents that contain both in the same fragments. The mixture of
superpositions is better suited for terms whose meaning should be combined, i.e.
forming phrases. Terms co-occurrence could, for instance, be used to determine
whether two terms should be combined to form a phrase, e.g. \social network".
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>From queries to structured queries</title>
        <p>In the previous section, we de ned an algebra, i.e. its constants and operators,
over weighted sets of IN aspects. In our QIR framework, a weighted set of IN
aspects V is used to compute to which extent a document is relevant to the
IN aspects in V. With this algebra, we can express more sophisticated, aka
structured, query representations, where the relationships between query terms
are accounted for. In this section, we describe how a query q (a set of terms) is
mapped to a structured query expressed as a weighted set V (q) in our algebra.</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], two main strategies were applied to construct the (non-structured)
query representation: (1) the query had only one aspect and the weighted sets of
each query term were combined using a mixture } or a mixture of superposition
s (2) each query term corresponded to an aspect or a tensor product (of n
aspects). The latter strategy has been shown to perform best for a variety of
TREC collections. Both strategies were, however, simple translations into our
algebra.
        </p>
        <p>
          Building a structured query involves two steps. First, we must indentify the
di erent IN aspects, i.e. compute how many aspects the query has, and which
query terms are linked to each. Second, we must characterise the relationships
among terms within each aspect. We therefore make use of query analysers. The
use of such tools is an active topic in IR, e.g. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. In this paper, the following
tools are employed:
        </p>
        <p>
          Chunkers Tools that splits a query into a series of chunks corresponding to
the query \concepts". In this paper, we experimented with Two-stage query
Segmentation (TSQS) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], a machine learning-based chunker, and META (Meta), a
Natural Language Processing tool [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ];
        </p>
        <p>
          Dependency parser A more powerful query analyser tool is MaltParser
(Malt ) [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], a dependency parser that builds a term graph based on grammatical
relationships. Particular to MaltParser is that term dependence can be between
non-adjacent terms. For example, in \the cat saw the black mouse", \saw" will be
connected to \cat" (subject) and \mouse" (object), and \mouse" to \black"
(modi er) and \the" (determinant).
5 In practice, we used a modi ed formula for computational reasons (see [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]).
        </p>
        <p>These tools were used to build a structured query representation V (q). A
number of strategies have been experimented with; we associate with each an
identi er (in bold between parenthesis), and a formula V (q) representing
the query q, expressed within our algebra, i.e. phrase constants V (p), mixture
}, mixture of superpositions s, and nally tensor product .</p>
        <p>Two-term queries We wanted to compare all possible uses of operators. Thus to
limit the number of possibilities, we restricted ourselves to queries composed of
only two terms, t1 and t26. We started with four such strategies (names pre xed
by 2t) three using our operators: tensor (2tT) V (q) = V (t1) V (t2), mixture
of superpositions (2tS) V (q) = V (t1) sV (t2), mixture (2tM) V (q) = V (t1) }
V (t2) and one considering the two terms as forming a phrase (2tP) V (q) =
V (t1t2). For the latter, we observed that in 30% of the cases, the phrase t1t2 does
not appear in the document collection, resulting in an empty weighted set. We
hence performed two additional experiments on two-term queries where V (t1t2)
is combined with either a tensor product (2tP-T) V (q) = V (t1t2) V (t1) V (t2)
or a mixture (2tP-M) V (q) = V (t1t2) }V (t1) } V (t2). In this way, V (t1t2) has
no e ect if the phrase t1t2 does not exist in the collection.</p>
        <p>
          Reference For all experiments as a reference, we use the representation where
each term corresponds to a di erent aspect, which was shown to be the most
successful in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The corresponding query representation is a tensor product
of weighted sets, one for each query term t. We have two such references, one
being speci c to TSQS as the terms used by TSQS to construct the query
representation is di erent to those used by all other strategies 7 (TSQS-R, R)
V (q) = Nt2q V (t).
        </p>
        <p>Chunks We used both TSQS and META to decompose queries into chunks.
These are sets of terms that can be considered as a whole, and usually de ne
a concept. Hence, we decided to map each identi ed chunk using a mixture of
superpositions, where each chunk is considered as an aspect of the topic. Thus
each aspect is represented as a mixture of superpositions of the terms in the
chunk. Denoting C the set of identi ed chunks, where each chunk is a set of
terms, we have (TSQS-S, Meta-S) V (q) = Nc2C st2cV (t).</p>
        <p>Dependency Parsing We used MaltParser to extract term relationships in
queries. In particular, we were interested by the \modi er" relationship, which
\describes a noun phrase or restricts its meaning in some way"8. For example, in
the TREC topic 357, \... Identify documents discussing international boundary
6 This meant, 89 out of the 250 topics from the TREC Robust track (the detail of the
test collection used in our experiments is given in Section 4).
7 TSQS makes use of all query elds (title, description and narrative), in addiction
to some external sources, to segment the original query into chunks. This results in
a di erent bag-of-word and justify the computation of a separate reference for this
strategy.
8 From the Oxford dictionary.
disputes...", international and boundary are modi ers of disputes. We denote
Hq the head terms of the query q, i.e. the terms that are not modi ers. In our
example, disputes is a head term. For each head term t, we follow the \modi ed
by" links of the dependency graph generated by MaltParser, and construct a
set Ph of all the possible maximal sequences t1 : : : tnh, i.e. those where t1 does
not have any \modi ed by" link associated to it. In our example, the phrases
\international disputes" and \boundary disputes" would be part of Ph.</p>
        <p>Obviously (as also observed with 2tP) not all phrases in Ph may occur in
the document collection. For such phrases, the corresponding weighted set will
be empty. We therefore adopt two alternative strategies. In the rst, an aspect
corresponds to a mixture of all the possible subsequences of phrases constructed
by following the \modi ed by" link. Formally (PM):</p>
        <p>(V(t1t2:::tnh) } V(t2:::tnh) } ::: } V(h))
V(q) = O</p>
        <p>O
t2Hq</p>
        <p>h2Hq
t1:::tnh2Ph
t2Hq</p>
        <p>h2Hq
t1:::tnh2Ph
The second strategy is to consider that there are as many aspects as subsequences
following the \modi ed by" link (PT):</p>
        <p>V(q) = O</p>
        <p>O</p>
        <p>V(t1t2:::tnh)O V(t2:::tnh) O ::: O V(h)</p>
        <p>Following the same rationale as for 2tP, we decided to perform an additional
experiment in which the sequences of phrases are combined with the tensor
product of all the query terms. In that case, a term that is a modi er can have
the same in uence on the nal query representation. For example, the query \A
B" (where B is modi ed by A) becomes VAB VA VB. Thus, a query q is
represented by (PT2):</p>
        <p>V(q) = OV(t)
t2q</p>
        <p>O
h2Hq
t1:::tnh2Ph</p>
        <p>V(t1t2:::tnh)O V(t2:::tnh) O ::: O V(tnh)</p>
        <p>Next, we experiment all these structured query constructions. While not
exhaustive, they provide a good starting point to study sophisticated query
representations in our QIR framework.
4
4.1</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <sec id="sec-4-1">
        <title>Experimental Setup</title>
        <p>
          We used the TREC ROBUST 2004 collection (topics 301-450, 601-700). All
topics were automatically converted into a query using their \title" part, with the
exception of the TSQS strategy (see Footnote 7). We experimented with the
various strategies de ned in Section 3.2. To de ne the weight f (p) of a phrase p
(a sequence of terms), following [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], we used a normalised inverse document
frequency. To create the document subspace and compute each term representation,
we de ne each fragment as a window of 10 words, to which we apply stemming
and remove stop-words before using a binary weighting scheme for the vector
representing the topical aspect covered by the fragment. We then follow [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] for
the computation of the document and query representation.
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Results</title>
        <p>Our performance results are reported using Mean Average Precision (MAP).
Results for our three sets of experiments, two-term queries, chunks, and
dependency parsing, are shown in Tables 1a, 1b, and 1c, respectively. We also show, for
comparison, results using our Reference (TSQS-R when comparing to TSQS-S,
and R when comparing with all other strategies), and the classic baseline BM25.</p>
        <p>
          Overall, none of the results (except for 2tP-T) have shown an improvement
over BM25. However, as already indicated in [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], although using mixture/mixture
of superpositions to construct query representations do not increase overall
performance, they increase the performance of speci c types of topics. We return
to this throughout this section.
        </p>
        <p>Generally, all strategies follow similar trends across all three sets of
experiments. The tensor product (2tT, TSQS-T, R) outperforms the mixture of
superpositions (2tS, TSQS-S, Meta-S). But again, when looking on a per topic
basis, in the chunks set of experiments for instance, we saw that, for 25% of the
queries, the mixture of superpositions gives the best result9.</p>
        <p>We now discuss results with respect to the use of phrases in constructing
query representations using our algebra. For two-term queries, the strategy using
the tensor product (2tT) led to better performance than that using phrases made
of the simple concatenation of the two terms (2tP). Furthermore, even when
using a query analyser tool (MaltParser) to indentify phrases (here restricted
to head and modi ers), the tensor product of all query terms (R) led to better
performance than the tensor product of weighted sets, one for each aspect, where
an aspect corresponds to a mixture of all phrases (the set Ph) associated with
the head term h (PM); this is even the case when the tensor product is used to
compute the weighted set for each query aspect, where each aspect corresponds
to a phrase in Ph (PT).</p>
        <p>Only two strategies making use of phrases (2tP-T, PT2) led to higher
performance compared to the tensor product only strategy (no phrases) was used
(2tT, R). Both 2tP-T and PT2 combine phrases with the tensor product of all
query terms. This would indicate that using the tensor product of the query
terms is important in representing structured queries.</p>
        <p>We performed an in-depth manual analysis of the topics used in our
twoterm queries experiments. Tensor product works well when the query expresses
distinct aspects that should be contained in the same document (e.g. \poppy
cultivation,"\L-tryptophan deaths"). Using the mixture of superpositions increases
9 For space constraint, we do not report performance results per topic.
e ectiveness for queries containing terms forming a concept (e.g., \space
station", \computer viruses", \radioactive waste"). Finally, although using phrases
(2tP, 2tP-M, 2tP-T) may allow capturing the narrower meaning coming from the
modi er (e.g. \home schooling", \marine vegetation"), only the combined use of
the tensor product on both the phrase and the two query terms (2tP-T) results
in e ective performance. It is not fully clear whether this comes from and only
the fact that many phrases, as such, do not appear in document fragments, or
whether their representation is not optimal (i.e., the right weighted set is built
due to approximations issues).</p>
        <p>Overall our results indicate that retrieval performance depends strongly on
how the relationships between query terms are represented within our algebra
(i.e. which operators are used and how). In future work, we will look at
automatic means to assign construction strategies to topics, depending on their type.
We however performed two upper-bound experiments, where we chose the best
performing strategy for each topic, and calculate MAP over these best
performing strategies. The rst experiment was conducted over the two-term queries
strategies (2tM, 2tS, 2tT, 2tP, 2tP-M and 2tP-T), and the resulting MAP value
is 0.2801, which corresponds to an increase of +13% over BM25 (0.2552). The
second experiment considered R, TSQS-S, TSQS-T, Meta-S, PM, PT, and PT2,
leading to a MAP value of 0.2726, which corresponds to an increase of +12.59
over BM25 (0.2421).</p>
        <p>To summarise, using phrases and the tensor product operator seems to be
the most e ective strategy (2tP-T, PT2). However, the best results are achieved
with short queries (2tP-T outperforms BM25 whereas PT2 although better than
R is outperformed by BM25). In future work, we intent to investigate whether
this comes from the tools used to analyse the queries (e.g. we saw that
MaltParser failed to identify a large number of head-modi er relationships), or their
representations using our algebra. As shown in our upper-bound experiments,
there is much room for improvement with respect to the latter.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>
        This paper describes a query algebra de ned within a quantum-inspired approach
to IR[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], where user information needs are represented as weighted sets of vectors.
The algebra, de ned over weighted sets, is used to construct more sophisticated,
aka structured, query representations, where relationships between terms are
accounted for through the consideration of phrases. We proposed several strategies
to transform a query (set of terms) into a structured query (set of terms and
phrases), using the operators from this algebra. Some of the strategies make use
of query analyser tools to extract proper phrases.
      </p>
      <p>Our experimental results indicate that each speci c operator improves
performance for a speci c type of queries (tensor for multi-aspect queries, mixture
of superpositions for concepts, and phrase operator for queries where a narrower
meaning is given by a modi er). Future work includes looking at automatic
means to assign operators (and strategies) to queries, in particular for
representing phrases. Other query analyser tools will be employed, as properly extracting
phrases is important to build e ective structured queries using our algebra.
Acknowledgements This research was supported by an Engineering and Physical Sciences Research
Council grant (Grant Number EP/F015984/2). This paper was written when Mounia Lalmas was a
Microsoft Research/RAEng Research Professor and based on the work done by Annalina Caputo as
an intern, both at the University of Glasgow.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1. van Rijsbergen,
          <string-name>
            <surname>C.J.</surname>
          </string-name>
          :
          <source>The Geometry of Information Retrieval</source>
          . Cambridge University Press, New York, NY, USA (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Piwowarski</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frommholz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lalmas</surname>
          </string-name>
          , M.,
          <string-name>
            <surname>van Rijsbergen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>What can quantum theory bring to IR? In: CIKM, ACM (</article-title>
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Piwowarski</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frommholz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lalmas</surname>
          </string-name>
          , M.,
          <string-name>
            <surname>van Rijsbergen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Exploring a multidimensional representation of documents and queries</article-title>
          . In: RIAO. (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Metzler</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Croft</surname>
          </string-name>
          , W.B.:
          <article-title>Combining the language model and inference network approaches to retrieval</article-title>
          .
          <source>IPM</source>
          <volume>40</volume>
          (
          <issue>5</issue>
          ) (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Spink</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolfram</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jansen</surname>
            ,
            <given-names>M.B.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saracevic</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Searching the web: The public and their queries</article-title>
          .
          <source>JASIST</source>
          <volume>52</volume>
          (
          <issue>3</issue>
          ) (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Clarke</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kolla</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cormack</surname>
            ,
            <given-names>G.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vechtomova</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ashkan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , Buttcher, S., MacKinnon, I.:
          <article-title>Novelty and diversity in information retrieval evaluation</article-title>
          .
          <source>In: Proceedings of the 31st ACM SIGIR</source>
          , Singapore,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>July 2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Croft</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bendersky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
          </string-name>
          , G., eds.
          <source>: SIGIR 2010 Workshop on Query Representation and Understanding</source>
          . (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Bendersky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Croft</surname>
            ,
            <given-names>W.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>D.A.</given-names>
          </string-name>
          :
          <article-title>Two-stage query segmentation for information retrieval</article-title>
          .
          <source>In: Proc. of the 32nd ACM SIGIR</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Basile</surname>
            , P., de Gemmis,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gentile</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iaquinta</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lops</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Semeraro</surname>
          </string-name>
          , G.:
          <article-title>METAMultilanguagE Text Analyzer</article-title>
          .
          <source>In: Proc. of LangTech</source>
          . (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Nivre</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hall</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nilsson</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanev</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eryigit</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , Kubler,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Marinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Marsi</surname>
          </string-name>
          , E.:
          <article-title>MaltParser: A language-independent system for data-driven dependency parsing</article-title>
          .
          <source>Natural Language Engineering</source>
          <volume>13</volume>
          (
          <issue>02</issue>
          ) (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>