<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>XY-PHOC Symbol Location Embeddings for Math Formula Retrieval and Autocompletion</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Robin Avenoso</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Behrooz Mansouri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Zanibbi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Rochester Institute of Technology</institution>
          ,
          <addr-line>1 Lomb Memorial Drive, Rochester, NY, USA 14623</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We describe the participation of the XY-PHOC team from The Document and Pattern Recognition Lab (DPRL) from the Rochester Institute of Technology (RIT, USA) in ARQMath 2021 Task 2 (Formula Retrieval). We generalize a one dimensional spatial encoding for word spotting in handwritten document images, the Pyramidal Histogram of Characters or PHOC, to obtain the two-dimensional XY-PHOC, which provides robust spatial embeddings of symbols with modest storage requirements. XY-PHOC symbol location embeddings capture the relative position of symbols without the need to generate or store explicit edges between symbols. For ARQMath 2021, the XY-PHOC model obtains competitive results in formula similarity search. We also present new results using XY-PHOC for the related task of formula autocompletion from visual queries, in which target formula symbols may be added to the query in an any order.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Character Embeddings</kwd>
        <kwd>Spatial Embeddings</kwd>
        <kwd>Formula Retrieval</kwd>
        <kwd>Formula Autocomplete</kwd>
        <kwd>Mathematical Information Retrieval (MIR)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The XY-PHOC team from The Document and Pattern Recognition Lab (DPRL) from the Rochester
Institute of Technology (RIT, USA) participated in ARQMath 2021 Task 2. In this task, the
participants are given a formula from questions in Task 1, and are asked to return the top-1000
relevant formulas for each topic [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. These retrieved formulas are from the ARQMath Math
Stack Exchange corpus, containing posted answers and questions from 2010-2018. Formula
topic queries are then taken from posts in 2019 (ARQMath 2020) and 2020 (ARQMath 2021).
      </p>
      <p>
        We provided one run for Task 2, using the XY-PHOC system [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. XY-PHOC uses a very simple
relative spatial encoding to capture the locations of symbols in a formula at diferent levels of
granularity. Embeddings for the location of individual symbols are stored in an inverted index,
which can then be compared with symbol location embeddings for the query formula using
cosine similarity.
      </p>
      <p>
        Current state-of-the-art systems for formula similarity search rely on graph-structured data,
such as in Tangent-S [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and Tangent-CFT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. These systems require special indexing strategies
in order to build inverted indexes on non-traditional keys. The scoring of these systems also
requires several operations (e.g., a subtree alignment step in Tangent-S) or trained embedding
models for tree paths (e.g., in Tangent-CFT).
      </p>
      <sec id="sec-1-1">
        <title>Level 2</title>
      </sec>
      <sec id="sec-1-2">
        <title>Level 3</title>
      </sec>
      <sec id="sec-1-3">
        <title>Level 4</title>
      </sec>
      <sec id="sec-1-4">
        <title>Level 5</title>
      </sec>
      <sec id="sec-1-5">
        <title>Level 1</title>
        <p>{2, 8, =, x, y, √}
{2, 8, =, y} {=, x, √}
{2, 8, y} {=} {x, √}</p>
      </sec>
      <sec id="sec-1-6">
        <title>Level 2'</title>
      </sec>
      <sec id="sec-1-7">
        <title>Level 3'</title>
        <p>{2, 8, y} {8, =} {=, √} {x, √} Level 4'
{2, y} {8, y} {=} {√} {x, √}</p>
        <p>Level 5'
{8}
{2, =, x, y, √}
{8}
{2, =, x, √}
{y}
{{}8}
{{}2, =, x, y, √}
{}
{8}
{2, =, √}
{{}x, y}</p>
        <p>The XY-PHOC representation requires less storage than does a graph-based model: in the
XY-PHOC symbol location embeddings index keys are symbols, and postings consist of 29-bit
binary vectors stored in 32-bit words, accompanied by an integer identifier for the associated
formula (Section 2 provides details). Storing standard types in the index will allow for standard
information retrieval tools and techniques to be utilized. The representation for XY-PHOC can
be put into a standard inverted index used for document retrieval with only small modifications
for scoring. Scoring matches for XY-PHOC embeddings is performed with an optimized
rankequivalent cosine similarity.</p>
        <p>Interestingly, the XY-PHOC retrieval model was designed with formula autocompletion
rather than formula retrieval in mind, as initially we believed that formula search using symbol
positions alone would be over-constraining and produce many false negatives/misses. The
ARQMath results obtained indicate that this is not the case, possibly because the presence of a
symbol anywhere in a formula is represented. The XY-PHOC symbol location retrieval model
can be applied as an autocompletion model or a similarity search model simply by switching
from conjunctive queries (requiring all query symbols) to disjunctive queries (requiring at least
one query symbol).</p>
        <p>Currently there are only two parameters in our model, controlling the number of regions
and the function used to identify which regions a symbol belongs to. Scoring is performed
using cosine similarity over binary vectors defined for individual symbols. Indeed, our symbol
location vectors may be understood as a form of ‘term location frequency’ vectors. Our retrieval
model is nearly identical to the TF portion of a standard TF-IDF retrieval model over words, but
where the presence of symbols in multiple overlapping regions are captured, rather than the
number of symbol occurrences.</p>
        <p>[1101110[011110111101000011111111]0001111]
[1011100[111011110100001111111100]0011110]
[1011100[111011110100001101111100]0010110]
(a) Formula Query</p>
        <p>(b) XY-PHOC Symbol Location Embeddings</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. XY-PHOC: Bidirectional Pyramidal Histogram of Characters</title>
      <p>
        In this section we describe the XY-PHOC model, which generalizes Pyramidal Histogram of
Character (PHOC) bit vectors representing character locations in words [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to capture horizontal
and vertical symbol positions within formulas. An illustration of XY-PHOC symbol locations
and their vector representation can be found in Figure 1, and Figure 2 illustrates how the bit
vector per symbol looks using the first four levels.
      </p>
      <p>
        The Pyramidal Histogram of Characters (PHOC) embedding model generalized for XY-PHOC
comes from the field of word spotting [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In word spotting, the goal is to retrieve word images
matching a query word given as either a character string (e.g., in UTF-8) or raster image. PHOC
is a binary vector indicating the presence of a character within each horizontal region. For each
level  there are  regions beginning with one word-level region (level 1), left and right regions
after splitting the formula in half (level 2), followed by additional identically-sized splits up to 
regions. The presence of a symbol in region(s) at each level gives the represention of locations
where the symbol appears. After region embeddings are concatenated, a total of ( + 1) − 1
regions are represented. So for example, using up to 5 uniform splits in the horizontal and
vertical directions, we obtain 5 · 6 − 1 = 29 regions, which will be represented as a 29 bit vector.
      </p>
      <p>This embedding was inspired by how the redundant spatial information in the character
embeddings make it suitable for capturing both coarse and precise locations for symbols in
space, making it useful for visual formula autocompletion. As Level 1 captures all symbols
present in a formula, this along with the redundant spatial information are well suited to finding
formulas similar to a query, starting from the symbols they share.</p>
      <p>The original PHOC embedding is very sparse, as it uses a vector the length of the alphabet per
region to indicate which symbols are in that region. For math the symbol vocabulary is much
larger than the Latin alphabet, making the vector even longer. To use XY-PHOC eficiently, we
use an inverted index over symbols with each posting as a pair (, ) containing a formula
identifier and a bit vector representing the XY-PHOC regions where that symbol appears.</p>
      <p>
        For spatial regions, we use five levels in both the horizontal (X) and vertical (Y) directions,
making a symbol-specific XY-PHOC vector 29 bits long, which we store in a 32-bit integer.
Formulas tend to be much wider than tall; we found in experiments that we could better
distinguish the vertical location of symbols using a single point, and that using a line to
represent the horizontal extent of symbols provided helpful redundancy [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>An example of a simple 4-level embedding is shown in Figure 2. When encoding the binary
XY-PHOC vectors, as seen in Figure 1, symbols are represented by a horizontal line that spans
the width of the symbol’s bounding box, with the line positioned at the vertical center of the
bounding box. If any point on this line is included in a horizontal region, the bit for that region
is set to 1. Vertical region memberships are identified by the vertical center of the box (a single
point). This encoding represents the presence of the symbol in each region, and if a symbol
appears in a region more than once it is still represented by a 1. For each horizontal level 
greater than 1, there will be a level ′ which represents the vertical splits at that level.</p>
      <p>We index each formula identified as visually distinct in the ARQMath corpus once using
a single instance (e.g., aiming to have just a single entry for 2). XY-PHOC symbol location
vectors are computed starting from formulas given in LATEX using the MathJax javascript library1.
Once formulas have been converted to SVG images, we convert these to lists of symbols with
bounding boxes, which are then used to compute our XY-PHOC symbol embeddings. In order
to index the large ARQMath collection, several tools are employed. We perform indexing using
Apache Spark2, using a distributed map-reduce implementation that ultimately produces the
index in a text file. The index is then loaded into a Redis 3 database for use in retrieval. A second
index maps formula ids to their original file, the normalization constant for each XY-PHOC
vector as described in Section 2.1, and the number of symbols in the formula.</p>
      <p>An important advantage of the reduced representation of XY-PHOC is that standard
information retrieval techniques and tools can be used to generate an eficient and robust system.
Other math formula retrieval systems that work on graph representations need to use custom
solutions in order to index the paths that make up formula graphs, whereas the XY-PHOC
embedding easily fits into standard tools for text-based search engines.
2.1. Scoring using XY-PHOC Symbol Location Vectors
We are able to retrieve and score formulas in a manner similar to conventional TF-IDF retrieval
over words - each query symbol is looked up in the index, XY-PHOC vectors from formulas
containing the symbol are retrieved, and then compared to the query symbol XY-PHOC vector.
Symbol location match counts are then accumulated across query symbols, and finally scaled
by candidate formula size, giving preference to smaller formulas when two or more candidate
formulas have the same number of matching symbol locations.</p>
      <p>More concretely, to score candidate formulas, we follow the work of Sudholt et al. [8] who
demonstrated that in the word-spotting context, cosine similarity works well for scoring words
represented as one-dimensional PHOCs.</p>
      <p>For query vector a and candidate formula vector b the cosine similarity is:
 =</p>
      <p>· 
|||| |||| = √︁∑︀=1 2 √︁∑︀=1 2
∑︀=1 
.</p>
      <p>(1)</p>
      <sec id="sec-2-1">
        <title>A faster rank-equivalent similarity metric  is defined as:</title>
      </sec>
      <sec id="sec-2-2">
        <title>1https://www.mathjax.org 2https://spark.apache.org 3https://redis.io</title>
        <p>(a, b) = (a, b) = √︀|1b|1 |a ∧ b|1</p>
        <p>The dot product of two binary vectors is the Hamming weight (number of 1s) in the logical
AND of the vectors, which is equivalent to the 1 norm (||1). The normalization factor for
query a is constant across candidate formulas, and so can be removed without altering the rank
ordering. To accelerate computation, the normalization factor for b is pre-computed and stored
in the formula (secondary) index for lookup at retrieval time.</p>
        <p>
          Our first retrieval implementation is very simple. We perform term-at-a-time scoring, and
do not make use of skip lists, score bounds (e.g., from MaxScore [9]) or alternative indexing
strategies (e.g., block-max [10]). For ARQMath-1 (2020) topics, this resulted in an average query
time of 566 seconds (i.e., roughly 9.5 minutes) per query using a Python implementation running
on a desktop Linux system with an Intel i7-8700K CPU (3.7GHz) and 32GB RAM, using Redis
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. For a rapid prototype this was workable, but in the future, we hope to produce a much
faster implementation.
(2)
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Formula Retrieval Results</title>
      <p>
        In this section we will present the results for the XY-PHOC retrieval model on the ARQMath
formula retrieval task [11]. We show results for both queries from ARQMath 2020 and ARQMath
2021. Note that our model has only two parameters (number of regions and the symbol region
membership function), and only a small amount of training/fitting on the ARQMath 2020
collection was performed [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>For formula similarity search, we use disjunctive queries (weak AND) over symbol location
vectors. This requires that at least one symbol present in the query is present in a candidate
for retrieval. This assumes that relevant candidates share at least one symbol with the query
formula.</p>
      <p>
        There are 45 ARQMath 2020 topics. For ARQMath-2, the top results from all teams are shown
in Table 1. Our model obtained the third-highest nDCG′ score. In the table we see that XY-PHOC
has a higher P′@10 than the baseline system and is within 1.7% in MAP′ and 8% in nDCG′
scores relative to the baseline system, Tangent-S [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Tangent-S is a much more complex model,
making use of path-based retrieval on both Symbol Layout Trees (SLTs) and Operator Trees
(OPTs), followed by aligning the query formula with candidates in both representations before
producing a final score. The best-performing system (LtRall) is obtained by re-ranking
TangentS results after including additional OPT and SLT tree embeddings and tree edit distances within
a learning-to-rank framework, adding complexity and computational expense.
      </p>
      <p>ARQMath 2021 has 60 topics. The top results from all teams is shown in Table 1. Our model
obtained the third-highest nDCG′, MAP′, and P′@10. It is interesting to see both the Tangent-S
system and LtRall systems performed less well then the XY-PHOC model, while two systems
that scored lower than XY-PHOC for ARQMath 2020 topics (MathDowsers and approach0)
obtained the highest metrics for ARQMath 2021 topics. Across the ARQMath 2020 and ARQMath
2021 topics, the performance of XY-PHOC model remains relatively stable. Among the top-3
runs for ARQMath 2021, all the systems had very similar scores, with the nDCG′ having a
0.488
0.450
0.433
0.419
0.317
0.197
0.217
 = √− 1
± == =±√√√−−1.1
 = √− −1,1
diference of 0.7%, MAP′ having a diference of 3.8% and P′@10 having a diference of 5.5%. It
is interesting that all three of these models use one formula representation for retrieval: OPT
for approach0, SLT for MathDowsers, and XY-PHOC for our system. It would be interesting
to see how diferent retrieved formulas are between these systems, and whether an ensemble
would produce stronger results.</p>
      <p>To understand the types of queries the XY-PHOC system performs well and poorly on,
we present some queries with their top-5 retrieval results from the 2020 ARQMath queries,
after removing the formulas that are not assessed and so not used in computing the prime (′)
evaluation metrics. In Table 2 topics B.27 and B.29 are shown. For both of these queries, the
top-5 results have an nDCG′@5 of 1. The strong performance is due to the types of formulas
marked with a relevance of High (3) - formulas with the same symbols, and roughly the same
placement as the queries. For both queries, the top result is the query formula. Formulas in
ranks 2-5 also contain the query symbols, with only slight changes in placement or additional
punctuation such as a ‘.’ or a ‘,’.</p>
      <p>Next we consider two queries that the system did not perform well on. XY-PHOC’s preference
for candidates with similar structure and symbols can be seen in the results for topic B.2 and
B.32 presented in Table 3. Both Topics have an nDCG′@5 of 0, meaning no formulas in the
collection annotated as relevant were returned in the top 5 after removing unassessed formulas.
[¬∀ ∈  (()] ⇐⇒ ∃ ∈  (¬()).
∀∃(() ∨ ()) ⇐⇒ ∃∀(() ∨ ())
 (Q)  ⇐⇒ ∃ ∈ Q︀( ,  ∈ )︀
 ∈ () ⇐⇒ () ⊆ ()
∀∃(, ) ⇐⇒ ∃∀∃(, , )
For topic B.32, after removing unrated formulas only 4 formulas remained in the results. To get
a better understanding of the the formulas returned, we present results for these topics without
removing unrated formulas in Table 4. Based solely on shared symbols and structure, it would
appear these filtered formulas may be more similar to the query, and it is possible that results
would be stronger if these formulas were annotated.</p>
      <p>Having shown the extremes for topics that obtained nDCG′ scores of 1.0 and 0, we now
present two query topics with an nDCG′@5 &gt; 0.7 . These results are presented in Table 5 using
topics B.12 and B.14. For both of these topics, the top retrieval result matches the query formula,
with a relevance of High (3). The relevance scores then decrease with rank. In these two queries,
relevant results are retrieved at a high rank based on the symbols locations.</p>
      <p>As originally hoped, the XY-PHOC model generally works well for exact item retrieval, as
often the embedding with the highest score will be the query formula. We apply our model to
the problem of formula autocompletion in the next Section.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Formula Autocompletion Results</title>
      <p>
        In this section we briefly present the results from formula autocompletion experiments reported
by Avenoso [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The XY-PHOC system was originally designed with the goal of autocompletion
in mind, and is the first math formula autocomplete system that allows symbols to be input in
any order. For example, a query such as shown in Figure 2 can be used to retrieve formulas
containing many diferent integrands between the integral and ‘dx.’ Further, the query symbols
can be either entered, or selected from an existing formula in any order. Query symbols do not
need to comprise a well-formed subexpression (e.g., an SLT), or to be entered as a text encoding
from left-to-right (e.g., for LATEX). Matching is purely spatial.
      </p>
      <p>For autocompletion, we make some small but important changes from our formula similarity
search: (1) we use conjunctive queries, requiring all symbols in the query to be present in all
candidates, and (2) the number of symbols in each candidate formula must be no smaller than
that for the query formula. This reflects that our autocompletion is intended to exactly match a
portion of the target formula, and so formulas that do not meet these requirements are pruned
during retrieval.</p>
      <p>
        To study the efect of inputting symbols in diferent orders on XY-PHOC autocompletion
performance, we tested four possible symbol input orders: Left-to-Right (a roughly ‘standard’
entry order), Right-to-Left, Outside-in alternating adding symbols from the left and right ends,
and Middle-out starting at the middle symbol and alternating between adding a symbol on
the left and right side of the query. The formulas used for autocompletion queries were taken
from the set of visually distinct ARQMath formulas that have been assessed [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Results for
the first 102 queries are shown in this section. In Figure 3 we show the Mean Reciprocal Rank
(MRR) grouped by the percentage of symbols input from the target formula, using groupings
of 10 percent. MRR reflects how close target formulas are to the top rank. In an autocomplete
system we generally want our target formula to obtain as high a rank as possible using as few
symbols as possible, so that the user can easily find it. Note that an MRR of 50% reflects that the
harmonic mean of the rank for the 102 query formulas is rank 2, 25% is the same for rank 4, etc.
      </p>
      <p>Looking at Figure 3, we see that the inside-outside order obtains substantially higher MRR
scores throughout the input size range. Interestingly, the left-right order obtains the lowest
score across the input size range, which might be caused by formulas being less unique at their
left end (particularly given the stronger performance for entering symbols right-left), but we
have not had time to confirm this. The improved performance for outside-in also reflects that
anchoring the left and right ends of the XY-PHOC symbol location vectors after the first two
symbols leads to more stable location vectors.</p>
      <p>In terms of how these diferent input orders for formula autocompletion might be used, we
see two possibilities. One is users placing symbols on a canvas, e.g., by recognizing a formula
image, using handwriting, or manually entering/placing symbols. Another is interactively
selecting symbols in displayed formulas to define queries such as shown in Figure 2. To achieve
this, we plan to integrate our XY-PHOC based autocompletion into the multimodal MathDeck
search interface [12].4</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>
        We have presented a bidirectional PHOC embedding [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] that has been applied to the ARQMath
2021 Task 2 (Formula Retrieval). The system uses a simple embedding for relative symbol
locations, in order to capture appearance without knowing the underlying writing or mathematical
content for formulas. This simplified embedding can easily be used in standard search engines
designed for text. For ARQMath-2, the XY-PHOC system ranked a close third based on highest
nDCG′, MAP′ and P′@10 metrics, using what is perhaps the most naive formula representation,
based only on the spatial arrangement of symbols. We also showed how XY-PHOC may be used
for autocompletion by switching from disjunctive (weak AND) queries to conjunctive queries.
The results from our autocompletion experiment are promising, and illustrate how a system
could be built using XY-PHOC for both formula retrieval and autocompletion.
      </p>
      <p>For similarity search, an obvious limitation of our approach is that formulas sharing no
symbols (e.g.,  and ) cannot be used to retrieve one another. We believe unified matches
can be implemented through additional symbols in the index that encode mathematical types
and symbol alphabets at diferent levels of granularity (e.g., variable, greek letter, set operation).
We also have not incorporated inverse term frequencies into our scoring model, but believe
that weighting rarer symbols higher in a TF-IDF or BM25 [13] manner may improve results
substantially.</p>
      <p>Our current naïve implementation is very slow for the whole ARQMath collection, but we
believe this can be improved with a more sophisticated implementation along with the use
of rank-safe (e.g., skip lists, MaxScore [9]) and non-rank-safe (e.g., thresholded traversal of
score-based sorted posting lists) optimizations.</p>
      <p>The XY-PHOC symbol location retrieval model requires only symbol positions for indexing
and search, without any representation of formula structure (visual, operation, or otherwise).
This suggests that our approach might be applied to retrieving other two-dimensional notations
and graphics (e.g., tables, plots, figures, chemical diagrams, etc.). It would also be worth
exploring how additional levels of spatial partitioning beyond the currently used five vertical
and horizontal splits improve or adversely impact retrieval efectiveness.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>Our thanks to Wei Zhong for notes from his early investigations into formula autocompletion,
and to Yancarlos Diaz and the RIT DPRL lab for helpful discussions. This material is based
upon work supported by the Alfred P. Sloan Foundation under Grant No. G-2017-9827 and the
National Science Foundation (USA) under Grant No. IIS-1717997. Any opinions, findings and
conclusions or recommendations expressed in this material are those of the author(s) and do not
necessarily reflect those of the National Science Foundation or the Alfred P. Sloan Foundation.
[8] S. Sudholt, G. A. Fink, Evaluating Word String Embeddings and Loss Functions for
CNNBased Word Spotting, in: 2017 14th IAPR International Conference on Document Analysis
and Recognition (ICDAR), volume 01, 2017, pp. 493–498. doi:10.1109/ICDAR.2017.87,
iSSN: 2379-2140.
[9] H. Turtle, J. Flood, Query evaluation: Strategies and optimizations, Information Processing
&amp; Management 31 (1995) 831–850. URL: https://www.sciencedirect.com/science/article/
pii/030645739500020H. doi:https://doi.org/10.1016/0306-4573(95)00020-H.
[10] S. Ding, T. Suel, Faster top-k document retrieval using block-max indexes, in: W. Ma,
J. Nie, R. Baeza-Yates, T. Chua, W. B. Croft (Eds.), Proceeding of the 34th International
ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR
2011, Beijing, China, July 25-29, 2011, ACM, 2011, pp. 993–1002. URL: https://doi.org/10.
1145/2009916.2010048. doi:10.1145/2009916.2010048.
[11] R. Zanibbi, D. W. Oard, A. Agarwal, B. Mansouri, Overview of ARQMath 2020: CLEF Lab
on Answer Retrieval for Questions on Math, in: A. Arampatzis, E. Kanoulas, T. Tsikrika,
S. Vrochidis, H. Joho, C. Lioma, C. Eickhof, A. Névéol, L. Cappellato, N. Ferro (Eds.),
Experimental IR Meets Multilinguality, Multimodality, and Interaction, volume 12260,
Springer International Publishing, Cham, 2020, pp. 169–193. URL: http://link.springer.
com/10.1007/978-3-030-58219-7_15. doi:10.1007/978-3-030-58219-7_15, series
Title: Lecture Notes in Computer Science.
[12] Y. Diaz, G. Nishizawa, B. Mansouri, K. Davila, R. Zanibbi, The mathdeck formula editor:
Interactive formula entry combining latex , structure editing, and search, in: Y. Kitamura,
A. Quigley, K. Isbister, T. Igarashi (Eds.), CHI ’21: CHI Conference on Human Factors in
Computing Systems, Virtual Event / Yokohama Japan, May 8-13, 2021, Extended Abstracts,
ACM, 2021, pp. 192:1–192:5. URL: https://doi.org/10.1145/3411763.3451564. doi:10.1145/
3411763.3451564.
[13] B. Croft, D. Metzler, T. Strohman, Search Engines: Information Retrieval in Practice, 1st
ed., Addison-Wesley Publishing Company, USA, 2009.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mansouri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Oard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zanibbi</surname>
          </string-name>
          ,
          <source>Advancing math-aware search: The arqmath2 lab at clef</source>
          <year>2021</year>
          ., in: European Conference on Information Retrieval, Springer,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>[2] Experimental ir meets multilinguality, multimodality, and interaction</article-title>
          .,
          <source>Proceedings of the Twelfth International Conference of the CLEF Association (CLEF</source>
          <year>2021</year>
          )
          <article-title>(????).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3] Working Notes of CLEF 2021 -
          <article-title>Conference and Labs of the Evaluation Forum (????).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Avenoso</surname>
          </string-name>
          ,
          <article-title>Spatial vs</article-title>
          .
          <source>Graph-Based Formula Retrieval</source>
          ,
          <source>Master's thesis</source>
          , Rochester Institute of Technology,
          <year>2021</year>
          . URL: https://scholarworks.rit.edu/theses/10784.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>K.</given-names>
            <surname>Davila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zanibbi</surname>
          </string-name>
          ,
          <article-title>Layout and Semantics: Combining Representations for Mathematical Formula Search</article-title>
          ,
          <source>in: Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Shinjuku Tokyo Japan,
          <year>2017</year>
          , pp.
          <fpage>1165</fpage>
          -
          <lpage>1168</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3077136.3080748. doi:
          <volume>10</volume>
          .1145/3077136. 3080748.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mansouri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rohatgi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Oard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L.</given-names>
            <surname>Giles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zanibbi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Tangent-CFT</surname>
          </string-name>
          :
          <article-title>An Embedding Model for Mathematical Formulas</article-title>
          ,
          <source>in: Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval</source>
          , ICTIR '19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , pp.
          <fpage>11</fpage>
          -
          <lpage>18</lpage>
          . URL: http://doi.org/10.1145/ 3341981.3344235. doi:
          <volume>10</volume>
          .1145/3341981.3344235.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sudholt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Fink</surname>
          </string-name>
          ,
          <article-title>PHOCNet: A Deep Convolutional Neural Network for Word Spotting in Handwritten Documents</article-title>
          ,
          <source>in: 2016 15th International Conference on Frontiers in Handwriting Recognition (ICFHR)</source>
          , IEEE, Shenzhen, China,
          <year>2016</year>
          , pp.
          <fpage>277</fpage>
          -
          <lpage>282</lpage>
          . URL: http://ieeexplore.ieee.org/document/7814076/. doi:
          <volume>10</volume>
          .1109/ICFHR.
          <year>2016</year>
          .
          <volume>0060</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>