<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Overview of ARQMath-2 (2021): Second CLEF Lab on Answer Retrieval for Questions on Math (Working Notes Version)</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Behrooz</forename><surname>Mansouri</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Rochester Institute of Technology</orgName>
								<address>
									<region>NY</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Richard</forename><surname>Zanibbi</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Rochester Institute of Technology</orgName>
								<address>
									<region>NY</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Douglas</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">University of Maryland</orgName>
								<address>
									<settlement>College Park</settlement>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anurag</forename><surname>Agarwal</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Rochester Institute of Technology</orgName>
								<address>
									<region>NY</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Overview of ARQMath-2 (2021): Second CLEF Lab on Answer Retrieval for Questions on Math (Working Notes Version)</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">358752FB097891503D5E72AF6835CF92</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:42+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Community Question Answering (CQA)</term>
					<term>Mathematical Information Retrieval (MIR)</term>
					<term>Math-aware search</term>
					<term>Math formula search</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper provides an overview of the second year of the Answer Retrieval for Questions on Math (ARQMath-2) lab, run as part of CLEF 2021. The goal of ARQMath is to advance techniques for mathematical information retrieval, in particular retrieving answers to mathematical questions (Task 1), and formula retrieval (Task 2). Eleven groups participated in ARQMath-2, submitting 36 runs for Task 1 and 17 runs for Task 2. The results suggest that some combination of experience with the task design and the training data available from ARQMath-1 was beneficial, with greater improvements in ARQMath-2 relative to baselines for both Task 1 and Task 2 than for ARQMath-1 relative to those same baselines. Tasks, topics, evaluation protocols, and results for each task are presented in this lab overview.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>This second Answer Retrieval for Questions on Math (ARQMath-2) lab <ref type="bibr" target="#b0">1</ref> at the Conference and Labs of the Evaluation Forum (CLEF) continues a multi-year effort to build new test collections for Mathematics Information Retrieval (Math IR) from content found on Math Stack Exchange, <ref type="bibr" target="#b1">2</ref> a Community Question Answering (CQA) forum. Using the question posts from Math Stack Exchange, participating systems are given a question or a formula from a question, and asked to return a ranked list of either potential answers to the question or potentially useful formulae (in the case of a formula query). Relevance is determined by the expected utility of each returned item. These tasks allow participating teams to explore leveraging math notation together with text to improve the quality of retrieval results. Table <ref type="table">1</ref> illustrates these two tasks, and Figure <ref type="figure" target="#fig_0">1</ref> shows the topic format for each task.</p><p>For the CQA task, 146,989 questions from 2020 that contained some text and at least one formula were considered as search topics, from which 100 were selected for use in ARQMath-2.</p><p>CLEF 2021 -Conference and Labs of the Evaluation Forum, September 21-24, 2021, Bucharest, Romania bm3302@rit.edu (B. Mansouri); rxzvcs@rit.edu (R. Zanibbi); oard@umd.edu (D. W. Oard); axasma@rit.edu (A. Agarwal)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Examples of relevant and not-relevant results for tasks 1 and 2 [1]. For Task 2, formulae are associated with posts, indicated with ellipses at right (see Figure <ref type="figure" target="#fig_0">1</ref> for more details). Query formulae are from question posts (here, the question at left), and retrieved formulae are from either an answer or a question post. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Relevant</head><p>You can use AM ≥ GM.</p><formula xml:id="formula_0">1 + 1 + • • • + 1 + √ 𝑛 + √ 𝑛 𝑛 ≥ 𝑛 1/𝑛 ≥ 1 1 − 2 𝑛 + 2 √ 𝑛 ≥ 𝑛 1/𝑛 ≥ 1 Relevant . . . lim 𝑛→∞ 𝑛 √ 𝑛 . . .</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Not Relevant</head><p>If you just want to show it converges, then the partial sums are increasing but the whole series is bounded above by</p><formula xml:id="formula_1">1 + ∫︁ ∞ 1 1 𝑥 2 𝑑𝑥 = 2 Not Relevant . . . ∑︁ 𝑘=1 1 𝑘 2 = 𝜋 2 6</formula><p>. . .</p><p>For the question answering task, the title and body of the question were provided to participating teams, although other associated data (e.g., comments, answers, and links to related questions) were excluded. For the formula search task, an individual formula from the question post is specified as the query, and systems return a ranked list of other potentially useful instances of formulae found in the collection. Each of the 58 formula queries is a single formula extracted from a question used in the CQA task. For both tasks, participating teams had the option to construct queries using only the text or math portions of each question, or to use both math and text. Following convention, we refer to both questions and formula queries as topics.</p><p>The ARQMath labs have three objectives:</p><p>1. Create test collections for training and evaluating Math IR systems. 2. Establish state-of-the-art results on those test collections to which future researchers can compare their results. 3. Foster the growth of the Math IR research community.</p><p>ARQMath-2 saw progress on each of these goals, roughly doubling the size of the available test collections, nearly doubling the number of participating teams, and demonstrating that substantial improvements over the results reported in ARQMath-1 are possible.</p><p>Task 1: Question Answering &lt; T o p i c s &gt; . . . &lt; T o p i c number = "A . 9 " &gt; &lt; T i t l e &gt; S i m p l i f y i n g t h i s s e r i e s &lt; / T i t l e &gt; &lt; Q u e s t i o n &gt; I need t o w r i t e t h e s e r i e s &lt; span c l a s s =``math − c o n t a i n e r ' ' i d =``q_52 ' ' &gt; $$ \ sum_ { n = 0 } ^N nx ^n$$ &lt;/ span &gt; i n a form t h a t d o e s n o t i n v o l v e t h e summation n o t a t i o n , f o r example &lt; span c l a s s =``math − c o n t a i n e r ' ' i d =``q_53 ' ' &gt; $ \ sum_ { i = 0 } ^n i ^2 = \ f r a c { ( n ^2+ n ) ( 2 n + 1 ) } { 6 } $ &lt;/ span &gt; Does anyone have any i d e a how t o do t h i s ? I have a t t e m p t e d m u l t i p l e ways i n c l u d i n g u s i n g g e n e r a t i n g f u n c t i o n s however no l u c k . &lt;/ Q u e s t i o n &gt; &lt; Tags &gt; s e q u e n c e s −and − s e r i e s &lt; / Tags &gt; &lt;/ T o p i c &gt; . . . &lt;/ T o p i c s &gt; Task 2: Formula Retrieval &lt; T o p i c s &gt; . . . &lt; T o p i c number = " B . 9 " &gt; &lt; F o r m u l a _ I d &gt; q_52 &lt; / F o r m u l a _ I d &gt; &lt; L a t e x &gt; \ sum_ { n = 0 } ^N nx ^n &lt; / L a t e x &gt; &lt; T i t l e &gt; S i m p l i f y i n g t h i s s e r i e s &lt; / T i t l e &gt; &lt; Q u e s t i o n &gt; . . . &lt;/ Q u e s t i o n &gt; &lt; Tags &gt; s e q u e n c e s −and − s e r i e s &lt; / Tags &gt; &lt;/ T o p i c &gt; . . . &lt;/ T o p i c s &gt; Figure <ref type="figure" target="#fig_0">1</ref>: XML Topic File Formats for Tasks 1 and 2. Formula queries in Task 2 are taken from questions for Task 1. Here, ARQMath-1 formula topic B.9 is a copy of ARQMath-1 question topic A.9 with two additional tags for the query formula identifier and L A T E X before the question post.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Math IR shares many commonalities with information retrieval more generally. For example, both exploratory search and refinding are common tasks, and query autocompletion and diversity ranking can be useful capabilities. Math IR is a special case of cross-modal retrieval, since both text and math can be used to express the same idea, and those two modalities can be productively used together in the query, the document, or both.</p><p>The nature of mathematics, however, introduces some unique challenges. Here we need to distinguish between mathematics as a field and mathematical notation as a language. The notion of relevance in Math IR is grounded in mathematics as a field, whereas many of the implementation details are grounded in mathematical notation as a language. To see the difference, consider the notion of equality: many people would consider that 3 + 2, 2 + 3, and 5 express the same idea, being equally happy to find formulae that contain any of those. However, many might regard 𝑐𝑜𝑠 2 (𝑥) + 𝑠𝑖𝑛 2 (𝑥) and 1 as different, despite their equality, because the first is specific to some uses of mathematics, and thus not an appropriate formulation for others.</p><p>Indeed, thinking of mathematics as a field is itself too reductionist -mathematics is used in many disciplines (e.g., computer science, physics, chemistry, quantum mechanics, economics, and nearly every branch of engineering). In some cases, relevance may be defined within one of those disciplines, with economists looking for other work in economics, for example. In other cases, relevance might be defined in a way that spans disciplines, such as when an engineer might be looking for the work of mathematicians that can help them to solve some specific problem, even when they don't know the notation the mathematicians would have used in formulating or solving that problem.</p><p>No single evaluation design could possibly model the full complexity of information needs for Math IR, so every evaluation has been specialized in some way. Mathematicians naturally find Math IR potentially interesting, and one motivation for early work on Math IR has been to support mathematics education. Students can use search engines to find references for assignments, to solve problems, increase knowledge, or clarify concepts. In general math-aware search can be used to find similarities between a piece of mathematics being developed, on the one hand, and proved theorems and well-developed theories in the same or different parts of mathematics, on the other hand.</p><p>Complementing this somewhat underdeveloped focus on task design among Math IR researchers is a quite well developed lower-level focus on mathematical notation within the Mathematical Knowledge Management (MKM) community, where the representations of, and operations on math notation have been carefully considered. Among other accomplishments, their activities informed the development of MathML<ref type="foot" target="#foot_2">3</ref> for math on the Web, and novel techniques for math representation and applications such as theorem proving. This community meets annually at the Conference on Intelligent Computer Mathematics (CICM) [2].</p><p>Math IR naturally draws on both of these traditions. Math formula search has been studied since the mid-1990's for use in solving integrals, and publicly available math+text search engines have been around since the DLMF <ref type="bibr" target="#b3">4</ref> system in the early 2000's [3, 4]. Prior to ARQMath, the most widely used evaluation resources for math-aware information retrieval were initially developed over a five-year period at the National Institute of Informatics (NII) Testbeds and Community for Information access Research (at NTCIR-10 [5], NTCIR-11 [6] and NTCIR-12 [7]). NTCIR-12 used two collections, one a set of arXiv papers from physics that is split into paragraph-sized documents, and the other a set of articles from English Wikipedia. The NTCIR Mathematical Information Retrieval (MathIR) tasks developed evaluation methods and allowed participating teams to establish baselines for both "text + math" queries (i.e., keywords and formulae) and isolated formula queries.</p><p>At NTCIR-11 and NTCIR-12, formula retrieval was considered in a variety of settings, including the use of wildcards and constraints on symbols or subexpressions (e.g., requiring matched argument symbols to be variables or constants). Our Task 2, Formula Retrieval, has similarities in design to the NTCIR-12 Wikipedia Formula Browsing task, but differs in how queries are defined and how evaluation is performed. In particular, relevance is defined contextually in AR-QMath, and ARQMath evaluation is based on visually distinct formulae, rather than all (possibly identical) formula instances, as had been done in NTCIR-12. The NTCIR-12 formula retrieval test collection also had a smaller number of queries, with 20 fully specified formula queries (plus 20 variants of those same queries with subexpressions replaced by wildcard characters). NTCIR-11 also had a formula retrieval task, with 100 queries, but in that case systems searched only for exact matches [8].</p><p>Another related effort was the SemEval 2019 [9] question answering task. Question sets from MathSAT (Scholastic Achievement Test) practice exams in three categories were used: Closed Algebra, Open Algebra and Geometry. A majority of the questions were multiple choice, with some having numeric answers. This is a valuable parallel development; the questions considered in the CQA task of ARQMath are more informal and open-ended, and selected from actual Math Stack Exchange user posts (a larger and less constrained set).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">The ARQMath Stack Exchange Collection</head><p>For ARQMath-2, we reused the test collection from the first ARQMath. The test collection was constructed using the March 1st, 2020 Math Stack Exchange snapshot from the Internet Archive. <ref type="bibr" target="#b4">5</ref> Questions and answers from 2010-2018 are included in the collection. The ARQMath test collection contains roughly 1 million questions and 28 million formulae. Formulae in the collection are annotated using &lt;span&gt; tags with the class attribute math-container, and a unique integer identifier given in the id attribute. Formulae are also provided separately in three index files for different formula representations (L A T E X, Presentation MathML, and Content MathML), which we describe in more detail below.</p><p>HTML views of question threads, similar to those on the Math Stack Exchange web site (a question, along with answers and other related information) are also included in the ARQMath test collection. The threads are constructed automatically from Math Stack Exchange snapshot XML files. The threads are intended for those performing manual runs, or who wish to examine search results (on queries other than evaluation queries) for formative evaluation purposes. These threads are also used by assessors during evaluation. The HTML thread files were intended only for viewing threads; participants were asked to use provided XML and formula index files to train their models.</p><p>Questions posted after 2018 are used to create test topics: questions from 2019 were used for the first ARQMath, and questions from 2020 are used for ARQMath-2. Additional details may be found in the ARQMath-1 task overview paper [10].</p><p>Formula Index Files and Visually Distinct Formulae. In addition to L A T E X, it is common for math-aware information retrieval systems to represent formulae as one or both of two types of rooted trees. Appearance is represented by the spatial arrangement of symbols on writing lines (in Symbol Layout Trees (SLTs)), and mathematical syntax (sometimes referred to as (shallow) semantics) is represented using a hierarchy of operators and arguments (in Operator Trees (OPTs)) [11, 12, 13]. The standard representations for these are Presentation MathML (SLT) and Content MathML (OPT).</p><p>To reduce effort for participants, and to maximize comparability across submitted runs, we used LaTeXML<ref type="foot" target="#foot_5">6</ref> 0.8.5 to generate Presentation MathML and Content MathML from L A T E X for each formula in the ARQMath collection. Some L A T E X formulae were malformed, and LaTeXML has some processing limitations, resulting in conversion failures for 0.14% of both SLTs and OPTs. <ref type="bibr" target="#b6">7</ref> Participants could elect to do their own formula extraction and conversions, although the formulae that could be submitted in system runs for Task 2 were limited to those with identifiers in the provided L A T E X formula index file.</p><p>During evaluation we learned that LaTeX formulae that could not be processed by LaTeXML had their visual identifiers assigned incorrectly, and this may have affected adjacent formulae in the formula index files. This had a small effect on evaluation metrics (our best estimate is that no more than 1.3 visually distinct formulae in the pool for each topic were affected).</p><p>ARQMath formulae are provided in L A T E X, SLT, and OPT representations, as Tab Separated Value (TSV) index files. Each line of a TSV file represents a single instance of a formula, containing the formula id, the id of the post in which the formula instance appeared, the id of the thread in which the post is located, a post type (title, question, answer or comment), and the formula representation in either L A T E X, SLT (Presentation MathML), or OPT (Content MathML).</p><p>For ARQMath-2, in the formula TSV index files we added a new field for visually distinct formula identifiers used in evaluation for Task 2 (Formula Retrieval). <ref type="bibr" target="#b7">8</ref> The idea is to identify formulae sharing the same appearance. So for example, two occurrences of 𝑥 2 in a TSV formula index have different formula instance identifiers, but the same visually distinct formula identifier. All ARQMath-2 formula index files provide visually distinct identifiers for each formula in the collection.</p><p>There are three sets of formula index files: one set is for the collection (i.e., for posts from 2018 and before), while the second and third sets are for search topics from 2020 (ARQMath-2), and 2019 (ARQMath-1). Only the collection index files have visually distinct formula identifiers.</p><p>Distribution. The Math Stack Exchange test collection was distributed to participants as XML files on Google Drive. <ref type="bibr" target="#b8">9</ref> To facilitate local processing, the organizers provided python code on GitHub<ref type="foot" target="#foot_9">10</ref> for reading and iterating over the XML data, and generating the HTML question threads.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Task 1: Answer Retrieval</head><p>The main task in ARQMath is the answer retrieval task. Participating systems are given a Math Stack Exchange question post from 2019, and return a ranked list of up to 1,000 answer posts from 2010-2018. System results ('runs') are evaluated using rank quality measures that characterize the extent to which annotated answers with higher relevance come before answers with lower relevance (e.g., nDCG ′ ). This makes Task 1 a ranking task rather than a set retrieval task.</p><p>In the following we describe the Task 1 search topics, runs from participant and baseline systems, the assessment and evaluation procedures used, and a summary of the results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Topics</head><p>In Task 1, participants were given 100 Math Stack Exchange questions posted in 2020 as topics. We used a sampling strategy similar to ARQMath-1, where we chose from questions containing text and at least one formula. To help ensure that most topics had relevant answers available in the collection, we calculated the number of duplicate and related posts for each question, and then chose the majority of topics (89 out of 100) from those with at least one duplicate or related post. <ref type="bibr" target="#b10">11</ref> To increase the difficulty and diversity of topics, we selected the remaining topics from those without annotated duplicates or related posts. <ref type="bibr" target="#b11">12</ref> Because we were interested in a diverse range of search tasks, we also calculated the number of formulae for each question. Finally, we noted the asker's reputation and the tags assigned for each question. We manually drew a sample of 100 questions stratified along those dimensions. In the end, pools for 71 of these questions were evaluated and found to have a sufficient number of relevant responses, and thus were included in the ARQMath-2 test collection.</p><p>The topics were selected from various domains to capture a broad spectrum of mathematical areas. The difficulty level of the topics spanned from easy problems that a beginning undergraduate student might be interested in to difficult problems that would be of interest to more advanced users. The bulk of the topics were aimed at the level of undergraduate math majors (in their 3rd or 4th year) or engineering majors fulfilling their math requirements.</p><p>As organizers, we labeled each question with one of three broad categories, computation, concept or proof. Out of the 71 assessed questions, 25 were categorized as computation, 19 as concept, and 27 as proof. We also categorized questions based on their perceived difficulty level, with 32 categorized as easy, 20 as medium, and 19 as hard. Our last categorization was based on whether a question is dependent on text, formula or both. 10 questions were (in our opinion) dependent on text, 21 on formula and 40 on both.</p><p>The topics were published as an XML file with the format shown in Figure <ref type="figure" target="#fig_0">1</ref>, where the topic number is an attribute of the Topic tag, and the Title, Question and asker-provided Tags are from the Math Stack Exchange question post. To facilitate system development, we provided python code that participants could use to load the topics. As in the collection, the formulae in the topic file are placed in 'math-container' tags, with each formula instance represented by a unique identifier and its L A T E X representation. And, as with the collection, we provided three TSV files, one each for the L A T E X, OPT and SLT representations of the formulae, in the same format as the collection's TSV files.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Participant Runs</head><p>Participating teams submitted runs using Google Drive. A total of 36 runs were received from a total of 9 teams. Of these, 28 runs were declared to be automatic, meaning that queries were automatically processed from the topic file, that no changes to the system had been made after seeing the queries, and that ranked lists for each query were produced with no human intervention. 8 runs were declared to be manual, meaning that there was some type of human involvement in generating the ranked list for each query. Manual runs can contribute diversity to the pool of documents that are judged for relevance, since their error characteristics can differ from those of automatic runs. The teams and submissions are shown in Table <ref type="table" target="#tab_0">2</ref>. Please see the participant papers in the working notes for descriptions of the systems that generated these runs. Important Note: All participant and baseline runs are available online. <ref type="bibr" target="#b12">13</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Baseline Runs: TF-IDF, Tangent-S, Linked Posts</head><p>The organizers ran four baseline systems for Task 1. <ref type="bibr" target="#b13">14</ref> These baselines were also run for ARQMath 2020, and we re-ran them on the same systems as last year, obtaining very similar run-times [10]. Here is a description of our baseline runs.</p><p>1. TF-IDF. A term frequency‚ inverse document frequency model implementation provided by the Terrier system [14]. Formulae are represented using their L A T E X strings. Default parameters from Terrier were used. 2. Tangent-S. Formula search engine using SLT and OPT formula representations [11]. One formula was selected from each Task 1 question title if possible; if there was no formula in the title, then one formula was instead chosen from the question's body. If there were multiple formulae in the selected field, a formula with the largest number of symbols (nodes) in its SLT representation was chosen; if more than one formula had the largest number of symbols, we chose randomly between them. 3. TF-IDF + Tangent-S. Averaging similarity scores from the TF-IDF and Tangent-S baselines. The relevance scores from both systems were normalized in [0,1] using min-max normalization, and then combined using an unweighted average. 4. Linked Math Stack Exchange Posts. This is a simple oracle "system" that is able to see duplicate post links from 2020 in the Math Stack Exchange collection (which were not available to participants). It returns all answer posts from 2018 or earlier that were in threads that Math Stack Exchange moderators had marked as duplicating the topic question post. Answer posts are sorted in descending order by their vote scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Assessment</head><p>Pooling. Participants were asked to rank up to 1,000 answer posts for each topic, which were then sampled for assessment using Top-𝑘 pooling. The top 45 results were combined from all primary runs. To this, we added the top 15 results from each alternate run. The baseline systems, TF-IDF + Tangent-S and Linked Math Stack Exchange Posts, were considered as primary runs and the other two (TF-IDF and Tangent-S) were considered as alternative. Duplicates were then deleted, and the resulting pool was sorted randomly for display to assessors. The pooling depth was designed to identify as many relevant answer posts as possible given our assessment resources. On average, the pools contained 448.12 answers per topic. Relevance definition. We used the same relevance definitions created for ARQMath-1. To avoid assessors needing to guess about the level of mathematical knowledge available to the Math Stack Exchange users who originally posted the questions, we asked assessors to base their judgments on the degree of usefulness for an expert (modeled in this case as a math professor), who might then try to use that answer to help the person who had asked the original question. We defined four levels of relevance, as shown in Table <ref type="table">3</ref>.</p><p>Assessors were allowed to consult external sources to familiarize themselves with the topic of a question, but relevance judgments were made using only information available within the ARQMath test collection. For example, if an answer contained a Math Stack Exchange link such as https://math.stackexchange.com/questions/163309/pythagorean-theorem, they could follow that link to better understand the intent of the person writing the answer, but an external link to the Wikipedia page https://en.wikipedia.org/wiki/Pythagorean_theorem would not be followed.</p><p>Training. Unlike ARQMath-1, for ARQMath-2 participants could use the 77 annotated topics for ARQMath-1 Task 1 as a training set [10, 15]. For sanity checking results and comparison, results were collected from participants for both the ARQMath-1 and ARQMath-2 topics, and results for both training (ARQMath-1) and testing (ARQMath-2) are provided at the end this document.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Relevance Assessment Criteria for Tasks 1 and 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Score Rating Definition</head><p>Task 1: Answer Retrieval 3 High Sufficient to answer the complete question on its own 2 Medium Provides some path towards the solution. This path might come from clarifying the question, or identifying steps towards a solution 1 Low Provides information that could be useful for finding or interpreting an answer, or interpreting the question 0 Not Relevant Provides no information pertinent to the question or its answers. A post that restates the question without providing any new information is considered nonrelevant</p><p>Task 2: Formula Retrieval 3 High Just as good as finding an exact match to the query formula would be 2 Medium Useful but not as good as the original formula would be 1 Low There is some chance of finding something useful 0 Not Relevant Not expected to be useful</p><p>Assessment System. For ARQMath-2, assessments were again performed using Turkle <ref type="bibr" target="#b14">15</ref> , a locally installed system with functionality similar to Amazon Mechanical Turk. As Figure <ref type="figure">4</ref> at the end of this document illustrates, there were two panels in the Turkle user interface. The question was shown on the left panel, with the Title on top in a grey bar; below that was the question body. There was also a Thread link, through which assessors could access the Math Stack Exchange post in context, with the question and all answers given for this question (in 2020). In the right panel, the answer to be judged was shown at the top, along with another thread link that allows assessors to view the original thread in which the answer post appeared, which could be helpful for clarifying the context of the answer post, for example by viewing the original question to which it had been posted as a response. Finally, below the answer in the right panel was where assessors selected relevance ratings.</p><p>In addition to four levels of relevance, two additional choices were available: 'System failure' indicated system issues such as unintelligible rendering of formulae, or the thread link not working (when it was essential for interpretation). If after viewing the threads, the assessors were still not able to decide the relevance degree, they were asked to choose 'Do not know'. The organizers asked the assessors to leave a comment in the event of a system failure or a 'Do not know' selection. As it happened, the ARQMath-2 assessors did not use these options for Task 1; for each answer, they decided a relevance degree.</p><p>Assessor Training. Seven paid undergraduate and graduate mathematics and computer science students from RIT and St. John Fisher College were paid to perform relevance judgments. One assessor had worked with us previously on ARQMath-1. Due to the COVID pandemic, all training sessions were performed remotely over Zoom. For ARQMath-1, relevance criteria had been developed interactively with assessors, leading to four rounds of training; we found the resulting relevance criteria worked well, and so we reused them for ARQMath-2. This allowed us to reduce assessor training time: the four assessors who worked exclusively on Task 1 participated in three meetings, and just two rounds of training assessments. The remaining three assessors initially worked on Task 2, and were later moved to Task 1 after Task 2 assessment was completed. Those three assessors had an initial training meeting when they returned to Task 1 to introduce the task, and then they performed a single round of training assessments, with subsequent discussion at a second meeting. One of those three assessors had previously worked on both Task 1 and Task 2 assessments for ARQMath-1.</p><p>At the first assessor meeting, the lab and administrative details were discussed. After this, the assessors were divided into two groups, for Task 1 and Task 2. After this we began training sessions. In the first Task 1 training session, the task was explained, making reference to specific topics and previously assessed answers from ARQMath-1. For each training/practice assessment round, the same 7 topics were assigned to every assessor and the assessors worked independently, thus permitting inter-assessor agreement measures to be computed. After completing training assessments, a meeting was held to discuss disagreements in relevance scores between with the organizers, along with clarifications of the relevance criteria. The assessors discussed the reasoning for their choices, with the fourth author of this paper (an expert Math Stack Exchange user) sharing their own assessment and reasoning. The primary goal of training was to help assessors make self-consistent annotations, as question interpretations will vary across individuals.</p><p>Some of the question topics would not be typically covered in regular undergraduate courses, so that was a challenge that required the assessors to get a basic understanding of those topics before they could do the assessment. The assessors found the question threads made available in the Turkle interface helpful in this regard (see Figure <ref type="figure">4</ref>).</p><p>Figure <ref type="figure" target="#fig_1">2</ref> shows agreement between assessors in our two groups over the course of the training process. As shown, collapsing relevance to be binary by considering high and medium as relevant and low and not-relevant as a not-relevant (henceforth "H+M binarization") yielded better agreement among the assessors. <ref type="bibr" target="#b15">16</ref> Assessment. A total of 81 topics were assessed for Task 1. 10 judgment pools (for topics A.208, A.215, A.216, A.221, A.230, A.236, A.266, A.277, A.278 and A.280) had zero or one posts with relevance levels of high or medium; these topics were removed from the collection because topics with no relevant posts cannot be used to distinguish between ranked retrieval systems, and because topics with only a single relevant post result in coarsely quantized values for the evaluation measures that we report. For the remaining 71 topics, an average of 447.7 answers were assessed, with an average assessment time of 83.3 seconds per answer post. The average number of answers labeled with any degree of relevance (high, medium, or low; henceforth "H+M+L binarization") was 49.0 per question, with the highest number of relevant answers being 134 (for topic A.237) and the lowest being 4 (for topic A.227).</p><p>Post Assessment. After assessments were completed for Task 1, each assessor was assigned one topic that had originally been completed by another assessor. <ref type="bibr" target="#b16">17</ref> We were particularly interested in confirming cases in which non relevant documents were found, so for each <ref type="bibr" target="#b15">16</ref> H+M binarization corresponds to the definition of relevance usually used in the Text Retrieval Conference (TREC). The TREC definition is "If you were writing a report on the subject of the topic and would use the information contained in the document in the report, then the document is relevant. Only binary judgments ('relevant' or 'not relevant') are made, and a document is judged relevant if any piece of it is relevant (regardless of how small the piece is in relation to the rest of the document). " (source: https://trec.nist.gov/data/reljudge_eng.html) <ref type="bibr" target="#b16">17</ref> One assessor (with id 7) was not able to continue assessment. assessor we selected the topic with the fewest relevant topics. Among the 6 dual-assessed topics, 4 had no high or medium relevant answers according to at least one of the two assessors <ref type="bibr" target="#b17">18</ref> ; meaningful values of kappa for binary relevance can not be calculated in such cases. Averaged over the remaining two questions, kappa was 0.21 on the four-way assessment task, and using H+M binarization it was 0.32.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Evaluation Measures</head><p>For a complex task where rich training data is not yet available, it is possible that a large number of relevant answers may be missed during assessment. Measures which treat unjudged documents as not relevant can be used when directly comparing systems contributing to the judgment pools, but non-contributing systems can be disadvantaged by treating unjudged documents as not relevant, which may prove to be relevant in later analysis. We therefore chose the nDCG ′ measure (read as "nDCG-prime") introduced by Sakai and Kando [16] as the primary measure for the task. nDCG is a widely used measure for graded relevance judgments, used to produce a single figure of merit over a set of ranked lists. For ARQMath, each retrieved document earns a gain value (relevance score) 𝑔 ∈ {0, 1, 2, 3}, discounted by a slowly decaying function of the rank position of each result. Discounted gain values are accumulated and then normalized to [0,1] by dividing by the maximum possible Discounted Cumulative Gain (i.e., from all relevant documents sorted in decreasing order of gain value). This results in normalized Discounted Cumulative Gain (nDCG).</p><p>The only difference when computing nDCG ′ is that unjudged documents are removed from the ranked list before performing the computation. It has been shown that nDCG ′ has somewhat better discriminative power and somewhat better system ranking stability (with judgement ablation) than the bpref measure [17] used recently for formula search (e.g., [12]). Moreover, nDCG ′ yields a single-valued measure with graded relevance, whereas bpref, Precision@k, and Mean Average Precision (MAP) all require binarized relevance judgments. In addition to nDCG ′ , we also compute Mean Average Precision (MAP) with unjudged posts removed (thus MAP ′ ), and Precision at 10 with unjudged posts removed (P ′ @10). <ref type="bibr" target="#b18">19</ref> For MAP ′ and P ′ @10 we used H+M binarization.</p><p>The ARQMath Task 1 evaluation script removes unjudged posts as a preprocessing step where required, and then computes evaluation measures using trec_eval.<ref type="foot" target="#foot_14">20</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.6.">Results</head><p>Table <ref type="table">4</ref> in the appendix shows the results for baselines along with teams and their systems ranked by nDCG ′ . nDCG ′ values can be interpreted as the average (over topics) of the fraction of the score for the best possible that was actually achieved. As can be seen, the best nDCG ′ value that was achieved was 0.434, by the MathDowsers team. MAP ′ with H+M binarization generally ranks systems in the same order as nDCG ′ does with graded relevance judgments. However, the results for P ′ @10 with H+M binarization differ, the TU_DBS team doing best among the participating teams by that measure (exceeded only by the Linked MSE posts baseline, which uses human-built links that were not available to participating teams). There are some noticeable differences in system orderings for several participating teams when using ARQMath-2 topics compared with what was seen when those same teams used the same systems (in 2021) on ARQMath-1 topics. Now comparing results from 2021 with results from 2020, we see that the best improvement over the strongest fully automated baseline in both years (TF-IDF + Tangent-S) was substantially larger in 2021 than in 2020. Specifically, in 2020 the MathDowsers team outperformed that baseline by 39% as measured by nDCG ′ ; in 2021 they outperformed that same baseline by 116% as measured by nDCG ′ .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Task 2: Formula Retrieval</head><p>In the formula retrieval task, participants were presented with one formula from a 2020 question used in Task 1, and asked to return a ranked list of up to 1,000 formula instances from questions or answers from the evaluation epoch (2018 or earlier). Formulae were returned by their identifiers in math-container tags and the companion TSV L A T E X formula index file, along with their associated post identifiers.</p><p>As with Task 1, ranked lists were evaluated using rank quality measures, making this a ranking task rather than a set retrieval task. Three key details differentiate Task 2 from Task 1:</p><p>1. Unlike Task 1, in Task 2 the goal is not answering questions, but to instead show the searcher formulae that might be useful as they seek to satisfy their information need. Task 2 is thus still grounded in the question, but the relevance of a retrieved formula is defined by a formula's expected utility, not just the post in which any one formula instance was found. 2. In Task 1 only answer posts were returned, but for Task 2 the formulae may appear in answer posts or in question posts. 3. For Task 2 we distinguish visually distinct formulae from instances of those formulae, and evaluate by the ranking of visually distinct formulae returned. We call formulae appearing in posts formula instances, and of course the same formula may appear in more than one post. By a visually distinct formula we mean a set of formula instances that are visually identical when viewed in isolation. For example, 𝑥 2 is a formula, 𝑥•𝑥 is a different visually distinct formula, and each time 𝑥 2 appears is an instance of the visually distinct formula 𝑥 2 . Although systems in Task 2 rank formula instances in order to support the relevance judgment process, the evaluation measure for Task 2 is based on the ranking of visually distinct formulae.</p><p>The remainder of this section describes for Task 2 the search topics, the submissions and baselines, the process used for creating relevance judgments, the evaluation measures, and the results.</p><p>Important Note: run data and code for baseline systems may be found at links provided in the previous Section.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Topics</head><p>In Task 2, participating teams were given 100 mathematical formulae, each found in a different Math Stack Exchange question from Task 1 (posted in 2020). They were asked to find relevant formulae instances from either question or answer posts in the test collection (from 2018 and earlier). The topics for Task 2 were provided in an XML file similar to those of Task 1, in the format shown in Figure <ref type="figure" target="#fig_0">1</ref>. Task 2 topics differ from their corresponding Task 1 topics in three ways:</p><p>1. Topic number. For Task 2, topic ids are in the form "B.x" where x is the topic number.</p><p>There is a correspondence between topic id in tasks 1 and 2. For instance, topic id "B.209" indicates the formula is selected from topic "A.209" in Task 1, and both topics include the same question post (see Figure <ref type="figure" target="#fig_0">1</ref>). 2. Formula_Id. This added field specifies the unique identifier for the query formula instance. There may be other formulae in the Title or Body of the question post, but the query is only the formula instance specified by this Formula_Id. 3. L A T E X. This added field is the L A T E X representation of the query formula instance as found in the question post.</p><p>Because query formulae are drawn from Task 1 question posts, the same L A T E X, SLT and OPT TSV files that were provided for the Task 1 topics can be consulted when SLT or OPT representations for a query formula are needed. Formulae for Task 2 were manually selected using a heuristic approach to stratified sampling over three criteria: complexity, elements, and text dependence. Formulae complexity was labeled low, medium or high by the fourth author. For example, 𝑑𝑓 𝑑𝑥 = 𝑓 (𝑥 + 1) is low complexity,</p><formula xml:id="formula_2">∑︀ 𝑛 𝑘=0 (︀ 𝑛 𝑘 )︀</formula><p>𝑘 is medium complexity, and</p><formula xml:id="formula_3">𝑥 − 𝑥 3 3 × 3! + 𝑥 5 5 × 5! − 𝑥 7 7 × 7! + • • • = ∞ ∑︁ 𝑛=0 (−1) 𝑛 𝑥 (2𝑛+1) (2𝑛 + 1) × (2𝑛 + 1)! is high complexity.</formula><p>Text dependence reflected the first author's opinion of the degree to which text in the Title and Question fields were likely to yield related search results. For instance, for one Task 2 topic, the query formula is 𝑑𝑓 𝑑𝑥 = 𝑓 (𝑥 + 1) whereas the complete question is: "How to solve differential equations of the following form: 𝑑𝑓 𝑑𝑥 = 𝑓 (𝑥 + 1) . " When searching for this formula, perhaps the surrounding text could safely be ignored. At most one formula was selected from each Task 1 question topic to produce Task 2 topics. For cases in which suitable formulae were present in both the title and the body of the Task 1 question, we selected the Task 2 formula query from the title.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Participant Runs</head><p>A total of 17 runs were received for Task 2 from a total of six teams, as shown in Table <ref type="table" target="#tab_0">2</ref>. Each run contained at most 1,000 formula instances for each topic, ranked in decreasing order of systemestimated relevance to that query. For each formula instance in a ranked list, participating teams provided the formula_id and the associated post_id for that formula. Please see the participant papers in the working notes for descriptions of the systems that generated these runs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Baseline Run: Tangent-S</head><p>We again used Tangent-S [11] as our baseline. Unlike Task 1, a single formula is specified for each Task 2 query, so no formula selection step was needed. This Tangent-S baseline makes no use of the question text. Timing was similar to the use of Tangent-S in ARQMath-1.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4.">Assessment</head><p>Pooling. The retrieved items for Task 2 are formula instances, but pooling was done based on the visually distinct formulae, and not individual formula instances. Visually distinct formulae were identified by clustering all formula instances in the collection. <ref type="bibr">21</ref> Pooling was performed by then proceeding down each results list until at least one instance of some number of visually distinct formulae had been seen. For primary runs and for the baseline run, the pool depth was the rank of the first instance of the 20th visually distinct formula; for alternate runs the pool depth was the rank of the first instance of the 10th visually distinct formulae. <ref type="bibr">22</ref> Clustering of visually distinct formulae instances was performed using the SLT representation when possible, <ref type="bibr">23</ref> and the L A T E X representation otherwise. We first converted the Presentation MathML representation to a string representation using Tangent-S, which performed a depthfirst traversal of the SLT, with each SLT node and edge generating a single character of the SLT string. Formula instances with identical SLT strings were considered to be the same formula; note that this ignores differences in font. For formula instances with no Tangent-S SLT string available, we removed the white space from their L A T E X strings and grouped formula instances with identical strings. This process is simple and appears to be reasonably robust, but it is possible that some visually identical formula instances were not captured due to LaTeXML conversion failures, or where different L A T E X strings produce the same formula (e.g., if subscripts and superscripts appear in a different order in L A T E X).</p><p>Assessment was done on formula instances: for each visually distinct formula we selected at most five instances to assess. We did this differently than last year; in order to prefer highlyranked instances and instances returned in multiple runs, we selected the 5 instances using a simple voting protocol, where each instance votes by the sum of its reciprocal ranks within each run, breaking ties randomly. Out of 8,129 visually distinct formulae that were assessed, 117 (1.4%) had instances in more than 5 pooled posts. <ref type="bibr">24</ref> Relevance definition. The relevance judgment task was defined for assessors as follows: for a formula query, if a search engine retrieved one or more instances of this retrieved formula, would that have been expected to be useful for the task that the searcher was attempting to accomplish?</p><p>Assessors were presented with formula instances in context (i.e., in the question or answer in which they had been found). They were then asked to decide their relevance by considering whether retrieving either that instance or some other instance of that formula could have helped the searcher to address their information need. To make this judgment, they were shown the query formula within the question post where it appeared. Each formula instance in the judgment pool was assigned one of four relevance levels as defined in Table <ref type="table">3</ref>.</p><p>For example, if the formula query was ∑︀ 1 𝑛 2+cos 𝑛 , and the formula instance to be judged is</p><formula xml:id="formula_4">∑︀ ∞ 𝑛−1 1</formula><p>𝑛 2 , the assessors could look at the formula's associated post, compare the formula's variable types and operations with the query, identify the area of mathematics it concerns, and then decide whether finding the second formula rather than the first would be expected to yield good results. Further, they could consider the content of the question post containing the query (and, optionally, the thread containing that question post) in order to understand the searcher's information need. Thus the question post fills a role akin to Borlund's simulated work task [18], although in this case the title, body and tags from the question post are included in the topic and thus can optionally be used by the retrieval system.</p><p>The assessor could also consult the post containing a retrieved formula instance (which may be another question post, or an answer post) along with the associated thread, to see if in that case the formula instance would indeed have been a useful basis for a search. Note, however, that the assessment task is not to determine whether the specific post containing the retrieved formula instance is useful, but rather to use that context as a basis for estimating the degree to which useful content would likely be found if this or other instances of the retrieved formula were returned by a search engine.</p><p>Although this definition of relevance was unchanged between ARQMath-1 and ARQMath-2, we did make one potentially significant change to the way this relevance definition was interpreted for ARQMath-2. In ARQMath-2, Task 2 assessment was done based on the context in which the formulas appear, exactly as described above. This means that it is possible that an exact match with the query formula may in some cases (depending on context) be considered not relevant. For example, for the formula query 𝑥 𝑛 + 𝑦 𝑛 + 𝑧 𝑛 (B.289) x, y, and z could be any real numbers in the original question post in which that formula had originally appeared. The assessors considered all exact matches in the pooled posts in which x, y, and z referred not to real numbers but specifically to integers as not relevant. On the other hand, formulas that do not share the same appearance or syntax with the query might be considered relevant. This is usually the case where they are both referring to the same concept. For the query formula It should be noted that the last author (a mathematician) reviewed these two examples and agreed with the assessors. In 2020, by contrast, assessors had been instructed during training that if the query and candidate formulae were the same (in appearance), then the candidate was certainly highly relevant. During assessor training in 2021 this issue received considerable attention and discussion, and we ultimately concluded that our guidance in 2020 had not been fully consistent with our relevance definition. We therefore clarified the interpretation of 'exact match' for ARQMath-2 annotations in 2021 to take the formula semantics and context directly into account even in the case of identical formulae (so for example, variables of different types would not be considered the same, even if variable names are identical). This change may affect the utility of some ARQMath-1 relevance judgments for training systems that will be evaluated using ARQMath-2 (or subsequent ARQMath) relevance judgments.</p><p>As in 2020, we defined the relevance score for a formula to be the maximum relevance score for any judged instance of that formula. This relevance definition essentially asks "if instances of this formula were returned, would we reasonably expect some of those instances to be useful?"</p><p>Assessment System. We again used Turkle to build the assessment system for Task 2. As shown in Figure <ref type="figure">4</ref> (at the end of this document), there are two main panels. In the left panel, the question is shown as in Task 1, but now with the formula query highlighted in yellow. In the right panel, up to five retrieved posts (question posts or answer posts) containing instances of the same retrieved formula are displayed, with the retrieved formula instance highlighted in each case. For example, the formula ∑︀ ∞ 𝑛=1 𝑎 𝑛 shown in Figure <ref type="figure">4</ref> was retrieved in two question posts. As in Task 1, buttons are provided for the assessor to record their judgment; unlike Task 1, judgments for each instance of the same retrieved formula (up to 5) are recorded separately, and later used to produce a single maximum score for each visually distinct formula.</p><p>Assessor training. Three assessors were assigned to to perform relevance judgements for Task 2, one of whom had also assessed Task 2 for ARQMath-1 in 2020. Three rounds of training were performed.</p><p>In the first training round, the assessors were familiarized with the task. To illustrate how for-mula search might be used, we interactively demonstrated formula suggestion in MathDeck [19]  and the formula search capability of Approach0 [13]. Then the task was defined using examples, showing a formula query with some retrieved results, talking through the relevance definitions and how to apply those definitions in specific cases. Two topics from ARQMath-1 (B. ) to independently assess. The formulae to assess were chosen manually using the same process as the first training round. After assessment, the assessors and organizers met by Zoom to discuss and resolve disagreements. The assessors used this opportunity to refine their understanding of the relevance criteria, and the application of those criteria to specific cases. Assessor agreement was found to be fairly good (kappa=0.281 over four relevance levels and kappa=0.417 with H+M binary relevance). The assessors were then each assigned another 7 ARQMath-1 topics (B.8, B.69, B.83, B.89, B.92, B.95, B.98) and a third round of assessment practice followed by discussion was performed. The average kappa on the these topics was 0.467 over four relevance levels, and 0.565 for H+M binary relevance, agreement levels consistent with those observed at the end of Task 2 assessor training in 2020 [10]. Figure <ref type="figure" target="#fig_3">3</ref>.(a) shows the Cohen's kappa coefficient values for each assessor in the last training round.</p><p>Assessment. A total of 60 topics were assessed for Task 2. Two queries (B.243 and B.266) had fewer than two relevant answers after H+M binarization and were removed. Of the remaining 58 queries, an average of 140.0 visually distinct formulae were assessed per topic, with an average assessment time of 39.5 seconds per formulae. The average number of formula instances labeled as relevant after H+M binarization was 30.9 per topic, with the highest being 107 for topic B.296 and the lowest being 3 for topics B.211 and B.255.</p><p>Post Assessment. After assessment for Task 2 was completed, each of the three assessors were assigned two topics, one of which had been assessed by each of the other two assessors. Figure <ref type="figure" target="#fig_3">3</ref> shows the Cohen's kappa coefficient values for each assessor. A kappa of 0.329 was achieved on the four-way assessment task, and with H+M binarization the average kappa value was 0.694.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.5.">Evaluation Measures</head><p>As for Task 1, the primary evaluation measure for Task 2 is nDCG ′ , and MAP ′ and P ′ @10 were also computed. Participants submitted ranked lists of formula instances, but we computed these measures over visually distinct formulae. The ARQMath-2 Task 2 evaluation script replaces each formula instance with its associated visually distinct formula, and then deduplicates from the top of the list downward, producing a ranked list of visually distinct formulae, from which our prime evaluation measures are then computed using trec_eval. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.6.">Results</head><p>Table <ref type="table">5</ref> in the appendix shows the results, with the baseline run shown first, and then teams and their systems ranked by nDCG ′ . For ARQMath 2 topics, we see that the best results by nDCG ′ were achieved by the Approach0 team, with the MathDowsers team doing almost as well by that measure, and the XY-PHOC-DPRL team a close third. The order between the best runs from each of those three teams is the same when evaluated on ARQMath-2 topics using MAP ′ and P ′ @10.</p><p>Comparing ARQMath-2 results from 2021 with the last year's (2020) ARQMath-1 results, we see that (as with Task 1) for Task 2 the performance relative to the baseline is substantially improved in 2021 over 2020. Specifically, in 2020 the team with the best nDCG ′ (DPRL) was 15% below the Tangent-S baseline by that measure; in 2021 the team with the best nDCG ′ (Approach0) outperformed the Tangent-S baseline by 13%, as measured by nDCG ′ .</p><p>We note, however, a substantial drop in MAP ′ and NDCG ′ between ARQMath-1 topics and ARQMath-2 topics for the Tangent-S baseline on Task 2. Interestingly, the P ′ @10 results are much less adversely affected than the rank-based metrics when moving from the ARQMath-1 to the ARQMath-2 Task 2 topics. It is possible that these differences might result from the different instructions that we gave relevance assessors in 2021 regarding assessment of identical formulae, since it seems that the 2021 Task 2 assessors did indeed take the question context and formula semantics more directly into account when performing assessment for Task 2 of ARQMath-2 than did the ARQMath-1 assessors in 2020.</p><p>Tangent-S linearly combines six similarity scores computed from OPTs and SLTs, with weights fit using the NTCIR-12 math formula search topics. NTCIR-12 assessors had considered both visual and operator syntax similarity when comparing formulas retrieved in isolation, but without checking the types and domains for variables and operators associated with query and candidate formulas. This approach had transferred reasonably well when evaluated using ARQMath-1 formula queries, but the additional variation in the way identical formulae were assessed in ARQMath-2 may have made the problem more challenging for Tangent-S. In a quick experiment with retuning Tangent-S, we found that simply using the absolute value of the combination weights (so that no similarity scores acted as penalties) increased nDCG' to 0.536. Without reading too much into that one simple experiment, we can at least say that the Tangent-S baseline that we report does not seem to be optimally tuned to the ARQMath-2 collection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>This second year of ARQMath resulted in an improved test collection, more participation, and better results. We anticipate continuing ARQMath for a third year, with participants in ARQMath benefiting from a mature evaluation infrastructure, a larger (and perhaps now also somewhat better) set of relevance judgements on which to train, and a larger and more diverse community of researchers with whom to share ideas.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>ARQMath-2 Task 1 (CQA) results. P indicates a primary run, M indicates a manual run, and (✓) indicates a baseline pooled at the primary run depth. For Precision′@10 and MAP′, H+M binarization was used. The best baseline results are in parentheses. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ARQMath</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Task 1 :</head><label>1</label><figDesc>Question Answering Task 2: Formula Retrieval Question I have spent the better part of this day trying to show from first principles that this sequence tends to 1. Could anyone give me an idea of how I can approach this problem?</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Inter-annotator agreement (Average Cohen' kappa) over 7 assessors during the last Task 1 training (7 topics from ARQMath-1); four-way classification (gray) and two-way (H+M binarized) classification (black). Chart (a) shows the agreement between the assessors who did only Task1 and had an additional training session. Chart (b) shows the agreement between the assessors who started with Task 2, and then moved to Task1.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>𝑆 𝑛 ≥ 𝑛 √ 𝑃 (B.277), formula 1+2+3+...+𝑛 𝑛 ≥ 𝑛 √ 𝑛! has medium relevance. Both formulae are referring to the AM-GM inequality (of Arithmetic and Geometric Means).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Inter-assessor agreement (Cohen's kappa) over 3 assessors. Chart (a) shows the agreement on the last training round (7 topics from ARQMath-1). Chart (b) shows the agreement after official Task 2 assessment. Each assessor evaluated two topics, one by each the other two assessors. Shown are four-way classification (gray) and two-way (H+M binarized) classification (black).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 2</head><label>2</label><figDesc>Submitted Runs for Task 1 (36 runs) and Task 2 (17 runs). Additional baselines for Task 1 (4 runs) and Task 2 (1 run) were also generated by the organizers.</figDesc><table><row><cell></cell><cell cols="2">Automatic</cell><cell cols="2">Manual</cell></row><row><cell></cell><cell>Primary</cell><cell>Alternate</cell><cell>Primary</cell><cell>Alternate</cell></row><row><cell>Task 1: Answers</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Baselines</cell><cell>2</cell><cell>2</cell><cell></cell></row><row><cell>Approach0</cell><cell></cell><cell></cell><cell>1</cell><cell>4</cell></row><row><cell>BetterThanG</cell><cell></cell><cell>2</cell><cell>1</cell><cell>2</cell></row><row><cell>DPRL</cell><cell>1</cell><cell>2</cell><cell></cell></row><row><cell>GoogolFuel</cell><cell>1</cell><cell>4</cell><cell></cell></row><row><cell>MathDowsers</cell><cell>1</cell><cell>1</cell><cell></cell></row><row><cell>MIRMU</cell><cell>1</cell><cell>4</cell><cell></cell></row><row><cell>MSM</cell><cell>1</cell><cell>4</cell><cell></cell></row><row><cell>PSU</cell><cell>1</cell><cell></cell><cell></cell></row><row><cell>TU_DBS</cell><cell>1</cell><cell>4</cell><cell></cell></row><row><cell>Task 2: Formulas</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Baseline</cell><cell>1</cell><cell></cell><cell></cell></row><row><cell>Approach0</cell><cell></cell><cell></cell><cell>1</cell><cell>4</cell></row><row><cell>DPRL</cell><cell>1</cell><cell>3</cell><cell></cell></row><row><cell>MathDowsers</cell><cell>1</cell><cell>1</cell><cell></cell></row><row><cell>NLP-NITS</cell><cell>1</cell><cell></cell><cell></cell></row><row><cell>TU_DBS</cell><cell>1</cell><cell>3</cell><cell></cell></row><row><cell>XY_PHOC_DPRL</cell><cell>1</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>1, B.18) were selected as examples. During the training session, the assessors saw different example results for topics and discussed their relevance based on criteria defined for them with the organizers. These examples were manually selected from ARQMath-1 relevance judgments having different relevance degrees, and included examples from dual-assessed topics that 2020 assessors had disagreements on. The assessors also received feedback from the fourth author of this paper, an expert Math Stack Exchange user. All three assessors were then assigned 7 other Task 2 topics from ARQMath-1 (B.29, B.32, B.33, B.41, B.59, B.62, B.70</figDesc><table /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://www.cs.rit.edu/~dprl/ARQMath</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://math.stackexchange.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://www.w3.org/Math</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://dlmf.nist.gov</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://archive.org/download/stackexchange</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://dlmf.nist.gov/LaTeXML</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">We thank Deyan Ginev and Vit Novotny for helping reduce LaTeXML failures: for ARQMath-1 conversion failures affected</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">8% of SLTs, and 10% of OPTs.<ref type="bibr" target="#b7">8</ref> We thank Frank Tompa for sharing this suggestion at CLEF 2020.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8"><ref type="bibr" target="#b8">9</ref> https://drive.google.com/drive/folders/1ZPKIWDnhMGRaPNVLi1reQxZWTfH2R4u3</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9"><ref type="bibr" target="#b9">10</ref> https://github.com/ARQMath/ARQMathCode</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">Participating systems did not have access to this information.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_11">In ARQMath-1, all topics had links to at least one duplicate or related post that were available to the organizers.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_12">https://github.com/hltcoe/turkle</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="19" xml:id="foot_13">Pooling to at least depth 10 ensures that there are no unjudged posts above rank 10 for any baseline, primary, or alternative run. Note that P ′ @10 cannot achieve a value of 1 because some topics have fewer than 10 relevant posts.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="20" xml:id="foot_14">https://github.com/usnistgov/trec_eval</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="21" xml:id="foot_15">This differs from the approach used for ARQMath-1, when only submitted formula instances were clustered. For ARQMath-2 the full formula collection was clustered to facilitate post hoc use of the resulting test collection.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="22" xml:id="foot_16">In ARQMath-2, Task 1 pools were not used to seed Task 2 pools.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="23" xml:id="foot_17">For ARQMath-1, 92% of formula instances had an SLT representation; for ARQMath-2 we reparsed the collection and improved this to 99.9%.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="24" xml:id="foot_18">  24  As mentioned in Section 3, a relatively small number of formulae per topic had incorrectly generated visual ids. In 6 cases assessors indicated that a pooled formula for a single visual id was 'not matching' the other formulae in hits grouped for a visual id, rather than assign a relevance score for the formula.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We thank our student assessors from RIT and St. John Fisher College: Josh Anglum, Dominick Banasick, Aubrey Marcsisin, Nathalie Petruzelli, Siegfried Porterfield, Chase Shuster, and Freddy Stock. This material is based upon work supported by the National Science Foundation (USA) under Grant No. IIS-1717997 and the Alfred P. Sloan Foundation under Grant No. G-2017-9827.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 5</head><p>ARQMath-2 Task 2 (Formula Retrieval) results, computed over visually distinct formulae. P indicates a primary run, and (✓) shows the baseline pooled at the primary run depth. For MAP ′ and P ′ @10, relevance was thresholded H+M binarization. All runs were automatic. Baseline results are in parentheses.</p><p>ARQMath </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Answer Post</head><p>Thread Title: Does the radius of convergence for a power series change when you multiply it by a constant?</p><p>Question:</p><p>I have to use differentiation to find power series representations for certain functions. I also have to find the radius of convergence. For part a) of the problem, I have to do this for . I then have to use the answer for that problem to find the power series representation for . However, in order to do so, I have to modify my answer from part a) and multiply it by . This brings me to my question: does the radius of convergence for a power series change when you multiply it by a constant? What if it wasn't a power series but some other type of series such as a rational or algebraic one, would it change then? Thread Title: Need to show the following function is uniformly continuous on R Question:</p><p>Could you please tell me how I am supposed to show that is uniformly continuous in . I did some pre-calculation and found that if So, is the selection of correct since I am supposed to find one single , so is or not?  Shown are hits for Formula Retrieval (Task 2). In the left panel, the formula query is highlighted. In the right panel, two question posts containing the same retrieved formula are shown. For Task 1, a similar interface was used, but without formula highlighting, and just one returned answer post viewed at a time.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Finding old answers to new math questions:the ARQMath lab at CLEF</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
	</analytic>
	<monogr>
		<title level="m">Intelligent Computer Mathematics -12th International Conference, CICM 2019</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">C</forename><surname>Kaliszyk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><forename type="middle">C</forename><surname>Brady</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Kohlhase</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Coen</surname></persName>
		</editor>
		<meeting><address><addrLine>Prague, Czech Republic</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">July 8-12, 2019. 2019</date>
			<biblScope unit="volume">11617</biblScope>
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A survey on retrieval of mathematical knowledge</title>
		<author>
			<persName><forename type="first">F</forename><surname>Guidi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Coen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CICM</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2015">9150. 2015</date>
			<biblScope unit="page" from="296" to="315" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Recognition and retrieval of mathematical expressions</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Blostein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal on Document Analysis and Recognition</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="331" to="357" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
	<note>IJDAR)</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">NTCIR-10 math pilot task overview</title>
		<author>
			<persName><forename type="first">A</forename><surname>Aizawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kohlhase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NTCIR</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">NTCIR-11 Math-2 task overview</title>
		<author>
			<persName><forename type="first">A</forename><surname>Aizawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kohlhase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schubotz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NTCIR</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="88" to="98" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Aizawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kohlhase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Topic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Davila</surname></persName>
		</author>
		<title level="m">NTCIR-12 MathIR task overview</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note>NTCIR</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Challenges of mathematical information retrieval in the NTCIR-11 Math Wikipedia Task</title>
		<author>
			<persName><forename type="first">M</forename><surname>Schubotz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Youssef</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Markl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">S</forename><surname>Cohl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR, ACM</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="951" to="954" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">SemEval-2019 Task 10: Math Question Answering</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hopkins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">Le</forename><surname>Bras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Petrescu-Prahova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Stanovsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hajishirzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Koncel-Kedziorski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th International Workshop on Semantic Evaluation</title>
				<meeting>the 13th International Workshop on Semantic Evaluation</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of ARQMath 2020: CLEF lab on answer retrieval for questions on math</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-58219-7_15</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-030-58219-7\_15" />
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -11th International Conference of the CLEF Association, CLEF 2020</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">A</forename><surname>Arampatzis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kanoulas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Tsikrika</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Vrochidis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Lioma</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Eickhoff</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">September 22-25, 2020. 2020</date>
			<biblScope unit="volume">12260</biblScope>
			<biblScope unit="page" from="169" to="193" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Layout and semantics: Combining representations for mathematical formula search</title>
		<author>
			<persName><forename type="first">K</forename><surname>Davila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1165" to="1168" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Tangent-CFT: An embedding model for mathematical formulas</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rohatgi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Giles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 ACM SIGIR International Conference on Theory of Information Retrieval (ICTIR)</title>
				<meeting>the 2019 ACM SIGIR International Conference on Theory of Information Retrieval (ICTIR)</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="11" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Structural similarity search for formulas using leaf-root paths in operator subtrees</title>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="116" to="129" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Terrier information retrieval platform</title>
		<author>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Amati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Plachouras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Johnson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="517" to="519" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">DPRL systems in the CLEF 2020 ARQMath lab</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2696/paper_223.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2020 -Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Eickhoff</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</editor>
		<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">September 22-25, 2020. 2696. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">On information retrieval metrics designed for evaluation with incomplete relevance assessments</title>
		<author>
			<persName><forename type="first">T</forename><surname>Sakai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kando</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="447" to="470" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Retrieval evaluation with incomplete information</title>
		<author>
			<persName><forename type="first">C</forename><surname>Buckley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Voorhees</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<meeting>the 27th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="25" to="32" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">The IIR evaluation model: a framework for evaluation of interactive information retrieval systems</title>
		<author>
			<persName><forename type="first">P</forename><surname>Borlund</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Research</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="8" to="11" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">MathSeer: A math-aware search interface with intuitive formula editing, reuse, and lookup</title>
		<author>
			<persName><forename type="first">G</forename><surname>Nishizawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Diaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dmello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="470" to="475" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
