<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Dowsing for Answers to Math Questions: Ongoing Viability of Traditional MathIR</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Yin</forename><forename type="middle">Ki</forename><surname>Ng</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">David R. Cheriton School of Computer Science</orgName>
								<orgName type="institution">University of Waterloo</orgName>
								<address>
									<postCode>N2L 3G1</postCode>
									<settlement>Waterloo</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dallas</forename><forename type="middle">J</forename><surname>Fraser</surname></persName>
							<email>dallas.fraser.waterloo@gmail.com</email>
							<affiliation key="aff1">
								<orgName type="institution">Knowledgehook Inc</orgName>
								<address>
									<addrLine>151 Charles St W</addrLine>
									<postCode>N2G 1H6</postCode>
									<settlement>Kitchener</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Besat</forename><surname>Kassaie</surname></persName>
							<email>bkassie@uwaterloo.ca</email>
							<affiliation key="aff0">
								<orgName type="department">David R. Cheriton School of Computer Science</orgName>
								<orgName type="institution">University of Waterloo</orgName>
								<address>
									<postCode>N2L 3G1</postCode>
									<settlement>Waterloo</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Frank</forename><forename type="middle">Wm</forename><surname>Tompa</surname></persName>
							<email>fwtompa@uwaterloo.ca</email>
							<affiliation key="aff0">
								<orgName type="department">David R. Cheriton School of Computer Science</orgName>
								<orgName type="institution">University of Waterloo</orgName>
								<address>
									<postCode>N2L 3G1</postCode>
									<settlement>Waterloo</settlement>
									<region>ON</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Dowsing for Answers to Math Questions: Ongoing Viability of Traditional MathIR</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">89A8CD91BE8313B12FE0875268F20FE2</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:45+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Community Question Answering (CQA)</term>
					<term>Mathematical Information Retrieval (MathIR)</term>
					<term>Symbol Layout Tree (SLT)</term>
					<term>Mathematics Stack Exchange (MSE)</term>
					<term>ARQMath Lab</term>
					<term>Tangent-L</term>
					<term>formula matching</term>
					<term>proximity</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We present our application of the math-aware search engine Tangent-L to the 2021 ARQMath Lab. This is a continuation of our MathDowsers submissions to last year's Lab, where we produced the best Task 1 participant run. Since then, we have improved the search engine's formula retrieval power by considering additional math features in the ranking function. This year, we also explore two approaches to incorporate proximity in evaluating the suitability of a document to be considered a match to a query.</p><p>For the 2021 ARQMath Lab, our primary run in Task 1 produces an nDCG ′ value of 0.434, which is nearly five points higher than that produced by the second-best participant run. An unsubmitted run, which corrects the setup of the primary run and preserves duplicate keyword terms during query term extraction, produces an even higher nDCG ′ of 0.462. Meanwhile, our primary run in Task 2 produces an nDCG ′ value of 0.552, which is the best automatic run and is comparable to the best participant run, a manual run from the Approach0 team.</p><p>The success of our runs continue to demonstrate that a traditional math information retrieval system remains a viable option for Community Question Answering specialized in the mathematical domain and for in-context formula retrieval.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The growing popularity of Community Question Answering (CQA) sites such as Math Stack Exchange 1 (MSE) and Math Overflow 2 demonstrates the need to find answers to mathematical questions, especially for questions posed in mathematical natural language. An effective question answering system capable of handling mathematical formulas and terminology would be of great interest to help serve this need.</p><p>The ARQMath Lab at CLEF 2021 <ref type="bibr" target="#b0">[1]</ref>, hereafter referenced as ARQMath-2, continues the previous year's Lab <ref type="bibr" target="#b1">[2]</ref> (ARQMath-1) by sponsoring an evaluation exercise centering around a CQA Task with questions involving math data. The Labs use a collection of questions and answers from MSE between 2010 and 2018 consisting of approximately 1.1 million question-posts and 1.4 million answer-posts. In this Lab series, Task 1 is the CQA Task in which participants are asked to return potential answers to unseen mathematical questions among existing answerposts in the collection. The closely related Task 2 considers formula retrieval in-context, in which formulas within questions serve as queries for matching relevant formulas from question-posts and answer-posts in the same collection.</p><p>In ARQMath-1, the Waterloo team of MathDowsers (Figure <ref type="figure" target="#fig_0">1</ref>) participated in Task 1, and our best run achieved an nDCG ′ value of 0.345 <ref type="bibr" target="#b2">[3]</ref>, which outperformed other participating systems <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>. Our approach was a three-stage Mathematics Information Retrieval (MathIR) system centered around the use of a math-aware search engine, Tangent-L <ref type="bibr" target="#b7">[8]</ref>: first, topics of mathematical questions were automatically transformed into formal queries consisting of keywords and formulas; then the formal queries were executed against a corpus of MSE questionanswer pairs by Tangent-L; finally, results were re-ranked based on a linear regression model trained on CQA metadata using mock relevance assessments. Submissions were made based on different configurations in each stage of the system, and the best run was produced without re-ranking, demonstrating success of a traditional math-aware query system in addressing a CQA task specialized in the mathematical domain.</p><p>For ARQMath-2, we participate again as the MathDowsers team for Task 1 and (for the first time) Task 2, with the goal to continue exploring the potential of a traditional math-aware query system in tackling both tasks. In particular, we are interested in further developing the formula matching capability of our core math-aware search engine Tangent-L, given that a satisfactory performance has been observed over formula-dependent questions in ARQMath-1 <ref type="bibr" target="#b8">[9]</ref>. With the empowered Tangent-L, we then refine our system for Task 1 and develop two baseline approaches for Task 2.</p><p>Our refinement is successful, with our primary run for Task 1 continuing to be the best participant run with respect to the primary measure nDCG ′3 . Our primary run for Task 2 turns out to be the most effective automatic run, essentially indistinguishable from the best participant run, a manual run from the Approach0 team. In this paper, we present:</p><p>• an updated Tangent-L with several avenues that improve its formula matching capability, 3 Normalized Discounted Cumulative Gain (nDCG) with unjudged documents removed </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Generated repetition tokens for the formula in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Token Type Tokens Generated Remark</head><p>Repeated symbols • a refinement of our system for mathematical answer retrieval with respect to query conversion and searching with Tangent-L, • two related approaches that are motivated by proximity, • for in-context formula retrieval, two simple baselines based on our developed system, • performance results for both Task 1 and Task 2 in ARQMath-2</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Improving Formula Matching with Tangent-L</head><p>Tangent-L is the cornerstone of our system for the tasks. It is a traditional math-aware query system built on the popular Lucene text search platform <ref type="bibr" target="#b9">[10]</ref>. During both index time and search time, it converts a formula into a bag of math tokens that each capture local characteristics of the Symbol Layout Tree (SLT) representation of a formula <ref type="bibr" target="#b10">[11]</ref>, so that mathematical documents can be matched against a query through text tokens and converted math tokens using a weighted BM25 + ranking <ref type="bibr" target="#b11">[12]</ref>.</p><p>The basic math tokens used by Tangent-L and the approach to weighting text against math tokens are described elsewhere <ref type="bibr" target="#b8">[9]</ref>. In this section, we describe improvements tested in this year's Lab.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Repeated Symbols</head><p>Repetitions of symbols are commonplace in a formula; for instance, 𝑥 repeats in the formula 𝑥 2 + 3 𝑥 + 𝑥, as does the operator +. Ideally, a search for either 𝑦 𝑥 − 𝑥 or 6𝑥 3 − 𝑦 + 𝑥 could match that formula because of the pattern of repetitions for 𝑥, and a search for 2𝑦 3 + 𝑦 + 5 could also match because of the repeated symbol +.</p><p>With this motivation, a new type of token-repetition tokens-is introduced into Tangent-L's formula representation to capture this characteristic. Repetition tokens are generated based on the relative positions of the repeated symbols in the formula's SLT representation. For every pair of repeated symbols:</p><p>1. if the pair of repeated symbols reside on the same root-to-leaf path of the SLT (that is, one is an ancestor of the other), then a repetition token {symbol, 𝑝} is generated, where 𝑝 represents the path between the repeated symbols; 2. otherwise, a repetition token {symbol, 𝑝 1 , 𝑝 2 } is generated where 𝑝 1 and 𝑝 2 represent the paths from the closest common ancestor in the SLT to each repeated symbol.</p><p>If a symbol repeats 𝑘 times where 𝑘 &gt; 1, (︀ 𝑘</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2</head><p>)︀ repetition tokens are generated for that symbol following the above procedure. For each of these tokens, an additional "location" token is generated with the augmentation of the path traversing from the root to the closest common ancestor of the pair. As such, a total of 2 • (︀ 𝑘</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2</head><p>)︀ repetition tokens are generated and indexed. Table <ref type="table" target="#tab_1">1</ref> shows the repetition tokens that would be indexed for the formula 𝑥 2 + 3 𝑥 + 𝑥 in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Revised Ranking Formula</head><p>With the introduction of repetition tokens, Tangent-L now generates three token types: text tokens, regular math tokens, and repetition tokens from documents or queries containing mathematical expressions. During a search, Tangent-L applies BM25 + ranking to the query terms and the document terms, using custom weights for each class of token as described here.</p><p>Let 𝑞 𝑡 be the set of text tokens, 𝑞 𝑚 be the set of regular math tokens, and 𝑞 𝑟 be the set of repetition tokens generated for the query terms. Let 𝑑 be a document represented by the set of all its indexed tokens. Then the revised ranking formula with the repetition tokens is:</p><formula xml:id="formula_0">BM25 + w (𝑞 𝑡 ∪ 𝑞 𝑚 ∪ 𝑞 𝑟 , 𝑑) = 𝛼 • 𝛾 • BM25 + (𝑞 𝑟 , 𝑑) + (1 − 𝛾) • BM25 + (𝑞 𝑚 , 𝑑) max(𝛾, 1 − 𝛾) + (1 − 𝛼) • BM25 + (𝑞 𝑡 , 𝑑)<label>(1)</label></formula><p>where 𝛼 and 𝛾 are parameters ranging from 0 to 1. The value of 𝛼 balances the weight of math features against keyword features, while the value of 𝛾 balances the weight of repetitions within math formulas against other math features. Both parameters can be tuned based on the target dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Formula Normalization</head><p>Mathematical expressions can be rewritten in numerous ways without altering their meaning. For example, 𝐴 + 𝐵 matches 𝐵 + 𝐴 semantically because of the commutative law. To accommodate such variability and increase recall, we equip Tangent-L with the ability to generate similar math features for two formulas with the same semantics. We consider the following five classes of semantic matches: The adjustment to handle the first two classes, Commutativity and Symmetry, are similar. Recall that originally Tangent-L generates a math token for each pair of adjacent symbols with their orders preserved. For example, two math tokens (𝐴, +, →) and (+, 𝐵, →) are generated for the expression 𝐴 + 𝐵, and two different math tokens (𝐵, +, →) and (+, 𝐴, →) are generated for the expression 𝐵 + 𝐴. In order for an exact match to take place for the two expressions, a simple adjustment to the math tokens is to ignore the order of a pair of adjacent symbols whenever commutative operators or symmetric relations are involved. With this approach, both expressions 𝐴 + 𝐵 and 𝐵 + 𝐴 generate the same pair of math tokens, (+, 𝐴, →) and (+, 𝐵, →), so that an exact match is made possible. <ref type="foot" target="#foot_0">4</ref>The next two classes, Alternative Notation and Operator Unification, can be easily accommodated by choosing a canonical symbol for each equivalence class of operators and consistently using only the canonical symbols in any math tokens generated as features.</p><p>The final class, Inequality Equivalence, can be handled by choosing a canonical symbol (for instance, choosing the symbol "≤" in preference to "≥") and then reversing the operands whenever necessary during math tokens generation. <ref type="foot" target="#foot_1">5</ref>For each of these five classes of semantic matches, Tangent-L provides a separate flag to control whether or not the class is to be supported, so that only those deemed to be advantageous are applied when math tokens are generated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Data Cleansing</head><p>For the ARQMath dataset, the original L A T E X formulas from the Math Stack Exchange collections are wrapped within an identifiable block (a span tag with class="math-container" and an id identifier), and the corresponding Presentation MathML representations are provided as separate files. Since the input to Tangent-L includes formulas encoded in Presentation MathML, its formula matching ability will be hindered when the quality of the MathML representation is poor or conversions from L A T E X are missing.</p><p>Thanks to the effort from the Lab organizers, coverage of the Presentation MathML for detected formulas has been increased from 92% for ARQMath-1 to over 99% for ARQMath-2 <ref type="bibr" target="#b12">[13]</ref>. However, further cleansing is still beneficial in preparation for search. We further improve the data cleansing in preparation for search as follows.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Correcting Conversion Errors:</head><p>The provided Presentation MathML, generated from L A T E X representation using LaTeXML<ref type="foot" target="#foot_2">6</ref> , contains conversion errors for formulas including either less-than "&lt;" or greater-than "&gt;"operators. In particular, when a L A T E X formula contains the operator "&lt;", it is first encoded as "&amp;lt;", but then erroneously escaped again to form"&amp;amp;lt;". This results in an erroneous encoding in Presentation MathML, as shown in Table <ref type="table" target="#tab_2">2</ref>.</p><p>As part of our data preparation, Presentation MathML encodings with doubly-escaped representations for "&lt;" and "&gt;" are recognized with regular expression matching and replaced by our own converted representations, improving 869,074 (∼ 3%) formulas.</p><p>Providing Missing Formula Identifiers: Approximately 10% of the annotated formulas in the postings are not correctly and completely captured, many missing their unique formula identifiers, as shown in Figure <ref type="figure" target="#fig_2">3</ref>. In this case, our program is unable to locate their Presentation MathML representation in the file provided by the Lab organizers.</p><p>Formulas such as those from Figure <ref type="figure" target="#fig_2">3</ref> are recognized as much as possible through regular expression matching for text within $ and $$ blocks. These are then checked against the formula file provided by the lab organizers to reverse-trace their formula-ids. As a result, our program is able to capture over 99% of the formulas, including the 10% that are improperly represented in math-container blocks without ids.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Task 1: Finding Answers to Math Questions</head><p>In Task 1, participants are given mathematical questions selected from MSE posts from either year 2019 (for ARQMath-1) or year 2020 (for ARQMath-2). Each question is formatted as a topic that contains a unique identifier, the title, the question body text, and the tags. Participant systems are asked to return the top-1000 potential answer-posts for each of the topics from the MSE collection. For ARQMath-2, we continue to use the three-stage system adopted for ARQMath-1 <ref type="bibr" target="#b8">[9]</ref>:</p><p>Stage 1 Conversion: Transform the input (a mathematical question posed on MSE) into a well-formulated query consisting of a bag of formulas and keywords.</p><p>Stage 2 Searching: Use Tangent-L, the math-aware search engine, to execute the formal query to find the best matches against an indexed document corpus created from the collection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Stage 3</head><p>Re-ranking: Re-order the best matches with a run-specific re-ranking model.</p><p>In this section, we describe various modifications we wished to explore. We first validate the benefits of each modification using the ARQMath-1 benchmark, and then we test them using the ARQMath-2 benchmark.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Conversion: Fine-tuning Keyword Extraction from Formulas</head><p>For ARQMath-1, our designed automated mechanism used to extract query keywords and formulas from the task topics was shown to be competitive with the human ability to select search terms <ref type="bibr" target="#b8">[9]</ref>, as it produces an result that is comparable to the manual set of query terms selected by the Lab organizers. For ARQMath-2, we fine-tune this automated mechanism using the ARQMath-1 benchmark for validation as follows:</p><p>1. Keywords within a formula representation are intentionally<ref type="foot" target="#foot_3">7</ref> retained and extracted, as a drop in nDCG ′ occurs if they are removed. For example, "mod" is a crucial keyword for topic 𝐴.7-Finding out the remainder of 11 10 −1 100 using modulus -but this word is present within a formula representation only and not anywhere else in the text. Similarly, "sin", "cos", "tan" can be extracted from \sin, \cos, \tan in formula representations after punctuation is removed. 2. Every term extracted by the automated mechanism should become part of the query, and their weight should be boosted naturally if they repeat. <ref type="foot" target="#foot_4">8</ref> On the other hand, restricting the number of keywords and formulas extracted from the mechanism (as we had hypothesized to be a possible improvement last year) does not show an improved result.</p><p>After fine-tuning the automated mechanism, results obtained for the ARQMath-1 benchmark can be observed to consistently outperform those obtained with the manual set of query terms, validating the potential of this mechanism. For ARQMath-2, we continue to use question-answer pairs as indexing units for the document corpus, as worse performance results for the ARQMath-1 benchmark if the content of the associated question is dropped and only text from each answer is indexed. In addition to the fields included for ARQMath-1, comments 9 associated with answers are also included. As a Table <ref type="table">3</ref> Various proximity measures <ref type="bibr" target="#b13">[14]</ref>, each of which can also be normalized by document length.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Searching: Enriching the Document Corpus</head><p>Span: length of the shortest document segment that covers all query term occurrences in a document, including repeated occurrences Normalized-Span: length of the shortest document segment that covers all query term occurrences in a document, including repeated occurrences, divided by the number of matched instances Min-Span: length of the shortest document segment that covers each matched query term at least once in a document Normalized-Min-Span: length of the shortest document segment that covers each matched query term at least once in a document, divided by the number of matched query terms Min-Distance: smallest distance value of all pairs of unique matched query terms Ave-Distance: average distance value of all pairs of unique matched query terms Max-Distance: largest value of all pairs of unique matched query terms</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Comparison of proximity measures on the ARQMath-1 benchmark for highly relevant (HR), relevant (R), partially relevant (PR), and non-relevant (NR) math answers, where Δ(𝑎, 𝑏) = prox(𝑎)−prox(𝑏) 0.5(prox(𝑎)+prox(𝑏)) . result, more formulas and more text words are available for matching. Figure <ref type="figure" target="#fig_3">4</ref> illustrates the fields indexed as part of each question-answer pair.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Δ(HR</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Re-ranking: Proximity</head><p>Whereas in ARQMath-1 we attempted re-ranking the retrieved answers from Tangent-L based on CQA metadata, for ARQMath-2 we investigate the possibility of re-ranking based on proximity. Proximity is a measure of distance between matched query terms as detailed in Table <ref type="table">3</ref>, which can be a strong signal for document relevancy. Following the experimental design used by Tao and Zhai <ref type="bibr" target="#b13">[14]</ref>, we measure the average proximity of search terms for highly relevant, relevant, partially relevant, and non-relevant documents in the ARQMath-1 benchmark. The experimental result is shown in Table <ref type="table">4</ref>. We observe strong signals from several measures that distinguish relevance with the correct order (marked in gradient orange), particularly for normalized-span which correctly orders all four levels of relevancy (a smaller normalized-span indicating a higher level of relevancy) without the need to be normalized by document length. Motivated by this finding, for ARQMath-2 we attempt re-ranking of the retrieved answers by Tangent-L in increasing order of normalized-span, breaking ties by a decreasing BM25 + score returned from Tangent-L.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Matching Formulas Holistically</head><p>Formula matching within Tangent-L is based on comparing a set of math tokens from the query to those from each document (Equation <ref type="formula" target="#formula_0">1</ref>). If we index a document that has multiple formulas, math tokens generated from all the formulas within the document are considered as a single unordered bag of terms. However, given the strong signal of proximity playing a role in document relevancy (Table <ref type="table">4</ref>), we hypothesize that matching each formula as a whole within a document, instead of matching math tokens irrespective of formulas that might scatter across a document, could produce a better result. <ref type="foot" target="#foot_6">10</ref> As such, as a post-experiment we design a holistic formula search as follows:</p><p>At preparation time, we first pre-build a formula corpus for Tangent-L that indexes all visually distinct formulas in the MSE dataset, each as a separate document with the formula's visual-id serving as a key. We define the formula similarity between two formulas to be the normalized BM25 + score for one formula when the other formula acts as a query. When indexing the question-answer corpus, rather than replacing each formula within the document by the set of math tokens generated for that formula, we represent each formula by a single holistic formula token that contains the formula's visual-id (that is, its key from the formula corpus). At query time, we first search for each query formula in the formula corpus and then replace the formula text in the query by the keys of the top-𝑘 most similar formulas, thus changing the query to search for those visual-ids (as well as whatever keywords are also part of the query, of course). Finally, the ranking formula for documents is revised to weight each match of a formula id by its formula similarity with respect to the original query formula.</p><p>In the following subsections, we describe these ideas in greater detail.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.1.">Formula Corpus</head><p>The formula corpus is built by extracting all visually distinct formulas from the document corpus described in Section 3.2-including formulas found within questions, answers, and comment posts. Each formula in this corpus is associated with the formula's visual-id, which serves as a key. The resulting corpus contains 8,595,899 out of 9,329,274 (∼ 92%) visually distinct formulas and is indexed by Tangent-L under the setup described in Section 2, each formula being considered as a document.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.2.">Normalized Formula Similarity</head><p>We define "formula similarity" as follows: Let 𝑓 𝑞 be an arbitrary formula used as a query, 𝐹 be the set of formulas in the formula corpus, and 𝑓 ∈ 𝐹 . Let RawScore(𝑓 𝑞 , 𝑓 ) represents the score obtained for formula 𝑓 when the query is 𝑓 𝑞 , using the following definition:</p><formula xml:id="formula_1">RawScore(𝑓 𝑞 , 𝑓 ) = (1 − 𝛾) • BM25 + (𝑞 𝑚 , 𝑓 ) + 𝛾 • BM25 + (𝑞 𝑟 , 𝑓 )<label>(2)</label></formula><p>where 𝑞 𝑚 is the set of regular math tokens and 𝑞 𝑟 is the set of repetition tokens in a query formula 𝑓 𝑞 . As in Equation 1, 0 ≤ 𝛾 ≤ 1 balances the weight of repetition tokens against regular math tokens.</p><p>The Normalized Formula Similarity of 𝑓 with respect to 𝑓 𝑞 is:</p><formula xml:id="formula_2">𝑁 (𝑓, 𝑓 𝑞 ) = RawScore(𝑓 𝑞 , 𝑓 ) max 𝜙∈𝐹 RawScore(𝑓 𝑞 , 𝜙)<label>(3)</label></formula><p>The value of 𝑁 (𝑓, 𝑓 𝑞 ) is in the range [0,1] and represents how well the query formula 𝑓 𝑞 is matched by 𝑓 relative to other formulas within the formula corpus.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.3.">Holistic formula token</head><p>A holistic formula token is a placeholder token that incorporates the formula's visual-id. Formulas in a question-answer document are replaced by their holistic formula tokens only, so that when searching the question-answer corpus, formulas can only be matched as a whole.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.4.">Ranking for Holistic Search</head><p>Let 𝑞 𝑡 be the set of keyword tokens and 𝑞 𝑓 be the set of query formulas. Let 𝑓 𝑞 ∈ 𝑞 𝑓 be a query formula and let 𝑆 𝑘 (𝑓 𝑞 ) be the set of keys for the top-𝜅 most similar formulas with respect to 𝑓 𝑞 , determined by Normalized Formula Similarity. Let 𝑑 be a document represented by the set of all its indexed tokens. When searching the document corpus, we adopt the following variant of BM25 + :</p><formula xml:id="formula_3">BM25 + w (𝑞 𝑡 ∪ 𝑞 𝑓 , 𝑑) = (1 − 𝛼) • BM25 + (𝑞 𝑡 , 𝑑) + 𝛼 • BM25 + (𝑞 𝑓 , 𝑑)<label>(4)</label></formula><p>and</p><formula xml:id="formula_4">BM25 + (𝑞 𝑓 , 𝑑) = ∑︁ 𝑓𝑞 ∈𝑞 𝑓 ∑︁ 𝑓 ∈(𝑑 ∩ 𝑆 𝑘 (𝑓𝑞 )) (︃ 𝑁 (𝑓, 𝑓𝑞) • (𝑘 + 1)tf 𝑓 𝑘 (︁ 1.0 − 𝑏 + 𝑏 |𝑑| 𝑑 )︁ + tf 𝑓 + 𝛿 )︃ log (︃ |𝐷| + 1 |𝐷 𝑓 | )︃<label>(5)</label></formula><p>where, as in Equation <ref type="formula" target="#formula_0">1</ref>, 0 ≤ 𝛼 ≤ 1 balances math features against keyword features. 11</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Task 1: Runs and Result</head><p>Parameter settings are chosen based on testing with the ARQMath-1 benchmark. For ARQMath-2, we prepared four automatic runs: 11 As usual for BM25 + <ref type="bibr" target="#b14">[15]</ref>, 𝑘, 𝑏, and 𝛿 are constants (following common practice, chosen to be 1.2, 0.75, and 1, respectively); tf 𝑓 is the number of occurrences of formula 𝑓 in 𝑑; |𝑑| is the total number of terms in 𝑑; 𝑑 = ∑︀ 𝑑∈𝐷 |𝑑| |𝐷| is the average document length; and |𝐷 𝑓 | is the number of documents in 𝐷 containing formula 𝑓 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 5</head><p>The setup for the primary run for ARQMath-2.</p><p>Repeated Symbols (Sect. 2.1) Repetition tokens are adopted. Revised Ranking Formula (Sect. 2.2) In Equation <ref type="formula" target="#formula_0">1</ref>, 𝛼 = 0.25 and 𝛾 = 0.1.</p><p>Formula Normalization (Sect. Query terms are (unintentionally) de-duplicated.</p><p>primary: A submitted run with most of the presumably best setup, based on tests on the ARQMath-1 benchmark, as described in Table <ref type="table">5</ref>.</p><p>proximityReRank: A submitted run based on Section 3.3. This uses the same setup as the primary run, but the top-1000 matches are subsequently re-ranked by proximity, using normalized span as the proximity measure.</p><p>holisticSearch: A post-experiment run that matches formulas holistically based on Section 3.4. When searching in the formula corpus, 𝛾 is set to 0.1 in Equation <ref type="formula" target="#formula_1">2</ref>and when searching in the document corpus, 𝛼 is set to 0.5 in Equation <ref type="formula" target="#formula_3">4</ref>and 𝜅 is set to be 300.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>duplicateTerms:</head><p>A post-experiment run sharing the same setup as the primary run, except that duplicate query terms are preserved as described in Section 3.1.</p><p>The results of these runs for ARQMath-2 are shown in Table <ref type="table">6</ref>, together with the baseline runs and our submissions from last year over the ARQMath-1 benchmark. In general, after parameter selection based on the ARQMath-1 benchmark, our updated system produces results that have a significant improvement compared with those from last year's system over the ARQMAth-1 topics. For instance, our primary setup evaluated over the ARQMath-1 benchmark achieves an nDCG ′ score of 0.433, which is nearly a 10-point gain over the nDCG ′ score of 0.345 produced by our best participant run (alpha05-noR) last year.</p><p>This parameter selection based on the ARQMath-1 benchmark helps our updated system to achieve equally good results for the new set of math topics in ARQMath-2. Our primary run produces an nDCG ′ of 0.434, which remains the best run among all participants <ref type="bibr" target="#b12">[13]</ref>. The unsubmitted run duplicateTerms, which corrects an oversight in the primary run and therefore reflects our intended "best" setup, scores even higher, with an nDCG ′ of 0.462.</p><p>The duplicateTerms run also has the highest values for the ARQMath-2 benchmark in all other evaluation measures, with the exception of P ′ @10 for the baseline run Linked MSE posts (which uses human-built links that were not available to participating teams <ref type="bibr" target="#b12">[13]</ref>). With a closer look to the effectiveness breakdown by topic category in Table <ref type="table" target="#tab_5">7</ref>, we observe that this run has a strong performance for Formula-dependent topics, Proof-like topics, and topics of Low-level difficulty. In spite of a different set of math topics being evaluated, these observed strengths are similar to the observed strengths of our best participant run last year <ref type="bibr" target="#b2">[3]</ref>.</p><p>On the other hand, our submitted alternative run proximityReRank, which tries to re-rank the results using the proximity signal Normalized-Span, does not perform well. For the ARQMath-1</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 6</head><p>Task 1: Evaluation of the MathDowsers runs and the baseline runs in ARQMath-2, compared with that over the ARQMath-1 benchmark. Parentheses indicate a result from an approach using privately held data not available to participants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ARQMath-1 (77 Topics)</head><p>ARQMath-2 (71 Topics) nDCG ′ MAP ′ † P ′ @10 † bpref † nDCG ′ MAP ′ † P ′ @10 † bpref † benchmark, this run shows a 6-point loss compared to the primary run (0.373 vs 0.433) and the loss is enlarged to nearly 10 points in ARQMath-2 (0.335 vs 0.434), indicating an unsatisfactory re-ranking. It seems that even for a measure that shows a strong signal for proximity in Table <ref type="table">4</ref>, the separation among documents based on proximity might be inadequate to reflect relevance. Finally, our unsubmitted run holisticSearch, which is an approach also motivated by proximity, performs fairly well. Compared to the primary run, the nDCG ′ score for the holisticSearch run shows a 3-point loss over the ARQMath-1 benchmark (0.405 vs 0.433) and similarly a 2-point loss in ARQMath-2 (0.414 vs 0.434). Notably, this run outperforms all other runs submitted by participants in ARQMath-2 and outperforms our primary run in the 𝑃 ′ @10 and bpref measures. However, this run is outperformed by the unsubmitted duplicateTerms run in all evaluation measures (with nearly a 5-point loss (0.414 vs 0.462 for nDCG ′ ), suggesting room for improvement for this approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Baselines</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Task 2: In-context Formula Retrieval</head><p>For Task 2, participants are asked to retrieve the top matching formulas, together with their associated posts, for each topic formula chosen from the set of topics used for Task 1. Relevancy of a retrieved formula is evaluated in context: both the associated post of a retrieved formula and the associated topic content of the topic formula are presented to the assessors for evaluation. Assessments are then aggregated so that each visually distinct formula is judged to be relevant if any of the corresponding formula occurrences are deemed to be relevant. The performance of a system is then determined by its performance with respect to visually distinct formulas only. For ARQMath-2, we propose two simple approaches that re-use two major components created for Task 1:</p><p>1. the Formula Corpus of all visually distinct formulas, as described in Section 3.4.1; 2. the results from Task 1 Answer-Ranking of the top 10,000 answer-posts for each topic, run with the primary setup as detailed in Table <ref type="table">5</ref>.</p><p>The rest of this section describes our two approaches built on these components.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Formula-centric: Selecting Visually Matching Formulas</head><p>The first straightforward approach is formula-centric, relying on Tangent-L's internal formula matching capability to find the matching formulas. To create a list of matching formulas for a topic, we first search for matches to the topic formula in the formula corpus of all visually distinct formulas. This gives us a ranking 𝑅 of visually distinct formulas. We then expand each element of 𝑅 with its set of formula occurrences: formulas that have the same visual-id but appearing in different posts. <ref type="foot" target="#foot_7">12</ref> We refer to a set of formula occurrences having the same visual-id as a visual group. The selection of formula occurrences to return is then governed by the rank of their associated posts in the answer retrieval task. In particular, 1. Formulas within the same visual group are ranked in the same order as the ranking of their associated posts in Task 1 for the corresponding topic. If the associated posts of formulas are question-posts that are not associated with any answer from Task 1, the formulas are assigned the lowest ranking. Finally, the lexical order of formula-ids is used to break ties. 2. For each of the top-20 visually distinct formulas in 𝑅, we select the top five formulas from its visual group (or all formulas in the visual group if there are fewer than five); for the remainder, we select the top formula only (if any have associated question or answer posts). 3. Sequentially considering the formulas in 𝑅 in order, selected formula occurrences from each visual group are appended to the final list of matching formulas until 1000 formula occurrences are selected in total.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Document-centric: Screening Formulas from Matched Documents</head><p>The second straightforward approach is document-centric, relying more on the results from the answer retrieval task. Based on the answer-ranking from Task 1, the final list of matching formula occurrences is selected from the answers as follows:</p><p>1.</p><p>For each matched answer-post for the corresponding topic in Task 1, we retrieve its question-answer document from the document corpus. If the document contains only one formula, that formula is selected. Otherwise, each formula from the document is mapped to its visual group, and its Normalized Similarity Score (Equation <ref type="formula" target="#formula_2">3</ref>) with respect to the topic formula is computed using 𝛾 = 0.1 in Equation 2 (but see below). Formulas having a score less than a threshold of 0.8 are screened out, and the rest are preserved and ranked accordingly. 2. Following the original answer-ranking, preserved formulas from each question-answer pair are appended to the final list until 1000 formulas are selected in total.</p><p>Formulas in an answer-post might correspond to visually distinct formulas any where in the formula corpus, but it is highly inefficient to compute the Normalized Similarity Score for every formula in the formula corpus, which requires retrieving over 8.5 million RawScores using Tangent-L. Therefore, for each topic, formulas in answer-posts that are not within the top 10,000 most similar formulas to the query formula are assigned a score of 0 and therefore screened out.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Task 2: Runs and Result</head><p>For ARQMath-2, we include two automatic runs: formulaBase: A submitted run selecting visually matching formulas as in Section 4.1; docBase: A submitted run selecting formulas from matched documents as in Section 4.2. The result of both runs in ARQMath-2 are shown in Table <ref type="table" target="#tab_6">8</ref>, together with the baseline run and the best participant runs for the ARQMath-1 and ARQMath-2 benchmarks. Our primary run formulaBase, with parameter selection based on the ARQMath-1 benchmark, achieves a very close performance to the best participant run Tangent-CFTED produced from the DPRL team last year (0.562 vs 0.563). However, on the ARQMath-1 benchmark, it does not perform as well as the ltrall run submitted this year by the DPRL team, having a 17-point loss on nDCG ′ over the same set of math topics (0.562 vs 0.735).</p><p>On the ARQMath-2 benchmark, however, with a new set of math topics, our primary run formulaBase performs approximately as well, with an nDCG ′ score of 0.552. This score is the best among all automatic runs, and it is almost indistinguishable from the best participant run P300 from the Approach0 team, which is a manual run. Notably, on the ARQMath-2 benchmark, it outperforms the ltrall run from the DPRL team by over 10 points (0.552 vs 0.445).</p><p>On the other hand, our alternative run docBase does not perform as well as expected. For the ARQMath-1 benchmark, this run shows nearly a 16-point loss with respect to our primary run (0.404 vs 0.562) and nearly a 12-point loss (0.433 vs 0.552) for the ARQMath-2 in terms of nDCG ′ . This run also achieves lower scores in all other evaluation measures, suggesting that simply selecting formulas from matching documents does not work well.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Efficiency</head><p>The machines used for our experiments have the following specifications: Note that data and index sizes show the values reported by the du command on Linux, which measures disk space usage based on blocks; thus the many small documents in the formula corpus require much more disk space than might be expected. (In fact, the total size of the data in the formula corpus is only 9.2 GB.)</p><p>Runs for ARQMath-2 were executed on Machine B with the following average, minimum, and maximum query times per topic as follows: The proximityReRank run uses Machine A to rerank the output from the primary run, thus requiring first the time shown for the primary run on Machine B and then an additional 8 hours to re-rank all topics on Machine A.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions and Further Work</head><p>We conclude that a traditional math-aware search system continues to be an efficient and effective approach to tackle the CQA task, which is proven by producing the best participant run in Task 1 again this year. In particular, a significant boost in effectiveness for Task 1 can be observed on both years' math topics after parameter selection based on tests on the ARQMath-1 benchmark. The best result is achieved through several aspects of improvement of the formula matching capability of Tangent-L, demonstrating the competitiveness of this math-aware search engine in handling text and mathematical notations together.</p><p>We also develop a simple but strong baseline for the in-context formula retrieval task. Being the best automatic run and competitive with the best participant run, our formula-centric run demonstrates again the strong formula matching ability of Tangent-L.</p><p>Nevertheless, several aspects of our runs turn out to be somewhat disappointing again. In the CQA task, we explore the incorporation of proximity in two approaches and the result does not improve effectiveness over using a bag-of-terms approach:</p><p>Proximity Re-Ranking: Re-ranking based on proximity is unsatisfactory, despite some proximity difference being observed based on the relevancy of judged documents. Perhaps proximity is a more important measure when the BM25 + score is low, and therefore it needs to be incorporated into the initial retrieval <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b15">16]</ref> rather than used for re-ranking. Alternatively, despite the percentage differences observed, the actual differences might be too small to serve as a reliable signal of relevance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Matching Formulas Holistically:</head><p>The proposed method to match formulas holistically shows some promise but does not perform as well as matching based on math tokens. Perhaps Equation 5 can be improved to make better use of the formula similarity scores returned from the formula corpus. Improvements here might also provide insights into further improving our formula-centric approach in Task 2.</p><p>Additionally, our proposed document-centric baseline for the in-context formula retrieval task, which selects formulas from top matching math answers, does not perform as well as expected given our strong result in the answer retrieval task. Investigation into the distribution of matching formulas among the top relevant answers might be helpful in further exploring this simple tactic for the task.</p><p>All in all, while our updated system with Tangent-L continues to excel in both tasks, there is still a huge room for improvement in how we might use the document relevancy signals observed from the ARQMath-1 benchmark to propose new approaches that might further improve effectiveness. In retrospect, approaches that we attempted through re-ranking did not benefit sufficiently from the raw signals obtained from the ARQMath-1 benchmark. With the additional new evaluation data available from the ARQMath-2 benchmark, we expect to gain better insights, and we are excited to continue exploring question answering for the mathematical domain.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Researcher dowsing for answers to math queries.</figDesc><graphic coords="2,213.06,84.19,166.68,117.14" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Symbol Layout Tree for 𝑥 2 + 3 𝑥 + 𝑥 with repetitions highlighted.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Partial text from an answer post (post-id 2653) including "math-container" blocks without "id" attributes, even though the corresponding formulas are included in the formula file with formula-ids from 2285 to 2296.</figDesc><graphic coords="7,89.29,84.19,416.71,80.69" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: An illustration of the revised indexing unit to create the document corpus. Each document is an HTML file containing a question-answer pair and its associated information.</figDesc><graphic coords="8,223.47,316.80,145.84,175.97" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>2 . 3 )</head><label>23</label><figDesc>Only semantic matches of Commutativity is supported. Data Cleansing (Sect. 2.4) Recognition of Presentation MathML is improved. Document Corpus (Sect. 3.2) Comments from answers are added to the indexing unit. Query Keyword Extraction (Sect. 3.1) Keywords within a formula representation are retained.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>1 .</head><label>1</label><figDesc>Commutativity: 𝐴 + 𝐵 should match 𝐵 + 𝐴 2. Symmetry: 𝐴 = 𝐵 should match 𝐵 = 𝐴 3. Alternative Notation: 𝐴 × 𝐵 should match 𝐴 𝐵, and 𝐴 ≯ 𝐵 should match 𝐴 ≤ 𝐵 4. Operator Unification: 𝐴 ≺ 𝐵 should match 𝐴 &lt; 𝐵 5. Inequality Equivalence: 𝐴 ≥ 𝐵 should match 𝐵 ≤ 𝐴 and simple adjustments are applied to Tangent L's regular math tokens to support these semantic matches.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Erroneous Presentation MathML for the formula "0.999... &lt; 1" (formula id 382).</figDesc><table><row><cell>Expected Presentation MathML</cell><cell>Erroneous Presentation MathML Provided</cell></row><row><cell>&lt;mrow&gt;</cell><cell>&lt;mrow&gt;</cell></row><row><cell>&lt;mrow&gt;</cell><cell>&lt;mrow&gt;</cell></row><row><cell>&lt;mn&gt;0.9999&lt;/mn&gt;</cell><cell>&lt;mn&gt;0.9999&lt;/mn&gt;</cell></row><row><cell>&lt;mi mathvariant="normal"&gt;...&lt;/mi&gt;</cell><cell>&lt;mo&gt;&lt;/mo&gt;</cell></row><row><cell>&lt;mo&gt;&amp;lt;&lt;/mo&gt;</cell><cell>&lt;mi mathvariant="normal"&gt;...&lt;/mi&gt;</cell></row><row><cell>&lt;mn&gt;1&lt;/mn&gt;</cell><cell>&lt;mo&gt;&lt;/mo&gt;</cell></row><row><cell>&lt;/mrow&gt;</cell><cell>&lt;mi mathvariant="normal"&gt;&amp;amp;&lt;/mi&gt;</cell></row><row><cell>&lt;mrow&gt;</cell><cell>&lt;mo&gt;&lt;/mo&gt;</cell></row><row><cell></cell><cell>&lt;mi&gt;l&lt;/mi&gt;</cell></row><row><cell></cell><cell>&lt;mo&gt;&lt;/mo&gt;</cell></row><row><cell></cell><cell>&lt;mi&gt;t&lt;/mi&gt;</cell></row><row><cell></cell><cell>&lt;/mrow&gt;</cell></row><row><cell></cell><cell>&lt;mo&gt;;&lt;/mo&gt;</cell></row><row><cell></cell><cell>&lt;mn&gt;1&lt;/mn&gt;</cell></row><row><cell></cell><cell>&lt;/mrow&gt;</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 7</head><label>7</label><figDesc>Category performance of the duplicateTerms run in ARQMath-2. The better performance measure for each sub-category and each evaluation measure is highlighted in bold.</figDesc><table><row><cell></cell><cell>Topic</cell><cell>duplicateTerms</cell></row><row><cell></cell><cell cols="2">Count nDCG ′ MAP ′ P ′ @10 bpref</cell></row><row><cell>Overall</cell><cell>71</cell><cell>0.462 0.187 0.241 0.163</cell></row><row><cell>Dependency</cell><cell></cell><cell></cell></row><row><cell>Text</cell><cell>10</cell><cell>0.423 0.158 0.260 0.142</cell></row><row><cell>Formula</cell><cell>21</cell><cell>0.516 0.235 0.319 0.204</cell></row><row><cell>Both</cell><cell>40</cell><cell>0.443 0.169 0.195 0.146</cell></row><row><cell>Topic Type</cell><cell></cell><cell></cell></row><row><cell>Calculation</cell><cell>25</cell><cell>0.455 0.189 0.200 0.165</cell></row><row><cell>Concept</cell><cell>19</cell><cell>0.429 0.160 0.232 0.137</cell></row><row><cell>Proof</cell><cell>27</cell><cell>0.492 0.204 0.285 0.178</cell></row><row><cell>Difficulty</cell><cell></cell><cell></cell></row><row><cell>Low</cell><cell>32</cell><cell>0.509 0.216 0.300 0.199</cell></row><row><cell>Medium</cell><cell>20</cell><cell>0.383 0.116 0.150 0.098</cell></row><row><cell>Hard</cell><cell>19</cell><cell>0.466 0.213 0.237 0.169</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 8</head><label>8</label><figDesc>Task 2: Evaluation of MathDowsers runs, the best participant runs, and baseline runs in ARQMath-2. MAP ′ † P ′ @10 † bpref † nDCG ′ MAP ′ † P ′ @10 † bpref †</figDesc><table><row><cell></cell><cell></cell><cell>ARQMath-1</cell><cell></cell><cell cols="2">ARQMath-2</cell></row><row><cell cols="3">nDCG ′ Baselines</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Tangent-S</cell><cell></cell><cell>0.691 0.446 0.453 0.412</cell><cell cols="4">0.492 0.272 0.419 0.290</cell></row><row><cell>MathDowsers</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>formulaBase</cell><cell cols="2">¶ 0.562 0.370 0.447 0.374</cell><cell cols="4">0.552 0.333 0.450 0.348</cell></row><row><cell>docBase</cell><cell>*</cell><cell>0.404 0.251 0.386 0.275</cell><cell cols="4">0.433 0.257 0.359 0.291</cell></row><row><cell>Best Participant Run</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Approach0-P300</cell><cell cols="2">*M 0.507 0.342 0.441 0.343</cell><cell cols="4">0.555 0.361 0.488 0.362</cell></row><row><cell>DPRL-ltrall</cell><cell cols="2">¶ 0.738 0.525 0.542 0.495</cell><cell cols="4">0.445 0.216 0.333 0.228</cell></row><row><cell cols="3">Best Participant Run (year 2020)</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">DPRL-Tangent-CFTED *</cell><cell>0.563 0.388 0.436 0.372</cell><cell>-</cell><cell>-</cell><cell>-</cell><cell>-</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table><note>¶ submitted primary run * submitted alternate run M manual run † using H+M binarization</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_0">Our simple implementation suffers from the fact that math tokens handle only a pair of adjacent symbols at a time. For a longer expression, such as 𝐴 + 𝐵 ×</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_1">, the overly simplistic approach generates the same set of math tokens as the expression 𝐵 + 𝐴 × 5, failing to consider the priority of operators. nevertheless, we have chosen to take this approach because correct treatment requires that the math formulas be parsed properly, which is difficult to achieve when the input of Tangent-L-Presentation MathML-captures layout only.<ref type="bibr" target="#b4">5</ref> Similar to commutative operations and symmetric relations, the reversion of operands is implemented simplistically over a pair of adjacent symbols at a time. Thus the generated set of math tokens might equally well represent a semantically distinct formula.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_2">https://dlmf.nist.gov/LaTeXML</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_3">Keywords were not intended to be extracted from within formula representations in the original design for ARQMath-1, but turned out to be a valuable "mistake" that helped boost performance.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_4">In the submission for ARQMath-1, duplicate terms were extracted, but their weights were not boosted accordingly because of an oversight in our implementation.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_5"><ref type="bibr" target="#b8">9</ref> When extracting the comments, the file Comments.V.1.0.xml is used instead of the more recently released Comments.V.1.2.xml because the former contains approximately three times as many comments as the latter. Note, however, that the former file contains more "noise" that requires cleansing as discussed in Section 2.4.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_6">Note, however, that this ignores proximity among keywords and between keywords and formulas.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_7">Only question-posts and answer-posts are of concern in the task, so any returned formulas from commentposts are ignored.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research has been funded by the Waterloo-Huawei Joint Innovation Lab and NSERC, the Natural Science and Engineering Research Council of Canada. The NTCIR Math-IR dataset used for earlier benchmarks and as a source of relevant keywords was made available through an agreement with the National Institute of Informatics.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>(F. Wm. Tompa) https://www.linkedin.com/in/kiking0501/ (Y. K. Ng); https://uwaterloo.ca/scholar/bkassaie/home (B. Kassaie); http://www.cs.uwaterloo.ca/~fwtompa/ (F. Wm. Tompa) 0000-0002-1907-9535 (F. Wm. Tompa</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of ARQMath-2</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Second CLEF lab on answer retrieval for questions on math</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="volume">12880</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2021</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">): CLEF lab on answer retrieval for questions on math</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Dowsing for math answers with Tangent-L</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">K</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Fraser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kassaie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Labahn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Marzouk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">W</forename><surname>Tompa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">DPRL Systems in the CLEF 2020 ARQMath Lab</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Three is Better than One Ensembling Math Information Retrieval Systems</title>
		<author>
			<persName><forename type="first">V</forename><surname>Novotný</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sojka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Štefánik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lupták</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">PSU at CLEF-2020 ARQMath Track: Unsupervised Re-ranking using Pretraining</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rohatgi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Giles</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">ARQMath Lab: An Incubator for Semantic Formula Search in zbMATH Open?</title>
		<author>
			<persName><forename type="first">P</forename><surname>Scharpf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schubotz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Greiner-Petter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ostendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Teschke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Gipp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CLEF 2020</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Choosing math features for BM25 ranking with Tangent-L</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Fraser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">W</forename><surname>Tompa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">DocEng</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page">10</biblScope>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Dowsing for math answers</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">K</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Fraser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kassaie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">W</forename><surname>Tompa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CLEF 2021</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">12880</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Białecki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Muir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ingersoll</surname></persName>
		</author>
		<title level="m">SIGIR 2012 Workshop on Open Source Information Retrieval</title>
				<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="17" to="24" />
		</imprint>
	</monogr>
	<note>Apache Lucene 4</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Recognition and retrieval of mathematical expressions</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Blostein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Int. J. Document Anal. Recognit</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="331" to="357" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Lower-bounding term frequency normalization</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lv</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CIKM&apos;11</title>
				<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="7" to="16" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Advancing math-aware search: The ARQMath-2 lab at CLEF</title>
		<author>
			<persName><forename type="first">B</forename><surname>Mansouri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">W</forename><surname>Oard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zanibbi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECIR 2021</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="volume">12657</biblScope>
			<biblScope unit="page" from="631" to="638" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">An exploration of proximity measures in information retrieval</title>
		<author>
			<persName><forename type="first">T</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGIR 2007</title>
				<imprint>
			<date type="published" when="2007">2007</date>
			<biblScope unit="page" from="295" to="302" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">The probabilistic relevance framework: BM25 and beyond</title>
		<author>
			<persName><forename type="first">S</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zaragoza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and Trends in Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="333" to="389" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Term proximity scoring for keyword-based retrieval systems</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Rasolofo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Savoy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th European Conference on IR Research (ECIR 2003)</title>
				<meeting>the 27th European Conference on IR Research (ECIR 2003)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="volume">2633</biblScope>
			<biblScope unit="page" from="207" to="218" />
		</imprint>
	</monogr>
	<note>Advances in Information Retrieval</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
