<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Extending Monolingual Asymmetric Semantic Search Models for Multilingual Query Processing using Knowledge Distillation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Iryna</forename><surname>Yurchuk</surname></persName>
							<email>a.yurchuk@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Taras Shevchenko National University of Kyiv</orgName>
								<address>
									<addrLine>Volodymyrska Street, 60</addrLine>
									<postCode>01033</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Danylo</forename><surname>Boiko</surname></persName>
							<email>danylo.boiko@knu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">Taras Shevchenko National University of Kyiv</orgName>
								<address>
									<addrLine>Volodymyrska Street, 60</addrLine>
									<postCode>01033</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Extending Monolingual Asymmetric Semantic Search Models for Multilingual Query Processing using Knowledge Distillation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">509FF72F1CBA4609D4C609806384E9A5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:49+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>asymmetric semantic search</term>
					<term>multilingual embedding</term>
					<term>knowledge distillation1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Semantic search is a key task in today s world, where the amount of data is growing rapidly. In this work, the main role is devoted to cases when long answers must be found to a short query (known as asymmetric search). The teacher model with 30,522 tokens of vocabulary and the student model with 119,547 tokens of vocabulary are basic for training a multilingual asymmetric semantic search model using multilingual knowledge distillation. The authors used the reciprocal rank (RR), the mean average precision (MAP), and the normalized discounted cumulative gain (NDCG) to evaluate the obtained model.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Semantic search is a set of search algorithms that work based on understanding the meaning of text. This approach effectively handles synonyms, abbreviations, and spelling errors, unlike keyword search engines that rely on exact lexical matches to find documents. It is useful for grading and assessing academic work <ref type="bibr" target="#b0">[1]</ref>, integrating search functionalities on e-commerce <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>, information retrieval in the petrochemical sector using a fusion of video transcript data with other data sources <ref type="bibr" target="#b3">[4]</ref>, helping Human Resources employees to target relevant people for their events and trainings <ref type="bibr" target="#b4">[5]</ref>, performing social search <ref type="bibr" target="#b5">[6]</ref>, IoT systems <ref type="bibr" target="#b6">[7]</ref>, and other different fields <ref type="bibr" target="#b7">[8]</ref>.</p><p>Ontologies and knowledge graphs are classical approaches to the design and implementation of semantic search systems, but recently these approaches have been enriched or replaced by statistical algorithms <ref type="bibr" target="#b8">[9]</ref> and AI techniques <ref type="bibr" target="#b9">[10]</ref> for query expansion, generating question-precise training statistics, semi-supervised gaining knowledge, etc.</p><p>This work aims to study the possibility of using the bi-encoder architecture to implement asymmetric semantic search when storage contains indexed data in English with the necessary to search information in Ukrainian.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background and Algorithms</head><p>The intention behind semantic search is to transform passages into a multidimensional vector space. During the search, the query is similarly embedded in this vector space, which allows to identify the 2 desired quantity of most relevant matches. This approach ensures that even if the wording is different, the meaning remains the same and the system can provide accurate retrievals (Figure <ref type="figure" target="#fig_0">1</ref>). There are two main types of semantic search that work with different types of data. For symmetric search, queries and passages in the corpus have approximately the same length and content. In turn, asymmetric search typically uses short queries (e.g., a question or a few keywords) and longer passages answering those queries.</p><p>Nowadays, a lot of trained asymmetric semantic search models based on both bi-encoder and cross-encoder architectures <ref type="bibr" target="#b10">[11]</ref> are available for English. Bi-encoder models are efficient for largescale retrieval due to their ability to use precomputed passage embeddings from storage, making them ideal for speed-critical tasks, while the cross-encoder models are known for higher accuracy by directly capturing similarity between query and passages, suited for accuracy-sensitive scenarios like reranking where computational cost is less of a concern.</p><p>To use asymmetric semantic search for less common languages or even a mix of them, we need to train new models, which will require a lot of data and computational power. Fortunately, there is a way to facilitate this using multilingual knowledge distillation <ref type="bibr" target="#b11">[12]</ref> (Figure <ref type="figure" target="#fig_1">2</ref>). This approach requires a teacher model for the source language and a set of pairs (each pair includes a sentence in the source language and its translation). A new student model attempts to approximate the output of a teacher model for both source and target sentences using the mean squared error (MSE) loss. The student model could have the structure and weights of the teacher model, or it could be a different network architecture since the student model learns representation of the teacher model. This allows the student model to achieve robust generalization across languages.</p><p>It is critical that identical sentences in different languages have similar vector representations. That is why the vector space properties of the source language, obtained from the teacher model, must be applied to other languages.</p><p>From the information above, it is evident that the input parameters of semantic search models are queries and passages. In the field of multilingual search, there are three main paradigms:</p><p>• Multilingual-to-monolingual: this approach allows to accept queries in multiple languages and compare them with passages in a single language. • Monolingual-to-multilingual: conversely, the idea behind this paradigm is to accept queries in a single language and return passages in multiple languages. • Multilingual-to-multilingual: provides the highest search adaptability and allows users to formulate queries and peruse passages across multiple languages. However, this flexibility may reduce the average accuracy due to the large number of possible language pairs. This paper explores the scenario when a storage contains indexed data in English and it becomes necessary to search information in other languages, particularly Ukrainian. We will take the original model (data in the index must be produced by this model) trained for asymmetric semantic search, using the bi-encoder architecture, and add capabilities to handle multilingual queries (in our case, bilingual, i.e., English and Ukrainian).</p><p>Using knowledge distillation will significantly reduce the training time since we will mimic the teacher model on translated queries. To avoid unnecessary training on passages and to ensure consistency between previously indexed and new documents, we will continue to use the teacher model to create embeddings for new passages. The output of both models can be evaluated using a similarity metric to find the most relevant passages (Figure <ref type="figure" target="#fig_2">3</ref>). To determine how similar two vectors are, we can use different metrics (cosine similarity, Euclidean distance, and dot product). Measuring Euclidean distance for high dimensional vectors becomes impractical, as they will be very far apart simply because of the vastness of the space they inhabit. Using cosine similarity, which measures the angle between two vectors by paying attention to their direction and ignoring magnitude, or dot product, which measures the overall congruence .</p><p>In practice, we can generate sentence embeddings using specially designed modifications of the Bidirectional Encoder Representations from Transformers (BERT) model, known as Sentence Transformers (SBERT). It provides two types of state-of-the-art asymmetric semantic search models, one tuned for cosine similarity, the other for dot product. Cosine similarity tuned models prefer to retrieve shorter passages, while dot product tuned models prefer to retrieve longer passages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Datasets</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">MS MARCO</head><p>Microsoft Machine Reading Comprehension (MS MARCO) <ref type="bibr" target="#b12">[13]</ref> is a large-scale dataset created by Microsoft for training and evaluating information retrieval systems. It is widely used to benchmark the performance of various models on tasks such as reading comprehension, question answering, and passage ranking.</p><p>The dataset comprises over a million anonymous questions offering a collection of short, real-world, natural language queries split into the train (808,731 queries), development (101,093 queries), and evaluation (101,092 queries) subsets. We will use the train and development subsets to train and evaluate the student model respectively, so it makes sense to look at their length and word patterns (Figure <ref type="figure" target="#fig_3">4</ref>). Furthermore, the dataset contains 8,841,823 passages that are required to provide natural language answers (a question may have multiple answers or no answers at all). Although passages are not involved in the knowledge distillation for multilingual-to-monolingual models, it still seems reasonable to analyze them to figure out what the teacher model was trained on (Figure <ref type="figure" target="#fig_4">5</ref>). The length distribution of the combination of queries from the train and development subsets is markedly skewed, with the majority being quite brief, typically under 100 characters, predominantly concentrated between 15-25 characters. The frequency drops significantly for longer queries, with very few extending beyond 100 characters. The most common bigrams (two-word sequence of words) indicate that the majority of queries are fact-oriented. These patterns reveal that a significant number of queries are short questions asking , , , The length distribution of the passages has a right-skewed plot with the noticeable peak around 200-300 characters. The character type distribution shows that letters are mostly used, with a median of about 250 characters per passage and a wide interquartile range reflecting considerable variation. Digits and punctuation marks are used less frequently, but there are outliers indicating some passages with a large number of such characters.</p><p>The scale and real-world nature of the dataset makes it attractive for training and evaluating machine learning models, but the original MS MARCO contains only English search queries. To train and evaluate models for other languages, the dataset must be translated in one of the available ways. For Ukrainian, the OPUS-MT English to East Slavic neural machine translation model <ref type="bibr" target="#b13">[14]</ref> demonstrated high-quality results in a relatively short time.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">TREC 19</head><p>In 2019, the National Institute of Standards and Technology (NIST), in collaboration with Microsoft, organized the Text Retrieval Conference (TREC) deep learning track <ref type="bibr" target="#b14">[15]</ref> benchmark competition. This event was aimed to foster research in the information retrieval direction using deep learning techniques. As the official evaluation dataset, organizers provided a list of 200 queries and a pool of documents and passages, labeled by NIST assessors using multi-graded judgments.</p><p>Although the deliberate selection of the queries at an intuitive level implies high-quality data and a lack of outliers, it still seems reasonable to check directly the length and word patterns (Figure <ref type="figure" target="#fig_5">6</ref>). The length distribution peaks at 20-25 characters, indicating that most queries are concise and right-skewed, with fewer longer queries beyond 60 characters. The most frequent bigrams begin with common phrases such as "what," "definition," and "how," suggesting that queries are structured as questions, often seeking definitions or factual information.</p><p>Looking at the queries from both datasets, it is clear that TREC 19 queries are generally shorter and more focused on concrete evidence, while MS MARCO queries might be significantly longer and more complex. It is also important to note that TREC 19 does not have outliers, unlike MS MARCO, where some queries can exceed 400 characters.</p><p>These days, the TREC 19 has become a recognized benchmark for various information retrieval and deep learning tasks such as document and passage retrieval. In the task of full search, we can evaluate up to 1,000 passages for each query based on their estimated likelihood of containing the answer.</p><p>To evaluate the trained model, all 200 queries were manually translated into Ukrainian by a native speaker, which guarantees the veracity of the results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Training and Evaluation</head><p>We will use the DistilBERT base multilingual (cased) <ref type="bibr" target="#b15">[16]</ref> as a student model. This model was trained on the concatenation of Wikipedia in 104 different languages (including English and Ukrainian), has 6 layers, 768 dimensions, and 12 heads, totalizing 134 million parameters (compared to 177 million parameters for the BERT base multilingual).</p><p>The MS MARCO DistilBERT base v4 <ref type="bibr" target="#b16">[17]</ref> was chosen as a teacher model. It embeds text into a 768-dimensional vector space and can be used for clustering and semantic search. This model was fine-tuned on the original MS MARCO passage ranking dataset and optimized to generate embeddings for queries and passages.</p><p>We trained the student model for 5 epochs with a batch size of 24, 10,000 warm-up steps, and a learning rate of 2e-5. The entire training process on the train and development (used for intermediate evaluation) subsets took about 8 hours using the Apple M3 Max chip (16-ore CPU, 40-ore GPU). To measure the difference between computed and target query embeddings we used the MSE loss (Figure <ref type="figure" target="#fig_6">7</ref>). The teacher model has a vocabulary of 30,522 tokens, the same as the original BERT base. This vocabulary includes common words, sub-words, and special tokens to deal with English. The student model, on the other hand, is a distilled version of the BERT multilingual. The extended vocabulary of 119,547 tokens covers a variety of multiple symbols from different languages, allowing to efficiently process and understand text in different linguistic contexts.</p><p>Any dataset for semantic search or information retrieval systems has selection bias. It can be related to the data source, the date of publication, and the personal preferences of the publisher. Achieving high performance on cross-lingual tasks depends on the ability to seamlessly map sentences from different languages into a single vector space. The similarity between alphabets also significantly affects accuracy. Languages with similar alphabets, such as English and German, typically produce more accurate results than languages with dissimilar alphabets, such as English and Ukrainian.</p><p>Using the original English and machine-translated Ukrainian search queries, we achieved a reasonable level of accuracy (about 93%) for English to Ukrainian and Ukrainian to English translation tasks, indicating a high level of proficiency in both directions (Figure <ref type="figure" target="#fig_7">8</ref>). It is very important to monitor performance on the subset that did not participate in training to avoid overfitting. MSE on the evaluation subset steadily decreases, indicating that the student model learns to more accurately mimic the results of the teacher model (Figure <ref type="figure" target="#fig_8">9</ref>). Evaluating the performance of information retrieval systems is a critical step in improving their efficiency. NIST assessors labeled the TREC 19 using multi-graded judgments, making it easy to measure all the necessary and widely used state-of-the-art metrics.</p><p>Reciprocal rank (RR) calculates the score for the first relevant passage in the ranked list. This metric is very important when we need to evaluate the occurrence of the most relevant passage:</p><formula xml:id="formula_0">𝑅𝑅 = 1 𝑟𝑒𝑙 ,<label>(1)</label></formula><p>where rel the rank of the first relevant passage in the list. Mean average precision (MAP) evaluates both the relevance of the suggested passages and the position of the most relevant passages at the top. For each query, the average precision (AP) is determined by calculating the arithmetic mean of the precision scores for every position in which a relevant passage was found. This metric focuses on the ability to distinguish relevant and irrelevant items:</p><formula xml:id="formula_1">𝑀𝐴𝑃@𝑘 = 1 𝑄 ∑ 𝐴𝑃@𝑘 𝑖 𝑄 𝑖=1 ,<label>(2)</label></formula><p>where Q the total number of queries. Normalized discounted cumulative gain (NDCG) measures the ability of machine learning algorithms to sort passages by relevance. NDCG is determined by dividing the discounted cumulative gain (DCG) by the ideal DCG, representing the best version of the rating. This is useful in many scenarios where we expect passages to be sorted by relevance:</p><formula xml:id="formula_2">𝐷𝐶𝐺@𝑘 = 𝐷𝐶𝐺@𝑘 𝐼𝐷𝐶𝐺@𝑘 = ∑ 𝑟𝑒𝑙 𝑖 𝑙𝑜𝑔 2 (𝑖 + 1) 𝑘 𝑖=1 ∑ 𝑟𝑒𝑙 𝑖 * 𝑙𝑜𝑔 2 (𝑖 + 1) 𝑘 𝑖=1 ,<label>(3)</label></formula><p>where k the number of items considered in the calculation. Both MAP and NDCG reflect ranking quality, but account for rank reduction in different ways. MAP gives more weight to relevant passages at the top of the list because this metric is based on precision, making it more sensitive to changes in early positions. DCG assigns decreasing weights to passages proceeding down the hierarchical order, but they are logarithmic and reduce the contribution of passages not very quickly.</p><p>In the case of models based on the bi-encoder architecture, it makes sense to evaluate up to 100 passages using the above metrics (Table <ref type="table" target="#tab_0">1</ref>). If the obtained performance is not sufficient, a reranking model based on the cross-encoder architecture can be used to improve results. The bilingual model trained using knowledge distillation inherited the capabilities of the monolingual teacher model and shows reasonable results for both languages. Obtained 768-dimensional vectors are optimized to work with cosine similarity as expected.</p><p>We can notice a slight decrease in performance for English, which is not critical since we got the ability to search passages using multiple languages. A higher score does not necessarily mean higher performance in production. At some point, models can become too specialized on the MS MARCO and its selection bias.</p><p>It is impossible to create a perfect dataset for semantic search. Manual creation of a subset even for evaluation is quite expensive and unfortunately always has some selection bias. This is a longrecognized problem, but there is no good solution for it, especially at the scale of millions of queries and passages.</p><p>The number of languages may not be limited to two, but it is better to reasonable add only justified languages and keep a balance between performance and multilingual search capabilities. Otherwise, new languages to the model degrades performance because the capacity of the model remains the same.</p><p>The performance of the trained model was also affected by the quality of the data, which in our case is a combination of original and machine-translated queries. Cloud solutions such as Google Translate or DeepL could marginally improve the results but would not compare to manual translation by native speakers. The time taken to create pairs using the neural machine translation model was about 50% longer than the time required to train the student model. Intuitively, it makes the knowledge distillation less of a training and more of a translation task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>For multilingual processing of asymmetric semantic search queries, the DistilBERT base multilingual (cased) as a student model and the MS MARCO DistilBERT base v4 as a teacher model can be used as components in the scenario when storage contains indexed data in English with the necessity to search for information in Ukrainian up to the reciprocal rank (RR), the mean average precision (MAP), and the normalized discounted cumulative gain (NDCG) as its evaluating.</p><p>The authors achieved a reasonable level of accuracy (about 93%) for English to Ukrainian and Ukrainian to English translation tasks, indicating a high level of proficiency in both directions. In the future, this result will be improved by the quality of the data, which can be a combination of original queries and manual translations by native speakers. The resulting model is useful in industries such as finance, healthcare, and e-commerce, where huge data sets are prevalent and asymmetric semantic search plays a key role to quickly retrieve relevant information.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Search in the vector space.</figDesc><graphic coords="2,135.82,104.94,328.58,178.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The idea of multilingual knowledge distillation.</figDesc><graphic coords="2,110.20,511.31,380.25,138.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Pipeline of using the trained and original model.</figDesc><graphic coords="3,111.10,412.08,378.43,119.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Length and word patterns of the queries from the train and development subsets.</figDesc><graphic coords="4,82.33,249.10,435.61,186.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Length and character type distribution of the passages.</figDesc><graphic coords="4,83.68,538.32,433.30,186.04" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Length and word patterns of the queries from the evaluation dataset.</figDesc><graphic coords="5,86.80,441.45,426.95,181.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: MSE loss on the train and evaluation subsets.</figDesc><graphic coords="6,127.82,377.99,344.54,208.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Accuracy of translation between English and Ukrainian.</figDesc><graphic coords="7,126.20,146.32,348.19,213.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: MSE on the evaluation subset.</figDesc><graphic coords="7,125.13,449.75,350.34,210.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Performance of the original and trained model on TREC 19</figDesc><table><row><cell>Model</cell><cell></cell><cell cols="2">TREC 2019 EN-EN</cell><cell></cell><cell cols="2">TREC 2019 UK-EN</cell></row><row><cell></cell><cell cols="6">RR MAP@100 NDCG@100 RR MAP@100 NDCG@100</cell></row><row><cell>msmarco-distilbert-base-</cell><cell>0.96</cell><cell>0.35</cell><cell>0.59</cell><cell></cell><cell></cell><cell></cell></row><row><cell>v4</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>msmarco-distilbert-</cell><cell>0.92</cell><cell>0.29</cell><cell>0.53</cell><cell>0.78</cell><cell>0.25</cell><cell>0.46</cell></row><row><cell>multilingual-en-uk</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Declaration on Generative AI</head><p>The authors have not employed any Generative AI tools.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">An Intelligent Semantic Vector Search Model for Grading and Assessing Students</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kgosietsile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">U</forename><surname>Okike</surname></persName>
		</author>
		<idno type="DOI">10.1109/i-COSTE60462.2023.10500811</idno>
	</analytic>
	<monogr>
		<title level="m">2023 International Conference on Sustainable Technology and Engineering (i-COSTE)</title>
				<meeting><address><addrLine>Nadi, Fiji</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Semantic Search for Sustainable Platforms Using Transformers</title>
		<author>
			<persName><forename type="first">S</forename><surname>Shirol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kulkarni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Agarwal</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICETCI58599.2023.10331079</idno>
	</analytic>
	<monogr>
		<title level="m">2023 International Conference on Emerging Techniques in Computational Intelligence (ICETCI)</title>
				<meeting><address><addrLine>Hyderabad, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="112" to="118" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Deep Learning-based Semantic Search Techniques for Enhancing Product Matching in E-commerce</title>
		<author>
			<persName><forename type="first">F</forename><surname>Aamir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sherafgan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Arbab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jamil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">N</forename><surname>Bhatti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Hameed</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICMI60790.2024.10586148</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE 3rd International Conference on Computing and Machine Intelligence (ICMI)</title>
				<meeting><address><addrLine>Mt Pleasant, MI, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Unveiling Deeper Petrochemical Insights: Navigating Contextual Question Answering with the Power of Semantic Search and LLM Fine-Tuning</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">P</forename><surname>Saikia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mahapatra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nandy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Das</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICCCIS60361.2023.10425564</idno>
	</analytic>
	<monogr>
		<title level="m">2023 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS)</title>
				<meeting><address><addrLine>Greater Noida, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="881" to="886" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Employee Data</title>
		<author>
			<persName><forename type="first">D</forename><surname>Sheth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Gupta</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICCICA52458.2021.9697114</idno>
	</analytic>
	<monogr>
		<title level="m">2021 International Conference on Computational Intelligence and Computing Applications (ICCICA)</title>
				<meeting><address><addrLine>Nagpur, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Semantic Social Searching-An Ontology Based Approach</title>
		<author>
			<persName><forename type="first">I</forename><surname>Sindhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Shamsi</surname></persName>
		</author>
		<idno type="DOI">10.1109/IMCERT57083.2023.10075145</idno>
	</analytic>
	<monogr>
		<title level="m">International Multi-disciplinary Conference in Emerging Research Trends (IMCERT)</title>
				<meeting><address><addrLine>Karachi, Pakistan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Semantic Search System For Real Time Occupancy</title>
		<author>
			<persName><forename type="first">P</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Acharya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Beliatis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Presser</surname></persName>
		</author>
		<idno type="DOI">10.1109/IoTaIS53735.2021.9628719</idno>
	</analytic>
	<monogr>
		<title level="m">2021 IEEE International Conference on Internet of Things and Intelligence Systems (IoTaIS)</title>
				<meeting><address><addrLine>Bandung, Indonesia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="49" to="55" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">An Analysis of the Technical Trend of Semantic Search in Natural Language Processing</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICNISC60562.2023.00033</idno>
	</analytic>
	<monogr>
		<title level="m">9th Annual International Conference on Network and Information Systems for Computers (ICNISC)</title>
				<meeting><address><addrLine>Wuhan, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="51" to="53" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Blending Weighted TF-IDF &amp; BERT Improving Semantic Search</title>
		<author>
			<persName><surname>Md</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICARC54489.2022.9753875</idno>
	</analytic>
	<monogr>
		<title level="m">2022 2nd International Conference on Advanced Research in Computing (ICARC)</title>
				<meeting><address><addrLine>Belihuloya, Sri Lanka</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="154" to="159" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Ameliorating Semantic Search Through Advanced AI Techniques</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">V</forename></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dhabliya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mathur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">B</forename><surname>Rao</surname></persName>
		</author>
		<idno type="DOI">10.1109/SMARTGENCON60755.2023.10442780</idno>
	</analytic>
	<monogr>
		<title level="m">2023 3rd International Conference on Smart Generation Computing, Communication and Networking (SMART GENCON)</title>
				<meeting><address><addrLine>Bangalore, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">FBC: Fusing Bi-Encoder and Cross-Encoder for Long-Form Text Matching</title>
		<author>
			<persName><forename type="first">J</forename><surname>Liao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Duan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.3233/faia230426</idno>
	</analytic>
	<monogr>
		<title level="m">Frontiers in Artificial Intelligence and Applications</title>
				<meeting><address><addrLine>Krakow, Poland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1473" to="1480" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Making Monolingual Sentence Embeddings Multilingual using Knowledge Distillation</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.emnlp-main.365</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</title>
				<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="4512" to="4525" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Marco</surname></persName>
		</author>
		<ptr target="https://microsoft.github.io/msmarco" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">OPUS-MT Building open translation services for the World</title>
		<author>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.eamt-1.61" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd Annual Conference of the European Association for Machine Translation</title>
				<meeting>the 22nd Annual Conference of the European Association for Machine Translation<address><addrLine>Lisboa, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="479" to="480" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">TREC Deep Learning Track: Reusable Test Collections in the Large Data Regime</title>
		<author>
			<persName><forename type="first">N</forename><surname>Craswell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mitra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Yilmaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Campos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Voorhees</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Soboroff</surname></persName>
		</author>
		<idno type="DOI">10.1145/3404835.3463249</idno>
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;21: The 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2369" to="2375" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter</title>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1910.01108</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1410</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3973" to="3983" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
