<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the Second Shared Task on Spoken Query Cross-Lingual Information Retrieval for Indic Languages SqCLIR at FIRE 2025</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bhargav Dave</string-name>
          <email>R@100</email>
          <email>R@1000</email>
          <email>bhargavdave1@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Prasenjit Majumder</string-name>
          <email>prasenjit.majumder@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Debasis Ganguly</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Evangelos Kanoulas</string-name>
          <email>ekanoulas@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dhirubhai Ambani University</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>GreenAI Services Pvt. Ltd</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Amsterdam</institution>
          ,
          <addr-line>Amsterdam</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Glasgow</institution>
          ,
          <addr-line>Scotland</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>This paper presents an overview of the Second Shared Task on Spoken Query Cross-Lingual Information Retrieval for Indic Languages (SqCLIR 2025), organised as part of FIRE 2025. The task focuses on developing and evaluating systems capable of retrieving relevant textual documents given spoken queries in Indic languages. Building on the first edition conducted in 2024, SqCLIR 2025 introduced significant enhancements, including the adoption of the IndicMSMARCO dataset as the retrieval collection and the use of spoken queries (audio in .wav format) across five languages - Gujarati, Hindi, Bengali, Kannada, and English. The shared task comprised two subtasks: (1) Spoken Query Ad-Hoc Retrieval (Monolingual), focusing on retrieving documents in the same language as the spoken query; and (2) Spoken Query Cross-Lingual Retrieval, targeting document retrieval across diferent source and target languages. In addition to the IndicMSMARCO text collection, spoken queries were derived from TREC DL 19 and 20 topics, recorded in Indic languages to simulate realistic voice-search scenarios. System performance was evaluated using standard IR metrics, including nDCG@10, MRR, and recall at multiple depths. A total of six teams registered, though only one team submitted a valid run. Despite limited participation, the task successfully established a foundation for spoken cross-lingual retrieval in low-resource Indic settings, highlighting challenges related to ASR accuracy, language diversity, and speech variability.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Spoken Query</kwd>
        <kwd>Information Retrieval</kwd>
        <kwd>Indic Language</kwd>
        <kwd>Cross-Lingual</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Voice-based interfaces have emerged as a dominant means of accessing digital information, driven by
the widespread use of smartphones, virtual assistants, and speech-enabled applications. In multilingual
societies such as India, users frequently express their information needs through speech in their native
languages rather than text. This trend underscores the importance of developing spoken information
retrieval systems that can accurately interpret speech queries and retrieve relevant textual content across
languages. However, most traditional IR systems are designed for text-based and monolingual settings,
which limits their applicability for diverse, multilingual users. Spoken Query Cross-Lingual Information
Retrieval (SqCLIR) addresses this gap by combining speech recognition, language translation, and
retrieval to enable systems that can process spoken queries in one language and retrieve documents in
the same or another language.</p>
      <p>Developing efective SqCLIR systems for Indic languages presents several challenges. India’s linguistic
landscape is characterised by vast diversity, and many languages have rich morphology, distinct scripts,
and limited computational resources. The scarcity of parallel corpora and annotated data restricts the
development of robust translation and retrieval models. Additionally, Automatic Speech Recognition
(ASR) systems for Indic languages face high word error rates due to factors such as accent variation,
background noise, and frequent code-mixing. These issues make spoken cross-lingual retrieval a
complex, multi-stage problem that remains underexplored in low-resource multilingual contexts.</p>
      <p>
        Research in IR has evolved from traditional lexical-matching approaches such as BM25 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to dense
retrieval models that learn semantic representations using dual-encoder architectures like DPR [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and
ColBERT [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These models have significantly improved multilingual and cross-lingual retrieval when
combined with language-agnostic encoders such as LaBSE [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Work on Cross-Lingual Information
Retrieval (CLIR) has been extensively studied in evaluation forums such as TREC, CLEF, and FIRE, where
approaches based on query translation, document translation, or shared multilingual embeddings have
been compared [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. In parallel, Spoken Information Retrieval (SIR) and Spoken Document Retrieval
(SDR) have investigated the retrieval of textual or audio content from speech inputs [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], although most
prior studies relied on ASR transcripts rather than raw audio queries. Within FIRE, early CLIR tasks
were purely text-based, and only recently have shared tasks such as SqCLIR 2024 [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ] introduced
real spoken queries for Indic languages. More recently, the study SqCLIRIL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] further advanced this
direction by exploring end-to-end spoken query retrieval approaches for Indian languages, providing
valuable insights that complement the present SqCLIR 2025 shared task 1.
      </p>
      <p>
        The SqCLIR shared task series, organised under the FIRE initiative, aims to foster research in
speechdriven retrieval for Indic languages. The second edition (SqCLIR 2025) introduced several key
enhancements over the first edition. We adopted the IndicMSMARCO [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] dataset as the retrieval collection,
providing a large-scale, multilingual benchmark suitable for both monolingual and cross-lingual
retrieval. Spoken queries were derived from TREC DL’19 [12] and DL’20 [13] queries and re-recorded
in multiple Indic languages to simulate realistic voice-search conditions. The task expanded to five
languages—Gujarati, Hindi, Bengali, Kannada, and English—making it one of the most comprehensive
spoken IR evaluations in the Indic context.
      </p>
      <p>The primary objectives of SqCLIR 2025 were to:
• Establish a benchmark platform for evaluating spoken query retrieval systems across multiple</p>
      <p>Indic languages.
• Encourage research in monolingual and cross-lingual spoken information retrieval.
• Provide standardised datasets and evaluation protocols to ensure reproducibility and fair
comparison.
• Identify key challenges in integrating speech recognition, translation, and retrieval in low-resource
multilingual settings.</p>
      <p>To achieve these goals, two subtasks were designed to address distinct retrieval scenarios:
• Task 1: Spoken Query Ad-Hoc Retrieval (Monolingual) – Retrieve documents in the same
language as the spoken query.
• Task 2: Spoken Query Cross-Lingual Retrieval – Retrieve documents written in a diferent
target language from the spoken query.</p>
      <p>Together, these subtasks aim to provide a unified and realistic benchmark for evaluating spoken
query retrieval in multilingual Indic environments and to stimulate further research on speech-driven
cross-lingual retrieval systems.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Definition</title>
      <p>The SqCLIR 2025 shared task focuses on developing and evaluating systems that can efectively retrieve
relevant textual documents given spoken queries in Indic languages. Participants are provided with a
text-based query along with its corresponding spoken version in wav format. They are encouraged
to utilize the provided spoken queries or generate additional spoken samples recorded under varied
environmental conditions to test system robustness. The shared task is divided into two subtasks
designed to evaluate both monolingual and cross-lingual retrieval capabilities.</p>
      <sec id="sec-2-1">
        <title>2.1. Task 1: Spoken Query Ad-Hoc Retrieval (Monolingual)</title>
        <p>In this subtask, participants are required to develop a spoken query retrieval system capable of handling
monolingual queries. Both the spoken query and the target document collection belong to the same
language, making the retrieval process comparatively straightforward. The objective is to accurately
interpret spoken queries and retrieve the most relevant documents from the text corpus in the same
language. For SqCLIR 2025, the monolingual task includes five Indic and English languages: Gujarati,
Hindi, Bengali, Kannada, and English.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Task 2: Spoken Query Cross-Lingual Retrieval</title>
        <p>The second subtask focuses on cross-lingual retrieval, where the spoken query and the document
collection are in diferent languages. Participants are required to design retrieval systems that can
interpret a spoken query in one language and return the most relevant documents written in another
language. This task introduces additional challenges such as translation ambiguity, speech recognition
errors, and cross-lingual semantic alignment. The task involves the same five languages—Gujarati,
Hindi, Bengali, Kannada, and English—and allows for any combination of query–document language
pairs. This flexible setup enables participants to explore a variety of cross-lingual retrieval strategies
and evaluate system performance under multilingual and speech-driven conditions.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset and Resources</title>
      <p>
        The SqCLIR 2025 shared task employed large-scale multilingual datasets for text retrieval and spoken
query evaluation across Indic languages. The primary text collections provided to the participants
were the IndicMSMARCO [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] dataset for Gujarati, Hindi, Bengali, and Kannada, and the original
MSMARCO [14] passage ranking dataset for English. IndicMSMARCO, an extension of MSMARCO,
ofers a multilingual benchmark consisting of query document pairs translated from English into
multiple Indic languages. It comprises over 8.8 million passages, enabling consistent evaluation across
both monolingual and cross-lingual retrieval tasks. The inclusion of both Indic and English collections
ensured that retrieval experiments could be performed under realistic conditions, supporting a wide
range of research scenarios within the SqCLIR 2025 task.
      </p>
      <p>
        The spoken query dataset was derived from the TREC DL’19 and DL’20 query sets, consisting of a
total of 97 queries. The translated and recorded spoken queries were sourced directly from the SqCLIRIL
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] study, which originally prepared these resources for spoken query cross-lingual retrieval research.
For each language: Gujarati, Hindi, Bengali, Kannada, and English, one male and one female speaker’s
recordings were taken from SqCLIRIL to ensure balanced gender representation and natural variability
in pronunciation and acoustic characteristics. The recordings were distributed in wav format with
consistent sampling rates and controlled durations. To further test system robustness, participants were
encouraged to record their own spoken versions of the same queries in the specified format and under
varied acoustic conditions for selected languages, as outlined in the task guidelines.
      </p>
      <p>All datasets and resources were made available through the FIRE 2025 SqCLIR portal. The release
package included the text queries, spoken query files, and query relevance judgments (qrels) for
evaluation. Together, these datasets and tools form a comprehensive benchmark for advancing research
in spoken query and cross-lingual information retrieval for low-resource Indic languages.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation Setup</title>
      <p>For both tasks, system performance was evaluated using standard information retrieval metrics. The
primary evaluation metric was nDCG@10, which measures ranking quality based on graded relevance.
Additional metrics included MAP, MRR, Recall@100, and Recall@1000, providing a comprehensive
view of retrieval efectiveness across diferent evaluation depths. This metric suite ensured consistent
and comparable assessment of spoken and cross-lingual retrieval systems submitted to the SqCLIR 2025
shared task.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Discussion</title>
      <p>A total of six teams registered for the SqCLIR 2025 shared task; however, only one team submitted a
valid run corresponding to the Hindi monolingual track. The submitted system evaluated retrieval
performance using both traditional and neural baselines, including BM25 and IndicBERT, across four
query conditions: text queries, spoken queries (for males and females), and participant-recorded spoken
queries.</p>
      <p>As presented in Table 1, the results obtained by the participant team [15] were the best retrieval
efectiveness achieved for text queries, which represent the upper bound of system performance in
the absence of speech-related errors. For BM25, text queries achieved an nDCG@10 of 0.2024 and a
MRR of 0.4497, while IndicBERT achieved an nDCG@10 of 0.1638 and an MRR of 0.3618. In contrast,
performance for spoken queries was substantially lower due to acoustic variability and ASR errors.
Among spoken inputs, the female query recordings consistently outperformed male recordings across
both retrieval models, possibly due to clearer articulation and less background noise in the recorded
samples. The participant-recorded queries produced comparable results to the provided spoken queries,
indicating a degree of robustness in the evaluation design.</p>
      <p>Despite limited participation, the results provide a valuable reference point for future research in
spoken information retrieval for Indic languages. The performance gap between text and spoken queries
highlights the challenges of ASR accuracy, pronunciation diversity, and domain mismatch core issues
that must be addressed to advance cross-lingual speech-based retrieval in low-resource settings.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>The SqCLIR 2025 shared task advanced research in spoken query–based cross-lingual information
retrieval for Indic languages, extending the first edition by adopting the large-scale IndicMSMARCO
dataset and incorporating spoken queries across five languages—Gujarati, Hindi, Bengali, Kannada, and
English. Despite limited participation, the task successfully established a benchmark framework for
evaluating both monolingual and cross-lingual retrieval using spoken input. The findings reveal that
text queries consistently outperform spoken ones, largely due to automatic speech recognition errors,
pronunciation variability, and background noise. Nonetheless, SqCLIR 2025 demonstrates the feasibility
of large-scale spoken query retrieval in low-resource Indic settings and provides a strong foundation
for future work focusing on robust ASR integration, cross-lingual modeling, and end-to-end neural
speech retrieval systems.</p>
      <p>Query Type
Text query
Spoken query (female)
Spoken query (male)
Participant Recorded query (male)</p>
      <p>Model
BM25
IndicBERT
BM25
IndicBERT
BM25
IndicBERT
BM25
IndicBERT
0.2024
0.1638
0.0951
0.0715
0.0751</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We sincerely thank the organizers of FIRE 2025 for providing the opportunity to host the SqCLIR track
as part of the conference. We also express our heartfelt gratitude to the native speakers who contributed
to the creation of the spoken query dataset—their support and dedication were instrumental in the
successful development of this resource.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and
take(s) full responsibility for the publication’s content.
2: Short Papers), Association for Computational Linguistics, Bangkok, Thailand, 2024, pp. 501–509.
doi:10.18653/v1/2024.acl-short.46.
[12] N. Craswell, B. Mitra, E. Yilmaz, D. Campos, E. M. Voorhees, Overview of the TREC
2019 deep learning track, CoRR abs/2003.07820 (2020). URL: https://arxiv.org/abs/2003.07820.
arXiv:2003.07820.
[13] N. Craswell, B. Mitra, E. Yilmaz, D. Campos, Overview of the TREC 2020 deep learning track,</p>
      <p>CoRR abs/2102.07662 (2021). URL: https://arxiv.org/abs/2102.07662. arXiv:2102.07662.
[14] T. Nguyen, M. Rosenberg, X. Song, J. Gao, S. Tiwary, R. Majumder, L. Deng, MS MARCO: A
human generated machine reading comprehension dataset, CoRR abs/1611.09268 (2016). URL:
http://arxiv.org/abs/1611.09268. arXiv:1611.09268.
[15] P. TT, T. KK, B. B, Spoken-query cross-lingual information retrieval for the indic languages (sqclir)
using bm25 and indic-bert (2025).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zaragoza</surname>
          </string-name>
          ,
          <article-title>The probabilistic relevance framework: Bm25 and beyond</article-title>
          ,
          <source>Found. Trends Inf. Retr</source>
          .
          <volume>3</volume>
          (
          <year>2009</year>
          )
          <fpage>333</fpage>
          -
          <lpage>389</lpage>
          . doi:
          <volume>10</volume>
          .1561/1500000019.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Oguz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Min</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Edunov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          , W.-t. Yih,
          <article-title>Dense passage retrieval for open-domain question answering</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <source>ACL</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>6769</fpage>
          -
          <lpage>6781</lpage>
          . doi:
          <volume>10</volume>
          .18653/ v1/
          <year>2020</year>
          .emnlp-main.
          <volume>550</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaharia</surname>
          </string-name>
          ,
          <article-title>Colbert: Eficient and efective passage search via contextualized late interaction over bert</article-title>
          ,
          <source>in: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>48</lpage>
          . doi:
          <volume>10</volume>
          .1145/ 3397271.3401075.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Arivazhagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Language-agnostic bert sentence embedding</article-title>
          , arXiv preprint arXiv:
          <year>2007</year>
          .
          <year>01852</year>
          (
          <year>2022</year>
          ). URL: https://arxiv.org/abs/
          <year>2007</year>
          .
          <year>01852</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Oard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Diekema</surname>
          </string-name>
          ,
          <article-title>The state of the art in cross-language information retrieval</article-title>
          ,
          <source>in: Proceedings of the American Society for Information Science Annual Meeting</source>
          , volume
          <volume>35</volume>
          ,
          <year>1998</year>
          , pp.
          <fpage>237</fpage>
          -
          <lpage>246</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Grefenstette</surname>
          </string-name>
          ,
          <string-name>
            <surname>Cross-Language Information</surname>
            <given-names>Retrieval</given-names>
          </string-name>
          , Springer Science &amp; Business
          <string-name>
            <surname>Media</surname>
          </string-name>
          ,
          <year>2012</year>
          . doi:
          <volume>10</volume>
          .1007/
          <fpage>978</fpage>
          -94-011-5206-1.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Garofolo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stanford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Pallett</surname>
          </string-name>
          , W. M. Fisher,
          <string-name>
            <given-names>N. L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <article-title>The trec spoken document retrieval track: A success story</article-title>
          ,
          <source>in: Proceedings of the Content-Based Multimedia Information Access Conference (RIAO)</source>
          ,
          <year>2000</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>20</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>B.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ganguly</surname>
          </string-name>
          , E. Kanoulas,
          <article-title>Overview of the fire 2024 sqclir track: Spoken query cross-lingual information retrieval for the indic languages, in: Proceedings of the Forum for Information Retrieval Evaluation (FIRE</article-title>
          <year>2024</year>
          )
          <article-title>Working Notes</article-title>
          , volume
          <volume>3810</volume>
          , CEUR-WS.org,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3810</volume>
          /paperXXX.pdf, available at CEUR Workshop Proceedings.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>B.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ganguly</surname>
          </string-name>
          , E. Kanoulas,
          <article-title>Findings of shared task on spoken query crosslingual information retrieval for the indic languages at fire 2024, in: Proceedings of the 16th Annual Meeting of the Forum for Information Retrieval Evaluation</article-title>
          , FIRE '24,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2024</year>
          . doi:
          <volume>10</volume>
          .1145/3734947.3735669.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Dave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          , Sqcliril:
          <article-title>Spoken query cross-lingual information retrieval in indian languages, Pattern Recognition Letters (</article-title>
          <year>2025</year>
          ). URL: https://www.sciencedirect.com/science/article/ pii/S0167865525003071. doi:https://doi.org/10.1016/j.patrec.
          <year>2025</year>
          .
          <volume>08</volume>
          .022.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Haq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Chhaya</surname>
          </string-name>
          , P. Bhattacharyya,
          <article-title>IndicIRSuite: Multilingual dataset and neural information models for Indian languages</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics</source>
          (Volume
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>