<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>IIR</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>An Analysis of a Methodology and Experimental Results for the Retrieval of Clinical Trials</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Discussion Paper</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giorgio Maria Di Nunzio</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Guglielmo Faggioli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefano Marchesin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Engineering, University of Padua</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>13</volume>
      <fpage>0000</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>In this paper, we present our previous and current work about the the methodology and the experimental analysis of a query reformulation, pseudo-relevance feedback, and document filtering approach. In particular, we present a summary of two studies carried out in the context of the TREC Precision Medicine track. The two original papers are [1] and [2]. The Clinical Trial Task1 at the Text REtrieval Conference (TREC)2 is an information retrieval challenge in the medical domain where the query is the a synthetic patient descriptions and the corpus is a large set of clinical trial descriptions. This task is part of a long run challenge related to the evaluation of Clinical Decision Support systems3 started in 2014. These tasks have sought to provide benchmark datasets and evaluate information retrieval systems focused on many of the most important information access problems in biomedicine. The dataset provided by the organizers of the task (in both 2021 and 2022) consists of a set of topics, a brief patient case description, and a set of documents, a snapshot of ClinicalTrials.gov. This collection consists of a snapshot of all the clinical trials available on ClinicalTrials.gov on April 27, 2021. The data is available as XML, with this specific snapshot containing 375,581 clinical trial descriptions. In this extended abstract, we summarize the methodologies and the experimental results that we achieved in the last two editions of the Clinical Trial Task, 2021 and 2022, where the main</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Precision medicine</kwd>
        <kwd>Query reformulation</kwd>
        <kwd>Pseudo Relevance Feedback</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>focus was the retrieval of clinical trials given a lengthy query that describes the patient case
that simulates an admission statement in an electronic health record.</p>
      <p>
        Our participation to these two editions focused on the evaluation of a mixture of query
reformulation, rank fusion, and document filtering approaches optimized on the experimental
analyses of our previous participations to this track [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. In addition, we also performed
experiments with pseudo-relevance feedback [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The objective of these studies is to continue the
evaluation of this longitudinal study of diferent combinations of approaches. These results
were reported in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology</title>
      <p>The methodology we employed in both participations was mainly based on a query reformulation
approach, the merging of the ranking lists provided by the diferent retrieval methods using (or
not) summarized queries and applying query expansion based on pseudo-relevance feedback.
In the following sections, we describe the approach for each element of the retrieval pipeline.</p>
      <sec id="sec-2-1">
        <title>2.1. Query reformulation</title>
        <p>
          In the 2021 edition, we use either BART [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] or T5 [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] models to perform summarization over the
original, lenghty queries. In the 2022 edition, we used two manual summarization approaches: i)
Natural Language Summary (NLS), where we reduced the original query keeping the structure
of the language; and ii) Keyword Summary (KS), where we kept only terms that are likely to be
relevant. As an additional experiment, we also tried a two-step summarization where we further
summarize NLS summaries using the transformer-based [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] T5 model [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. After summarization,
we applied both pseudo-relevance feedback and document filtering.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Query expansion</title>
        <p>
          We used the RM3 model to implement a pseudo-relevance feedback strategy including query
expansion [
          <xref ref-type="bibr" rid="ref4 ref8">4, 8</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Retrieval models:</title>
        <p>
          For each query, we run the Okapi BM25 retrieval model [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Filtering</title>
        <p>After the retrieval step, we filter out from the list of candidate trials those for which a patient
is not eligible based on their demographic data – that is, age and gender. In other words, we
automatically extract patient’s age and gender from queries and filter out trials whose eligibility
criteria do not allow for the extracted age and gender values. In those cases where part of
the demographic data are not specified, a clinical trial is kept or discarded on the basis of the
remaining demographic information. For instance, if the clinical trial does not specify a required
minimum age, then it is kept or discarded based on its maximum age and gender required
values.</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Ranking fusion</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <p>
        Given diferent ranking lists, we used the CombSUM [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] approach with minmax normalization
to merge them.
      </p>
      <p>The organizers of the TREC tasks usually provide the summary of the results in terms of best,
median, and worst value for each topic for three evaluation measures: Normalized Discount
Cumulative Gain (NDCG), precision at 10 (P@10), and Reciprocal Rank (RecipRank). In the
following sections, we summarize the details of each experiment and the results for each year.
3.1. Results of 2021
In 2021, we submitted five runs:
• RM3Filtered: run with RM3 expansion, using BM25 as the first and second stage retrieval
model. After both the first and the second retrieval stages, results have been filtered to
remove trials with unfeasible age or sex attributes;
• T5RM3Filt: Prior to the retrieval, queries are summarized using the T5 summarization
algorithm with a summary length - chosen by T5 - between 30 and 130 words. The same
model as RM3Filtered is used to retrieve documents;
• BARTRM3Filt: Prior to the retrieval, queries are summarized using the BART
summarization algorithm with a summary length - chosen by BART - between 30 and 130 words.</p>
      <p>The same model as RM3Filtered is used to retrieve documents;
• imsFused1: additive fusion of runs obtained with T5 summarizations with exact lengths
20, 50, 100, 150 and a run with T5 summarizations in the range 0-150. BM25 is used as
the retrieval model. results with unfeasible values of age or sex have been removed;
• imsFused2: CombSUM fusion with min-max normalization of imsFused1, RM3Filtered,</p>
      <p>T5RM3Filt, and BARTRM3Filt;</p>
      <p>In Table 2, we report the median values of the three measures averaged across topics, as well
as the averaged results of the five submitted runs.</p>
      <p>The results show that all the runs perform better than median values. In particular, the RM3
Filtered run performs significantly better than median (statistical analyses will be provided in
the final version of the paper), followed by the imsFused2 run and the BART RM3 filtered rank.
Given these promising results, we plan to investigate the integration of re-ranking components
in the retrieval pipeline.</p>
      <p>measure
infNDCG
P@10
RecipRank</p>
      <p>In Table 2, we report the median values of the three measures averaged across topics, as well
as the averaged results of the five submitted runs.</p>
      <p>The results show that the runs have mixed performances compared with median values.
Among the diferent approaches, those using keyword-based summaries seem to achieve higher
performance. On the other hand, the impact of RM3 to expand queries is not clear, and might
hinder the performance – as for the ims_RM3Filtered_s run. Given these mixed results, we plan
to deepen the investigation on manual summarization to understand what is the proper tradeof
between NLS and KS summaries.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In this extended abstract, we have presented a brief overview of the experimental results
obtained in the last two editions of the TREC Clinical Trial task. The results of the proposed
approach were well above the median results of the participants in both editions. For some
evaluation measures, our results were included in the top 5 performing systems.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G.</given-names>
            <surname>Di Nunzio</surname>
          </string-name>
          , G. Faggioli,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchesin</surname>
          </string-name>
          , Filter, transform, expand, and fuse,
          <source>in: Proceedings of the Thirtieth Text REtrieval Conference</source>
          , TREC 2021, Gaithersburg, Maryland, USA, November
          <volume>15</volume>
          -
          <issue>19</issue>
          ,
          <year>2021</year>
          , NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Di Nunzio</surname>
          </string-name>
          , G. Faggioli,
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchesin</surname>
          </string-name>
          ,
          <article-title>Summarize and expand queries in clinical trials retrieval. the iiiaunipd at trec 2022 clinical trials</article-title>
          ,
          <source>in: Proceedings of the Thirty-First Text REtrieval Conference</source>
          , TREC 2022, Gaithersburg, Maryland, USA, November
          <volume>15</volume>
          -
          <issue>19</issue>
          ,
          <year>2022</year>
          , NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Marchesin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. M.</given-names>
            <surname>Di Nunzio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Agosti</surname>
          </string-name>
          ,
          <article-title>Simple but efective knowledge-based query reformulations for precision medicine retrieval</article-title>
          ,
          <source>Information</source>
          <volume>12</volume>
          (
          <year>2021</year>
          ). URL: https://www. mdpi.com/2078-2489/12/10/402. doi:
          <volume>10</volume>
          .3390/info12100402.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Lavrenko</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>Relevance based language models</article-title>
          ,
          <source>in: Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '01,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2001</year>
          , pp.
          <fpage>120</fpage>
          -
          <lpage>127</lpage>
          . URL: https://doi.org/10.1145/383952.383972. doi:
          <volume>10</volume>
          .1145/383952.383972.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer,
          <article-title>BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</article-title>
          ,
          <source>ACL 2020, Online, July</source>
          <volume>5</volume>
          -
          <issue>10</issue>
          ,
          <year>2020</year>
          , Association for Computational Linguistics,
          <year>2020</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          . URL: https: //doi.org/10.18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>J. Mach. Learn. Res</source>
          .
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <volume>140</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>140</lpage>
          :
          <fpage>67</fpage>
          . URL: http://jmlr.org/papers/v21/
          <fpage>20</fpage>
          -
          <lpage>074</lpage>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9</source>
          ,
          <year>2017</year>
          , Long Beach, CA, USA,
          <year>2017</year>
          , pp.
          <fpage>5998</fpage>
          -
          <lpage>6008</lpage>
          . URL: https://proceedings.neurips.cc/ paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Jaleel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Allan</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Diaz</surname>
            ,
            <given-names>L. S.</given-names>
          </string-name>
          <string-name>
            <surname>Larkey</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          <string-name>
            <surname>Smucker</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Wade</surname>
          </string-name>
          , Umass at TREC 2004:
          <article-title>Novelty and HARD</article-title>
          , in: E. M.
          <string-name>
            <surname>Voorhees</surname>
            ,
            <given-names>L. P.</given-names>
          </string-name>
          Buckland (Eds.),
          <source>Proceedings of the Thirteenth Text REtrieval Conference</source>
          , TREC 2004, Gaithersburg, Maryland, USA, November
          <volume>16</volume>
          -
          <issue>19</issue>
          ,
          <year>2004</year>
          , volume
          <volume>500</volume>
          -261 of NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>2004</year>
          . URL: http://trec.nist.gov/pubs/trec13/papers/ umass.novelty.hard.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zaragoza</surname>
          </string-name>
          ,
          <article-title>The probabilistic relevance framework: BM25 and beyond</article-title>
          ,
          <source>Foundations and Trends in Information Retrieval</source>
          <volume>3</volume>
          (
          <year>2009</year>
          )
          <fpage>333</fpage>
          -
          <lpage>389</lpage>
          . URL: https://doi.org/ 10.1561/1500000019. doi:
          <volume>10</volume>
          .1561/1500000019.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Fox</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Shaw</surname>
          </string-name>
          , Combination of Multiple Searches,
          <source>NIST special publication SP</source>
          <volume>243</volume>
          (
          <year>1994</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>