<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Zaragoza, Spain
$ roi.santos.rios@udc.es (R. Santos-Ríos); adrian.lopez.gude@udc.es (A. Gude); francisco.prado.valino@udc.es
(F. Prado-Valiño); ana.ezquerro@udc.es (A. Ezquerro); jesus.vilares@udc.es (J. Vilares)
 https://dunque.github.io/ (R. Santos-Ríos); https://adrian-gude.github.io/ (A. Gude); https://anaezquerro.github.io/
(A. Ezquerro); https://www.grupolys.org/~jvilares/ (J. Vilares)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>LyS at IberLEF 2025 Task PRESTA: Zero-Shot Code Generation for Spanish Tabular QA</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Roi Santos-Ríos</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adrián Gude</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francisco Prado-Valiño</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ana Ezquerro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jesús Vilares</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidade da Coruña, CITIC, Departamento de Ciencias de la Computación y Tecnologías de la Información</institution>
          ,
          <addr-line>Campus de Elviña s/n, 15071, A Coruña</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>9</fpage>
      <lpage>0009</lpage>
      <abstract>
        <p>This paper describes our participation in PRESTA, an IberLEF 2025 task, focused on Tabular Question Answering. We developed a zero-shot pipeline that leverages an Large Language Model to generate functional code capable of extracting the relevant information from tabular data based on an input question. Our approach consists of a modular pipeline where the main code generator module is supported by additional components that identify the most relevant columns and analyze their data types to improve extraction accuracy. In the event that the generated code fails, an iterative refinement process is triggered, incorporating the error feedback into a new generation prompt to enhance robustness. Our results show that zero-shot code generation is a valid approach for Tabular QA, achieving rank 4 out of 7 in the test phase despite the lack of task-specific fine-tuning.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Tabular Question Answering</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Zero-Shot</kwd>
        <kwd>Code Generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Tabular Question Answering (Tabular QA) has huge potential in real-world applications such as financial
analysis, business intelligence, and scientific data exploration, where structured databases serve as
the primary source of information. Unlike traditional text-based Question Answering (QA), which
primarily deals with unstructured data, Tabular QA requires extracting information from structured
tables to be able to answer the input questions, thus involving reasoning about diverse table schemas,
column relationships, and heterogeneous data types.</p>
      <p>
        Complex supervised systems have been proposed to deal with the structured nature of Tabular QA,
either leveraging structured prediction with language representations [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] or by formulating the
task as a sequence-to-sequence problem [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4, 5</xref>
        ]. However, with the rise of instruction-based Large
Language Models (LLM) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], recent approaches have shifted away from reliance on large annotated
datasets, instead reframing the task as a zero-shot generation problem [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>In this work, we further explore instruction-based LLMs to dynamically generate code functions
capable of retrieving relevant data from tables based on the input question in a zero-shot manner. To
enhance accuracy and reliability, we developed a modular three-staged pipeline that includes: (i) a
column selection mechanism to determine the most relevant columns and their data-type, (ii) a code
generation module responsible for producing executable code and (iii) an iterative error handling
module that, in case the initial code execution fails, tries to fix the generated code accordingly.</p>
      <p>
        Our group tested this approach within IberLEF’s [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] PRESTA task [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which provided a diverse
dataset featuring real-world tabular data.1 The competition required models to produce answers in
multiple formats, including boolean, categorical, numerical, and list-based outputs. Our model was
designed to generalize across diferent table structures, making it adaptable to various datasets beyond
the shared task, ensuring robustness and broad applicability. Although our approach demonstrated
strong performance in code generation and execution, subsequent analysis revealed that the model
struggles with columns containing complex data types (lists, dictionaries, etc.) and ambiguous queries,
particularly for list-based responses.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        Question Answering (QA) has been gaining significant attention in recent years, driven by the need
for models capable of reasoning over structured data. Early tasks in QA mainly focused on retrieving
information from unstructured text sources [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], but the increasing availability of structured datasets
has led to new challenges in understanding and querying tabular data. Unlike classic text-based QA,
where answers are retrieved from free-form text, Tabular QA requires a higher level of interpretation
and robustness to map questions to relevant columns and rows, handle missing values, and compute
statistics when necessary.
      </p>
      <p>
        In parallel, several datasets have been introduced to benchmark Tabular QA models, including
WikiTableQuestions [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], SQA [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], and the more recent DataBenchSPA dataset [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], which provides
real-world tabular data for evaluating models in diferent scenarios.
      </p>
      <p>
        Structured Tabular QA Most state-of-the-art approaches for Tabular QA leverage a pretrained
language model —equipped with an specialized encoding module to represent tabular information—
tailored for structured prediction. For example, TaPas [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] feeds both the input question and the flattened
table into BERT [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] as a single sequence, and finetunes the architecture to select relevant columns and
predict an aggregation function. Similarly, TaCube [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] combines a cube constructor with BART [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
to predict the real answers based on the input question and the results of the cube operations.
Generative Tabular QA To address the rigidity of structured approaches, recent works have explored
generative models for program synthesis, where an LLM is finetuned to generate executable programs
or instructions (in the form of SQL queries, for example) to be applied against tabular sources. Zhong
et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] proposed Seq2SQL, a sequence-to-sequence model to translate natural language into SQL
syntax, incorporating query-space pruning to significantly simplify and enhance the generative task.
Later, Yin et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] joined both concepts by optimizing tabular embeddings that fit both generative and
structured purposes.
      </p>
      <p>
        Zero-Shot Code Generation More recently, advancements in code generation have enabled a
paradigm shift in Tabular QA, driven by powerful multipurpose LLMs with strong coding capabilities,
such as Qwen [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and Mistral’s Codestral [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. These models facilitate a zero-shot approach to program
synthesis, eliminating the need for predefined templates or large annotated datasets. Instead, zero-shot
generation allows the system to dynamically adapt to diferent schemes without explicit prior knowledge
of the table structure [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], thus providing flexibility and scalability.
      </p>
      <p>Despite its potential, zero-shot code generation models still face big challenges, particularly in
error handling, runtime execution failures, and schema variability. Building on this approach, our
work extends an instruction-based model with error awareness, enabling it to detect and recover from
execution failures in an iterative error-recovery mechanism, where the model dynamically analyzes
execution failures and regenerates code based on error feedback.</p>
    </sec>
    <sec id="sec-3">
      <title>3. System Overview</title>
      <p>
        Our approach for the PRESTA task iterates upon the code generation approaches for Tabular QA, where
the core component is a pretrained LLM responsible of generating executable code to extract the answer
from the tables. To build upon prior works [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we incorporated a module that helps selecting the
columns relevant to the question, while also identifying the data types of their content. Moreover, we
      </p>
      <p>Answer</p>
      <p>No</p>
      <p>Code error?
Answer Execute CFioxdeer Yes
incorporaCotdee an error-fixing module that attempts to catch runtime errors and integrates them as part of
a new prompt, guiding the LLM to refine its code generation.</p>
      <p>Figure 1 shows an schematic view of the architecture of our system. We have designed a modular
pipeline that features three main components, which we describe below: (i) a column selector, (ii) an
answer generator and (iii) a code fixer.</p>
      <p>Input question
¿Cuál es la edad máxima de las mujeres participantes en la encuesta?
Database</p>
      <p>Column Selector Instead of relying on manually crafted heuristics or embedding similarity measures,
the first component of our system leverages an instruction-based LLM tasked to identify the most
relevant columns of a tabular source from an input question in natural language form. Our template
provides the list of column names and instructs the model to return only those that are essential for
answering the query.2
Answer Generator Once the relevant columns are identified, the second component of our pipeline
is instructed to generate executable code that retrieves the answers from the tabular source using both
the input query and the relevant columns extracted in the previous step. As part of our prompt, we
guided the LLM to generate Python programming code and postprocessed the output to ensure that
only Python lines were passed throught the next module. Python language was chosen since it is widely
used in data analysis and has extensive support for tabular data processing through libraries such as
Pandas.</p>
      <p>Code Fixer The final component of our pipeline captures execution errors that might occur due to
incorrect syntax, schema mismatches, or runtime exceptions. This module captures the error messages
and re-generates a corrected function by feeding the error context back into the LLM. To achieve this,
we used a structured prompt that includes the code that causes an error with the corresponding error
description.</p>
      <p>Preprocessing Since our system strongly relies on a well-formatted prompt, we manually designed a
preprocessing step to ensure a consistent format to feed our system. We standardized column names
for simplified versions (removing emoji and all non-alphanumerical characters except punctuation
symbols) to prevent possible errors in the Answer Generator caused by mismatches between the table
2All our prompts are available in the code publicly available at GitHub, as well as in the Annex section.
structure and the generated code. We identified enum-like column types, such as the case of categorical
attributes with a finite amount of strings as a value (e.g. a “Survey” column that only contains “Yes”,
“No” or “Maybe”), and inferred a common scheme so to ensure consistency across diferent attributes,
thus reducing errors related to unexpected variations in categorical values.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <p>Our system relies on open-source LLMs for zero-shot code generation. This way, no explicit training nor
ifnetuning was conducted. Instead, we used the available training phase datasets to validate diferent
LLMs and select the best performing one for the final test phase.</p>
      <p>Dataset The dataset provided for the task is divided into three sets: training, development (aka dev),
and test. In our case, since we had opted for a zero-shot approach, the training set remained unused
during the development phase, using only the dev set for our experiments. During this stage we tried
diferent LLMs to compare their ability to generate the adequate Python code to answer the input
questions. To do that, we analyzed the accuracy obtained with respect to the ground truth of the
validation set, together with manual checks to assess the quality of the generated code.
Evaluation The oficial evaluation consists of checking if the system is able to retrieve the answer
from the tables, comparing its output with the expected one. Still, in this competition the evaluation
scripts allow for non-meaningful diferences in the outputs; i.e. the system outputs 125.0 (float), and the
expected result is 125 (int). In this case it is counted as a correct response.</p>
      <p>
        System Setup We conducted experiments with diferent open-source LLMs adjusted to our hardware
limitations, specifically pretrained for instruction-based code generation: Qwen-2.5-Coder [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] (with 7B
and 32B versions), Mistral-7B and Codestral-22B —the later two from Mistral [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>To run the generated code we relied on Python 3.10.12 with Pandas 2.2.3 as a requirement. Due to
VRAM constraints, all models were executed with 4-bit quantization, using a greedy generation strategy
with a temperature of 0.7.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Analysis of Results</title>
      <p>In this section, we present the evaluation of our system on the task. We first report performance during
the development phase ($5.1), where we experimented with diferent models on the validation dataset,
followed by the final test phase ($5.2), where our system was evaluated on the test dataset through
CodaBench submissions.3</p>
      <sec id="sec-5-1">
        <title>5.1. Development Phase</title>
        <p>As explained before, during the development phase we focused on selecting the best performing LLM
just using the dev set; that is, dismissing the training set. At this first stage, our pipeline was conformed
by only the Answer Generator module.</p>
        <p>
          The results obtained for this original setup, presented in Table 1, show that the only model able to
outperform the baseline system [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] is Qwen-2.5-Coder32B.
        </p>
        <p>The majority of code answers outputted from the Qwen-2.5-Coder7B, Mistral7B and Codestral22B
models stem from trying to use the function "split()", which cannot be used with the majority of
datatypes present in the columns of the dataset.</p>
        <p>The smaller LLM models needed more postprocessing in order to be able to extract the code they
generated from the rest of the response. They usually add unnecessary textual descriptions of the code,
even though it is stated in the prompt that none of that is necessary, and will make the execution fail.
05.56
00.00</p>
        <p>An example of this behavior with an output of Codestral22B:
‘ ‘ ‘ python
i m p o r t p a n d a s a s pd
d e f answer ( d f : pd . DataFrame ) −&gt; l i s t :
column_name = ’ c o n s i d e r a n d o</p>
        <p>una e s c a l a de 0 a 1 0 , donde 0 s i g n i f i c a ’
nada , en a b s o l u t o ’ y 10 ’ t o t a l m e n t e ’ , digame , p o r f a v o r , s i d u r a n t e
l a</p>
        <p>u l t i m a semana s e ha s e n t i d o . . . _ F e l i z ’
r e s p o n s e , and ‘ n l a r g e s t ( 3 ) ‘ i s u s e d t o
s e l e c t t h e t h r e e
most common
r e s p o n s e s . The ‘ i n d e x ‘ a t t r i b u t e i s u s e d t o
g e t t h e
a c t u a l r e s p o n s e s ,
and ‘ t o l i s t ( ) ‘ i s u s e d t o c o n v e r t t h e i n d e x t o a l i s t .</p>
        <p>Meanwhile, Qwen-2.5-Coder32B does not make these kind of errors, and just outputs the desired code
without need of further postprocessing. It’s important to note that the Qwen-2.5-Coder32B model was
able to perform better with list datatypes rather than with numbers, when the former datatype tends to
stem from more dificult or complex queries.</p>
        <p>Ablation Study</p>
        <p>We relied on the results displayed in Table 1 to select the best performing LLM, which
served as the foundation for integrating the additional modules that could further enhance performance
(see Figure 1). Table 2 shows the results when varying the components of the pipeline while maintaining
Qwen-2.5-Coder32B as backbone. The AG (Answer Generator only) setup corresponds to the result
displayed in Table 1, from which the extra components of our pipeline where compared to see if there
was an actual improvement when introducing error-awareness and column pre-selection. The AG+CS
(AG with Column Selector) setup shows a clear improvement of 8 points with respect to the AG-only
model, outlining the importance of first asking the LLM to filter the relevance of the input attributes.
Lastly, when integrating the Code Fixer (CF) with an enhanced column selection (ECS) to feed richer
information about feature variations to the prompt, our final system setup (AG+ECS+CF) maintains
almost the same performance as the setup with (AG+ECS). It deals better with some datatypes, and
worse or equal with others. This means that the model is powerful enough to not input erroneous code,
thus the mistakes it makes are from not interpreting the question correctly and giving a wrong answer.</p>
        <p>AG
. AG
lid AG+CS
VAG+ECS+CF
a
t
se AG+CS
T AG+ECS+CF</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Final Test Phase</title>
        <p>The best performing configuration is almost a draw between AG+ECS and AG+ECS+CF, but we decided
to go with the latter one to participate in the competition, just in case the code fixer is able to correct
possible coding mistakes. Our zero-shot approach reached 79 points of accuracy in the task, which
ranked us in the 4th position out of 7 participants.</p>
        <p>Our results during the development phase (69 points) benefited from a significant increase of 10
points in accuracy with respect to the validation results, likely due to the complexity of the questions
present in the test set.
boolean accuracy reaches more than 80 points, while list-like types do not surpass 75 points. This might
indicate that the LLM is not able to infer these complex schemes on the test set, producing errors that
are propagated from the Column Selector module to the Answer Generator.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and Future Work</title>
      <p>In this work we propose a zero-shot approach for Tabular QA that demonstrated a strong performance
for the PRESTA task, ranking among the best systems in the development phase, although sufering from
a performance drop in the test phase. Still, our system shows that an instruction-based approach allows
to dynamically adapt to diferent dataset schemes without requiring additional training or finetuning,
surpassing the baseline model even with limited hardware resources available.</p>
      <p>Future work will focus on further refining prompt templates, improving schema adaptation,
optimizing execution eficiency or incorporating a voting system with diferent LLMs. Improving the
detection of complex datatypes is also critical, as they allow the model to answer questions on less
structured tables —which constitute the majority of online data—, ultimately making the system more
generalizable.</p>
    </sec>
    <sec id="sec-7">
      <title>Hardware Setup</title>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>Our hardware resources are somewhat limited by today’s standards. We had shared access to an Intel
Core i9-10920X at 3.50 GHz with 258 GiB RAM and two integrated NVIDIA RTX 3090, so we opted to
perform zero-shot instead of finetuning the LLMs.</p>
      <p>We
acknowledge
grants</p>
      <sec id="sec-8-1">
        <title>SCANNER-UDC (PID2020-113230RB-C21) funded by</title>
        <p>CIU/AEI/10.13039/501100011033;
GAP
(PID2022-139308OA-I00)
funded
by</p>
        <p>MI</p>
      </sec>
      <sec id="sec-8-2">
        <title>MICIU/AEI/10.13039/501100011033/ and ERDF, EU; LATCHING (PID2023-147129OB-C21) funded by MICIU/AEI/10.13039/501100011033 and ERDF, EU; CIDMEFEO funded by the Spanish National Statistics Institute (INE); as well as funding by Xunta de Galicia (ED431C 2024/02), and Centro de</title>
        <p>Investigación de Galicia “CITIC”, funded by the Xunta de Galicia through the collaboration agreement
between the Consellería de Cultura, Educación, Formación Profesional e Universidades and the Galician
universities for the reinforcement of the research centres of the Galician University System (CIGUS).</p>
        <p>CITIC, as a center accredited for excellence within the Galician University System and a member
of the CIGUS Network, receives subsidies from the Department of Education, Science, Universities,
and Vocational Training of the Xunta de Galicia. Additionally, it is co-financed by the EU through the
FEDER Galicia 2021-27 operational program (Ref.ED431G 2023/01)</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT for minor copy-editing, and GitHub
copilot for code autocompletion. After using these tool(s)/service(s), the author(s) reviewed and edited
the content as needed and take(s) full responsibility for the publication’s content.
arXiv:2310.06825.
P. Wang, S. Wang, W. Wang, S. Wu, B. Xu, J. Xu, A. Yang, H. Yang, J. Yang, S. Yang, Y. Yao, B. Yu,
H. Yuan, Z. Yuan, J. Zhang, X. Zhang, Y. Zhang, Z. Zhang, C. Zhou, J. Zhou, X. Zhou, T. Zhu, Qwen</p>
    </sec>
    <sec id="sec-10">
      <title>A. Prompts used</title>
      <sec id="sec-10-1">
        <title>A.1. Answer Generator</title>
        <p>R o l e and C o n t e x t
You a r e a Python −powered T a b u l a r Data Q u e s t i o n − Answering System . Your c o r e e x p e r t i s e l i e s i n u n d e r s t a n d i n g t a b u l a r
d a t a s e t s and c r a f t i n g Python s c r i p t s t o g e n e r a t e p r e c i s e s o l u t i o n s t o u s e r q u e r i e s .</p>
        <p>Task D e s c r i p t i o n :
G e n e r a t e Python code t o a d d r e s s a query b a s e d on t h e p r o v i d e d d a t a s e t . The o u t p u t must :
− Use t h e d a t a s e t and query a s given , a v o i d i n g any e x t e r n a l a s s u m p t i o n s .
− Adhere t o s t r i c t s y n t a x r u l e s f o r Python , e n s u r i n g t h e code r u n s f l a w l e s s l y w i t h o u t e x t e r n a l m o d i f i c a t i o n s .
− R e t a i n t h e o r i g i n a l column names o f t h e d a t a s e t i n your s c r i p t .</p>
        <p>I n p u t S p e c i f i c a t i o n
d a t a s e t : A Pandas DataFrame c o n t a i n i n g t h e d a t a t o be a n a l y z e d .</p>
        <p>q u e s t i o n : A s t r i n g o u t l i n i n g t h e s p e c i f i c query .
I n p u t :
column_names : { column_names }
q u e s t i o n : { q u e s t i o n }</p>
      </sec>
      <sec id="sec-10-2">
        <title>A.3. Code Fixer</title>
        <p>R o l e and C o n t e x t
You a r e a Python −powered T a b u l a r Data Q u e st i on − Answering System . Your c o r e e x p e r t i s e l i e s i n u n d e r s t a n d i n g t a b u l a r d a t a s e t s
and c r a f t i n g Python s c r i p t s t o g e n e r a t e p r e c i s e s o l u t i o n s t o u s e r q u e r i e s .</p>
        <p>Task D e s c r i p t i o n :
F i x t h e Python code t o a d d r e s s a query b a s e d on t h e p r o v i d e d d a t a s e t . The o u t p u t must :
− Use t h e d a t a s e t and query a s given , a v o i d i n g any e x t e r n a l a s s u m p t i o n s .
− Adhere t o s t r i c t s y n t a x r u l e s f o r Python , e n s u r i n g t h e code r u n s f l a w l e s s l y w i t h o u t e x t e r n a l m o d i f i c a t i o n s .
− R e t a i n t h e o r i g i n a l column names o f t h e d a t a s e t i n your s c r i p t .</p>
        <p>I n p u t S p e c i f i c a t i o n
code : The Python code t h a t n e e d s t o be f i x e d .</p>
        <p>e r r o r : The e r r o r message t h a t r e s u l t s from r u n n i n g t h e code .</p>
        <p>Output S p e c i f i c a t i o n</p>
        <p>R e t u r n o n l y t h e Python code t h a t s o l v e s t h e query i n t h e f u n c t i o n , e x c l u d i n g any i n t r o d u c t o r y e x p l a n a t i o n s or comments .</p>
        <p>The f u n c t i o n must :
I n c l u d e a l l e s s e n t i a l i m p o r t s .</p>
        <p>Be c o n c i s e and f u n c t i o n a l , e n s u r i n g t h e s c r i p t can be e x e c u t e d w i t h o u t a d d i t i o n a l m o d i f i c a t i o n s .</p>
        <p>Use t h e d a t a s e t and r e t u r n a r e s u l t o f t y p e number , c a t e g o r i c a l v a l u e , b o o l e a n v a l u e , or a l i s t o f v a l u e s .
Code :</p>
        <p>Below i s t h e p i e c e o f code t h a t n e e d s t o be f i x e d , a l o n g with t h e e r r o r message t h a t r e s u l t s from r u n n i n g t h e code :
{ r e s p o n s e }
E r r o r : { e r r o r }</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Herzig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Nowak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Piccinno</surname>
          </string-name>
          , J. Eisenschlos, TaPas: Weakly Supervised Table Parsing via Pre-training, in: D.
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Chai</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Schluter</surname>
          </string-name>
          , J. Tetreault (Eds.),
          <article-title>Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>4320</fpage>
          -
          <lpage>4333</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>398</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>398</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Yin</surname>
          </string-name>
          , G. Neubig, W.-t. Yih, S. Riedel,
          <article-title>TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data</article-title>
          , in: D.
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Chai</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Schluter</surname>
          </string-name>
          , J. Tetreault (Eds.),
          <article-title>Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>8413</fpage>
          -
          <lpage>8426</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>745</volume>
          /. doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>745</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <article-title>Seq2SQL: Generating Structured Queries from Natural Language using Reinforcement Learning</article-title>
          ,
          <year>2017</year>
          . URL: https://arxiv.org/abs/1709.00103. arXiv:
          <volume>1709</volume>
          .
          <fpage>00103</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , D. Radev,
          <article-title>TypeSQL: Knowledge-Based Type-Aware Neural Text-to-SQL generation</article-title>
          , in: M.
          <string-name>
            <surname>Walker</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Stent (Eds.),
          <source>Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>2</volume>
          (
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , New Orleans, Louisiana,
          <year>2018</year>
          , pp.
          <fpage>588</fpage>
          -
          <lpage>594</lpage>
          . URL: https://aclanthology.org/N18-2093/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N18</fpage>
          -2093.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>V.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          , E. Kanoulas, M. de Rijke,
          <article-title>MultiTabQA: Generating Tabular Answers for MultiTable Question Answering</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>6322</fpage>
          -
          <lpage>6334</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>348</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>348</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herbert-Voss</surname>
          </string-name>
          , G. Krueger,
          <string-name>
            <given-names>T.</given-names>
            <surname>Henighan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Child</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ziegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Winter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hesse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          , E. Sigler,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chess</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Berner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>McCandlish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Sutskever</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language Models are Few-Shot Learners</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2020</year>
          , pp.
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2020/file/ 1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fried</surname>
          </string-name>
          ,
          <article-title>API-Assisted Code Generation for Question Answering on Varied Table Structures</article-title>
          , in: H.
          <string-name>
            <surname>Bouamor</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Pino</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Bali (Eds.),
          <source>Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Singapore,
          <year>2023</year>
          , pp.
          <fpage>14536</fpage>
          -
          <lpage>14548</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .emnlp-main.
          <volume>897</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .emnlp-main.
          <volume>897</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Osés-Grijalba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Ureña-López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Cámara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-Collados</surname>
          </string-name>
          , Overview of PRESTA at IberLEF 2025:
          <article-title>Question Answering Over Tabular Data In Spanish, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Rajpurkar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lopyrev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          , Squad:
          <volume>100</volume>
          ,000+ Questions for Machine Comprehension of Text,
          <year>2016</year>
          . URL: https://arxiv.org/abs/1606.05250. arXiv:
          <volume>1606</volume>
          .
          <fpage>05250</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Salakhutdinov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          ,
          <article-title>HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering</article-title>
          , in: E.
          <string-name>
            <surname>Rilof</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Chiang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Hockenmaier</surname>
          </string-name>
          , J. Tsujii (Eds.),
          <source>Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Brussels, Belgium,
          <year>2018</year>
          , pp.
          <fpage>2369</fpage>
          -
          <lpage>2380</lpage>
          . URL: https://aclanthology.org/D18-1259/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D18</fpage>
          -1259.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Pasupat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Compositional Semantic Parsing on Semi-Structured Tables</article-title>
          , in: C.
          <string-name>
            <surname>Zong</surname>
          </string-name>
          , M. Strube (Eds.),
          <source>Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing</source>
          (Volume
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Beijing, China,
          <year>2015</year>
          , pp.
          <fpage>1470</fpage>
          -
          <lpage>1480</lpage>
          . URL: https://aclanthology.org/P15-1142/. doi:
          <volume>10</volume>
          .3115/v1/
          <fpage>P15</fpage>
          -1142.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Iyyer</surname>
          </string-name>
          , W.-t. Yih, M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <article-title>Search-based Neural Structured Learning for Sequential Question Answering</article-title>
          , in: R. Barzilay, M.-Y. Kan (Eds.),
          <source>Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Vancouver, Canada,
          <year>2017</year>
          , pp.
          <fpage>1821</fpage>
          -
          <lpage>1831</lpage>
          . URL: https://aclanthology.org/P17-1167/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P17</fpage>
          -1167.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J. Osés</given-names>
            <surname>Grijalba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Ureña-López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. Martínez</given-names>
            <surname>Cámara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-Collados</surname>
          </string-name>
          ,
          <article-title>Question Answering over Tabular Data with DataBench: A Large-Scale Empirical Evaluation of LLMs</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            , M.-
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Kan</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Hoste</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sakti</surname>
          </string-name>
          , N. Xue (Eds.),
          <source>Proceedings of the 2024 Joint International Conference on Computational Linguistics</source>
          ,
          <article-title>Language Resources and Evaluation (LREC-COLING 2024), ELRA</article-title>
          and
          <string-name>
            <given-names>ICCL</given-names>
            ,
            <surname>Torino</surname>
          </string-name>
          , Italia,
          <year>2024</year>
          , pp.
          <fpage>13471</fpage>
          -
          <lpage>13488</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          . lrec-main.
          <volume>1179</volume>
          /.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/N19-1423/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Dong</surname>
          </string-name>
          , Z. Cheng, F. Cheng, S. Han,
          <string-name>
            <surname>D</surname>
          </string-name>
          . Zhang, TaCube:
          <article-title>Pre-computing Data Cubes for Answering Numerical-Reasoning Questions over Tabular Data</article-title>
          , in: Y.
          <string-name>
            <surname>Goldberg</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Kozareva</surname>
          </string-name>
          , Y. Zhang (Eds.),
          <source>Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing</source>
          , Association for Computational Linguistics, Abu Dhabi, United Arab Emirates,
          <year>2022</year>
          , pp.
          <fpage>2278</fpage>
          -
          <lpage>2291</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .emnlp-main.
          <volume>145</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          . emnlp-main.
          <volume>145</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, BART:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          , in: D.
          <string-name>
            <surname>Jurafsky</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Chai</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Schluter</surname>
          </string-name>
          , J. Tetreault (Eds.),
          <article-title>Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Online,
          <year>2020</year>
          , pp.
          <fpage>7871</fpage>
          -
          <lpage>7880</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          /. doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2020</year>
          .acl-main.
          <volume>703</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Ge</surname>
          </string-name>
          , Y. Han,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <source>Technical Report</source>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2309.16609. arXiv:
          <volume>2309</volume>
          .
          <fpage>16609</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A. Q.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sablayrolles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bamford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Chaplot</surname>
          </string-name>
          , D. de las Casas,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bressand</surname>
          </string-name>
          , G. Lengyel,
          <string-name>
            <given-names>G.</given-names>
            <surname>Lample</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Saulnier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. R.</given-names>
            <surname>Lavaud</surname>
          </string-name>
          , M.
          <article-title>-</article-title>
          <string-name>
            <surname>A. Lachaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Stock</surname>
            ,
            <given-names>T. L.</given-names>
          </string-name>
          <string-name>
            <surname>Scao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lavril</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Lacroix</surname>
            ,
            <given-names>W. E.</given-names>
          </string-name>
          <string-name>
            <surname>Sayed</surname>
          </string-name>
          , Mistral 7B,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2310.06825.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J. O.</given-names>
            <surname>Grijalba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A. U.</given-names>
            <surname>López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-Collados</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Cámara</surname>
          </string-name>
          ,
          <article-title>Towards quality benchmarking in question answering over tabular data in spanish</article-title>
          ,
          <source>Proces. del Leng. Natural</source>
          <volume>73</volume>
          (
          <year>2024</year>
          )
          <fpage>283</fpage>
          -
          <lpage>296</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>