=Paper= {{Paper |id=Vol-3178/CIRCLE_2022_paper_25 |storemode=property |title=On the Usability of Transformers-based models for a French Question-Answering task - abstract |pdfUrl=https://ceur-ws.org/Vol-3178/CIRCLE_2022_paper_25.pdf |volume=Vol-3178 |authors=Oralie Cattan,Christophe Servan,Sophie Rosset |dblpUrl=https://dblp.org/rec/conf/circle/CattanSR22 }} ==On the Usability of Transformers-based models for a French Question-Answering task - abstract == https://ceur-ws.org/Vol-3178/CIRCLE_2022_paper_25.pdf
On the Usability of Transformers-based models for a
French Question-Answering task - abstract
Oralie Cattan1,2,*,† , Christophe Servan1,2,† and Sophie Rosset2,†
1
    Paris-Saclay University, CNRS, LISN
2
    QWANT


                                         Abstract
                                         Transformers have sparked a paradigmatic shift in question-answering training practices by simplifying
                                         its architectures. As models became larger and better important usability shortcomings appeared. This
                                         includes their computational costs and degraded performance with limited training data (e.g., domain-
                                         specific or low-resourced language tasks). Considering this resource trade-off, we (i) explore training
                                         strategies such as data augmentation, hyperparameter optimization, cross-lingual transfers and cross-
                                         dataset mixing, (ii) perform an in-depth analysis to understand the contribution of each on model
                                         performance maintenance and (iii) provide a question-answering corpus and a compressed pre-trained
                                         model for French1 . Our experimental results attest to the merit of a flexible paradigm for a low-resource
                                         scenario.

                                         Keywords
                                         question answering, transformer architectures, pre-trained models and scalability, language resources




1
 Both are available at: https://huggingface.co/qwant
CIRCLE (Joint Conference of the Information Retrieval Communities in Europe) 2022, July 04–07, 2022, Samatan, France
*
  Corresponding author.
†
  These authors contributed equally.
$ oralie.cattan@lisn.fr (O. Cattan); christophe.servan@lisn.fr (C. Servan); sophie.rosset@lisn.fr (S. Rosset)
 0000-0003-2805-5620 (O. Cattan); 0000-0003-2306-7075 (C. Servan); 0000-0002-6865-4989 (S. Rosset)
                                       © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
    CEUR
    Workshop
    Proceedings
                  http://ceur-ws.org
                  ISSN 1613-0073
                                       CEUR Workshop Proceedings (CEUR-WS.org)