<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>AIInFunds. Intelligent Services for Monitoring Tendencies of Alternative Investment Funds</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>José Antonio García-Díaz</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Antonio Miñarro-Giménez</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ángela Almela</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gema Alcaraz-Mármol</string-name>
          <email>Gema.Alcaraz@uclm.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>María José Marín-Pérez</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francisco García-Sánchez</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rafael Valencia-García</string-name>
          <email>valencia@um.es</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Departamento de Filología Moderna, Universidad de Castilla La Mancha</institution>
          ,
          <addr-line>45071</addr-line>
          ,
          <country>España</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Facultad de Informática, Universidad de Murcia, Campus de Espinardo</institution>
          ,
          <addr-line>30100 Murcia</addr-line>
          ,
          <country>España</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Facultad de Letras, Universidad de Murcia</institution>
          ,
          <addr-line>Campus de la Merced, 30001, Murcia</addr-line>
          ,
          <country>España</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Alternative investment funds have high strategic value but high uncertainty associated with them because they are relatively new and dificult to monitor using traditional approaches. In addition, being such specialized products there is little public information. The main objective of the AIInFunds project is to use Natural Language Processing and Semantic Web technologies to extract information about alternative assets and apply entity extraction techniques and sentiment and emotion analysis to measure social perception towards these alternative assets. AIInFunds is being developed by the TECNOMOD group of the University of Murcia and is financed by the Spanish National Research Agency and by the European Union NextGenerationEU/PRTR through the Prueba de Concepto 2021 projects.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Large Language Models</kwd>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Emotion Analysis</kwd>
        <kwd>Semantic Web</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction and main objective</title>
      <p>
        from news and social networks were also implemented
and a large number of financial corpora were collected,
This project is funded by the Prueba de Concepto 2021 which were only partially annotated.
call1 and is based on a previous project KBS4FIA [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], Alternative investment funds have a high strategic
whose objective was the development of Natural Lan- value and have become fashionable in this sector. Some
guage Processing and knowledge extraction technologies examples of these assets are venture capital and certain
within the financial domain to enable a better manage- physical assets such as infrastructure or real estate. One
ment of unstructured information for decision-making of the reasons for the success of alternative assets is that
within this domain. Knowledge extraction technologies the foreseeable future of traditional investment options is
based on Spanish texts were developed including named worse, which is why investors have opted for assets with
entity recognition, text classification, semantic annota- higher return expectations, although riskier. However,
tion and ontology population. Deep-learning technolo- for the eficient use of these assets certain challenges
gies were also developed to process subjective language should be considered. First, since they have not been
in Spanish, which involved sentiment analysis in difer- used traditionally, investors are not familiar with the
ent domains, including the financial domain. In addition, risks involved. Second, these assets require a high degree
technologies for the eficient extraction of information of specialization, as well as extensive knowledge of the
existing legislation on them. Third, these assets are
often illiquid, which makes it dificult both to determine
their market price and sell them quickly. Fourth, they
require a long-term investment, which, in some cases,
can be as long as 10 to 15 years. Additionally, these assets
have high management costs, including higher fees than
traditional assets, which may also put of some
potential investors. Finally, these assets raise doubts among
investors who are wary due to the little historical and
analytical information available. The scarce number of
people interested in investing in such specialized
products has caused a relative lack of information, which, to
top it all of, is often not made public. In fact, in Spain
there are very few companies dedicated to managing or
advising on this type of assets.
      </p>
      <p>
        To assess the opinion and experience of Spanish in- data is being recorded in diferent categories to be able to
vestors with alternative assets, AFI Trust conducted a carry out diferent studies, such as feelings or emotions.
survey among a group of entities with diferent profiles Tools such as web crawlers and the UMUCorpusClassifier
and investment objectives: (1) insurance companies and [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] are being used for the compilation and annotation
pension funds, (2) banks private, (3) foundations or asso- processes. The second work package consists of training
ciations, and (4) non-financial corporations. The report and optimizing deep learning-based models for
concep[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] highlights liquidity restrictions and the dificulty of tual aspect modeling and emotion analysis in the domain
analysis among the main obstacles when it comes to of alternative investment funds. The last two work
packinvesting in this type of assets. Due to the motivation ages consist of the development of a global platform for
described above, this project consists in the creation of a the monitoring of alternative assets based on a
microserplatform that allows investors to create a series of alterna- vices architecture. The final work package is concerning
tive assets and facilitates the monitoring of information the customization and validation of the global platform
found in documents on the Internet, as well as in publi- in two scenarios related to alternative investment funds.
cations on social networks or specialized news. In this Some of the most important milestones reached so far
way, we aim to fill the non-specialized knowledge gap in the project are described next.
that currently exists about alternative assets.
      </p>
      <p>We intend to use Natural Language Processing tech- 2.1. FINA
nologies to, first, identify concepts related to these assets
and, second, apply sentiment analysis techniques focused
on measuring people’s perception of these issues. This
will be available through a customizable control panel
that will allow to filter information geographically using
time intervals. At a technological level, this platform is
integrating the solutions developed during the execution
of the KBS4FIA project, which are being optimized to
be eficient and scalable through their implementation
on a real salable platform. In addition, a microservices
architecture is being built to commercialize the platform
and be able to integrate and adapt these technologies to
software development companies.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Project status</title>
      <p>The AIInFunds project has the following specific
objectives: (OB1) Development and annotation of linguistic
resources in the alternative investment funds domain,
(OB2) Development and optimization of deep
learningbased models for conceptual aspect modelling and
emotion analysis in the alternative investment funds domain,
(OB3) Integration and optimization of modules in a global
platform, (OB4) Deployment of the global platform in a
Software-as-a-Service commercial and scalable
environment, and (OB5) Customization and validation of the
global platform in various scenarios.</p>
      <p>
        This project was scheduled in 24 months starting the
1st of December 2021 and was divided into two
management and five development work packages. The five
development-related work packages are described next.
The first work package consists of compiling and
annotating linguistic resources in the domain of alternative
investment funds. Here, diferent resources and lexicons
from the financial domain, including news corpus, social
networks and domain ontologies, are being compiled for
the semantic exploitation of the results. The compiled
One of the objectives of this project is the development
of a language model, namely FINA, focused on the
domain of economics and finances. To do this, we trained
a model based on the RoBERTa [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] architecture with a
corpus compiled from 5 gigabytes of financial news. To
compile the corpus we used the Spatie crawler2. We
selected more than 100 newspapers with financial content.
Some examples of these newspapers are: Expansión3,
Invertia4, ABC 5 or ModaEs6. It is worth noting that not all
newspapers are from Spain. Some of them are from other
Spanish-spearking countries such as Diario Financiero7
from Chile, or El Financiero8 from Mexico. As not all the
newspapers are focused solely on providing financial
information, the news were extracted by applying diferent
iflters. The first filter was based on regular expressions
within the news URLs. Thus, it was possible to focus
on news items that were located within sections such
as /finanzas/ or /economia/ of the processed web
portals. The second filter was based on CSS rules, so
that we could keep the relevant content, discarding from
the HTML code advertisements, links to related news or
unrelated content.
      </p>
      <p>
        The HTML tags were removed from the compiled news
items, which were then converted into Markdown format.
We chose this format because we wanted to keep the basic
structural elements of each news item, such as headlines
or sections. The next step was cleaning the corpus. To
do this, a script was developed that removes irrelevant
data from the news items (e.g., dates or information from
the authors who have written the journalistic piece).
2https://github.com/spatie/crawler
3https://www.expansion.com
4https://www.elespanol.com/invertia
5https://www.abc.es
6https://www.modaes.com/
7https://www.df.cl/
8https://www.elfinanciero.com.mx/
2.3. Targeted Sentiment Analysis
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Also diferent pre-trained transformer-based
models were evaluated, namely, (i) BETO [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], (ii) ALBETO
and (iii) DistillBETO, which are light variants of BETO
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], (iv) MarIA [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and (v) Bertin [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], both based on the
RoBERTa architecture, and (vi) multilingual BERT [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]
and (vii) XLM [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], two multilingual models.
      </p>
      <sec id="sec-2-1">
        <title>We are currently training two configurations based on</title>
        <p>RoBERTa. The first model has 12 attention heads and 12
hidden layers. The second model also has 12 attention
heads, but only 6 hidden layers. The tokenizer model
is Byte-Level BPE with the following parameters: a
maximum of 512 tokens, a vocabulary size of 52,262 (the
RoBERTa default), and a minimum frequency of 2. The
dataset is processed paragraph by paragraph, that is, we
extract multiple texts from each news. Finally, it is worth
mentioning that the model is trained with the Masked
Language Modeling task, with a probability of 15% and
for 5 epochs.</p>
        <p>
          To validate the suitability of the model and its
applicability to diferent tasks, we are testing its performance
on a sentiment analysis task with a financial corpus. We
are comparing the results with more general Spanish
pretrained models (e.g., BETO [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], MarIA [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], BERTIN [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ])
and other multilingual models.
        </p>
        <p>
          Another of the tasks involved in this project is the
development of a targeted sentiment analysis approach in
the financial domain. To do this, we compiled a corpus
with close to 80,000 texts, coming from social networks
and news headlines. The idea behind this targeted
classification method is to predict the sentiment polarities
of diferent targets in the same text. In classic sentiment
analysis system, only the sentiment towards the main
entity or topic is calculated. In this approach, however,
the model is trained to obtain the sentiments towards
2.2. Sentiment analysis and evaluation of three types of entities: (1) towards the main economic
target (MET), (2) towards other companies, and (3)
tolanguage models wards society in general. In addition, this model is also
In [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] we explore the impact of combining diferent fea- capable of recognizing the entity (i.e., the MET) using
ture sets for sentiment analysis in financial texts in Span- a sequence classification model, in the style of Named
ish. To do this, we compiled a corpus of 15,915 tweets and Entity Recognition.
annotated them as either positive, negative, or neutral. For the purposes of this experiment, in the first place,
Then, features based on word embeddings were evalu- a subset of the corpus of news and tweets was selected
ated along with linguistic features [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. These features and the MET and sentiment towards these three entities
were evaluated both individually and combined using en- (i.e., the MET, other companies, and society in general)
semble learning and knowledge integration. The ensem- were manually annotated. Then, diferent classifiers were
ble learning strategy consists into generate a new output trained evaluating various language models, both specific
based on the outputs of the rest of the models. For this, we to Spanish and multilingual. The results of this study can
evaluate the mode of the predictions, averaging probabili- be found at [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
ties, selecting the label with the highest probability along Another contribution in this field is the organization of
all the models and a weighted mode based on custom vali- the FinancES 2023 [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] shared-task, which is held within
dation results. The second evaluated strategy, knowledge the IberLEF 2023 workshop. This shared task uses a
integration, is the retrain of a new multi-input neural subset of the compiled corpora and proposes two tasks
network model that combines all the feature sets and to the participants. On the one hand, to identify the
outpus a unique result. The best approach achieved an main entity that appears in the text and its sentiment
F1 of 73.15880%, combining the features evaluated using and, on the other hand, the sentiments towards other
the knowledge integration strategy. The non-contextual companies and society as a whole. The FinancES 2023
word and sentence embeddings considered in our study dataset contains 6359 documents for training and 1621
are as follows: GloVe [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], Word2Vec [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] and fastText documents for testing. Table 1 contains some examples
of the dataset, the MET and the sentiments towards this
target (S. MET), other companies (S. Companies), and
society (S. Society). This competition is hosted in the
Codalab platform9.
        </p>
        <p>
          In order to facilitate the participation we have allowed
the participants to send their results for both subtasks
independently. Besides, we prepared two notebooks to
show to the participants how to train a baseline model
based on TF–IDF features trained with logistic regression
for the sentiment polarity detection task, and a model
based on Spacy [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] for detecting the MET. These
notebooks also contain instructions about how to prepare the
submission files required to participate in the
competition.
        </p>
        <p>At the end of the evaluation stage, a total of 10 teams
have participated, achieving competitive results in both
subtasks.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Further work</title>
      <sec id="sec-3-1">
        <title>As future work, we are working on developing a global</title>
        <p>
          platform in a commercial environment with the aim to
deliver the functionality through a Software-as-a-Service
model. Once developed, we will validate it based on
common use cases to assess its usefulness in real scenarios.
One major line of improvement is to incorporate
opendata sources to the platform. As these data are usually
available through diferent APIs, we are exploring the
reliability of using code generation tools based on LLMs
models to automate this process [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ].
        </p>
        <p>Once we finish with the evaluation of the FINA model
in several NLP tasks, we will release two versions (a large
one with 12 hidden layers and a lightweight one with
only 6 hidden layers) for the research community and
industry in Huggingface10 together with a manuscript
with the details of its compilation and the analysis and
benchmark of this model applied to the diferent NLP
tasks.</p>
        <p>
          Concerning the organization of shared tasks, next year
we will focus on NLP tasks dealing with financial data
in Spanish. However, we want to incorporate texts that
include comments and opinions from blogs in order to
introduce a more informal speech and the presence of
ifgurative language [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Acknowledgments</title>
      <sec id="sec-4-1">
        <title>This work is part of the research projects</title>
        <p>AIInFunds (PDC2021-121112-I00) funded by
MCIN/AEI/10.13039/501100011033 and by the
European Union NextGenerationEU/PRTR.</p>
      </sec>
      <sec id="sec-4-2">
        <title>9https://codalab.lisn.upsaclay.fr/competitions/10052 10https://huggingface.co</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>García-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Paredes-Valverde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Alcaraz-Mármol, Á. Almela, KBS4FIA: leveraging advanced knowledge-based systems for financial information analysis</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>59</volume>
          (
          <year>2017</year>
          )
          <fpage>145</fpage>
          -
          <lpage>148</lpage>
          . URL: http://journal.sepln.org/sepln/ojs/ojs/index. php/pln/article/view/5507.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>AFI</given-names>
            <surname>: Analistas Financieros</surname>
          </string-name>
          <string-name>
            <surname>Internacionales</surname>
          </string-name>
          ,
          <article-title>Inversión en activos alternativos</article-title>
          , https://media.afi.es/webcorporativa/2022/07/ Estudio-activos
          <article-title>-alternativos-Afi-Aberdeen_ DEF_Actualizacion-JUN22</article-title>
          .pdf ,
          <year>2022</year>
          . Accessed:
          <fpage>2023</fpage>
          -05-25.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          , Á. Almela,
          <string-name>
            <given-names>G.</given-names>
            <surname>Alcaraz-Mármol</surname>
          </string-name>
          , R. Valencia-García,
          <article-title>UMUCorpusClassifier: Compilation and evaluation of linguistic corpus for natural language processing tasks</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>65</volume>
          (
          <year>2020</year>
          )
          <fpage>139</fpage>
          -
          <lpage>142</lpage>
          . URL: http://journal.sepln.org/sepln/ojs/ojs/index. php/pln/article/view/6292.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cañete</surname>
          </string-name>
          , G. Chaperon,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fuentes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pérez</surname>
          </string-name>
          ,
          <article-title>Spanish pre-trained BERT model and evaluation data</article-title>
          ,
          <source>in: PML4DC at ICLR</source>
          <year>2020</year>
          ,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Gutiérrez-Fandiño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Armengol-Estapé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pàmies</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Llop-Palao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Silveira-Ocampo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Carrino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Armentano-Oller</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>RodriguezPenagos, A</article-title>
          .
          <string-name>
            <surname>Gonzalez-Agirre</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Villegas, MarIA: Spanish language models</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>68</volume>
          (
          <year>2022</year>
          )
          <fpage>39</fpage>
          -
          <lpage>60</lpage>
          . URL: http://journal.sepln.org/sepln/ojs/ojs/index.php/ pln/article/view/6405.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>J. de la Rosa</surname>
            ,
            <given-names>E. G.</given-names>
          </string-name>
          <string-name>
            <surname>Ponferrada</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Romero</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Villegas</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>González de Prado Salas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Grandury, BERTIN: eficient pre-training of a spanish language model using perplexity sampling</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>68</volume>
          (
          <year>2022</year>
          )
          <fpage>13</fpage>
          -
          <lpage>23</lpage>
          . URL: http://journal.sepln.org/sepln/ojs/ojs/index. php/pln/article/view/6403.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>García-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. ValenciaGarcía</surname>
          </string-name>
          ,
          <article-title>Smart analysis of economics sentiment in spanish based on linguistic features and transformers</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>14211</fpage>
          -
          <lpage>14224</lpage>
          . URL: https://doi.org/10.1109/ACCESS.
          <year>2023</year>
          .
          <volume>3244065</volume>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2023</year>
          .
          <volume>3244065</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Vivancos-Vicente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Almela</surname>
          </string-name>
          , R. Valencia-García,
          <article-title>UMUTextStats: A linguistic feature extraction tool for spanish</article-title>
          ,
          <source>in: Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>6035</fpage>
          -
          <lpage>6044</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , Glove:
          <article-title>Global vectors for word representation</article-title>
          ,
          <source>in: Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP)</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Eficient estimation of word representations in vector space</article-title>
          ,
          <source>arXiv preprint arXiv:1301.3781</source>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>E.</given-names>
            <surname>Grave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <article-title>Learning word vectors for 157 languages</article-title>
          , arXiv preprint arXiv:
          <year>1802</year>
          .
          <volume>06893</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cañete</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Donoso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bravo-Marquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Carvallo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Araujo</surname>
          </string-name>
          ,
          <article-title>ALBETO and DistilBETO: Lightweight spanish language models</article-title>
          ,
          <source>arXiv preprint arXiv:2204.09145</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Unsupervised crosslingual representation learning at scale</article-title>
          , CoRR abs/
          <year>1911</year>
          .02116 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1911</year>
          .02116. arXiv:
          <year>1911</year>
          .02116.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          , CoRR abs/
          <year>1810</year>
          .04805 (
          <year>2018</year>
          ). URL: http://arxiv.org/abs/
          <year>1810</year>
          .04805. arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Garcia-Sanchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Evaluation of transformer models for financial targeted sentiment analysis in spanish</article-title>
          ,
          <source>PeerJ Computer Science</source>
          <volume>9</volume>
          (
          <year>2023</year>
          )
          <article-title>e1377</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>García-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Valencia</given-names>
            <surname>García</surname>
          </string-name>
          , Overview of FinancES 2023:
          <article-title>Financial targeted sentiment analysis in spanish (to appear</article-title>
          ),
          <source>Procesamiento del Lenguaje Natural</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>X.</given-names>
            <surname>Schmitt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kubler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Robert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Papadakis</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <article-title>LeTraon, A replicable comparison study of NER software: StanfordNLP, NLTK</article-title>
          , OpenNLP, SpaCy, Gate, in: 2019
          <source>Sixth International Conference on Social Networks Analysis, Management and Security (SNAMS)</source>
          , IEEE,
          <year>2019</year>
          , pp.
          <fpage>338</fpage>
          -
          <lpage>343</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Yetiştiren</surname>
          </string-name>
          , I. Özsoy,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ayerdem</surname>
          </string-name>
          , E. Tüzün,
          <article-title>Evaluating the code quality of ai-assisted code generation tools: An empirical study on github copilot, amazon codewhisperer, and chatgpt</article-title>
          ,
          <source>arXiv preprint arXiv:2304.10778</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>M. del Pilar</surname>
            Salas-Zárate,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Alor-Hernández</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>Sánchez-Cervantes</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          <string-name>
            <surname>Paredes-Valverde</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <string-name>
            <surname>García-Alcaraz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Valencia-García</surname>
          </string-name>
          ,
          <article-title>Review of English literature on figurative language applied to social networks</article-title>
          ,
          <source>Knowledge and Information Systems</source>
          <volume>62</volume>
          (
          <year>2020</year>
          )
          <fpage>2105</fpage>
          -
          <lpage>2137</lpage>
          . doi:
          <volume>10</volume>
          .1007/ s10115-019-01425-3.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>