<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>E. L. Estevanell-Valladares)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Automatically Designing Machine Learning Models out of Natural Language</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ernesto Luis Estevanell-Valladares</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Software and Computer Systems, University of Alicante</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Mathematics and Computer Science, University of Havana</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>The popularity of artificial intelligence has led to an increasing need for machine learning models tailored to specific needs. AutoML aims to automate the process of creating efective machine-learning solutions, but current systems need to be more versatile to meet the demand. While more flexible and extensible heterogeneous systems have overcome many limitations of traditional AutoML systems, they lack accessibility due to their programmatic interfaces. We propose a research project to address this issue to develop a heterogeneous AutoML system that can produce optimal machine-learning pipelines using a natural language interface.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;AutoML</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Large Language Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Machine learning has expanded rapidly, presenting researchers and practitioners with many
new algorithms and data sets. However, selecting the most suitable strategy for a given issue
has become increasingly complex, requiring extensive experimentation and technical expertise.
AutoML has emerged as a solution to this problem by providing powerful tools to search through
large machine-learning pipelines [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Nevertheless, the range of possible techniques for natural
language processing is vast, making it hard to combine and compare diferent algorithms. Thus,
AutoML algorithms must agree on a standard protocol for sharing outputs as inputs for any
other algorithm.
      </p>
      <p>
        To achieve the primary goal of AutoML, the systems must have interfaces that are easy to
use for those with limited computer science and machine learning knowledge. Furthermore,
the systems should have strong generalization capabilities to create tools that can be utilized in
various scenarios and produce machine learning models that can be applied to a wide range
of applications. However, current AutoML systems focus on a specific set of algorithms, often
tailored to a library or toolkit [
        <xref ref-type="bibr" rid="ref4">2, 3, 4</xref>
        ]. This reduces their ability to explore various algorithms
from diferent domains and find optimal solutions to complex, multifaceted problems. In
contrast, Heterogeneous AutoML systems generate learning solutions by mixing techniques
from diferent domains [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. However, they do not provide natural interfaces for novice users
in programming or AutoML, and their execution requires preparation of the environment, the
definition of the problem in appropriate terms, and the provision of data in specific formats [
        <xref ref-type="bibr" rid="ref4 ref6 ref7 ref8">2,
3, 4, 6, 7, 8, 9</xref>
        ].
      </p>
      <p>
        Using natural language as a user interface can significantly enhance the accessibility and
user-friendliness of AutoML. Large Language Models (LLMs) are trendy for their ability to
process raw text and efectively identify patterns and connections in data [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. However,
these models can sometimes generate incorrect responses or need help with inference tasks
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Recent studies suggest that incorporating external knowledge significantly improves the
performance of LLMs [13].
      </p>
      <p>Researchers are exploring combining these techniques to address the limitations of both
AutoML and LLMs. Shen et al. [14] employed an LLM to process queries and generate learning
task planning using pre-trained models. This approach has successfully integrated image, audio,
and text prediction, classification, and processing capabilities into a single system. However,
it does not focus on optimizing model selection or hyperparameter optimization. It does not
produce tuned learning models in standalone programs or allow for the export of inference
power to arbitrary environments.</p>
      <p>The research project consists of producing a Heterogeneous AutoML system that integrates
natural language processing as its primary interface. The ultimate goal is to design a tool to
generate optimal machine learning models that are flexible and adaptable to diferent contexts
and heterogeneous situations. This leads to our Main Research Question: ”In what way can
we integrate Natural Language into a Heterogeneous AutoML process?”.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Most AutoML systems only use limited algorithms specific to a particular library or toolkit.
This limitation hinders their ability to solve complex problems by exploring various algorithms
from diferent areas. On the other hand, Heterogeneous AutoML systems combine techniques
from multiple domains to create better learning solutions. However, they require a user-friendly
interface for those new to programming or AutoML. This setup involves defining the problem
correctly, providing data in specific formats, and setting up the environment.</p>
      <p>Table 1 contrasts several existing AutoML systems with the system proposed in this research
regarding their capabilities of dealing with heterogeneous scenarios. This evaluation focused
on their capacity to handle diverse scenarios encompassing multiple algorithms. It is worth
noting that this assessment was solely based on their ability to handle heterogeneous algorithms
without consideration for their overall performance, capacity, or applicability.</p>
      <p>
        AutoML systems vary in capabilities and limitations, depending on the specific learning
libraries they are built on. Some, like Auto-Sklearn [3], Auto-Weka [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and Auto-Keras[2], are
restricted to using Scikit-learn [15], Weka [
        <xref ref-type="bibr" rid="ref13">16</xref>
        ], and Keras [
        <xref ref-type="bibr" rid="ref14">17</xref>
        ], respectively. Other systems,
such as RECIPE [9] and Hyperopt [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], can incorporate algorithms from diferent libraries but
require a concrete implementation. TPOT [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and ML-Plan [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] provide a more flexible approach,
combining technologies from multiple learning libraries to create concrete implementations of
learning pipelines. AutoML systems are mostly focused on supervised learning, but some ofer
the potential to integrate unsupervised learning functionality, like Hyperopt.
      </p>
      <p>AutoML Systems
Hyperopt
Auto-Weka 2.0
RECIPE
TPOT
ML-Plan
Auto-Keras
Auto-Sklearn 2.0
AutoGOAL
≈
≈
✓
✓
✓
s
e
i
r
a
r
b
i
l
e
l
p
i
t
l
u
M
✓
✓
s
m
e
l
b
o
r
p
L
M
e
l
p
i
t
l
u
M
✓
✓
✓
✓
✓
c
i
t
s
i
l
i
b
a
b
o
r
P
✓
✓
✓
✓
✓
✓
e
l
b
i
s
n
e
t
x
E
✓
✓
y
r
e
v
o
c
s
i
D
c
i
t
a
m
o
t
u
A
✓
≈
✓
✓
✓
✓
✓
e
v
i
t
c
e
j
b
o
i
t
l
u
M
✓
✓
✓
✓
✓
d
e
t
u
b
i
r
t
s
i
D
✓
✓
✓
✓
✓
✓</p>
      <p>Systems like Auto-Sklearn and Auto-Keras benefit from a unified underlying API, while
modularly designed systems such as ML-Plan allow the addition and modification of algorithms.
Several systems allow balancing diferent objectives or metrics during optimization, which is
relevant in multiple development and research scenarios. For example, Auto-Keras and ML-Plan
use a weighted sum approach to combine multiple evaluation metrics into a single objective
function. These systems allow you to specify both the metrics and the weights assigned to each
metric, which allows you to control their relative importance in the optimization process.</p>
      <p>
        Hyperopt, ML-Plan, Auto-WEKA, and Auto-Sklearn include mechanisms to distribute search
processes and resources among multiple computers, optimizing search time and generating
learning pipelines more quickly. To meet the goal of AutoML, system solutions should be easily
usable in arbitrary environments and applicable as portable machine learning algorithms outside
the AutoML context. Systems such as Hyperopt, TPOT, and Auto-Sklearn allow exporting the
best pipeline found during the search process as a Python script, while Auto-WEKA can export
to JAVA [
        <xref ref-type="bibr" rid="ref15">18</xref>
        ] code. Auto-Keras allows exporting the models in various formats, including
TensorFlow [
        <xref ref-type="bibr" rid="ref16">19</xref>
        ], PyTorch [20], and Keras [
        <xref ref-type="bibr" rid="ref14">17</xref>
        ].
      </p>
      <p>Using probabilistic models to describe the space of possible pipelines is another interesting
feature of AutoML systems. AutoML systems based on Bayesian optimization build an internal
representation of the space of possible algorithm pipelines, which can be interpreted as assigning
a probability distribution to each particular pipeline. This feature describes the algorithm
pipeline space and allows researchers to gather additional information by analyzing which
regions have higher or lower probabilities of generating efective Pipeline components and
allow researchers to gather additional information.</p>
      <p>
        Previous research addressed the limitations of AutoML systems. Estevez-Velarde et al. [
        <xref ref-type="bibr" rid="ref18">21</xref>
        ]
introduced the concept of Heterogeneous AutoML, a more general formulation of the AutoML
Problem. Additionally, they introduced AutoGOAL, a flexible and eficient system for
heterogeneous AutoML implemented in Python. With AutoGOAL, users can describe a specific
machine problem, input and output requirements, and a set of objectives. The system then
automatically finds the best pipeline of algorithms from various libraries, including Scikit-learn
[15], NLTK [
        <xref ref-type="bibr" rid="ref19">22</xref>
        ], Keras [
        <xref ref-type="bibr" rid="ref14">17</xref>
        ], and Gensim [
        <xref ref-type="bibr" rid="ref20">23</xref>
        ]. It is also customizable, allowing users to add and
integrate new algorithms into the existing pipelines. AutoGOAL uses a Pareto Front approach
to multiobjective optimization and an optimization process based on probabilistic grammatical
evolution for context-free grammar [
        <xref ref-type="bibr" rid="ref21">24</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Research</title>
      <p>AutoGOAL has achieved state-of-the-art performance against other AutoML systems and has
been able to solve machine-learning tasks outside of supervised learning. Additionally, it can
build complex pipelines targeting dificult NLP tasks like Named Entity Recognition, being
able to connect algorithms of diferent natures. This research project will use AutoGOAL as a
baseline for its capabilities regarding Heterogeneous AutoML.</p>
      <sec id="sec-3-1">
        <title>3.1. Heterogeneous AutoML</title>
        <p>
          According to Estevez-Velarde et al. [
          <xref ref-type="bibr" rid="ref18">21</xref>
          ], the Heterogeneous AutoML problem’s space of all
possible pipelines can be represented as a  graph. This graph consists of nodes representing
each known algorithm , and edges exist between all pairs ,  such that their corresponding
output and input types are compatible. Given a machine-learning task defined as a function
that transforms an input type into an output type, we can build a specific search space graph
′ that only models valid pipelines. To extract ′, we must introduce two additional nodes to
: Input and Output. These nodes are then connected to all algorithms capable of producing
the desired output from the specific input. By identifying any path in ′ that connects the
Input and Output nodes, we can obtain a pipeline that addresses the machine-learning problem
at hand.
        </p>
        <p>A suitable computational implementation of this process requires solving the following
problems:
1. Defining each algorithm and their respective input and output, such that it is
computationally feasible to determine if two algorithms can be connected and construct the
graph.
2. Designing an optimization strategy that can efectively search in the space of all pipelines,
algorithms, and their hyperparameters, given restricted computational resources.</p>
        <p>AutoGOAL utilizes a set of Semantic Type objects to implement this compatibility function.
Each semantic data type is a Python class that belongs to a hierarchy in which object inheritance
directly represents the relation for type compatibility (e.g., Word can also be interpreted as
Text, a more general type). The data types have a semantic interpretation beyond their
underlying computational structure. For example, a string in computational terms can either
be a Document, a Sentence, or a Word. Each algorithm is implemented as a class with a
run(input: Tin) -&gt; Tout method that performs the corresponding processing, potentially
wrapping an underlying implementation from a machine learning library. Each algorithm’s
input and output types are specified using Semantic Types and represented by the Tin and
Tout annotations.</p>
        <p>While this method for computing compatibility has advantages, it is rigid and dificult to
maintain. Due to the closed nature of the type system, precise type definitions must be matched
for any new algorithms identified and added to the AutoGOAL search space. Also, because
adding new Semantic Types does not automatically update current algorithm annotations, users
need to check on every existing algorithm to identify which should be annotated accordingly.
Moreover, this mechanism assumes a tree-like structure of type compatibility when there might
be more complex relationships (e.g., a Stem might also be considered a Word, albeit these two
types do not inherit from each other). This leads to an interesting question: ”Can we model
algorithm compatibility more openly?”.</p>
        <p>
          Recent proposals suggest using natural language to store information describing algorithms.
Shen et al. [14] uses Jsons, mainly text-based, to store information about pre-trained models.
They parse natural language prompts into multiple tasks that are matched with suitable
algorithms using an LLM. However, this tool does not address the AutoML problem, as it does not
optimize model selection or hyperparameter configurations for algorithms. In contrast, Zhang
et al. [
          <xref ref-type="bibr" rid="ref22">25</xref>
          ] aims to develop an AutoML system called AutoML-GPT, which uses LLMs to train
models on datasets with user inputs and descriptions automatically. The LLMs serve as an
automatic training system to establish connections with models and process inputs.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Research Questions</title>
        <p>In this research, we propose the integration of description cards based on text data for algorithms
in AutoGOAL. By leveraging the power of LLMs, we can match algorithms based on their
description. This method adds a vital factor of generalization that might solve the previous
limitations of AutoGOAL. Moreover, by substituting the current Semantic Type system with
natural language, we open the tool’s interface to accept user text prompts.</p>
        <p>To achieve our main objective, we must address various questions within the proposal:
1. Which LLM should we use?
2. What language should we support in the interface?
3. What machine learning tasks should we target?</p>
        <p>To answer the first question, we must conduct experiments to determine which LLM will be
most suitable for our project. One idea is to use two diferent LLMs that are each fine-tuned
for specific purposes. For example, one LLM can help determine the compatibility between
algorithms used during the optimization process. At the same time, the other can identify
problem definitions (input and output types and objective functions) out of natural language
for better user interaction with the system.</p>
        <p>For the second question, we aim to develop an interface not tailored to a specific language to
make it fully inclusive. However, the performance of LLMs can vary significantly in diferent
languages due to diferences in available training data. Therefore, we will compare the
performance of multilingual models against specific-language models in our objective tasks before
deciding which approach to follow.</p>
        <p>Finally, our main objective is to extend existing tools, specifically AutoGOAL, to saturate
the definition of Heterogeneous AutoML, thus achieving more flexibility and integrating more
tasks seamlessly. We can achieve this by using the compatibility function to discover algorithms
that were once bound to a specific machine-learning problem but can be part of the solution of
another one. The diversity and amount of algorithms determine the limit of tasks we can solve
in the system.</p>
        <p>The proposed system has valuable scientific, economic, and social implications. It can enhance
our understanding of artificial intelligence and apply it to robotics and process automation.
Economically, it can speed up the development of applications and decrease the cost and time
required for building machine learning solutions.</p>
        <p>From a social standpoint, an AutoML system based on natural language can improve the
accessibility and ease of use of machine learning, especially in critical areas such as healthcare
and education. Furthermore, automating the building of learning models can lessen the need
for human intervention in repetitive and monotonous tasks, thus reducing the carbon footprint
associated with computer system operations.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Proposed Experimentation</title>
      <p>To evaluate the potential of our proposed system, we plan to develop a benchmark that
incorporates challenging tasks from multiple domains such as vision, NLP, audio, and others. We will
perform ablation studies to comprehend the significance of diferent LLMs and optimization
strategies in the overall performance of our system. In addition, to make the evaluation more
comprehensive, we will compare our new system with its previous version and other
stateof-the-art AutoML systems. By providing more flexibility, we aim to test the capabilities of
our system against human adversaries in various challenges. Furthermore, we will explore the
possibility of integrating human feedback into the learning process, which can provide valuable
insights and lead to further improvement.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and Future work</title>
      <p>The purpose of this publication is to present the research framework for a thesis that aims to
investigate the intersection between AutoML and Large Language Models (LLMs). Our objective
is to improve AutoML systems, making them more accessible, user-friendly, and versatile. In
order to achieve this goal, we will begin by examining the current state of the art in this field.
Subsequently, we will develop corresponding description cards for each algorithm available
in AutoGOAL and also include new algorithms, such as pre-trained models from Hugging
Face along with their respective cards. The next step will be integrating an LLM to model the
compatibility function between algorithms, thereby enabling a natural language interface for
user interaction. Ultimately, our aim is to pave the way for more inclusive and eficient machine
learning applications in various domains.
[13] Baolin Peng et al. “Check your facts and try again: Improving large language models with
external knowledge and automated feedback”. In: arXiv preprint arXiv:2302.12813 (2023).
[14] Yongliang Shen et al. “HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in</p>
      <p>HuggingFace”. In: arXiv preprint arXiv:2303.17580 (2023).
[15] Fabian Pedregosa et al. “Scikit-learn: Machine Learning in Python”. In: Journal of Machine</p>
      <p>Learning Research 12.85 (2011), pp. 2825–2830.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Frank</given-names>
            <surname>Hutter</surname>
          </string-name>
          , Lars Kotthof, and
          <string-name>
            <given-names>Joaquin</given-names>
            <surname>Vanschoren</surname>
          </string-name>
          .
          <source>Automated Machine Learning</source>
          . Springer,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Haifeng</given-names>
            <surname>Jin</surname>
          </string-name>
          , Qingquan Song, and Xia Hu. “
          <string-name>
            <surname>Auto-Keras</surname>
          </string-name>
          :
          <article-title>An Eficient Neural Architecture Search System”</article-title>
          .
          <source>In: ACM</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1946</fpage>
          -
          <lpage>1956</lpage>
          . doi:
          <volume>10</volume>
          .1145/3292500.3330648.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Matthias</given-names>
            <surname>Feurer</surname>
          </string-name>
          et al. “
          <article-title>Auto-Sklearn 2.0: The Next Generation”</article-title>
          .
          <source>In: arXiv: Learning</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Chris</given-names>
            <surname>Thornton</surname>
          </string-name>
          et al. “
          <article-title>Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms”</article-title>
          . In: ACM,
          <year>2013</year>
          , pp.
          <fpage>847</fpage>
          -
          <lpage>855</lpage>
          . doi:
          <volume>10</volume>
          .1145/2487575.2487629.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Suilan</given-names>
            <surname>Estevez-Velarde</surname>
          </string-name>
          et al. “
          <article-title>Automl strategy based on grammatical evolution: A case study about knowledge discovery from text”</article-title>
          .
          <source>In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          .
          <year>2019</year>
          , pp.
          <fpage>4356</fpage>
          -
          <lpage>4365</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Randal</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Olson</surname>
          </string-name>
          and
          <string-name>
            <surname>Jason H. Moore</surname>
          </string-name>
          . “
          <article-title>TPOT: A Tree-Based Pipeline Optimization Tool for Automating Machine Learning”</article-title>
          . In: vol.
          <source>2019</source>
          . Springer, Cham,
          <year>2018</year>
          , pp.
          <fpage>66</fpage>
          -
          <lpage>74</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -05318-
          <issue>5</issue>
          _
          <fpage>8</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Brent</given-names>
            <surname>Komer</surname>
          </string-name>
          , James Bergstra, and Chris Eliasmith. “
          <string-name>
            <surname>Hyperopt-Sklearn</surname>
          </string-name>
          :
          <article-title>Automatic Hyperparameter Configuration for Scikit-Learn”</article-title>
          . In: (
          <year>2013</year>
          ), pp.
          <fpage>32</fpage>
          -
          <lpage>37</lpage>
          . doi:
          <volume>10</volume>
          .25080/MAJORA14BD3278-006.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Felix</given-names>
            <surname>Mohr</surname>
          </string-name>
          , Marcel Dominik Wever, and Eyke Hüllermeier.
          <article-title>“ML-Plan: Automated machine learning via hierarchical planning”</article-title>
          .
          <source>In: Machine Learning 107.8</source>
          (
          <issue>2018</issue>
          ), pp.
          <fpage>1495</fpage>
          -
          <lpage>1515</lpage>
          . doi:
          <volume>10</volume>
          .1007/S10994-018-5735-Z.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Alex</given-names>
            <surname>Guimarães Cardoso de Sá</surname>
          </string-name>
          et al. “
          <article-title>RECIPE: A Grammar-Based Framework for Automatically Evolving Classification Pipelines”</article-title>
          . In: Springer, Cham,
          <year>2017</year>
          , pp.
          <fpage>246</fpage>
          -
          <lpage>261</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -55696-3_
          <fpage>16</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Bonan</given-names>
            <surname>Min</surname>
          </string-name>
          et al.
          <source>Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey</source>
          .
          <year>2021</year>
          . arXiv:
          <volume>2111</volume>
          .01243 [cs.CL].
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Frank</surname>
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Xu</surname>
          </string-name>
          et al. “
          <article-title>A Systematic Evaluation of Large Language Models of Code”</article-title>
          .
          <source>In: (Feb</source>
          .
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .48550/ARXIV.2202.13169. arXiv:
          <volume>2202</volume>
          .13169 [cs.PL].
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Abubakar</surname>
            <given-names>Abid</given-names>
          </string-name>
          , Maheen Farooqi, and
          <string-name>
            <given-names>James</given-names>
            <surname>Zou</surname>
          </string-name>
          .
          <source>Persistent Anti-Muslim Bias in Large Language Models</source>
          .
          <year>2021</year>
          . doi:
          <volume>10</volume>
          .1145/3461702.3462624.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Geofrey</given-names>
            <surname>Holmes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Donkin</surname>
          </string-name>
          , and
          <string-name>
            <surname>Ian</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Witten</surname>
          </string-name>
          . “
          <article-title>WEKA: a machine learning workbench”</article-title>
          .
          <source>In: IEEE</source>
          ,
          <year>1994</year>
          , pp.
          <fpage>357</fpage>
          -
          <lpage>361</lpage>
          . doi:
          <volume>10</volume>
          .1109/ANZIIS.
          <year>1994</year>
          .
          <volume>396988</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>François</given-names>
            <surname>Chollet</surname>
          </string-name>
          .
          <article-title>Keras: The Python Deep Learning library</article-title>
          .
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>James</given-names>
            <surname>Gosling</surname>
          </string-name>
          et al.
          <article-title>The Java language specification</article-title>
          .
          <source>Addison-Wesley Professional</source>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [19] [20]
          <string-name>
            <given-names>Martín</given-names>
            <surname>Abadi</surname>
          </string-name>
          et al. “
          <article-title>TensorFlow: a system for large-scale machine learning”</article-title>
          .
          <source>In: USENIX Association</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>283</lpage>
          . doi:
          <volume>10</volume>
          .5555/3026877.3026899.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Adam</given-names>
            <surname>Paszke</surname>
          </string-name>
          et al. “
          <article-title>PyTorch: An Imperative Style, High-Performance Deep Learning Library”</article-title>
          . In: vol.
          <volume>32</volume>
          . hgpu.org,
          <year>2018</year>
          , pp.
          <fpage>8026</fpage>
          -
          <lpage>8037</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Suilan</given-names>
            <surname>Estevez-Velarde</surname>
          </string-name>
          et al. “
          <article-title>Automatic discovery of heterogeneous machine learning pipelines: An application to natural language processing”</article-title>
          .
          <source>In: Proceedings of the 28th International Conference on Computational Linguistics</source>
          .
          <year>2020</year>
          , pp.
          <fpage>3558</fpage>
          -
          <lpage>3568</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Edward</given-names>
            <surname>Loper</surname>
          </string-name>
          and
          <string-name>
            <given-names>Steven</given-names>
            <surname>Bird</surname>
          </string-name>
          . “
          <article-title>NLTK: the natural language toolkit”</article-title>
          .
          <source>In: arXiv preprint cs/0205028</source>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>Petr</given-names>
            <surname>Sojka</surname>
          </string-name>
          . “
          <article-title>Gensim-statistical semantics in python”</article-title>
          . In:
          <article-title>Retrieved from genism</article-title>
          .
          <source>org</source>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Hyun-Tae Kim</surname>
          </string-name>
          and
          <article-title>Chang Wook Ahn. “A new grammatical evolution based on probabilistic context-free grammar”</article-title>
          .
          <source>In: Proceedings of the 18th Asia Pacific Symposium on Intelligent and Evolutionary Systems-Volume 2</source>
          . Springer.
          <year>2015</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Shujian</surname>
          </string-name>
          Zhang et al. “
          <article-title>AutoML-GPT: Automatic Machine Learning with GPT”</article-title>
          .
          <source>In: arXiv preprint arXiv:2305.02499</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>