<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Not only ChatGPT: educational approaches for Embodied Artificial Intelligence</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Monica Casella</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Clara Nobile</string-name>
          <email>clara.nobile@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Federico Diano</string-name>
          <email>federico.diano@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Raffaella Esposito</string-name>
          <email>raffaella.esposito3@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Luongo</string-name>
          <email>maria.luongo@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alessio Manfredini</string-name>
          <email>alessiomanfredini17@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicola Milano</string-name>
          <email>nicola.milano@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Onofrio Gigliotta</string-name>
          <email>onofrio.gigliotta@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Davide Marocco</string-name>
          <email>davide.marocco@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michela Ponticorvo</string-name>
          <email>michela.ponticorvo@unina.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Natural and Artificial Cognition Lab, Department of Humanities, University of Naples Federico II</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The increasing interest in Artificial Intelligence has been translated in a growing number of educational proposals on machine learning and large language models. These proposals have addressed different targets, ranging from students to professionals, and adopted different approaches that cover both the theoretical foundations and practical applications of the technology. In the field of AI, robotics applications are gaining more and more appeal and a need for education in this specific field is emerging. In the present contribution, we outline some challenges in teaching Embodied Artificial Intelligence and report three experiences of its teaching to children and kids, post-graduate students and professionals.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Embodied Artificial Intelligence</kwd>
        <kwd>Education for AI</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>By combining these approaches, educators can offer a well-rounded AI curriculum that is both
theoretically sound and practically applicable. In the next section we will introduce some educational
pathways on AI.</p>
    </sec>
    <sec id="sec-2">
      <title>2. LLMs educational pathway</title>
      <p>In current educational proposal on LLMs, represented graphically in fig.1, the starting point is
often the foundational knowledge in mathematics and programming.</p>
      <p>
        AI relies heavily on linear algebra, calculus, probability, and statistics. Building a strong
mathematical foundation helps students understand algorithms, optimization techniques, and data
analysis [
        <xref ref-type="bibr" rid="ref8">10</xref>
        ]. At the same time, it is important to provide students with programming tools, for
example in languages like Python, R, and frameworks like TensorFlow and PyTorch. Teaching
programming skills is essential for implementing AI models.
      </p>
      <p>The next step is to provide tools to understanding Machine Learning (ML) algorithms, particularly
Supervised Learning (algorithms like linear regression, decision trees, and support vector machines,
which rely on labeled data to make predictions) and Unsupervised Learning (techniques like
clustering and dimensionality reduction enable to find patterns in unlabeled data).</p>
      <p>
        These steps prepare the exploration of Large Language Models (LLMs) which represent a highly
advanced frontiers of AI applications. In this case, it is important to introduce Natural Language
Processing (NLP), such as tokenization, sentiment analysis, and text generation, which are key in
training and fine-tuning LLMs like GPT and BERT [
        <xref ref-type="bibr" rid="ref9">11</xref>
        ].
      </p>
      <p>Along with this algorithmic approach, there are some fields of AI where the focus is put on the
modelling meaning of AI, on how it represents a simulation of natural behaviors, natural nervous
systems and adaptation: we will focus on these pathways in next section.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Embodied Artificial Intelligence educational pathway</title>
      <p>
        Embodied Artificial Intelligence (E-AI) refers to a branch of AI that focuses on the development of
intelligent systems or agents that interact with the physical world through a body, such as robots or
virtual agents [
        <xref ref-type="bibr" rid="ref10 ref11">12, 13</xref>
        ]. Unlike traditional AI, which often exists solely in software environments and
deals primarily with abstract data, embodied AI is concerned with integrating perception, movement,
and decision-making to navigate and perform tasks in real-world environments [
        <xref ref-type="bibr" rid="ref12 ref13">14, 15</xref>
        ].
      </p>
      <p>
        The concept of embodiment implies that intelligence arises not just from cognitive processing but
from the interaction between the body and the environment. This idea draws from cognitive science
theories that suggest intelligent behavior is shaped by the way an organism’s body perceives and
responds to its surroundings [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">16, 17, 18</xref>
        ].
      </p>
      <p>
        In the educational pathways on E-AI, the first step is to introduce perception and sensorimotor
integration; indeed, Embodied AI systems use sensors (e.g., cameras, microphones, touch sensors) to
perceive their environment and actuators (e.g., motors, robotic arms) to interact with it [
        <xref ref-type="bibr" rid="ref12">14</xref>
        ]. The
integration of sensory input with motor control is crucial for these systems to learn and adapt to
dynamic environments. The next point consists in determining the control system, that, following
the metaphor of natural agents, mainly consists of artificial neural networks, with artificial nodes
linked by weighted connections and specific activation functions, organized in layers and with
possible recursive connections [
        <xref ref-type="bibr" rid="ref11">13</xref>
        ].
      </p>
      <p>
        The third step is devoted to the learning via interaction. Embodied AI often employs learning
techniques, such as reinforcement learning, where the agent improves its performance by interacting
with the environment, receiving feedback, and refining its actions or evolutionary algorithms where
the metaphor is once again derived from natural evolution [
        <xref ref-type="bibr" rid="ref13">15</xref>
        ].
      </p>
      <p>
        These steps lead to robotics, which represents a core element of E-AI and is widely used in the
context of educational robotics. Autonomous robots and humanoids rely on Embodied AI to perform
tasks such as moving through complex environments, manipulating objects, and interacting with
humans [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">12, 13, 14</xref>
        ].
      </p>
      <p>Embodied AI aims to create autonomous and adaptive systems that can operate in the physical
world much like humans and animals do. This field is particularly important for applications where
real-world interaction is critical. By grounding AI in physical experiences, Embodied AI allows to
bring us closer to creating machines that not only think in abstract but also act and learn in ways that
are more natural and intuitive in human-centered environments.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Comparison between LLMs and E-AI teaching</title>
      <p>
        Sketching the differences in these two pathways, we must start from the domain. LLMs like GPT,
BERT, and others are designed to process and generate human language. The teaching of LLMs
centers around Natural Language Processing (NLP), which includes tasks like text generation,
translation, summarization, sentiment analysis, and more [
        <xref ref-type="bibr" rid="ref17">19</xref>
        ]. On the other hand, E-AI focuses on
systems that interact with the physical world through bodies, such as robots. Teaching E-AI is about
sensorimotor integration, physical interaction, and how agents learn from their environment. It deals
with perception, navigation, manipulation, and physical tasks in the real or simulated world [
        <xref ref-type="bibr" rid="ref11">13</xref>
        ].
      </p>
      <p>
        Teaching LLMs emphasizes text-based data processing, including tokenization and embeddings
(transforming words into vectors), language model architectures (e.g., transformers), pre-training and
fine-tuning on large text corpora, attention mechanisms, sequence modeling, and text generation [
        <xref ref-type="bibr" rid="ref9">11</xref>
        ].
      </p>
      <p>
        Learning Environments are also different, as LLMs typically learn in virtual environments with
large text datasets. The focus is on how models process and predict language patterns using labeled
and unlabeled data, whereas E-AI happens in real-world environments or simulations, where the
agent physically interacts with objects, learns through sensory feedback, and adapts to its
surroundings. This could include robotic experiments or simulations where agents perform tasks like
walking, grasping, or navigating [
        <xref ref-type="bibr" rid="ref12">14</xref>
        ].
      </p>
      <p>
        Moreover, LLMs are trained primarily using supervised learning and self-supervised learning
techniques [
        <xref ref-type="bibr" rid="ref18">20</xref>
        ]. The models are trained on massive datasets of text to predict the next word or
sentence, with no direct physical interaction. Feedback is often provided through loss functions (e.g.,
cross-entropy) that measure the difference between predicted and actual outputs and E-AI systems
typically learn through Reinforcement Learning (RL) or Evolutionary Algorithms, where feedback
comes from interactions with the environment [
        <xref ref-type="bibr" rid="ref11 ref12">13, 14</xref>
        ]. The agent receives rewards or penalties based
on how well it performs a task (e.g., reaching a goal or manipulating an object), learning through trial
and error.
      </p>
      <p>In summary, LLMs focus on language, dealing with text-based data and algorithms to understand,
generate, or manipulate human language. Teaching LLMs revolves around NLP tasks, large-scale
data, and model architectures like transformers.</p>
      <p>
        Embodied AI deals with intelligent agents interacting with the physical world. Teaching Embodied
AI involves robotics, sensorimotor systems, and algorithms like Reinforcement Learning, focusing on
how systems learn to navigate, manipulate, and perform tasks in dynamic environments [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">12, 13, 14</xref>
        ].
      </p>
      <p>These distinct approaches reflect the different challenges and goals in creating intelligent systems
that either understand and generate language or physically interact with their environment.</p>
      <p>Whereas there are nowadays many proposals of educational pathways on Artificial Intelligence,
there is still less attempts to build educational pathways on E-AI. We propose here some examples of
how E-AI can be introduced to different target groups.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Some examples of E-AI educational pathways</title>
      <p>Embodied Artificial Intelligence (E-AI) represents a transformative approach in AI, emphasizing
the integration of physical bodies or agents with cognitive capabilities. By embedding AI systems
within physical forms, E-AI aims to achieve more robust interactions with real-world environments,
supporting advanced learning, adaptation, and responsiveness. Understanding the diverse pathways
through which E-AI is realized is crucial for advancing applications in robotics, autonomous systems,
and human-robot interaction. In the following subsections, we introduce some key examples of these
pathways, each illustrating distinctive strategies for introducing and teaching the fundaments of
EAI, including artificial neural networks, learning algorithms, and evolutionary strategies.</p>
    </sec>
    <sec id="sec-6">
      <title>5.1.Introducing Embodied Artificial Intelligence to post-graduate students of humanities</title>
      <p>In the last years, in the field of AI, we have witnessed the overcoming and solution of many
technical challenges, including the possibility to rely on stronger and stronger computational
resources, but, at the same time, many challenges raised concerning the human factor in AI [21, 22].</p>
      <p>For this reason, there is a growing tendence to widen the interdisciplinary AI teams to include
specialists in human disciplines, thus recovering the foundational approach [23]. Teaching AI to
people in humanistic fields enables them to bring a human-centered perspective to AI development
and use, especially in ethical awareness [24], and informed critique and policy influence [25].</p>
      <p>Moreover, this trend allows to improve communication between disciplines: AI is increasingly
interdisciplinary, requiring collaboration between technical and non-technical experts. Humanists
who understand AI can bridge communication gaps between technical developers and broader
society, translating complex technological concepts into language that non-experts can understand
and ensuring public discourse around AI is accessible [26]. On the other hand, humanists contribute
to creativity and innovative thinking that can lead to new, imaginative applications of AI. When AI
is taught in the humanistic context, it opens possibilities for its use in fields beyond traditional tech
spaces.</p>
      <p>At the University of Naples “Federico II”, under the supervision of Laboratory for Natural and
Artificial Cognition “Orazio Miglino”, there have been two editions of an innovative post-graduation
course devoted to Artificial Intelligence.</p>
      <p>The key distinguishing features of these educational pathways lie in their ability to merge the
humanistic backgrounds of participants with a strong and comprehensive technical foundation in
Artificial Intelligence (AI) tools and methodologies. This interdisciplinary approach ensures that
learners not only understand the theoretical and ethical dimensions of AI but also gain the technical
proficiency needed to apply AI in real-world contexts. Programming languages, such as Python, serve
as the critical instruments through which participants can create complex simulations, develop
sophisticated models, and build artificial environments, with the goal of modeling and replicating
aspects of reality in a controlled, computational setting [27].</p>
      <p>
        Furthermore, delving deeply into the study of AI models and their practical applications, as well
as exploring various systems and methods in Machine Learning (also referred to as artificial learning),
computational linguistics, and advanced data analysis, lays the groundwork for more specialized
fields such as Evolutionary Robotics and neurorobotics [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">12, 13, 14</xref>
        ]. These emerging areas of research
and application are particularly significant in the realm of machine autonomy, where robots and
artificial agents are endowed with the ability to adapt, evolve, and make independent decisions based
on their interactions with the environment.
      </p>
      <p>Integrating Embodied AI education into academic and professional training is essential for
addressing emerging challenges at the intersection of cognitive science, robotics, and human-machine
interaction. As artificial intelligence systems increasingly operate in dynamic, real-world
environments, the need for embodied intelligence—wherein agents perceive, reason, and interact
through physical or simulated bodies—has grown substantially [28]. Unlike traditional AI, which
often relies on static data or limited environmental feedback, Embodied AI leverages sensory-motor
capabilities and real-time environmental responses to foster adaptive and context-sensitive
interactions [29, 30]. Teaching Embodied AI in educational settings has therefore become essential
for equipping the next generation of AI researchers and practitioners with the knowledge and skills
necessary to design, evaluate, and ethically implement these systems.</p>
      <p>Moreover, incorporating Embodied AI into advanced curricula enriches students' understanding
of not only computational algorithms but also the biological, psychological, and sociocultural
dimensions of intelligence. Students trained in Embodied AI can better understand concepts such as
perception, proprioception, and motor skills that are foundational to both human cognition and
robotic functionality. This interdisciplinary approach encourages innovative research, bridging fields
that traditionally operate in silos, thus fostering a comprehensive view of intelligence in natural and
artificial systems.</p>
      <p>The integration of Embodied AI into educational frameworks, also exploiting on-line collaboration
(for an example, see https://www.moowcode.eu/) represents a promising evolution in AI education,
preparing post-graduate students to address complex challenges in interdisciplinary and
humancentered ways. As AI systems increasingly interact with the physical world and affect human lives,
the importance of Embodied AI education becomes evident.</p>
    </sec>
    <sec id="sec-7">
      <title>5.2.ADA (All Digital Academy): an educational pathway for adult training on AI</title>
      <p>The All Digital Academy (ADA) project serves as a comprehensive educational platform aimed at
upskilling adult educators in key areas of Artificial Intelligence (AI) and digital technologies. By
aligning with the European Commission’s Digital Education Action Plan (DEAP) [31], ADA targets
the growing demand for digital literacy in a society where only 56% of individuals aged 16-74
currently possess basic digital skills [32, 33]. The project recognizes the critical role educators play in
shaping a digitally competent citizenry and addresses this through structured learning experiences
designed to foster understanding, application, and ethical consideration of AI.</p>
      <p>The ADA initiative utilizes a blended learning approach, incorporating various resources such as
MOOCs, webinars, and an extensive repository of educational materials. Designed around the ADDIE
instructional model (Analysis, Design, Development, Implementation, and Evaluation), ADA ensures
that its courses are both adaptable and responsive to the evolving needs of educators [34, 35].
Moreover, ADA's curriculum is built upon the Digital Competence Framework for Citizens (DigComp
2.1) [36], which serves as the foundation for developing AI-related skills across different facets,
including Information and Data Literacy, Communication and Collaboration, Digital Content
Creation, Security, and Problem Solving.</p>
      <p>The ADA curriculum emphasizes a holistic and hands-on approach to AI education. This approach
encompasses foundational concepts, technical know-how, practical experimentation, and ethical
considerations—elements crucial to the responsible integration of AI into everyday practice. In the
introduction, the importance of understanding AI’s history, mechanics, applications, and ethical
implications was highlighted. ADA addresses these aspects through its structured educational
pathways, providing educators with a deep dive into AI topics and competences.</p>
      <p>For example, the ADA curriculum stresses the importance of understanding what AI systems can
and cannot do. This includes recognizing that AI systems, despite their autonomous appearance, are
fundamentally products of human intelligence and decision-making. Educators are guided to explore
how AI systems process data, make predictions, and influence both real and virtual environments.
Additionally, ADA underscores the relevance of understanding the mechanics behind AI systems,
such as the use of statistics and algorithms, as well as the principles of machine learning, to provide
educators with a solid technical grounding in AI concepts.</p>
      <p>Furthermore, ADA places a strong emphasis on the application of AI across various domains,
supporting educators in identifying where AI can bring benefits to everyday life. By linking
theoretical knowledge with practical experimentation, ADA equips educators to implement AI-based
tools in their teaching practices. The program encourages hands-on experiences with model building
and training, helping educators gain confidence in applying AI to solve real-world problems, thus
mirroring the project-based and collaborative learning methods mentioned in the introduction.</p>
      <p>To further support its mission, ADA fosters a Community of Practice (CoP) [37, 38] where
educators can engage, share knowledge, and collaborate on AI-related topics. This community-driven
approach aligns with the introduction's mention of collaborative learning and hackathons, providing
an avenue for continuous learning and professional development. Through this digital environment,
educators are encouraged to exchange best practices, troubleshoot challenges, and stay updated with
advancements in the field of AI, enhancing both their own skills and those of their learners.</p>
      <p>The ADA project, through its focus on a comprehensive AI curriculum, hands-on experimentation,
and an ethical framework, aligns with the key elements introduced in AI education. It not only equips
educators with the necessary technical skills but also fosters an understanding of AI's broader
implications.</p>
    </sec>
    <sec id="sec-8">
      <title>5.3.TEACH E-AI 2C (Teaching Embodied Artificial Intelligence to</title>
    </sec>
    <sec id="sec-9">
      <title>Children): an educational proposal for young learners on E-AI</title>
      <p>The Teach E-AI 2C (Teaching Embodied Artificial Intelligence to Children) project [39] was born
with the objective of filling the gap between AI growing progress and the users’ knowledge of even
basic principles [40]. It aims to create tailored educational resources that enable children, aged 9-13,
to learn about AI by using AI itself. It is in an ongoing phase of development and assessment. Future
evaluations are essential to ensure its applicability and educational potential.</p>
      <p>The project follows an Embodied AI approach [41] and is grounded in the instructional design
principles of the 4C/ ID model [42]. This educational model is designed to cope with complex learning
tasks. Also, it emphasizes the importance of practice in realistic context where learners can integrate
both knowledge and skills [42].</p>
      <p>These features, together with E-AI approach, will make it possible to: (i) moving beyond a strictly
algorithm-driven approach; (ii) understanding biological systems to replicate their functions in
artificial systems; (iii) clarifying principles for intelligent behavior; (iv) enforcing these principles to
artificial systems that interact with the physical world; (v) encouraging hands-on learning methods.</p>
      <p>To achieve the educational aims, the project comprises three Learning Units (LUs), which
introduce E-AI concepts to children and early adolescents. Consistent with the 4C/ ID model,
Supportive Information for these LUs provides permanent source material on core AI topics, including
Artificial Neural Networks, Genetic Algorithms, E-AI, Robotics, and Evolutionary Robotics, alongside
foundational concepts such as Learning processes, Evolution, and Darwin's Theory. The LUs consist
mainly of text written in a simplified form, to make complex AI topics accessible to young learners.
Hyperlinks direct learners to additional content from sources like the Treccani Vocabulary and
Encyclopedia, allowing for deeper exploration of fundamental concepts. Visual aids, links to
educational videos, games, activities, and audio-pills are incorporated to enhance engagement and
meet diverse learning preferences.</p>
      <p>For Learning Tasks, the project incorporates case studies, mini-projects, and problem-solving
activities based on real-world scenarios. These tasks lead learners to play with unplugged practice
and creativity, allowing them to master individual concepts progressively, and improve knowledge
retention. In addition, children are supported with Procedural Information, in the form of ‘how to’ or
‘step-by-step’ instructions on using E-AI.</p>
      <p>Finally, the Part-task Practice has been filled with the development of the Teach E-AI 2C robotic
farm, in order to provide hands-on activity on E-AI. It is an integrated hardware-sofware platform
developed for Thymio [43] and AlphAI [44] robots. This software simulates artificial evolution,
enabling users to breed and evolve a population of robots, even without technical expertise [45].
Starting with a given first generation, users can observe robot performance and select some of them
to produce subsequent offspring. The selection for reproduction can either be made by the system,
selecting the "best robots", or made by the users. Human users, however, simply choose the robots
they believe performed best.</p>
      <p>Once the selection process is complete, the system creates clones of the selected robots,
introducing random mutations into their control systems. This selection/cloning/mutation cycle can
be repeated until the "breeder" finds a particularly capable robot. At this point, the brain of the
simulated robot can be uploaded to a real robot, allowing the user to observe its performance in the
real world. The system also tracks population fitness over generations and offers various
customization options through its interface, making it accessible and user-friendly for young learners
experimenting with artificial evolution. With the Teach E-AI 2C robotic farm, children can conduct
their own artificial evolution experiments firsthand, observe theory in action, play in competitive
groups, and foster collaborative learning.</p>
    </sec>
    <sec id="sec-10">
      <title>6. Acknowledgements</title>
      <p>MOOW-CODE is funded by the Erasmus + Programme of the European Union, KA220-HED
Cooperation partnerships in higher education, KA220-HED-5BA32084 [46].</p>
      <p>ADA is funded by the European Union by European Education and Culture Executive Agency
(EACEA). Project No: 101049118.</p>
      <p>Teach E-AI is funded by Fondazione Compagnia di San Paolo and Cassa Depositi e Prestiti under
the Artificial Intelligence Call AI-LEAP: LEArning Personalization with AI and of AI
(D13C23001280007) project, which includes Teach E-AI 2C sub-project.</p>
    </sec>
    <sec id="sec-11">
      <title>7. References</title>
      <p>[1] M. Van Mechelen, R. C. Smith, M. M. Schaper, M. Tamashiro, K- E. Bilstrup, M. Lunding, M.</p>
      <p>Graves Petersen, and O. Sejer Iversen, Emerging Technologies in K12 Education: A Future
HCI Research Agenda, in: ACM Trans. Computer-Human Interaction 30.3, 2023.
[2] W. Yang, Artificial Intelligence education for young children: Why, what, and how in
curriculum design and implementation, in: Computers and Education: Artificial Intelligence
3, 2022.
[21] L. Sanneman &amp; J. A. Shah, The situation awareness framework for explainable AI (SAFE-AI)
and human factors considerations for XAI systems, International Journal of Human–
Computer Interaction, 2022, 38(18-20), 1772-1788.
[22] U. Ehsan, P. Wintersberger, Q. V. Liao, E. A. Watkins, C. Manger, H. Daumé III, ... &amp; M. O.</p>
      <p>Riedl, Human-Centered Explainable AI (HCXAI): beyond opening the black-box of AI, in: CHI
conference on human factors in computing systems extended abstracts, 2022, pp. 1-7.
[23] J. Moor, The Dartmouth College artificial intelligence conference: The next fifty years. Ai</p>
      <p>Magazine, 2006, 27(4), 87-87.
[24] S. C. Kong, W. M. Y. Cheung, &amp; O. Tsang, Evaluating an artificial intelligence literacy
programme for empowering and developing concepts, literacy and ethical awareness in
senior secondary students. Education and Information Technologies, 2023, 28(4), 4703-4724.
[25] J. C. Heilinger, The ethics of AI ethics. A constructive critique. Philosophy &amp; Technology,
2022, 35(3), 61.
[26] L. Agustí-Cullel &amp; M. Schorlemmer, A humanist perspective on artificial intelligence, 2021.
[27] D. Parisi, Simulazioni: la realtà rifatta nel computer. Società editrice il Mulino, 2010.
[28] R. Pfeifer &amp; J. Bongard, How the body shapes the way we think: A new view of intelligence.</p>
      <p>MIT Press, 2007.
[29] R. A. Brooks, Intelligence without representation. Artificial Intelligence, 1991, 47(1-3),
139159.
[30] A. Clark, An embodied cognitive science? Trends in Cognitive Sciences, 1999, 3(9), 345-351.
[31] European Commission, “Digital Education Action Plan 2021-2027,” COM(2020) 624 final,
2020. Accessed: May 29, 2024. [Online]. Available:
https://eur-lex.europa.eu/legalcontent/EN/TXT/?uri=CELEX%3A52020DC0624
[32] Organisation for Economic Cooperation and Development, TALIS 2018 Results (Volume I):
Teachers and School Leaders as Lifelong Learners, TALIS, OECD Publishing, Paris, 2019. doi:
10.1787/1d0bc92a-en.
[33] J. Fraillon, J. Ainley, W. Schulz, T. Friedman, and D. Duckworth, Preparing for Life in a Digital</p>
      <p>World, Springer International Publishing, Cham, 2020. doi: 10.1007/978-3-030-38781-5.
[34] R. Mahajan, K. Gupta, P. Gupta, S. Kukreja, and T. Singh, “Multimedia Instructional Design
Principles: Moving from Theoretical Rationale to Practical Applications,” Indian Pediatrics,
vol. 57, no. 6, pp. 555–560, 2020. doi: 10.1007/s13312-020-1854-2.
[35] M. K. Khalil and I. A. Elkhider, “Applying learning theories and instructional design models
for effective instruction,” Advances in Physiology Education, vol. 40, pp. 147–156, 2016. doi:
10.1152/advan.00138.2015.
[36] European Commission, Joint Research Centre, S. Carretero, R. Vuorikari, and Y. Punie,
DigComp 2.1 – The Digital Competence Framework for Citizens with Eight Proficiency
Levels and Examples of Use, Publications Office, 2017. doi: 10.2760/38842.
[37] E. Wenger-Trayner and B. Wenger-Trayner, “An introduction to communities of practice: a
brief overview of the concept and its uses,” Accessed: May 30, 2024. [Online]. Available:
https://www.wenger-trayner.com/introduction-to-communities-of-practice
[38] E. Wenger, Communities of Practice: Learning, Meaning, and Identity, Cambridge University</p>
      <p>Press, Cambridge, 1998. doi: 10.1017/CBO9780511803932.
[39] C. Nobile, D. Marocco, O. Gigliotta, and M. Ponticorvo, Teaching Embodied Artificial
Intelligence to Children (Teach E-AI 2C): an educational proposal for young learners (in
press), in: Proceeding of the 2024 IEEE International Conference on Metrology for eXtended
Reality, Artificial Intelligence and Neural Engineering (MetroXRAINE).
[40] European Commission. "Digital Economy and Society Index (DESI)." European Commission,
https://digital-strategy.ec.europa.eu/it/policies/desi. Accessed: May 30, 2024. [Online].
[41] O. Miglino, and M. Ponticorvo, Exploring the Roots of Spatial Cognition in Artificial and
Natural Organisms: The Evolutionary Robotics Approach, in P. A. Vargas, E. A. Di Paolo, I.</p>
      <p>Harvey, and P. Husbands, The Horizons of Evolutionary Robotics, CogNet, 2014.
[42] J.J. Van Merriënboer and P.A. Kirschner, Ten steps to complex learning: a systematic
approach to four-component instructional design. Oxon: Routledge, 2013.
[43] F. Mondada, M. Bonani, F. Riedo, M. Briod, L. Pereyre, P. Retornaz, S. Magnenat, "Bringing
Robotics to Formal Education: The Thymio Open-Source Hardware Robot," in IEEE Robotics
&amp; Automation Magazine, vol. 24, no. 1, pp. 77-85, March 2017.
[44] Learning Robots discover and teach AI with the AlphAI robot. https://learningrobots.ai/ (Last
access 15 April 2024).
[45] O. Miglino, O. Gigliotta, M. Ponticorvo, and S. Nolfi, Breedbot: An edutainment robotics
system to link digital and real world, in: Knowledge-Based Intelligent Information and
Engineering Systems: 11th International Conference, KES 2007, XVII Italian Workshop on
Neural Networks, Vietri sul Mare, Italy, September 12-14, 2007. Proceedings, Part II 1, pp.
7481, Springer Berlin Heidelberg.
[46] M. Luongo, L. S. Sica, L.S., M. Ponticorvo, An In-Depth Look at MOOW's Innovative Approach
in Higher Education (in press), in: Proceeding of the 2024 IEEE International Conference on
Metrology for eXtended Reality, Artificial Intelligence and Neural Engineering
(MetroXRAINE).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ma</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <article-title>Paving the Way for Novices: How to Teach AI for K-12 Education in China</article-title>
          ,
          <source>in: Proceedings of the 36th AAAI Conference on Artificial Intelligence</source>
          , AAAI Press, Palo Alto, California,
          <year>2022</year>
          : pp.
          <fpage>12852</fpage>
          -
          <lpage>12857</lpage>
          . doi:
          <volume>10</volume>
          .1609/aaai.v36i11.
          <fpage>21565</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>X.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Jin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Arcat: A tangible programming tool for DFS algorithm teaching</article-title>
          ,
          <source>in: Proceedings of the 18th ACM International Conference on Interaction Design and Children</source>
          (IDC),
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY,
          <year>2019</year>
          , pp.
          <fpage>533</fpage>
          -
          <lpage>537</lpage>
          . doi:
          <volume>10</volume>
          .1145/3311927.3325308.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. K.</given-names>
            <surname>Mortensen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <article-title>SmileyCluster: supporting accessible machine learning in K-12 scientific discovery, in: Proceedings of the Interaction design and children conference (IDC), Association for Computing Machinery</article-title>
          , New York, NY,
          <year>2020</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>35</lpage>
          . doi:
          <volume>10</volume>
          .1145/3392063.3394440.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Khalid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Iivari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kinnula</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <source>Familiarizing Children with Artificial Intelligence</source>
          ,
          <source>in: Proceedings of the 25th International Academic Mindtrek Conference, Association for Computing Machinery</source>
          , New York, NY,
          <year>2022</year>
          , pp.
          <fpage>372</fpage>
          -
          <lpage>376</lpage>
          . doi:
          <volume>10</volume>
          .1145/3569219.3569390.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Bojer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Meldgaard</surname>
          </string-name>
          ,
          <article-title>Kaggle forecasting competitions: An overlooked learning opportunity</article-title>
          , in:
          <source>International Journal of Forecasting 37.2</source>
          ,
          <year>2021</year>
          ,
          <fpage>587</fpage>
          -
          <lpage>603</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Sintov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hoffman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lyet</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Tambe</surname>
          </string-name>
          ,
          <article-title>Keeping it real: Using real-world problems to teach AI to diverse audiences</article-title>
          ,
          <source>in: AI Magazine 38.2</source>
          ,
          <year>2017</year>
          ,
          <fpage>35</fpage>
          -
          <lpage>47</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R.</given-names>
            <surname>Gennari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Melonio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Pellegrino</surname>
          </string-name>
          , and
          <string-name>
            <surname>M. D'Angelo</surname>
          </string-name>
          ,
          <article-title>How to Playfully Teach AI to Young Learners: a Systematic Literature Review in: Proceeding of the 15th Biannual Conference of the Italian SIGCHI Chapter (CHItaly</article-title>
          <year>2023</year>
          ),
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          . doi:
          <volume>10</volume>
          .1145/3605390.3605393.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xin-She</surname>
          </string-name>
          ,
          <article-title>Introduction to algorithms for data mining and machine learning</article-title>
          , 1st. ed., Academic Press,
          <year>2019</year>
          . doi:
          <volume>10</volume>
          .1016/C2018-0
          <article-title>-02034-4</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>E.</given-names>
            <surname>Kasneci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Seßler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Küchemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bannert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dementieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fischer</surname>
          </string-name>
          , &amp; G. Kasneci,
          <article-title>ChatGPT for good? On opportunities and challenges of large language models for education</article-title>
          ,
          <source>in: Learning and individual differences</source>
          ,
          <year>2023</year>
          ,
          <volume>103</volume>
          ,
          <fpage>102274</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>O.</given-names>
            <surname>Miglino</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Ponticorvo</surname>
          </string-name>
          ,
          <article-title>Exploring the Roots of Spatial Cognition in Artificial and Natural Organisms: The Evolutionary Robotics Approach</article-title>
          , in: P. A.
          <string-name>
            <surname>Vargas</surname>
            ,
            <given-names>E. A.</given-names>
          </string-name>
          <string-name>
            <surname>Di Paolo</surname>
            ,
            <given-names>I. Harvey</given-names>
          </string-name>
          , and P. Husbands (Eds.),
          <source>The Horizons of Evolutionary Robotics</source>
          , 1st. ed., Mit press, CogNet,
          <year>2014</year>
          , pp.
          <fpage>93</fpage>
          -
          <lpage>123</lpage>
          . doi:
          <volume>10</volume>
          .7551/mitpress/8493.003.0006.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ponticorvo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Di</given-names>
            <surname>Ferdinando</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Marocco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Miglino</surname>
          </string-name>
          ,
          <article-title>Approaches to Embed Bioinspired Computational Algorithms in Educational and Serious Games</article-title>
          ,
          <source>in: Proceedings of the 1st International Workshop on Cognition and Artificial Intelligence for Human-Centred Design</source>
          <year>2017</year>
          ,
          <article-title>CEUR-WS</article-title>
          .org,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ponticorvo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Walker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Miglino</surname>
          </string-name>
          ,
          <article-title>Evolutionary robotics as a tool to investigate spatial cognition in artificial and natural systems</article-title>
          ,
          <source>in: Artificial cognition systems</source>
          ,
          <year>2007</year>
          , pp.
          <fpage>210</fpage>
          -
          <lpage>237</lpage>
          ,
          <string-name>
            <given-names>IGI</given-names>
            <surname>Global</surname>
          </string-name>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>G.</given-names>
            <surname>Paolo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalez-Billandon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kégl</surname>
          </string-name>
          ,
          <article-title>A call for embodied AI</article-title>
          ,
          <source>in: Proceedings of the 41st International Conference on Machine Learning, PMLR</source>
          ,
          <year>2024</year>
          . URL:https://proceedings.mlr.press/v235/paolo24a.html.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.</given-names>
            <surname>Varela</surname>
          </string-name>
          , E. Thompson, and E. Rosch, “
          <source>The Embodied Mind: Cognitive Science and Human Experience”</source>
          , Cambridge: Massachusetts Institute of Technology,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Clark</surname>
          </string-name>
          , “Being There - Putting
          <string-name>
            <surname>Brain</surname>
          </string-name>
          , Body and World Together Again”. Cambridge, MA: MIT Press,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>L.</given-names>
            <surname>Shapiro</surname>
          </string-name>
          , “Embodied Cognition”, Routledge,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          , K. Zhu, ... &amp;
          <string-name>
            <surname>X. Xie</surname>
          </string-name>
          ,
          <article-title>A survey on evaluation of large language models</article-title>
          ,
          <source>in: ACM Transactions on Intelligent Systems and Technology</source>
          ,
          <year>2024</year>
          ,
          <volume>15</volume>
          (
          <issue>3</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>45</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>