<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Proceedingsofthe1stInternational workshoponMultimodalArtificial IntelligenceinEducation</article-title>
      </title-group>
      <abstract>
        <p>onlinefrom Utrecht,TheNetherlands</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>IntelligenceinEducation
atthe22ndInternationalConferenceonArtificial
IntelligenceinEducation
Please refer to these proceedings as:</p>
      <p>Daniele Di Mitri, Roberto Martinez-Maldonado, Olga C. Santos, Jan
Schneider, Khaleel Asyraaf Mat Sanusi, Mutlu Cukurova, Daniel Spikol,
Inge Molenaar, Michail Giannakos, Roland Klemke &amp; Roger Azevedo
(eds.): Proceedings of the 1st International Workshop on Multimodal
Arti cial Intelligence in Education. At the 22nd International Conference
on Arti cial Intelligence in Education. Online, Utrecht, The Netherlands
June 14, 2021, CEUR-WS.org/Vol-2902, ISSN 1613-0073.
© 2021 for the individual papers by the papers' authors. Copying permitted
for private and academic purposes. Re-publication of material from this volume
requires permission by the copyright owners.</p>
      <p>Address of rst editor:
Daniele Di Mitri
DIPF | Leibniz Institute for Research and Information in Education
Rostocker Str. 6, 60323 Frankfurt am Main, Germany.
dimitri@dipf.de</p>
    </sec>
    <sec id="sec-2">
      <title>Preface</title>
      <p>
        Learning is a complex multidimensional process that involves multiple sources of
data. In current Arti cial Intelligence (AI) in Education research, however, many
modalities of data are neglected to a large extent. Traditionally, the AI-based
systems designed for education, typically Intelligent Tutoring Systems (ITS),
rely on click-stream data generated from computer interaction with a mouse
and keyboard or screen interaction in the case of mobile devices. The variety of
interactions, multi-sensor devices and multimodal data can provide a more
detailed digital representation of the learner. There is an increasing interest in the
data-driven educational research communities in multimodal and multi-sensor
interaction methods as an opportunity. The physiological sensors can provide a
wealth of information that can be used further to contextualise the learning
performance or strategies [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The analysis of multimodal data for learning, however,
poses a series of promises [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] as well as challenges [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Multimodal data have low
semantic value and increasing complexity. For this reason, it is harder for human
intelligence to scout for patterns and regularities in multidimensional datasets.
      </p>
      <p>
        Existing studies using Multimodal AI for education approaches can be
divided into two groups. The rst group aims at extending computer-based
learning activities with physical sensors for tracking learners' behaviour, including,
for example, learner's eye-gaze, facial expressions or physiological responses such
as heart rate or neural activation (e.g. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]). The second group is concerned with
tracing learning activities \beyond mouse and keyboard" which belong to the
domain of learning of psychomotor learning, i.e. practical learning activities that
require levels of physical coordination. This emerging research direction was
grounded in the Special Issue \The Next 25 Years: How Advanced Interactive
Learning Technologies will Change the World" of the International Journal of
AIED [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>In AI, there is a philosophical discussion on whether AI systems are set to
augment human abilities or replace them entirely. This discussion carries over
to Multimodal AI in Education, across the degree of intervention that the AI
assumes in the learning processes and to what extent it is set to replace human
support and feedback.</p>
      <p>
        On the \Multimodal augmentation" side, the AI algorithms are used for
nding patterns and regularities in the data, which are then communicated to human
actors in the learning process. This approach is advocated by the \CrossMMLA"
community { learning across physical and digital spaces using multimodal
learning analytics, a particular interest group within the Society of Learning Analytics
Research [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>The \intelligent tutoring" aims instead to the automatic provision of
automatic feedback using intelligent and autonomous agents has been a longstanding
discussion and topic of investigation within the Arti cial Intelligence in
Education (AIED) community. Although only recently, the community has started to
look into more diverse data sources alternative to learner-computer interaction.</p>
      <p>With this context in mind, in this rst edition of the International Workshop
on multimodal Arti cial Intelligence in Education (MAIEd 2021), we have
compiled ten research studies that go from early stages of developments to present
empirical studies where novel experimental designs, theoretical contributions and
practical demonstrations. MAIEd 2021 took place on June 14th, 2021 and was
run virtually in conjunction with the 22nd International Conference on Arti cial
Intelligence in Education (AIED 2021).</p>
      <p>In particular, at the MAIEd workshop, we have discussed which scienti c,
state-of-the-art ideas and approaches are being pursued and which impacts we
expect on educational technologies and education, targeting the intersection of
these two elds of AI and multimodal interaction. We tried to advance state of
art in theories, technologies, methods, and knowledge towards the development
of multimodal intelligent tutors with the following workshop topics:
{ Multimodal Intelligent Tutoring Systems or User Interfaces
{ Multimodality in Augmented, Virtual, and Mixed Reality
{ Multimodal Learner Modeling and A ective Computing
{ Adaptive Feedback, Guidance, and Process in Multimodal Learning
{ Arti cial Intelligence for Learning Analytics
{ Big Data-driven Visual Analytics for Learning
{ Error detection and classi cation for multimodal data
{ Cognitive Load in Multimodal Interaction with Intelligent Tutoring Systems
{ Explainability, Trust, and Safety in Multimodal Intelligent Tutoring Systems
{ Multimodal data for Self-Regulated learning</p>
      <p>The website of the workshop can be found here https://maied.edutec.science/.
A video recording of the event is avaialble upon request.</p>
    </sec>
    <sec id="sec-3">
      <title>Contributions</title>
      <p>A peer-reviewed process was carried out to select the workshop papers. At least
three members of the Program Committee with expertise in the area reviewed
each paper. As a result, ten submissions were accepted (out of 13 submissions),
which discuss ideas and progress on several interesting topics, such as multimodal
analytics for a ective computing, multimodal intelligent tutoring, self-regulated
learning, multimodal intelligence augmentation.</p>
      <p>Jiang et al. investigate if an ITS gives supportive, empathetic, or motivational
feedback messages to the learner to alter the learner's emotional state and detect
the change.</p>
      <p>Huang, Fridolin and Whitelock carry a comparative analysis of holographic
intelligent agents by analysing nine `Holographic AIs' characters. They derive
various design dimensions and principles for future research.</p>
      <p>Zhou, Wannapon and Cukurova explore self-regulated learner clusters'
engagement behaviours at individual, group and cohort activities. Their study
analyses the relationship between students' SRL competence and their
learning engagement behaviours observed in multimodal data. The results revealed
that students with di erent SRL competence clusters might exhibit di erent
behaviours in individual, group, and cohort level activities.</p>
      <p>Lim and Leinonen propose an experimental design for fostering creativity
with AI in multimodal and sociocultural learning environments. In the proposed
Creative Peer System, humans and machines learn from each other in a
multimodal learning environment and develop original artefacts.</p>
      <p>Ronda et al. focus on simulated teamwork practice and how multimodal
data processing can be performed to identify if the arousal levels matched the
teachers' expectations regarding the students' emotional situation in the di erent
phases in which the teachers' teamwork practice can be divided according to the
instructional design.</p>
      <p>Yang, Cukurova and Porayska-Pomsta focus on dyadic joint visual attention
interaction in face-to-face collaborative problem-solving at K-12 Maths
Education. Their results indicate that the multimodal approach can bring more insights
into students' problem-solving. In addition, they propose a method for capturing
gaze convergence by considering eye xations, eye blinks, and the overlapping
time between two eye gazes.</p>
      <p>Echeverria and Santos introduce KUMITRON, a multimodal psychomotor
intelligent learning system that can provide personalised support when training
karate combats.</p>
      <p>Howell-Munson et al. move preliminary steps towards detecting proactive and
reactive control states during learning with fNIRS brain signals. They distinguish
between proactive and reactive control using fNIRS brain imaging in a controlled
continuous performance task. They also propose integrating the fNIRS
datastream with the ITS to create a multimodal system to detect the user's cognitive
state and adapt the environment to promote better learning strategies.</p>
      <p>Gupta et al. focus on multimodal and multi-task stealth assessment for
re ection-enriched game-based learning. They present a stealth assessment
framework that takes as input multimodal data streams (e.g., game trace logs, pre-test
data, natural language responses to in-game re ection prompts) to predict
posttest scores and written re ection depth scores jointly.</p>
      <p>Lee-Cultura, Sharma and Giannakos propose a multimodal AI agent to
support students' motion-based educational gameplay. The AI agent identi es and
delivers appropriate feedback mechanisms to support a student's play learning
experience. A Dashboard visualises the measurements to keep teachers informed
of a student's progress.</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusions</title>
      <p>MAIEd 2021 workshop aimed at gathering new insights around the use of Arti
cial Intelligence (AI) systems and autonomous agents for education and learning
leveraging multimodal data sources. It built upon the CrossMMLA workshop
series at the Learning Analytics &amp; Knowledge Conference and called for new
empirical studies, even if in their early stages of development. As a result, ten
papers have been accepted for presentation at the workshop that consists of
novel experimental designs, theoretical contributions, and practical
demonstrations which can prove the use of multimodal and multi-sensor devices \beyond
mouse and keyboard" in learning contexts with the purpose of automatic
feedback generation, adaptation, and personalisation in learning. Through the
organisation of this rst edition of the workshop, we sought to engage the scienti c
community in opening up the scope of AI in Education towards novel and diverse
data sources.</p>
      <p>The MAIEd 2021 chairs would like to thank the authors for their submissions
and the AIED workshop chairs for their advice and guidance during the MAIEd
workshop. The MAIEd chairs also served as Program Committee that reviewed
high quality reviews for the received submissions.</p>
      <p>The following project has partially supported the organisation of the MAIEd
2021 workshop: \INTelligent INTra-subject development approach to improve
actions in AFFect-aware adaptive educational systems" INT2AFF funded under
Grant PGC2018-102279-B-I00 (MCIU/AEI/FEDER, UE) by the Spanish
Ministry of Science, Innovation and Universities, the Spanish Agency of Research
and the European Regional Development Fund (ERDF).</p>
    </sec>
    <sec id="sec-5">
      <title>Measuring the E ect of ITS Feedback Messages on Students' Emotions</title>
      <p>Han Jiang, Zewelanji Serpell, and Jacob Whitehill . . . . . . . . . . . . . . . . . . . . . . 7</p>
    </sec>
    <sec id="sec-6">
      <title>Design Dimensions for Holographic Intelligent Agents: A Comparative</title>
    </sec>
    <sec id="sec-7">
      <title>Analysis</title>
      <p>Xinyu Huang, Fridolin Wild, and Denise Whitelock . . . . . . . . . . . . . . . . . . . . 17</p>
    </sec>
    <sec id="sec-8">
      <title>Di erent modality, di erent design, di erent results: Exploring selfregulated learner clusters' engagement behaviours at individual, group and cohort activities</title>
      <p>Qi Zhou, Wannapon Suraworachet, and Mutlu Cukurova . . . . . . . . . . . . . . . 27</p>
    </sec>
    <sec id="sec-9">
      <title>Creative Peer System: An Experimental Design for Fostering Cre</title>
      <p>ativity with Arti cial Intelligence in Multimodal and Sociocultural</p>
    </sec>
    <sec id="sec-10">
      <title>Learning Environments</title>
      <p>Jeongki Lim and Teemu Leinonen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40</p>
    </sec>
    <sec id="sec-11">
      <title>Towards Exploring Stress Reactions in Teamwork using Multimodal</title>
    </sec>
    <sec id="sec-12">
      <title>Physiological Data</title>
    </sec>
    <sec id="sec-13">
      <title>Dyadic joint visual attention interaction in face-to-face collaborative problem-solving at K-12 Maths Education: A Multimodal Approach</title>
      <p>Chiao-Wei Yang, Mutlu Cukurova, and Kaska Porayska-Pomsta . . . . . . . 60</p>
    </sec>
    <sec id="sec-14">
      <title>KUMITRON: A Multimodal Psychomotor Intelligent Learning System to Provide Personalized Support when Training Karate Combats</title>
      <p>Jon Echeverria and Olga C. Santos . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70</p>
    </sec>
    <sec id="sec-15">
      <title>Preliminary steps towards detection of proactive and reactive control states during learning with fNIRS brain signals</title>
      <p>Alicia Howell-Munson, Deniz Sonmez Unal, Erin Walker, Catherine
Arrington, and Erin Solovey . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82</p>
    </sec>
    <sec id="sec-16">
      <title>Multimodal Multi-Task Stealth Assessment for Re ection-Enriched</title>
    </sec>
    <sec id="sec-17">
      <title>Game-Based Learning</title>
      <p>Anisha Gupta, Dan Carpenter, Wookhee Min, Jonathan Rowe, Roger Azevedo,
and James Lester . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92</p>
    </sec>
    <sec id="sec-18">
      <title>Multimodal AI Agent to Support Students' Motion-Based Educational Game Play</title>
      <p>Serena Lee-Cultura, Kshitij Sharma, and Michail Giannakos . . . . . . . . . . 102</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Cukurova</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giannakos</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martinez-Maldonado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>The promise and challenges of multimodal learning analytics</article-title>
          .
          <source>British journal of educational technology</source>
          ,
          <volume>51</volume>
          (
          <issue>5</issue>
          ),
          <fpage>1441</fpage>
          -
          <lpage>1449</lpage>
          .https://doi.org/10.1111/bjet.13015
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Di</given-names>
            <surname>Mitri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Specht</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            , &amp;
            <surname>Drachsler</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          (
          <year>2018</year>
          ).
          <article-title>From signals to knowledge: A conceptual model for multimodal learning analytics</article-title>
          .
          <source>Journal of Computer Assisted Learning</source>
          ,
          <volume>34</volume>
          (
          <issue>4</issue>
          ),
          <volume>338</volume>
          {
          <fpage>349</fpage>
          . https://doi.org/10.1111/jcal.12288
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Giannakos</surname>
            ,
            <given-names>M. N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pappas</surname>
            ,
            <given-names>I. O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kostakos</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Velloso</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Multimodal data as a means to understand the learning experience</article-title>
          .
          <source>International Journal of Information Management</source>
          ,
          <volume>48</volume>
          (February),
          <volume>108</volume>
          {
          <fpage>119</fpage>
          . https://doi.org/10.1016/j.ijinfomgt.
          <year>2019</year>
          .
          <volume>02</volume>
          .003
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Giannakos</surname>
            ,
            <given-names>M. N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spikol</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molenaar</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Di</given-names>
            <surname>Mitri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Ochoa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            , &amp;
            <surname>Hammad</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Proceedings of CrossMMLA in practice: Collecting, annotating and analyzing multimodal data across spaces co-located with 10th International Learning</article-title>
          and Analytics &amp; Knowledge
          <string-name>
            <surname>Conference</surname>
          </string-name>
          (LAK
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Santos</surname>
            ,
            <given-names>O. C.</given-names>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>Training the Body: The Potential of AIED to Support Personalized Motor Skills Learning</article-title>
          .
          <source>International Journal of Arti cial Intelligence in Education</source>
          ,
          <volume>26</volume>
          (
          <issue>2</issue>
          ),
          <volume>730</volume>
          {
          <fpage>755</fpage>
          . https://doi.org/10.1007/s40593-016-0103-2 Miguel A.
          <string-name>
            <surname>Ronda-Carracao</surname>
          </string-name>
          , Olga C. Santos, Gloria Fernandez-Nieto, and
          <string-name>
            <surname>Roberto</surname>
            <given-names>Martinez-Maldonado . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48</given-names>
          </string-name>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>