<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Toward Quantum Social Robotics: a Hybrid Architecture for Emotion and Coping Management</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Barbara Castrignano</string-name>
          <email>barbara.castrignano@herovision.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Berardina Nadja De Carolis</string-name>
          <email>berardina.decarolis@uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Fiore</string-name>
          <email>andrea.fiore@dotandmedia.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marco Romei Longhena</string-name>
          <email>marco.longhena@dotandmedia.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vito Nicola Losavio</string-name>
          <email>v.losavio5@studenti.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maria Grazia Miccoli</string-name>
          <email>mariagrazia.miccoli@herovision.it</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giuseppe Palestra</string-name>
          <email>giuseppe.palestra@herovision.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Bari “Aldo Moro"</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>DotandMedia srl</institution>
          ,
          <addr-line>Milan</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Hero srl</institution>
          ,
          <addr-line>Apulia</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>This paper presents a novel architecture for quantum-enhanced social robotics designed to support users emotionally through the personalized suggestion of coping strategies. The system follows a classical-quantum-classical processing pipeline: user inputs (text or multimodal) are encoded into embeddings and compressed via autoencoders trained with triplet loss. An 8-qubit variational quantum classifier then detects the underlying emotional state, which is used to search a set of potential coping strategies using Grover's algorithm. Each user is associated with a personalized probability table inspired by Q-learning, allowing the system to adapt over time through feedback. The selected strategy is communicated to the user using a language model that tailors responses empathetically based on the detected emotion. Future developments include the integration of incremental learning mechanisms to support the addition of new strategies and the management of data drift in emotion classification. The proposed solution demonstrates the potential of using quantum computing to create more adaptive and emotionally responsive human-robot interactions.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Quantum Social Robotics</kwd>
        <kwd>VQC</kwd>
        <kwd>Grover's algorithm</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Robotics is one of the research fields undergoing continuous development. It was initially established
as a framework to enable machines to become autonomous in perceiving their environment and
performing actions aimed at achieving specific goals. Over time, with the objective of assisting humans
in a broad range of tasks, robotics has evolved into an interdisciplinary field, driven by the introduction
of new technologies and frameworks that have shaped the advancement of today’s highly sophisticated
machines [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Concurrently, Quantum Computing is undergoing rapid development, with the promise
of unlocking transformative opportunities across a broad spectrum of technological domains. Among
these, robotics stands out as a key area of exploration and potential innovation. As highlighted by Yan
et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] in their comprehensive study, quantum computing, based on the fundamental principles of
quantum mechanics such as superposition and entanglement, ofers significantly increased processing
speed, enabling the resolution of problems that classical frameworks have struggled to address over time.
While Computational Complexity, also known as runtime complexity, remains the most well-known
advantage of quantum algorithms, it is by no means the only one. Alchieri et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] have identified
additional dimensions for evaluating the benefits of quantum approaches, particularly within quantum
machine learning,the discipline most frequently applied in robotics:
1. Sample complexity: Refers to the number of training examples required for algorithms to efectively
generalize.
2. Noise robustness: Describes the ability of models to remain stable even when the dataset contains
corrupted or suboptimal data.
3. Model complexity: Relates to the model’s expressive power.
      </p>
      <p>
        In light of this, growing attention has been devoted to the application of quantum information
processing across diverse domains, including Social Robotics. This subfield of robotics focuses on
developing machines capable of interacting with humans in a natural, emotionally resonant manner,
ofering assistance in various aspects of everyday life. Within this context, quantum computing has been
explored for tasks such as emotion recognition, decision-making, cognitive modelling, and pathfinding.
Nonetheless, as highlighted by De Carolis et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], a major challenge in this area remains the complexity
of implementing quantum circuits, largely due to current hardware limitations.
      </p>
      <p>The present work is part of the broader QUADRI Project (QUAntum enhanceD human-Robot
Interaction: Pioneering Intelligent Social Robotics), which aims to advance the state of the art in quantum
social robotics by proposing innovative solutions that leverage quantum paradigms in human-robot
interaction.</p>
      <p>
        Drawing on the framework proposed by Yan et al. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the psychological structure of human beings
can be understood as comprising three interrelated components: cognition, emotion, and behaviour.
Cognition pertains to how individuals perceive and interpret objects or situations; emotion involves the
subjective experience of feelings such as happiness, anger, or sadness, often triggered by the satisfaction
or frustration of personal needs; and behaviour refers to the broad set of actions and routines that
characterize daily life.
      </p>
      <p>To be perceived as truly empathic while remaining functionally efective, a robot must be able to
acquire contextual information, process it to identify appropriate actions, and interpret emotional cues
to tailor its responses accordingly.</p>
      <p>In this light, the study introduces a novel architecture for quantum social robots designed to provide
emotional support by suggesting personalized ways to cope with distress or reinforce positive states,
while also addressing computational constraints and the limitations of current quantum hardware. The
proposed framework leverages quantum algorithms, such as Grover’s search, performed via simulation
on classic machines, ensuring compatibility with available resources. The system operates by detecting
the user’s emotional state through textual and visual inputs, classifying it via a quantum model, and
selecting the most suitable response from a curated set of coping options using Grover’s algorithm.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Within the field of Quantum Social Robotics, two research areas have emerged as particularly prominent:
the development of cognitive modelling frameworks and the implementation of systems for emotion
recognition and decision-making. In both domains, quantum computing techniques are employed to
simulate or enhance human-like cognitive processes and emotional understanding. These approaches
aim to exploit the probabilistic and high-dimensional nature of quantum systems to model complex
human behaviours and emotional states more efectively than classical methods. In fact, from a
neuroscientific perspective, one of the most credited theories posits that human reasoning is inherently
Bayesian, operating through probabilistic inference mechanisms. While our framework relies on
quantum probabilistic models, these can be viewed as conceptually complementary to Bayesian approaches,
potentially ofering new ways to model uncertainty in human-like cognition [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        A particularly comprehensive approach to modelling afective processes in quantum social robotics
is ofered by Ho et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] through the development of Quantum Coppélia (Q-Coppélia). This framework
adopts quantum algorithms rather than fuzzy logic to improve the simulation of brain-like emotional
and decision-making processes, and builds upon a previous cognitive model known as Silicon Coppélia.
The latter employs fuzzy logic to simulate afective-cognitive dynamics by interpreting environmental
features, aligning them with internal goals, and selecting appropriate responses based on use intentions
and engagement. By transitioning to quantum computation, Q-Coppélia enhances this architecture
with mechanisms such as superposition and entanglement, enabling the representation of emotional
ambiguity, probabilistic decisions, and parallel processing, key aspects in modelling complex human-like
emotional behavior. Despite its conceptual richness, the authors specify that Q-Coppelia will encounter
technical dificulties in executing simulations due to its high number of qubits.
      </p>
      <p>
        A more implementation-oriented contribution is QUATRO [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a suite of full, computationally hard
classical and quantum-theoretic models, mapped to real quantum hardware. While it represents
a concrete step toward practical quantum cognition, the framework is still limited by the current
technologies. Another relevant efort is the model proposed by Ulyanov et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] that introduces a
control framework grounded in quantum computing, capable of inferring human emotional states and
modulating robotic behavior in response.
      </p>
      <p>
        Focusing a little more on the emotion recognition module, De Carolis et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] ofer a broad and
up-to-date overview of the state of the art, highlighting numerous attempts to apply quantum computing
to afective computing. The work ranges from conceptual models to more concrete implementations:
from early experiments on expressive humanoid robots and quantum Bayesian networks for predicting
social decisions, to the use of Quantum SVM and Quantum CNN to improve the recognition of facial,
vocal or multimodal emotions. However, despite the variety and creativity of the contributions, many
of these studies have some significant limitations. In fact, most models remain confined to theoretical
simulations or limited experiments, without systematic validation in real-world contexts, and often
lack critical reflection on the dificulties associated with scalability and hardware integration.
      </p>
      <p>
        Consequently, in line with the emerging trend in the state of the art, this article proposes the
architecture of a hybrid framework designed to ofer emotional support to users through personalised
suggestions for coping strategies. The definition of coping adopted is inspired by the concept introduced
by Lazarus and Folkman [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], according to which coping represents the set of behaviours implemented
by an individual to deal with situations perceived as relevant and potentially superior, or in any
case burdensome, compared to their emotional, cognitive or behavioural resources. Furthermore, the
proposed architecture aims to maintain a high degree of implementation flexibility, avoiding constraints
that would limit its adoption. It has been designed to be accessible and usable through simulators,
as well as to ensure good performance in terms of eficiency and adaptability in diferent application
contexts.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>As noted by Mennone et al. [9], the human mind remains partly irrational and dificult to fully
formalize, especially in its emotional and instinctive aspects. A viable approach, they argue, is to design
computational models that imitate typical human responses to external stimuli. For instance, in their
work, Mennone et al. adopt the cognitive model by D’Ariano and Faggin [10] to guide the behavior
of a dancing robot. Their system follows a classical–quantum–classical flow: it processes classical
inputs through a quantum-inspired layer to generate classical outputs. Following their example, the
proposed architecture is structured around three main components: the collection of classical inputs,
quantum-based processing, and the generation of classical outputs.</p>
      <p>In the context of Quantum Social Robotics, where the goal is to develop robots or assistive platforms
that can support individuals not only eficiently but also empathetically, our objective is to design a
framework capable of interpreting emotional states from user input. Specifically, the system takes
either text alone or a combination of text and image as input, and classifies the underlying emotion
using quantum classifiers. Based on the detected emotion, it then searches a dataset for the coping
strategy the user is most likely to adopt in response. At this point, the process returns to the classical
domain, where the selected strategy is communicated to the user in an empathic and supportive manner,
either to help them navigate dificult moments or to reinforce positive experiences. Figure 1 provides
an overview of the various modules that make up the architecture.</p>
      <sec id="sec-3-1">
        <title>3.1. Classical data pre-processing</title>
        <p>The first module (A in Figure 1) of the system is dedicated to the collection and preprocessing of classical
input data, which can consist of:
1. Textual data, used to capture the user’s verbal expressions and infer the associated emotional
state.
2. Visual data (images), used to extract emotions from facial expressions.</p>
        <p>The framework is designed to operate in either unimodal (text only) or multimodal (text plus image)
mode. Multimodal processing is activated only with the user’s explicit consent to being photographed.
Once the input is received, the data undergo standard preprocessing procedures. For textual data, this
includes classic NLP steps such as tokenization, stop-word removal, and normalization. For images, the
preprocessing step includes resizing and standardization to ensure compatibility with the embedding
extraction model. Following the initial cleaning steps, both text and image data are transformed
into high-dimensional embedding vectors using domain-specific, fine-tuned transformers trained to
capture emotional features. However, quantum classifiers in our pipeline operate on input vectors of 8
dimensions, therefore, a dimensionality reduction phase is required.</p>
        <p>To achieve this, we employ autoencoders trained independently for each modality. These networks
learn to compress the high-dimensional embeddings into an 8-dimensional latent space while preserving
the most relevant emotional features. In the multimodal case, the text and image embeddings (already
reduced to 8 dimensions each) are concatenated into a 16-dimensional vector. This vector is then passed
through a fusion layer, trained to reduce it to 8 dimensions for quantum processing.</p>
        <p>However, emotional data often exhibit high intra-class variability and inter-class overlap, making it
dificult for classifiers to distinguish between nuanced emotions. To address this, we incorporate Triplet
Loss into the training objective of the autoencoders and of the layer.</p>
        <p>Triplet Loss The triplet loss is a powerful objective function introduced to learn embeddings that
reflect structured relationships among samples. Originally formalized in the context of large-margin
nearest neighbor classification [ 11] and later popularized by FaceNet [12], it has proven particularly
efective for tasks such as face verification and emotion classification.</p>
        <p>In simple terms, the triplet loss forces the model to pull embeddings of similar samples (anchor and
positive) closer together, while pushing embeddings of dissimilar samples (anchor and negative) farther
apart by at least a margin . The concept is illustrated in the Figure 2.</p>
        <p>Mathematical formulation. Formally, let  () ∈ R denote the embedding of an input  into a
-dimensional Euclidean space, normalized such that ‖ ()‖2 = 1. Given a triplet (, , ), where
 is an anchor sample,  is a positive sample from the same class, and  is a negative sample from a
diferent class, the objective is to ensure that the distance between the anchor and positive embeddings
is smaller than the distance between the anchor and negative embeddings. In other words, we want
|| −  ‖22 +  &lt; ‖  −  ‖22,
∀(, , ) ∈  .</p>
        <p>where  is a margin enforced between positive and negative pairs, and  is the set of all possible
triplets in the training set with cardinality  .</p>
        <p>The loss being minimized is then defined as follows: it sums over all triplets, penalizing cases where
the distance between the anchor and positive (plus the margin) exceeds the distance between the anchor
and negative:</p>
        <p>= ∑︁ [︀ ‖ ( ) −  ( )‖22 − ‖ ( ) −  ( )‖22 +  ]︀ + .</p>
        <p>=1</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Quantum emotion Classifier</title>
        <p>Once the input data has been compressed into an 8-dimensional latent vector, it is passed to the second
module: the emotion classification module (B in Figure 1). Depending on the operational mode, this
module may receive embeddings from textual input only (unimodal) or from both text and image
sources (multimodal). In both cases, the goal is to identify the user’s emotional state.</p>
        <p>Our classification scheme follows the model of fundamental emotions devised by Paul Ekman [ 13],
which posits the universality of a discrete set of afective states. The classifier is therefore trained
to recognize the following 7 classes: Anger, Disgust, Fear, Happiness, Sadness, Surprise and Neutral,
added to represent the absence of strong emotional arousal. To perform this task, we take advantage
of Variational Quantum Classifier (VQC) operating on an 8-qubit quantum circuit, which matches the
dimensionality of the input embedding.</p>
        <p>VQC As described by Yen-Chi Chen et al. [14], a Variational Quantum Circuit (VQC) (called also
Parameterized Quantum Circuit (PQC)),is often used as an implementation of Quantum Neural Network
(QNN). VQC is generally structured into three main stages:</p>
        <p>where Θ collects all trainable parameters across the layers.
• Encoding circuit: This part, denoted by  (⃗), maps a classical input vector ⃗ into a quantum
state. It does so by applying a unitary transformation to the initial state |0⟩⊗ (where  is the
number of qubits), resulting in the encoded state
• Parameterized (variational) circuit: Represented by  (Θ) , this circuit consists of multiple
layers of trainable quantum gates. Each layer can be viewed as a sub-circuit  (⃗ ) with its own
parameter vector ⃗ . The entire variational circuit is then expressed as</p>
        <p>(⃗) |0⟩⊗
 (Θ) =
1
∏︁  (⃗ ),
=</p>
        <p>• Measurement: After the encoding and variational circuits, measurements are performed on
ˆ
specific observables  to extract classical information from the quantum state.</p>
        <p>Putting these stages together, the final quantum state prepared by the VQC becomes:
This quantum architecture realizes a quantum function:
|Ψ⟩ =  (Θ) (⃗) |0⟩</p>
        <p>⊗ .
→−− − −
 (⃗; Θ)⃗ =</p>
        <p>︀( ⟨ˆ1⟩, . . . , ⟨ˆ⟩︀)
where each expectation value is given by:
⟨⟩ = ⟨0|  †(⃗) †(Θ) ˆ (Θ) (⃗) |0⟩</p>
        <p>ˆ</p>
        <p>Such expectation values are typically estimated through repeated measurements (shots) on quantum
hardware or via quantum simulators running on classical computers. In Figure 3, we can see an example
of VQC architecture.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Quantum Search to Obtain Coping Strategy</title>
        <p>Reinforcement Learning, and more recently its quantum counterpart, has been extensively applied in
the context of robotics, particularly for autonomous decision-making. Among the most prominent
algorithms in this area is Q-learning, a value-based method in which an agent learns an optimal
policy through interaction with the environment [15]. As Skolik et al. discussed in their article [16],
Q-learning relies on the construction and iterative update of a Q-table, which stores the estimated utility
(Q-values) of performing a given action in a given state. This table is updated using feedback in the
form of rewards, gradually improving the agent’s decision-making over time.</p>
        <p>While our architecture does not implement Q-learning in the strict sense, its conceptual foundation
inspired the design of the quantum search module (C in the Figure 1). In particular, we adopt the notion
of a personalized Q-table to guide the selection of coping strategies tailored to each user.</p>
        <p>Each user maintains a custom table structured as follows:
• Columns: a set of general-purpose coping strategies;
• Rows: the seven emotional states defined by Ekman’s theory (anger, disgust, fear, happiness,
sadness, surprise, neutral);
• Cells: probabilities indicating the likelihood that a given strategy (column) will be appropriate
and well-received when the user is experiencing a specific emotion (row).</p>
        <p>Initially, all strategies are considered equally probable, with each cell initialized to 1/ , where  is
the number of available coping strategies. As the system interacts with the user and receives feedback,
these probabilities are updated accordingly:
(4)
(5)
(6)
• If the user rejects a suggested strategy, the associated probability for that emotion is decreased;
• If the user accepts the suggestion, the probability is increased.</p>
        <p>To ensure diversity and prevent convergence to a single repetitive suggestion, upper and lower bounds
are imposed on these probabilities, and normalization is enforced so that the sum of all probabilities for
a given emotion always equals 1.</p>
        <p>The final selection of the coping strategy is performed using a quantum search method based on
Grover’s algorithm. Originally developed for unstructured database search, Grover’s algorithm
provides a quadratic speedup over classical search, allowing the identification of a marked item among
 elements in approximately (√ ) queries [17].</p>
        <p>Grover’s algorithm operates by amplifying the probability amplitude of the desired state through
iterative applications of two operations:
1. Oracle operator: marks the desired item (in our case, the strategy with highest probability);
2. Difusion operator: inverts all amplitudes about their average to reinforce the marked state.</p>
        <p>In our setting, the search space corresponds to the set of candidate coping strategies for a given
emotional state, and the oracle is defined to mark the strategy with the highest current acceptance
probability in the user’s table. By applying Grover iterations, the system amplifies the most suitable
option and retrieves it with high probability after a limited number of queries, thereby reducing the
search complexity while preserving adaptability to user-specific patterns.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Empathic Output Generation</title>
        <p>Once the most appropriate coping strategy has been identified through quantum search, it must be
communicated to the user in a supportive and personalized manner. The last module of the pipeline (D
in the Figure 1) is handled by the classical processing of the empathic response module, which leverages
a natural language model to convey the suggestion.</p>
        <p>To ensure privacy, the language model operates using only two inputs:
• the detected emotional state, classified in the previous module;
• the selected coping strategy, retrieved via Grover’s search algorithm.</p>
        <p>No other user-specific or sensitive information is transmitted to the language model, ensuring both
compliance with ethical principles and respect for the user’s personal data. So, this layer transforms
the abstract coping strategy into a concrete, human-understandable suggestion, thereby closing the
interaction loop and supporting the user in navigating their emotional experience.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Preliminary Experiments</title>
      <p>To provide initial empirical support to the proposed architecture, we conducted preliminary experiments
focusing on the modules described in Sections 3.1 and 3.2, i.e., the classical pre-processing and the
quantum emotion classification. A multimodal dataset was constructed by combining textual data from
GoEmotions [18] and facial expressions from KDEF [19], mapped into Ekman’s seven basic emotions.
Textual and visual embeddings were compressed into an 8-dimensional latent space via autoencoders
trained with Triplet Loss.</p>
      <p>An 8-qubit Variational Quantum Classifier (VQC) was then trained on these embeddings. With the
optimized multimodal embeddings and advanced quantum feature maps, classification performance
reaches 93% test accuracy in the fused (text+image) setting and 95% in the text-only case. These
results, although preliminary, indicate the feasibility of integrating quantum classifiers into an emotion
recognition pipeline and validate the design choices of the proposed architecture.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>The architecture presented in this work lays the foundation for an emotionally aware quantum-enhanced
assistive system capable of suggesting personalized coping strategies. While the current implementation
provides a complete pipeline, from emotion recognition to empathic communication, future
developments will aim to introduce an incremental learning component. On one hand, the quantum search
module will be extended to support the dynamic addition of new coping strategies over time. This
requires updating both the user-specific probability tables and the oracle used in Grover’s algorithm,
ensuring that the system remains adaptable as user needs evolve or as new strategies are introduced. On
the other hand, the emotion classification module will be enhanced to handle data drift and user-specific
patterns via incremental updates to the quantum classifier. By incorporating adaptive mechanisms, such
as continual learning of the variational parameters, the model will maintain robustness and accuracy in
dynamic environments where the user’s emotional responses may shift gradually. This dual incremental
design supports long-term personalization and aligns with the goals of sustainable, human-centric
social robotics.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work has been financed by the Research Program “National Quantum Science and Technology
Institute” (PE0000023-NQSTI). It is supported within the framework of the National Recovery and
Resilience Plan (NRRP), Mission 4 “Education and Research” – Component 2 “From Research to Business”,
with explicit reference to funding provided by the European Union through the NextGenerationEU
initiative.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used GPT-4o in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and
take(s) full responsibility for the publication’s content.
[9] M. Mannone, A. Chella, G. Pilato, V. Seidita, F. Vella, S. Gaglio, Modeling robotic thinking and
creativity: A classic–quantum dialogue, Mathematics 12 (2024). URL: https://www.mdpi.com/
2227-7390/12/5/642. doi:10.3390/math12050642.
[10] G. M. D’Ariano, F. Faggin, Hard Problem and Free Will: An Information-Theoretical
Approach, Springer International Publishing, Cham, 2022, pp. 145–192. URL: https://doi.org/10.
1007/978-3-030-85480-5_5. doi:10.1007/978-3-030-85480-5_5.
[11] K. Q. Weinberger, L. K. Saul, Distance metric learning for large margin nearest neighbor
classification., Journal of machine learning research 10 (2009).
[12] F. Schrof, D. Kalenichenko, J. Philbin, Facenet: A unified embedding for face recognition and
clustering, in: Proceedings of the IEEE conference on computer vision and pattern recognition,
2015, pp. 815–823.
[13] P. Ekman, T. Dalgleish, M. Power, Basic emotions, San Francisco, USA (1999).
[14] S. Y.-C. Chen, Z. Liang, Introduction to quantum machine learning and quantum architecture
search, in: 2025 IEEE International Symposium on Circuits and Systems (ISCAS), IEEE, 2025, pp.
1–6.
[15] L. Hu, C. Wei, L. Yin, Fuzzy a quantum multi-stage q-learning artificial potential field for path
planning of mobile robots, Engineering Applications of Artificial Intelligence 141 (2025) 109866.
[16] A. Skolik, S. Jerbi, V. Dunjko, Quantum agents in the gym: a variational quantum algorithm
for deep q-learning, Quantum 6 (2022) 720. URL: http://dx.doi.org/10.22331/q-2022-05-24-720.
doi:10.22331/q-2022-05-24-720.
[17] P. R. Giri, V. E. Korepin, A review on quantum search algorithms, Quantum Information Processing
16 (2017) 315.
[18] D. Demszky, D. Movshovitz-Attias, J. Ko, A. Cowen, G. Nemade, S. Ravi, Goemotions: A dataset of
ifne-grained emotions, arXiv preprint arXiv:2005.00547 (2020).
[19] M. G. Calvo, D. Lundqvist, Facial expressions of emotion (kdef): Identification under diferent
display-duration conditions, Behavior research methods 40 (2008) 109–115.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Iliyasu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Salama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Hirota</surname>
          </string-name>
          ,
          <article-title>Quantum robotics: a review of emerging trends</article-title>
          ,
          <source>Quantum Machine Intelligence</source>
          <volume>6</volume>
          (
          <year>2024</year>
          )
          <fpage>86</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Alchieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Badalotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bonardi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bianco</surname>
          </string-name>
          ,
          <article-title>An introduction to quantum machine learning: from quantum logic to quantum deep learning</article-title>
          ,
          <source>Quantum Machine Intelligence</source>
          <volume>3</volume>
          (
          <year>2021</year>
          )
          <fpage>28</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B. N.</given-names>
            <surname>De Carolis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Loglisci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Miccoli</surname>
          </string-name>
          , G. Palestra,
          <string-name>
            <given-names>S.</given-names>
            <surname>Violante</surname>
          </string-name>
          ,
          <article-title>Quantum-enhanced social robotics: The quadri project</article-title>
          ,
          <source>in: Adjunct Proceedings of the 33rd ACM Conference on User Modeling, Adaptation and Personalization</source>
          ,
          <year>2025</year>
          , pp.
          <fpage>151</fpage>
          -
          <lpage>155</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T. L</given-names>
            <surname>Grifiths</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kemp</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. B Tenenbaum</surname>
          </string-name>
          ,
          <article-title>Bayesian models of cognition (</article-title>
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Hoorn</surname>
          </string-name>
          ,
          <article-title>Quantum afective processes for multidimensional decision-making</article-title>
          ,
          <source>Scientific reports 12</source>
          (
          <year>2022</year>
          )
          <fpage>20468</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R. P.</given-names>
            <surname>Pothukuchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lufkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Simon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Thorstenson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. E.</given-names>
            <surname>Trevisan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Foxman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. S.</given-names>
            <surname>Pothukuchi</surname>
          </string-name>
          , et al.,
          <article-title>The quatro application suite: Quantum computing for models of human cognition</article-title>
          ,
          <source>arXiv preprint arXiv:2309.00597</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S. V.</given-names>
            <surname>Ulyanov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Kurawaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. S.</given-names>
            <surname>Ulyanov</surname>
          </string-name>
          , T. Hagiwara,
          <article-title>Unconventional cognitive intelligent robotic control: Quantum soft computing approach in human being emotion estimation-qcoptkb toolkit application</article-title>
          ,
          <source>arXiv preprint arXiv:2307.06858</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Biggs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Brough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Drummond</surname>
          </string-name>
          ,
          <article-title>Lazarus and folkman's psychological stress and coping theory, The handbook of stress and health: A guide to research and practice (</article-title>
          <year>2017</year>
          )
          <fpage>349</fpage>
          -
          <lpage>364</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>