=Paper= {{Paper |id=Vol-3827/invited2 |storemode=property |title=Neurosymbolic Visual Commonsense: On Integrated Reasoning and Learning about Space and Motion in Embodied Multimodal Interaction |pdfUrl=https://ceur-ws.org/Vol-3827/invited2.pdf |volume=Vol-3827 |authors=Mehul Bhatt |dblpUrl=https://dblp.org/rec/conf/strl/Bhatt24 }} ==Neurosymbolic Visual Commonsense: On Integrated Reasoning and Learning about Space and Motion in Embodied Multimodal Interaction== https://ceur-ws.org/Vol-3827/invited2.pdf
                         Neurosymbolic Visual Commonsense
                         On Integrated Reasoning and Learning about Space and Motion in
                         Embodied Multimodal Interaction

                         Mehul Bhatt
                         School of Science and Technology, Örebro University – Sweden
                         CoDesign Lab EU (Artificial and Human Intelligence).,
                         Cognitive Vision and Perception » https:// codesign-lab.org/ cognitive-vision


                                           Abstract
                                           We present recent and emerging advances in computational cognitive vision addressing artificial visual and spatial intelligence at
                                           the interface of (spatial) language, (spatial) logic and (spatial) cognition research. With a primary focus on explainable sensemaking
                                           of dynamic visuospatial imagery, we highlight the (systematic and modular) integration of methods from knowledge representation
                                           and reasoning, computer vision, spatial informatics, and computational cognitive modelling. A key emphasis here is on generalised
                                           (declarative) neurosymbolic reasoning & learning about space, motion, actions, and events relevant to embodied multimodal interaction
                                           under ecologically valid naturalistic settings in everyday life. Practically, this translates to general-purpose mechanisms for computational
                                           visual commonsense encompassing capabilities such as (neurosymbolic) semantic question-answering, relational spatio-temporal
                                           learning, visual abduction etc.
                                           The presented work is motivated by and demonstrated in the applied backdrop of areas as diverse as autonomous driving, cognitive
                                           robotics, design of digital visuoauditory media, and behavioural visual perception research in cognitive psychology and neuroscience.
                                           More broadly, our emerging work is driven by an interdisciplinary research mindset addressing human-centred responsible AI through
                                           a methodological confluence of AI, Vision, Psychology, and (human-factors centred) Interaction Design.

                                           Keywords
                                           Cognitive vision, Knowlede representation and reasoning (KR), Machine Learning, Integration of reasoning & learning, Commonsense
                                           reasoning, Declarative spatial reasoning, Relational Learning, Computational cognitive modelling, Human-Centred AI, Responsible AI



                         1. Motivation                                                                                                  spectrum of high-level human-centred sensemaking capa-
                                                                                                                                        bilities. These capabilities encompass operational functions
                         Multimodality in embodied interaction is an inherent aspect                                                    such as:
                         of human activity, be it in social, professional, or every-                                                         • Visuospatial conception formation, common-
                         day mundane contexts. Next-generation human-centred                                                                   sense/qualitative generalisation,  analogical
                         AI technologies, operating in such contextualised every-                                                              inference;
                         day settings, will require an inherent foundational capacity
                         to “make sense” of —e.g., perceive, understand, explain,                                                            • Hypothetical reasoning, argumentation, explana-
                         anticipate— everyday, naturalistic interactional multimodal-                                                          tion, counterfactual reasoning;
                         ity. This would be essential towards successfully achieving                                                         • Event based episodic maintenance & retrieval for
                         technology mediated (“human-in-the-loop” ) collaborative                                                              perceptual narrativisation.
                         assistance, as well as ensuring compliance with emerging
                         human-centred ethical and legal requirements, performance                                                      The afore enumeration is by no means exhaustive: in
                         benchmarks, and inclusive usability expectations. It is there-                                                 essence, in scope of artificial visual intelligence are diverse
                         fore crucial that the foundational building blocks of such                                                     high-level cognitive visuospatial sensemaking capabili-
                         next-generation systems be semantically aligned with the                                                       ties —be it mundane, analytical, or creative— that humans
                         descriptive, analytical, and explanatory characteristics and                                                   acquire developmentally or through specialised training,
                         complexity of human task conceptualisation, performance                                                        and are routinely adept at performing seamlessly in their
                         benchmarks, and usability expectations. Against this back-                                                     everyday life and work (e.g., driving a vehicle, tracking
                         drop, we define artificial visual intelligence [1] as:                                                         moving objects, navigating a crowded urban environment,
                                                                                                                                        engaging in sports, interpreting subtle cues in everyday
                                  » The computational capability to seman-                                                              people-communication from visual / gestural and auditory
                                  tically process and interpret diverse forms                                                           signals).
                                  of visual stimuli (typically, but not necessar-
                                  ily) emanating from sensing embodied mul-                                                             Our central focus is on the development of general,
                                  timodal interactions of / amongst humans                                                              domain-independent methods that may be seamlessly
                                  and other artefacts in diverse naturalistic                                                           integrated as part of hybrid computational cognitive system,
                                  situations of everyday life and work.                                                                 or even within computational cognitive models / cognitive
                                                                                                                                        architectures [2]. We also contextualise and demonstrate in
                         Within the scope of artificial visual intelligence are a wide-                                                 the backdrop of applications in autonomous driving, cog-
                                                                                                                                        nitive robotics, visuoauditory media design, and cognitive
                         International Joint Conference on Artificial Intelligence (IJCAI)., STRL 24:                                   psychology (e.g. [3, 4, 5, 6], [7, 8] ). Through applied case-
                         Third International Workshop on Spatio-Temporal Reasoning and Learning                                         studies, we provide a systematic model and general method-
                         (STRL), IJCAI 2024 – 5 August 2024, Jeju, South Korea                                                          ology showcasing the integration of diverse, multi-faceted
                         $ mehul.bhatt@oru.se (M. Bhatt)                                                                                AI methods pertaining Knowledge Representation and Rea-
                         € https://mehulbhatt.org (M. Bhatt)
                                   © 2024 CoDesign Lab EU. Use permitted under Creative Commons License Attribution 4.0 International   soning, Computer Vision, Machine Learning, and Visual
                                   (CC BY 4.0).



CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
Perception towards realising practical, human-centred, com-       spatial systems [10] where integrated reasoning about ac-
putational visual intelligence.                                   tion and change [11, 12] is involved:

                                                                         • interpolation and projection of missing informa-
2. Neurosymbolic Visual                                                    tion, e.g., what could be hypothesised about missing
                                                                           information (e.g., moments of occlusion [13]); how
   Commonsense: Integrated                                                 can this hypothesis support planning an immediate
   Reasoning and Learning about                                            next step?
   Space, Motion, and Inter(A)ction                                      • object identity maintenance at a semantic level,
                                                                           e.g., in the presence of occlusions, missing and noisy
In the present status quo, our research in (computational)                 quantitative data, error in detection and tracking
neurosymbolic visual commonsense categorically addresses
                                                                         • ability to make default assumptions, e.g., pertain-
three key questions:
                                                                           ing to persistence objects and/or object attributes
    I. What kind of (relational) abstraction mechanisms
                                                                         • maintaining consistent beliefs respecting (domain-
        are needed to computationally “make-sense” of em-
                                                                           neutral) commonsense criteria, e.g., related to com-
        bodied multimodal interaction ?
                                                                           positionality & indirect effects, space-time continu-
    II. How can (and why should) abstraction mechanisms                    ity, positional changes resulting from motion
        (such as in I) be founded on behaviourally estab-
        lished cognitive human- factors emanating from nat-              • inferring / computing counterfactuals [14], in a
        uralistic empirical observation in real-world applied              manner akin to human cognitive ability to perform
        contexts?                                                          mental simulation for purposes of introspection
   III. How to articulate behaviourally established abstrac-               about the past or anticipation of the future, or per-
        tion mechanisms, preferences (etc) as formal declar-               forming “what-if” reasoning tasks etc
        ative models suited for computational modelling
                                                                  We particularly emphasise the abilities to abstract, learn,
        aimed at operational“sensemaking” (encompassing
                                                                  and reason with cognitively rooted structured characterisa-
        capabilities such as abduction, relational learning,
                                                                  tions of commonsense knowledge about space and motion,
        counterfactual inference) ?
                                                                  encompassing visuospatial question-answering, abduction,
                                                                  and relational learning:
Present work is particularly aimed at developing general
methods for the semantic interpretation of (multimodal) dy-
                                                                  I.  Visuospatial Question-Answering. Focus is on a com-
namic visuospatial imagery with an emphasis on the ability
to neurosymbolically perform abstraction, reasoning, and          putational framework for semantic-question answering
learning with cognitively rooted structured characterisa-         with video and eye-tracking data founded in constraint logic
tions of commonsense knowledge pertaining to space and            programming; we also demonstrate an application in cogni-
motion. Here, we specifically emphasise:                          tive film & media studies, where human perception of films
                                                                  vis-a-via cinematographic devices is of interest.
     • General foundational commonsense abstractions of                                                           » [4, 6, 7, 8]
       space, time, and motion needed for representation
       mediated (grounded) reasoning and learning with            II.  Visuospatial Abduction. Focus is on a hybrid archi-
       dynamic visuospatial stimuli (e.g., emanating from         tecture for systematically computing robust visual explana-
       multimodal human behavioural signals in modali-            tion(s) encompassing hypothesis formation, belief revision,
       ties such as RGB(D), video, audio, eye-tracking and        and default reasoning with video data (for active vision
       possibly even bio signals [9]);                            for autonomous driving, as well as for offline processing).
     • Deep (visuospatial) semantics, entailing systemat-         The architecture supports visual abduction with space-time
       ically formalised declarative (neurosymbolic) rea-         histories as native entities, and founded in (functional) an-
       soning and learning with aspects pertaining to             swer set programming based spatial reasoning.
       space, space-time, motion, actions & events, spatio-                                               » [3, 13, 15][16, 17]
       linguistic conceptual knowledge. Here, it is of the
       essence that an expressive ontology consisting of,         III.  Relational Visuospatial Learning. Focus is on a gen-
       for instance, space, time, space-time motion primi-        eral framework and pipeline for: relational spatio-temporal
       tives as first-class ‘neurosymbolic’ objects is accessi-   (inductive) learning with an elaborate ontology supporting
       ble within the (declarative) programming paradigm          a range of space-time features; and generating semantic,
       under consideration; and                                   (declaratively) explainable interpretation models in a neu-
                                                                  rosymbolic pipeline demonstrated for the case of analysing
     • Explainable models of computational visuospatial           visuospatial symmetry in visual art.
       commonsense based on a systematic integration of                                                         » [18][5][19]
       symbolic/relational methods on the one hand, and
       neural techniques aimed at low level quantitative
                                                                  Formal semantics and computational models of deep seman-
       (e.g., visual) data processing on the other;
                                                                  tics manifest themselves as neurosymbolic spatio-temporal
At a higher level of abstraction, deep (visualspatial) se-        extensions of established declarative AI frameworks such
mantics (or deep semantics for short) entails inherent sup-       as Constraint Logic Programming (CLP) [20], Inductive
port for tackling a range of challenges concerning epistemo-      Logic Programming (ILP) [21], and Answer Set Program-
logical and phenomenological aspects relevant to dynamic          ming (ASP) [22]. The more foundational aspects pertaining
declarative spatial reasoning (built on top of CLP, ILP, ASP)       neural learning techniques, or otherwise.
independent of its relationship to cognitive vision research
                                                                    In this invited position statement, we have attempted to
may be consulted in [23], [16, 24], [18].
                                                                    summarise our mindset and ongoing work in the CoDesign
                                                                    Lab towards:

3. Discussion                                                              » Establishing a human-centric foundation
                                                                           and roadmap for the development of neu-
The vision that drives our scientific methodology is:                      rosymbolically grounded inference about
                                                                           embodied multimodal interaction as iden-
        » To shape the nature and character of                             tifiable in a range of real-world application
        (machine-based) artificial visual intelligence                     contexts.
        with respect to human-centred cognitive
        considerations, demonstrating an exemplar                   This summary is not meant to be a comprehensive literature
        for developing, applying, and disseminating                 review; this may be obtained through the cited works. For
        such methods in socio-technologically rele-                 key technical details and to obtain a summary of open di-
        vant application areas where:                               rections, we direct interested readers to select publications
                                                                    as follows: a compact starting point may be obtained via
        (a) embodied (multimodal) human interac-                    the comprehensive summary in [1], or through the shorter/-
        tion is inherent;                                           focussed components in [15, 5, 4, 13, 3]. Longer summaries
                                                                    in the form of (recent) doctoral dissertations are available
        (b) human-in-the-loop collaborative work is
                                                                    in [29] and [30, 31].
        of the essence; and
        (c) normative ethico-legal compliance based
        on regulatory requirement and human-                        Acknowledgments
        factors driven inclusive or universal design
        criteria is to be ensured.
                                                                    We acknowledge funding by the Swedish Research Council
Towards realising this vision, we adopt an interdisciplinary        (VR - Vetenskapsrådet) - https://www.vr.se, and the Swedish
approach –at the confluence of Cognition, AI, Interaction,          Foundation for Strategic Research (SSF – Stiftelsens för
and Design– which we deem necessary to better appreci-              Strategisk Forskning) - https://strategiska.se. Previously,
ate the complexity and spectrum of varied human-centred             this research has been supported by the German Research
challenges for the design and (usable) implementation of            Foundation (DFG – Deutsche Forschungsgemeinschaft) -
(explainable) artificial visual intelligence solutions in diverse   https://www.dfg.de.
human-system interaction contexts.
One of the key technical driving forces in our work is that of
“representation mediated multimodal sensemaking”.
                                                                    References
In essence, we consider (neurosymbolic) representation me-           [1] M. Bhatt, J. Suchan, Artificial visual intelligence:
diated grounding as being significant in semiotic construc-              Perceptual commonsense for human-centred cogni-
tion, e.g., enabling high-level meaning-making. This view                tive technologies, in: Human-Centered Artificial
stems from the long-established value of “grounding” in                  Intelligence: Advanced Lectures, Springer-Verlag,
Artificial Intelligence and related disciplines [25]. Our re-            Berlin, Heidelberg, 2023, p. 216–242. URL: https://
search advances the theoretical, methodological, and ap-                 doi.org/10.1007/978-3-031-24349-3_12. doi:10.1007/
plied understanding of “grounded representation” mediated                978-3-031-24349-3_12.
multimodal sensemaking of embodied human interaction
at the interface of spatial language, spatial logic, and spatial     [2] S. Jones, J. Laird, Anticipatory thinking in cogni-
cognition. In our view, the significance of this form of (neu-           tive architectures with event cognition mechanisms,
rosymbolic) grounding must now be reiterated, re-asserted                in: A. Amos-Binks, D. Dannenhauer, R. E. Cardona-
even, in view of recent advances in neural machine learning              Rivera, G. A. Brewer (Eds.), Short Paper Proc. of Work-
and the well-recognised “explainability” and “interpretabil-             shop on Cognitive Systems for Anticipatory Thinking
ity” requirements from the viewpoint of human-centred AI                 (COGSAT 2019), AAAI Fall Symp., volume 2558, 2019.
[26, 27, 28]. We believe that research in knowledge represen-            URL: http://ceur-ws.org/Vol-2558/short1.pdf.
tation and reasoning (KR) has, since its inception, concerned
                                                                     [3] J. Suchan, M. Bhatt, S. Varadarajan, Common-
itself with the “hard” problem of semantics, emphasising
                                                                         sense visual sensemaking for autonomous driv-
explainability, formal verification and diagnosis, elabora-
                                                                         ing - on generalised neurosymbolic online abduc-
tion tolerance amongst other things. Research in KR, and
                                                                         tion integrating vision and semantics, Artif. In-
more broadly in symbolic AI and semantics, and their role
                                                                         tell. 299 (2021) 103522. URL: https://doi.org/10.1016/
and contribution towards large-scale hybrid “human-in-the-
                                                                         j.artint.2021.103522. doi:10.1016/j.artint.2021.
loop” intelligence is of even greater significance now than
                                                                         103522.
ever before given the tremendous synergistic opportunities
afforded by the widely demonstrated power of deep learning           [4] J. Suchan, M. Bhatt, Semantic question-answering
driven techniques in computer vision (and beyond). The                   with video and eye-tracking data: AI foundations for
onus now, we posit, is on KR research to drive itself towards            human visual perception driven cognitive film stud-
developing methods that can seamlessly integrate (and be                 ies, in: S. Kambhampati (Ed.), Proceedings of the
“usable”) with other kinds of AI methods, be data-centric                Twenty-Fifth International Joint Conference on Ar-
     tificial Intelligence, IJCAI 2016, New York, NY, USA,      [15] J. Suchan, M. Bhatt, P. A. Walega, C. P. L. Schultz, Vi-
     9-15 July 2016, IJCAI/AAAI Press, 2016, pp. 2633–2639.          sual explanation by high-level abduction: On answer-
     URL: http://www.ijcai.org/Abstract/16/374.                      set programming driven reasoning about moving ob-
                                                                     jects, in: 32nd AAAI Conference on Artificial Intel-
 [5] J. Suchan, M. Bhatt, S. Vardarajan, S. A. Amirshahi,
                                                                     ligence (AAAI-18), USA, AAAI Press, 2018, pp. 1965–
     S. Yu, Semantic Analysis of (Reflectional) Visual Sym-
                                                                     1972.
     metry: A Human-Centred Computational Model for
     Declarative Explainability, Advances in Cognitive          [16] P. A. Walega, M. Bhatt, C. P. L. Schultz, ASPMT(QS):
     Systems 6 (2018) 65–84. URL: http://www.cogsys.org/             non-monotonic spatial reasoning with answer set pro-
     journal.                                                        gramming modulo theories, in: F. Calimeri, G. Ianni,
                                                                     M. Truszczynski (Eds.), Logic Programming and Non-
 [6] J. Suchan, M. Bhatt, The geometry of a scene: On deep
                                                                     monotonic Reasoning - 13th International Conference,
     semantics for visual perception driven cognitive film,
                                                                     LPNMR 2015, Lexington, KY, USA, September 27-30,
     studies, in: 2016 IEEE Winter Conference on Applica-
                                                                     2015. Proceedings, volume 9345 of Lecture Notes in
     tions of Computer Vision, WACV 2016, Lake Placid,
                                                                     Computer Science, Springer, 2015, pp. 488–501. URL:
     NY, USA, March 7-10, 2016, IEEE Computer Soci-
                                                                     https://doi.org/10.1007/978-3-319-23264-5_41. doi:10.
     ety, 2016, pp. 1–9. URL: https://doi.org/10.1109/WACV.
                                                                     1007/978-3-319-23264-5\_41.
     2016.7477712. doi:10.1109/WACV.2016.7477712.
                                                                [17] P. A. Walega, C. P. L. Schultz, M. Bhatt, Non-
 [7] J. Suchan, M. Bhatt, Deep Semantic Abstractions of
                                                                     monotonic spatial reasoning with answer
     Everyday Human Activities: On Commonsense Rep-
                                                                     set programming modulo theories,           Theory
     resentations of Human Interactions, in: ROBOT 2017:
                                                                     Pract. Log. Program. 17 (2017) 205–225. URL:
     Third Iberian Robotics Conference, Advances in Intel-
                                                                     https://doi.org/10.1017/S1471068416000193.
     ligent Systems and Computing 693, 2017.
                                                                     doi:10.1017/S1471068416000193.
 [8] M. Spranger, J. Suchan, M. Bhatt, Robust Natural Lan-
                                                                [18] J. Suchan, M. Bhatt, C. P. L. Schultz, Deeply semantic
     guage Processing - Combining Reasoning, Cognitive
                                                                     inductive spatio-temporal learning, in: J. Cussens,
     Semantics and Construction Grammar for Spatial Lan-
                                                                     A. Russo (Eds.), Proceedings of the 26th Interna-
     guage, in: IJCAI 2016: 25th International Joint Confer-
                                                                     tional Conference on Inductive Logic Programming
     ence on Artificial Intelligence, AAAI Press, 2016.
                                                                     (Short papers), London, UK, 2016, volume 1865, CEUR-
 [9] M. Bhatt, K. Kersting, Semantic interpretation of multi-        WS.org, 2016, pp. 73–80.
     modal human-behaviour data - making sense of events,
                                                                [19] K. S. R. Dubba, A. G. Cohn, D. C. Hogg, M. Bhatt,
     activities, processes, Künstliche Intell. 31 (2017) 317–
                                                                     F. Dylla, Learning Relational Event Models from
     320. URL: https://doi.org/10.1007/s13218-017-0511-y.
                                                                     Video, J. Artif. Intell. Res. (JAIR) 53 (2015) 41–
     doi:10.1007/S13218-017-0511-Y.
                                                                     90. URL: http://dx.doi.org/10.1613/jair.4395. doi:10.
[10] M. Bhatt, S. W. Loke, Modelling dynamic spatial                 1613/jair.4395.
     systems in the situation calculus, Spatial Cogni-
                                                                [20] J. Jaffar, M. J. Maher, Constraint logic programming:
     tion & Computation 8 (2008) 86–130. URL: https:
                                                                     A survey, The journal of logic programming 19 (1994)
     //doi.org/10.1080/13875860801926884. doi:10.1080/
                                                                     503–581.
     13875860801926884.
                                                                [21] S. Muggleton, L. D. Raedt, Inductive logic program-
[11] M. Bhatt, H. W. Guesgen, S. Wölfl, S. M. Hazarika,
                                                                     ming: Theory and methods, Journal of Logic Program-
     Qualitative spatial and temporal reasoning: Emerging
                                                                     ming 19 (1994) 629–679.
     applications, trends, and directions, Spatial Cognition
     & Computation 11 (2011) 1–14. URL: https://doi.org/10.     [22] G. Brewka, T. Eiter, M. Truszczyński, Answer set
     1080/13875868.2010.548568. doi:10.1080/13875868.                programming at a glance, Commun. ACM 54 (2011)
     2010.548568.                                                    92–103. doi:10.1145/2043174.2043195.
[12] M. Bhatt, Reasoning about space, actions and change:       [23] M. Bhatt, J. H. Lee, C. P. L. Schultz, CLP(QS): A declar-
     A paradigm for applications of spatial reasoning, in:           ative spatial reasoning framework, in: M. J. Egenhofer,
     Qualitative Spatial Representation and Reasoning:               N. A. Giudice, R. Moratz, M. F. Worboys (Eds.), Spa-
     Trends and Future Directions, IGI Global, USA, 2012.            tial Information Theory - 10th International Confer-
                                                                     ence, COSIT 2011, Belfast, ME, USA, September 12-16,
[13] J. Suchan, M. Bhatt, S. Varadarajan, Out of sight but
                                                                     2011. Proceedings, volume 6899 of Lecture Notes in
     not out of mind: An answer set programming based
                                                                     Computer Science, Springer, 2011, pp. 210–230. URL:
     online abduction framework for visual sensemaking in
                                                                     https://doi.org/10.1007/978-3-642-23196-4_12. doi:10.
     autonomous driving, in: S. Kraus (Ed.), Proc. of 25th
                                                                     1007/978-3-642-23196-4_12.
     Intnl. Joint Conference on Artificial Intelligence, IJ-
     CAI 2019, 2019, pp. 1879–1885. doi:10.24963/ijcai.         [24] C. P. L. Schultz, M. Bhatt, J. Suchan, P. A. Walega,
     2019/260.                                                       Answer Set Programming Modulo Space-Time, in:
                                                                     C. Benzmüller, F. Ricca, X. Parent, D. Roman (Eds.),
[14] R. Byrne, Counterfactual thought, Annual Re-
                                                                     Rules and Reasoning - Second International Joint Con-
     view of Psychology 67 (2016) 135–157. URL: https:
                                                                     ference, RuleML+RR 2018, Luxembourg, September 18-
     //doi.org/10.1146/annurev-psych-122414-033249.
                                                                     21, 2018, Proceedings, volume 11092 of Lecture Notes
     doi:10.1146/annurev-psych-122414-033249.
                                                                     in Computer Science, Springer, 2018, pp. 318–326. URL:
     pMID: 26393873.
                                                                     https://doi.org/10.1007/978-3-319-99906-7_24. doi:10.
      1007/978-3-319-99906-7\_24.                                         (Eds.), 2013 Workshop on Computational Models of
                                                                          Narrative, CMN 2013, August 4-6, 2013, Hamburg,
[25] S. Harnad, The symbol grounding problem, Physica
                                                                          Germany, volume 32 of OASIcs, Schloss Dagstuhl -
     D 42 (1990) 335–346.
                                                                          Leibniz-Zentrum für Informatik, 2013, pp. 24–29. URL:
[26] AI HLEG, High-level expert group on artificial intel-                https://doi.org/10.4230/OASIcs.CMN.2013.24. doi:10.
     ligence: Ethical guidelines for trustworthy ai, 2019.                4230/OASICS.CMN.2013.24.
     URL: https://www.aepd.es/sites/default/files/2019-12/
                                                                     [36] C. Lewis, Representation, Inclusion, and Innova-
     ai-ethics-guidelines.pdf.
                                                                          tion: Multidisciplinary Explorations, Synthesis Lec-
[27] EU Commission, Communication: Building trust in                      tures on Human-Centered Informatics, Morgan &
     human centric artificial intelligence, 2019.                         Claypool Publishers, 2017. URL: https://doi.org/10.
                                                                          2200/S00812ED1V01Y201710HCI038. doi:10.2200/
[28] EU Commission, Proposal for a regulation of the euro-                S00812ED1V01Y201710HCI038.
     pean parliament and of the council laying down har-
     monised rules on artificial intelligence (artificial intelli-
     gence act) and amending certain union legislative acts,
     2021. URL: https://eur-lex.europa.eu/legal-content/
     EN/TXT/?uri=CELEX:52021PC0206.
[29] J. Suchan, Declarative Reasoning about Space and Mo-
     tion in Visual Imagery - Theoretical Foundations and
     Applications, Ph.D. thesis, Universität Bremen, 2022.
     URL: https://elib.dlr.de/188919/.
[30] V. Kondyli, Behavioural Principles for the Design of
     Human-Centred Cognitive Technologies : The Case
     of Visuo-Locomotive Experience, Ph.D. thesis, Örebro
     University, School of Science and Technology, 2023.
[31] V. Nair, The Observer Lens: Characterizing Visuospa-
     tial Features in Multimodal Interactions, Ph.D. thesis, ,
     School of Informatics, Informatics Research Environ-
     ment, 2024.
[32] M. Bhatt, J. Suchan, Cognitive vision and percep-
     tion, in: G. D. Giacomo, A. Catalá, B. Dilkina, M. Mi-
     lano, S. Barro, A. Bugarín, J. Lang (Eds.), ECAI 2020 -
     24th European Conference on Artificial Intelligence,
     29 August-8 September 2020, Santiago de Compostela,
     Spain, August 29 - September 8, 2020 - Including 10th
     Conference on Prestigious Applications of Artificial
     Intelligence (PAIS 2020), volume 325 of Frontiers in Ar-
     tificial Intelligence and Applications, IOS Press, 2020, pp.
     2881–2882. URL: https://doi.org/10.3233/FAIA200434.
     doi:10.3233/FAIA200434.
[33] V. Kondyli, M. Bhatt, D. Levin, J. Suchan, How do
     drivers mitigate the effects of naturalistic visual com-
     plexity? on attentional strategies and their impli-
     cations under a change blindness protocol, Cogni-
     tive Research: Principles and Implications 8 (2023).
     doi:10.1186/s41235-023-00501-1.
[34] K. S. R. Dubba, M. Bhatt, F. Dylla, D. C. Hogg, A. G.
     Cohn, Interleaved inductive-abductive reasoning for
     learning complex event models, in: S. H. Muggle-
     ton, A. Tamaddoni-Nezhad, F. A. Lisi (Eds.), Inductive
     Logic Programming - 21st International Conference,
     ILP 2011, Windsor Great Park, UK, July 31 - August 3,
     2011, Revised Selected Papers, volume 7207 of Lecture
     Notes in Computer Science, Springer, 2011, pp. 113–
     129. URL: https://doi.org/10.1007/978-3-642-31951-8_
     14. doi:10.1007/978-3-642-31951-8\_14.
[35] M. Bhatt, J. Suchan, C. Schultz, Cognitive interpre-
     tation of everyday activities - toward perceptual nar-
     rative based visuo-spatial scene interpretation, in:
     M. A. Finlayson, B. Fisseni, B. Löwe, J. C. Meister