<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Exploratory Analysis of Users' Interactions with AR Data Visualisation in Industrial and Neutral Environments</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Franciszek Sobiech</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Natalia Walczak</string-name>
          <email>natalia.walczak@dokt.p.lodz.pl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aleksandra Buczek</string-name>
          <email>aleksandra0buczek@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mathias Jeanty</string-name>
          <email>mathias.jeanty@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kamil Kupiński</string-name>
          <email>kamil.kupinski@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zbigniew Chaniecki</string-name>
          <email>zbigniew.chaniecki@p.lodz.pl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrzej Romanowski</string-name>
          <email>andrzej.romanowski@p.lodz.pl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Krzysztof Grudzień</string-name>
          <email>krzysztof.grudzien@p.lodz.pl</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Arts et Métiers Institute of Technology</institution>
          ,
          <addr-line>151, Boulevard de l'Hôpital 75013 Paris</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lodz University of Technology</institution>
          ,
          <addr-line>116 Żeromskiego Street 90-924 Lodz</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Industrial and neutral environments bring multiple challenges in case of data analysis and cooperation while working with data visualization. This exploratory study aims to provide insights into user-3D data visualization interaction using Microsoft HoloLens 2 headset. The results were obtained in a 20-person study in a neutral environment - a classroom in a university building, and in a semi-industrial process tomography laboratory. Based on thematic analysis, this study presents initial guidelines for AR applications for working with 3D data visualizations.</p>
      </abstract>
      <kwd-group>
        <kwd>1 gesture interaction</kwd>
        <kwd>augmented reality headset</kwd>
        <kwd>3D data visualization</kwd>
        <kwd>process tomography</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Technological progress significantly influences the development of systems supporting access to
information and data visualisation. Augmented reality (AR) fulfills the function of supporting activities
performed physically by providing the user with a visual perception beyond the real world [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ]. In
terms of technical issues, including tools for AR visualisation, can be analysed systems where tools
need hands to use them and free-hands systems [
        <xref ref-type="bibr" rid="ref2 ref4">2, 4</xref>
        ]. This solution increases the possibility of the user
interacting with the AR world using hands based on generated AR data. In each of the applications
areas, hand gestures, as an integral part of human behaviour, can be used to control, edit, or interact.
The relatively short history of dealing with AR systems means that people do not know how to interact
with the AR world. Most AR applications are dedicated to information visualisation, which humans use
to react to current situations.
      </p>
      <p>
        Over time, awareness, and the need to use gestures in AR has evolved, and research has been
developed to show which gestures are intuitive, in what cases do gestures work, and what other solutions
should be considered [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Therefore, this work refers to analysing human gestures invoked during
interaction with 3D data in AR systems, where users directly modify the data. This would be especially
helpful for monitoring and maintaining the reach of ongoing processes in measurement data using
headmounted display (HMD) devices when hands are occupied. So, we propose an exploratory study to
investigate how users would interact with the measurement data of a 3D nature with the aid of HoloLens
HMD augmented reality device. We designed a user study aimed to investigate how users use gestures
in augmented reality during interaction with 3D process tomography data.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Study Scenario and Experimental description</title>
      <p>The dataset used in the prototype came from an electrical capacitance tomography (ECT) system.
This industrial data visualization technique allows one to take a picture of the industrial processes in
real time, which is essential for monitoring the process and understanding the phenomena inside an
industrial silo. In the presented work, 3D tomography data is a visualization of granular material
distribution in silo during process discharging. The observations and interviews were conducted during
20 experiment sessions, each with a different participant. They were not selected based on specific
characteristics or personality types and were mostly recruited from university students at Lodz
University of Technology. Data on gender, age, and prior experience with AR and ECT process
tomography was collected from each participant through a demographic questionnaire.</p>
      <p>
        Fifteen experiments were conducted in a neutral environment – a classroom in a university building.
In the remaining ones, participants would be observed using the AR system in the Process Tomography
Laboratory (TomoLab), a semi-industrial environment in a real silo used in the ECT process
tomography experiments [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Every session started with an introduction to the tomography system and
image interpretation; consisted of a general overview of what is electrical capacitance tomography,
where the data come from, how looks like the silo flow process, and what the 3D data visualization
looks like. Based on this knowledge, the participant had to complete the 4 tasks (example depicted on
Fig.1), without a time constraint and finished with a short semi-structured interview and a task-load
index (TLX) survey.
      </p>
    </sec>
    <sec id="sec-3">
      <title>2.1.Tasks Description &amp; Data Analysis Procedure</title>
      <p>Four tasks were planned for the study participants. Visual representation of the conception of
implementation of tasks is shown in Figure 1, and more detailed descriptions are given in Tab. 1. Study
participants were encouraged to speak out loud about what and how they wanted to perform, which
allowed researchers to help them and gather more data by making notes of their comments.</p>
      <p>The semi-structured interviews were recorded using a smartphone app and the observations were
noted using pen and paper. Apart from that, each session was recorded from the point of view of the
Microsoft HoloLens 2 headset using a built-in camera. The recordings were converted into transcripts
for easier analysis using an automatic audio file to text software. Data collected from the three sources
were then analysed using a thematic analysis approach. The first step was conducted by two researchers
and consisted of going through all the written data (transcripts of sessions and interviews) and notes
from the videos, highlighting similar phrases and ideas and giving names to those groups. This created
two sets of codes which were then compiled into one set and finally, with input from all of the
researchers, the codes were grouped into themes. The results of our analysis are presented in detail in
section 3. This approach was used to highlight the most important issues presented in the interviews
and actions of the participants and to draw conclusions directly from the user.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Results and discussion</title>
      <p>To better understand the study's results, it is important to know the participants' experience level
with using the Augmented Based on the collected surveys, the participant group is 70% occasional users
of virtual reality headsets, and 30% had never used such technology. Additionally, none of them had
any prior knowledge of process tomography. Most of them, 95%, are aged between 18 and 35 years
old, and most are students. After observations and interviews, we were able to categorize the data, using
the thematic analysis method and identified several distinct themes. The observations identified are
described below, categorized either by the individual or group theme.</p>
    </sec>
    <sec id="sec-5">
      <title>3.1. Strategies and attitudes connected to using gestures</title>
      <p>With n=20 participants overall there were numerous different attitudes towards using gestures and
generally using the app as well as strategies that the participant adopted in using it. Five participants
expressed that they felt a disconnect between their expectations and the way the app was working and
this led to discomfort and some frustrations. In some cases, users tried to use solutions from different
apps or devices, for example, a mobile and VR game “Fruit Ninja” or comparing the gestures to using
a touchpad on a computer.</p>
      <p>Most users tried the same gesture multiple times as the system feedback was not easily understood
by the user. Participants were not sure if their inconsistent results were caused by the problems of the
software and hardware, or they were not good at using the application. There were two competing
attitudes between participants to tasks involving moving and rotating the 3D cylinder. The first one was
treating the immaterial 3D object in AR as a physical object and trying to come closer and touch it with
open palms. This was especially noticeable when users tried different gestures for rotating. The second
approach was treating the cylinder similar to a 2D object on a touchscreen and utilized gestures similar
to those from a touchscreen. They would try to move by pointing at it or making a one-handed gesture.</p>
    </sec>
    <sec id="sec-6">
      <title>3.2. Physical &amp; Spatial Comfort and Intuitiveness of Gestures</title>
      <p>A crucial theme present both in the interviews and actions of the participants was the role of physical
and spatial comfort understood as providing a comfortable environment that would enable the user to
focus on the data.</p>
      <p>Users reported discomfort connected with wearing the headset. Even though this is something
HoloLens addresses with the option to change the tightness of the grip in the headset, this mild
frustration was present in the interviews after the session. Apart from this issue there were reports of
problems with field of view, which was not big enough for some participants. This required them to
move their head around and was especially problematic if the cylinder was placed high above the person
or was in the full-size version of the silo in TomoLab.</p>
      <p>Moreover, a topic found in the interviews and recordings from HoloLens was the importance of
having enough space around the user and the placement of 3D data relative to the environment and the
user. Some users would come close to the cylinder and move around it to look at all the sides while
some preferred to move the dataset closer to them. This second option was especially noticeable in the
sessions conducted at TomoLab, because of the equipment present around and sharp edges of some of
the silo parts.</p>
      <p>The cutting gesture that was used in the HoloLens application required the user to move their hand
through the augmented reality object. This proved difficult for some of the participants at first and even
more problematic during the research sessions at TomoLab because of the size of the silo. On the other
hand, some users instinctively wanted to come closer to the 3D object and touch it as if it was a physical
thing even in tasks that did not require it.</p>
      <p>One important aim of this study was to find out if users will find thinking of gestures to use, without
any prior instructions from the app or the researcher. Generally, this was not difficult for them, as they
reported thinking of gestures was “easy” and “natural”. Several participants compared the gestures,
both those that they thought of and those that were implemented in the app, to doing something in “real
life”. However, four users felt that finding the gesture set in the application would be impossible or near
impossible without the guidance of the researcher. Two other users also expressed the need for guidance
and were glad that they could redo the gesture after getting a hint.</p>
    </sec>
    <sec id="sec-7">
      <title>3.3. Problems: Software &amp; Hardware, Visual and Data Recognition</title>
      <p>The biggest sources of frustration for the users were the problems and limitations connected with
tested software and the hardware HoloLens HMD. The inconsistencies in gesture detection limited
some users from freely trying to complete the tasks. There were also problems with the visibility of
menus that were supposed to be anchored to the users’ hands. Besides the already mentioned comments,
an important issue was the field of view. The issues were not as often in the case of tests in a classroom
as in the TomoLab, because the cylinder was bigger, and the users wanted to see it all at once, which
was not possible in some cases. The importance of the system running smoothly is further underlined
by the fact that participants mention software and hardware problems even when speaking about
different aspects of the experience. In the first task (called Task 0 in the app) the user was asked to look
at the cylinder, with the option to walk around it to look at the other side and familiarize with the 3D
object. That is also when the researcher would ask the user about some of the information given during
the introduction to process tomography given before the tasks. Most of the answer about the overall
understanding of what is happening were correct, but there were some when it was not clear for the
user. The participants were also asked to comment on the colours of the visualization, with most of
them commenting on the “inverted” colour scheme – the red colour indicated a lower value. There were
also some hardware problems as it was hard to see some differences between colours because of the
semi-transparent look of the 3D object. In tasks that required cutting the object in half some of the users
expressed the desire to see the line where the cut will happen before they make the gesture.</p>
    </sec>
    <sec id="sec-8">
      <title>3.4.NASA TLX results</title>
      <p>Notably, on average the physical and temporal demand scores were not the highest. Furthermore,
performance was the highly self-esteemed and weighted category, which indicates that most users were
preoccupied with their performance during the tasks. The frustration is also highly weighted, which
indicates that for a lot of users the frustration was important, which supports the findings from the
thematic analysis.</p>
    </sec>
    <sec id="sec-9">
      <title>Conclusions and further work</title>
      <p>
        The scope of this study was to investigate the behaviour of a group of users while interacting
gesturally with a virtual 3D object containing data. We conducted a task-based study using an AR
prototype app on HoloLens HMD, in which specific, industrial tomography measurement data was
visualized as a 3D object. Several types of interaction as well as some problems, were revealed based
on qualitative data analysis. Even though this study used a specific configuration, some issues were
noticed or outliers among the data overall, the analysis of this data allowed us to see that the gestures
used to interact with the 3D object were naturally similar for most users and mostly corresponded to
the gestures available on the HoloLens, with some differences depending on the testing settings. To
verify the results, our application should be tested on a larger scale with more complex tasks, especially
based on temporal and spatio-temporal data [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and it would be interesting to employ eye tracking
to explore the design space for such complex AR interactions [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>The study aimed to identify convenient gestures to analyse and view data in an industrial
environment. Whereas there are already natural gestures used in augmented reality systems, it is crucial
to note the different approach when interacting with data in the AR space in comparison to traditional
interfaces. Based on the quantitative and qualitative data obtained during the survey of 20 people, it can
be concluded that users tend to use gestures resembling interaction with real objects. Furthermore, the
gestures they performed were similar but not identical. Future findings will be used for the development
of more complicated data analysis actions and process complex data in other branches.</p>
    </sec>
    <sec id="sec-10">
      <title>5. Acknowledgements</title>
      <p>This work was partially supported by the project “Multimodal system supporting remote
collaborative work in industrial settings” SKN/SP/535708/2022 funded by the Polish Ministry of
Education and Science within programme “Student Engineering Clubs Innovate”.</p>
    </sec>
    <sec id="sec-11">
      <title>6. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Nowak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Woźniak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pieprzowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Romanowski</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Mixed Reality in Action - Exploring Applications for Professional Practice</article-title>
          .
          <source>In proc. ISDA'18. Advances in Intelligent Systems and Computing</source>
          , vol
          <volume>941</volume>
          . pp.
          <fpage>1085</fpage>
          -
          <lpage>1095</lpage>
          , Springer, Cham. https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -16660-1_
          <fpage>106</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Aromaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Aaltonen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Kaasinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Elo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and I.</given-names>
            <surname>Parkkinen</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Use of wearable and augmented reality technologies in industrial maintenance work</article-title>
          .
          <source>In proc. AcademicMindtrek '16</source>
          . ACM, NY, USA,
          <fpage>235</fpage>
          -
          <lpage>242</lpage>
          . https://doi.org/10.1145/2994310.2994321 missing
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R. T.</given-names>
            <surname>Azuma</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>A survey of augmented reality</article-title>
          .
          <source>Presence: Teleoper. Virtual Environ</source>
          .
          <volume>6</volume>
          ,
          <issue>4</issue>
          (
          <year>August 1997</year>
          ),
          <fpage>355</fpage>
          -
          <lpage>385</lpage>
          . https://doi.org/10.1162/pres.
          <year>1997</year>
          .
          <volume>6</volume>
          .4.
          <fpage>355</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Feigl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Porada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Steiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Löffler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Mutschler</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Philippsen</surname>
          </string-name>
          ,
          <year>2020</year>
          . Localization Limitations of ARCore, ARKit, and
          <article-title>Hololens in Dynamic Large-scale Industry Environments</article-title>
          .
          <source>In Proceedings of the 15th International Joint Conference on Computer Vision</source>
          , Imaging and
          <source>Computer Graphics Theory and Applications - GRAPP</source>
          ,
          <fpage>307</fpage>
          -
          <lpage>318</lpage>
          . DOI:
          <volume>10</volume>
          .5220/0008989903070318
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Pham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vermeulen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L. M.</given-names>
            <surname>Vermeulen</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Scale Impacts Elicited Gestures for Manipulating Holograms: Implications for AR Gesture Design</article-title>
          .
          <source>In Proceedings of the 2018 Designing Interactive Systems Conference (DIS '18)</source>
          .
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA,
          <fpage>227</fpage>
          -
          <lpage>240</lpage>
          . https://doi.org/10.1145/3196709.3196719
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Romanowski</surname>
          </string-name>
          ,
          <article-title>"Big Data-Driven Contextual Processing Methods for Electrical Capacitance Tomography,"</article-title>
          <source>in IEEE Transactions on Industrial Informatics</source>
          , vol.
          <volume>15</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>1609</fpage>
          -
          <lpage>1618</lpage>
          ,
          <year>March 2019</year>
          , doi: 10.1109/TII.
          <year>2018</year>
          .
          <volume>2855200</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Romanowski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chaniecki</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Koralczyk</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Woźniak</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Nowak</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kucharski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; Jaworski,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Malaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Rózga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ;
            <surname>Grudzień</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <year>2020</year>
          .
          <article-title>Interactive Timeline Approach for Contextual Spatio-Temporal ECT Data Investigation</article-title>
          .
          <source>Sensors</source>
          <year>2020</year>
          ,
          <volume>20</volume>
          , 4793. https://doi.org/10.3390/s20174793
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Sulikowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; Zdziebko,
          <string-name>
            <surname>T.</surname>
          </string-name>
          ,
          <year>2020</year>
          ,
          <article-title>Deep Learning-Enhanced Framework for Performance Evaluation of a Recommending Interface with Varied Recommendation Position and Intensity Based on Eye-Tracking Equipment Data Processing</article-title>
          .
          <source>Electronics</source>
          <year>2020</year>
          ,
          <volume>9</volume>
          ,
          <fpage>266</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Hampel</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Babout</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Banasiak</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Schleicher</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Soleimani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wondrak</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vauhkonen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lähivaara</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hoyle</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Penn</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <article-title>A Review on Fast Tomographic Imaging Techniques and Their Potential Application in Industrial Process Control</article-title>
          .
          <source>Sensors</source>
          <year>2022</year>
          ,
          <volume>22</volume>
          , 2309. https://doi.org/10.3390/s22062309
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>