=Paper= {{Paper |id=Vol-2618/paper3 |storemode=property |title=Design and Implementation of Asynchronous Remote Support |pdfUrl=https://ceur-ws.org/Vol-2618/paper3.pdf |volume=Vol-2618 |authors=Irene Reisner-Kollmann,Andrea Aschauer }} ==Design and Implementation of Asynchronous Remote Support== https://ceur-ws.org/Vol-2618/paper3.pdf
           Design and Implementation of Asynchronous Remote Support
                                          Irene Reisner-Kollmann*                  Andrea Aschauer†

                                                   University of Applied Sciences Upper Austria



                                                                               We provide a system, where multiple users can access and edit
                                                                            spatial data at different times. We provide different interfaces for
                                                                            mobile users on-site and desktop users at a remote location. The
                                                                            system is mainly based on annotations on 2D snapshots (see Fig. 1),
                                                                            but other content types are supported as well. For mobile users, the
                                                                            content is spatially registered to 3D landmarks. For desktop users
                                                                            the content is attached to 2D images, as they do not have the full
                                                                            spatial information.
                                                                               We identified the following goals for our asynchronous remote
                                                                            support system:
                                                                               • Exchange information between local and remote locations

Figure 1: Screenshot of the mobile app, where the annotations are              • Create a documentation and exchange information over time
directly overlaid in instant mode.
                                                                               • Include existing data from other sources in order to be easily
                                                                                 accessed on-site
A BSTRACT                                                                      • Keep the mobile app simple by moving time-consuming tasks
We present an AR-based system that allows spatial collaboration for              to the desktop app
multiple users. The main contribution is that on-site and remote users      The main benefit of asynchronous collaboration is that not all partic-
can work together in an asynchronous way. On the one hand, mobile           ipants have to attend simultaneously.The participants can conduct
users can create a new project by documenting the environment               their work at their preferred time slot. They have the possibility
on-site and saving this data for asynchronous processing. On the            to carefully plan and review their input. Another practical prob-
other hand they can access previously saved data on-site with an            lem, which demands asynchronous methods, is the poor internet
augmented reality view where annotations and other content are              connection in some peripheral industrial sites.
directly visualized on top of the real world. Desktop users can                In this work, we leverage standard hardware devices and software
access the data remotely and have more sophisticated ways for               components for creating a sophisticated collaboration system. This
editing. Both, mobile and desktop users can revisit, view and edit          allows to quickly get advantage of the system as it does not require
a collaboration project anytime. We show how the system can be              any costly investments or disruptive organizational changes.
setup with standard hardware and software modules and share details
about the implementation.                                                   2 R ELATED WORK
Index Terms: Human-centered computing—Human computer                        Remote collaboration using a live video stream and augmented
interaction (HCI)—Interaction paradigms—Mixed / augmented                   reality has been an active research topic in recent years [1] and
reality; Human-centered computing—Collaborative and social                  has also led to commercial applications [8, 9]. The systems usually
computing—Collaborative and social computing systems and tools—             connect a user with a mobile device and a remote expert. We use a
Asynchronous editors;                                                       very similar concept and transfer it to an asynchronous paradigm.
                                                                               There are different variations of remote support systems [3, 6, 7].
1 I NTRODUCTION                                                             Annotations can be stabilized in world coordinates or fixed to the
For many services in industrial environments it is necessary to ac-         video frame. They can be persistent or vanish after a short time. The
cess a remote site multiple times and consolidate information from          expert’s view is freely selectable or it is bound to the current view
different people. For example, consider the following workflow for          of the local user.
the maintenance of a large machine at a customer site. First, a cus-           In one of the first papers, Gauglitz et al. [4] present a remote
tomer service representative gathers all information on-site including      collaboration system for mobile devices. They coarsely reconstruct
pictures, videos and notes from the customer. Based on this infor-          the 3D scene in the remote application for correctly placing 3D
mation, employees in the back office determine what exactly will be         annotations. The remote user can select any of the previous view
repaired and create an offer for the customer. Additionally, they add       points provided by the local user.
more information like manuals or specific instructions. Finally, a             A comprehensive review of collaboration techniques in aug-
technician executes the maintenance and retrieves all information           mented and virtual reality is given by Ens et al. [2]. They analyze
from the project. If the machine needs maintenance again at a later         mixed reality systems according to concepts of Computer Supported
time, it is beneficial to access the information again.                     Cooperative Work (CSCW) research. The vast majority of the ana-
                                                                            lyzed papers, namely 95 %, focus on synchronous collaboration.
   * e-mail: irene.reisner-kollmann@fh-steyr.com                               Irlitti et al. [5] describe theoretical considerations and challenges
   † e-mail:andrea.aschauer@fh-hagenberg.com                                for asynchronous collaboration using augmented reality. They sug-
                                                                            gest the importance of comprehensive information from users and
                                                                            for the spatial and temporal organization of the data. In contrast
Copyright © 2020 for this paper by its authors. Use permitted under         to our work, they mainly focus on local collaboration and not on
Creative Commons License Attribution 4.0 International (CC BY 4.0).         remote users.
Figure 2: Overview of collaboration system: Mobile and desktop users
share the same content on the server.
                                                                                 Figure 3: Screenshot of the mobile app in icon mode.

3     D ESIGN CONSIDERATIONS
Fig. 2 shows the structure of our asynchronous collaboration system.      Instant mode As shown in Fig. 1, the 2D annotations are directly
All users access the same data on the server. All data is organized in         overlaid to the live camera image by rendering them as quads
anchors, which contain a 3D pose, a 2D snapshot and the content                at the anchor’s pose in the three-dimensional space. Multiple
itself. Only mobile users can create new anchors, because the 3D               annotations are shown at the same time.
information from the augmented reality framework is needed. Delet-
ing anchors is only possible for desktop users, because we want to        Icon mode An icon is shown at the 3D location of the anchor. When
keep the mobile app as simple as possible.                                     clicking the icon, the original 2D snapshot and the annotations
                                                                               are shown full-screen. This mode is especially useful in case
3.1    Content types                                                           of poor registration (see Sect. 4.1), because direct overlays are
                                                                               not accurate enough. It also allows the user to see whether
The most important content are 2D annotations. Both, mobile and
                                                                               something has changed in the scene compared to a previous
desktop users can edit the annotations with a standard drawing tool
                                                                               session.
on top of a 2D snapshot of the scene. Annotations are stored as
separated images with a transparent background. Annotations allow         Other content types are always indicated with an icon in the 3D
to highlight specific objects in the scene or to scribble additional      scene and are opened by clicking on the icon. We use standard
instructions.                                                             software for showing videos and documents.
   A special form of the 2D annotations is text. The visualization is        The user can create new anchors by tapping one of the buttons in
similar to other 2D drawings, but it is important to store is as plain    the bottom (see Fig. 3) or directly onto the video stream. The 3D
text for editing and searching.                                           position of the anchor is computed by raycasting the reconstructed
   Audio content allows the mobile user to easily and quickly cap-        3D points and planes from the augmented reality framework. Op-
ture information. Video is useful to provide an overview of the           tionally, the user can create and edit 2D annotations to the current
whole scene or to include external data such as instructional videos.     snapshot with a simple drawing tool. Besides drawing lines with the
Documents allow to attach further information, e.g. the user guide        finger tip, the user can select a color and clear the whole screen.
for a specific machine. Documents are basically stored as links to           While our system is designed for connecting on-site and remote
external files, i.e. they can contain almost any type of data. The only   users, it is not mandatory to have a remote user. The mobile app
prerequisite is that a viewer is installed on all devices.                can be used on its own and allows users to view and edit the data
   The system can be easily extended to any other content type.           multiple times on-site.
We’d like to incorporate interactive elements in the future such as
animated instructions or checklists.                                      3.3 Desktop app
                                                                          The desktop app is aimed for remote users. They can review the
3.2    Mobile app
                                                                          data from the object or environment to work on, which has been
The mobile app is used for creating, viewing and editing content on-      provided by the mobile users. Further they can edit the data and
site. We currently support handheld devices, but other AR displays        provide additional information. Fig. 4 shows the main parts of the
such as head-mounted displays could be used as well. Fig. 3 shows         user interface for remote users.
the main user interface of the mobile app.                                   The desktop app visualizes the data only in 2D. Therefore, the
   We use the standard augmented reality view, i.e. the user sees the     2D snapshots are the main point of interaction. The user can flip
live video of the device’s camera together with the AR annotations.       through all 2D snapshots. Annotations are always overlaid if they
The information from the AR framework such as points and planes           are present. Other content types can be opened with buttons.
can be optionally visualized. This gives the user a better understand-       Users can edit the 2D annotations with a similar drawing tool as
ing whether the scene has been captured well enough. If it disturbs       in the mobile app. They can add any number of additional content
the user from the actual information, it can be simply hidden.            and assign it to one of the existing anchors. It is not possible to
   As noted before, our main content are 2D annotations. There are        create new anchors in the desktop app, because the 3D information
two modes for visualizing them in the mobile AR view:                     is not available.
                                                                         on wrong objects. In this case it is recommended to fallback to
                                                                         the icon mode and show annotations only on top of the original
                                                                         snapshot.
                                                                         4.2 Data storage
                                                                         An important request from industrial enterprises is that data can
                                                                         be easily shared with other software units. Therefore we decided
                                                                         to store all content in conventional data types. For example, all
                                                                         2D snapshots and annotations are stored as individual image files.
                                                                         The same is done for audio, video and documents. Additional data
                                                                         from our system is stored in json-files. This approach makes it easy
                                                                         to create importers for other software systems or to view the data
                                                                         directly with off-the-shelf software.
                                                                         5 C ONCLUSIONS AND F UTURE WORK
                                                                         We showed a system that allows on-site and remote users to exchange
                                                                         information on a specific site. Mobile users can see the information
                                                                         directly overlaid to the real object of interest in an augmented reality
                                                                         view. Remote users have a 2D view on the data and have more ways
                                                                         for adding additional data. We showed that the system can be simply
                                                                         implemented with standard hardware. We think that asynchronous
             Figure 4: Screenshot of the desktop app.                    remote collaboration has a big potential of improving collaboration
                                                                         on visual data over distance and time.
                                                                            In the future we plan a detailed evaluation of the system. We like
   Usually, the goal on-site is to capture as much information as        to compare it to traditional data sharing and to synchronous remote
possible because it’s often difficult to access it again. The desktop    support systems.
app offers various ways for organizing large amounts of data. There         An obvious extension will be the support of head mounted dis-
is a rating mechanism in order to optionally hide less important         plays. Another big improvement will be to make 3D data accessible
anchors. The remote user has also the possibility to delete specific     to the remote user. The remote user might get a better overview and
content or complete anchors.                                             has the possibility to create new anchors. An important topic will
                                                                         also be large scene changes between sessions and the handling of
4 I MPLEMENTATION DETAILS                                                incorrect scene registration.
The system is implemented in Unity 3D, which has the benefit that
many parts can be shared between the mobile and the desktop app.         ACKNOWLEDGMENTS
We use Unity’s integrated augmented reality framework ARFounda-          This work was funded by the FFG under the project Mixed Reality
tion. ARFoundation provides a unified API for ARCore and ARKit           Based Collaboration 4 Industry (Collective Research).
and thus supports both, Android and iOS. ARFoundation is also
necessary for using Microsoft’s Azure spatial anchors in Unity.          R EFERENCES
                                                                         [1] R. A. J. de Belen, H. Nguyen, D. Filonik, D. Del Favero, and T. Bednarz.
4.1 Registration                                                             A systematic review of the current state of collaborative mixed reality
An important factor for the AR application is to register the cur-           technologies: 2013–2018. AIMS Electronics and Electrical Engineering,
rent session to the existing spatial data. We provide three different        3(2):181–223, 2019.
approaches:                                                              [2] B. Ens, J. Lanir, A. Tang, S. Bateman, G. Lee, T. Piumsomboon, and
                                                                             M. Billinghurst. Revisiting collaboration through mixed reality: The
Cloud anchors Azure spatial anchors are created by capturing the             evolution of groupware. International Journal of Human-Computer
    surrounding scene and can be shared with other Android, iOS              Studies, 131:81–98, 2019.
    and Hololens devices. When the scene is viewed again, the            [3] O. Fakourfar, K. Ta, R. Tang, S. Bateman, and A. Tang. Stabilized
                                                                             annotations for mobile remote assistance. In Proceedings of the 2016
    pose of the anchor is re-established without user interaction.
                                                                             CHI Conference on Human Factors in Computing Systems, pp. 1548–
    This approach fails if there were too many changes in the scene.
                                                                             1560, 2016.
    In our system, we use one Azure spatial anchor per project.          [4] S. Gauglitz, C. Lee, M. Turk, and T. Höllerer. Integrating the physical
                                                                             environment into mobile remote collaboration. In Proceedings of the
Marker Aligning the scene to a known marker is supported by
                                                                             14th international conference on Human-computer interaction with
    all established AR frameworks. Usually, it is not possible to
                                                                             mobile devices and services, pp. 241–250, 2012.
    permanently mount markers at customers site. In this case, the       [5] A. Irlitti, R. T. Smith, S. Von Itzstein, M. Billinghurst, and B. H. Thomas.
    marker location can be inferred from screenshots of previous             Challenges for asynchronous collaboration in augmented reality. In
    sessions. The same printed marker can be used for each support           2016 IEEE International Symposium on Mixed and Augmented Reality
    case and the user positions it again in the scene according to           (ISMAR-Adjunct), pp. 31–35. IEEE, 2016.
    the screenshot.                                                      [6] A. Jakl, L. Schöffer, M. Husinsky, and M. Wagner. Augmented reality
                                                                             for industry 4.0: Architecture and user experience. In FMT, pp. 38–42,
Manual registration The user selects a screenshot from a previous            2018.
   session and manually overlays the current camera view. This           [7] S. Kim, M. Billinghurst, and G. Lee. The effect of collaboration styles
   procedure is clearly very inaccurate. We included it as a fall-           and view independence on video-mediated remote collaboration. Com-
   back when there is no internet connection and markers have                puter Supported Cooperative Work (CSCW), 27(3-6):569–607, 208.
   not been used in previous sessions.                                   [8] Microsoft. Remote Assist. https://dynamics.microsoft.com/de-de/mixed-
                                                                             reality/remote-assist/.
None of these registration methods is absolutely free from defects. If   [9] PTC. Vuforia Chalk. https://chalk.vuforia.com/.
the registration is not good enough, the annotations will be overlaid