=Paper=
{{Paper
|id=Vol-2779/paper5
|storemode=property
|title=Individual and Collaborative Cross-reality Immersive Analytics - Initial Ideas
|pdfUrl=https://ceur-ws.org/Vol-2779/paper5.pdf
|volume=Vol-2779
|authors=Nanjia Wang,Amir Aminbeidokhti,Frank Maurer
|dblpUrl=https://dblp.org/rec/conf/tabletop/WangAM20
}}
==Individual and Collaborative Cross-reality Immersive Analytics - Initial Ideas==
Individual and Collaborative Cross-reality Immersive Analytics – Initial Ideas Nanjia Wang Amir Aminbeidokhti Frank Maurer nanjia.wang1@ucalgary.ca amir.aminb97@gmail.com frank.maurer@ucalgary.ca University of Calgary University of Calgary University of Calgary Calgary, Alberta Calgary, Alberta Calgary, Alberta ABSTRACT This paper presents research questions on supporting analytics work using cross reality (CR) and the initial idea of the investigation we want to conduct. KEYWORDS cross reality, augmented reality, virtual reality, immersive analytics, data visualization Figure 1: Diagram showing a representation of Milgram et Reference Format: al. Reality Virtuality Continuum. Nanjia Wang, Amir Aminbeidokhti, and Frank Maurer. 2020. Individual and Collaborative Cross-reality Immersive Analytics – Initial Ideas. In Cross-Reality (XR) Interaction, ACM ISS 2020 (International Workshop on XR Interaction 2020). CR is a f eld that aims to create interactive environments that allow users to seamlessly interact with information across the virtuality spectrum. Data can be placed on traditional displays, in a physical space using AR tools or in a virtual space using VR approaches. Dierent users can chose in which space they want to utilize to engage with their information. Collaborative CR applications allow dierent users to use dierent interaction modalities. Our work aims to develop the processes and guidelines to create CR applica- tions oer users a beer understanding of complex datasets across immersive environments. 1 INTRODUCTION Figure 2: Screenshot of immersive analytics data visualiza- CR is a eld that aims to create interactive environments that allow tions from the SIERA project [2] users to bring virtual content into the physical or bring physical world objects into the virtual world to help users to have a better perception of data and information. Our work aims to develop the Mounted Displays (HMDs).[5] The term "cross-reality" (CR) is rela- CR application to oer users a beer understanding of complex tively new and describes a new kind of software systems that bridge datasets in the immersive environment. between two or more points on the Milgram spectrum(Figure 1). Such applications allow their users to move data or visualizations The term mixed reality has been around for decades , but real- from a PC to a physical space and/or a virtual space and back. world deployment has recently exploded due the availability of In our studies, we want to explore how to support analytics work commodity hardware. Hardware is available for applications tar- using CR from both individual and collaborative perspective.The geting the whole Milgram reality-virtuality spectrum. Aordable SIERA project done by Lawton et al. is an example of using the vir- hardware and software enable researchers and developers to create tual reality application for understanding complex datasets(Figure augmented-reality (AR), extended reality (XR), mixed-reality (MR) 2).[2] We consider the following Research Questions (RQs) as and virtual-reality (VR) applications on smartphones, tablet or Head objectives of our studies: • RQ1: How can we support analytics work by moving Copyright © 2020 for this paper by its authors. Use permied under Creative visualizations across the Milgram continuum ? Commons License Aribution 4.0 International (CC BY 4.0). Cross-Reality (XR) Interaction, ACM ISS 2020, November 8 2020, Lisbon, Portugal Data analysis is the process of discovering useful informa- tion by cleaning, inspecting, transforming, and modeling data.[6] Data visualization helps data analysis and interpre- tation. In a CR environment, information can be presented at any point of the Milgram spectrum and the application or International Workshop on XR Interaction 2020, November 8 2020, Lisbon, Portugal Wang, et al. the user can determine where it should be presented to �t are more useful for lateral and altitude �ight path tracking. In the their needs.[4] In our study, we want to explore information meantime, 2D displays have an advantage in the accurate mea- visualization across the continuum to determine what kind surement of airspeed.[1] We will conduct case studies to explore of information is best presented at speci�c levels of virtuality what visualizations should be moved to 3D AR space to support to provide better support to analytics work. individual and team collaboration. We will also investigate how user want to interact with CR applications. • RQ2: What interactions do users naturally use for cross- reality applications? With the support of XR, users could interact with data in 2.2 Impact of mixing AR and VR collaborators 3D space without the limitation of the 2D screen, keyboard, We want to explore mixing analytic data collaborators with dif- and mouse. The XR environment that will be created for ferent systems work in the same extended reality space. We will our study will be a mix of the physical world and the virtual investigate the potential impact of collaborators using devices with world. Our study will explore under what circumstance users di�erent degrees of virtuality across Milgram’s continuum. We are want to move visualization from monitors in the physical particularly interested in studying how collaboration works out world to virtual space and how they will do it. We also want when some participants are using AR while others are using VR to study how users want to interact with visualization in spaces. XR space naturally when they have more options, including hand gestures and eye gazing. 2.3 Natural gestures to support interactions • RQ3: How does control over a shared reality impact We will conduct an elicitation study to investigate how individuals team collaboration? and teams want to interact with the CR system with natural gestures Collaboration in the physical world can make reasonable intuitively. He hope that a common set of gestures and interactions assumptions about a shared reality: when two people work will materialize so that toolkits can be developed to support them. together in a room, they have an intuitive understanding of what the other person can see and not see. They also know that if they both look at an object in the room, they 2.4 Engineering guidelines both see the same object. This creates an intuitive shared In the long run, we would like to develop engineering guidelines to understanding of the world around them. CR applications provide evidence-based advice to developers on how to design CR can break this intuition: the CR system can show one object applications and which kind of data should be presented on speci�c to one person, and a di�erent object - in the same place - to places of the virtuality continuum. the other person. This "unsharing" of real-world intuition can be accidental (e.g. incorrect calibration resulting in a shift of an object), on purpose (e.g. orienting text towards 2.5 "Unshared" experiences in cross-reality people on opposite sides or showing additional information visualizations to only a domain expert) or even malicious (e.g. a hacker As CR systems can overcome physical limitations, we are interested creating a misunderstanding). We want to investigate how to investigate if creating realted but di�erent realities can support to balance between shared and unshared reality to enhance collaboration of hinder them. We will conduct studies to explore team collaboration e�ciency. how to present user-speci�c perspectives on shared information while users are in di�erent locations in a virtual space. Questions include: How far can a virtual representation deviate from intuitive 2 INITIAL IDEAS TO INVESTIGATE physical norms without hindering collaboration? Can we enhance CR is a new research area with plenty of unknowns, creating im- collaboration by "unsharing" visualizations of information? mense research opportunities ranging from questions related to cognition all the way to engineering processes and development ACKNOWLEDGMENTS guidelines. In this spectrum, we will be focusing on a small number The authors would like of thank Bryson Lawton and the ASE team of speci�c research topics. for providing a screenshot of SIERA, a Seismic Information Ex- tended Reality Analytics tool. We are grateful the support of NSERC, 2.1 Move information from a 2d screen to a 3d the Natural Sciences and Engineering Research Council of Canada, space and back for our work. To better support analytic work, we want to investigate applying CR to increase productivity and enhance sense-making and com- REFERENCES [1] Ian D. Haskell and Christopher D. Wickens. 1993. Two- and Three-Dimensional munication. A previous study shows that work stations with more Displays for Aviation: A Theoretical and Empirical Comparison. The Interna- monitors could potentially increase productivity.[3] We want to tional Journal of Aviation Psychology 3, 2 (1993), 87–109. https://doi.org/10.1207/ investigate how to move information from a 2D monitor to a 3D s15327108ijap0302_1 [2] Bryson Lawton, Hannah Sloan Sloan, Patrick Abou Gharib, Frank Maurer, Marcelo space and back in order to overcome the limitation of the number Guarido De Andrade, Ali Fathalian, and Daniel Trad. 2020. SIERA: The Seismic of the physical display. Haskell and Wickens show that 3D displays Information Extended Reality Analytics Tool. In In Companion Proceedings of the Individual and Collaborative Cross-reality Immersive Analytics – Initial Ideas International Workshop on XR Interaction 2020, November 8 2020, Lisbon, Portugal 2020 Conference on Interactive Surfaces and Spaces (ISS ’20 Companion). Lisbon, [4] Paul Milgram, Haruo Takemura, Akira Utsumi, and Fumio Kishino. 1994. Aug- Portugal. https://doi.org/10.1145/3380867.3426223 mented reality: A class of displays on the reality-virtuality continuum. Telemanip- [3] Chen Ling, Alex Stegman, Chintan Barhbaya, and Randa Shehab. 2016. Are Two ulator and Telepresence Technologies 2351 (January 1994). https://doi.org/10.1117/ Better Than One? A Comparison Between Single- and Dual-Monitor Work Stations 12.197321 in Productivity and User’s Windows Management Style. International Journal [5] Steven Vi. 2020. Exploring User Experience Guidelines for Designing HMD Extended of Human-Computer Interaction 33 (September 2016). https://doi.org/10.1080/ Reality Applications. Master’s thesis. University of Calgary, Calgary, Canada. 10447318.2016.1231392 [6] Belle Xia and Peng Gong. 2014. Review of business intelligence through data analysis. Benchmarking: An International Journal 21 (April 2014), 300–311. https: //doi.org/10.1108/BIJ-08-2012-0050