=Paper=
{{Paper
|id=Vol-91/paper-3
|storemode=property
|title=Abstract and Concrete Interaction with Mixed Reality Systems / The case of the mini screen, a new interaction device in Computer-Assisted Surgery
|pdfUrl=https://ceur-ws.org/Vol-91/paperD3.pdf
|volume=Vol-91
|dblpUrl=https://dblp.org/rec/conf/mixer/NigayMT04
}}
==Abstract and Concrete Interaction with Mixed Reality Systems / The case of the mini screen, a new interaction device in Computer-Assisted Surgery==
Abstract and Concrete Interaction
with Mixed Reality Systems
The case of the mini screen, a new interaction device
in Computer-Assisted Surgery
Benoit Mansoux, Laurence Nigay Jocelyne Troccaz
Laboratoire CLIPS-IMAG Laboratoire TIMC-IMAG
BP 53 I. I. I. S. -Faculté de Médecine
38041 Grenoble cedex 9 38706 La Tronche cedex
{mansoux, nigay}@imag.fr jocelyne.troccaz@imag.fr
ABSTRACT INTRODUCTION
In this paper we focus on the design of mixed reality (MR) In this paper, we focus on the design of Mixed Reality
systems. We propose two design spaces that can be useful (MR) systems. We focus on MR systems that assist a user
in a top-down (abstract to concrete) design method for MR in performing a task on a physical object (a class of MR
systems. The first design space consists of an organized systems called “Augmented Reality” in [2]). One of our
framework of abstract interaction situations for describing main application domains for such MR systems is
mixed systems. Each situation is depicted by an ASUR Computer Assisted Surgery (CAS), in the context of a
diagram and describes the exchange of information between multidisciplinary project that involves the HCI and the
the entities involved in a mixed system. The situations are CAS research groups of the University of Grenoble. The
abstract because they are independent of the interaction main objective of CAS systems is to help a surgeon in
modalities (both interaction languages and devices). The defining and executing an optimal surgical strategy based
abstract interaction situations are illustrated by several on a variety of multi-modal data inputs. MR systems play
Computer-Assisted Surgery (CAS) systems. Such a a central role in the CAS domain because the key point of a
framework is useful for the designer in order to CAS system is to "augment" the physical world of the
systematically explore the set of possibilities at an early surgeon: the operating theater, the patient, the surgical
stage of the interaction design, without being biased by a tools etc., by providing pre-operative information during
particular technology. With the interaction situation the surgery. MR systems are now entering many surgical
described, the designer can then focus on the modalities to specialties and such systems can take on the most varied
be used: both passive and active modalities can be elected. forms. Although many CAS systems have been developed
This design stage consists of concretizing the interaction and provide real clinical improvements, their design is ad-
situation by selecting the modalities. For this stage of the hoc and principally driven by technologies.
design, we propose a design space that characterizes the In this context, our research aims at providing elements
possible usages of one particular innovative interaction useful for the design of usable MR systems by focusing on
device for CAS systems: a mini screen. We illustrate the the interaction between the user and the MR system. We
complementarity of our two design spaces by presenting present two design spaces that can be useful in a top-down
two CAS systems that embed a mini screen for different design method for MR systems. The first design space,
purposes in the interaction: one system is based on a presented in the second section of the paper, consists of an
localized mini screen fixed on the surgical tool while the organized framework of abstract interaction situations for
other involves the surgeon handling the mini screen on top describing MR systems. This first result is useful at an
of the patient’s body. early stage of the design of MR systems: indeed it enables
Keywords the designer to systematically explore the set of
Mixed Reality, Computer Assisted Surgery, Design Space, possibilities without being biased by the available
Interaction Device, Mini screen technologies. While this first design space focuses on
abstract interaction (i.e., independent of the interaction
technologies), our second design space, presented in the
third section of the paper, characterizes the possible usages
of one particular interaction device, a mini screen. Our two
design spaces are therefore complementary and address
different stages of a top-down design method of MR
systems: abstract versus concrete interaction. Before
presenting our two design spaces, we first clarify the two
interaction design steps, i.e. the design of the abstract and
concrete interaction.
ABSTRACT AND CONCRETE INTERACTION Entities (R and A) Relationships
We call interaction situation, an abstract description of the characteristics characteristics
interaction involved in an MR system. Such a description
- Perceptual/Action sense - Interaction language
is independent of the interaction modalities. We define in
and location
[7] a modality as the coupling of a physical device with an
interaction language. After describing the interaction
Table 1: Characteristics of ASUR entities and
situation, the following step in the design consists of
relationships.
concretizing the abstract situation by choosing the
modalities: the description of the interaction is then Two levels of abstraction in describing
concrete. For describing the abstract and concrete interaction using ASUR
interaction, we use the ASUR notation [2][3]. In the In [3], we explained how we use the ASUR notation during
following paragraph, we summarize the main characteristics the requirements definition phase for describing usage
of the notation. We then describe how to use the ASUR scenarios and during the external specification phase for
notation for describing the abstract and concrete interaction. describing the concrete designed interaction. Going one
ASUR notation step further, we define here two levels of abstraction in
ASUR [2][3] stands for "Adapter", "System", "User", "Real describing the interaction in an MR system, as part of a
objects". In user-centered MR systems are described in top-down (abstract to concrete) method for designing the
terms of entities (A, S, U, R) taking part in the interaction interaction. Interaction situations are described using
and the relations between those entities. Between the user ASUR at the most abstract level. Nevertheless, for
(U) and the computer system (S), the adapters bridge the analytical reasons, we describe the two levels of interaction
gap between the physical world and the digital one. They description in reverse order, from the concrete one to the
could be input adapters (Ain) (e.g., a mouse, a localization abstract one.
mechanism) or output ones (Aout) (e.g., a video projector, (1) The most concrete description is the final stage of the
audio speakers). Physicality is one key feature of MR external specification phase. Interaction is fully depicted by
systems: real objects are involved in the task. Within the a set of ASUR entities and relations that are described by
ASUR notation we distinguish physical objects that are the ASUR characteristics. The interaction modalities
tools (Rtool) for performing the task, from the ones that are (devices and languages) are therefore chosen: we distinguish
the objects of the task (Robject). two types of modalities in an MR system, the active and
Three kinds of relationship between two ASUR entities are passive modalities. Active and passive modalities are
identified: defined for the MR systems we are concerned with in this
paper: the object/target of the main task is physical, for
• Exchange of data is represented by an arrowed line example the patient for CAS systems.
between two ASUR entities (AÆB).
• For inputs, active modalities are used by the user to
• Physical activity triggering an action: a double-line issue a command to the computer such as a pedal to
arrow (AfiB) denotes the fact that when the entity A move a laparoscope in a CAS system. Passive modalities
meets a given spatial constraint with respect to entity B, are used to capture relevant information for enhancing the
data will be exchanged along another specified realization of the task, information that is not explicitly
relationship (CÆD). expressed by the user to the computer ("perceptual user
• Physical collocation is represented by a non-directed interfaces" [9]). For example, in our CASPER (Computer
double line (A=B). This refers to a persistent physical ASsisted PERicardial puncture) system, presented in
proximity of two entities. Figure 1, a system for computer assistance in pericardial
Finally, the ASUR entities and relationships are described punctures, a passive modality is used for tracking the
by a set of characteristics. Table 1 presents some of them. position of the puncture needle.
For example the first characteristic induced by the use of a • For outputs, active modalities, conveying information
real object (R) or an adapter (A) is the human sense from the computer to the user, imply that the user
involved in perceiving data from such an entity or in explicitly switches attention from her/his current task
performing actions using such entity. The most common focus to a new focus in order to perceive the provided
used ones are the haptic, visual and auditory senses. A information. For example in our CASPER system,
second characteristic is the location where the user has to visual guidance information during the puncture task is
focus with the required sense, in order to displayed on a screen. While using CASPER (Figure 1),
perceive/manipulate the real entity as well as to manipulate the surgeon must always shift between looking at the
the adapter or perceive the data provided by it. In addition screen and looking at the patient and the needle (i.e., the
one characteristic of a relation between two ASUR entities task environment). As opposed to active modalities,
is the interaction language used to express data carried by passive output modalities convey information to the user
the relation. If we refer to our definition of a modality [7] that is integrated in her/his task environment, for
as the coupling of a physical device with an interaction example displaying anatomical information onto the
language, the device is described by an ASUR entity while patient’s body during a surgery. For the case of passive
the interaction language is a characteristic of the relation output modalities, the user does not have to switch
from this entity (device) to another ASUR entity.
attention from her/his current task focus in order to is represented by two mobile crosses, while one stationary
perceive the provided information. cross represents the planned trajectory. A complete
description of the concrete interaction in ASUR can be
found in [2].
(2) A more abstract level of description of interaction
consists of focusing on the exchange of information
between the involved entities during interaction. By doing
so, we describe what we call the interaction situation.
Interaction modalities are not yet chosen but the elementary
tasks are identified. The role of the adapters are therefore
defined (for example, a localization mechanism, a data
presenter) but the concrete adapters (physical devices) as
well as the forms of the data conveyed along the relation
are not yet defined. In addition the physical setting is not
yet defined, physical relationships between entities are not
decided. In conclusion, such level of description consists of
an ASUR diagram:
Fig. 1: CASPER in use during the intervention.
• without characterization of the entities and relations.
In Figure 2, we illustrate this level of interaction
description, by presenting the ASUR diagram of the • with one kind of relation: exchange of data (AÆB).
CASPER system. During the surgery, CASPER assists the Figure 3 illustrates this level of description using our
surgeon (U) by providing in real time the position of the CASPER system. Figure 3 is therefore a more abstract
puncture needle (Rtool) according to the planned trajectory. description of the interaction described in Figure 2.
Two adapters (Ain, Aout) are necessary: The first one (Aout) is
the screen for displaying guidance to the surgeon, and the
second one (Ain) is dedicated to tracking the needle position
S
and orientation as well as the patient’s body (Robject). The
localization of the needle is possible within a predefined
volume near the patient’s body. Such a constraint is
represented in Figure 2 by an ASUR relation fi (physical A1in A2in Aout U
activity triggering an action).
S
Rtool
Ain Aout
U
Robject
Robject : Patient Rtool Robject : Patient
Rtool : Puncture needle Rtool : Puncture needle
U : Surgeon U : Surgeon
Ain : Cameras+diodes A1in : Localizer
Aout : Screen Robject A2in : Localizer
S : Computer System Aout : Data presenter
S : Computer System
Fig. 2: ASUR diagram of the concrete interaction in
CASPER. For a complete ASUR description, the Fig. 3: ASUR description of the abstract interaction in
diagram is completed by the characteristics of each CASPER. As opposed to Figure 2, the interaction
entity and relation (see [2]). modalities as well as the physical relationships are not
yet defined at this stage of the design.
The concrete interaction description of Figure 2 is not
complete. The ASUR diagram is completed by the In a top-down (abstract to concrete) design method, the
characteristics of the identified entities and relations. For designer first focuses on the interaction situation (i.e.,
example the interaction language (one of the characteristics) abstract description of the interaction) and will then select
used to convey the guidance information on screen (Aout) the modalities for concretizing the interaction. Our first
must be described: Using CASPER, in the same window design space identifying a set of interaction situations is
on screen, the current position and orientation of the needle therefore useful at an early stage of the interaction design
for reasoning on the interaction without being biased by the For these two situations that involve passive modalities,
interaction technologies. Our second design space we suggest that the user and the object of the task are
characterizes the possible usages of one particular physically together. In the case of telesurgery for example,
innovative interaction device (output adapter) for CAS the surgeon (user) and the patient (object of the task) are
systems: a mini screen. This second design space is distant. Such situations are described using ASUR by
therefore useful for designing concrete interactions adding an ASUR chain that comprises the computer system
involving a mini screen. (S) between:
INTERACTION SITUATION DESIGN SPACE • the user (U) and the tool ([Rtool, R object]) for Class III-
Our design space is made of interaction situations that are input,
independent of the interaction modalities. A situation is • the user (U) and the object of the task (Robject) for Class
dedicated to a particular task. For example, in Figure 3, the IV-input.
diagram depicts the interaction situation for the task of
The ASUR chain to be added is either:
pericardial puncture while using CASPER. A situation
describes both the abstract input and output interaction. (a) (AinÆSÆAout)
Our framework is composed of input and output situations. (b) (RtoolÆAinÆSÆAout)
Our approach for establishing the framework of interaction
situations draws from our distinction of active and passive The two ASUR chains differ by the way the user interacts
modalities. with the computer system (S). The two chains (a) and (b)
respectively correspond to Class I-input and Class II-input.
Input interaction situations
We therefore obtain four classes:
For inputs (user to computer), we identify four situations,
two of them involve active modalities while the other two Class III-input-a (U and Robject distant):
involve passive modalities. UÆ(AinÆSÆAout)Æ[Rtool, Robject]ÆAinÆS
(1) The two situations, Class I-input and Class II-input, Class III-input-b (U and Robject distant):
involve active modalities. In these situations, the user
UÆ(RtoolÆAinÆSÆAout)Æ[Rtool, Robject]ÆAinÆS
explicitly issues a command to the computer system. The
user must switch attention from the task’s focus (Robject) to Class IV-input-a (U and Robject distant):
a new focus in order to interact with the computer. As a UÆ(AinÆSÆAout)ÆRobjectÆAinÆS
consequence, in the ASUR diagram that depicts these two
situations, there is no R object involved. Without R object, the Class IV-input-b (U and Robject distant):
two remaining possibilities are: UÆ(RtoolÆAinÆSÆAout)ÆRobjectÆAinÆS
Class I-input: UÆAinÆS For example, the input interaction situation of the
Class II-input: UÆRtoolÆAinÆS telesurgery system described in [5] belongs to Class III-
input-b: The surgeon (U) remotely controls a slave robot
The first situation (Class I-input) depicts a classical (Aout), that holds the surgical tools (AoutÆ[Rtool, R object]),
interaction with a computer, for example using a mouse. by manipulating force-feedback arm-mounted tools
The second situation (Class II-input) describes the case (UÆRtoolÆAin).
where the user manipulates a physical object (Rtool) to
interact with the computer via an adapter that captures the Output interaction situations
manipulations. Examples of such input situations are the For outputs (computer to user), we identify four situations,
physical icons that are physical handles to digital objects, two involving active modalities and two involving passive
“coupling the bits with everyday physical objects and ones. This is the symmetric case of input situations.
architectural surfaces” [6]. Class I-output and Class II-output correspond to situations
(2) We identify two situations that involve passive involving active modalities. The user must switch attention
modalities. The user is performing a task in the physical (explicit action of the user) from her/his current task focus
world on an R object while the computer captures relevant (Robject) to a new focus in order to perceive the provided
information for enhancing the realization of the task, thanks information carried by the active modalities. The ASUR
to passive modalities. Two situations are possible whether diagrams of these two situations therefore do not comprise
the user manipulates R object using a tool ([Rtool, R object]) or an entity Robject.
directly manipulates Robject. Class I-output: SÆAoutÆU
Class III-input: UÆ[Rtool, Robject]ÆAinÆS Class II-output: SÆAoutÆRtoolÆU
Class IV-input: UÆRobjectÆAinÆS A Class I-output example is the CASPER output situation
A Class III-input example is the CASPER input situation described in Figure 3: During the puncture task, the
described in Figure 3: During the puncture task, the surgeon perceives guidance information displayed on a
surgeon is handling the puncture needle (Rtool) that touches screen. An example of Class II-output situation would
the patient body ([Rtool, R object]). Both the needle and the correspond to a CAS system that displays information on
patient are localized by the system via adapters. the wall of the operating theater: Although a surface of the
physical environment is used for displaying information
(Rtool), it implies that the surgeon consciously switch possibilities at an early stage of the interaction design,
attention from the environment of the task (the operating without being biased by a particular technology. With the
field) to the wall in order to perceive the information. interaction situation described, the designer can then focus
(2) As for inputs, two output situations involve passive on the modalities (device and language) that are passive or
modalities. These situations describe the cases where the active according to the situation, as well as on the physical
user is perceiving the information provided by the system setting (physical relations described in ASUR). From an
within her/his task environment (Robject). The ASUR abstract interaction situation, several concrete interaction
diagrams that describe these two situations therefore solutions can be designed. In the following paragraph, we
involve an Robject. focus on concrete interaction involving a particular device:
a mini screen.
Class III-output: SÆAoutÆ[Rtool, Robject]ÆU
CONCRETE INTERACTION INVOLVING A MINI
Class IV-output: SÆAoutÆRobjectÆU SCREEN
The output situation using the PADyC (Passive Arm with The transition from interaction situation to concrete
Dynamic Constraints) system [8] belongs to Class-III- interaction is difficult because the set of possibilities in
output. Indeed using PADyC, the surgeon is handling a terms of modalities (device and language) is huge. As a
surgical tool that is linked to a passive arm (Aout). The first step for accompanying this transition, we propose a
programmable arm enables us to provide haptic guidance design space that describes the possible modalities that
information (touch feedback) to the surgeon while involve a mini screen.
performing the surgery. Another output situation of this Small devices are increasingly being used in MR systems
class that involves a mini screen will be described in the as in [10], and offer new interaction techniques, like the
last section of the paper. Embodied User Interfaces defined in [4]. For CAS systems,
A Class IV-output example is the situation using the a small screen is an innovative device.
second version of CASPER [2] that involves a see-through Beyond standard technical features of an LCD screen like
head-mounted display (HMD), instead of a screen as in the size, weight, resolution, frame rate, number of colors,
first version of CASPER (Figure 1). Thanks to the HMD, luminance, viewing angle, and thickness, we propose a
the surgeon directly perceives the guidance information design space based on more interaction-centered
displayed on top of the patient. Another example is the characteristics, that are inspired from our situation design
Image Overlay system [1] presented in Figure 5. The space. As shown in Figure 4, our framework is comprised
guidance information is displayed onto a see-through of four dimensions, namely Input, Output, Manipulation
surface located in between the surgeon and the patient’s and DOF.
body. Such an interaction situation belongs to Class IV-
output. Input
The same reasoning as the one for inputs can be applied for The Input dimension is used to characterize how the screen
studying the case where the user and the object of the task is used by the user to convey information to the computer
are distant. The two chains to be added to Class III-output system. Five values are identified along this dimension:
and Class IV-output, in between Robject and U are: none, tactile, pressure, acceleration, localization. The value
none means that the screen is not used as part of an input
(a) (AinÆSÆAout) modality. Tactile is the common input modality with a
(b) (AinÆSÆAoutÆRtool) PDA (tactile screen). Moreover sensors can be embedded
within the device. Thus pressure or acceleration can be
For example:
detected as in [4]. Finally the localization of the screen can
Class IV-output-a: SÆAoutÆRobjectÆ(AinÆSÆAout) ÆU be known by the computer system thanks to a tracking
One example of such a situation will be the following one: mechanism.
a telesurgery system displays anatomical information on Output
top of the patient’s body (SÆAoutÆRobject), while a camera The Output dimension is used to describe how the device
(Ain) facing the patient’s body enables the distant surgeon conveys information to the user. We focus here on visual
(U) to see on her/his screen (Aout) the image of the patient data but other non visual interaction languages can be used,
enhanced by the anatomical information. including haptic feedback. Along this dimension, two
Completeness of the situation design space values are identified showing whether the displayed data are
For each input as well as output situation, we described all dependent on the screen's position or not. For instance, if
the combination possibilities of ASUR entities, making the screen is tied to a tool handled by the surgeon, and it
the design space complete. Nevertheless for each situation conveys guidance information, then the output data may be
the described ASUR chain is the minimal one. While dependent on the screen's position: the displayed data
concretizing the abstract situation, some ASUR entities change according to the screen's positions over the patient's
may be inserted in the minimal chain. body. Other kind of data (e.g., blood pressure, body
temperature) may be independent on the screen's position in
The completeness of the framework makes it a useful tool that same case.
for the designer to systematically explore the set of
stationary translation rotation free
(DOF)
none
none
indirect
tactile
Screen position
dependent data
direct
pressure
Manipulation
acceleration
Captions
Screen position
independent data : Guidance system
localization
: Overlay system
Input interaction Output interaction
Fig. 4: Mini screen design space.
Manipulation Guidance system
The Manipulation dimension expresses the context of use An immediate usage of the mini screen consists of using it
of the screen. Two values, direct and indirect, are identified as an output adapter to display guidance information. We
along this dimension. The manipulation is direct if the user therefore obtain the same situation as in CASPER
holds the device. The manipulation is indirect when the presented in Figure 3 (Class III-input and Class I-output).
device is bound to another entity (e.g., an automatic arm), While concretizing the interaction, we decided that the mini
which itself is manipulated by the user. screen will be tied directly to the tool (e.g., a drill) or a
Degree of Freedom (DOF) tool guide if the tool itself is too fragile (e.g., a needle).
The DOF dimension is used to describe the number of The ASUR description of the concrete interaction therefore
different ways in which the screen can move. The screen includes: (Aout=R tool). Within the mini screen design space,
can be stationary, move only in translation or in rotation, this design decision is described by (none, screen position
or accept free motions. Those values are always based on a independent data, indirect, stationary) as shown in Figure
referential. For instance, if a screen is tied to a surgical tool 4. Taking this design decision was driven by the need to
(e.g., a drill), the screen is stationary in the tool's reduce the perceptual discontinuity as defined in [2] and
referential, but freely mobile in a more global referential: experimentally observed in CASPER. Linking the screen
Its position and orientation are therefore tool-dependent. and the tool may indeed reduce the perceptual
That referential is often defined thanks to the context of use discontinuity.
(Manipulation). As a CAS system to integrate our prototype, we have
Two CAS systems involving a mini screen chosen puncture applications, either pericardial or renal.
We present two usages of a mini screen that we designed Guidance information in these systems is limited and easy
and are currently developing. They correspond to different to represent (tool direction, tool orientation, and tool
interaction situations as well as characterizations within our depth).
mini screen design space.
Overlaying data system As on-going work, we are studying the interaction
Another possible usage of a mini screen consists of not situations of other types of MR systems and not only the
reducing it to an output adapter only, as in the previous ones that assist a user in performing a task on a physical
system, but allowing it to be manipulated as a tool by the object as in CAS systems.
surgeon. The mini screen can then be used as a magnifying During the workshop we would like to discuss the
glass or "magical lens" on top of the patient’s body. Our completeness of the mini screen design space and apply our
design is inspired from the Image Overlay system [1] interaction situation design space for describing the
presented in Figure 5. As opposed to the interaction interaction situations of the presented systems.
situation of the Image Overlay system where the surface is
an output adapter (Aout), the mini screen in our system is ACKNOWLEDGMENTS
both an Aout and a R tool. Indeed the surgeon is no longer This work is supported by the French Minister of Research
manipulating surgical tools but the mini screen. The under contract MMM. Special thanks to C. Marmignon for
interaction situation therefore belongs to Class III-output as the CASPER picture and to G. Serghiou for reviewing the
opposed to Class IV-output for the Image Overlay system. paper.
For inputs, the interaction situation corresponds to the REFERENCES
same one as in CASPER. The mini screen as a tool (Rtool) 1. Blackwell, M., Nikou, C., DiGiola, A.M., and Kanade,
is localized by an input adapter. T. An Image Overlay System for Medical Data
While concretizing the interaction, the same localizer (Ain) Visualization. Proceedings of MICCAI'98, (1998),
can be used for both the patient and the mini screen (as in LNCS 1496, Springer-Verlag, 232-240.
CASPER). We fixed the value “translation on top of the 2. Dubois, E., Nigay, L., and Troccaz, J. Consistency in
patient’s body” along the dimension DOF, as shown in Augmented Reality Systems. Proceedings of EHCI'01,
Figure 4. The ASUR description of the concrete interaction (2001), IFIP WG2.7 (13.2) Conference, LNCS 2254,
therefore includes: (Rtool fiR object). Springer-Verlag, 117-130.
3. Dubois, E., Gray, P., Nigay, L. ASUR++: a Design
Notation for Mobile Mixed Systems. Interacting With
Computers (2003), IWC, 15(4), Elsevier Science, 497-
520.
4. Fishkin, K.P., Moran, T.P. and Harrison, B.L.
Embodied User Interfaces : Towards Invisible User
Interfaces. Proceedings of EHCI’98, (1998), IFIP
WG2.7 (13.2) Conference , Kluwer academic, 1-18.
5. Green, P. S., Jensen, J. F., Hill, J. W., Shah, A.
Mobile Telepresence Surgery. Proceedings of
MRCAS'95 (1995), International Symposium on
Medical Robotics and Computer Assisted Surgery,
Wiley, New-York, 98-103.
6. Ishii, H., Ullmer, B. Tangible Bits: Towards Seamless
Interfaces between People, Bits and Atoms. Proceedings
of CHI’97 (1997), ACM Press, 234-241.
7. Nigay, L., Coutaz, J. A Generic Platform for
Addressing the Multimodal Challenge. Proceedings of
Fig. 5 : Image Overlay system. CHI'95 (1995), ACM Press, 98-105.
CONCLUSION 8. Troccaz, J., Delnondedieu, Y. Semi-Active Guiding
In this paper we presented two design spaces for MR Systems in Surgery: A Two-DOF Prototype of the
systems. Passive Arm with Dynamic Constraints (PADyC).
Mechatronics (1996), 6(4), Elsevier Science, 399-421.
• The interaction situation design space is useful for the
designer in order to systematically explore the set of 9. Turk, M., Robertson, G., Eds. Perceptual user
possibilities at an early stage of the interaction design, Interfaces. Communications of the ACM (2000), 43(3),
without being biased by a particular technology. ACM Press, 32-70.
• The mini screen design space helps the transition 10. Wagner, D., and Schmalstieg, D. First steps Towards
between the abstract and concrete interaction by Handheld Augmented Reality. Proceedings of ISWC03
characterizing possible usages of one particular innovative (2003), IEEE International Symposium on Wearable
interaction device for CAS systems: a mini screen. Computers, IEEE Computer Society.