=Paper=
{{Paper
|id=Vol-91/paper-9
|storemode=property
|title=A Mixed Reality System for Design Engineering: Thinking, Issues and Solutions
|pdfUrl=https://ceur-ws.org/Vol-91/paperE3.pdf
|volume=Vol-91
|dblpUrl=https://dblp.org/rec/conf/mixer/NellyJ04
}}
==A Mixed Reality System for Design Engineering: Thinking, Issues and Solutions==
A Mixed Reality System for Design Engineering:
Thinking, Issues and Solutions
Nelly de Bonnefoy Jean-Pierre JESSEL
EADS Corporate Research Centre Toulouse Computer Science Research Institute
Toulouse Computer Science Research Institute (I.R.I.T.)
(I.R.I.T.) Paul Sabatier University
31700 Blagnac, France 31400 Toulouse, France
33(0)5 61 58 48 51 33(0)5.61.55.63.11
Nelly.de-bonnefoy@eads.net jessel@irit.fr
ABSTRACT exchanges around the physical object aims at taking some
Despite the extended use of Digital Mock-Up (DMU) decisions concerning the future product configurations such
applications during the development of complex as systems segregation, ergonomics, and physical
aeronautical products, some major changes can be arrangement of components. Consequently, during those
requested in order to correct deficiencies or to improve reviews, debates between the different designers can lead to
further products performances, especially when the first one a request for an engineering change of the product.
comes out production. These changes are discussed upon Usually, information exchanges are made verbally helped
the current product configuration and imply trades-off with hand made sketches. In those cases, it might be
between different competencies involved in the product’s difficult for a designer to make his engineering change
lifecycle. Based on a specific use case, we first highlight the proposal understood by others team members. This
underlying issues associated to the development and situation highlights a requirement implied by this kind of
introduction of Mixed Reality (MR) systems. We tackle meeting: the need to visualise the proposed modification, in
these issues from cognitive and ergonomic points of view as order to assess the potential changes impacts.
well as technical. Then, we propose several MR solutions,
which will be studied further during the two remaining Mixed Reality (MR) systems can efficiency support inter-
years of our research project. Theses solutions will then be participants information sharing around a physical object.
evaluated against the specific needs and requirements of the As matter of fact, MR systems can bring some more
aeronautic industry. engineering information about the physical object through
“optical see through”-based visualization applications. Part
Keywords changes are made visible by adding complementary
Mixed Reality, Interactions, Design Review. information to the current part («Augmented reality»
INTRODUCTION concepts)(Cf. figure1). This “augmentation of reality” can
When the first product comes out production during the be realised with the image incrustation in the users field of
development of complex aeronautical products, to correct view. Incrustation are performed through, optical see-
deficiencies or to improve further products performances, through head mounted displays. Those display devices
some major changes can be requested. They are discussed allow users to keep their environmental perception while
upon the current product configuration. These changes also having intuitively access to more contextual information.
imply trades-off between different competencies involved Therefore, such systems leverage the typical limitations of
in the product’s lifecycle. It is typically a collaborative paper-based systems, and add value to exchanges and trade-
work situation where a team of designers gather around a off through of their intuitive and fast access mode to
table to perform a product design review. Information pertinent information.
sharing and negotiation movements during those review
sessions are strongly influenced by the respective speciality,
knowledge and experience of participants. Information
Figure 1 - The Context
SYSTEM DESCRIPTION · the final change.
The investigated system allows users to virtually simulate
Depending on the nature of the information to be
the integration of an engineering change by modifying
communicated to the user, relevant output interaction
designers’ perception of the real object.
modalities have been identified. The next section present
During the meeting, all participants share the visualisation the various output modalities used and for each, the
of the actions performed on the augmented model, but the different types of information associated to them, and the
access to the model itself must remain. So that conflicts can best user-centred utilization.
be avoided. Practically, a designer must have the knack if
he desires to create virtual modifications. One design
sequence could be performed like this: he selects a virtual Output modalities
geometric form in one menu, it displays in front of him, and Three different output modalities have been considered in
he can modify its attributes by using the sub-menu order to make more perceptive actions:
associated to the virtual form. He is free to place the virtual · the visualisation through a Head Mounted Display
modification of the real object when and where he wants. (HMD),
The introduction of such systems within the review · the text-to-speech system,
environment must preserve natural communications and
social cues between participants. Indeed, the review process · and the haptic force feedback.
and user behaviour must not be perturbed by a complex MR Display of the HMD
system and its multiple accessories. The use of such system The optical see-through HMD is a display device allowing
has to be as natural and intuitive as possible, this approach users to keep their environment perception while visualising
is similar to the “Natural User Interface” (NUI) developed more information in their field of view, in context.
by [1].
Hardware
This system comes also under the definition of
The figure 2 describes the basic principle of the optical see-
Collaborative Mixed Reality systems. Some references
through Head Mounted Display:
already exist in the literature:
· “Build-it”: a collaborative mixed reality system
which tend to be a NUI [2],
· “Arthur”: the development of an interactive task
oriented collaboration environment based on
augmented reality [3],
· “MARE”: a multi-users augmented reality
environment on table set-up [4].
Figure 2 - Optical see-through HMD
INFORMATION SHARING
Information diversity There are two main devices interesting, the Sony Glasstron
Different types of information exchanges take place during (no more on the market) and the Nomad from Microvision.
review meetings, the next section focuses on system/user Equipped with this device, each designer visualizes all
information transactions. information in his field of vision, such as his environmental
We considered two major typologies of information to be information and system information. The system must
manipulated by the user: textual information, and graphical perform display in the most user-friendly and ergonomic
information. Both type can be broken down as follows: way so the user cannot be lost within the information
display density.
Text information of consists of:
For that, we discuss the general exploitation of the field of
· interface information (menus, etc.), view and we detail requirements display for each kind of
· system messages, information.
· annotations the user wants to add. Usually, people fields of view consist of one main area that
Graphical Information consists of: is a zone in the sight direction where people see all things
clearly, and around it, the field of view area where people
· arrows, chips, or others in order to show have to move eyes to see distinctly information (Cf. Figure
something to advantage, 3):
· classic geometric shape the user will manipulate in
order to obtain the final modification,
Field of view Interface information are in this area because users do not
have to focus their attention upon it. Users do not need to
keep this type of information at time in the centre of their
field of view. This area is visible only if the user has the
knack.
Direct sight access to information
System message
System messages (2) are important for users to navigate
within the application. Users must perceive this kind of
Indirect sight access to information
information as quick as possible. That is why they will be
displayed in the centre of their field of view. A special
colour has to be selected, in order to increase messages
Figure 3 - Field of view organisation
ergonomic and cognitive values.
One can be noticed that, most of the time, when people Annotation and geometric shape
have to look at something on the right (or left), even if it is
close to them, the head moves more than eyes. People tend At the beginning of the process, annotation (3) and
to put the subject of their attention in the centre of their geometric shapes (5) do not need, necessarily, to be
field of view. connected with the real object. The system will display this
kind of information in front of the user field of view, in the
Each type of information does not require the same display “direct sight access to information” area. They could be
modes. Displays do not have the same duration, neither the modified, moved, deleted in the users own way.
same location imperatives (Cf. Table 1).
Final change and Arrows, chips, etc.
LOCATION
The information category that concerned final change (6)
Specific area, Specific area, No specific and arrows, chips, etc (4), is used to emphases particular
Link with no link with area, no link items. For example, it will be used to precise a place
part part with part through an arrow or to show the real part modification
Interface (1) X through a virtual change.
Syst messag. (2) X Their location in the users’ environment is important. They
depend on the real part location in the meeting room and
Annotation (3) X X not on one specific user field of view area. The
Arrows… (4) X modification part, which is geometric shape, has an exact
location on the real object; however, arrows and others only
Geom. Shape (5) X X require an accurate location associated to the real object.
Final change (6) X General
Different colours will be used: each designer will have his
Table 1 - Information type location own colour, in order to differentiate users’ actions.
These information displays are developed hereafter. Different colours are associated to system information, one
Interface information for menus and one for messages. The selected menu section
will have different background than others, and if the user
In order to get designers used to the location of interface
goes down in sub-menus the navigation path will be
information (1), specific zones are associated to this type of
displayed as a key words list.
information (Cf. Figure 4). It is located as follows:
Field of view Moreover, users have the opportunity to remove display; all
visual information will be hidden if he desires so.
Menu
Text to speech
Sub This output modality is the sense of hearing one.
-
Menu The hardware is a headphone.
The most important part, with this modality is to have a
good thought concerning its use. If text-to-speech is used
permanently, for all information in text form, it becomes
unbearable for users very fast. And in that case, users put
Figure 4 – Interface, menu and information location the sound off.
This modality has to be used only if it gives an added value This choice has been made because interactions with the
to the system situation. That is why it is generally fitted for system must remain as intuitive as possible.
classic system messages: system announcement or error As we already mentioned there are different types of
messages. Users’ attention has to be focused on this type of information to be manipulated. The user has to interact with
information. all of them.
Haptic force feedback Speech recognition
This output modality is used to modify, and move the In this system, the integration of speech recognition is done
geometric shape that will become the final change and also in two stages.
to touch the final change. As matter of fact, the haptic force
feedback is very important to manipulate virtual objects. To First, speech recognition is used to browse menus and sub-
make the system more realistic, users must perceive tactile menus. In this case, the system must be able to recognize
information. But today, without special devices, it is almost words rather than sentences. Commercial off-the-shelf
impossible for users to catch this information from virtual applications are enough efficient to perform these functions.
objects. But a great deal of attention must be paid to the design of
menus and to the selection of a clear and concise
Hardware vocabulary.
There are two main different accessories: the glove and the In the second integration stage, the speech recognition will
Phantom, which is a computer device most closely related be used to navigate in menus, to modify virtual changes and
to the mouse. Their function is to interact with objects in a integrate them on the real product. This will imply that the
three dimensional environment. During the last few years, system will perform sentence recognition making the
many research have been carried out in this domain, and system more friendly and intuitive.
accessories progresses in ergonomics are significant.
There is, for example, the Cyberglove
(Cf. Figure5) with a vibrotactile Gestures recognition
feedback [5]. There are small Like speeches, gestures are a spontaneous mean for people
vibrotactile stimulators on each finger to communicate with other actors. The use of gestures in
and the palm of the cyberglove. They multimodal applications facilitates users interactions, in
can produce complex tactile feedback particular in noisy environments. Moreover, users tend to
patterns. Even if this glove is an execute gestures for manipulation operations rather than
accessory, the ergonomic material Figure 5 – state them or access to them by using classical interfaces
aspect is studied: users cannot feel that Cyberglove [6] like window, icon, etc.
they have a robot hand, they just wear The system use gestures recognition to interact with
an ordinary glove. interface information and to modify the shape of the
Because the system has to be as intuitive as possible and changed geometric form. This feature, in some cases, will
because there is no method without accessories, ergonomics be used simultaneously with speech recognition features.
issues play a major role. The devices that will be selected The goal is to identify and track the gestures of the user that
must allow users to keep their usual meeting behaviours. has the knack. As people know, there are different ways to
For example, they should catch objects if they want, such as make gestures recognition.
a pen, a glass, etc. Method with Digital Gloves
Flexion angles measurements, which are obtained with an
optic fibre positioned on each finger, give fingers
In an information exchange there are two communication
configuration and position. These angles are determined by
ways, in our case: “system (S) to user (U)” and “user to
the luminous signal intensity sent in the fibre and with its
system”. The last section explains how the user perceives
intensity in tip of finger. A tracker is located on the hand in
information from the system (SàU). The next section deals
order to process the hand position and orientation. This
with the User to System communication (U à S), which are
method gives accurate results, but it constraints users to
kinds of interaction the user can perform.
wear a glove generally linked to a system (depending on
technologies employed). Users do not have hands free.
INTERACTION – INPUT MODALITIES Visual Methods [6] [7]
Human-system interactions have the objective of These methods are based on computer vision and on image
developing models, concepts, tools and methods, in order to processing techniques. Hands movements are recorded with
realise systems that answer users’ needs and aptitude. one or more video cameras. Then different techniques can
Reproducing usual human-human communication modes, be used to process images depending on the gestures
the modalities used in this system are: voice and gestures. recognition method used. It is more difficult to use this kind
of method but users get rid of physical accessories. Most of As people know, there are different sorts of trackers:
the processing techniques consist of four operations, which · Electromagnetic trackers (alternating current,
are acquisition, segmentation, characteristics extraction and direct current, compass)
classification. They can be realised in different ways: based
on markers, on three-dimension model, on visual · Acoustic trackers (distance measurements
appearance. The main advantage of these methods is that determined by ultrasonic time of flight),
users do not have to wear physical artefacts. · Optical trackers (with punctual receptors
This input modality will be use to navigate in interface (phototransistors), or video based tracking),
menu and to modify and move virtual geometric shape. · Mechanical trackers (inclinometer, gyroscope,
accelerometer…),
Now that all modalities have been defined, the tracking · GPS trackers.
system, which has a primordial role in Mixed Reality The good configuration of trackers has to be found. But
System, must be tackled. some considerations have to be taken into account:
In order to make our system as natural as possible, the use
TRACKING AND REGISTRATION of peripheral is limited and bulky peripheral are proscribed.
To perform an incrustation of a virtual object in the user The less the system uses devices, the better it is.
field of view, scene components need to be located For instance, as one device used for the haptic force
accurately. Indeed, to make a good registration, MR feedback, could be also used for gestures recognition as
systems need trackers with approximately one millimetre well.
accuracy in position and a low fraction of degree in
Moreover, this use case takes place in a specific meeting
orientation. Most of commercially available trackers answer
room that is a well-defined place and where the luminosity
one of the two conditions but not both.
is constant.
Tracked elements
Considering the huge advances made in video technologies
In order to offer an intuitive and a free visualisation of part during the last few years, it is now possible to find little
modification, the system will track continuously and with cameras with very good resolution. This greatly improves
accuracy different environment elements: the quality of image processing for markers recognition.
· the physical part, Moreover these video cameras are now equipped with USB
· each designer, communication port, providing a good data quality and
speed transfer. Practically all web cams have these
· users points of view, characteristics today, so a good camera could be found for a
· users hands. reasonable price. To finish with, little cameras start to be
equipped of IEEE communication port, which offers the
The first three elements are tracked to make an efficient
best transfer speed and quality.
registration of the virtual modification on the real object in
each user field of vision. The last one is tracked so users
can realize virtual changes with gesture recognition, move For all these, the video based tracking has been selected for
the virtual object, and perform haptic force feedback. our system. Possible video based tracking methods are
The tracking system is designed to create relationships presented in the next section.
between each tracked element configuration (Cf. figure 6).
Video based tracking methods [8]
As people know there are two main video-tracking
configurations:
· Inside out (Cf. Figure 7)
One or many video cameras are on the moving target. They
watch markers fixed in the environment which are
references.
· Outside in (Cf. Figure 8)
One or many video cameras are fixed in the environment,
they are references; they watch the target movements, on
Figure 6 - Referential example
which markers have been affixed.
Glasstron and the Nomad of Microvision. Their display
length on the field of vision can be considered as small.
Devices Resolution Display length
Sony Glasstron 800 x 225
Microvision HMD 800 x 600 23° x 17°
There are other display devices on market, based on the
head mounted display concept, like video-see through
Figure 7 - Inside-out Figure 8 - Outside-in
HMD and some video screens that can be clipped on
Markers should be classic draws or LEDs. Once the glasses. However, the first one does not allow users to see
configuration is chosen, there are different manners to their close environment directly, and the second type
calculate the target location. The first uses the two or more proposes a small display of a computer screen, PDA screen,
cameras in order to calculate the target marker positions, by etc.
using the triangulation for example. The target orientation
should be calculated by using several markers on it. The
second use the pattern recognition techniques; there is only Real object location issue
one video camera and some target markers geometric The real part is on a Location of the First extension
2nd extension
knowledge. table. There is a
visualisation problem
The fact that the inside-out method gives more accurate
(Cf. figure 9) if
results and a better orientation resolution than the outside-in
different changes are
method should be noticed.
made in different
locations of the real
Registration part (one of the bottom Figure 9 - problem visualisation
The registration is one recurrent problem in such system. In and another at the top
order to make the visualisation of the real object and its for instance).
virtual change realistic, an accurate registration is required. The real object location needs a thorough thought in order
Two types of errors can be encountered: the static and the to permit users the visualisation of different changes.
dynamic ones [9].
The static errors are due to the optical distortion, errors of Gestures recognition issues
the tracking system and differences between models or Gestures recognition using visual approach, in particular
material specifications and real material physical properties. without the use of markers, is not a mature technology.
This kind of errors is perceived even if the user does not Some systems work in real time but in very specific
move. conditions, such as with a uniform background, a small
The dynamic errors are due to the processing time lag that vocabulary, etc. Even if a system that recognises gestures,
is the delay between measurements made by the tracking with a visual approach without markers, in real time, in a
system and the display of the virtual entity. In fact, there are normal environment would add a tremendous value to
due to all processing time devices and systems. Different general MR systems, it remains, for the moment, a research
ways have been explored to reduce this dynamic error, by perspective.
reducing the system lag or the perceptible delay [11], by Moreover, as people know, in the human-human
making location prediction [11], by image matching. communication, the use of speech, gestures, and facial
To make a good registration the tracking system must expressions contributes to the information exchange. In
perform good locations. particular during argumentative phases or solution
negotiation activities, the frequency of gestures and facial
To resolve the static errors a calibration have to be made,
expressions significantly increases. The case study
and as many system, ours will use Kalman filtering to
presented in this paper deals with one specific situation
reduce dynamic errors.
where people will try to negotiate integrating several
different points of view. So, making the difference between
SYSTEM ISSUES social communication gestures with actual gestures
performed to interact with the system is primordial in our
Hardware issues: Optical see-through Head Mounted
Display case. The system must have the possibility to distinguish
For the moment this technology is not mature enough. As these two types of gesture, in order to avoid false
we saw upper, there are two main devices, the Sony interpretations.
Virtual modifications tracking system and issues met. All these bases will allow
Object that people can bring in a meeting room are us to build a prototype.
generally medium-sized. Moreover, changes made on a part
during the review sessions are not consistent; the general
object shape is not called into question as modifications are REFERENCES
localised and concern only small areas. Therefore, to keep 1. Rauterberg, M., From Gesture to Action: Natural User
the visualisation as realistic as possible, the registration of Interfaces. Eindhoven University of technology; Mens-
the virtual change has to be very accurate. Machine Interactie, Diesrede 1999, 15-25.
Part modification is not necessary an extension of the real 2. Fjeld, M., Bichsel, M. and Rauterberg, M.: BUILD-IT:
object, it could be a suppression of a little part area. To Intuitive plan layout mediated by natural interaction.
make an extension the system proposes to the user to mould Arbete Människa Miljö & Nordisk Ergonomi (work,
a classic geometric form and after to register it on the real human being, environment), 1/99, 49-56.
object. But to make small area suppression, it is more 3. Romell, O., Arthur (augmented Round Table for
difficult to make it realistic, the representation is more Architecture and Urban Planning) – An augmented
elaborated. Reality, Task Oriented Collaboration Environment. AVR
The system has the three-dimension model of the real II and CONVR 2001 Conferences at Chalmers,
object. This will help us to make this kind of modification Gothenburg, Sweden, October 4-5, 2001, 29-37.
as realistic as possible. 4. Grasset., R., and Gascuel, J.D.: MARE: Multiuser
Our first solution is to use a dark shape to simulate the Augmented Reality Environment on table set-up.
suppression area. But the realistic visualisation depends too 5. http://www.immersion.com/3d/products/cyber_touch.
much on the user’s location (Cf. Figure10).
6. http://www.immersion.com/
Dark shape
7. Ying Wu, Thomas S. Huang, "Vision-Based Gesture
Recognition: A Review", A. Braffort et al. (Ed.) Lecture
Notes in Artificial Intelligence 1739, Gesture-Based
Communication in Human-Computer Interaction
(International Gesture Workshop, GW'99, Gif-sur-
Yvette, France, March 1999).
8. Ying Wu, "Vision and Learning for Intelligent Human-
Figure 10 - different modification views Computer Interaction", Ph.D. Dissertation, 2001,
Department of Electrical and Computer Engineering,
University of Illinois at Urbana-Champaign.
Occlusions issues
9. Rolland, P.J., Baillot, Y., and Goon A., A survey of
Occlusion is the typical issue of this type of MR system.
tracking technology for virtual environments. Center for
All actors are situated in a close room area. In our study research and education in optics lasers (CREOL),
case, the design review is made around a table; so all users university of central Florida, Orlando FL 32816, 1999.
move and perform their task, in a confined place. In this
10. Azuma, R., A survey of augmented reality. In presence:
context, occlusions will be frequent.
teleoperators and virtual environment 6, 4, August
Solutions must be found in the room layout, in cameras 1997, pp.355-385
positioning and in tracking methods.
11. Kijima, R., and Ojika, T., Reflex HMD to Compensate
Lag and Correction of Derivative Deformation, IEEE
CONCLUSION Virtual Reality Conference, Orlando, Florida, March 24
This paper deals with information sharing in our MR - 28, 2002, pp.172-182.
system for design engineering, the modalities used, the