<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>iron Syst</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>M udio-actile T</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>O.</institution>
          ,
          <addr-line>In: and Langue QC C., Egenhofr, Moratz, Notes in</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>asitnce Theun</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>at Hburg ,am Departmn of Informatics</institution>
          ,
          <addr-line>oln-Straße 30, 22527 Hburg ,am Gy erman kerzel</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2012</year>
      </pub-date>
      <fpage>25</fpage>
      <lpage>30</lpage>
      <abstract>
        <p>erst¨ Univ ogt -V h sytem blind the asitng inspred y b td presn tesd A and b impared with visualy effns ectiv to facilte haptic wledg kno asitng tacile devic. trled con discu expts rimn a a [3 sytem c , tha genrats erp otd y prote user uterancs en giv tacile eralizd as virtual tacile effns ectiv [1 tha as w explors maps alutd ev of h suc 0 lop dev acording actile T e w subtie maps</p>
      </abstract>
      <kwd-group>
        <kwd />
        <kwd>p</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>,
M
a
t
t
h
i
a
s
K
e
r
z
e
l
a
n
d
C
h
r
i
s
t
o
p
h
e
r
H
a
b
e
l
n
t
n
t
{
l
no-visual
increas
ey</p>
    </sec>
    <sec id="sec-2">
      <title>Genratio,</title>
    </sec>
    <sec id="sec-3">
      <title>Maps</title>
      <p>wo</p>
    </sec>
    <sec id="sec-4">
      <title>Acesibl</title>
      <p>rd
M
s
aps,
ulti-M</p>
    </sec>
    <sec id="sec-5">
      <title>Maps,</title>
      <p>o dal
erbaly V
M
aps,</p>
    </sec>
    <sec id="sec-6">
      <title>Virtual</title>
    </sec>
    <sec id="sec-7">
      <title>Et virome n M aps, Naturl</title>
      <p>n
Maps are imp ortant external
urban-area maps in every-d
bakery or physician in enviro
limited for blind and visually
substitutes. Acquiring knowl
by vision: On the one hand, i
than that provided by visua
b e integrated over time. To
maps, providing additional
11] we suggested that appr
by developing a system im
Tactile Maps (VAVETaM).
assisting utterances, which
the prop er names of the m
drawbacks stemming from
utterances are based on a
representations of space. F
ay scenarios such as plann
nments novel to them. The
impaired p eople, therefore,
edge from maps haptically
nformation provided by tac
l maps; on the other hand
reduce some difficulties in
information via the auditor
oaches towards audio-tactil
plementing Verbal ly Assisti
The suggested system gener
not only inform the explorer
ap ob jects, but include more
the sequentiality of tactile-ma
corpus study (see [9] for a
or example, humans use
ing a route to the next
access to visual maps is
tactile maps are used as
is different from doing it
tile maps is more sparsely
, this information has to
haptic comprehension of
y channel is useful. In [3,
e maps can b e extended
ng Virtual-Environment
ates situation-dep endent
of the tactile map ab out
information that reduces
p reading. The generated
discussion of the corpus
P
r
o
c
e
e
d
2
and the set of assisting utterances). For example, the assisting utterances include
information about spatial relations between map objects. The Sensable Phantom
Omni device allows the map user to perceive virtual tactile maps (see Fig. 1). The
device can be thought of as a reverse robotic arm that generates force depending
on the position of the pen-like handle. Streets and landmarks such as buildings
are marked as indentations that can be felt with the device.</p>
      <p>In this paper, we present a prototype implementation of the VAVETaM system.
The goal of the development of this prototype is to show the technical feasibility
of the suggested system; that is, that the prototype works. The feedback gained
from the user study presented in this paper will be used for further development.</p>
      <p>In the remainder, we discuss relevant literature (Section 2) and we briefly
introduce the prototype (Section 3). Furthermore, we present a user study with
the prototype (Section 4) before concluding the paper.
2</p>
      <p>
        State of the Art
Audio-tactile maps as improvements of uni-modal tactile maps have successfully
been developed and tested for usability. Some approaches use physical (‘printout’)
overlay maps on touch screens (e.g., [13, 14]). The Phantom device used in our
prototype has been successfully used to present virtual tactile maps with
audiofeedback [
        <xref ref-type="bibr" rid="ref1">12, 1, 4</xref>
        ]. Existing approaches link sounds or brief verbal information
(mostly, names) to map objects. In a corpus study, we found that humans asked
to assist map explorers include more information in their assisting utterances,
which can be described as brief descriptions of the local surroundings [9]. The
VAVETaM approach extends earlier systems by generating assisting utterances
more similar to human performance. That non-visual knowledge acquisition by
direct perception of indoor environments can be facilitated by situated verbal
descriptions was shown by [2].
      </p>
      <p>We have positively evaluated the multi-modal interface that is described here
prior to implementation in Wizard-of-Oz-like experiments (i.e., the assisting
utterances were controlled by the experimenter) with both blindfolded sighted
and visually impaired people [10, 7]. Participants received assisting utterances
as suggested compared to a condition in which they were only informed about
the names of objects. Spatial knowledge acquisition of virtual tactile maps was
facilitated by verbal assistance.
3</p>
      <p>The Prototype System</p>
      <sec id="sec-7-1">
        <title>Haptic Device Haptic Event Detection</title>
      </sec>
      <sec id="sec-7-2">
        <title>Generation Verbal Assistance</title>
        <p>To generate assisting utterances, two important tasks have to be solved. First,
the stream of position information from the haptic device is analyzed in order to
detect semantic—that is, meaningful—exploration events. Haptic event detection,
for example detects when a map user explores a street with the haptic device.
This information is the input to the generation component. Second, based on this
input, appropriate information is selected from a knowledge base and prepared
for verbalization; that is, natural language is generated.</p>
        <p>For each of these tasks, components were developed. Haptic event detection
is described in more detail in [6, 5]. A generation component and the interface
between event detection and generation is discussed in [9, 8].</p>
        <p>We connected the two components to a fully working prototype system.
The utterances produced by the prototype are similar to those evaluated in
previous experiments (see Section 2). For example, when the movements the user
performs with the device indicate that he or she is interested in a street called
‘Amselweg’, the system produces an output which can be translated as follows:
‘This is Amselweg. This street is parallel to Blumenstraße. It forms a corner with
Dorfstraße at the top end and towards the bottom it is restricted by the map
frame. Furthermore, it crosses Hochstraße.’ See [9], for a thorough discussion of
the content of the utterances. The generation of the utterances is not based on
canned text. Both, the verbalization history and the current map exploration,
are taken into account: When information was already verbalized, repetitions
leading to unnecessary redundancy are avoided. When the map user starts to
explore another map object—for example, a landmark called ‘Rathaus’ (town
hall)—the verbalization is stopped after the ongoing sentence is finished and the
map user receives currently relevant utterances.</p>
        <p>The generation system and the haptic event detection are implemented in
Java. We use the Chai 3D toolkit for haptic rendering of the virtual tactile map.
The Chai 3D toolkit and the haptic event detection are interfaced using JAVA
Native Interface. Speech synthesis was realized using Mary TTS.1
1 tp:h//w.hai3cd.org and tp:h//.marydfki.de
.</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>Ealution :v</title>
    </sec>
    <sec id="sec-9">
      <title>Mean Resp on</title>
      <p>I
refd.</p>
    </sec>
    <sec id="sec-10">
      <title>The sytem</title>
      <p>t en w
,
e
r
z
e
l
ea
p o
.70
.10
.50
.80
.70
.50
.90
.80
1
.
7
0
1
.
5
0</p>
      <p>S</p>
      <p>D
r
o
c
e
e
d
i
n
g
s
otyp e system, we asked 13 participants (university students,
rse credit or monetary; mean age: 24.7 years, SD = 7.2 years,
prototyp e and to give us feedback by indicating agreement
it and by taking part in semi-structured interviews.
aining with the haptic device, participants were blindfolded
the VAVETaM prototyp e. After they rep orted that they
avior of the prototyp e, they were instructed to learn the
ould b e able to sketch it afterwards. This instruction was
clear goal for the exploration of the map. Participants
s they wanted to explore and memorize the map (overall
= 14:16 min, SD = 4:46 min; interaction after instruction:
3:42 min). After they sketched the map, they were asked
nt to a set of statements ab out the prototyp e. Selected
n in Table 1.2 They were given the list of statements in
swers were given on a 1–5 Likert-typ e scale (1 corresp onds
y’ and 5 to ‘I do not agree at all’). The list of statements
her with the exp erimenter in order to enable immediate
ant p oints (discussions were audio-recorded).
ses to the statements indicate that participants considered
to supp ort tactile maps with natural langauge (B) and
b ehavior of the prototyp e comprehensive (J). Furthermore,
aps were considered understandable (A, F). The resp onses
ossible to lo cate landmarks, although with more difficulty
ts (G). A p ossible explanation is that the connected network
2</p>
      <p>Due
of the
orse w
from:
other
neutral.
tp :h
of
y prote
the
pro duce
pa er,
a
The
/
ful
/
list of staemn and
w.informatk.uni-burg .hamde
mean
/
resp on can
WSV/
b e d retiv
AVV
Verbally Assisted Virtual-Environment Tactile Maps: A Prototype System
of streets was easy to follow, while landmarks were not connected with each other
and had to be explored one by one.</p>
      <p>Participants considered the timing of assisting utterances appropriate to
establish reference to the map object to which the assistances belonged (I). They
considered the utterances helpful and they indicated that they knew what was
meant (D, E). The ratings whether utterances were understandable (C) are
comparably low. This seems somewhat contradictory to the high agreement with
assertion D. In the discussion and the interview, participants indicated that
problems with understandability were in general limited to proper names.</p>
      <p>Currently, once sentences are started they cannot be stopped or changed
anymore. From our own experience with the system, we expected that this is
potentially confusing. In fact, most participants considered the inability of the
prototype to change current utterances confusing (H).</p>
      <p>In the interviews conducted subsequently, the points discussed above were
elaborated and extended. Participants in general considered the set of assisting
utterances appropriate and did neither miss important information, nor did they
indicate that superfluous information was given. When asked for suggestions on
how to improve the system, participants considered options for customization
(changing the haptic presentation of the map, switching certain kinds of assisting
utterances on and of) and the quality of the synthesis as potential improvements.
Some participants indicated that they would have changed the content and
frequency of certain utterances—however, this feedback was not consistent.
5</p>
      <p>Conclusion
We presented the Verbally Assisting Virtual-Environment Tactile Maps
(VAVETaM) prototype that solves the task to generate assisting utterances for tactile
map explorations. The prototype is based on the interaction of two components:
Haptic event detection and assistance generation. The effectiveness of the system
was previously evaluated positively in controlled experiments [10, 7]. We presented
a user study with the prototype in order to show that it works. Participants
considered the prototype behavior comprehensive and the utterances and haptic
presentation understandable. Potential improvements pointed out by the users
are (1) the quality of the speech synthesis with respect to proper names, (2) the
possibility to customize the behavior of the system, and (3) the implementation
of a possibility to change ongoing utterances when the user moves to another part
of the map. Aside from these potential improvements, both, with respect to the
assisting utterances and to their timing, participants considered the prototype
system usable, comprehensive, and helpful.</p>
      <p>Acknowledgments. The research reported in this paper has been partially
supported by DFG (German Research Foundation) in IRTG 1247 ‘Cross-modal
Interaction in Natural and Artificial Cognitive Systems’ (CINACS). We thank
Martin Christof Kindsmu¨ller for discussion. We thank the anonymous reviewers
for their highly useful commentaries.
e
e
e
c</p>
      <p>s
., Rena,</p>
    </sec>
    <sec id="sec-11">
      <title>In: orld W 15. , B holgia Psyc</title>
      <p>3
K
In:
0.
K
In:</p>
    </sec>
    <sec id="sec-12">
      <title>Conferc .</title>
    </sec>
    <sec id="sec-13">
      <title>Pro cedings</title>
      <p>F.,</p>
    </sec>
    <sec id="sec-14">
      <title>Haptics C</title>
    </sec>
    <sec id="sec-15">
      <title>Tsukba (20 akdsh, J.Z.,</title>
    </sec>
    <sec id="sec-16">
      <title>Atolic,</title>
      <p>onferc. p.
07
L
)
eg,
7–358
M.,
of AA</p>
    </sec>
    <sec id="sec-17">
      <title>Menlo ark ,P</title>
    </sec>
    <sec id="sec-18">
      <title>Mhna ,osc</title>
      <p>(200</p>
    </sec>
    <sec id="sec-19">
      <title>Lohman, K.</title>
    </sec>
    <sec id="sec-20">
      <title>AI orkshp W CA .P, ars ,Tzo v ,</title>
      <p>.
,
D
e ,istan
A
.
:
A
aplicton
G
.</p>
      <p>E
.
:
ords :w
7
:
)
erbal V asitnce
tacile-map
aps:
ards w to
1
.
2
.
3
.
4
.
.
.
.
.
.
.
.
.
.
.
rendig
(
es ,P ark
t men
Pro cedings
disable
ang ,W
aces
erst¨ Univ</p>
    </sec>
    <sec id="sec-21">
      <title>Ehorn ,ic</title>
      <p>y prote sytem.
orkshp W on Sph ec</p>
    </sec>
    <sec id="sec-22">
      <title>Mtreal ,on</title>
      <p>Eba ,sc hen
exploratins. In:
iudce, N.,</p>
    </sec>
    <sec id="sec-23">
      <title>9. Lectur</title>
      <p>Hab el, C</p>
    </sec>
    <sec id="sec-24">
      <title>Pro cedings of Spatil Cognit</title>
    </sec>
    <sec id="sec-25">
      <title>Kerzl,</title>
      <p>I.,
orkshp W on
0)
olas ,Nik G
IE
.:
2
M</p>
    </sec>
    <sec id="sec-26">
      <title>Bergman,</title>
    </sec>
    <sec id="sec-27">
      <title>Multimo da Outp</title>
      <p>D
In:
visualy
oNtingham
,
maps
of the
p eol.</p>
      <p>Z.
digtal
1th
4
“NO A
informat partily ted sigh p eol.</p>
    </sec>
    <sec id="sec-28">
      <title>2nd teraol in confer on maps and graphics for</title>
      <p>(1988)
Li, B.,
ofr p eol with visual t .imparen In:
teraol in ACM
3–50. A
:</p>
    </sec>
    <sec id="sec-29">
      <title>Genratig situaed asitng B</title>
      <p>aumn,
the</p>
    </sec>
    <sec id="sec-30">
      <title>Pro cesing for</title>
      <p>(2
H
Ae sitv hnolgis ec T
01
ab el, C
p erction
lingusM
.</p>
      <p>(eds.
)
Pro</p>
    </sec>
    <sec id="sec-31">
      <title>Hburg am (in preation) ,</title>
      <p>R.</p>
    </sec>
    <sec id="sec-32">
      <title>Computer</title>
    </sec>
    <sec id="sec-33">
      <title>Extend</title>
      <p>01
.,
K</p>
    </sec>
    <sec id="sec-34">
      <title>Genratio orbys ,W</title>
    </sec>
    <sec id="sec-35">
      <title>Scien,</title>
      <p>erbal v asitnce faciltes
2
H
.,
2
Springe,
wledg kno
pres)
C
H
1
:
.</p>
      <p>M.</p>
    </sec>
    <sec id="sec-36">
      <title>Berlin,</title>
      <p>(eds.)</p>
    </sec>
    <sec id="sec-37">
      <title>Heidlb rg</title>
      <p>1
)
G
C.
enratig
,
rbeal v
.
.,
E
K
E
ostp ul,
M
K
ultimeda
.
:
Haptic
”:
to l
acquiston,
mangeT.
,
Hen ,a v
T.
:
A
,
C</p>
    </sec>
    <sec id="sec-38">
      <title>Pitsburg,</title>
      <p>C
AP
E
(
and
)
A
.cesiblt y</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>4 erzl,</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>