<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Barcelona, Spain | September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Text Entry in VR and Introducing Speech and Gestures in VR Text Entry</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jiban Adhikary</string-name>
          <email>jiban@mtu.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Michigan Technological University Houghton</institution>
          ,
          <addr-line>MI 49931</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>3</volume>
      <issue>2018</issue>
      <fpage>15</fpage>
      <lpage>17</lpage>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Permission to make digital or hard copies of part or all of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for profit or commercial advantage and that copies bear this notice and the full citation
on the first page. Copyrights for third-party components of this work must be honored.
For all other uses, contact the owner/author(s).</p>
      <p>Copyright held by the owner/author(s).</p>
      <p>MobileHCI, 2018 Barcelona, Spain.</p>
    </sec>
    <sec id="sec-2">
      <title>Biography</title>
      <p>I am a Computer Science PhD student at Michigan
Technological University, Houghton, Michigan, USA. I
have a Bachelor of Science degree in Computer Science
and Engineering from University of Dhaka, Bangladesh.
My current research focuses on designing interactive
systems for text entry in midair and virtual reality (VR)
Currently I am working under the supervision of Dr.
Keith Vertanen who is a renowned researcher in the
field of text entry.</p>
    </sec>
    <sec id="sec-3">
      <title>Summary of Related Past Works</title>
      <p>
        A lot of works have been done on text entry techniques
in personal computers and mobile devices. However,
works related to text entry in midair and VR
environments are barely sufficient. This is mainly
because there have always been challenges to design
and implement text input surfaces in midair or VR, to
track or sense user’s actions and to map the interaction
between the user’s actions and the input surface.
We have been working on designing an interface for
entering text in the virtual environment for a year. Text
entry in VR environment is different from computers or
mobile devices because there is no physical keyboard
or touchscreen in this environment to interact with.
Entering text in VR environment can be achieved by
speech, gestures or virtual keyboard. Our work focuses
on entering text in a virtual keyboard. We have been
able to create a prototype of a virtual keyboard that
enables user to input text in the virtual environment.
This prototype senses finger movements of the user
along the keyboard by using a Leap Motion Sensor. We
have also incorporated a sentence based decoder
named VelociTap[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] with this prototype for auto
correction and the prototype provides audio feedback
for better interaction. We plan to extend the prototype
to a usable interactive system and investigate its
limitations and benefits by conducting user studies.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Speech and Gestures in VR Text Entry</title>
      <p>While working with our prototype, we implemented a
single gesture (a thumbs up) as a delimiter of an
interaction in virtual environment. For example, the
thumbs up gesture could be used to mark the end of a
sentence. For our current prototype, it does not
necessarily require many gestures to fulfil our main
objective but it will really be interesting to explore how
multiple gestures or even speech can be introduced in
virtual reality text entry.</p>
      <p>Although it seems exciting to incorporate speech and
gestures in VR text entry, there are a few limitations. To
make a gesture in VR it would necessarily require a body
part to make a gesture. For example, a user wearing a
head mounted display device in VR might move his head
to make a gesture or if he is wearing a hand tracker then
he can make a gesture using his fingers. However, these
kind of head or finger interactions need users to move
their head or upper arm of the hand frequently which may
result in pain and fatigue.</p>
      <p>
        Too few works have explored gesture based text entry in
VR. Chun et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] investigated the feasibility of
headbased text entry for HMDs. In their work they used head
rotation to control a point on the keys of a virtual
keyboard. They investigated three techniques: TapType,
DwellType and GestureType. TapType resembled tap
typing in smartphones and users moved a pointer with
head rotation and selected a button by tapping a key. In
DwellType, users dwelled over a key to select that key
and in GestureType users performed a word-level input
using a gesture based typing style. They achieved the
best entry rate of 24.73 WPM in GestureType by
improving the gesture-word recognition algorithm and
incorporating head movement pattern which was recorded
during the study.
      </p>
      <p>
        While there have been a few works in mid-air text entry
using hand gestures, text entry in VR using gestures is
also rare. AirTap[
        <xref ref-type="bibr" rid="ref6">7</xref>
        ], Wilson et al. [8] and Vogel et al.[
        <xref ref-type="bibr" rid="ref7">9</xref>
        ]
used tap and pinch gestures to simulate button clicking in
virtual environment for text entry purposes. Vulture[3],
AirStroke[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Feit et al.[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] used hand gestures to enter
text in midair. The ideas described in these works can be
incorporated to enter text in VR as well.
      </p>
      <p>
        A mid-air word-gesture keyboard has been proposed in
Vulture[3]. The idea of a word-gesture keyboard is that a
user can draw a pattern formed by the letters of a word in
a touch surface rather than typing the letters. Swype,
SlideIT and ShapeWriter are examples of word-gesture
keyboards. In Vulture, the idea of implementing
wordgesture keyboard in midair instead of touch or
stylusbased surface was introduced. It used a large
highresolution display and users wore a glove with reflective
markers.
AirStroke[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Grafitti[
        <xref ref-type="bibr" rid="ref8">10</xref>
        ] are two examples of stroke
based text entry techniques. Stroke based techniques are
mainly used in stylus based interfaces. In these interfaces
users need to use the stylus to make a distinctive gesture
to mark the entry of a character.
      </p>
      <p>In comparison to the gesture based works in midair and
VR, text entry using speech has remained unexplored and
to my knowledge there has not been a single work related
to this idea. McGlashan et al.[6] investigated technical and
design issues in manipulating virtual reality with a speech
interface and proposed a prototype that provides users to
control specialized functions using speech.</p>
      <p>In conclusion, gesture and speech based text entry is still
in the budding process. Fortunately, the invention of
effective sensing devices and motion trackers (e.g. Leap
Motion Sensor, VICON tracker and HMDs etc.) is attracting
researchers and is paving the way to design and
implement new interactive systems. Hopefully, in the next
few years this line of research will flourish and we will
have better speech and gesture based text entry systems
in VR.</p>
      <sec id="sec-4-1">
        <title>Markussen, Anders, Mikkel Rønne Jakobsen, and</title>
      </sec>
      <sec id="sec-4-2">
        <title>Kasper Hornbaek. "Vulture: a mid-air word-gesture</title>
        <p>keyboard." Proceedings of the 32nd annual ACM
conference on Human factors in computing systems.</p>
        <p>ACM, 2014.</p>
      </sec>
      <sec id="sec-4-3">
        <title>McGlashan, Scott, and Tomas Axling. "A speech</title>
        <p>interface to virtual environments." In Proc.,</p>
      </sec>
      <sec id="sec-4-4">
        <title>International Workshop on Speech and Computers.</title>
        <p>1996.</p>
      </sec>
      <sec id="sec-4-5">
        <title>Wilson, Andrew D. "Robust computer vision-based</title>
        <p>detection of pinching for one and two-handed gesture
input." In Proceedings of the 19th annual ACM
symposium on User interface software and technology,
pp. 255-258. ACM, 2006.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Keith</given-names>
            <surname>Vertanen</surname>
          </string-name>
          , Haythem Memmi, Justin Emge, Shyam Reyal, and Per Ola Kristensson.
          <year>2015</year>
          .
          <article-title>VelociTap: Investigating Fast Mobile Text Entry using SentenceBased Decoding of Touchscreen Keyboard Input</article-title>
          .
          <source>In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15)</source>
          . ACM, New York, NY, USA,
          <fpage>659</fpage>
          -
          <lpage>668</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Yu</surname>
            <given-names>Chun</given-names>
          </string-name>
          , Yizheng Gu, Zhican Yang, Xin Yi, Hengliang Luo, and
          <string-name>
            <given-names>Yuanchun</given-names>
            <surname>Shi</surname>
          </string-name>
          .
          <article-title>"Tap, dwell or gesture?: Exploring head-based text entry techniques for hmds."</article-title>
          <source>In Proceedings of the 2017 CHI Conference on Human 3.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>Factors in Computing Systems</source>
          , pp.
          <fpage>4479</fpage>
          -
          <lpage>4488</lpage>
          . ACM,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Ni</surname>
            , Tao,
            <given-names>Doug</given-names>
          </string-name>
          <string-name>
            <surname>Bowman</surname>
          </string-name>
          , and Chris North.
          <article-title>"AirStroke: bringing unistroke text entry to freehand gesture interfaces</article-title>
          .
          <source>" Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Feit</surname>
            , Anna Maria, Srinath Sridhar,
            <given-names>Christian</given-names>
          </string-name>
          <string-name>
            <surname>Theobalt</surname>
            , and
            <given-names>Antti</given-names>
          </string-name>
          <string-name>
            <surname>Oulasvirta</surname>
          </string-name>
          .
          <article-title>"Investigating multi-finger gestures for mid-air text entry." Korea(</article-title>
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          7. Vogel, Daniel, and
          <string-name>
            <given-names>Ravin</given-names>
            <surname>Balakrishnan</surname>
          </string-name>
          .
          <article-title>"Distant freehand pointing and clicking on very large, high resolution displays." In Proceedings of the 18th annual ACM symposium on User interface software and technology</article-title>
          , pp.
          <fpage>33</fpage>
          -
          <lpage>42</lpage>
          . ACM,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          9.
          <string-name>
            <surname>Bowman</surname>
            ,
            <given-names>Doug A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chadwick</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Wingrave</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <string-name>
            <surname>Campbell</surname>
            ,
            <given-names>V. Q.</given-names>
          </string-name>
          <string-name>
            <surname>Ly</surname>
            , and
            <given-names>C. J.</given-names>
          </string-name>
          <string-name>
            <surname>Rhoton</surname>
          </string-name>
          .
          <article-title>"Novel uses of Pinch Gloves™ for virtual environment interaction techniques." Virtual Reality 6</article-title>
          , no.
          <issue>3</issue>
          (
          <year>2002</year>
          ):
          <fpage>122</fpage>
          -
          <lpage>129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          10.
          <string-name>
            <surname>Castellucci</surname>
            , Steven J.,
            <given-names>and I. Scott</given-names>
          </string-name>
          <string-name>
            <surname>MacKenzie</surname>
          </string-name>
          .
          <article-title>"Graffiti vs. unistrokes: an empirical comparison."</article-title>
          <source>In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems</source>
          , pp.
          <fpage>305</fpage>
          -
          <lpage>308</lpage>
          . ACM,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>