<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Pointing to an invisible object behind a wall: Comprehension of pointing with a bent index finger</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tetsuya YASUDA (t-yasuda@jumonji-u.ac.jp)</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Human Developmental Psychology, Jumonji University Sugasawa Niiza</institution>
          ,
          <addr-line>Saitama, 352-8510</addr-line>
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Harumi KOBAYASHI</institution>
        </aff>
      </contrib-group>
      <fpage>477</fpage>
      <lpage>481</lpage>
      <abstract>
        <p>Pointing is a gesture that people use to specify and convey information about objects in the environment. Previous research has mostly explored peoples' comprehension and production of others' “straight” pointing gestures, that is, pointing at an object with arm and index finger kept straight while aiming at a visible object. However, we seem to use various types of pointing in addition to typical straight pointing to denote both visible and invisible objects. This study examined comprehension of pointing with a “bent” index finger at an invisible object behind a wall. The experimenter pointed either at an object in front of a wall or one behind a wall with a straight index finger or a bent index finger, and the participants guessed which object was being denoted. The results were that when the participants looked at straight pointing, they thought objects in front of the wall were being denoted. However, when they looked at bent pointing, they thought objects behind the wall could be denoted. The study suggested that people have “common ground” in terms of interpretation of different types of pointing.</p>
      </abstract>
      <kwd-group>
        <kwd>gesture</kwd>
        <kwd>declarative pointing</kwd>
        <kwd>common ground</kwd>
        <kwd>non-linguistic information</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        When we communicate with others, we often draw their
attention to objects about which we wish to communicate
        <xref ref-type="bibr" rid="ref10">(Tomasello, 2008; 2014)</xref>
        . Pointing is a gesture that people
use to specify and convey information about referents. For
example, when one asks a friend about a landmark in the
city, she will point at the landmark using her index finger or
her hand.
      </p>
      <p>
        <xref ref-type="bibr" rid="ref3">Clark (2003)</xref>
        discussed use of attention-getting gestures
in various cases. He noted that pointing at a referent and
placing a referent are both useful ways to convey
information about referents, but that people index objects
differently. In pointing, a person directs the addressee’s
attention to the referent object; for example, a customer may
point at a package of a medicine that is difficult for her to
reach but is easy for the clerk. In placing, a person puts a
referent object in the area of an addressee’s attention; for
example, a customer may place a package of medicine on
the checkout counter where a clerk waits. These
communications are possible without saying any words. In
order to communicate smoothly, people must share mutual
understanding of pointing at referents and placing referents
in different situations.
        <xref ref-type="bibr" rid="ref4">Clark, Schreuder, &amp; Buttrick (1983</xref>
        )
and
        <xref ref-type="bibr" rid="ref2">Clark (1996)</xref>
        proposed that people use “common ground”
as implicit mutual knowledge in human communication.
Common ground can include a variety of information: how
people convey information using words and sentences,
knowledge about the history and culture of the speakers and
addressees, mutually shared knowledge about specific
people and events, and knowledge of what is going on in the
current communication. Thus, common ground includes
both general knowledge about the world and specific
knowledge of the specific task that the conversation partners
are conducting. Previous research has focused mostly on
language and verbally describable information included in
common ground. Non-verbal information such as gestures
must also be comprehended using common ground as to
how people use gestures in different situations; however,
usage of gestures as common ground has not yet been
thoroughly explored.
      </p>
      <p>
        Some research has explored peoples’ comprehension and
production of pointing gestures when they use some
language such as demonstratives such as “this” or “that”
while indicating objects in the environment
        <xref ref-type="bibr" rid="ref1 ref5 ref6">(Bangerter,
2004; Coventry, Valdés, Castillo, &amp; Guijarro-Fuentes,
2008; Coventry, Griffiths, &amp; Hamilton, 2014)</xref>
        . Most of
pointing studies have examined typical pointing gestures,
we named it “straight” pointing because pointing at an
object is done with the addresser’s arm and index finger
kept relatively straight while aiming at a visible object in the
environment
        <xref ref-type="bibr" rid="ref6 ref8 ref9">(Coventry et al., 2008; Doherty, Anderson, &amp;
Howieson, 2004; Jaswal &amp; Hansen, 2006; Kobayashi, 2007)</xref>
        .
In this situation, the addresser can easily share information
about visible objects, using visual joint attention and
common ground.
      </p>
      <p>How, then, can we point at invisible objects such as
objects behind obstacles? In the authors’ lab, we observed a
person pointing at a magnet pin that was attached on the
other side of a steel board. The addresser bent his index
finger when he pointed at the invisible, but known, magnet
pin. We observers immediately understood the meaning of
his pointing gesture, although pointing with a bent index
finger seems relatively unusual. We might have common
ground with regard to non-verbal gestures, or at least
knowledge about how we should interpret others’ various,
occasionally unusual, gestures. In the case of referring to
invisible objects, sharing information about the referent may
be more difficult for both the addresser and the addressee
because visual joint attention is difficult to establish. The
role of common ground in human non-verbal gesturing
seems to be more important when people refer to invisible
objects.</p>
      <p>We examined people’s common ground regarding
pointing gestures. This study focused on comprehension of
pointing with a bent index finger at an invisible object
behind a panel. Because this is the first study to examine the
role of a bent index finger, we did not examine the
possibility that the general posture of the arm and the index
finger as a whole may have a role in this study. The
experimenter pointed at an object in front of a panel or
behind a panel with a straight index finger or a bent index
finger. There were objects either in front of or behind the
panel, and the participants guessed which object was being
indicated. We expected that if the experimenter used a bent
index finger in pointing, the participant would interpret this
as referring to the object behind the panel, but if he used a
straight index finger in pointing, the participant would
interpret this as referring to the object in front of the panel.
The reason is that the bent index finger seems to suggest
that the “pointing trajectory” (imaginary trajectory of
pointing gesture) can go over the panel.</p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <sec id="sec-2-1">
        <title>Participants</title>
        <p>Twenty Japanese undergraduate university students (M
age = 21.2 years; 3 females) participated. The experiment
was conducted in accordance with Tokyo Denki
University’s code of ethics.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Procedure</title>
        <p>The experimental conditions consisted of two types of
pointing (straight vs. bent) and obstacle placement (with vs.
without).</p>
        <p>With regard to pointing condition, “straight pointing” was
when the experimenter pointed at the referent with her arm
extended horizontally and her index finger kept straight (Fig.
1a), “Bent pointing” was when the experimenter pointed at
the referent with her arm extended slantwise and her index
finger kept somewhat bent (Fig. 1b). The “with-obstacle”
condition was when there was a small black opaque panel
(W: 25 cm x H: 40 cm) on the table. “Without-obstacle”
was when there was no panel on the table.</p>
        <p>Fig.2 shows the experimental setup. On the table, there
were 4 small bottles (W: 2.3 cm × H: 8 cm) designated 1, 2,
3, and 4, respectively, on a label of each bottle. The
experimenter sat on one side of the table, wearing black
sunglasses during the experiment so that participants could
not see the experimenter’s gaze direction. The participant
sat at the table at a right angle to the experimenter. In the
with-obstacle condition, the panel was placed in the middle
of the table between object #2 and object #3. Participants
were randomly assigned to all conditions.</p>
        <p>Fig. 2: Experimental setup in the with-obstacle condition.</p>
        <p>Each object was placed 10 cm apart from the adjacent bottle.</p>
        <p>First, the experimenter and the participant looked at all
the bottles placed on the table. Each bottle was placed 10
cm apart from the adjacent bottle. Bottle #1 was placed 30
cm away from the edge of the side of the table where the
experimenter sat. Then, the participant sat on the
experimenter’s chair and looked at the table. Then, the
experimenter put the panel in between bottles #2 and #3,
and the participant again looked at the table. Thus, the
participant experienced the experimenter’s view in both
with the obstacle and without the obstacle conditions (Fig.
3). Then, in the straight pointing with the obstacle condition,
the experimenter put the panel in between bottles #2 and #3
and said to the participant, “I cannot see bottles #3 and #4.</p>
        <p>Now, I will point at one of the four bottles.” Then, the
experimenter pointed at bottles using either with the straight
index finger or the bent index finger. With each pointing
gesture, the experimenter said, “Now I am pointing at
something. Which bottle would you guess I am pointing at?
Please answer with the number of the bottle.” The
participant responded orally using the bottle number. The
bottle number corresponded to the distance from the edge of
the table: Bottle #1’s distance was 30 cm; #2, 40 cm; #3, 50
cm; and #4, 60 cm, respectively. In addition, the bottle
numbers corresponded to visible or invisible status within
the with-obstacle condition: bottles #1 and #2 were visible,</p>
        <p>Fig. 3: Experimenter’s view during the experiment in the without-obstacle condition (left) and in the with-obstacle
condition (right). Before the experiment, each participant first sat on the experimenter’s chair and looked at the table with
and without the obstacle to know the experimenter’s views of the both conditions.
and bottles #3 and #4 were invisible. Consequently, the
participants answered using a scale of 1, 2, 3, and 4. In the
without-obstacle condition, all four bottles were visible.</p>
        <p>Because the experimenter wore sunglasses, the participant
could not see the experimenter’s eye gaze.</p>
        <p>In the bent-pointing with-obstacle condition, the
procedure was the same as with the straight pointing
withobstacle condition except that bent pointing was used. In the
without-obstacle conditions, after both the experimenter and
the participant looked at the four bottles, the experimenter
pointed at a bottle in random order, and the participant
guessed which bottle was being pointed at.</p>
        <p>There were two pointing conditions (straight and bent),
and in each pointing condition, there were two obstacle
conditions (with obstacle, without obstacle); in each
pointing and obstacle combination, there were four bottle
(distance) trials. There were 4 blocks in the order of
pointing, straight-pointing and with-obstacle,
straightpointing and without-obstacle, bent-pointing and
withobstacle, bent-pointing and without-obstacle. In each block,
the order of the bottle was random, and there were totally 16
trials. Overall, the order of these blocks was
counterbalanced between the participants.</p>
        <p>The experimenter was trained to show the same pointing
gesture in either the bent or the straight conditions in the
aspects of speed of the movement, the height of the wrist,
and the distance from the participant’s body.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <p>Fig. 4 shows the participant’s responses when the
experimenter pointed at each object in each condition. A 2
(Pointing: straight, bent) × 2 (Obstacle: with, without) × 4
(Referent: #1, #2, #3, #4) ANOVA was performed with the
number of the bottle that the participant responded as the
dependent measure. There were significant main effects of
Pointing, F(1,19) = 78.042, p &lt; .001, ηp2= 0.804; Obstacle,
F(1,19) = 6.163, p &lt; .01, ηp2= 0.245; and Distance, F(3,57)
= 160.457, p &lt; .001, ηp2= 0.894. There were also significant
interactions of Pointing × Obstacle, F(1,19) = 6.935, p
&lt; .005, ηp2 = 0.300, and Pointing × Distance, F(3,57) =
6.935, p &lt; .005, ηp2 = 0.148.</p>
      <p>To explore the significant Pointing × Obstacle interaction,
the simple main effects of Pointing within each Obstacle
condition and the simple main effects of Obstacle within
each Pointing condition were analyzed. Pointing differences
in pointing with the obstacle (F(1,38) = 34.139, p &lt; .001, ηp2
= 0.473.) and without the obstacle (F(1,38) = 80.265, p
&lt; .001, ηp2 = 0.679.) were significant. Obstacle differences
in the straight pointing were significant, F(1,39) = 13.023, p
&lt; .001, ηp2 = .255. To explore the significant Obstacle ×
Distance interaction, the simple main effects of Pointing
within each Obstacle condition and the simple main effects
of Obstacle within each Pointing condition were analyzed.
Obstacle difference was significant for object #4, F(1,76) =
12.552, p &lt; .001, ηp2 = 0.142. Distance differences in the
straight and the bent pointing conditions were significant,
except when the experimenter pointed at objects #3 and #4
in the without-obstacle condition (p &lt; .05).</p>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>This study examined how people interpret the
experimenter’s pointing with a bent index finger at an
invisible object behind a panel. The experimenter pointed at
bottles that were placed either in front of the panel or behind
the panel using a straight index finger or a bent index finger,
and the participants guessed which object was being
indicated. The results were that in the with-obstacle
condition, straight pointing tended to be interpreted as
referring to objects #1 and #2, whereas in the
withoutobstacle condition, straight pointing tended to be interpreted
as referring to all objects, including objects #3 and #4.</p>
      <p>However, interestingly, the situation was different when
bent pointing was used. In the with-obstacle condition, bent
pointing tended to be interpreted as referring to all the
60 cm, respectively. The bottle numbers also correspond to visible or invisible status in the with-obstacle condition:
bottles #1 and #2 were visible, and bottles #3 and #4 were invisible in this condition. In the without-obstacle condition, all
four bottles were visible. Y-axis shows the number of the bottle the participant responded.
objects, including #3 and #4, and there was no difference
between the with-obstacle condition and the
withoutobstacle condition. Thus, the results indicate that
participants interpret the straight pointing as referring to all
objects when the panel was not present and the objects in
front of the panel when the panel was present. In contrast,
they tend to think the bent pointing as referring to all
objects irrespective of the presence or absence of the panel.</p>
      <p>The current experiment did not disentangle if the effect
could be due to the bent index finger or to the general
posture of the arm and the index finger as a whole. In future
research, the roles of the bent index finger per se and it and
the arm as a whole must be examined. In addition, the
current experiment did not perfectly control the speed of the
pointing gesture. The speed of pointing may have an effect
in estimating the distance of the “imaginary trajectory.”
Future research must examine this issue.</p>
      <p>In conclusion, the study showed that people could
interpret pointing at an invisible object when bent index
finger was used in pointing. It suggests that people know
the meaning of the “bent” index finger based on “common
ground” in their interpretation of different types of pointing.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>We would like to thank all participants who participated
in our experiment. We thank Kota Kuwayama for data
collection. This study was partially supported by the Japan
Society for the Promotion of Science (JSPS): Grant-in-Aid
for Scientific Research (C) #24530793 (H.K.), and
Grant-inAid for Young Scientists (B) # 26870549 (T.Y.).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bangerter</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2004</year>
          ).
          <article-title>Using pointing and describing to achieve joint focus of attention in dialogue</article-title>
          .
          <source>Psychological Science</source>
          ,
          <volume>15</volume>
          ,
          <fpage>415</fpage>
          -
          <lpage>419</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>H. H.</given-names>
          </string-name>
          (
          <year>1996</year>
          ).
          <article-title>Using language</article-title>
          . Cambridge: Cambridge University Press.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>H. H.</given-names>
          </string-name>
          (
          <year>2003</year>
          ).
          <article-title>Pointing and placing</article-title>
          . In S. Kita (Ed.). Pointing:
          <article-title>Where language, culture</article-title>
          , and cognition meet (pp.
          <fpage>243</fpage>
          -
          <lpage>268</lpage>
          ). Mahwah, NJ: Lawrence Erlbaum.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>H. H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schreuder</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Buttrick</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>1983</year>
          ).
          <article-title>Common ground at the understanding of demonstrative reference</article-title>
          .
          <source>Journal of Verbal Learning and Verbal Behavior</source>
          ,
          <volume>22</volume>
          ,
          <fpage>245</fpage>
          -
          <lpage>258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Coventry</surname>
            ,
            <given-names>K. R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Griffiths</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Hamilton</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>Spatial demonstratives and perceptual space: Describing and remembering object location</article-title>
          .
          <source>Cognitive Psychology</source>
          ,
          <volume>69</volume>
          ,
          <fpage>46</fpage>
          -
          <lpage>70</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Coventry</surname>
            ,
            <given-names>K. R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valdés</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castillo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>GuijarroFuentes</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>Language within your reach. Nearfar perceptual space and spatial demonstratives</article-title>
          .
          <source>Cognition</source>
          ,
          <volume>108</volume>
          ,
          <fpage>889</fpage>
          -
          <lpage>895</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Doherty</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Howieson</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>The rapid development of explicit gaze judgment ability at 3 years</article-title>
          .
          <source>Journal of Experimental Child Psychology</source>
          ,
          <volume>104</volume>
          ,
          <fpage>296</fpage>
          -
          <lpage>312</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Jaswal</surname>
            ,
            <given-names>V. K.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Hansen</surname>
            ,
            <given-names>M. B.</given-names>
          </string-name>
          (
          <year>2006</year>
          ).
          <article-title>Learning words: children disregard some pragmatic information that conflicts with mutual exclusivity</article-title>
          .
          <source>Developmental Science</source>
          ,
          <volume>9</volume>
          ,
          <fpage>158</fpage>
          -
          <lpage>165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Kobayashi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>The effect of touching object parts on learning novel object part names among young children and adults</article-title>
          .
          <source>Studies in Language Sciences</source>
          <volume>6</volume>
          ,
          <fpage>61</fpage>
          -
          <lpage>76</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Tomasello</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>Origins of human communication</article-title>
          . Cambridge: MIT press.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Tomasello</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>A Natural History of Human Thinking</article-title>
          . Harvard University Press.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>