<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Revisit to the Incorporation of Context - awareness in A ective Computing Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Aggeliki Vlachostergiou</string-name>
          <email>aggelikivl@image.ntua.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefanos Kollias</string-name>
          <email>stefanos@cs.ntua.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Image Video and Multimedia Systems Laboratory School of Electrical and Computer Engineering National Technical University of Athens Iroon Polytexneiou 9</institution>
          ,
          <addr-line>15780 Zografou</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The research eld of Human Computer Interaction (HCI) is moving towards more natural, sensitive and intelligent paradigms. Especially in domains such as A ective Computing (AC), incorporating interaction context has been identi ed as one of the most important requirements towards this concept. Nevertheless, research on modeling and utilizing context in a ect-aware systems is quite scarce. In this paper, we revisit the de nition of contexts in AC systems and propose a context incorporation framework based on semantic concept detection, extraction enrichment and representation in cognitive estimation, to further clarify and assist interpretation of contextual e ects in A ective Computing systems.</p>
      </abstract>
      <kwd-group>
        <kwd>HCI</kwd>
        <kwd>A ective Computing</kwd>
        <kwd>Context</kwd>
        <kwd>Context modeling</kwd>
        <kwd>Contextaware systems</kwd>
        <kwd>SEMAINE</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        A ective Computing (AC) systems have been well developed in the past decades
as an e ective solution for recognizing, interpreting, processing, and
simulating human a ects. Additionally, context-aware A ective Computing systems [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
emerged as a novel way of shifting from context-aware A ective Computing
systems to a ective aware intelligent Human Computer Interaction systems to
overcome the fact that contextual information can not be discounted in doing
automatic analysis of human a ective behavior. Thus, the fundamental assumption
of context-aware A ective Computing systems is that context is able of shaping
how people interpret the high and complex expressions of people and machines.
The variation of these expressions in human behavior arise not only from a
subject's internal psychological or cognitive state but also from other subjects or the
environment. For instance, frowning behavior may be an indicator of anger or it
may be due to concentration depending on the contextual interactional setting.
Context-based a ect aware analysis needs to clarify the preliminary selection of
contextual variables in order to further assist intepretation of contextual e ects
in A ective Computing systems. As a result, how to incorporate contexts into
A ective Computing systems is always a research question in this domain.
However, prior to that, which variables should be considered as contexts is still under
question.
      </p>
      <p>
        Currently, several context-aware A ective Computing systems [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] have been
developed but very few research went back to discuss the de nition of contexts.
And researchers simply blend location, identities of people around the user,
environment, social interaction etc. and other variables together to consider them
as contexts, which further creates the confusion on which should be the rst
decisions to be made prior to creating context-aware automatic a ect analysis
systems. The de nition and exploration of context are not only related to the
selection of contextual variables in A ective Computing systems, but are also
relevant to the interpretation of contextual e ects based on the outcomes or
ndings in the experiments. It is obvious that the academic area focuses more
on the development of e ective context-aware A ective Computing systems, but
ignores the identi cation of contexts and interpretation of contextual e ects in
recent years.
      </p>
      <p>This paper is organized as follows: Section 2 provides an overview of how the
term context has been studied so far in various disciplines. Section 3 presents
an overview of my research and the preliminary ndings in my previous work.
Section 4 discusses my plan for achieving my overall objective for the remainder
of my doctoral work and nally, Section 5 concludes by summarizing the impact
and relevance of the proposed approaches to the eld of context-aware A ective
Computing.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Actually contexts have been studied in various disciplines, such as ubiquitous
computing, contextual advertising, social signal processing, HCI, gaming, mental
health etc., where the de nition di ers resulting in a di erent understanding of
contexts among those areas. In context-aware A ective Computing systems, the
earliest research papers [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] may bring us to look back upon almost ten years ago;
however, the eld has yet to agree on the de nition of context. Several researchers
simply blend verbal content (semantics), knowledge of the general interactional
setting, discourse and social situations and other variables together and consider
all of them as contexts, which further creates confusion on which should be the
most appropriate contextual parameter w.r.t context-aware A ective Computing
systems (Figure 1).
      </p>
      <p>
        The most commonly used de nition is the one given by Abowd et al. in 1999
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], \context is any information that can be used to characterize the situation
of an entity. An entity is a person, place, or object that is considered relevant
to the interaction between a user and an application, including the user and
applications themselves." This de nition hardly limits the set of variables that
can be taken into account, and it is still ambiguous without clear guidelines to
select appropriate variables in AC systems.
      </p>
      <p>
        Apparently, the de nition and selection of contexts is a domain-speci c
problem, where classi cation of contextual variables is a typical way to put di erent
variables in categories but it is still not general enough and not exible in
interpreting the contextual e ects. The debate or discussion may be nally ended
by the idea proposed by Duric et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which is known as W5+ formalization
and incorporates context corresponing to the answering of the following
important aspects of context: Who you are with (e.g. dyadic/multiparty interactions),
What is communicated (e.g., (non)-linguistic message), How the information is
communicated (the persons a ective cues), Why, i.e., in which context the
information is passed on, Where the user is, What his current task is, How he/she
feels (has his mood been polarized changing from negative to positive?) and
which (re)action should be taken to satisfy humans needs).
      </p>
      <p>
        Unfortunately, so far the e orts on human a ective behavior understanding
are usually context independent [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In light of these observations, understanding
the process of a natural progression of context-related questions when people
interact in a social environment could provide new insights into the mechanisms of
their interaction context and a ectivity. The Who, What, Where context-related
questions have been mainly answered either separately or in groups of two or
three using the information extracted from multimodal input streams [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Thus,
as of date, no general W5+ formalization exist, due to the fact that current
systems which answer to most of the W questions are founded on di erent
psychological theories of emotion. Recent research on progressing to the questions of
\Why" and \How" has led to the emerging eld of sentiment analysis, through
mining opinions and sentiments from natural language, which involves a deep
understanding of semantic rules proper of a language.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Preliminary Work</title>
      <p>The key objective of my PhD research is to computationally identify,
automatically extract and incorporate contextual information into a ect aware
recognition frameworks, with the aim of identifying context-aware emotional-speci c
patterns.
3.1</p>
      <sec id="sec-3-1">
        <title>Context identi cation Framework</title>
        <p>
          To ful ll the need of understanding whether and how context is incorporated
in automatic analysis of human a ective behavior, we propose a novel
contextaware incorporation framework (Fig. 2) which (i): includes detection and
extraction of semantic context concepts, (ii): better enriches a number of Psychological
Foundations with sentiment values and (iii): enhances emotional models with
context information and context concept representation in appraisal estimation
[
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
        <p>
          As a rst step, our preliminary results are focused on bridging the gap at
concept level by exploiting semantics, cognitive and a ective information,
associated with the image verbal content (semantics), which for the needs of our
research is the contextual interactional information between the user and the
operator of the SEMAINE database [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], keeping xed the \Where" context-related
question. This context concept-based annotation method, that we are
examining, allows the system to go beyond a mere syntactic analysis of the semantics
associated with xed window sizes1. In most of traditional annotation methods,
emotions and contextual information are not always inferred by appraisals and
thus contextual information about the causes of that emotion is not taken into
account [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Approach</title>
        <p>
          In this section, our preliminary results of our proposed semantic context concept
extractor, are described in details in [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], and their application to indicative
examples validate our proposed approach are presented:
A. Data Corpus: The model here is confronted with the SEMAINE corpus [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
This audiovisual corpus comprises manually-transcribed sessions with natural
emotional displays. These sessions are recordered from two individuals, an
operator and a user, interacting through teleprompter screens from two di erent
rooms. The emotions were elicited with the sensitive arti cial listener (SAL)
framework, where the operator assumes four personalities aiming to elicit
positive and negative emotional reactions from the user. Agent's utterances are
constrained by a script, however, some deviations to the script occur in the
database.
1 The window length corresponds to 16 conversational turns and is displayed on gures
for future visualization purposes.
        </p>
        <p>B. Pre-Processing : The pre-processing submodule rstly interprets all the a
ective valence indicators usually contained in the verbal content of transcriptions,
such as special punctuation, complete upper-case words, exclamation words and
negations. Handling negation is an important concern in such scenario, as it
can reverse the meaning of the examined sentence. Secondly, it converts text
to lower-case and, after lemmatizing it, splits the sentence into single clauses
according to grammatical conjunctions and punctuation.</p>
        <p>
          These n-grams are not used blindly as xed word patterns but are exploited
as reference for the module, in order to extract multiple-word concepts from
information-rich sentences. So, di erently from other shallow parsers, the
module can recognize complex concepts also when irregular verbs are used or when
these are interspersed with adjectives and adverbs, for example, the concept \buy
easter present" in the sentence \I bought a lot of very nice Easter presents".
C. Semantic context concept parser: The aim of the semantic parser is to break
sentences into clauses and, hence, deconstruct such clauses into concepts. This
deconstruction uses lexicons which are based on sequences of lexemes that
represent multiple-word concepts extracted from ConceptNet, WordNet [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] and other
linguistic resources.
        </p>
        <p>Under this view, the Stanford Parser 2 has been used according to Python
NLTK 3; a general assumption during clause separation is that, if a piece of text
contains a preposition or subordinating conjunction, the words preceding this
function or are interpreted not as events but as objects. Secondly, dependency
2 http://nlp.stanford.edu:8080/parser/
3 http://nltk.org
structure elements are processed by means of Stanford Lemmatizer for each
sentence. Each potential noun chunk associated with individual verb chunks is
paired with the stemmed verb in order to detect multi-word expressions of the
form \verb plus object". The pos-based bigram algorithm extracts concepts, but
in order to capture event concepts, matches between the object concepts and the
normalized verb chunks are searched. It is important to build the dependency
tree before lemmatization as swapping the two steps result in several
imprecisions caused by the lower grammatical accuracy of lemmatized sentences. Each
verb and its associated noun phrase are considered in turn, and of more concepts
is extracted from these.</p>
        <p>
          D. Opinion and Sentiment Lexicon: Current approaches to concept-level
sentiment analysis mainly leverage on existing a ective knowledge bases such as
ANEW, WordNet-A ect and SentiWordNet [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. However, for the needs of our
current work, we use the SentiWordNet, which is a concept-level opinion lexicon
and contains multi-word expressions labeled by their polarity scores. 4
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Future Research</title>
      <p>My future work will focus on the integration of previous preliminary ndings and
insights on context a ect aware methodologies w.r.t. the production,
intepretation and analysis level respectively. To achieve this goal, we have to deal with
the following speci c challenges:
1) Extraction and Representation: How can we extract and computationally
measure contextually rich features w.r.t. the verbal content?
2) Learning : What are the proper learning methods needed to build models?
3) Incorporation: At what time unit should we make an emotion inference (e.g.
at the frame or utterance level) and how should we measure the performance of
our system?
A. Extraction and Representation</p>
      <p>
        In A ective Computing, all of the state-of-the-art algorithms perform well on
individual sentences without considering any context information, but their
accuracy is dramatically lowered because they fail to consider context and the
syntactic structure of the verbal content (transcriptions) at the same time. Based on
the assumption that the context around a sentence or pair of sentences also plays
an important role in determining sentiment, I plan to employ in my experiments
a conditional random eld (CRF) model [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to capture syntactic, structural and
contextual features of sentences. To this aim, I propose to utilize the SEMAINE
and RECOLA datasets [
        <xref ref-type="bibr" rid="ref10 ref8">8, 10</xref>
        ] for both human-agent and human-human social
4 To avoid the Sentiwordnet's multi-interpretations a combination of the following
methods have been examined: a) Pos tagging to reduce the number of candidate
senses, b) Cosine similarity between the sentence and the gloss of each sense of
the word in WordNet and c) the \SenseRelate" method to measure the \WordNet
similarity" between di erent senses of the target word and its surrounding words.
and situational interactions5, as they provide continuous-time dimensional labels
(valence-activation dimensional space).
      </p>
      <p>B. Learning</p>
      <p>
        In the learning stage, I plan to use discriminative learning methods and learn
patterns of contextual emotion-speci c segments in a supervised manner. My
hypotesis is that training contextual rich classi cation systems using segments
lacking clarity may lead to lower analysis rates, and using only segments with
high clarity will lead to improved performance as well as more e cient training
due to the decrease in training data. I will rst extract contextual rich
audiovisual descriptors of the above mentioned emotional corpora [
        <xref ref-type="bibr" rid="ref10 ref8">8, 10</xref>
        ] and then I
will utilize these \socially contextually rich" segments with consistent emotional
cues and their emotional labels to train a discriminative model. In the case of
the RECOLA dataset, in which only the rst 5 minutes have been annotated, I
will use the active learning method to build classi ers using less tarining data
for expanding labeled data pool.
      </p>
      <p>C. Incorporation</p>
      <p>Finally, the goal at the interpretation stage is to show that incorporating
context can both improve the system's performance and disambiguate multimodal
behaviors. To this aim, I propose to rst identify \socially contextually rich"
segments of the test data and incorporate emotion at the segment level using
the learned weights of discriminative models. I plan to conduct both classi
cation and regression tasks. I hypothesize that the classi cation of the \socially
contextually rich" segments can increase the regression of all the data, since
the captured and longer-range information may be more useful. I will predict
continuouas-valued labels in dimensional space using regression models such as
Support Vector Regression. Also, I will cluster and de ne new emotion classes
from the dimensional labels and identify classes using Support Vector Machines
(SVM). The SVM outputs will be combined to infer dimensional space outputs.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this paper, we point out the motivation and importance of identifying or
de ning the incorporation of context in A ective Computing systems, especially
when the disambiguation of a multimodal behavior depends on the contextual
interactional setting. Afterwards, we propose the context incorporation framework
based on semantic concept detection, extraction enrichment and representation
in cognitive estimation. And nally, we provide relevant analysis and conclude to
future work that needs to be developed. In future work, we would like to explore
the interpretation of contextual e ects on a ect production, interpretation and
analysis respectively.
5 In the SEMAINE corpus, situation is determined by the user's response while in
RECOLA corpus the situational context is de ned by the roles during the
collaboratorive, intensive and interactional task.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This research has been co- nanced by the European Union (European Social
Fund - EFS) and Greek national funds through the Operational Program
\Education and Lifelong Learning" of the National Strategic Reference Framework
(NSRF) - Research Funding Program: Thales. Investing in knowledge society
through the European Social Fund.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Abowd</surname>
            ,
            <given-names>G.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dey</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brown</surname>
          </string-name>
          , P.J.,
          <string-name>
            <surname>Davies</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smith</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steggles</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Towards a better understanding of context and context-awareness</article-title>
          .
          <source>In: Handheld and Ubiquitous Computing</source>
          . pp.
          <volume>304</volume>
          {
          <fpage>307</fpage>
          . Springer (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Dey</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          :
          <article-title>Context-aware computing: The cyberdesk project</article-title>
          .
          <source>aaai 1998 spring symposium on intelligent environments</source>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Duric</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>W.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heishman</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosenfeld</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schoelles</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schunn</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wechsler</surname>
          </string-name>
          , H.:
          <article-title>Integrating perceptual and cognitive modeling for adaptive and intelligent human-computer interaction</article-title>
          .
          <source>Proceedings of the IEEE</source>
          <volume>90</volume>
          (
          <issue>7</issue>
          ),
          <volume>1272</volume>
          {
          <fpage>1289</fpage>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Esuli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sebastiani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Sentiwordnet: A publicly available lexical resource for opinion mining</article-title>
          .
          <source>In: Proceedings of the 5th Conference on Language Resources and Evaluation (LREC06)</source>
          . vol.
          <volume>6</volume>
          , pp.
          <volume>417</volume>
          {
          <issue>422</issue>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Graves</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fernandez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidhuber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Bidirectional lstm networks for improved phoneme classi cation and recognition</article-title>
          .
          <source>In: Proceedings of the 15th International Conference on Arti cial Neural Networks: Formal Models and Their</source>
          Applications - Volume
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , pp.
          <volume>799</volume>
          {
          <fpage>804</fpage>
          . ICANN'
          <volume>05</volume>
          ,
          <string-name>
            <surname>Springer</surname>
          </string-name>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Gunes</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schuller</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Categorical and dimensional a ect analysis in continuous input: Current trends and future directions</article-title>
          .
          <source>Image and Vision Computing</source>
          <volume>31</volume>
          (
          <issue>2</issue>
          ),
          <volume>120</volume>
          {
          <fpage>136</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>La</surname>
            <given-names>erty</given-names>
          </string-name>
          , J.D.,
          <string-name>
            <surname>McCallum</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira</surname>
            ,
            <given-names>F.C.N.</given-names>
          </string-name>
          :
          <article-title>Conditional random elds: Probabilistic models for segmenting and labeling sequence data</article-title>
          .
          <source>In: Proceedings of the Eighteenth International Conference on Machine Learning</source>
          . pp.
          <volume>282</volume>
          {
          <fpage>289</fpage>
          . ICML '
          <fpage>01</fpage>
          , Morgan Kaufmann Publishers Inc. (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>McKeown</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valstar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cowie</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pantic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schroder</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The semaine database: annotated multimodal records of emotionally colored conversations between a person and a limited agent</article-title>
          .
          <source>IEEE Transactions on A ective Computing</source>
          <volume>3</volume>
          (
          <issue>1</issue>
          ),
          <volume>5</volume>
          {
          <fpage>17</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          :
          <article-title>Wordnet: a lexical database for english</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>38</volume>
          (
          <issue>11</issue>
          ),
          <volume>39</volume>
          {
          <fpage>41</fpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Ringeval</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sonderegger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sauer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lalanne</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Introducing the recola multimodal corpus of remote collaborative and a ective interactions</article-title>
          .
          <source>In: 10th International Conference on Automatic Face and Gesture Recognition (FG)</source>
          . pp.
          <volume>1</volume>
          {
          <issue>8</issue>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Vlachostergiou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caridakis</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raouzaiou</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kollias</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Hci and natural progression of context-related questions</article-title>
          .
          <source>In: Human-Computer Interaction: Design and Evaluation</source>
          , pp.
          <volume>530</volume>
          {
          <fpage>541</fpage>
          . Springer (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>