<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>CruzAffect at AffCon 2019 Shared Task: A feature-rich approach to characterize happiness</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jiaqi Wu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ryan Compton</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Geetanjali Rakshit</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marilyn Walker</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pranav Anand</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Steve Whittaker</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>UC Santa Cruz</institution>
          ,
          <addr-line>1156 High Street, Santa Cruz 95064, California</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We present our system, CruzAffect, for the CL-Aff Shared Task 2019. Cruz-Affect consists of several types of robust and efficient models for affective classification tasks. We utilize both traditional classifiers, such as XGBoosted Forest, as well as a deep learning Convolutional Neural Networks (CNN) classifier. We explore rich feature sets such as syntactic features, emotional features, and profile features, and utilize several sentiment lexicons, to discover essential indicators of social involvement and control that a subject might exercise in their happy moments, as described in textual snippets from the HappyDB database. The data comes with a labeled set (10K), and a larger unlabeled set (70K). We therefore use supervised methods on the 10K dataset, and a bootstrapped semisupervised approach for the 70K. We evaluate these models for binary classification of agency and social labels (Task 1), as well as multi-class prediction for concepts labels (Task 2). We obtain promising results on the held-out data, suggesting that the proposed feature sets effectively represent the data for affective classification tasks. We also build concepts models that discover general themes recurring in happy moments. Our results indicate that generic characteristics are shared between the classes of agency, social and concepts, suggesting it should be possible to build general models for affective classification tasks.</p>
      </abstract>
      <kwd-group>
        <kwd>affective classification well being theory social connections</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The overall goal of the CL-Aff Shared Task [4] is to understand what makes people
happy, and the factors contributing towards such happy moments. Related work has
centered around understanding and building lexicons that focus on emotional
expressions [5, 9], while Reed et al. [7] learn lexico-functional linguistic patterns as reliable
predictors for first-person affect, and constructed a First-Person Sentiment Corpus of
positive and negative first-person sentences from blog journal entries. Wu et al. [12]
propose a synthetic categorization of different sources for well-being and happiness
targeting the private micro-blogs in Echo, where users rate their daily events from 1
to 9. These work aim to identify specific compositional semantics that characterize the
sentiment of events, and attempt to model happiness at a higher level of generalization,
however finding generic characteristics for modeling well-being remains challenging.
In this paper, we aim to find generic characteristics shared between different affective
classification tasks. Our approach is to compare state-of-the-art methods for linguistic
modeling to prior lexicons’ predictive power. While this body of work is broader in
scope than the goals we are trying to address, they do include annotated sets of words
associated with happiness as well as additional categories of psychological significance.</p>
      <p>The aim of this work is to address the two tasks that are part of the CL-Aff Shared
Task. The data provided for this task comes from the HappyDB dataset [1]. Task 1
focuses on binary prediction of two different labels, social and agency. The intention is to
understand the context surrounding happy moments and potentially find factors
associated with these two labels. Task 2 is fairly open-ended, leaving it to the participant’s
imagination to model happiness and derive insights from their models. Here, we predict
the concepts label using multi-class classification. We explore various approaches to
determine which models work best to characterize contextual aspects of happy moments.
Though the predictions of agency and social sound simpler than concepts, we expect
that the best models for agency and social prediction could generate similarly optimal
performance for concepts, assuming that the classes of social, agency, and concepts
share common characteristics. To validate our assumptions, we build different models
for general affective classification tasks and then try to gain a deeper understanding
of the characteristics of happy moments by interpreting such models with the Riloff’s
Autoslog linguistic-pattern learner [8, 12].
2</p>
    </sec>
    <sec id="sec-2">
      <title>Agency and Social Classification</title>
      <p>
        This work utilizes a bootstrapping approach to conduct semi-supervised learning
experiments. This involves a three-step procedure: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) train a model on the labeled data;
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) use the trained model to make predictions on the unlabeled data; and (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) train a
new model using the combination of the labeled data and the predictions on the
unlabeled data. Training each model involves a 10-fold cross-validation to evaluate the
performance, while guaranteeing that the test set for each fold consists of gold-standard
hand-labelled instances.
2.1
      </p>
      <sec id="sec-2-1">
        <title>Feature Extraction</title>
        <p>We explore different features to find those most informative for the prediction task.
We aim to understand how syntactic features and emotional features compare to word
embeddings, and whether the profile features improve the prediction results.</p>
        <p>Syntactic Features: Our syntactic features are limited to Part of Speech (POS)
tagging, by applying a POS tagger to count the relative frequencies of syntactic nouns,
verbs, adjectives and adverbs, use of questions as well as tense and aspect information
[10]. There are 36 POS features.</p>
        <p>Emotional Features: We use 4 different types of emotional features. LIWC v2007
[9] is a lexicon providing frequency counts of words indexing important psychological
constructs, as well as relevant topics (Leisure, Work). The Emotion Lexicon (EmoLex)
[5] contains 14,182 words classified into 10 emotional categories: Anger, Anticipation,
Disgust, Fear, Joy, Negative, Positive, Sadness, Surprise, and Trust. The Subjectivity
Lexicon is part of OpinionFinder [11]. It consists of 8222 stemmed and unstemmed
words, annotated by a group of trained annotators as either strongly or weakly
subjective. Our last feature is our own regression model from prior work on predicting the
level of factual and emotion language. Details about this model can be found in [3].
There are 94 features in total.</p>
        <p>Word Embedding: We utilize GloVe [6] 100 dimension word vectors for word
representation. GloVe is expected to encode distributional aspects of meaning.</p>
        <p>Profile Features: The corpus include demographic features collected via a survey:
age, country, gender, married, parenthood, reflection, and duration. To reduce sparsity,
we convert the country feature into language feature, assuming that the people who
speak the same language might share similar culture, and thus similar happy moments.
After this conversion, we have 48 different languages from the 70 countries, the largest
group is English (80:3%), then Hindi (15:9%), corresponding to 79.8% examples from
USA and 15.87% examples from IND. We also bin age into groups, assuming that
different age groups would have different general happy moments. The age groups,
illustrated in Table 1, include kid (age &lt; 10), teenager (age &lt; 18), youth (age &lt; 24),
young adult (age &lt; 40), middle age (age &lt; 65) and elderly (age &gt;= 65). There are 70
features after the feature preprocessing. We aim to test whether the features extracted
from text are sufficient for affective sentiment analysis, and whether the profile features
improve performance.
In this section, we compare the performance of traditional machine learning methods,
such as Logistic Regression and XGBoosted Random Forest [2] and a Convolutional
Neural Network (CNN) model [13].</p>
        <p>
          Supervised Learning. For modeling the profile features, we apply logistic regression
with a liblinear solver and balanced class weights. For modeling the syntactic features
and emotional features, we use XGBoosted Random Forest with out-of-the-box values
for all parameters except for the following: 250 number of estimators, a learning rate of
0.05, and a maximum tree depth of 6. For the CNN model with word embedding, we
explore its performance with different parameter settings. The best hyperparameters of
the CNN model include filter size 3, multiple region size (
          <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
          ) and max pooling size
1, or filter size 4, multiple region size (
          <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2, 3, 4, 5</xref>
          ) and max pooling size 1. The region
size implies a windows size for N-grams. After getting the best hyperparameters, we
train the model with word embeddings, and word embeddings concatenated with
syntactic and emotional features, to test whether syntactic and emotional features improve
performance. Figure 1 illustrates the CNN model with region size (
          <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
          ).
        </p>
        <p>Table 2 shows the 10-fold cross-validation results for the 10,560 labeled data. The
macro-f1 score is reported along with the Precision, Recall, and Accuracy per label type.
The Logistic Regression model, with profile features, yields F1-score 0.53 for agency
prediction and 0.56 for social prediction. The CNN with word embedding outperforms
Logistic Regression with F1-score 0.80 for agency prediction and 0.90 for social
prediction. These results demonstrate that the happy moment contains enough information for
the affective classification without profile features. The XGBoosted Forest with
syntactic and emotional features also reaches a competitive F1-score of 0.78 for agency
prediction and 0.90 for social prediction, meaning that these features are as
representative as the word embedding. The different results of social and agency imply that the
prediction of social label might not rely on the text input, and it’s easier to predict. An
additional experiment run to explore this was the addition of the top 1000 unigrams as
features for the XGBoosted Forest, however this led to little (1-2%) to no increase in
predictive power. To further explore the feature set, we incrementally add the syntactic
features and emotional features, followed by the profile features to the CNN model. The
results show that adding the syntactic and emotional features leads to a slight drop for
agency. Though adding the profile features to the CNN model might lead to small
improvements, we mainly focus on the word embedding, syntactic features and emotional
features for the semi-supervised learning.</p>
        <p>Semi-Supervised Learning. After getting the best models from the supervised
learning, we generate the pseudo labels for the 72,324 unlabeled data using the XGBoosted
Forest and the CNN models. Then we combine the labeled training data with the
pseudolabel data to train the semi-supervised models via 10-fold cross-validation. The
validation set is always held out during the training. Performance of our models are reported
in Table 3.</p>
        <p>
          We had expected performance improvements from semi-supervised learning, but
notice that for the CNN model, the additional 70k pseudo-labeled data does not improve
performance. Compare Table Table 2 and Table 3. In Table 2 for agency prediction, the
best model CNN region (
          <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2, 3, 4, 5</xref>
          ) with GloVe has an F1-score of 0.80 and in Table 3
its F1-score remains 0.80. Similarly, the best CNN model with embeddings, syntactic
and emotional features for social prediction gets an F1-score of 0.90 in Table 2 as well
as Table 3.
        </p>
        <p>Note also that the XGBoosted Forest provides good performance with syntactic
and emotional features, and its performance improves slightly after semi-supervised
learning, e.g. for agency prediction, it has an F1-score of 0.78 for supervised learning,
and 0.79 for semi-supervised learning. For the social prediction, its F1-score is 0.90
for supervised learning and 0.91 for semi-supervised learning. These results encourage
us to investigate the further impact of syntactic and emotional features for affective
prediction tasks.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Concepts Modeling</title>
      <p>This work extends the modeling procedures described above to predicting the
concepts label within the HappyDB. We are interested in the concepts features since they
represent the theme of different types of happy moments. However, we expect that this
is a much more difficult task as it is a multi-class problem. For the concepts modeling
task, we are interested in both improving the prediction results, and interpreting the
performance of the models.</p>
      <sec id="sec-3-1">
        <title>XGBoosted Forest Model</title>
        <p>For this task the labeled 10k data set is split into a training set containing 67% of the
data and the remaining 33% is used for the test set. Within the training set, a 10-fold
cross-validation procedure is used.</p>
        <p>There are 15 unique concepts in the corpus, which are shown in Figure 2, however
they are commonly associated with each other as some instances within HappyDB have
multiple concepts attached. To simplify and examine if concepts are distinguishable
from each other, we only model the cases where a singular concepts tag has been
applied. Using the same feature set and modeling procedure for the XGBoosted Forest,
Table 4a shows the performance of the model on each unique concepts tag. The rows of
the table are ordered by model performance.</p>
        <p>Overall the model shows some promising performance across all concepts. The
model shows some good performance in the top 3 concepts, despite the small
sample size for Religion, it appears to be performing the best. However all other concepts
with lower than 100 instances show much poorer performance.</p>
        <p>One possibility for the poor performance in some of these concepts may be the
association between them that is already present within the HappyDB, as many
concepts are used together. Future work could take these common associations and look at
potentially making a more hierarchical modeling procedure.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>CNN Model</title>
        <p>
          Since the CNN model handles multiple classes, we convert the value of concepts
into a one-hot vector with 15 dimensions, allowing multiple concepts attached to a
happy moment. We explore the performance of CNN region size (
          <xref ref-type="bibr" rid="ref2 ref3 ref4 ref5">2, 3, 4, 5</xref>
          ) with
syntactic features and emotional features by 10-fold cross-validation. The overall accuracy
of the model is 0.596 and F1-score is 0.629. The metrics for each concepts are
demonstrated in Table 4b.
        </p>
        <p>Table 4b shows that the CNN model is generally better than the XGBoosted Forest
Model for the concepts prediction. For example, the highest F1 for CNN is 0.94 for
Religion, and the lowest F1-score is 0.64 for Technology; while the highest F1-score for
XGBoosted Forest is 0.9 and lowest is 0.2, meaning that the CNN model is more robust
and steady for multi-class prediction. The concepts that are improved more than 30% by
CNN are Weather, Party, Education, Exercise, Vacation, and Technology. We suggest
that the different performances are caused by the word information that is missing in
XGBoosted Forest. The POS and LIWC features used within the XGBoosted Forest
appear to be sufficient enough to cover general patterns only within Religion, Food,
Entertainment, Career, and Family.</p>
        <p>Besides the above features, we also explored adding the profile features. Its
overall accuracy is 0.601 and F1-score is 0.626. Adding the profile features to the model
doesn’t provide large improvements, but there are small improvements (1%) for most
of the concepts, with the biggest improvement in Exercise which increased by 3%.
Intuitively, some profile features can be good markers for concepts prediction such as age,
married, and parenthood. For instance, young people tend to discuss events of
Education, while the parents are likely to be happy for a Family theme. On the other hand, the
concepts that drop 1% after adding the profile features are Shopping, Weather, Party,
Conversation. The biggest changes include Technology, whose F1-score drops by 4%.</p>
        <p>Though these models have different performance, they all illustrate the difficulty
of predicting certain concepts and share a similar prediction trend. The trend is not
affected by the size of the training data as in Figure 2. For example, both models agree
that Religion and Food are much easier to predict than Romance and Technology, but
both Religion and Technology have small training sets, suggesting that perhaps Religion
contains many discriminative patterns that make it easier to predict. The performance
also implies that Religion and Food might have some distinctive patterns with global
agreement to represent the happy moment while Romance and Technology might vary,
meaning that people might have very different views of what causes happiness in these
concept themes.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Syntactic Pattern Analysis</title>
        <p>To further interpret the performance of the above models, we apply AutoSlog [8, 12],
a weakly supervised linguistic-pattern learner, to collect the compositional syntactic
patterns for the 10k labeled data. Table 6 illustrates the most frequent syntactic patterns
in the data. For each pattern, we list the top 3 concepts with the highest probability
(no less than 10%) given the pattern. In the top 15 list, there are 3 patterns include
MY (FAMILY MEMBER). This might explain why social prediction is easier than the
other tasks. As for the concepts, a Family theme usually dominates the pattern of MY
(FAMILY MEMBER), which implies that when Romance and Family co-occur, the
classifier would tend to predict Family, leading to a low recall for Romance.
Freq
1395
1303
1153
605
571
518
500
495
476
458</p>
        <p>Moreover, the concepts of Family, Career, Food, and Shopping contain distinctive
syntactic patterns, consistent with classification performance. To validate our
assumptions about the discriminative level of Technology and Religion, we look at the most
frequent patterns. For example, the most common pattern for Technology is (BOUGHT)
with 36 examples, and Religion’s common pattern is (WENT TO) with 148 examples.
Similar to lexical-diversity, we define the pattern-diversity as the number of unique
patterns divided by the total number of patterns. Technology contains 1078 syntactic
patterns, the average frequency for each pattern is 2, the standard deviation of frequency is
3, and the pattern-diversity is 0.55, whereas Religion includes 304 patterns, the average
count is 3, the standard deviation is 10, and the pattern-diversity is 0.51. As for Family,
which has general syntactic patterns, the average count is 3, the standard deviation is
10, and the pattern-diversity is 0.37. A larger standard deviation implies more typical
patterns, and a smaller pattern-diversity implies the concepts tends to include stronger
syntactic patterns. Since Technology has a similar number of examples as Religion,
this suggests that Religion is more easily identified because it has many typical
syntactic patterns. Our observation suggests that the syntactic patterns can be strong markers
for affective classification tasks even when the profile features are missing.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion and Future Work</title>
      <p>We explored the features and models with supervised learning and semi-supervised
learning for social, agency and concepts in order to answer some of the questions during
the experiments:
– Whether syntactic features and emotional features, which are generated from the
text, are representative? Our experiments show that the syntactic features and
emotional features are informative features as they are competitive to word embeddings,
and could outperform the word embeddings for some tasks.
– Whether the profile features are representative? How to convert them into a
meaningful granularity? The profile features include large background information of
the writer, therefore, they might represent some general happiness group. However,
some features should be mapped to a meaningful group, such as country and age.
Though our experiment results are relatively low for the profile features, it could
slightly improve the predictions of social and some concepts when combined with
other features. On the other hand, the text input and its extracted features perform
well on the prediction task, which indicates that the demographic information is not
necessary to achieve decent performance.
– Whether the neural network model outperforms the traditional machine learning
model? Whether the best models for agency and social prediction give good
performance for concepts prediction? The CNN models provide promising performance
for multi-class classification and semi-supervised learning, while the traditional
machine learning method XGBoosted Forest also generates competitive or even
better results for binary class prediction and semi-supervised learning. Our best
feature sets and models for social and agency prediction also provide good
performance for concepts prediction. Our syntactic pattern analysis also demonstrates
that these tasks share common characteristics. Therefore, we believe that they are
generally robust models for affective classification tasks.</p>
      <p>During the experiments, we realize that the variation and the definition of happy
moments might affect the performance of concepts modeling, and imply the
generalization level of different concepts. Some generic characteristics, which are implied by
the syntactic and emotional features, are shared between the classes of agency, social
and concepts. The linguistic-pattern learner AutoSlog provides insightful syntactic
patterns for us to interpret the models. In future work, we will focus on utilizing syntactic
patterns for other affective classification tasks and identify common patterns. We also
hope to explain such patterns and the concepts theme with psychological theories.
Finally, we are curious to see if there are generic characteristics or common compositional
semantic patterns for modeling happy or unhappy moments with cross-domain data.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Asai</surname>
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Evensen</surname>
            <given-names>S.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Golshan</surname>
            <given-names>B.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Halevy</surname>
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Li</surname>
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lopatenko</surname>
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Stepanov</surname>
            <given-names>D.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Suhara</surname>
            <given-names>Y.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Tan</surname>
            <given-names>W.C.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Xu</surname>
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Happydb: A corpus of 100,000 crowdsourced happy moments</article-title>
          .
          <source>In Proceedings of LREC 2018. European Language Resources Association(ELRA)</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Guestrin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Xgboost: A scalable tree boosting system</article-title>
          .
          <source>In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining</source>
          , pp.
          <fpage>785</fpage>
          -
          <lpage>794</lpage>
          . ACM (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Compton</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Haber</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Badenes</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Whittaker</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Just the Facts: Exploring the Relationship between Emotional Language and Member Satisfaction in Enterprise Online Communities</article-title>
          .
          <source>In Proceedings of the 11th International Conference on Web and Social Media. AAAI</source>
          . (
          <year>2017</year>
          ) https://aaai.org/ocs/index.php/ICWSM/ICWSM17/paper/view/15664
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Jaidka</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Mumick</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Chhaya</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Ungar</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , :
          <article-title>The CL-Aff Happiness Shared Task: Results and Key Insights</article-title>
          .
          <source>In Proceedings of the 2nd Workshop on Affective Content Analysis @ AAAI (AffCon2019)</source>
          . Honolulu,
          <string-name>
            <surname>Hawaii</surname>
          </string-name>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Mohammad</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Turney</surname>
          </string-name>
          , P. D.:
          <article-title>Emotions evoked by common words and phrases: Using Mechanical Turk to create an emotion lexicon</article-title>
          .
          <source>In Proceedings of the NAACL HLT</source>
          <year>2010</year>
          <article-title>workshop on computational approaches to analysis and generation of emotion in text</article-title>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>34</lpage>
          . Association for Computational (
          <year>2010</year>
          )
          <article-title>Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Manning</surname>
          </string-name>
          , C. D.:
          <article-title>GloVe: Global Vectors for Word Representation</article-title>
          .
          <source>In Proceedings of EMNLP</source>
          <year>2014</year>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          . (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Reed</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Oraby</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Anand</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Walker</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Learning Lexico-Functional Patterns for First-Person Affect</article-title>
          .
          <article-title>Association for Computational Linguistics (ACL) (</article-title>
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Riloff</surname>
          </string-name>
          , E.:
          <article-title>Automatically generating extraction patterns from untagged text</article-title>
          . pp.
          <fpage>1044</fpage>
          -
          <lpage>1</lpage>
          -
          <fpage>49</fpage>
          .
          <source>Proceedings of the Thirteenth National Conference on Artificial Intelligence (AAAI-96)</source>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Tausczik</surname>
            ,
            <given-names>Y. R.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Pennebaker</surname>
            ,
            <given-names>J. W.:</given-names>
          </string-name>
          <article-title>The psychological meaning of words: LIWC and computerized text analysis methods</article-title>
          .
          <source>Journal of language and social psychology</source>
          ,
          <volume>29</volume>
          (
          <issue>1</issue>
          ), pp
          <fpage>24</fpage>
          -
          <lpage>54</lpage>
          . (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Singer</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          (
          <year>2003</year>
          , May).
          <article-title>Feature-rich partof-speech tagging with a cyclic dependency network</article-title>
          .
          <source>In Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology-Volume 1</source>
          , pp.
          <fpage>173</fpage>
          -
          <lpage>180</lpage>
          . Association for Computational Linguistics. (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11. Wilson,
          <string-name>
            <given-names>T.</given-names>
            &amp;
            <surname>Hoffmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            &amp;
            <surname>Somasundaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            &amp;
            <surname>Kessler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            &amp;
            <surname>Wiebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            &amp;
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            &amp;
            <surname>Patwardhan</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.:</surname>
          </string-name>
          <article-title>OpinionFinder: A system for subjectivity analysis</article-title>
          .
          <source>In Proceedings of hlt/emnlp on interactive demonstrations</source>
          . pp.
          <fpage>34</fpage>
          -
          <lpage>35</lpage>
          . Association for Computational Linguistics (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Walker</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Anand</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Whittaker</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Linguistic Reflexes of Well-Being and Happiness in Echo</article-title>
          .
          <source>The 8th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis (WASSA)</source>
          (
          <year>2017</year>
          ),
          <source>Empirical Methods in Natural Language Processing (EMNLP)</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Wallace</surname>
            ,
            <given-names>B. C.</given-names>
          </string-name>
          :
          <article-title>A Sensitivity Analysis of (and Practitioners' Guide to) Convolutional Neural Networks for Sentence Classification</article-title>
          .
          <source>IJCNLP</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>