=Paper= {{Paper |id=Vol-1391/122-CR |storemode=property |title=INAOE's Participation at PAN'15: Author Profiling task |pdfUrl=https://ceur-ws.org/Vol-1391/122-CR.pdf |volume=Vol-1391 |dblpUrl=https://dblp.org/rec/conf/clef/CarmonaLMPE15 }} ==INAOE's Participation at PAN'15: Author Profiling task== https://ceur-ws.org/Vol-1391/122-CR.pdf
INAOE’s participation at PAN’15: Author Profiling task
                        Notebook for PAN at CLEF 2015


             Miguel A. Álvarez-Carmona, A. Pastor López-Monroy,
     Manuel Montes-y-Gómez, Luis Villaseñor-Pineda, and Hugo Jair Escalante

            Language Technologies Laboratory, Department of Computer Science,
                    Instituto Nacional de Astrofísica, Óptica y Electrónica,
                  Luis Enrique Erro No. 1, C.P. 72840, Pue. Puebla, México
       {miguelangel.alvarezcarmona,pastor, mmontesg, villasen, hugojair}@inaoep.mx



       Abstract In this paper, we describe the participation of the Language Technolo-
       gies Lab of INAOE at PAN 2015. According to the Author Profiling (AP) liter-
       ature. In this paper we take such discriminative and descriptive information into
       a new higher level exploiting a combination of discriminative and descriptive
       representations. For this we use dimensionality reduction techniques on the top of
       typical discriminative and descriptive textual features for AP task. The main idea
       is that each representation, using the full feature space, automatically highlights
       the different stylistic and thematic properties in the documents. Specifically, we
       propose the joint use of Second Order Attributes (SOA) and Latent Semantic
       Analysis (LSA) techniques to highlight discriminative and descriptive properties
       respectively. In order to evaluate our approach, we compare our proposal against
       a standard Bag-of-Words (BOW), SOA and LSA representations using the PAN
       2015 corpus for AP. Experimental results in AP show that the combination of
       SOA and LSA outperforms the BOW and each individual representation, which
       gives evidence of its usefulness to predict gender, age and personality profiles.
       More importantly, according to the PAN 2015 evaluation, the proposed approach
       are in the top 3 positions in every dataset.



1 Introduction

The Author Profiling (AP) task consists in knowing as much as possible about an un-
known author, just by analysing a given text [4]. The interest in AP tasks has captured
the attention of the scientific community in recent years. This is due, in part, to the po-
tential of the huge amount of the user-generated textual information on the internet. In
this context, several applications related to AP are emerging, some of them have to do
with e-commerce, computer forensics and security. There are several ways to address
the AP task. One of them is to approach it as a single-label multiclass classification
problem, where the target specific profiles (e.g., male and female) represent the classes
to discriminate.
   Broadly speaking, in text classification tasks there are three general-key procedures;
i) the extraction of textual features, ii) the representation of the documents, and iii) the
application of a learning learning algorithm. In the context of the AP tasks, for the first
step, specific lexical (e.g., simple words, function words) [4] and syntactical features
(e.g., POS tags) [14] have proven to be highly discriminative for some specific profiles.
Regarding to the last two steps, the representation of documents and the learning algo-
rithm that are the most common-effective approaches for AP tasks consist in using the
Bag-of-Words formulation (e.g., histograms of the presence/absence of textual features)
[15] and Support Vector Machines [2][5] respectively.
   According to the AP task literature, most of the work has been devoted to the first
step: to identify the most useful-interesting textual features for the target profiles. In
spite of the usefulness of previous interesting textual features and the good results
achieved by the configuration BoW-SVM, the research community has put little effort
to deepen in the second and third steps: alternative representations and learning algo-
rithms for the AP task. The main shortcomings of the BoW-SVM approach are well
known from other text mining task.
    To overcome the latter shortcomings, in this paper we focus in the second step, i.e.
the representation of the documents, in order to improve the representation of tweets.
The main goal of our approach is to compute high quality discriminative and descrip-
tive features built on the top of the state-of-the-art typical textual features (e.g., con-
tent words, function words, punctuation marks, etc.). For this, we propose to combine
two state-of-the-art dimensionality reduction techniques that best contribute to auto-
matically stress the contribution of the discriminative and descriptive textual features.
According to the literature the most frequent textual features (e.g., function words, stop-
words, punctuation marks) provide important clues about the discrimination of the au-
thors. For this we need a representation highly based in term frequencies, that stresses
the contribution of such discriminative attributes and produces highly discriminative
document representations. To capture this information contained among textual fea-
tures we use Second Order Attributes (SOA) computed as in [8]. On the other hand,
relevant thematic information usually are in descriptive terms, terms that are frequent
only in some specific documents or classes. In this way, to represent documents we
bring ideas from the information retrieval field exploiting the Latent Semantic Analysis
(LSA) [16]. LSA represents terms and documents into a new semantic space. This is
done performing a singular value decomposition using a Term Frequency Inverse Doc-
ument Frequency (TFIDF) matrix. The descriptive terms and documents representation
are stressed under the LSA formulation throwing out the noise, but emphasizing strong
patterns and trends. To the best of our knowledge, the idea of representing documents
using the combination discriminative and the descriptive high-level features through
dimensionality reduction techniques have never been explored before in AP task. Thus,
it is promising to bring together two of the best document representations to better im-
prove the AP; that is precisely the propose of this work.
    The rest of this paper is organized as follows: Section 2 introduces the proposed
representation, in Section 3 some characteristics of the corpus PAN15 are explained
briefly, Section 4 explains how we performed the experiments and the results we ob-
tained, finally Section 5 shows our conclusions.


2 Exploiting Discriminative and Descriptive features

Along this section we briefly describe each representation and the proposed strategy
to compute the final representation of documents. In Section 2.1 we explain the SOA
representation to get the discriminative features. In Section 2.2 we explain the LSA
algorithm with which we intend to get descriptive features. Finally, in Section 2.3 we
explain how we join these representations for the AP task.


2.1 Computing Discriminative Features

The stylistic textual features have proven to be useful for AP task [11]. A plenty of the
style textual attributes in text mining tasks (e.g., Author Profiling, Authorship Attribu-
tion, Plagiarism Detection) have been associated with highly frequent terms [12]. For
example, observing the frequency of stopwords and punctuation marks exposes clues
about the author of a document. In gender identification observing the distribution of
specific function words and determiners have proven to be also useful [11]. Second Or-
der Attributes (SOA) proposed in [8] is a supervised frequency based approach to build
document vectors in a space of the target profiles. Under this representation, each value
in the document vector represents the relationship of each document with each target
profile.
    The representation as described in [8] has two keys steps. i) To build words vectors
in an space of profiles and ii) to build documents vectors in an space of profiles. In the
former step, for each vocabulary term tj , a tj = htp1j , ..., tpmj i vector is computed.
Where each tpmj is a frequency-based-value that represents the relationship between
term tj and the profile pm . In the latter step the representation of documents is built us-
ing a weighted by frequency aggregation of the term vectors contained in the document
(see Equation 1).

                                       X          tfkj
                               dk =                        tj                           (1)
                                               lenght(dk )
                                      tj ∈Dk

    where Dk is the set of terms that belongs to document dk . For more details please
refer to [8].

2.2 Computing Descriptive Features
Besides the usefulness of stylistic features, thematic information has proven to be an
important aspect for the AP Task [11]. For example, several works have shown evi-
dence that groups of people of the same age and gender write generality about the same
topics. For this reason we exploit the Latent Semantic Analysis (LSA). LSA is a tech-
nique that can associate words and it contribution to automatically generated concepts
(topics) [1]. This is usually named the latent space, where documents and terms are
projected to produce a reduced topic based representation. We hypothesises that under
the aforementioned latent space, we can better expose descriptive relevant information
for the AP task.
  LSA is a method to extract and represent the meaning of the words and documents.
LSA is built from a matrix M where mij is typically represented by the TFIDF [13]
of the word i in document j. LSA uses the Singular Value Decomposition (SVD) to
decompose M as follows.

                                      M = UΣVT                                          (2)

   Where The Σ values are called the singular values and U and V are the left and right
singular vectors respectively. U and V contains a reduced dimensional representation
of words and documents respectively. U and V emphasizes the strongest relationships
and throws away the noise [6]. In other words, it makes the best possible reconstruction
of the M matrix with the less possible information [7]. Using U and V computed only
from the training documents, words and documents are represented for training and test.
For more details please refer to [16].

2.3 Exploiting the jointly use of Discriminative-Descriptive Features
The idea is to use the representations built under the whole feature space to automati-
cally highlight the discriminative and descriptive properties in documents. The intuitive
idea is to take advantage of both approaches in a representation using early fusion. Let
xj be the j − th training instance-profile under LSA representation with k dimensions
and yj be the same instance-profile under the SOA representation with m dimensions,
the final representation is show in Expression 3.


                            zj = hxj1 , . . . , xjk , yj1 , . . . , yjm i                  (3)

The collection of training documents are finally represented as:

                                               [
                                      Z=             hzj , cj i                            (4)
                                             dj ∈D

     Where cj is the class of the j − th training instance-profile.


3 Data Collection

We have approached the PAN 2015 AP task as a classification problem. PAN 2015
corpora is composed by 4 datasets in different languages (Spanish, English, Italian and
Dutch). Each dataset has labels of gender (male, female), age 1 (18-24, 25-34, 35-49,
50-xx) and five personality traits values (extroverted, stable, agreeable, conscientious,
open) between -0.5 and 0.5. In Table 1 we show the number of Author-Profiles per
language.
     For personality identification Table 2 shows the relevant information (in terms of
classes). For each language it shows the range and the number of the classes for each
trait2 . For personality we consider each trait value in the train corpus as a class. For
example, if only two values (e.g., 0.2 and 0.3) are observed in the training corpus, then
we built a two class classifier (e.g., 0.2 and 0.3) 3 .


4 Experimental Evaluation

4.1 Experimental Settings

We use for each experiment the following configuration: i) for terms we use words,
contractions, words with hyphens, punctuation marks and a set of common emoticons,
 1
   Age data for Italian and Dutch languages are not available.
 2
   The ranges with asterisk indicate that a value between the range is missing. For example, in
   Spanish (extroverted and conscientious) the -0.1 is missing.
 3
   For each personality trait in each language the number of the classes are variables between
   them, see Table 2
                            Table 1. Description of the dataset

                          Language             Author-Profiles
                           English                 152
                           Spanish                 100
                            Italian                 38
                            Dutch                   34

                  Table 2. The personality traits information by language

               English         Spanish        Italian      Dutch
    Trait       Range Classes Range Classes Range Classes Range Classes
 Extroverted [-0.3,0.5] 9    [-0.3,0.5]* 8 [0.0,0.5]*  5  [0.0,0.5] 6
   Stable     [-0.3,0.5] 9    [-0.3,0.5] 9  [-0.1,0.5] 7 [-0.2,0.5] 8
 Agreeable [-0.3,0.5] 9       [-0.2,0.5] 8 [-0.1,0.5]* 6 [-0.1,0.4] 6
Conscientious [-0.2,0.5] 8   [-0.2,0.5]* 7   [0.0,0.4] 5 [-0.1,0.4] 6
    Open      [-0.1,0.5] 7    [-0.1,0.5] 7  [0.0,054]  6  [0.1,0.5] 5


ii) we consider the terms with at least 5 occurrences in the corpus, iii) the number of
concepts for LSA is set to k = 100. We perform an stratified 10 cross fold validation
(CFV) using the training PAN15 corpus and a LibLINEAR classifier [3]. In order to
determine the full profile of a document (gender, age and the 5 personality traits) we
built one classifier to predict each target profile for each language.


4.2 Experimental Results

The aim of this first experiment is to analyse the performance of LSA, SOA and the
BOW approach in the AP tasks. We experiment with LSA and SOA separately and
finally with the two approaches together. We are interested in observing the contribu-
tion of discriminative-stylistic (captured by SOA) and descriptive-thematic (captured by
LSA) information in the AP task. For gender prediction, in Table 3 we can see that con-
sidering the individual representations, LSA obtains the best results, which outperforms
the BOW approach in every language. When LSA and SOA are together the result only
improves in English, which is an important remark since the English language is the
bigger-robust collection (see Table 1). The following conclusions can be outlined from
Table 3:

 – The descriptive information captured by LSA is the most relevant information for
    gender prediction in PAN 2015 AP dataset. This is because LSA obtained the best
    average individual performance.
                        Table 3. Detailed classification accuracy to gender

            Language            BOW            SOA           LSA              LSA+SOA
             English            74.00          70.86         74.34              78.28
             Spanish            84.00          74.00         91.00              91.00
              Italian           76.31          73.68         86.84              86.84
              Dutch             82.35          91.07         91.17              91.17


 – The pure discriminative information captured by SOA only outperforms BOW in
   Dutch documents. But the combination of LSA and SOA obtained an improvement
      of around 4% in accuracy for English gender detection. We think, SOA could im-
      prove the results if more documents are available 4 .


    For age prediction, Table 4 shows the experimental results. Recall that the age data is
available only for English and Spanish languages. As in the last experiment LSA obtains
the best individual performance, but in this experiment the combination of LSA and
SOA obtains an improvement in both collections. It is worth noting that despite of the
small datasets, for age prediction SOA could contribute to improve the classification5 .


                          Table 4. Detailed classification accuracy to age

            Language           BOW            SOA           LSA              LSA+SOA
             English           74.83          68.21         78.94              79.60
             Spanish           80.00          74.00         81.00              82.00



    Finally for personality prediction Table 5 shows the performance of BOW and LSA
plus SOA performance by language in the personality detection task. For this experi-
ment, although the results seems promising they should be taken with caution. This is
due to the lack of data and the number of classes that we consider (one class for each
observed value) one correct/wrong predicted instance is enough to change the results
considerably. For this specific experiment in personality, we built a representation on
the entire dataset, then we evaluate using a 10CFV. In general, the results suggest that
the combination of LSA plus SOA gets similar or better results than the typical BOW
approach. Given evidence of the usefulness of the discriminative features and the de-
scriptive features.

 4
     SOA has proven outstanding results in recent years in the PAN AP tracks [9,10].
 5
     The best results for SOA in previous PAN AP editions have been for age prediction
                    Table 5. Detailed classification accuracy for personality

             English     Spanish      Italian     Dutch
    Trait     BOW LSA+SOA BOW LSA+SOA BOW LSA+SOA BOW LSA+SOA
 Extroverted   64    87    62    87     65    94   64    91
   Stable      56    85    69    91     52    94   61    94
 Agreeable     60    80    62    84     71    92   61    88
Conscientious 61     78    62    86     57    94   67    91
    Open       65    86    62    74     55    84   64    97


5 Conclusions

In this paper, we have explored a new combination of document representations for
AP task. The main aim of this work was to experiment the with descriptive (LSA)
and discriminative (SOA) features. We found that the descriptive information is very
useful, which confirms several findings in the literature. Moreover, we also find that
discriminative information could improve the results when it is combined with descrip-
tive information. This indicates that LSA captures very important information which in
turn can be complemented with the SOA stylistic information.


Acknowledgment This work was partially funded by the program
SEP-CONACYT-ANUIES-ECOS Nord under the project M11-H04. Álvarez-Carmona
thanks for doctoral scholarship CONACyT-Mexico 401887. Also, López-Monroy
thanks for doctoral scholarship CONACyT-Mexico 243957.


References

 1. Dumais, S.T.: Latent semantic analysis. Annual review of information science and
    technology 38(1), 188–230 (2004)
 2. Estival, D., Gaustad, T., Pham, S.B., Radford, W., Hutchinson, B.: Author profiling for
    english emails. In: Proceedings of the 10th Conference of the Pacific Association for
    Computational Linguistics (PACLING’07). pp. 263–272 (2007)
 3. Fan, R.E., Chang, K.W., Hsieh, C.J., Wang, X.R., Lin, C.J.: Liblinear: A library for large
    linear classification. The Journal of Machine Learning Research 9, 1871–1874 (2008)
 4. Koppel, M., Argamon, S., Shimoni, A.R.: Automatically categorizing written texts by
    author gender. Literary and Linguistic Computing 17(4), 401–412 (2002)
 5. Koppel, M., Schler, J., Zigdon, K.: Determining an author’s native language by mining a
    text for errors. In: Proceedings of the eleventh ACM SIGKDD international conference on
    Knowledge discovery in data mining. pp. 624–628. ACM (2005)
 6. Landauer, T.K., Foltz, P.W., Laham, D.: An introduction to latent semantic analysis.
    Discourse processes 25(2-3), 259–284 (1998)
 7. Landauer, T.K., McNamara, D.S., Dennis, S., Kintsch, W.: Handbook of latent semantic
    analysis. Psychology Press (2013)
 8. Lopez-Monroy, A.P., Gomez, M.M.y., Escalante, H.J., Villasenor-Pineda, L.,
    Villatoro-Tello, E.: Inaoe’s participation at pan’13: Author profiling task. In: CLEF 2013
    Evaluation Labs and Workshop (2013)
 9. Rangel, F., Rosso, P., Chugur, I., Potthast, M., Trenkmann, M., Stein, B., Verhoeven, B.,
    Daelemans, W.: Overview of the 2nd author profiling task at pan 2014. In: Proceedings of
    the Conference and Labs of the Evaluation Forum (Working Notes) (2014)
10. Rangel, F., Rosso, P., Koppel, M., Stamatatos, E., Inches, G.: Overview of the author
    profiling task at pan 2013. Notebook Papers of CLEF pp. 23–26 (2013)
11. Schler, J., Koppel, M., Argamon, S., Pennebaker, J.W.: Effects of age and gender on
    blogging. In: AAAI Spring Symposium: Computational Approaches to Analyzing Weblogs.
    vol. 6, pp. 199–205 (2006)
12. Stamatatos, E.: A survey of modern authorship attribution methods. Journal of the
    American Society for information Science and Technology 60(3), 538–556 (2009)
13. Turney, P.: Mining the web for synonyms: Pmi-ir versus lsa on toefl (2001)
14. Van Halteren, H.: Linguistic profiling for author recognition and verification. In:
    Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics. p.
    199. Association for Computational Linguistics (2004)
15. Wang, G., Hoiem, D., Forsyth, D.: Building text features for object image classification. In:
    Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on. pp.
    1367–1374. IEEE (2009)
16. Wiemer-Hastings, P., Wiemer-Hastings, K., Graesser, A.: Latent semantic analysis. In:
    Proceedings of the 16th international joint conference on Artificial intelligence. pp. 1–14.
    Citeseer (2004)