<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Network Based Approach for the Visualization and Analysis of Collaboratively Edited Texts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tobias Hecking</string-name>
          <email>hecking@collide.info</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>General Terms</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>H. Ulrich Hoppe</string-name>
          <email>hoppe@collide.info</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Algorithms</institution>
          ,
          <addr-line>Visualization, Experimentation</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Duisburg-Essen</institution>
          ,
          <addr-line>Lotharstraße 63/65, 47048 Duisburg</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes an approach for network text analysis and visualization for collaboratively edited documents. It incorporates network extraction from texts where nodes represent concepts identified from the words in the text and the edges represent relations between the concepts. The visualization of the concept networks depicts the general structure of the underlying text in a compact way. In addition to that, latent relations between concepts become visible, which are not explicit in the text. This work concentrates on evolving texts such as wiki articles. This introduces additional complexity since dynamic texts lead to dynamic concept networks. The presented method retains the user information of each revision of a text and makes them visible in the network visualization. In a case study it is demonstrated how the proposed method can be used to characterize the contributors in collaborative writing scenarios regarding the nature of concept relations they introduce to the text.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Network Visualization</kwd>
        <kwd>Network Analysis</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Collaborative Writing</kwd>
        <kwd>Learning Analytics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Network text analysis is the task of extraction and analysis of
networks from text corpora. In those networks the nodes are
concepts identified from the words in the text and the edges
between the nodes represent relations between the concepts. The
visualization of concept networks can help to depict the general
structure of the underlying text in a compact way. In addition to
that, latent relations between concepts become visible, which are
not explicit in the text. Thus, approaches for visualizing texts as
networks allow analysts to concentrate on important aspects
without reading large amounts of the texts. Several network
analysis techniques can be applied to identify important concepts,
perform concept clustering, as well as comparative analysis of
different texts [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        Existing applications for network text analysis include the
identification of key phrases [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], mining of relations between
real world entities [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], as well as the extraction of complete
concept ontologies and concept maps with labelled edges [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
This work concentrates on the relations between concepts that can
be found in evolving and collaboratively edited texts such as wiki
articles. This introduces additional complexity since dynamic
texts lead to dynamic concept networks. The presented method
retains the user information of each revision of a text which
allows for characterizing the contributors in collaborative writing
scenarios regarding the nature of concept relations they introduce
to the text. The resulting visualization is a concept network with
colored edges where each edge color is allocated uniquely to a
specific contributor. In further analysis steps, network centrality
measures are calculated that give additional information about the
contribution of each editor.
      </p>
      <p>The outline of this paper is as follows: Section 2 gives the
theoretical background of this work and highlights significant
research work in the area of network text analysis. The general
idea of our visualization and analysis approach is presented in
section 3. Section 4 focuses on the concrete implementation. This
incorporates the applied natural language processing chain, as
well as the description of network analysis methods.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
    </sec>
    <sec id="sec-3">
      <title>2.1 Collaborative Writing Activities in</title>
    </sec>
    <sec id="sec-4">
      <title>Education</title>
      <p>
        Collaborative writing activities are a common task in educational
scenarios [
        <xref ref-type="bibr" rid="ref13 ref3">3, 13</xref>
        ]. Users can learn actively by creating artefacts but
can also learn passively by consuming artefacts created by others
[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        It could be shown that user generated content is relevant to
learners in addition to tutor provided content [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. With the
emergence of online communities such as Wikipedia collaborative
knowledge building takes place with open scale in terms of the
number of contributors. There is some evidence that individual
and collective knowledge co-evolves through collaborative
editing of epistemic artefacts in open online environments [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In
general collaborative writing requires different rhetorical and
organizational skills of the editors [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and thus, the learner
generated artefacts are a valuable data source for analysis.
This motivates the development of methods that makes
collaborative writing processes visible in order to understand and
improve the application of collaborative text writing in
educational settings.
      </p>
    </sec>
    <sec id="sec-5">
      <title>2.2 Visualization Approaches for</title>
    </sec>
    <sec id="sec-6">
      <title>Collaborative Writing</title>
      <p>
        Several methods have been developed to represent evolving texts
with multiple editors in a visual way. One of the first approaches
for the visualization of evolving wiki articles is the History Flow
method [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. In this approach each contributor has assigned a
unique color. Each revision of the evolving text is then
represented as a sequence of blocks that represent the sections of
the document. The blocks are colored according to the author who
has edited the section and the size of the block corresponds to the
amount of text. This does not only depict the insertion and
removal of text sections by the users but additionally allow for the
identification of edit wars between authors. In contrast to this
page centric view, the iChase method [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] visualizes activities of
a set of authors across multiple wiki articles as heatmaps.
Southavilay et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] extend the pure depiction of the amount
and location of text edits done by a user by incorporating topic
modeling. Therefore, they apply latent dirichlet allocation [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] in
order to identify the contributions of users to the particular topics
covered in a document. Based on the identified topics the
evolution of topics as well as collaboration networks of users on
particular topics can be analyzed.
      </p>
    </sec>
    <sec id="sec-7">
      <title>2.3 Representing Mental Models as Graphs</title>
      <p>
        Networks are a common representation for relations between
entities of various kinds. Schvaneveldt et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] argue that
networks between entities based on proximities induced by people
have a psychological interpretation. They assume that cognitive
concepts such as memory organization and mental categories are
reflected in the network structure. The pathfinder algorithm [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]
derives a network of concepts from proximity data. Such
proximities could be induced, for example, by associations made
by a person. In general, it is also possible to derive such proximity
data between concepts described in natural language texts [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
One of the first approaches that utilize computational tools to
extract mental models from text has been described by Carley [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
After the identification of relevant words in a text, the words are
linked based on syntactical analysis of the sentences of a text.
This approach has been further developed by Diesner et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
and implemented in the software tool Automap where an analyst
can specify a metamatrix of concepts and concept classes. This
enables the identification of relations between entities of different
types from text corpora, for example, people and organizations.
      </p>
    </sec>
    <sec id="sec-8">
      <title>3. Visualization Approach</title>
      <p>This paper extends network extraction from texts to dynamically
evolving and collaboratively edited documents. When networks
extracted from texts are considered as the author’s mental model
of the domain, as described in section 2.3, the aggregation of the
networks extracted from several revisions of a collaboratively
edited text can be interpreted as the joint representation of the
individual mental models of all authors.</p>
      <p>The basic assumption is that different authors introduce different
concepts and relations to the text. In order to make these
differences visible the author information is additionally
incorporated into the network representation.</p>
      <p>Each connection between concepts that can be extracted from the
text can be labeled with the author who established it. In the small
example in Figure 1 the little piece of text was produced by two
different authors. Each author has assigned a unique color - in this
case blue and red. The edges of the resulting network can then be
colored according to the author who was the first who introduced
the concept relation in the text.</p>
      <p>This not only allows for a characterization of the underlying
document in terms of concept relations but also a characterization
of the contributors. Central concepts that are used by different
authors but linked to different other concepts indicate different
associations or views of the authors. Furthermore, the
visualization approach additionally depicts which authors
concentrate on thematic areas and which authors tend to relate
concepts from different sub topics, for example, by writing a
summary.
By calculating network measures on the concept network a further
quantitative characterization of the authors is possible as
described in section 4.3.</p>
    </sec>
    <sec id="sec-9">
      <title>4. Implementation</title>
      <p>This section outlines details of the implementation in two
perspectives. In particular, these are word network extraction
using natural language processing, and network analysis.</p>
    </sec>
    <sec id="sec-10">
      <title>4.1 Extracting Concept Networks from Texts</title>
      <p>
        The extraction of networks from text requires several natural
language processing components. In this work the DKPro toolkit
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] was used. It is based on the Apache UIMA1 framework and
provides a large variety of natural language processing algorithms
that can be combined in a flexible way. The process of the
extraction of word networks from a single document is depicted in
Figure 2. First, a preprocessing step is often required for text
gathered from the web in order to remove wiki or HTML markup.
Further, in this step irrelevant content can be filtered from the
document. For example, Wikipedia pages often contain a large
reference section and a list of related web resources. These parts
are important for the wiki article itself but are a source of noise
when the actual content of the article should be analyzed. In the
1 https://uima.apache.org/
second step, the phrases representing concepts in the text have to
be identified, and after that, connected to a network by using a
proximity measure in step 3. Since the result might contain
phrases with slightly different spelling which actually refer to the
same semantic concept the entity resolution step merges those
candidate phrases to a single concept. Concepts and relations can
then be encoded as a network that is used for further processing.
In the following the steps 2 to 4 are described in more detail.
      </p>
      <sec id="sec-10-1">
        <title>4.1.1 Concept Extraction</title>
        <p>For the identification of the concepts in the input text noun phrase
chunking was applied. First, the text is segmented into its
sentences. Then part-of-speech (POS) tagging (using the Stanford
PSO tagger2) is applied to label each word according to its
function in its sentence. A naive solution for the extraction of
concepts from the text would be to take each noun identified by
the POS tagging as one concept. However, often one concept is
described by more than one word. For example the phrase
“Approach [NN] for [for] teaching [NN]” would result in two
concepts, namely “Approach” and “Teaching”, which does not
really reflect the meaning of the phrase. Thus, noun phrase
chunking is applied where the POS labeled words are chunked to
meaningful noun phrases. This is done with the OpenNLP
chunker3, which identifies noun phrases according to certain rules.</p>
        <sec id="sec-10-1-1">
          <title>2 http://nlp.stanford.edu/software/tagger.shtml</title>
        </sec>
        <sec id="sec-10-1-2">
          <title>3https://opennlp.apache.org/documentation/1.5.2</title>
          <p>incubating/manual/opennlp.html
For example, the words “Approach [NN] for [for] teaching [NN]”
are then identified as one single noun phrase.</p>
        </sec>
      </sec>
      <sec id="sec-10-2">
        <title>4.1.2 Relation Extraction</title>
        <p>
          After all concepts in the text are identified they have to be
connected to a concept network according to a certain proximity
measure. In this work, an edge between two concepts becomes
established if the concepts co-occur in a sliding window of n
words in at least one of the sentences in the text. This approach is
straight forward but works well in practice [
          <xref ref-type="bibr" rid="ref10 ref6">6, 10</xref>
          ].
        </p>
      </sec>
      <sec id="sec-10-3">
        <title>4.1.3 Entity Resolution</title>
        <p>
          As already mentioned entity resolution is necessary in order to
identify nodes in the network that represent the same concept and
to merge them into single nodes. For example the noun phrases
“Wiki” and “The Wikis” can be merged to the same concept
“Wiki”. In order to solve this problem, first all noun phrases have
to be normalized using lemmatization. After that the concepts are
compared pairwise by substring similarity [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. If the similarity
exceeds a value of 0.7 the concepts are merged and labeled with
the shorter label of the two concepts.
        </p>
      </sec>
    </sec>
    <sec id="sec-11">
      <title>4.2 Networks from Different Revisions</title>
      <p>In order to extract an aggregated network from different revisions
of a collaboratively edited text, the process chain described in
section 4.1 is applied to each revision of the text in temporal order
from the oldest to the latest revision. Each revision of the text was
done by a single author. The edges in the network of the first
revision are labeled with the author of this initial revision. Then in
the first aggregation step all edges that are part of the network
extracted from the second revision but do not exist in the network
of the first revision are labeled with the author of the second
revision and added to the previously extracted network. This
proceeds until each revision has been processed. As described in
section 3 the author information attached to the edges can then be
visualized by using different colors for each author.</p>
      <p>
        Since the aggregated network contains every noun phrase that has
been used by the authors as a concept node, the network can be
very large and likely contains concepts that are not relevant for
the domain. Those concepts are often not well connected. Thus, in
a preprocessing step the k-core [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] of the network is computed
such that the resulting network contains only concepts with at
least k connections to other concepts of the core. The resulting
network has a reduced number of nodes, and the visualization
concentrates on the most important concepts according to the
connectedness to other core concepts in the network.
      </p>
    </sec>
    <sec id="sec-12">
      <title>4.3 Quantitative Characterization of</title>
    </sec>
    <sec id="sec-13">
      <title>Contributors</title>
      <p>
        For quantitative analysis the nodes (concepts) and edges can be
ranked according to network centrality measures [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. In this
work concepts are ranked according to eigenvector centrality and
betweenness centrality. The eigenvector centrality is a recursive
measure and assigns a weight to each node according to the
number its neighbors while the connections are weighted
according to the centrality of the neighbors. This gives high
weight to concepts that have many connections to other important
concepts.
      </p>
      <p>Edges are ranked according to the edge-betweenness centrality.
The edge-betweenness centrality assigns high weights to edges
that often occur on shortest paths between any pair of nodes.
In order to use the network measures for a characterization of the
authors of the document an aggregation is necessary. For the node
centric centralities, namely node-betweenness and eigenvector
centrality the centrality contribution of an author A can be
calculated by equation 1:
This result is the average centrality of nodes that are incident to
edges labeled with author A.</p>
      <p>The edge-betweenness contribution of author A is the average of
all edges labeled with author A (equation 2):
An author with a high contribution in terms of edge-betweenness
centralilty could be interpreted as someone who relates different
parts of the text and introduces relations between concepts of
different sections. This could, for example, be someone who
creates a comprehensive summary of a longer wiki article.
Authors with high contribution to the eigenvector centrality of the
concepts can be those who work on important sections of the text
and establish many relations between important domain concepts.
options
(1)
(2)</p>
    </sec>
    <sec id="sec-14">
      <title>5. Case Study</title>
      <p>As a case study the described method was applied to a wiki article
on media economy created during a master level university course
in a study program on Applied Cognitive Science and Media
Science. The relations between the concepts are based on a sliding
window with the size of 4 words. Figure 3 depicts the 5-core of
the resulting aggregated concept network. The size of the nodes
corresponds to the number of connections in order to support the
visual discovery of important concepts. It can be directly seen
from the visualization that the concept “media combination” is
most central. Four of the six authors relate this concept to other
concepts as it can be seen by counting the different colors of the
incident edges. The highest coverage of the edges has the author
who has pink as assigned color. Other contributors relate concepts
more according to certain sub topics like communication (see blue
edges).</p>
      <p>The results for the quantitative characterization of the contributors
are presented in Table 1. It is important to mention that reducing
the network to its 5-core has mainly presentation purposes. Thus,
for more reliable results the calculations were performed on the
2core of the network in which more concept are present.</p>
    </sec>
    <sec id="sec-15">
      <title>Author</title>
      <p>Student 1 has by far the highest contribution to the edge
betweenness centrality. This is reasonable because this student did
a reworking of large parts of the article and was highly involved
in the shaping of the particular sections of the text. Student 2 has
the highest scores regarding the node based centrality measures.
However, the average edge-betweenness centrality is only
moderate. This indicates that this student concentrated on the core
topic of the article. This can also be seen in Figure 3 where the
red edges of student 2 are all incident to the central concept.</p>
    </sec>
    <sec id="sec-16">
      <title>6. CONCLUSION AND FURTHER WORK</title>
      <p>The research presented in this paper describes an approach for the
extraction of concept networks from text that incorporates author
information in the visualization. In contrast to other existing
visualizations of evolving texts our approach focuses rather on the
relations between concepts than on the amount of text that is
produced by individual authors. The case study has shown that the
method is promising and can contribute to the analysis of
collaborative text writing. In educational scenarios the proposed
method enables tutors to investigate how students relate important
domain concepts, and therefore, gain insights into their (possibly
different) mental conceptualization. Thus, different views and
focuses of students become visible. In future work the
visualization will be integrated in an interactive application that
supports the visual exploration of the resulting network through
improved node and edge highlighting as well as facilities for data
gathering and network reduction using k-core analysis. Regarding
the interpretation and the analysis of the extracted networks the
concept extraction can be adapted in such a way that the concepts
and relations can be weighted by an expert according to their
importance for the domain. This would result in more compact
networks. In further evaluation the student characterizations
derived from the colored word network can be related to
selfassessment and characterizations made by a tutor.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Bär</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zesch</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <article-title>DKPro Similarity: An Open Source Framework for Text Similarity</article-title>
          .
          <source>In Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics: System Demonstrations</source>
          . (Sofia, Bulgaria).
          <source>Association for Computational Linguistics</source>
          ,
          <year>2013</year>
          ,
          <fpage>121</fpage>
          -
          <lpage>126</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Bader</surname>
            ,
            <given-names>G. D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hogue</surname>
            ,
            <given-names>C. W.</given-names>
          </string-name>
          <article-title>An automated method for finding molecular complexes in large protein interaction networks</article-title>
          .
          <source>BMC Bioinformatics</source>
          ,
          <volume>4</volume>
          ,
          <issue>1</issue>
          (
          <year>2003</year>
          ),
          <fpage>2</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Belanger</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Thornton</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Bioelectricity</surname>
            :
            <given-names>A Quantitative</given-names>
          </string-name>
          <string-name>
            <surname>Approach Duke University's First</surname>
            <given-names>MOOC</given-names>
          </string-name>
          . (
          <year>2013</year>
          )
          <article-title>Technical Report</article-title>
          , Duke University.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Blei</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
          </string-name>
          , A. Y. and
          <string-name>
            <surname>Jordan</surname>
            ,
            <given-names>M. I.</given-names>
          </string-name>
          <article-title>Latent dirichlet allocation</article-title>
          .
          <source>J.Mach.Learn.Res.</source>
          ,
          <volume>3</volume>
          (mar
          <year>2003</year>
          ),
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Carley</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Palmquist</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Extracting, representing, and analyzing mental models</article-title>
          .
          <source>Social forces</source>
          ,
          <volume>70</volume>
          ,
          <issue>3</issue>
          (
          <year>1992</year>
          ),
          <fpage>601</fpage>
          -
          <lpage>636</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Diesner</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Carley</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M.</surname>
          </string-name>
          <article-title>Revealing social structure from texts: meta-matrix text analysis as a novel method for network text analysis</article-title>
          .
          <source>Causal mapping for information systems and technology research: Approaches</source>
          , advances, and illustrations,
          <year>2005</year>
          , pp.
          <fpage>81</fpage>
          -
          <lpage>108</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Eckart de Castilho</surname>
          </string-name>
          , R. and
          <string-name>
            <surname>Gurevych</surname>
          </string-name>
          , I.
          <source>In Proceedings of the Workshop on Open Infrastructures and Analysis Frameworks for HLT. Association for Computational Linguistics</source>
          ,
          <year>2014</year>
          ,
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Flower</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hayes</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          <article-title>A Cognitive Process Theory of Writing</article-title>
          .
          <source>College Composition and Communication</source>
          ,
          <volume>32</volume>
          ,
          <issue>4</issue>
          (
          <year>1981</year>
          ), pp.
          <fpage>365</fpage>
          -
          <lpage>387</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Harrer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moskaliuk</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kimmerle</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Cress</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <article-title>Visualizing wiki-supported knowledge building: coevolution of individual and collective knowledge</article-title>
          .
          <source>In Anonymous International. Symposium on Wikis</source>
          .
          <year>2008</year>
          19:
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          :
          <fpage>9</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          and Tarau,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <article-title>TextRank: Bringing order into texts</article-title>
          .
          <source>In Proceedings of the EMNLP. Association for Computational Linguistics</source>
          , (Barcelona, Spain),
          <year>2004</year>
          ,
          <fpage>404</fpage>
          -
          <lpage>411</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Paranyushkin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Identifying the pathways for meaning circulation using text network analysis</article-title>
          .
          <source>Technical Report Nodus Labs</source>
          , Berlin, (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Riche</surname>
            ,
            <given-names>N. H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Chevalier</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <article-title>iChase: Supporting Exploration and Awareness of Editing Activities on Wikipedia</article-title>
          .
          <source>In Proceedings of the International Conference on Advanced Visual Interfaces</source>
          . (Roma, Italy). ACM, New York, NY, USA,
          <year>2010</year>
          ,
          <fpage>59</fpage>
          -
          <lpage>66</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Sabrina</given-names>
            <surname>Ziebarth</surname>
          </string-name>
          and Hoppe,
          <string-name>
            <surname>H. U.</surname>
          </string-name>
          <article-title>Moodle4SPOC: A Resource-Intensive Blended Learning Course</article-title>
          .
          <source>In Proceedings of the European Conference on Technology Enhanced Learning. (Graz, Austria)</source>
          ,
          <year>2014</year>
          ,
          <fpage>359</fpage>
          -
          <lpage>372</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Scardamalia</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bereiter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <article-title>Computer Support for Knowledge-Building Communities</article-title>
          .
          <source>The Journal of the Learning Sciences, 3</source>
          ,
          <issue>3</issue>
          (
          <year>1993</year>
          ), pp.
          <fpage>265</fpage>
          -
          <lpage>283</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Schvaneveldt</surname>
            ,
            <given-names>R. W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Durso</surname>
            ,
            <given-names>F. T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Dearholt</surname>
            ,
            <given-names>D. W.</given-names>
          </string-name>
          <article-title>Network structures in proximity data</article-title>
          .
          <source>Psychol</source>
          . Learn. Motiv.,
          <volume>24</volume>
          (
          <year>1989</year>
          ),
          <fpage>249</fpage>
          -
          <lpage>284</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Southavilay</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yacef</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reimann</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>R. A.</given-names>
          </string-name>
          <article-title>Analysis of collaborative writing processes using revision maps and probabilistic topic models</article-title>
          .
          <source>In Proccedings of the Learning Analytics and Knowledge Conference. (Leuven, Belgium)</source>
          ,
          <year>2013</year>
          ,
          <fpage>38</fpage>
          -
          <lpage>47</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Viegas</surname>
            ,
            <given-names>F. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wattenberg</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Dave</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Studying cooperation and conflict between authors with history flow visualizations</article-title>
          .
          <source>In Proceedings of the SIGCHI conference on Human factors in computing systems. ACM</source>
          ,
          <year>2004</year>
          ,
          <fpage>575</fpage>
          -
          <lpage>582</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Villalon</surname>
            ,
            <given-names>J. J.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Calvo</surname>
            ,
            <given-names>R. A.</given-names>
          </string-name>
          <string-name>
            <surname>Concept Map</surname>
          </string-name>
          <article-title>Mining: A definition and a framework for its evaluation</article-title>
          .
          <source>In Proceedings of the International Vonference on Web Intelligence and Intelligent Agent Technology</source>
          ,
          <year>2008</year>
          , IEEE,
          <year>2008</year>
          ,
          <fpage>357</fpage>
          -
          <lpage>360</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Wasserman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Faust</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Social Network</surname>
          </string-name>
          <article-title>Analysis: Methods and Applications</article-title>
          . Cambridge University Press,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Wild</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bulow</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Monitoring conceptual development with text mining technologies: CONSPECT</article-title>
          .
          <source>In Proceedings of eChallenge</source>
          ,
          <year>2010</year>
          ,
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>