=Paper= {{Paper |id=Vol-2226/paper6 |storemode=property |title=Quantifying Collaboration in Synchronous Document Editing |pdfUrl=https://ceur-ws.org/Vol-2226/paper6.pdf |volume=Vol-2226 |authors=Adrian Pace,Louis Baligand,Stian Håklev,Jennifer K. Olsen,Nore de Grez,Bram De Wever |dblpUrl=https://dblp.org/rec/conf/swisstext/PaceBHOGW18 }} ==Quantifying Collaboration in Synchronous Document Editing== https://ceur-ws.org/Vol-2226/paper6.pdf
               Quantifying Collaboration in Synchronous Document Editing
Adrian Pace1 , Louis Baligand1 , Stian Håklev1 , Jennifer K. Olsen1 , Nore de Grez2 , Bram De Wever2

                      1
                          École Polytechnique Fédérale de Lausanne, Lausanne, Switzerland
                                          2
                                            Ghent University, Ghent, Belgium
                          {adrian.pace, louis.baligand, stian.haklev, jennifer.olsen}@epfl.ch
                                       {nore.degrez, bram.dewever}@ugent.be

                                                                       85% of university writing and company reports are collab-
                                                                       oratively written (Ede and Lunsford, 1990).
                           Abstract                                        Within educational sciences, researchers have looked at
                                                                       the effect of collaborative writing on the quality of the writ-
     Collaborative synchronous writing tools like                      ten product, but also its effect on students learning. For ex-
     Google Docs and Etherpad let multiple users edit                  ample, Storch (2005) found evidence that the students who
     the same document and see each others edits in                    write collaboratively produce shorter but better documents
     near real-time to simplify collaboration and avoid                in respect to accuracy, complexity and cogency. However,
     merge-conflicts. These tools are used extensively                 it is still an open question what collaborative writing pro-
     across many domains, including education, in                      cesses may be effective and how to measure these pro-
     both research and industry. The very nature of                    cesses, which we begin to explore in this paper.
     needing to constantly synchronize state between                       Collaborative writing generates very granular data,
     multiple users means that very granular editing                   which has a great potential for further analysis, but is very
     data is automatically captured and stored. In the-                difficult to collate and analyse. An example of a single
     ory, this data could provide important insights into              database entry might be ”John inserted the letter ’a’ at posi-
     the editing process, the contributions of the dif-                tion 380”, and during an editing session, thousands of these
     ferent users, how the text developed over time,                   entries are generated. Different editors also store the data
     and other questions relevant to researchers study-                slightly differently. For example, ShareDB encodes data
     ing writing from different theoretical and method-                into a JSON object, for example:
     ological angles. However, this extreme granular-
     ity of the data (down to individual key presses),                 {"seq":20,"v":20,"op":[{"p":["text",19],
     makes analysis very complex. Most of the re-                      "si":"d"}],"m":{"ts":1526393562757},
     search focused on automatic analysis of collabo-                  "d":"doc"}
     rative writing to date has focused on asynchronous                   Whereas Etherpad compresses the data about an opera-
     writing, and looked at the ”diffs” between one                    tion into a string like this:
     editing session and the next. In this paper, we
     present a method and a tool to construct infor-                   Z:z>1|2=m=b*0|1+1\$\n
     mative operations from text data, as well as pre-
                                                                          In this paper, we introduce a tool written in Python1 ,
     liminary metrics for measuring the collaborative
                                                                       that can parse the native database formats of several col-
     writing process. Additionally, our method adds to
                                                                       laborative editing tools or libraries (currently we support
     previous work in that it can be used to assess the
                                                                       Etherpad and ShareDB, but adding support for other for-
     writing during the writing process rather than just
                                                                       mats would require minimal changes). Beyond parsing
     being applied to an end product.
                                                                       individual operations, we combine atomic operations into
                                                                       larger meaningful operations such that a sentence written
 Introduction                                                          consecutively, without any pause larger than n seconds, and
 Collaborative writing is used extensively in many domains,            without moving to another location on the page, goes from
 such as families planning a trip, students collaborating on           being a number of atomic events with time-stamps, to being
 an essay, researchers writing grant proposals, and company            a single operation, with a start- and an end-time.
 staff coordinating their work. Studies have even shown that              However, to properly assign user intention to atomic
                                                                       events, we need to make decisions about cut-off points –
 In: Mark Cieliebak, Don Tuggener and Fernando Benites (eds.):         what is the n seconds threshold appropriate for deciding
 Proceedings of the 3rd Swiss Text Analytics Conference (Swiss-
                                                                          1
 Text 2018), Winterthur, Switzerland, June 2018                               https://github.com/chili-epfl/FROG-analytics




                                                                  50
that a sentence was written in two writing events as op-              Literature Review
posed to in a single event? If I am writing a sentence,
                                                                      Technology for collaborative writing
quickly backtrack to fix a spelling mistake, and continue
writing, should this be modelled as a series of small editing         The two main technologies for collaborative writing are
operations, or a single operation whose fundamental goal it           wikis, and text editors or word processors that allow for
is to insert the sentence?                                            synchronous editing. Wikis do not support synchronous
                                                                      editing. Rather, each user locks a page for a given time,
   To some extent, there can never be a perfect answer to
                                                                      makes the edits he or she wishes, and commits this change,
this, since it will depend on the research questions of the
                                                                      often with an edit summary, and a way for others to see a
researcher. If I am investigating editing behaviour at a very
                                                                      ”diff” of their changes. Given the high granularity of this
granular level, I might need access to all the spelling mis-
                                                                      data and the accessibility of this data through public sites
take corrections and backtracks. However, if I am inter-
                                                                      such as Wikipedia, there is a large body of research cover-
ested in how people add and reorganize information, coor-
                                                                      ing collaborative work in wikis. However, because the edits
dinate the development of a document and negotiate ideas,
                                                                      are made asynchronously, the research in this area does not
then I might prefer the writing events to be tracked at a
                                                                      provide insights into the collaborative writing process of
larger grain-size.
                                                                      students who are working on a document at the same time.
    By analyzing authentic data from students collabo-                    Tools that allow for synchronous editing are typically
ratively editing summaries, we have begun to explore                  based on the theory of Operational Transforms (OT, Sun
whether we can detect natural breakpoints for different op-           et al., 2004), which requires each edit to be decomposed
erations. We present an overview of the informed deci-                down to a set of atomic operations, such as ”add the char-
sions we have made, which we believe would be appropri-               acter M at position 480”, delete, move, change format (for
ate and useful to the majority of researchers. Nevertheless,          rich text editors), etc. Since the synchronization of data
our code is designed in such a way that it is very easy to            will never be perfectly real-time, we might have two users,
customize these settings as appropriate.                              where user1 deletes character 480, and user2 inserts a new
                                                                      character at position 480 at the same time. If user2’s edit
   Once we are able to detect these meaningful operations,            arrives at the server before user1’s, the result would be
we can replay through the entire history of a document, and           for user1’s operation to delete the newly inserted charac-
for each meaningful operation, also annotate it with addi-            ter, which would not correspond to user1’s intention.
tional contextual information, so that instead of merely stat-            To prevent this, OT uses the concept of document ver-
ing that ”John wrote the following sentence at 3PM”, we               sions. Each client keeps track of its latest document ver-
can add that ”John added a sentence to the end of a para-             sion, and reports this to the server with the operation. If
graph that had been authored by Chen Li”, or that ”John               the server receives two operations run against the same
added a sentence, while Rachida was editing a paragraph               document version, it needs to transform them into a sin-
two pages away”.                                                      gle operation that does not cause any conflict (in this case,
                                                                      delete the existing character at position 480, and then in-
   From this contextual data, we can begin to determine
                                                                      sert the new one, thus preserving both user1 and user2’s
metrics and indicators to summarize different phenomena
                                                                      intentions).
of interest. In the final section, we introduce a set of early
                                                                          To manage the coordinated version numbers, the OT ap-
indicators that may be predictive of quality of collabora-
                                                                      proach requires a central server and might run into prob-
tion or editing style, and we look at whether there is any
                                                                      lems with scaling above a certain number of people in
correlation between the editing preferences of the students,
                                                                      the same document. Conflict-free Replicated Data Types
and the metrics extracted from their collaboratively edited
                                                                      (CRDTs) have been suggest as a solution. These are data
documents.
                                                                      types that contain enough context that the merge could hap-
   The main contribution of our work is to create scores              pen on any of the clients, without requiring a central co-
that can provide insight into the collaborative performance           ordinating server (André et al., 2013). Currently, our ap-
of the authors of a document. It is important to note that            proach is focused on OT data, as that is the most widely
we do not know if a high score is good or not in terms of             supported approach, but it could possibly be extended to
collaboration. However, our aim was to measure numeri-                support CRDT-systems in the future.
cally the aspects that could help assess the collaboration in
the writing process.                                                  Automatic analysis of collaborative writing
   We begin by introducing the data we used. We then de-              One significant thread of research in the field of collabora-
scribe the implementation of our approach. Finally, we de-            tive writing has been centered around providing feedback
fine the metrics and look at how they behave on genuine               to teachers in order to assess the collaborative writing ef-
data collected in an experiment.                                      forts and performances of their students. However, most




                                                                 51
of the data has been collected at a very coarse granular-                The students were then randomly split into groups of
ity on long writing sequences. In the case of Wikipedia,              three (n=17) and asked to provide a synthesis based on
researchers look at a few snapshots during the writing pro-           three provided sources within 90 minutes. The instructions
cess and analyze the differences within an article at differ-         asked for a synthesis text of 800-1000 words, which would
ent points in time (Hu et al., 2016).                                 summarize the most important information from the source
    With Google Docs, a number of papers have conducted               texts in an integrated and well-organized manner. As a
fairly sophisticated analyses of collaborative writing based          preparation for this task, each group member was provided
on automatically requesting the entire document snapshot              with another source text which they needed to summarize
each minute, or even more frequently, and extracting the              in advance.
user actions by calculating the ”diffs” between each snap-               The students were not allowed to talk during the session,
shot (McNely et al., 2012). One example of this approach is           but communicated with each other via chat and comments
WriteProc, based on a taxonomy of the collaborative writ-             in Etherpad (the writing environment). Finally, students
ing process by Lowry et al. (2004), which traces semantic             were asked to fill in an online questionnaire about their ex-
changes between different versions and changing concepts              perience. In this paper, we use data from the collaborative
(Southavilay et al., 2009, 2010, 2013). However, these are            writing process that occurred within the second stage of
also asynchronous documents where authors rarely wrote                the experiment to conduct an unsupervised analysis of the
simultaneously (Wang et al., 2015; Sun et al., 2014)                  traces from the collaborative writing sessions.
    The only paper we found that looked directly at the in-
dividual operations was from Liu et al. (2017), which used            Data Logs
regular expressions to classify the compressed Etherpad
                                                                      Our data was collected from students working on a set of
operations as either add, delete or move, and then quan-
                                                                      Etherpad documents. Etherpad2 is a highly customizable,
tified the number of different types of operations over dif-
                                                                      open source, online editor that provides collaborative edit-
ferent time windows.
                                                                      ing in real-time. It allows multiple users to edit the same
    To summarize, existing research has been tied to a spe-           text document through a web page interface.
cific tool (Google Docs API, Etherpad database), and there
                                                                          Etherpad operations are saved within a database as ad-
has not been any attempt to create an abstract intermedi-
                                                                      ditions, deletions, or formatting of text (e.g., bolding text).
ate representation that would allow us to apply the same
                                                                      Each change (also referred to as a writing event) is defined
higher level analysis to data from multiple platforms. Also,
                                                                      by its author, a timestamp, the document version (incre-
most analyses are done by comparing ”diffs” from snap-
                                                                      mented at each writing event), and the modification, which
shots, rather than directly accessing the operations, which
                                                                      consists of the position in the document at which the event
leads to data loss in situations of synchronous editing (e.g.,
                                                                      takes place and what characters to delete or add. This al-
impossible to tell who has edited what). Finally, the only
                                                                      lows a view of the document to be reconstructed over time
attempt at looking at the actual operations did not try to ex-
                                                                      rather than only keeping the final product.
tract semantics, but rather focused on quantifying the kinds
                                                                          For our analysis, we are interested in how users collab-
of operations over time.
                                                                      orate in writing content, so we only focus on the addition
    In this paper, we work on a much finer granularity and            and deletion actions taken within the document. Although
look at individual edits instead of differences between the           our analysis focused on Etherpad data, our implementation
document in various points in time. Instead of focusing on            works with any software that could log the same types of
the semantics, we created metrics based on the location in            data streams as long as the writing events are stored in fine
the document of the authors’ contributions and how bal-               granularity. Depending on the format the database users to
anced the writing styles of the users are.                            store the changes, a simple parsing function would have to
                                                                      be coded to fetch the document changes from the database
Data Collection                                                       into our system.

Context                                                               Implementation
To acquire authentic synchronous writing data, we orga-               Our system is an application written in Python that sends
nized an experimental session with 50 Master students                 the metrics (discussed below) for a selection of documents
in educational sciences. The session consisted of three               defined by the user to a server. At regular configurable
phases. The students first completed an online individ-               intervals (5 seconds by default), it looks for changes in
ual questionnaire that queried learners implicit writing be-          tracked documents and send the updated metrics.
liefs measured with the Writing Beliefs Inventory (White                 At each time point, the application collects the relevant
and Bruning, 2005) and students’ individual writing strate-           document changes that happened since the last update from
gies based on the Writing Style Questionnaire (Kieft et al.,
                                                                         2
2008).                                                                       http://etherpad.org/




                                                                 52
the editor database. For each of the updates, relevant in-                   ate a Write from simply copying and pasting text into
formation is stored around the position of the event, text                   the document.
added/deleted, author, and timestamp. The information
collected at each stage is used to infer the writing oper-              In order for the operations to be relevant, we need to care-
ations that occurred and subsequently, the metrics on the               fully select the parameters that distinguish them from each
writing process.                                                        other (e.g., the threshold number of added characters to ei-
                                                                        ther classify an addition as a Write or as an Edit).
                                                                            To determine this threshold between Write and Edit, we
Writing Operations                                                      plotted the distribution of the number of Write, Delete,
The data collected from Etherpad is very granular, and each             Paste and Edit operations with respect to the threshold
data point taken separately does not have meaning, besides              number of added characters to classify the operation as
contributing to our ability to reconstruct the entire docu-             Edit. We selected a length of 15 characters because the
ment at any point in time. To give more meaning to the                  distribution stays relatively constant from this point. The
writing process, we used the saved changes within the doc-              average number of characters in an English word is 5.1
ument to calculate meaningful operations that begin to cor-             (Bochkarev et al., 2012), which supports our decision as it
respond with actual behavioural and cognitive processes.                roughly means that deleting or adding less than three words
   The small changes that are saved within Etherpad (i.e.,              is an Edit.
additions and deletions) can be grouped together when they                  The application also groups the writing events into para-
occur continuously, to provide writing behaviors. If users              graphs, collections of writing events that are currently lo-
stop for a coffee break or begin editing at another location            cated on the same line. This gives more insight into the
in the document, a new operation is formed. If the author               context in which an operation was written: Is this opera-
has taken a short break, but not begun a new operation by               tion occurring in a paragraph written by a single author? Is
writing somewhere else in the document, he or she can go                it a significant change to the paragraph given the length of
back to the piece of text that was being edited, and continue           the paragraph?
editing without starting a new operation.                                   This context is further extended with various details
   We classify the operations into four different types:                such as its length compared to the document, whether its
                                                                        the first operation of the day, and whether there were other
  • Write: An operation is classified as a Write if the au-             people writing in the document at the time.
    thor enters more characters than a set threshold. We
    consider this type of operation as representing draft-              Writing Metrics
    ing the bulk of the text. It occurs, for example, when
                                                                        To begin to assess the collaborative writing process, we cre-
    authors begin writing an essay and are adding ideas.
                                                                        ated eight different metrics that could be applied to the doc-
    The Write operation contains mostly addition changes
                                                                        ument at any point in time (Table 1). The metrics are de-
    but may have some deletions, as they form part of the
                                                                        rived from understanding the operation type that took place
    writing process.
                                                                        and the context that the operation is within. These metrics
  • Edit: An Edit is similar to a Write in that it can con-             fall into three basic categories: time, operations, and inte-
    sist of both addition and deletion changes. However,                gration. Except when noted, each of the metrics has a score
    an operation is classified as an Edit when the number               between 0 and 1.
    of character changes falls below the threshold for a                    The timing metrics are all related to when students were
    Write. Edits often occur when the authors review an                 writing in the document. Within this category, the metrics
    essay and fix typos or change words.                                include the count of day breaks (DB), the count of short
                                                                        breaks, and the amount of synchronicity in the document.
  • Delete: If an operation consists of an author remov-                The number of day breaks tracks whether the document
    ing more than a certain number of characters (deletion              was written across multiple days or in a single day. If there
    change), then it is classified as a Delete. The idea of             was more than eight hours between any two changes, the
    this type of operation is to observe when an author                 count of the day breaks is incremented. This metric allows
    does not simply remove a word (this is considered as                us to have a rough estimate of duration for the editing pro-
    an Edit or as part of a Write), but when he or she re-              cess.
    moves a significant amount of characters, such as a                     The second timing metric is the number of short breaks
    whole sentence or paragraph.                                        (SB) that were taken. The number of short breaks is cal-
                                                                        culated by assessing the number of instances where there
  • Paste: If the writer adds more than several characters              are no changes for at least 10 minutes (but less than 8
    (addition change) in one single writing event (which                hours). The short breaks are important to measure, because
    are sampled every few milliseconds), then the opera-                they can be an indicator of different writing styles, where
    tion is classified as a Paste. This is useful to differenti-        some people prefer to plan before writing and others do




                                                                   53
                                                                    of these operations over time, for the whole group, or for
     Table 1: Metrics to assess collaborative writing.
                                                                    individual users, helps us to understand the editing flow.
                     Time Metrics
                                                                       The overall operations (OO) metric captures the divi-
 Day Breaks                 Number of days the docu-                sion of different metrics across the entire document. For
                            ment was written over                   each type of operation, the proportion is calculated given
 Short Breaks               Number of pauses for longer             the total number of operations in the document.
                            than 10 minutes
 Synchronicity              Percentage of text that was                  # of char. classified as {delete, edit, paste, write}
                            written in a synchronous                                        # of characters
                            manner
                  Operation Metrics                                 A score of 0 indicates that there are no operations of this
 Overall Operations         Proportion each operation               type in the document while a score of 1 indicates that all of
                            was used in the document                the operations are of this type. This metric may help us to
 User Operations            Proportion that an author               better understand the writing style of the users.
                            used an operation normal-                   The second metric in the operations category is the user
                            ized by the total use of the            operations (UO), which reflects the distribution of the dif-
                            operation                               ferent operations across the users in the pad. This metric
                 Integration Metrics                                reflects how much the different users are balancing the dif-
                                                                    ferent operations among themselves and, as with the over-
 User Contribution          Balance of contributions be-
                                                                    all operations metric, there is a separate number for each
                            tween authors
                                                                    operation.
 Paragraph Integration      Interleaved or blocked para-
                            graphs by different authors                       1           X                                 1
 Paragraph Contribution Balance of contributions                                                    proptype (i) log
                                                                        log(#authors)                                  proptype (i)
                            within paragraphs between                                   i∈authors

                            authors                                     For example, if one user does all of the editing, then
not. These different writing styles may impact the success          there would be a score of 0 for the editing. However, if
of collaboration process depending upon the alignment of            there is an even distribution among the users, then there
the writing styles between the members of the group.                would be a score of 1. This metric may be beneficial in un-
                                                                    derstanding the roles that the users take within their group.
                                                                        The third category of metrics involves the integration
   The final timing metric is the synchronicity (SYN) of            of writing within the document, including user contribu-
the writing in the document. The synchronicity is measured          tion, paragraph integration, and within paragraph integra-
taking the number of synchronously written characters nor-          tion. The user contribution (UC) metric measures the bal-
malized over the total number of characters written.                ance of the writing contributions between users. The met-
           # of synchronously typed characters                      ric is close to 1 when the participation is equal between the
                                                                    users, and is close to 0 when a single author has done the
                     # of characters
                                                                    majority of the writing. This metric is important to track in
A synchronous character is classified as synchronous if two         a collaborative setting because it can be indicative of social
or more users are writing within three minutes of each              loafing within the group.
other. When all of the text was written in a synchronous
                                                                                  1            X                          1
manner, then the score is 1, and if there is no overlap be-                                              prop(i) log
tween the writing times of the users, the score is 0. The                   log(#authors)
                                                                                             i∈authors
                                                                                                                       prop(i)
synchronicity of authoring is an important measure of the
collaborative writing dynamics, especially when the users              The second integrative metric is the paragraph integra-
work in an otherwise synchronous setting, because it may            tion (PI) within the document. The paragraph integration
offer insights into the roles and dynamics of the group.            measures how interleaved the paragraphs written by differ-
   The second categorization of the metrics involves the            ent users are. The measure will be 1 if the main author of
operation types that are present within the document. As            each paragraph is alternating, or close to 0 if many blocks
mentioned above, to better understand how the users were            in a sequence have the same primary author (for example,
writing their document, we rolled-up the basic additions            one author mostly wrote the four first paragraphs, and the
and deletions stored by Etherpad into four types of opera-          other author wrote the four next ones). This metric may be
tions: Delete, Write, Edit, and Paste. Depending upon the           an indicator of the collaboration that is occurring within the
role of a user in a group or the stage of the writing that          group. When writing text, some group members will divide
the group is currently engaged in, the relative frequency of        the labor and have each member write a different section of
the types of operations may change. The relative frequency          the paper in a more cooperative style. Other groups will




                                                               54
work in a more collaborative style with all group members
contributing to each section.
       # of author alternations between Paragraphs
                   # of Paragraphs − 1
    The final metric, within paragraph integration, is a mea-
sure of the extent to which the users in a group contributed
to each of the paragraphs equally (PC). It will be close to 0
if a paragraph was written by a single user, and 1 if all au-
thors contributed equally to a paragraph. As with the pre-            Figure 1: Evolution of the PI over the 90-minute assign-
vious metric, within paragraph integration can be used as a           ment on all documents
measure of the collaboration between the group members.
              X                         
                          length(p)                1
           p∈paragraphs
                        length(total)     log(#authors in p)
                                                     
                  X                           1
                             propp (i) log           
              i∈authors in p
                                           propp (i)


    To investigate if our eight metrics are measuring differ-
ent aspects of the document, we checked the correlations
between each of the metrics. We did not find any signifi-
cant correlations between the metrics, indicating that they
are all measuring separate aspects of the document writing            Figure 2: Evolution of the PC over the 90-minute assign-
process. Separately, each of our eight metrics can provide            ment on all documents
insights into the group writing process. However, the met-               Similarly, the PC metric (Fig.2), which measures the
rics do not need to be looked at in isolation. By analyzing           sharing of operations within a paragraph, also shows a
several of the metrics together, we may be able to detect ad-         growth pattern over time. Like the PI metric, this metric
ditional processes of the collaboration process that would            begins at 0 and grows to about 0.15 by the end of the ses-
not be evident from a single metric.                                  sion. However, the pattern over time is very different. For
                                                                      the first half of the session, the PC metric remains relatively
Metrics Over Time                                                     close to 0, indicating that users work on their own individ-
Although all of our metrics can be applied to a finished              ual part of the document, without attempting to integrate
writing product, one of the strengths of our approach is that         their work.
the different metrics can be calculated in real-time, to both            In the second half of the session, the score begins to
eventually adapt to students, and to track how the writing            grow at a more linear rate, indicating that there is much
process changes as time progresses.                                   more sharing between paragraphs. Although the main au-
    In order to study the behavior and evolution of our met-          thor of the paragraphs may have stayed the same (as in-
rics over time, we calculated the metrics of each pad at              dicated by not much of a rise in the PI score), the group
given intervals. We split the lifespan of the document, i.e.          members were looking at each others work more, as indi-
90 minutes, in 32 linearly separated time-slices. For each            cated by the higher PC score.
time-slice, we display a box plot showing the spread of the
different metrics across the seventeen groups. Below we
present the analysis of three different metrics (paragraph in-
tegration, paragraph contribution, and synchronicity), and
how they develop over time.
    The PI (Fig.1) measures the amount that the members of
the group are interleaving their different paragraphs. In our
data set, we can see that the score begins at 0, when no one
has yet written. As the authors begin to add more text, we
can see that the score rises to about 0.15 by the end of the
first half of the session. In the second half of the session,         Figure 3: Evolution of the synchronicity over the 90-
the metric remains relatively constant, indicating that there         minute assignment on all documents
is no additional integration of paragraphs between group
members.                                                                 Finally, we investigated how the synchronicity met-




                                                                 55
ric changed over time (Fig.3). Because the groups were               live visualizations of the collaborative process in multiple
formed and were asked to produce their summary within a              groups for workshop organizers or teachers, and in the fu-
90-minute timeframe, we would expect to see that the doc-            ture we might consider adding user-facing visualizations or
ument was written very synchronously. From the data, this            even prompts and adaptive interfaces to support group col-
is what we find. At the beginning of the document when the           laboration.
group is just getting started, the score is close to 0. How-             Finally, the analyses described in this paper only con-
ever, in Fig.3, we can see that it quickly moves up to around        cern themselves with behaviour and could be applied to any
0.4 within a few time-slices. For the remainder of the doc-          language. However, by adding semantic context to the data
ument, the scores remains between 0.6 and 0.8, which is an           – for example using word2vec-like methods to measure se-
indicator of high synchronicity.                                     mantic distance between two users’ contributions over time
                                                                     (is there convergence?) – we might be able to further un-
Discussion and Conclusion                                            derstand the collaborative process.
These results can give an overall idea of the behavior of the
metrics and of their importance in gaining insight of the
document writing process. As we can observe, the metrics
                                                                     References
evolve through time. The initial increase in the PI at the           André, L., Martin, S., Oster, G., and Ignat, C.-L. (2013). Sup-
beginning of the document writing process indicates that               porting adaptable granularity of changes for massive-scale col-
                                                                       laborative editing. In Collaborative Computing: Networking,
authors begin by editing separate parts of the document,               Applications and Worksharing (Collaboratecom), 2013 9th In-
with very little intra- or inter-paragraph coordination.               ternational Conference Conference on, pages 50–59. IEEE.
   As the writing processes proceeds, the authors stop cre-
ating new paragraphs and begin editing the existing ones,            Bochkarev, V. V., Shevlyakova, A. V., and Solovyev, V. D. (2012).
which naturally leads to more collaboration and textual in-            Average word length dynamics as indicator of cultural changes
                                                                       in society. CoRR, abs/1208.6109.
tegration. This insight can help the supervisor of the au-
thors in determining if each writer write their own block of         Ede, L. and Lunsford, A. (1990). Singular Texts/plural Authors:
text or whether they have contributed in every main aspects             Perspectives on Collaborative Writing. Southern Illinois Uni-
of the document.                                                        versity Press.
   In addition, as time increases, the evolution of the PC           Hu, X., Ng, T.-D. J., Tian, L., and Lei, C.-U. (2016). Automating
shows that authors tend to write more inside of each oth-              assessment of collaborative writing quality in multiple stages:
ers paragraph. It would seem they write their own block                The case of wiki. In Proceedings of the Sixth International
of text at the beginning and then start gradually collaborat-          Conference on Learning Analytics & Knowledge, LAK ’16,
ing by writing in each others contributions. The score still           pages 518–519, New York, NY, USA. ACM.
remains relatively low meaning that there still is a main            Kieft, M., Rijlaarsdam, G., and van den Bergh, H. (2008). An
author per paragraph and that the contributions from other              aptitude-treatment interaction approach to writing-to-learn.
authors stay small.                                                     Learning and Instruction, 18(4):379 – 390.
   Moreover, the Synchronous score evolution shows that
the documents are written by at least two authors at the             Liu, M., Pardo, A., and Liu, L. (2017). Using learning analytics
                                                                        to support engagement in collaborative writing. International
same time during the overall duration of the assignment                 Journal of Distance Education Technologies, pages 79–98.
except at the very beginning.
   The aim of our metrics is to provide real-time insights           Lowry, P., Curtis, A., and Lowry, M. (2004). A taxonomy of
on the way authors contribute in writing an online docu-               collaborative writing to improve empirical research, writing
                                                                       practice, and tool development. Journal of Business Commu-
ment. They were designed to analyze whether the authors                nication, pages 66–99.
are collaborating between them and thus focus on the inter-
action between the writers. We are currently planning new            McNely, B. J., Gestwicki, P., Hill, J. H., Parli-Horne, P., and John-
experiments that will give us more detailed data about how             son, E. (2012). Learning analytics for collaborative writing:
collaborative writing processes manifest themselves in the             A prototype and case study. In Proceedings of the 2Nd Inter-
                                                                       national Conference on Learning Analytics and Knowledge,
data.                                                                  LAK ’12, pages 222–225, New York, NY, USA. ACM.
   We are also planning to analyze the quality of the written
documents, as well as the knowledge post-test, to see if we          Southavilay, V., Yacef, K., and Calvo, R. (2009). Writeproc: A
can correlate the proposed quality metrics in this document,            framework for exploring collaborative writing processes. In
with actual differences in quality or learning gains.                   ADCS 2009: Proceedings of the Fourteenth Australasian Doc-
                                                                        ument Computing Symposium, 4 December 2009, pages 543–
   The tool we have developed can provide these metrics                 548.
live during an editing session, and it will be integrated
in our collaborative learning platform FROG3 , to provide            Southavilay, V., Yacef, K., and Calvo, R. (2010). Process mining
                                                                        to support students’ collaborative writing. Educational Data
   3
       https://github.com/chili-epfl/FROG                               Mining, pages 257–266.




                                                                56
Southavilay, V., Yacef, K., Reimann, P., and Calvo, R. (2013).
   Analysis of collaborative writing processes using revision
   maps and probabilistic topic models. In Proceedings of the
   Third International Conference on Learning Analytics and
   Knowledge, pages 38–47. ACM.

Storch, N. (2005). Collaborative writing: Product, process, and
   students reflections. Journal of Second Language Writing,
   14(3):153 – 173.

Sun, D., Xia, S., Sun, C., and Chen, D. (2004). Operational trans-
   formation for collaborative word processing. In Proceedings
   of the 2004 ACM conference on Computer supported cooper-
   ative work, pages 437–446. ACM.
Sun, Y., Lambert, D., Uchida, M., and Remy, N. (2014). Collab-
   oration in the cloud at Google. In WebSci 2014 - Proceedings
   of the 2014 ACM Web Science Conference.
Wang, D., Olson, J. S., Zhang, J., Nguyen, T., and Olson, G. M.
  (2015). Docuviz: Visualizing collaborative writing. In Pro-
  ceedings of the 33rd Annual ACM Conference on Human Fac-
  tors in Computing Systems, CHI ’15, pages 1865–1874, New
  York, NY, USA. ACM.

White, M. J. and Bruning, R. (2005). Implicit writing beliefs and
  their relation to writing quality. Contemporary educational
  psychology, 30(2):166–189.




                                                                     57