<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>International Conference "Internet and Modern Society", June</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>An Interoperable Platform for Multi-Grain Text Annotation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Svetlana Sheremetyeva</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>South Ural State University</institution>
          ,
          <addr-line>76 Pr. Lenina, Chelyabinsk, 454080</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>2</volume>
      <fpage>4</fpage>
      <lpage>26</lpage>
      <abstract>
        <p>In this paper, we describe an interoperable platform for creating annotated corpora in different languages and domains. It focuses on two most widely used for practical information processing tasks levels of linguistic annotations, - morphological and conceptual, that can be performed separately or combined. The platform consists of two main modules, a program shell and a knowledge base. The program shell is universal and features flexible settings that ensure its adaptation to multilingual corpora of various domains and different levels of annotation. It is provided with several interfaces for knowledge acquisition and annotation control. The annotation platform knowledge base includes language-independent and language-dependent linguistic information. The language-independent information is presented by multilingual domain ontology, while the core of the language-dependent component of the platform knowledge base includes unilingual onto-lexicons. The annotation process consists in the practical realization of ontological analysis. In performing the annotation task, the NLP techniques are used to automatically support, rather than completely replace human judgment. The platform is multifunctional, and in addition to corpora annotation, it can directly be used for different types of theoretical linguistic research, e.g., terminology analysis, cross-linguistic comparative studies, etc. The paper covers both, the platform design and its application in the frame of a real project on the conceptual annotation of the "Terrorism" domain corpora in the Russian, English and French languages.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Annotation platform</kwd>
        <kwd>interoperability</kwd>
        <kwd>domain ontology</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Corpora annotations are a prerequisite for any succession of text processing steps and its accuracy
to a large extent defines the quality of the final NLP output. It is therefore the focus of many
international theoretical and applied linguistic studies. While many practical texts processing tasks
nowadays rely on morphological labelling, conceptual annotation is becoming increasingly used as
explicit semantics is starting to play a more prominent role in computer technologies targeted to
intelligent processing of unstructured information (automatic classification, intelligent content and
trend analyzes, machine learning, machine translation, etc.) [1]. By conceptual annotation (which in
many practical projects is called “semantic”) we understand that type of semantic annotation, which is
developed for solving specific information tasks within a particular domain, and use the term to
distinguish this particular type of annotation from the high level semantic mark-up such as “human”,
“animated”, etc. For example, in the “Terrorism” domain the English lexeme “car” will be
conceptually annotated as “means of attack”, rather than “concrete”, “non-animated”, etc. We also
believe that given the ambiguity of natural language on all levels, combining different types of
annotations, e.g. morphological-syntactic and conceptual might provide a feature space that would
enhance the chances to resolve annotation ambiguity.</p>
      <p>Information processing projects that strive for high quality results require annotating
comprehensive corpora, which with any level of tags, let alone conceptual, as a starting point of
research and development is nowadays mostly done manually and on its own is a hard, costly and
time-consuming task. Taking advantage of pre-developed resources that could allow skipping the
annotation stage is quite problematic. Annotated corpora are quite sparse and often cannot be
accessed at all, because the developers restrict or completely forbid their public use. In addition, the
volume and construction principles of most existing annotated resources are non-standardized and are
tuned to only a limited number of domains and information processing tasks. The situation puts in
focus the issues of developing automated annotation tools and their interoperability to save
development effort.</p>
      <p>This paper attempts just that and presents an automated interoperable platform for creating
multigrain annotations of corpora in different languages and domains. The platform is ontology-based and
is supported by the NLP technology that complements human annotation effort. The tool is
multifunctional. In addition to automated corpora annotation, it can directly be used for different types
of theoretical linguistic research, e.g., terminology and corpora analysis, cross-linguistic comparative
studies, etc. The description covers both, the platform design and its application in the frame of a real
project on the conceptual annotation of the "Terrorism" domain e-news in the English, Russian and
French languages.</p>
      <p>The paper is structured as follows. Section 2 overviews the related work. Section 3 describes the
platform design. Section 4 is devoted to a case-study, in the frame of which the platform
development and its use is described as applied to the multilingual corpora of the "Terrorism" domain
in English, Russian and French. We conclude with the summary and future work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        While all annotated corpora created to date necessarily contain morphological markup, since the
problem of automatic (or automated) morphological analysis for a large number of languages has now
been largely solved, the need to speed up and save human effort in corpora annotation for intelligent
text processing applications prompted studies specially devoted to the development of automated
concept annotation tools. Some attempts are made to apply unsupervised approaches and completely
exclude human labor [
        <xref ref-type="bibr" rid="ref1">2</xref>
        ]. However, most popular are semi-automatic approaches that rely on NLP
techniques [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ], document structure analysis [
        <xref ref-type="bibr" rid="ref3">4</xref>
        ] or learning that requires training sets or supervision
[
        <xref ref-type="bibr" rid="ref4">5</xref>
        ]. Some works to automate annotation rely on information extraction [
        <xref ref-type="bibr" rid="ref5 ref6">6, 7</xref>
        ]. Most modern
semiautomatic annotation tools are based on ontologies where the annotation procedure is performed by
the technique of ontological analysis that results in the identification concept instances from the
ontology in texts [
        <xref ref-type="bibr" rid="ref7">8</xref>
        ]. Notwithstanding whether ontology-based annotation is done manually or
involves automation, it has a very serious limitation, - the availability of an appropriate pre-defined
and well-established ontologies. Though quite a number of ontological libraries are now publicly
available, their suitability for every particular R&amp;D project involving ontology-based conceptual
annotation is, as a rule, problematic. Most works on ontology-based annotation therefore assume the
availability of an already existing ontology [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ] or include the creation of an ontological resource as
part of annotation problem solution. Ontologies are mostly created for conceptual annotation of
domain corpora in one (often, English) language and are tuned to specific information processing
tasks, - medical record analysis [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ], personalized filtration of eNews [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ], “Terrorism” domain
content analysis [
        <xref ref-type="bibr" rid="ref11">12</xref>
        ]. Much less research can so far be found on the ontology-based annotation of
corpora in other languages. For example, in [
        <xref ref-type="bibr" rid="ref12">13</xref>
        ] research on the semantic (in fact, conceptual)
annotation of the Russian e-service domain corpus is described as presented in e-news, the system
presented in [
        <xref ref-type="bibr" rid="ref13">14</xref>
        ] focus on the conceptual annotation of the French corpus. Most often, the
methodologies for the ontology based annotation include a combination of automated technics and
manual tagging (see e.g., the works cited above).
      </p>
      <p>Given the amount of effort and time needed to construct ontologies for language-specific corpora
processing, multilingual ontologies that could be interoperable cross-linguistically got in the circle of
research interest. There is no consensus on how to understand multilingualism in ontologies. Within
one approach, ontological multilingualism is treated as understandability (or adaptation) of the
ontological labels for the users who speak different national languages. In another approach, ontology
is taken to be multilingual, if it can be applied to processing texts in different languages no matter
what language was used for concept labels. These interpretations of ontological multilingualism
directly rely on ontology definition as either a language-independent or language-dependent resource.</p>
      <p>
        Language-dependent ontologies, a well-known example of which is the famous WordNet [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ], are
thesaurus-like structures defined by the properties of a particular language. Transition to
multilingualism there is treated as the localization of ontological concept labels. The localization itself
can be approached in different ways, as a) linking the word senses of different national languages to
ontological concepts by means of a specially developed model [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ], b) translation of the ontological
concept labels from one language into another [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ] and c) manual annotation of ontological concepts
with labels worded in different languages [
        <xref ref-type="bibr" rid="ref17">18</xref>
        ]. Among other ontology-related works in the frame of
interoperability are, for example, a research devoted to the creation of universal tools for
semiautomatic building of unilingual ontologies [
        <xref ref-type="bibr" rid="ref18">19</xref>
        ] and the studies to suggest interoperable
methodologies for cross-referencing the data and meta-data of unilingual ontologies [
        <xref ref-type="bibr" rid="ref19">20</xref>
        ].
      </p>
      <p>
        Language-independent ontologies, such as Mikrokosmos [
        <xref ref-type="bibr" rid="ref20">21</xref>
        ], SUMO [
        <xref ref-type="bibr" rid="ref21">22</xref>
        ] and BFO [
        <xref ref-type="bibr" rid="ref22">23</xref>
        ], per
definition allow multilingualism in the sense of the capability to process texts in different languages,
cross-linguistic conceptual annotation included, which is provided by building lexicons of specific
languages and mapping them into the concepts of one and the same multilingual ontology.
      </p>
      <p>
        One of the annotation challenges, which is discussed in the literature, is a way to find the best set
of tags for different levels of tagging from morphological tags up to conceptual labels. The main thing
here is to decide on the amount of information coded in a single tag, and on the size of the tagset.
Though most of the discussions on the tag subject concern morphological and syntactic tagging, the
main ideas of such discussions are worth to be taken in consideration for conceptual tagging as well.
For example, in [
        <xref ref-type="bibr" rid="ref23">24</xref>
        ], the external and internal criteria in a tagset design are suggested. The external
criterion demands the tags to be able to code the distinctions in the linguistic features that are required
by the processing task. The internal tag design criterion concerns making the tagging process as
precise as possible. It is believed that a smaller and simpler tagset should improve the accuracy of
tagging, while a large number of tags causes problems for creating reliable taggers. However, larger
amount of information included in the tagset may help tag ambiguity resolution. In [
        <xref ref-type="bibr" rid="ref24">25</xref>
        ], it is claimed
that tagging precision (or accuracy) depends crucially on using a wide range of linguistic features
including lexical ones. There is thus the eternal trade-off: tag coverage versus tag precision. Another
way to significantly reduce the number of tags and nevertheless take advantage of additional linguistic
knowledge for raising annotation accuracy is the use of supertags. In general, a supertag can code a
wide range of features (morphological, syntactic, semantic and conceptual thus providing for
significant gain in tagger performance [
        <xref ref-type="bibr" rid="ref25">26</xref>
        ]. Certain attempts have been made to develop
multilingually universal tagsets. Thus, the results of the experiments carried out on different language
families (Roman vs. Slavic) are reported and the most challenging linguistic phenomena for the task
are defined. Another suggestion is to use a coarse tagset consisting of twelve cross-language lexical
categories [
        <xref ref-type="bibr" rid="ref27">28</xref>
        ]. In the frame of the MULTEXT-East (MTE) project, an attempt is made to standardize
the tagset for a range of Slavic languages, such as Romanian, Croatian, Slovenian, Czeck and,
currently, Macedonian and Russian [
        <xref ref-type="bibr" rid="ref28">29</xref>
        ]. However, many studies aimed at developing real world
applications point out that general-text tagsets usually fail on domain specific texts, and therefore,
tagsets should be domain- and application-specific [
        <xref ref-type="bibr" rid="ref29">30</xref>
        ].
      </p>
      <p>
        As noted in [
        <xref ref-type="bibr" rid="ref30">31</xref>
        ], current applications using concept tags (or codes) show three different
approaches for concept tag definition, - conventional, directed and summative that mainly differ in the
tag origin. In the conventional approach, conceptual tagging categories are derived directly from the
text data. The directed approach for the initial set of concept tags relies on a theory or relevant
research findings. Concept tags within the summative approach coincide with preliminary extracted
text keywords. Most often, conceptual tag set design concerns the ontology size and granularity. In
[
        <xref ref-type="bibr" rid="ref31">32</xref>
        ] the ontological granularity is treated in terms of ontological levels, while the reduction of the
number of concept tags is suggested by using specific levels of the so-called multilevel ontologies
which would allow meeting the interoperability demand with multi-layer corpus annotation. One
more way to save annotation effort concerns the development of cross-platform interoperability for
collaboration in automated text annotation [
        <xref ref-type="bibr" rid="ref32">33</xref>
        ]. However, in spite of the development of increasingly
convivial and hardware-independent annotation tools, the need to create intuitive, user-friendly
interfaces, which can make the annotation tools more accessible to users without special technical
skills (for example, linguists or domain experts) is more and more emphasized [
        <xref ref-type="bibr" rid="ref33 ref34">34, 35</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Design</title>
    </sec>
    <sec id="sec-4">
      <title>3.1. Overview</title>
      <p>Our research and development effort is defined by the intersection of the following criteria: (i)
domain and cross-language interoperability (ii) increase of annotation quality, (iii) automation, (iv)
user-friendliness for linguists and domain expert’s with-out special technical skills, (v) annotation
multi-granularity from morphology up to semantic and conceptual mark-up.</p>
      <p>The requirements of annotation interoperability and multi-granularity were answered by defining
the annotation methodology as the practical realization of ontological analysis based on a
domainspecific multilingual ontology, a universal program shell and a reusable tagset. In defining our tagset
features we aimed at providing a) balance between the features’ annotation relevancy and realistic
expectations to detect them automatically, b) possibility to disambiguate the tags using both statistical
measures and local context linguistic rules as the quality of annotations depends upon the judicious
application of NLP technology, and c) possibility to share the tagset between languages within a
particular domain. The integration of these methodological and technological solutions determined the
architecture of the annotation platform, which consists of two main components - a knowledge base
and a program shell. The overall architecture of the annotation platform is shown in Figure 1.
3.2.</p>
    </sec>
    <sec id="sec-5">
      <title>The knowledge base</title>
      <sec id="sec-5-1">
        <title>The annotation platform knowledge base has the following main components:</title>
        <p> language-independent semantic (conceptual) knowledge of a particular domain presented in
the domain ontology;
language-dependent linguistic knowledge of the domain in question that includes
domainrelevant unilingual lexicons of one- and multicomponent units with assigned parts-of-speech
and other morphological features relevant for each language;
 linking knowledge on mapping the domain-relevant lexical units into the ontology concepts.</p>
        <p>
          The ontology as the core of the platform knowledge is built based on the following methodological
assumptions:
 Ontology is a language-independent resource and serves intermediary between unilingual
lexicons.
 Domain ontology is integral part of upper-level ontology, Mikrokosmos [
          <xref ref-type="bibr" rid="ref20">21</xref>
          ] in our case.
 The acquisition of the domain ontological knowledge is data-driven based on multilingual
comparable domain corpora using mixed (top-down/bottom-up) acquisition techniques.
        </p>
        <p>Building the knowledge base includes extraction of domain-relevant lexemes from training
multilingual corpora followed by grouping the resulted sets into semantic (conceptual) categories
according to the sense closeness within the one language, and across languages. Thus defined
semantic categories are taken to be the seed concepts of the domain ontology and following the
Mikrokosmos structure are divided into interrelated classes of the OBJECTS, EVENTS, and
PROPERTIES top concepts. The concept labels are worded in English, while the concept meanings
are specified by concept definitions. The unilingual lists of domain-related lexemes grouped into
conceptual categories are further called onto-lexicons and cover the linking knowledge.</p>
        <p>The interoperable annotation platform program shell consists of two main blocks: a knowledge
administration and storage module and a tagger (see Fig. 1).
3.3.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>The program main modules</title>
      <p>
        The main modules of the annotation platform program are a knowledge administration and storage
module, further TransDict, and a tagger that are two updated and reused components of the earlier
developed text processing platform described in [
        <xref ref-type="bibr" rid="ref36">37</xref>
        ] that to a large extent meets our design
requirements and allowed us reducing the development effort.
      </p>
      <p>TransDict is structured as a set of unilingual lexicons with cross-referenced entries of translation
equivalents. The linguistic information associated with every unilingual entry is formalized as a tree
of features:
[semantic class/concept [language [part-of-speech [other morphology [tag]]]]]</p>
      <p>The morphological zone of the module entries contains a full wordform paradigm of a unilingual
lexeme, each associated with a supertag that codes conceptual and morphological knowledge. The
entry is meant for one sense of a lexical unit. TransDict has a powerful environment for the automated
acquisition and administrations of multilingual lexical and ontological knowledge by means of a user
interface, which visualizers the platform knowledge (Fig.1) and gives access to the following built-in
supporting tools:</p>
      <p>Configuration block that creates and edits the TransDict feature settings such as semantic classes
(concepts), languages, parts of speech, word forms and their tags; any change in the settings will
automatically propagates to all the entries in a corresponding language.</p>
      <p>Defaulter that automatically assigns entry structures and some of the feature values to new entries
according to the user-set parameters and values; for example, all semantic classes and some of the
knowledge of the English entry are automatically ported to a lexicon in another language, when
added; the knowledge can be edited.</p>
      <p>Data importer/merger that imports wordlists and/or feature values from external files and
applications both, in batch mode and individually.</p>
      <p>Data exporter that exports wordlists and/or feature values from TransDict to external files and
applications.</p>
      <p>Copy-entry module that copies all, or individual fields of one entry into another</p>
      <p>Morphological generator that automatically generates wordforms for a given word and fills the
morphological fields of the entry automatically assigning the tags specified in the configuration
settings.</p>
      <p>Content and format checker, which reveals incomplete and/or ill formatted entries.</p>
      <p>Look-up tool that performs a wild card search on one or any combination of specified parameters
(letters, language, semantic (conceptual) classes and part-of-speech; it is also possible to filter the
whole sets of TransDict entries according to a specified lists of lexemes, incompletely filled entries,
entries of repeated tokens, etc. The use of the Look-up tool allows identification of knowledge gaps
and gives a lot of opportunities for analyzing the qualitative and quantitative linguistic characteristics
of the domains, which are either language specific, or hold across languages, and can be used to
develop metrics for resolving tag ambiguity (unavoidable in annotations) or for contrastive linguistic
research.</p>
      <p>To provide for a collaborative setup for sharing knowledge acquisition tasks, TransDict is
programmed in two versions: the MASTER version with the full range of built-in tools activated and
the LIGHT version, - an empty TransDict program shell configured as MASTER but with the
Configuration block disabled for consistency of the acquired knowledge. Acquirers can individually
fill LIGHTs with new lexical-ontological knowledge, which is then imported into MASTER on a
regular basis.</p>
      <p>The platform tagger gets a "raw" text as input and outputs its annotated version at a specified level,
- with either conceptual tags only or supertags. The main blocks of the tagger program are as follows:
Configuration block configures the tagger to a specific language and markup level.
Lexicon look-up module tags text with TransDict (super) tags of a selected level
Data importer imports texts from external files and from TransDict knowledge.</p>
      <p>Data exporter has two functions: it exports the annotated text to external files and interactively
exports lexical units tagged as “unknown” to the TransDict knowledge.</p>
      <p>Control interfaces for visualizing tagger output to control the annotation quality.</p>
      <p>Disambiguation rules interpreter integrates the rule-based NLP techniques into the annotation
process; the interpreter has a user-friendly interface for writing tag disambiguation rules in the simple
IF-THEN-ELSE-ENDIF formalism that does not require programmer’s skills. The rules are written
over the lexical knowledge and TransDict tagset and, when saved, are automatically compiled into the
program. The tagger disambiguation interpreter is fully functional and with a good rule coverage
insures the high quality of annotation. Of course, though the interpreter has a lot of effort saving
functionalities, the inherent problem of all rule-based NLP techniques (knowledge bottle-neck) cannot
be avoided. The interpreter module is therefore made optional and its use depends on the user's
willingness to invest a sufficient amount of effort in the disambiguation rule acquisition.</p>
    </sec>
    <sec id="sec-7">
      <title>4. Case study: the “Terrorism” domain annotation platform</title>
    </sec>
    <sec id="sec-8">
      <title>4.1. Knowledge handling</title>
      <p>To be applied in practice, the annotation platform program shell should be filled with domain
knowledge along the lines given in Section 2.1. We further describe this process as done in the frame
of the real on-going project on content analysis of the “Terrorism” domain e-news in the English,
Russian and French languages. The major project task requires the conceptual level of annotation as a
must prerequisite.</p>
      <p>
        The main parts of the platform knowledge base, - the “Terrorism” domain multilingual ontology
and unilingual English, Russian and French onto-lexicons were built in parallel on the data of three
comparable corpora of e-news on terrorist acts of 500,000 words each. The knowledge acquisition
details are described in [
        <xref ref-type="bibr" rid="ref37">38</xref>
        ]. We here concentrate on its presentation and handling in the TransDict
program. A fragment of the TransDict main interface is shown in Fig.2.
      </p>
      <p>In Fig.2, the screenshot of the main TransDict interface displays the entry of the highlighted
lexeme. In the center, the pop-up window of its English equivalent entry is shown as called by
clicking on the “English” bookmark. The interface buttons are self-explanatory. All fields are
interactive and can be edited. On the left pane (from left to right), shown are the interactive list of the
Russian onto-lexicon units, corresponding ontology concept codes (SC) and parts-of speech (PoS).
Every entry contains a lexeme linked to one ontological concept. In case a lexeme can be mapped into
different ontological concepts it appears in different TransDict entries (one per each conceptual
meaning). That explains the lexical duplications in the list.</p>
      <p>The content of a lexical entry opens on clicks on the lexeme and the bookmark of the language of
interest. The knowledge put in the highlighted entry appears on the right pain. The concept, language
and part-of-speech with their codes are located on the top of the wright pane, under which the
morphological zone containing a full paradigm of a lexeme wordforms with supertags is shown. The
TransDict supertags and parts-of-speech are the unified sets of the combinations of task-tuned
linguistic features of the Russian, English and French languages; the number of fields in the
morphological zone is different and defined according to the grammas of corresponding languages.
The new knowledge can be exported to TransDict in a batch mode or individually as follows. A click
on the “Add” button over the lexeme list calls the pop-up interactive menu of concepts; the selection
of a concept opens the part-of speech menu (see Fig.3), which, in turn opens a new TransDict entry
with the selected structure and all the knowledge but the morphological paradigm filled out. The
acquirer needs to fill only one wordform in the paradigm field, the rest word-forms will be generated
automatically. The content of every entry zone is editable and can be copied from one entry to
another. All settings are configurable; the setting changes automatically propagate to the lexical
entries.</p>
      <p>The “Terrorism” domain corpora-based lexemes exported to the TransDict unilingual lexicons is
aligned as translation equivalents; the translation gaps are filled out by the acquirers. This augmented
onto-lexicons and made the platform useful for machine translation-related tasks as well. The number
of aligned lexicon entries is thus the same but the number of unique unilingual lexemes differs due to
the different levels of synonymy in each language. The explicit list of lexemes’ paradigms in the
TransDict entries allows skipping many analysis problems and annotating the input text by a simple
look-up in the TransDict morphological zones. The look-up procedure goes from left to right, longer
units first. The results of such look-up can be displayed in the tagger interface on the level of concept
tags only (see Fig.4) or on the level of supertags, if necessary.</p>
    </sec>
    <sec id="sec-9">
      <title>The annotation platform as a research tool</title>
      <p>The developed annotation platform due to its advanced search functions accessible through the
TransDict main interface can also be used as a research tool. We did just that in an attempt to find
quantitative disambiguation metrics that could complement or even substitute the disambiguation
rules. As a first step on this way, we sorted out all lexemes that were linked to multiple ontology
concepts and thus lead to conceptual multi-tags after the TransDict look-up. Analysis of both, the
sorted out lists and the domain corpora showed that multi-tags are caused by two different
phenomena, that of lexical conceptual ambiguity and that of conceptual syncretism. The unilingual
lexemes are conceptually ambiguous, if in the domain corpora, they can function in different mutually
exclusive conceptual meanings, like, for example, the English word “car” and its Russian and French
equivalents “автомобиль” and “voiture”, correspondingly ( annotated with the multi-tag ~P~C) :</p>
      <p>CONSEQUENCES-DAMAGE (P): The terrorist attack damaged about 50 cars. / В результате
атаки террориста повреждено около 50 автомобилей/ L'attaque terroriste a endommagé environ 50
voitures.</p>
      <p>MEANS OF ATTACK (C): A car hit people on Westminster Bridge. / На Вестминстерском
мосту автомобиль наехал на людей/ Une voiture a heurté des gens sur le pont de Westminster.</p>
      <p>The unilingual lexemes are conceptually sincretical, if they have several conceptual meanings that
do not contradict each other. Most often, but not exclusively, conceptual syncretism was detected in
multicomponent domain-relevant lexemes. For example, in the English noun phrase "airport shooting
suspect" the word "shooting" contains information about the type of attack, the word "airport"
indicates the place where the attack occurred, the word "suspect" has two conceptual meanings at
once - "assumption" and " perpetrator of a terrorist act ”. Therefore, after the tagger look-up this
lexeme will be conceptually annotated as {airport shooting suspect} ~T~L~I~A.</p>
      <p>In the multi-tag syncretism case no ambiguity resolution is required as the meanings of the
individual conceptual tags in a multi-tag are complimentary. On the contrary, multi-tags that are
caused by conceptual ambiguity need to be disambiguated. We tried to answer the question whether it
is possible to automatically identify syncretical multi-tags to exclude them from the disambiguation
procedure.</p>
      <p>To reduce the volume of annotator tasks, we conducted the research on relatively small portions of
the unilingual e-news corpora of 35,000 wordforms each, which were automatically annotated by the
tagger TransDict look-up and manually post-edited to the gold standard. We then calculated the
frequencies of the multi-tags, which “survived” the postediting and thus were sincretical per
definition. The threshold for cutting the top frequency list of the syncretical multi-tags to be
excluded from the disambiguation procedure can be defined empirically. We currently experimented
with the 10 top sincretical multi-tags in every language. We further introduced a heuristic concept
usage relevancy (CUR) measure. The heuristics is: the higher the concept CUR value, the more
prioritized its tag can be in the set of the other tags assigned to the same lexical unit. The empirical
formula we use at the current stage of research to calculate the CUR value is:</p>
      <p>CUR = (RCF*w1+Cf*w2) / (w1+w2), were</p>
      <p>RCF is the ratio of concept fillers; it accounts for the variety of the lexical units mapped into a
concept and is calculated as</p>
      <p>RCF = n/N, where
Cf = (Cfs +Cfm ) / F, where
n is the number of unique (different) unilingual corpus lexical units mapped into a particular
concept in the corpus and N is the total number of ontology-mapped lexemes in the corpus;
Cf is the concept frequency index calculated as</p>
      <p>Cfs is the frequency of the concept in the corpus as a single tag, Cfm is the frequency of the
concept in the corpus as a component of a multi-tag; F is the total number of conceptual tags (single
and multiple) in the corpus; w1 and w2 are arbitrary weights; we so far experimented with w1= 10
and w2=1. The denominator (w1+w2) in the CUR formula is used to normalize the CUR value to the
common percentage scale.</p>
      <p>The suggested disambiguation measures are supposed to be crosslinguistically universal, while
their values are obviously language-dependent. The scope of this paper does not permit to give the
details of the calculations (it takes a forthcoming paper), we here therefore present the preliminary
results of using the CUR values in the annotation workflow, which we defined to be performed in the
following order:</p>
      <sec id="sec-9-1">
        <title>Automatic text annotation with the tagger TransDict look-up, Automatic exclusion of the top 10 of always syncretical multi-tags from disambiguation, Automatic disambiguation of the rest of the multi-tags based on concept usage relevancy (CUR) values,</title>
        <p>Manual postediting of the resulting annotations.</p>
        <p>
          In assessing the conceptual annotation accuracy we used the temporal post-editing effort
quantitative measure [
          <xref ref-type="bibr" rid="ref38">39</xref>
          ]. Participants in the evaluation experiment were the project members who
acquired the platform knowledge and students of the South Ural State University (Russia) enrolled in
a translation studies program and familiar with the computational linguistics concepts and post-editing
techniques. They were given same-size portions of raw and automatically annotated texts (stage 3
output of the annotation workflow) and were asked to report on the time they spent on producing the
gold annotations of the raw and automatically annotated texts. To make the evaluation less subjective,
the raw and automatically annotated texts given to each participant were taken from different corpora.
The reported time values were then summarized and normalized. The results showed that the
postediting time spent on the automatically annotated texts was on average 35% less than the time needed
to conceptually annotate the raw text, which shows our annotation framework to be viable.
        </p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>5. Conclusions</title>
      <p>We have presented an interoperable platform for multi-grain annotation of multilingual domain
corpora. The platform is a stand-alone PC application realized for Windows in the C++ programming
language. The interoperability of the platform is provided by the tagset that includes conceptual
information specified in the language-independent domain ontology and a universal tagging
algorithm. The latter is defined to consist of two main successive procedures: ontological analysis
(text-to-ontology mapping) and multi-tag disambiguation, for which both the rule-based NLP
technique and/or quantitative measures can be applied. The paper covers the platform general design
and its application for the conceptual annotation of the "Terrorism" domain corpora in English
Russian and French. The potential of the developed interoperable platform as a research tool to define
quantitative metrics for tag disambiguation is also demonstrated on the example of the
conceptuallevel annotation. The suggested quantitative metrics account for a) the frequency of the concept
usage in unilingual corpora annotations and b) the variety of the unilingual lexical units mapped into a
multilingual ontological concept. The specificity of the approach is that a) the unit of the ontological
analysis is taken to be a multicomponent phrase rather than a single word and b) tag disambiguation
can supported by the rule-based NLP technology through the fully functional platform tagger
interpreter and/or by quantitative measures. The case study assessment of the conceptual tagging
effort with the suggested an-notation workflow steps and quantitative tag disambiguation measures
(without rule-based disambiguation) showed on average the 35% gain in tagging time, which proves
the legitimacy of the proposed interoperable multilingual annotation frame-work. We are fully aware
that more research should be done on disambiguation metrics and see it as our future work. In
parallel, we will proceed with enlarging both the depth and the breadth of the multilingual ontology
and the coverage of the onto-lexicons both in terms of size and the number of languages.</p>
    </sec>
    <sec id="sec-11">
      <title>6. References</title>
      <p>[1] L. Stojanović, N. Stojanovic, J. Ma. On the Conceptual Tagging: An Ontology Pruning Use
Case. WI '07: Proceedings of the IEEE/WIC/ACM International Conference on Web
Intelligence, 2007, pp. 344–350.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Buitelaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramaka</surname>
          </string-name>
          .
          <article-title>Unsupervised ontology-based semantic tagging for knowledge markup</article-title>
          .
          <source>Proceedings of the Workshop on Learning in Web Search at Learning in Web Search at 22nd International Conference on Machine Learning</source>
          , Bonn, Germany,
          <fpage>26</fpage>
          -
          <lpage>32</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Charniak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M</given-names>
            <surname>Berland</surname>
          </string-name>
          .
          <article-title>Finding Parts in Very Large Corpora</article-title>
          .
          <source>In Proceedings of the 37th Annual Meeting of the ACL</source>
          ,
          <year>1999</year>
          , pp.
          <fpage>57</fpage>
          -
          <lpage>64P</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Glover</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Tsioutsiouliklis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lawrence</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pennock</surname>
          </string-name>
          ,
          <string-name>
            <surname>G</surname>
          </string-name>
          , Flake,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Using Web Structure for Classifying and Describing Web Pages</article-title>
          .
          <source>In Proc. of the 11thWWW Conference</source>
          , pp.
          <fpage>562</fpage>
          -
          <lpage>569</lpage>
          , ACM Press,
          <year>2002</year>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Reeve</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hyoil</surname>
          </string-name>
          .
          <article-title>Survey of Semantic Annotation Platforms</article-title>
          .
          <source>In SAC '05</source>
          , pp.
          <fpage>1634</fpage>
          -
          <lpage>1638</lpage>
          , ACM Press, NY, USA,
          <year>2005</year>
          , ISBN 1-58113-964-0
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kiryakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Popov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Terziev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Manov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ognyanoff</surname>
          </string-name>
          . Semantic Annotation, Indexing, and
          <string-name>
            <surname>Retrieval</surname>
          </string-name>
          .
          <source>Elseviers Journal of Web Semantics</source>
          , Vol.
          <volume>2</volume>
          ,
          <issue>2005</issue>
          , No. 1, http://www.ontotext.com/kim/semanticannotation.html.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Kogut</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Holmes</surname>
          </string-name>
          . AeroDAML:
          <article-title>Applying Information Extraction to Generate DAML Annotations from Web Pages</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Gauch</surname>
          </string-name>
          .
          <article-title>Scientific method in practice</article-title>
          . Cambridge : Cambridge University Press. 435 P,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Ceausu</surname>
          </string-name>
          and
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Despr`es. Learning Term to Concept Mapping Through Verbs: A Case Study</article-title>
          .
          <source>Proceedings of the Semantic Authoring, Annotation and Knowledge Markup Workshop (SAAKM2007) located at the 4th International Conference on Knowledge Capture (KCap</source>
          <year>2007</year>
          ), Whistler, British Columbia, Canada,
          <source>October 28-31</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gaizauskas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hepple</surname>
          </string-name>
          , G. Demetriou,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Setzer</surname>
          </string-name>
          .
          <article-title>Building a semantically annotated corpus of clinical texts</article-title>
          .
          <source>Journal of Biomedical Informatics</source>
          . - Vol.
          <volume>42</volume>
          (
          <issue>5</issue>
          ),
          <fpage>950</fpage>
          -
          <lpage>966</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Tenenboim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shapira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shoval</surname>
          </string-name>
          .
          <article-title>Ontology-Based Classification of News in an Electronic Newspaper</article-title>
          .
          <source>International Book Series “Information Science and Computing”</source>
          ,
          <fpage>89</fpage>
          -
          <lpage>97</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>U.</given-names>
            <surname>Inyaem</surname>
          </string-name>
          , Ch. Haruechaiyasak, Ph. Meesad,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tran</surname>
          </string-name>
          .
          <source>Ontology-Based Terrorism Event Extraction Proceedings of the 1st International Conference on Information Science</source>
          and Engineering. P.
          <volume>912</volume>
          -
          <fpage>915</fpage>
          ,
          <year>2009</year>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.V.</given-names>
            <surname>Dobrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Dobrova N. L.</given-names>
            ,
            <surname>Soms</surname>
          </string-name>
          <string-name>
            <given-names>N. L.</given-names>
            ,
            <surname>Chugunov</surname>
          </string-name>
          <string-name>
            <surname>A.V.</surname>
          </string-name>
          <article-title>Semanticheskij analiz novostnyh soobshchenij po teme «Elektronnye uslugi»: opyt primeneniya metodov ontologicheskoj semantiki Trudy XVIII ob"edinennoj konferencii «Internet i sovremennoe obshchestvo» (IMS2015</article-title>
          ).
          <fpage>120</fpage>
          -
          <lpage>125</lpage>
          ,
          <year>2015</year>
          . (in Russian)
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Djemaa</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Candito</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Muller</given-names>
            <surname>Ph</surname>
          </string-name>
          .,
          <string-name>
            <surname>Vieu</surname>
            <given-names>L</given-names>
          </string-name>
          .
          <article-title>Corpus annotation within the French FrameNet: a domain-by-domain methodology</article-title>
          .
          <source>Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC</source>
          <year>2016</year>
          ),
          <source>May</source>
          <year>2016</year>
          , Portorož, Slovenia
          <volume>3794</volume>
          -380.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Beckwith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Fellbaum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. J.</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Introduction to WordNet: An Online Lexical Database</article-title>
          .
          <source>International Journal of Lexicography</source>
          <volume>3</volume>
          (
          <issue>4</issue>
          ),
          <fpage>235</fpage>
          -
          <lpage>244</lpage>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>E.</given-names>
            <surname>Montiel-Ponsoda</surname>
          </string-name>
          , G. Aguado de Cea,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gómez-Pérez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Peters</surname>
          </string-name>
          . Modelling Multilinguality in Ontologies.
          <source>Proceedings of COLING</source>
          <year>2008</year>
          , Companion volume -
          <source>Posters and Demonstrations</source>
          .
          <volume>67</volume>
          -
          <fpage>70</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Espinoza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gómez-Pérez</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Mena E.</surname>
          </string-name>
          <article-title>Enriching an Ontology with Multilingual Information</article-title>
          .
          <source>The Semantic Web: Research and Applications. ESWC Lecture Notes in Computer Science</source>
          . - Springer, Berlin, Heidelberg. - Vol.
          <volume>5021</volume>
          .
          <fpage>333</fpage>
          -
          <lpage>347</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Chaves</surname>
          </string-name>
          , M and
          <string-name>
            <surname>Trojahn</surname>
            <given-names>C.</given-names>
          </string-name>
          (
          <year>2010</year>
          ).
          <article-title>Towards a Multilingual Ontology for Ontology-driven Content Mining in Social Web Sites</article-title>
          ,
          <year>2010</year>
          , - URL: https://goo.
          <source>gl/sZKmS2(19.02</source>
          .
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Alatrish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tošić</surname>
          </string-name>
          ,
          <string-name>
            <surname>N. Milenkov N.</surname>
          </string-name>
          <article-title>Building Ontologies for Different Natural Languages</article-title>
          .
          <source>Building Computer Science and Information Systems</source>
          . - Vol.
          <volume>11</volume>
          (
          <issue>2</issue>
          ).
          <fpage>623</fpage>
          -
          <lpage>64</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Embley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. W.</given-names>
            <surname>Liddle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Lonsdale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tijerino</surname>
          </string-name>
          .
          <article-title>Multilingual Ontologies for CrossLanguage Information Extraction</article-title>
          and
          <string-name>
            <given-names>Semantic</given-names>
            <surname>Search</surname>
          </string-name>
          ,
          <year>2019</year>
          . https://pdfs.semanticscholar.
          <source>org/6884/41a96b6da61295c7df39b70db2f28531370a.pdf (last accessed 21.02</source>
          .
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Nirenburg</surname>
          </string-name>
          , V. Raskin V.
          <article-title>Ontological Semantics</article-title>
          . MIT Press, Cambridge,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>I.</given-names>
            <surname>Niles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pease</surname>
          </string-name>
          .
          <article-title>Linking Lexicons and Ontologies: Mapping WordNet to the Suggested Upper Merged Ontology</article-title>
          .
          <source>Proceedings of the 2003 International Conference on Information and Knowledge Engineering</source>
          ,
          <fpage>412</fpage>
          -
          <lpage>416</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>R.</given-names>
            <surname>Arp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.D.</given-names>
            <surname>Spear</surname>
          </string-name>
          .
          <article-title>Building Ontologies with Basic Formal Ontology</article-title>
          . MIT Press, Cambridge,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>D.</given-names>
            <surname>Elworthy</surname>
          </string-name>
          .
          <article-title>Tagset design and inflected languages</article-title>
          .
          <source>In 7th Conference of the European Chapter of the Association for Computational Linguistics (EACL)</source>
          ,
          <source>From Texts to Tags: Issues in Multilingual Language Analysis SIGDAT Workshop</source>
          , 1-
          <fpage>10</fpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>J.</given-names>
            <surname>Nivre</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Boguslavsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Iomdin</surname>
          </string-name>
          .
          <article-title>Parsing the SynTagRus treebank of Russian</article-title>
          .
          <source>In Proceedings of the 22nd International Conference on Computational Linguistics (Coling</source>
          <year>2008</year>
          ),
          <fpage>641</fpage>
          -
          <lpage>648</lpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gnasa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Woch</surname>
          </string-name>
          .
          <article-title>Architecture of a knowledge based interactive Information Retrieval System</article-title>
          .
          <year>2002</year>
          , http://konvens2002.dfki.de/cd/pdf/12P-gnasa.
          <source>pdf (last accessed 19.02</source>
          .
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>A.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jirka</surname>
          </string-name>
          , Ch. Brew.
          <article-title>A cross-language approach to rapid creation of new morphosyntactically annotated resources</article-title>
          .
          <source>In Procedings. LREC</source>
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>S.</given-names>
            <surname>Petro</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. Das</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
          </string-name>
          .
          <article-title>A universal part-of-speech tagset</article-title>
          .
          <source>Proceedings of the Conference on Language Resources and Evaluation (LREC</source>
          <year>2012</year>
          ). Istanbul, Turkey,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>T.</given-names>
            <surname>Erjavec</surname>
          </string-name>
          .
          <article-title>Multext-east version 4: Multilingual morphosyntactic specifications, lexicons and corpora</article-title>
          .
          <source>In Proceedings of the Seventh conference on International Language Resources and Evaluation (LREC'10)</source>
          , Valletta, Malta,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>G.</given-names>
            <surname>Orosz</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Nov´ak,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Pr´osz´eky. Lessons Learned from Tagging Clinical Hungarian</article-title>
          .
          <source>IJCLA</source>
          , vol.
          <volume>5</volume>
          , no.
          <issue>1</issue>
          ,
          <fpage>129</fpage>
          -
          <lpage>145</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>H. F.</given-names>
            <surname>Hsieh. Three Approaches to Qualitative Content Analysis / H.-F. Hsieh</surname>
          </string-name>
          , S.E. Shannon // Qualitative Health Research. Vol.
          <volume>15</volume>
          (
          <issue>9</issue>
          ). - P.
          <fpage>1277</fpage>
          -
          <lpage>1288</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Carvalho</surname>
            <given-names>V.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almeida</surname>
            <given-names>J.P.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fonseca</surname>
            <given-names>C.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guizzardia</surname>
            <given-names>G.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Multi-level ontology-based conceptual modeling</article-title>
          .
          <source>Data &amp; Knowledge Engineering</source>
          .Volume
          <volume>109</volume>
          ,
          <fpage>3</fpage>
          -
          <lpage>24</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Ménard</surname>
            <given-names>P.A</given-names>
          </string-name>
          ,, Barrière C.
          <article-title>PACTE : a collaborative platform for textual annotation</article-title>
          -URL: https://www.aclweb.org/anthology/W17-7410.pdf (last accessed
          <year>2021</year>
          /03/08).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Stenetorp</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pyysalo</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Topic</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ohta</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ananiadou</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jun'ichi Tsujii J. BRAT</surname>
          </string-name>
          <article-title>: a Web-based Tool for NLP-Assisted Text Annotation</article-title>
          .
          <source>In: Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics</source>
          , pp.
          <fpage>102</fpage>
          -
          <lpage>107</lpage>
          . Avignon, France,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Zagorul'ko M. Yu</surname>
          </string-name>
          .,
          <string-name>
            <surname>Kononenko</surname>
            <given-names>I. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sidorova</surname>
            <given-names>E. A.</given-names>
          </string-name>
          <article-title>Sistema semanticheskoj razmetki korpusa tekstov v ogranichennoj predmetnoj oblasti [System for Semantic Annotation of DomainSpecific Text Corpora]. Proceeding of the international conference Komp'yuternaya lingvistika i intellektual'nye tekhnologii</article-title>
          , Bekasovo, May 30 - June 3,
          <year>2012</year>
          . Moscow, RSUH, vol.
          <volume>11</volume>
          (
          <issue>18</issue>
          ),
          <fpage>674</fpage>
          -
          <lpage>683</lpage>
          .
          <year>2012</year>
          . (in Russian)
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>P.</given-names>
            <surname>Stenetorp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pyysalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Topic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ohta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ananiadou</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Tsujii.</surname>
          </string-name>
          <article-title>BRAT: a Web-based Tool for NLP-Assisted Text Annotation</article-title>
          .
          <source>Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics, April</source>
          <volume>23</volume>
          -27, Avignon, France,
          <fpage>102</fpage>
          -
          <lpage>107</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sheremetyeva</surname>
          </string-name>
          .
          <source>Universal Computational Formalisms and Developer Environment for RuleBased NLP. Lecture Notes in Computer Science</source>
          , vol
          <volume>10761</volume>
          . Springer, Cham,
          <year>2018</year>
          . https://link.springer.com/chapter/10.1007/978-3-
          <fpage>319</fpage>
          -77113-7_5DOI https://doi.org/10.1007/978- 3-
          <fpage>319</fpage>
          -77113-
          <issue>7</issue>
          _
          <fpage>5</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sheremetyeva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zinovyeva</surname>
          </string-name>
          .
          <source>On Modelling Domain Ontology Knowledge for Processing Multilingual Texts of Terroristic Content. Communications in Computer and Information Science</source>
          ,
          <volume>859</volume>
          . Springer, Cham, pp.
          <fpage>368</fpage>
          -
          <lpage>379</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>A.</given-names>
            <surname>Zaretskaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Corpas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Seghiri</surname>
          </string-name>
          .
          <article-title>Measuring Post-editing Time and Effort for Different Types of Machine Translation Errors</article-title>
          . New Voices in Translation Studies.
          <volume>15</volume>
          ,pp.
          <fpage>63</fpage>
          -
          <lpage>91</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>