<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Domain Ontology Learning using Link Grammar Parser and WordNet</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dmytro Dosyn</string-name>
          <email>dmytro.h.dosyn@lpnu.ua</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yousef Ibrahim Daradkeh</string-name>
          <email>daradkehy868@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vira Kovalevych</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mykhailo Luchkevych</string-name>
          <email>luchkevychmm@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yaroslav Kis</string-name>
          <email>yaroslav.p.kis@lpnu.ua</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Engineering and Networks, College of Engineering at Wadi Addawasir 11991, Prince Sattam Bin Abdulaziz University</institution>
          ,
          <addr-line>KSA</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of the theory of wave processes and optical systems of diagnostics, Karpenko Physico-Mechanical Institute of the National Academy of Sciences of Ukraine</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Information Systems and Network Department, Institute of Computer Science and Information Technologies, Lviv Polytechnic National University (LPNU)</institution>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2009</year>
      </pub-date>
      <volume>105</volume>
      <fpage>110</fpage>
      <lpage>113</lpage>
      <abstract>
        <p>The problem of knowledge discovery that comes down to information pertinence evaluation still stay unsolved because of absence of practical effective methods and means of ontology learning. Despite active development of natural language processing tools, they mostly reach the level of semantics, named entity recognition and sentiment analysis but not pragmatics. On the other hand, ontology is the only instrument, “measuring ruler” to compare and estimate the usefulness of information for some particular user, which knowledge could be represented by such ontology as his hierarchical task network (HTN). The need to build separate HTN ontology for each user puts on the agenda the task of design of the automated ontology learning from text. With aim to solve this task the system of automated and semiautomated ontology learning from text had been developed using Carnegie Mellon Link Grammar Parser and WordNet API. Two approaches to distinguish semantic relations in natural language text (NLT) sentences were adopted: analysis of the sentence constituent trees - for explicit relations recognition and Naïve Bayes supervised learning - for recognition implicit semantic relations which need not only verb phrase but other parts of sentence due to its ambiguity. Developed approach was implemented in the Java desktop application using OWL API and Protégé-OWL API. Experimental results were compared to expert analysis and had shown good recognition reliability.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Ontology Learning</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Naïve Bayes Classifier</kwd>
        <kwd>Protégé-OWL API</kwd>
        <kwd>Knowledge Discovery</kwd>
        <kwd>Pertinence Estimation</kwd>
        <kwd>Link Grammar Parser</kwd>
        <kwd>WordNet</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        Traditionally, computer linguistics (CL) presents a text document as a "bag of words", i.e., the
model of a document in terms of CL is the distribution of terms according to the frequency of their
occurrence in the text of this document. Content as a logical essence is not discussed in this approach,
instead only lexical analysis is performed, so it is difficult to distinguish between important and
unimportant documents, look for borrowings from other documents, make short annotations, translate
their content into other languages. Historically, the vast majority of information retrieval systems
have been built on this approach, as the simplest and most obvious. The corresponding information
model of the document is called a vector-spatial model. In this interpretation, the document is
searched in the space of vectors of frequency of occurrence of terms and their combinations. Such
search methods are widely described in special and popular literature (see, for example, [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]), however,
their main drawback is the already mentioned approach, in which the model of a text document is
exclusively lexical and stochastic. This is due to the fact that such models are easy to implement, and
therefore this approach to solving the practical problem of information retrieval is reminiscent of a
situation where a drunkard, returning from a party, looking for lost keys under a lantern only because
it is better visible. But a simple solution is not always correct. And it is intuitively clear that the metric
requires choosing an object commensurate with its dimension as a measure of some phenomenon.
      </p>
      <p>Thus, the model of a text document, as a message of one intellectual agent to another has a certain
functional structure, caused by its purpose to convey certain important from the point of view of the
author information to its recipient. That is why the message usually consists of two parts – a
concluding introductory, in which the author indicates to the addressee known to the latter, in his
opinion, the context of basic information, and the main, constructive, in which the author transmits to
the recipient the information that in a given context will be potentially demanded by the addressee
(See Fig. 1). This means that the recipient will supplement the knowledge base with this new
information.</p>
      <p>The task of the intelligent agent linguistic analysis subsystem is to recognize the true context of the
message contained in the text document, provided that it is correctly specified. To simplify, we must
assume that this is so. What is the process of context recognition? Obviously, in the localization in the
agent's knowledge base of entities and connections, as well as possibly registered in the agent's
knowledge base, the facts referred to in this document, ie, to which there is an explicit or implicit
reference, but which must be recognized.</p>
      <p>To do this, it is necessary to provide the possibility of eliminating ambiguity in the identification
of the concepts mentioned in the text. In particular, this is achieved by comparing the attributes of
such a concept used in the text with the semantic connections of the corresponding entity that are
valid in the agent's ontology. For this purpose it is necessary to perform syntactic-semantic analysis of
sentences of this text with their transformation into predicates of descriptive logic, the names of which
will be semantic connections between concepts represented in the text mainly by verbs (verb phrases)
and adjectives (verb adjectives), and attributes – the concepts themselves, represented by nouns,
pronouns or noun groups.</p>
      <p>The intention of the author of the message to pass it to the correspondent gives grounds to believe
that the author has provided the necessary set of features in the first part of the message to
unambiguously identify the context according to the above scheme. After determining the context, the
next part of the message can be used to identify in the text of unknown (missing in the knowledge
base of the agent) facts and enter them into the knowledge base.</p>
    </sec>
    <sec id="sec-2">
      <title>2. THE TASK STATEMENT</title>
      <p>The need to build a separate HTN ontology for each user puts on the agenda the task of design of
the methods and means of automated ontology learning from text. Those tools should be able to
separate the part of input information that can be relevant to the specific user domain, distinguish
right semantics of the concepts to be placed at the right position in the ontology’s class hierarchy and
reconstruct appropriate specific semantic relations between concepts of the learned ontology from a
complex verb phase.</p>
      <p>
        Context information should be extracted consequently from the learned ontology in a part relevant
to the analyzed text fragment and compared to the discovered context of the whole analyzed
document and/or its information source. The most relevant context should be chosen [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. RELATED WORKS</title>
      <p>As of 2020, the processing of text documents written in natural language (NLP) is based on
statistical models of the use of words and phrases in the text as a whole or the terminological structure
of each sentence regardless of the rest of the text, i.e. without reference to the context of the entire
message. Thus both the systems of rules developed by means of machine learning of statistical
models, and neural network technologies can be used.</p>
      <p>
        The existing ontology population approaches and already available software had been investigated.
In particular, similar studies were conducted at Carnegie Mellon University with financial support
from DARPA, Google, NSF, and CNPq under the NELL project (Never-Ending Language Learning,
[
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]). According to the developers, "If successful, this will result in a knowledge base (i.e., a
relational database) … that mirrors the content of the Web” (see:
http://rtw.ml.cmu.edu/rtw/overview). Initially, the NELL ontology contained several hundred
categories of concepts, 280 types of semantic relations between them, and 10 to 15 sample sentences
for each of them. The system processed about 500 million Internet pages by selecting possible copies
of the categories and connections known to it by several hundred recognition methods available to the
system and identified 390 thousand concepts and 350 thousand types of semantic links. Stuart Russell
commented the project in his book “Human Compatible: Artificial Intelligence and the Problem of
Control”: “Unfortunately NELL has confidence in only 3 percent of its beliefs…” and as a
conclusion: “Reading requires knowledge and knowledge (largely) comes from reading” [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Another similar project, TextRunner, was previously implemented at the University of
Washington. During its implementation, about 9 million web pages, 133 million individual sentences
were processed, of which 7.8 million tuples were selected [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Ontology learning consists in knowledge discovery which differs significantly from data mining
and even from knowledge base population. In the case of data mining statistical analysis is used to
discover some trends and irregularities in big sets of uniform represented data. The knowledge base
population updates the knowledge base by new facts according to already existing formal model of
the domain without changing it. Instead ontology learning besides updating separate facts by new data
updates and rebuilds whole model.</p>
      <p>We define knowledge as useful information. Usefulness anticipates existing an intelligent agent
which is able to make decisions taking into account some information which could be (or not be)
useful. An agent strategy has almost the same parameter – expected utility. Such coincidence is not
accidental – in both cases we are talking about agent’s informed decision making. Information that
increases agent’s strategy expected utility is useful therefore he uses it as knowledge “how to increase
his performance”. Information usefulness as an agent knowledge called pertinence. Knowledge
metrics problem still stay open because of absence of adequate measure instrument which could be
ontology itself, but only in the case when it will contain an optimal strategy of the ontology owner.
Formal description in an ontology of optimal strategy with states, rewards, actions, goal and evaluated
expected utility creates an agent's knowledge what to do to reach max utility (reward) in observed
conditions.</p>
      <p>How to build an agent optimal strategy by means of RDF-OWL-SWRL-SPARQL-PDDL ontology
– still stays an open problem, but some prospective approaches had been already developed [7…9].</p>
      <p>Each separate word and any unordered set of words can’t represent the same statement as a logical
sentence with right syntax, similarly no one set of not ordered sentences (and correspondent DL
predicates) can’t be distinguished as a valuable new knowledge without asserting them into existing
DL model. The key problem of discovering new knowledge in a text document written in natural
language is the development of a method of automatic learning of ontology as a formal-logical model
of the world, in which new information found in the text can be transformed into knowledge.</p>
      <p>
        According to information theory by Claude Shannon [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] any knowledge as a kind of information
could be measured in terms of changing entropy of the system. But it could be shown that not all
equal entropy changing gives to an intelligent agent equal increasing his performance or equal
opportunity to reach closer to his goal. Exactly the change of expected utility of the agent strategy
after updating his knowledge base by this new information could be used as its usefulness i.e.
pertinence. Therefore to find if this information is useful for this agent, other words, to estimate the
pertinence of the information we should include it into ontology strategy of the agent, rebuild an
optimal strategy, taking into account new information, evaluate new expected utility, compare it to
previous one and take the difference as a value of pertinence of a new information, i.e. obtained
knowledge value. For this purposes an adaptive self-learned strategy planning ontology of the
intelligent agent, based on PEAS architecture is needed together with ontology learning technics
similar to described here.
      </p>
      <p>In general, the solution of almost any automatic word processing problem includes the following
levels of analysis:</p>
      <p>• graphematic analysis – to be concise if simplified – the division of the text into individual words
(tokens) and sentences;
• phonetic analysis, applicable in the processing of acoustic data;
• morphological analysis – selection of the grammatical basis of the word, definition of parts of
speech, reduction of words to the dictionary form;</p>
      <p>• parsing – identifying syntactic connections between words in a sentence, building the syntactic
structure of the sentence;
• semantic analysis – identification of semantic connections between words and syntactic groups;
• pragmatic analysis – identification of statements – facts and rules, assessment of the importance
of the identified statements for the overall software model.</p>
      <p>Each such analysis is set and implemented as a separate independent task, which is used in
conjunction with others to solve specific applied and theoretical problems.</p>
    </sec>
    <sec id="sec-4">
      <title>A. TECHNOLOGIES</title>
      <p>From the point of view of practical realization, the most developed and, accordingly, the most
widespread are the following methods and corresponding NLP technologies:</p>
      <p>
        1. Named entity recognition (NER) is one of the most popular methods of semantic analysis.
According to it, the algorithm takes a phrase or paragraph as input and identifies all existing nouns or
names by classifying them. The task of fully recognizing a named entity can be divided into two
different problems: identifying names and classifying names according to the type of entity they refer
to (e.g., person, organization, location, and others). Hierarchies of the named types of entities are
developed. For example, BBN categories [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] consists of 29 types and 64 subtypes, Sekin's extended
hierarchy consists of 200 subtypes [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Ritter used a hierarchy based on conventional Freebase entity
types, applying NER to texts from social media [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Approaches with the use of social networks data
are considered promising [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>2. Tokenization or lexical parsing is the process of converting a sequence of symbols into a
sequence of tokens (groups of symbols corresponding to certain patterns), and determining their
types. The process of displaying texts as a set of characters in a set of ribbons, a set of ribbons in a set
of words are the main steps of any PMT processing task. Various methods and libraries are available
for tokenization, in particular, such as NLTK, Gensim, Keras.</p>
      <p>
        A tokenizer based on Penn TreeBank [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] allows to recognize punctuation, auxiliary words (words
that occur only in combination with other words) and words with a hyphen. Moses’ tokenizer
developed as part of the project of the automatic translator of the same name
(http://www.statmt.org/moses/) based on statistical training on parallel text corpora.
      </p>
      <p>spaCy (https://spacy.io/api/tokenizer) – another tokenizer which provides flexibility in defining
special markers.</p>
      <p>
        Word particle tokenization is also used. This tokenization is very useful for a specific application,
where it is important to take into account frequently used particles [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
      <p>3. Stemming and lemmatization. In linguistic morphology, stemming is the process of removing
affixes in order to obtain the basis of a word, which will serve as its most general formulation. The
base found does not have to be identical to its root. Since 1968, in the field of computer science,
appropriate stemming algorithms have appeared. Many search engines use stemming to highlight the
base in order to supplement the queries with synonyms which is called a conflation.</p>
      <p>
        Lematization is aimed at solving the same problem with the only difference that special
dictionaries (for example, WordNet [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]) and/or corpora of texts are used to find the basis in order to
find the correct vocabulary word form – a lemma, as some canonical form of the studied token.
      </p>
      <p>
        4. "Bag of words"/TF-IDF/N-grams model – NLP procedures, aimed at further its lexical and
statistical analysis as a vector-space model [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for using machine learning methods. This procedures
preferably includes mentioned stages of tokenization, stemming, and lemmatization. This approach
ignoring context but is simple therefore widely popular.
      </p>
      <p>
        5. Sentiment analysis has been actively developing since the 2000s. Most commercially developed
NLP systems include a subsystem to analyze their tonality (see, for example, [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]). In such an
analysis, it is assumed that the textual information contains both facts and judgments of the author, the
tone of which should be determined. Knowledge-based techniques, statistical methods, and hybrid
approaches are used to distinguish the polarity of the studied text messages [20].
      </p>
      <p>6. Immediate constituent analysis (analysis of successive layers – constituents) is a method of
sentence analysis, which was first formulated by Leonard Bloomfield, and developed by Rollon Wells
and Noam Chomsky [21, 22]. The method uses successive division of a sentence into subparts –
“immediate constituents” up to last word to obtain a layered sentence structure of recursive
inclusions, which should represent some its internal logic to be further analyzed. Now this practice is
widely used.</p>
      <p>7. Dependency analysis parses syntactic structure of a sentence to build a network of
subordinations dependencies between all separate words of the sentence. Used dependency grammar
lets to reveal the basic elements of the sentence, which fulfils the roles of subject, object and a
semantic relation between them as skeletal predicate structure of the sentence. Authors [23] used to
compare constituent and dependency analysis by parsing the same example for both methods (see
Figure 2).
(a)
(b)
Figure 2: Immediate constituent analysis (a) and dependency analysis (b) of the sentence “I saw
a fox” (From https://www.baeldung.com/cs/constituency-vs-dependency-parsing)
The result of dependency analysis can be represented by an oriented graph with a single root node
corresponding to the predicate, from which the edges named according to the type of connection are
directed to the dependent members of the sentence, and each word in the sentence, except that which
serves as a root node, must be at least one edge oriented to that word. Although syntactic
dependencies cannot be directly transformed into predictive logic predicates, they, in combination
with constituent analysis data that allow the identification of multi-word predicate attributes, can
serve as a sufficient basis for the application of machine learning methods to classify-recognize
known semantic connections and corresponding predicate constructions.</p>
    </sec>
    <sec id="sec-5">
      <title>B. TOOLS</title>
      <p>At the end of the second decade of the 21st century, it is the methods of automating the processing
of text written in natural language, or rather their imperfection, became a bottleneck in the
development of information systems towards their intellectualization: over-scale volumes of available
textual information remain unavailable for full analysis. The developed tools use vector-space model
methods such as "bag of words" or recognition of named entities and do not go beyond the analysis of
the tone of the text, automatic abstracting or translation. A number of companies and research groups
are actively developing and improving existing systems and technologies:
1. NLTK (Natural Language Toolkit) – open source library of entry level NLP tools, developed
with Python. It includes tools such as:
• tokenization;
• stemming;
• grammar markup;
• text classification;
• NER;
• sentiment analysis.</p>
      <p>
        In addition to the NLP processing tools themselves, NLTK includes more than 50 lexical databases
and text corpora, such as RST Treebank [25], Penn Treebank [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], WordNet 17], Lin thesaurus [26]
etc.
      </p>
      <p>Drawbacks include low performance and lack of a syntactic parser, which must be compensated by
additional tools [27], such as, for example, the statistical parser Bohnet and Nivre [28].</p>
      <p>2. Stanford Core NLP – multi-purpose “full-stack” tool for text analysis, including NER, text
tonality (sentiment analysis), construction of conversational interface. Like NLTK, Stanford CoreNLP
offers a wide range of NL text treatment tools. The main advantage of Stanford NLP tools is
scalability. Unlike NLTK, Stanford Core NLP is better suited for processing large amounts of data
and performing complex operations. Thanks to the architecture of the scalability system Stanford
CoreNLP effectively handles the processing of streaming data, such as extracting information from
open sources (social media, breaking news, running reviews generated by users), tone analysis (social
media, customer support), conversational interfaces chatbots), natural language generation (customer
support, e-commerce). This tool effectively recognizes named entities and provides fast syntactic
markup of terms and phrases.</p>
      <p>3. Apache OpenNLP – a multifunctional text analyzer similar to Stanford Core NLP, which
supports the main common tasks of natural language text processing, such as tokenization, sentence
segmentation, tagging of parts of speech, recognition of named entities, fragmentation, parsing,
language recognition and interpretation of current references (searches expressions that refer to the
same entity in the text). This analyzer, according to its documentation, is based on the use of machine
learning methods, such as classification by maximum entropy and perceptron modeling. Apache
OpenNLP is a library that consists of many components. The user can make from these components a
NLP conveyor according to his tasks.</p>
      <p>A comparative analysis of the Apache OpenNLP on the Internet with Stanford Core NLP shows a
significant functional similarity, which does not allow making an unambiguous decisive choice
between them: both systems perform their functions equally well.</p>
      <p>4. SpaCy – open source software library for NLP in Python. The library includes the following
components:
• tokenization
• POS tagging
• dependency parsing
• lemmatization
• SBD (sentence boundary detection)
• NER (named entity recognition)
• EL (entity linking)
• similarity
• text classification
• rule-based matching – search for patterns by given features and regular expressions.</p>
      <p>In general, this is a standard set of basic-level functions for the construction of specialized PMT
processing systems designed to solve specific problems in particular software. The main feature of the
library, which distinguishes it from the previous two, is the implementation in Python. Easy
integration of deep learning software is declared. It also includes proprietary parsing and NER
visualization tools.</p>
      <p>5. GATE – open-source software infrastructure and graphical IDE for reliable creation and
deployment of components and resources for NLP [29]. Due to the modular architecture, it can be
used element by element or complex. It performs an almost complete set of the above operations with
text. Initially it was made in Java but currently also supports Python. The low entry threshold for
system users is provided by the existing graphical shell, which ensures its widespread use not only by
experts in the field of NLP, but also by those who try to solve practical problems in other areas related
to streaming and processing text streams, such as automatic dialog systems.</p>
      <p>The advantages of GATE developers include the following properties of the system:
• automatic measurement of the performance of components;
• distinguishing between low-level tasks, such as data storage, data visualization, component
detection and loading, and high-level NLP tasks;
• clear separation between data structures and NLP algorithms;
• use of standard linguistic data exchange mechanisms for components and open standards such as
Unicode and XML;</p>
      <p>• correction of exotic data formats – GATE performs automatic format conversion and provides
alignment and standardization of NLP data;</p>
      <p>• providing a basic set of NLP components that can be expanded and/or replaced by users as
needed [30].</p>
      <p>An important positive feature of the GATE project is the presence of APIs for modeling ontologies
and manipulating them.</p>
      <p>6. Link Grammar Parser – a publicly available library of NLP software, originally built on ANSI
C as a formal grammar system Link Grammar, which in the process of development became part of
the Java distribution RelEx (https://wiki.opencog.org/w/RelEx_ Dependency_Relationship_Extractor) [31].
The system was developed in the 1990s by three linguists – David Temperley, John Lafferty and
Daniel Sleator [32]. This system includes all the elements of text analysis – from the initial
(graphematic) to the level that can be called the primary semantics of the English language.</p>
      <p>The system decomposes English sentences into a set of binary meta-semantic relations between
words or punctuation, thus forming some syntactic structure (similar to dependency parsing). To do
this, the authors have built a special dictionary, in which for each word of the language indicates
which meta-semantic relations it can be associated with other words in the sentences of this language.</p>
      <p>Link Grammar Parser (LGP) distinguishes over 100 different types of meta-semantic relations
between words in English sentences, providing the user with the ability to recognize detailed context.</p>
      <p>The advantages of the system include the ability to process the entire array of text in parallel,
provided by dictionary technology to identify meta-semantic relations between word pairs in
sentences with subsequent weight filtering by probability, which is fully consistent with the Big Data
view. At same time due to the exponential complexity of used algorithm it operation is limited in
time.</p>
    </sec>
    <sec id="sec-6">
      <title>4. METHODS</title>
      <p>Among all the above-mentioned NLP tools and related technologies, it is necessary to choose the
most appropriate for the task of knowledge discovery according to the criteria of reliability,
robustness, speed and ease of implementation.</p>
      <p>From the point of view of algorithmic language, the choice was made in favor of Java, mainly
given that most software libraries for work in the field of semantic networks (in particular,
ProtégéOWL API, OWL API, Apache Jena) were developed in this language. From the point of view of
specific distributions and relevant software libraries, it is quite difficult to choose one (see Table 2)
and, based on previous development experience, it is necessary to avoid such a choice as long as
possible, trying to choose the most standardized solutions and combine necessary elements of various
means (API) by including the required software libraries in the overall project. At the same time it is
necessary to pay special attention to possible conflicts between versions and implementations of the
involved libraries.</p>
      <p>Among other dependency parsers, RelEx attempts to obtain a greater degree of semantic
normalization: for questions, comparative data, entities, and sentence and subordinate clauses, while
other parsers (such as Stanford Core NLP) follow a literal representation of the text's syntactic
structure. In particular, RelEx pays special attention to determining when a sentence is hypothetical or
speculative, and isolating query variables from the question. Both of these aspects aim to make RelEx
suitable for question-answering and semantic understanding/reasoning systems. In addition, RelEx
performs word markup to denote a part of speech, a set of nouns, their gender, the temporal form of
verbs, and so on. According to the developers, RelEx analyzes text almost four times faster than
Stanford Core NLP, and it provides a "compatibility mode" in which it can parse sentences in the
same format as Stanford University's parser.</p>
      <p>Nevertheless, RelEx project seems to be already stopped, moreover, none of the mentioned
systems contains the means of analysis at the level of pragmatics of text documents. The vast majority
of them are about equally successful in analyzing the tone of the text. Heuristics based on statistical
criteria using machine learning methods are mainly used. And here there is nothing unusual because
text message pragmatics (its assessment in terms of usefulness-neutrality-harmfulness) cannot be
determined without the use of knowledge base formulated as an optimal strategy of the intellectual
agent, as there is no starting point – the person with his goals and/or damages to be assessed.</p>
      <p>Thus, it was decided to combine different libraries in the project as much as possible, preferring
the most standardized ones and those that do not impose unjustified restrictions on possible further
selection and development. As a core of the software package stays the OWL API library which
should help to reach the main aim – populate OWL ontology as a knowledge keeper and
simultaneously tool of text pragmatics evaluation.</p>
    </sec>
    <sec id="sec-7">
      <title>A. USED TECHNOLOGIES</title>
      <p>Despite the need to combine different NLP technologies due to the lack of full-fledged means of
extracting from the text description logic predicates necessary for ontology learning and/or knowledge
base population, one of them had to be taken as basic and supplemented by missing algorithms and
technologies. For the reasons stated above, the choice was made on the approach used in the Link
Grammar Parser. The main advantage of the method is simplicity and usability. Parsing mechanism is
separated from dictionaries describing language. Due to the use of dictionaries and the system of rules
of syntactic dependencies between words of the sentence, which can be dynamically edited and
supplemented in the learning process, this parser easily adapts to a given problem area, organically
combines with the ontology, actually becoming part of it. On the other hand, as mentioned earlier in
the parser's overview, the ability to recognize more than 100 types of meta-semantic relations between
words allows for detailed semantic analysis of the text.</p>
      <p>Link Grammar Parser focuses on parsing syntactic dependencies between sentence words.
However, there are a number of significant differences: the edges of the formed relations graph are
not oriented. The parser also analyzes the dependencies between phrases, thus combining the
advantages of both approaches – constituent analysis and dependency parsing. As a result, the
recognized relations between words and phrases are not only syntactically but also contextually
dependent. Therefore, the developers have used their own system of notation for such dependencies,
which would, in their opinion, best suit their role in shaping the semantics of a sentence formed from
related words and phrases. In this regard, it is appropriate to call the syntactic dependencies classified
by the developers of LGP as meta-semantic relations (MSR).</p>
      <p>As combined identifiers of such meta-semantic relations, special symbolic sequences are used –
connectors that correspond to a specific type of connection. An either ‘+’ or ‘-’ sign is attached to
each connector on the right to indicate the direction of relation. For example, if the word W1 is
assigned the connector A+ and the word W2 is assigned the connector A-, then in the syntactic
structure of a sentence consisting of two words W1 and W2, the connection A will be established
between the words W1 and W2. In this case, the sentence “W2 W1” will not receive any
interpretation, because W2 is assigned MSR A-, which forms a connection only to the left, and the
word W1 is assigned A+, which forms a connection only to the right.</p>
      <p>LinkParser performs the analysis in two stages: the first stage – construction of set of syntactic
representations of one sentence in English. The parser considers all variants of meta-semantic
relations between words, choosing among them those which satisfy criteria of projectivity
(connections should not intersect) and connectivity (the resulting graph should contain as few
components as possible).</p>
      <p>For example, for the following dictionary:
the: D+
white: A+
horse: D- &amp; A- &amp; S+
was: S- &amp; PP+
tired:
PPthe phrase “The white horse was tired” will be parsed as follows:
+------D-----+
| +---A--+-S--+-PP-+
| | | | |</p>
      <p>The white horse was tired</p>
      <p>Here, the algorithm distinguishes four pairs of words connected by the corresponding MSR: D (the
- horse), А (white - horse)), S (horse - was), PP (was - tired).</p>
      <p>In the second stage of analysis, the program divides the constructed graph into so-called domains,
each of which contains links belonging to the corresponding fragment of the sentence (atomic
statement). For example, in the sentence You think that I am happy, two domains "You think" and
"that I am happy" will be selected. Then for each domain conditions of type are checked: if in the
domain there is ‘A’ relation there should (or should not) be a relation ‘B’. Thus, it is possible to check
presence/absence in the domain of one relation depending on presence/absence of other relation.</p>
      <p>A few MSR can be attributed to one word. Their combination may be subject to certain conditions,
reflected in the dictionary by notation similar to that used in regular expressions, namely:
&amp; – "asymmetric conjunction": "W: A+ &amp; B+" means some word X, with which the word W
forms a connection A, must be earlier in the text than the word Y, with which the word W forms a
connection language B;</p>
      <p>or – "disjunction": "W: A+ or B-" means the word W can form either a connection A to the right
or a connection B to the left;</p>
      <p>{} – "optional": "W: A+ &amp; {B+}" means after the word W formed the right connection A, it may
or may not form the connection B;
@ – "unlimited" means that the connection can be formed an unlimited number of times.</p>
      <p>Formulas can be assigned to a single word or a whole class of words. For example, the connector
(A) that connects adjectives with nouns is assigned to all adjectives at once. Similarly, the formulas
are assigned to 23 categories of English words, distinguished by morphological (degrees of
comparison of adjectives and adverbs, grammatical number (singular-plural) nouns, etc.) and basic
syntactic features (transitivity, bi-transitivity, etc.).</p>
      <p>Traditional relation extraction approach uses extraction templates to distinguish predefined set of
relations, training documents and testing documents as well [33]. Here it is proposed to build
extraction templates using patterns, automatically constructed from set of most used word-MSR-word
pairs. LGP parser gives us all needed preliminary information to build such patterns.</p>
      <p>The choice of a particular method took into account the features of the problem facing the system
as a whole, the initial stage of learning, as well as the architecture of the system, namely the presence
of an ontology as both a means and objectives of such learning:</p>
      <p>1) terminological TBox facts should contain statistical templates of semantic connections, by
which the system will recognize a potential binary predicate;</p>
      <p>2) search-recognition-learning should be purposeful, in the direction of solving some specific
problem (set of problems) given by the ontology of the subject area;</p>
      <p>3) the applied algorithms should have the minimum complexity because of need of their repeated
in a long cycle of check of expected usefulness of the received fact.</p>
      <p>The processing of NLT in this way should consist in the sequential transformation of sentences of
the text into predicates of descriptive logic in terms of a pre-formed (learned in the same way)
ontology.</p>
    </sec>
    <sec id="sec-8">
      <title>B. ALGORITHM</title>
      <p>The general flowchart of data transformations for extracting semantic relations from text is
presented on the Figure 3, corresponding block-diagram of the algorithm is presented on the Figure 4.</p>
      <p>Software implementation of the method of ontology learning involves recognizing semantic
relations in NLT sentences, building on their basis DL predicates, identifying their logical
interdependence in order to form appropriate rules and assert the identified facts and rules into the
ontology.</p>
      <p>To solve this task it is necessary to take into account the following statements:
1. A prerequisite for identifying a semantic relation is the presence of a verb phrase. Its presence is
determined by means of direct constituent analysis, and the type – corresponding to the verb phrase
meta-semantic relation. All conjunctions to the right of the conjunction 'S *' indicate a verb phrase.</p>
      <p>2. The words that accompany (which are pointed to by) these symbolic sequences of
metasemantic relations determine the type of semantic relation, i.e, the name of the predicate encoded in
the sentence. These words are usually verbs: "knows", "has", "belongs", "refers", or verb phrases
"belongs to", "consists of", etc.</p>
      <p>3. Adjectives are interpreted as properties and can also be recognized in sentences through the
semantic relation "has a property".</p>
      <p>4. To recognize the semantic relation, the following steps should be performed:
- parse sentences using LGP;
- find a verb phrase through constituent parsing taking into account the connection symbols to the
right of "Ss" MSR;
- find the verbs indicated by these symbols;
- find the subject of the action (subject in the sentence), which indicates the symbol "Ss";
- find the object of action (obviously, the definition in the sentence), i.e., the subject to which the
action is directed;</p>
      <p>- check in the ontology the presence of this type of semantic connection and, in its absence, create
it;</p>
      <p>- check the presence in the ontology of entities that mean the object and subject of action. There are different
options (see Table 3 below).</p>
      <p>In case (5), obviously, the sentence is ignored. In case (4) also nothing to connect: so, for example,
if the known ontology connection ‘IS-A’ is used, which connects two unknown ontologies terms: X
and Y, such statement will also have to be ignored. In cases (2) and (3) it is possible to introduce the
concept into the ontology through a known semantic relation.</p>
      <p>5. A distinction should be made between ontology as a set of admissible semantic relations
between concepts and the knowledge base, as a set of facts and rules about a given model of reality.</p>
      <p>6. Relations can be unconditional and conditional. Conditional relations are written as rules –
implication of predicates. Unconditional connections are a partial case of conditional ones and are
recorded as facts in the form of separate predicates.</p>
      <p>7. It is necessary that the recognition of semantic relations is based on the data of the learned
ontology which, among other things, would contain the concept of semantic relation and its
featuresproperties.</p>
      <p>Based on the above arguments, a semi-controlled machine learning method using the naive Bayes
classifier was chosen. To do this, we used the LGP analysis of the NLT sentence, which resulted in a
graph of meta-semantic relationships between the words.</p>
      <p>In the first step, the ontology recognizes the words of this sentence as concepts of a given domain.
Next, the results of parsing identify the roles of recognized words in the sentence. The next step is to
build a feature vector of each of the recognized concepts in relation to its role in this sentence. This
vector of features recognizes the type of semantic relations in a sentence. The ontology statistics is
accordingly updated which is one of ontology learning stages. Since the basis of the semantic relation
in a sentence is a verb phrase, the analysis begins with it.</p>
      <p>When analyzing another sentence, the system recognizes or does not recognize the verb phrase
selected by the direct constituent analysis of this sentence. If it does not recognize, it enters it into the
ontology as a new type of relation and creates its pattern (see Figure 5) with the ability to calculate
statistics. If it recognizes, it adds an existing pattern to the statistics. If in the ontology there are
several semantic relations with the given verb, the system by means of procedure of recognition on
statistics of patterns decides to what type of relation this sentence (the given verb phase) concerns.</p>
      <p>The classical relation was used to recognize the type of semantic relation by the naive Bayes
classifier:</p>
      <p>P(hi | X j )=</p>
      <p>P(X j | hi )P(hi )</p>
      <p>P(X j )
(1)
where hi – i-th relation hypothesis; Xj – feature vector (pattern) of j-th semantic relation; P(hi ) –
unconditional probability (frequency of occurrence) of the i-th type of semantic relation; P(X j ) –
probability of appearance of the j-th feature vector (frequency of occurrence of the j-th pattern);
P(hi | X j ) – probability that the i-th type of semantic relation takes place in the presence of the j-th
feature vector; P(X j | hi ) – the probability of the appearance of the j-th feature vector, provided that
there is the i-th type of semantic relation.</p>
      <p>The naive Bayes classifier is used, for which a feature vector is considered statistically
independent, and therefore the probability of co-occurrence of features is considered equal to the
product of the probabilities of independent appearance of each of the features:</p>
      <p>N
P(X j | hi )∝ ∏ P(xk , j | hi ). (2)</p>
      <p>k=1
Then the probability of belonging to the class hi, provided that the set of features Xj is close to:</p>
      <p>N
P(hi | X j )∝ ∏ P(xk ,j | hi )P(hi ), (3)</p>
      <p>k=1
that is, we are looking for a hypothesis about the present semantic relation h*:</p>
      <p>N
h* = arg max ∏ P(xk ,j | hi )P(hi ), (4)</p>
      <p>i k=1
where N – the number of features of the pattern.</p>
      <p>This method was implemented as a Java software using the Protégé-OWL API and Link Parser
API as part of the overall CROCUS project. After preprocessing the text document is divided into
separate sentences and for each of the sentences it is sequentially parsed by LGP, the results of which
are passed to the procedure of ontology updating. According to the specification, all statistics of the
features of semantic relations are stored in the appropriate domain ontology, which provides the
ability to self-learning and appropriate adaptability to user needs. The scheme of storage of features is
given in Figure 6. Adaptability is provided by weighing of each stored features according to its
occurrence during ontology learning. For this purposes every element of a pattern descriptor has its
own counter: Wml – for MSR name; Sml – for the concept, which participates in a descriptor as a
subject; Oml – for the concept, which participates as an object (see Figure 6). All of them are stored in
ontology as an integer type datatype property (see Table 4). This way we can estimate all needed
conditional probabilities using only data from the learned ontology.</p>
      <p>A recursive procedure of adding of new concepts to the ontology during its learning was
elaborated and applied with help of WordNet API. Each time new word, not represented in the
learned ontology by correspondent concept, appears in a learning text the procedure retrieves all (but
no more than N most frequent) correspondent synsets from the WordNet database, then retrieves there
his hypernym and checks its presence in the ontology. If not, procedure recursively repeats until all
kind of the hypernyms of the original word will find their hypernym concept in the learned ontology.
For set of obtained concepts the total importance weight WT of each of them are compared to select
the highest one, for which all previously unknown hyponyms up to original word are sequentially
added to the ontology as correspondent subclasses.</p>
      <p>It is possible thanks to using the method of weighting of ontology concepts and relations,
developed by authors [34]. According to this method:</p>
      <p>W ji = Woij +Wsij +Wnij
2. At the moment of adding a new subclass to the i+1st level, it is assigned its own weight Woij+1
equal to half of the class's own weight of the upper (i-th) level. The weight of the class Wcij and all
parent classes up to the root increases by the weight of the newly created subclass:
1. The total weight</p>
      <p>of the ontology class is equal to the sum of its own weight Woij , the
weight of the subclasses Wsij and the weight of the adjacent non taxonomic classes Wnij (associated by
a semantic relation, other than "is-a"):
where: Wsij = ∑Wcki+1L j,k – the weight of k subclasses of the j-th class of the i-th level of the taxonomy;
k</p>
      <p>Wcki+1 = Woki+1 +Wski+1 – the weight of the class Wcki+1 ; Lj,k – the weight of the relation between
classes C j and Ck .
(5)
(6)</p>
      <p>Wc mj = Wc mj +Woij+1 ,∀m ≤ i .</p>
      <p>3. When establishing a relation Lk1 ,k2 between the concepts k1 and k2, an edge appears between the
corresponding vertices of the ontology graph, and the weight Wc2 is added to the weight of the
adjacent classes Wn1, and vice versa, the weight of the new adjacent class Wc1 is added to Wn2:
Wn j = ∑Wck L j,k (7)</p>
      <p>k
4. The weight of an instance in the knowledge base is equal to the total weight (5) of its class in
the ontology.</p>
      <p>5. The weight of a relation importance is equal to value of its mentions counter (see Figure 7).</p>
      <p>The developed method and correspondent algorithm is the basis for automatic recalculation of the
weight of ontology classes, instances and relations between them during its filling and adaptation to a
given subject area during operation with the aim to optimize structure and volume of automatically
generated ontology [35].</p>
      <p>Ontology learning tools and methods, described above, had been implemented as the software
CROCUS (Cognition of Relations Over Concepts Using Semantics).</p>
      <p>It is based on Protégé-OWL API and OWL API with functions, realized by LGP API, WordNet
API and some other Java libraries. It allows combining work with both OWL1 and OWL2 ontology
formats, depending of the task.</p>
      <p>The first recognition level – which concept corresponds to a NLT word – is built with help of
WordNet: all possible meaning of the word is represented by a set of synsets, which title name and
attributes are compared with the names and corresponding attributes of concepts in the ontology [36].
If it is absent CROCUS proposes two options: 1) to choose from the hypernyms, proposed by the
WordNet (see: Figure 8), or 2) to let the software choose one automatically based on criteria of most
general concept from the subset of known of the set of proposed by the WordNet.</p>
      <p>It is possible thanks to automated ontology concept weighting during learning process. A concept
is as high at the ontology taxonomy as more general it is. A level of concept place in the taxonomy of
the ontology is unambiguously determined by its Wo weight. Therefore Wo could be interpret as a
comparative level of concept generality.</p>
      <p>Often a hypernym concept of the looked NLT word is also absent at the ontology. In that case the
looking procedure is repeated recursively for each founded hypernym of each synset WordNet
representation of the word. As a result of such a search using the WordNet taxonomy, a search tree
with a high vertex index is formed (see, for example, the number of hypernyms, suggested by
WordNet for the word “thinking” in Figure 8), which leads to the need to calculate the semantic
similarity coefficients for the number of concepts equal to the number of terminal vertices of the
whole obtained search tree. For reasons of conciseness of the ontology the system chooses the branch
of the search tree with the highest Wo terminal node parameter. All the non-terminal concept nodes of
the search tree branch also then added to the ontology as subclasses in a backward sequence up to the
word from the NLT.</p>
      <p>All the learned parameters of the ontology, which provide it adaptability to a certain domain
and/or user needs, are saved inside this ontology as RDF datatype properties. The place of storage and
assignment of parameters is summarized in the Table 4.</p>
      <p>Carrier address*</p>
      <p>Aim</p>
      <sec id="sec-8-1">
        <title>Own weight</title>
      </sec>
      <sec id="sec-8-2">
        <title>Subclasses weight</title>
      </sec>
      <sec id="sec-8-3">
        <title>Lisa</title>
        <p>Wml
Oml</p>
      </sec>
      <sec id="sec-8-4">
        <title>Individual</title>
      </sec>
      <sec id="sec-8-5">
        <title>Individual</title>
        <p>"default"+&lt;MSR&gt;
"default"+&lt;sem.relation&gt;+&lt;MSR&gt;+”object”
+&lt;concept&gt;+&lt;POS&gt;
Sml</p>
      </sec>
      <sec id="sec-8-6">
        <title>Individual "default"+&lt;sem.relation&gt;+&lt;MSR&gt;+”subject” +&lt;concept&gt;+&lt;POS&gt; Mx</title>
      </sec>
      <sec id="sec-8-7">
        <title>Individual "default"+&lt;sem.relation&gt; * – all address elements are separated by underline sign</title>
      </sec>
      <sec id="sec-8-8">
        <title>Is-a relation weight</title>
      </sec>
      <sec id="sec-8-9">
        <title>MSR counter</title>
      </sec>
      <sec id="sec-8-10">
        <title>A counter of using a</title>
        <p>concept as an object in a</p>
      </sec>
      <sec id="sec-8-11">
        <title>MSR for a certain</title>
        <p>semantic relation</p>
      </sec>
      <sec id="sec-8-12">
        <title>A counter of using a</title>
        <p>concept as an subject in
a MSR for a certain
semantic relation</p>
      </sec>
      <sec id="sec-8-13">
        <title>Semantic relation counter</title>
        <p>The learning procedure is implemented in LearnSemRelation.java class, which contains the
methods:
- findSuperClass();
- addSubClass();
- setDataProperty();
- addPattern();
- readPattern(),
and a set of 15 other auxiliary methods like classExist(), getInstanceMaxNumber(),
relationStatisticsView() etc.</p>
        <p>The key method addPattern() takes a set of descriptors, produced by the
SupplementedLGParser.java class, and performs a sequence of steps:
1) needed standard concepts is adding to the ontology;
2) “default_”-kind auxiliary individuals is adding;
3) learning sentences representative individuals is generated;
4) new semantic relation ontology class is created;
5) correspondent default-individual is created with the datatype RDF property Mx=1;
6) an abstract MSR concept is created or updated;
For each obtained descriptor (see Figure 6):
- consequently MSR subject and object concepts are adding to an ontology recursively (as
described above) and/or, if any of them already exists, default auxiliary individuals, mentioned in
Table 4, are created;
- Wml, Oml and Sml datatype properties of the correspondent individuals are updating;
- importance weights Wo, Ws, Wn are recalculating with accordance with (5)…(7) to choose right
superclass during next learning steps.</p>
        <p>A set of all needed functions are implemented in the CROCUS system. The main user interface of
the system is presented on the Figure 9. The most commonly used functions of the system are moved
to the toolbox of the front panel:
- Process text is a complex procedure of ontology learning by parsing text using Link Grammar
parser as a set of sentences with the same semantic relation which should be sequentially
processed recursively updating the ontology with help of WordNet lexical database;
- Parse sentence is a similar to text processing procedure except it is manual and under one
separate sentence;
- Verify ontology opens separate dialog interface with possibility of observation of all main
values and parameters of the learned adaptive ontology: concepts, relations, weights, properties
values etc.;
- Estimate is a procedure of forced up-down recalculation of concept weights;
- Analyze – is a mode of automated recognition of semantic relation using Bayesian classifier,
described above.</p>
        <p>According to the main aim of the software CROCUS it assists in automated and semi
automated ontology learning by natural language text on the level of semantic relations,
terminological TBox axioms and facts/predicates. A learned adaptive ontology processed by such
learning procedures will be ready to distinguish semantic meaning of the NLT sentences without
an expert participation.</p>
        <p>Current state of a learned ontology can be monitored in a separate window, shown at the
Figure 10. There are hierarchical structure of the ontology, weights of the learned concepts Wo
and Ws, evaluated by the method, mentioned above on section 4, main ontology statistics etc.</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>6. EXPERIMEMTAL RESULTS</title>
      <p>An almost empty ontology had been used to test the process of ontology learning from text using
the Link Grammar parser and WordNet API. An initial metrics, provided by the Protégé 3.4. is
presented at the Figure 11.</p>
      <p>Learning texts had been selected from a corpus of scientific paper abstracts in the common
Material Science and Computer Engineering domain. Random first 20 sentences which using each
learned semantic relation had been adopted and applied to learn an ontology to recognize the 3 kind of
relations:
• consist-of;
• is-a;
• shows.</p>
      <p>The selection of semantic relation kind was random with taking into account the statistics of their
using in the domain abstracts: in the middle of 1/3 top popular relations (see: Figure 12).
1600 1456
1400
1200
1000
800
600
400
200
0
381
362
358317302
289
235230223204171160136127110103 99 88</p>
      <p>During a learning process the number of superclasses, which should be added to the ontology
simultaneously (previously) with the new concept from the learned text gradually decreases, like it
could be seen on the Figure 14. On the Figure 15 it is presented the data of learned descriptors on the
example of “is-a” semantic relation pattern, structure of which was shown at the Figure 6. The
correspondent numerical data for the learned three kind of relations after learning by 20 sentences for
each is provided in the Table 5.
25
20
15
10
5
0</p>
      <p>Testing classification experiments had been executed with aim to examine the developed software
and the algorithm, described above in previous section, based on the naïve Bayes classifier and the set
of learned semantic relation pattern descriptors, mentioned in the Table 5. The results are presented as
a print-screen copy of the results of the working CROCUS software (see Figure 16).</p>
    </sec>
    <sec id="sec-10">
      <title>7. CONCLUSION</title>
      <p>In this study, the system of automated and semi-automated ontology learning form text had been
developed using Carnegie Mellon Link Grammar Parser and WordNet API. A combination of two
different methods to distinguish semantic relations in NLT sentences had been applied: analysis of the
sentence constituent trees – by means of Link Grammar Parser for explicit meta-semantic relations
recognition and Naïve Bayes supervised learning – for direct recognition semantic relations with help
of additional meta-semantic data, obtained on the previous step. Developed approach was
implemented in the Java desktop application using WordNet API, OWL API and Protégé-OWL API.
Experimental results were compared to expert analysis and had shown good recognition reliability.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raghavan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Schultze</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          (
          <year>2008</year>
          ). Introduction to Information Retrieval. Cambridge University Press, Cambridge, UK.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Ivokhin</surname>
            ,
            <given-names>E.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oletsky</surname>
            ,
            <given-names>O.V.</given-names>
          </string-name>
          <article-title>Restructuring of the Model “State-Probability of Choice” Based on Products of Stochastic Rectangular Matrices</article-title>
          .
          <source>Cybern Syst Anal</source>
          . Vol.
          <volume>58</volume>
          (
          <issue>2</issue>
          ), pp.
          <fpage>242</fpage>
          -
          <lpage>250</lpage>
          (
          <year>2022</year>
          ). https://doi.org/10.1007/s10559-022-00456-z
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>[3] Toward an Architecture for Never-Ending Language Learning</article-title>
          . A.
          <string-name>
            <surname>Carlson</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Betteridge</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Kisiel</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>E.R. Hruschka</given-names>
          </string-name>
          <string-name>
            <surname>Jr</surname>
            . and
            <given-names>T.M.</given-names>
          </string-name>
          <string-name>
            <surname>Mitchell</surname>
          </string-name>
          .
          <source>In Proceedings of the Conference on Artificial Intelligence (AAAI)</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4] Mitchell,
          <string-name>
            <given-names>T.</given-names>
            &amp;
            <surname>Kisiel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            &amp;
            <surname>Krishnamurthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            &amp;
            <surname>Lao</surname>
          </string-name>
          , Ni &amp; Rivard, Kathryn &amp; Mohamed,
          <string-name>
            <given-names>T.</given-names>
            &amp;
            <surname>Nakashole</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            &amp;
            <surname>Platanios</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Emmanouil</given-names>
            <surname>Antonios</surname>
          </string-name>
          &amp;
          <string-name>
            <surname>Ritter</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Samadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Settles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Cohen</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Wang</surname>
            , Richard &amp; Wijaya,
            <given-names>D.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Gupta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Chen</surname>
            , Xinlei &amp; Saparov,
            <given-names>A.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Greaves</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Gardner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          . (
          <year>2018</year>
          ).
          <article-title>Never-ending learning</article-title>
          .
          <source>Communications of the ACM</source>
          .
          <volume>61</volume>
          .
          <fpage>103</fpage>
          -
          <lpage>115</lpage>
          .
          <fpage>10</fpage>
          .1145/3191513.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Russell</surname>
          </string-name>
          , S.
          <source>Human Compatible: Artificial Intelligence and the Problem of Control</source>
          .- Penguin Publishing Group. -
          <volume>352</volume>
          p.
          <article-title>- 2019</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Yates</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Banko</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Broadhead</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cafarella</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etzioni</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Soderland</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <article-title>TextRunner: Open Information Extraction on the Web</article-title>
          .
          <article-title>Proceedings of Human Language Technologies: The Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-HLT)</article-title>
          ,
          <year>Pp</year>
          .
          <fpage>25</fpage>
          -
          <lpage>26</lpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Behnke</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bercher</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biundo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glimm</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponomariov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schiller</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Integrating ontologies and planning for cognitive systems</article-title>
          .
          <source>In: Proceedings of the 28th International Workshop on Description Logic (DL</source>
          <year>2015</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>France-Mensah</surname>
          </string-name>
          , Jojo, and
          <string-name>
            <surname>William J. O'Brien.</surname>
          </string-name>
          “
          <article-title>A Shared Ontology for Integrated Highway Planning</article-title>
          .”
          <source>Advanced Engineering Informatics</source>
          <volume>41</volume>
          (
          <year>2019</year>
          )
          <article-title>: n. pag. Advanced Engineering Informatics</article-title>
          . Web.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Lucas</given-names>
            <surname>Martínez</surname>
          </string-name>
          , N.;
          <string-name>
            <surname>Martínez-Ortega</surname>
          </string-name>
          , J.-F.;
          <string-name>
            <surname>Castillejo</surname>
            ,
            <given-names>P.; Beltrán</given-names>
          </string-name>
          <string-name>
            <surname>Martínez</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <article-title>Survey of Mission Planning and Management Architectures for Underwater Cooperative Robotics Operations</article-title>
          .
          <source>Appl. Sci</source>
          .
          <year>2020</year>
          ,
          <volume>10</volume>
          ,
          <fpage>1086</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Shannon</surname>
            ,
            <given-names>C. E.</given-names>
          </string-name>
          (
          <year>2001</year>
          ).
          <article-title>A mathematical theory of communication</article-title>
          .
          <source>ACM SIGMOBILE mobile computing and communications review</source>
          ,
          <volume>5</volume>
          (
          <issue>1</issue>
          ),
          <fpage>3</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Brunstein</surname>
          </string-name>
          ,
          <article-title>Ada "Annotation Guidelines for Answer Types"</article-title>
          .
          <source>LDC Catalog. Linguistic Data Consortium</source>
          ,
          <year>2002</year>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Sekine</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <source>Kiyoshi Sudo and Chikashi Nobata. “Extended Named Entity Hierarchy.” LREC</source>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Ritter</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Mausam; Etzioni.,
          <string-name>
            <surname>O.</surname>
          </string-name>
          (
          <year>2011</year>
          ).
          <article-title>Named Entity Recognition in Tweets: An Experimental Study (PDF)</article-title>
          .
          <source>Proc. Empirical Methods in Natural Language Processing</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Derczynski</surname>
            , Leon and
            <given-names>Diana</given-names>
          </string-name>
          <string-name>
            <surname>Maynard</surname>
          </string-name>
          , Giuseppe Rizzo, Marieke van Erp,
          <string-name>
            <surname>Genevieve Gorrell</surname>
          </string-name>
          , Raphael Troncy, Johann Petrak, and
          <string-name>
            <surname>Kalian Botcheva</surname>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>«Analysis of named entity recognition and linking for tweets»</article-title>
          .
          <source>Information Processing and Management</source>
          <volume>51</volume>
          (
          <issue>2</issue>
          ): Pp.
          <fpage>32</fpage>
          -
          <lpage>49</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Taylor</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marcus</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Santorini</surname>
            <given-names>B.</given-names>
          </string-name>
          (
          <year>2003</year>
          )
          <article-title>The Penn Treebank: An Overview</article-title>
          . In: Abeillé A. (eds) Treebanks. Text,
          <source>Speech and Language Technology</source>
          , vol
          <volume>20</volume>
          . Springer, Dordrecht. https://doi.org/10.1007/
          <fpage>978</fpage>
          -94-010-0201-
          <issue>1</issue>
          _
          <fpage>1</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Sennrich</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haddow</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Birch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <source>“Neural Machine Translation of Rare Words with Subword Units”, &lt;i&gt;arXiv e-prints&lt;/i&gt;</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Miller</surname>
            <given-names>G.A.</given-names>
          </string-name>
          <string-name>
            <surname>Wordnet</surname>
          </string-name>
          :
          <article-title>A lexical database for English / G.A</article-title>
          . Miller // Communications of the ACM Vol.
          <volume>38</volume>
          , No.
          <volume>11</volume>
          . -
          <fpage>1995</fpage>
          . - Pp.
          <fpage>39</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Salton</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>C. S..</given-names>
          </string-name>
          <year>1975</year>
          .
          <article-title>A vector space model for automatic indexing</article-title>
          .
          <source>Commun. ACM</source>
          <volume>18</volume>
          ,
          <issue>11</issue>
          (Nov.
          <year>1975</year>
          ),
          <fpage>613</fpage>
          -
          <lpage>620</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Anh-Dung</surname>
            <given-names>Vo</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quang-Phuoc Nguyen</surname>
          </string-name>
          , and
          <string-name>
            <surname>Cheol-Young Ock</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Automatic Knowledge Extraction for Aspect-based Sentiment Analysis of Customer Reviews</article-title>
          .
          <source>In Proceedings of the</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>