<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using Non-Negative Matrix Factorization for Text Segmentation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Aliya Nugumanova</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Madina Mansurova</string-name>
          <email>mansurova01@mail.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yerzhan Baiburin</string-name>
          <email>ebaiburin@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yermek Alimzhanov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Al-Farabi Kazakh National University</institution>
          ,
          <addr-line>Almaty</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>D. Serikbayev East Kazakhstan State Technical University</institution>
          ,
          <addr-line>Oskemen</addr-line>
          ,
          <country country="KZ">Kazakhstan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>233</fpage>
      <lpage>242</lpage>
      <abstract>
        <p>The aim of this paper is to investigate whether non-negative matrix factorization (NMF) can be useful for semantic segmentation of large full-text documents. NMF is a universal technique that decomposes the monolithic structure of a massive dataset into diferent trends. In case of textual data these trends can be interpreted as topics. Thereby NMF can associate each document with topics covered in it, however, without linking topics to the certain parts of that document. In this paper, we complement this traditional NMF technique with a new goal: for a given full-text document we build a semantic map which links document's parts with topics covered in it.</p>
      </abstract>
      <kwd-group>
        <kwd>non-negative matrix factorization</kwd>
        <kwd>text segmentation</kwd>
        <kwd>topic modeling</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Text segmentation is a very interesting challenge in the field of natural language
processing. It arises in many information retrieval applications providing users
with quick access to document repositories. Since full-text documents stored in
such repositories are usually large to read and analyze, information retrieval
applications should be able to divide them into chunks and deliver the most
relevant chunks to users in accordance to their requests [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        In this paper, we focus on the task of segmentation of full-text documents
from a topic modeling perspective. In recent years, topic modeling is gaining
momentum in data mining in general [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], and in particular in the text segmentation
ifeld [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Recent work in this field has shown that using topic distribution over
documents instead of term distribution can significantly increase segmentation
performance [
        <xref ref-type="bibr" rid="ref4 ref5 ref6 ref7 ref8 ref9">4,5,6,7,8,9</xref>
        ].
      </p>
      <p>
        The segmentation mechanism drawn from topic modeling is very simple. At
ifrst, each document is divided on small segments (e.g., sentences or paragraphs).
At second, topics covered in this document are revealed, and each word in each
segment is associated with one topic; thereby, for each segment topic occurrences
are defined. At last, adjacent document’s segments sharing a certain number of
common topics are merged into topical chunks.
One of the most popular approaches to topic modeling is non-negative matrix
factorization (NMF). In general, NMF is a well-recognized technique due its
ability to extract relevant structures of data and may thus contribute to a deeper
understanding of data behavior [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. This technique, being applied to a collection
of full-text documents, maps it into a space of topics. For example, in Fig. 1,
NMF is applied to a co-occurrence matrix of a collection, which consists of 5
documents. As we can see from the figure, after matrix factorization, document
1 is represented as a combination of topic 1 and topic 3. Simultaneously, topic 3
is represented as a combination of term 2, term 4 and term 5, and term 5 is the
most significant for this topic.
      </p>
      <p>As we can see, NMF has two useful applications. Firstly, for each document it
defines the most weighted topics, which we call relevant topics. Secondly, for each
topic it finds the most weighted terms, which we call support terms. In this paper,
we use support terms for the semantic segmentation of full-text documents.
Our contribution is to complement the traditional NMF representation with
a new goal: the creation of a semantic map of the given document through
using support terms as map’s nodes. We suppose that, by linking these nodes to
the corresponding parts of the document, we achieve its smart and comfortable
segmentation.</p>
      <p>The rest of this paper is organized as follows. In Section 2, we discuss previous
work on text segmentation and explain our reasons to use NMF. In Section
3, we present proposed approach. In particular, we demonstrate how we link
support terms with the document parts, and present some experimental results.
In Section 4, we formulate conclusions and plans for our future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        The most simple and intuitive algorithm of text segmentation is TextTiling [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
It uses a sliding window to move through a document and capture text blocks
(tiles). The similarity between consecutive blocks are calculated on the base of
cosine metrics. The calculated values are used to draw a similarities curve that
tracks topics changes between consecutive blocks so that the segment boundaries
are chosen at the local minima of the curve. The main disadvantage of TextTiling
is low accuracy because of the sparsity of text blocks.
      </p>
      <p>
        Another simple algorithm of text segmentation is C99 [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. At first, it divides
the input document into minimal blocks (sentences) and for each block calculates
its rank based on the blocks similarities. Then it performs divisive clustering
starting with the whole document and splitting it to parts in accordance with
blocks’ ranks. In [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] C99 algorithm is improved by applying Latent Semantic
Analysis for calculating the blocks’ similarity matrix.
      </p>
      <p>
        C99 algorithm also is used in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This work addresses the issue of providing
topic driven access to full-text documents. Authors of this work apply C99 to
subdivide documents into smaller thematically homogeneous parts that can be
used as link targets. They try to perform segmentation as accurate as possible:
document parts should be of such sizes that "shrinking them would cause relevant
information to be left, and expanding them would bring in too much non-relevant
information" [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, they concentrate only on the segmentation phase
without details of designing a whole navigation system.
      </p>
      <p>
        As we have mentioned in the introduction, a considerable line of research
explores text segmentation methods based on topic modeling. The most popular
algorithms for topic modeling are Latent Dirichlet Allocation (LDA) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8,9,10</xref>
        ]
and Non-negative Matrix Factorization (NMF) [
        <xref ref-type="bibr" rid="ref10 ref11">10,11</xref>
        ]. Although output of LDA
is very similar to the output of NMF, these models are fundamentally diferent
in nature: LDA is based on a Bayesian probabilistic model; whereas NMF is
based on algorithms of linear algebra that fit root mean squared error. As it’s
shown in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], both LDA and NMF can discover concise and coherent topics and
demonstrate similar performance, however NMF learns more incoherent topics
than LDA. Authors of [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] also compare LDA and NMF, and conclude that NMF
better than LDA "from the perspectives of consistency from multiple runs and
early empirical convergence".
      </p>
      <p>
        We choose NMF here because of its basis sparseness [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Basis sparseness
means that NMF uses less basis features (terms) than LDA. This makes NMF
topics more overlapped, i.e. more semantically related to each other than LDA
ones (see an example represented by Table 1). We consider that these relations
are essential for the understanding of how the document’s semantic map should
be organized.
      </p>
      <p>Fig. 2 gives a visual interpretation of Table 1. The four topics are shown in
Fig.2, and some of them are linked with other topics through specific key terms.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Proposed Approach</title>
      <p>Our method of text segmentation consists of 5 steps. Firstly, we should subdivide
a given full-text document into units and build units-by-terms co-occurrence
matrix. Secondly, we should define a reasonable number of topics (K) and apply
NMF to factorize the co-occurrence matrix and obtain 2 matrices: units-by-topics
and topics-by-terms. Thirdly, for each extracted topic we should sort topic terms
# in RussiBaansis tinerEmnglish
1 горный mining
2 порода rock
3 поверхность surface
4 склон slope
5 процесс process
6 динамика dynamics
7 система system
8 зона zone
9 состав composition
10 кора crust
by their weights and select only the most weighted terms (no more than 10% of
all terms). We call these terms support terms.</p>
      <p>
        In our case we have used "Geology" textbook written in Russian [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and
divided it into 89 units (by number of chapters). Then we have chosen K=5
and extracted 5 topics and 500 support terms. By the way we have proposed
information about term distribution over these 5 topics to a geology expert for
analysis. Based on the information the expert has concluded that it is best to
name extracted topics as "Endogenous", "Solar system", "Oceans", "Exogenous"
and "Geochronology". Tables 2-6 represent top 10 support terms for each of 5
extracted topics.
      </p>
      <p>Through NMF Through LDA
# Russian English Russian English
1 вода water вода water
2 процесс process процесс process
3 порода rock динамика dynamics
4 экзогенный exogenous внешний external
5 внешний external экзогенный exogenous
6 динамика dynamics материал material
7 материал material склон slope
8 склон slope поверхность surface
9 выветривание erosion озеро lake
10 поверхность surface подземный underground</p>
      <p>The fourth step is the most important in our method. We should associate
our units with topics taking into account topics’ support terms. If we use only
the traditional NMF representation we miss opportunities to exploit the
distributional power of support terms.</p>
      <p>For example, let’s consider the Unit #30 in the given Geology textbook. The
unit describes the history of glaciations as well as the impact of glaciers on the
Earth’s crust (see Table 7). So the geology expert has associated this unit with
the topic "Exogenous" as well as with the topic "Geochronology". In contrast,
the traditional NMF algorithm evaluates highly the relation of this unit with the
topic "Exogenous" and very lowly the relation with the topic "Geochronology".
But if one analyses the support terms used in the Unit #30, one finds that the
topic "Geochronology" is well represented in the unit with the help of support
terms such as "period", "year", "history", "epoch", "time", "Holocene", "cycle"
etc.</p>
      <p>Therefore, in order to more accurately define topics for each document’s
unit we should complement the traditional NMF approach by analyzing support
terms distribution in this unit. We should analyze next 3 factors:
1. What support terms related to the certain topic are occurred in this unit?
2. How frequently they are occurred?
3. How important are they for the certain topic (how many their weights in the
topic)?</p>
      <p>As a result we should decide can we associate this unit with some support
terms and with some topics represented via these terms. The decision rule can
be summarizes as follows:
where  is a support term related to the topic,  (,   ) is its frequency in
the unit, ℎ (,   ) is its weight in the topic,  ℎ is a threshold value
above which the topic is recognized as related to the unit. In this paper we set
 ℎ = 0.1.</p>
      <p>At the fifth step, we construct a navigation (semantic) map that contains
three layers: layers of document units, layers of support terms and layers of
topics. In Fig. 3 a part of the Geology textbook’s map is illustrated. This map
consists of 5 top-level nodes which correspond to textbook topics and 500
middlelevel nodes which correspond to support terms. Middle-level nodes can be moved
up or down or rolled up and stored away until one activates them again. Active
middle-level nodes point out the related units which are bottom-level nodes. If
middle level nodes are rolled up, access to bottom-level units is enable directly
through top-level nodes.</p>
      <p>Endogenous</p>
      <p>Earthquake</p>
      <p>Unit 31
Textbook</p>
      <p>Oceans
Solar system
Exogenous</p>
      <p>Mantle
Earth</p>
      <p>Wave
Geochronology</p>
      <p>Magnetic</p>
      <p>Unit 11</p>
      <p>Unit 82
По сейсмологическим
данным, в Земле сегодня
выделяют границ раздела,
в целом свидетельствующих
о концентрическом
расслоенном строении ...
So, the main advantage of NMF in comparison with LDA is a great
"naturalness", i.e. the topics in NMF are more widely intersect with each other by a
number of representative terms. LDA tries to generate topics in such a way so
that intersect as less as possible, as a result, each topic has its own set of
"exclusive" words. In practice, such a result does not look natural because in each
document there is a number of general thematic terms which refer to the subject
in a whole and can not belong only to one definite topic within the subject. For
example, in geology such terms are the words "rock", "process", "dynamics",
etc. NMF associates such terms with several topics, while LDA tries to assign
each such term to only one topic, as a result its stability sufers. In the new
series of experiments, a set of words referred by LDA to one topic can significantly
difer from the set of words referred to the same topic in the previous series of
experiments. In our experiments, NMF proved to be a more stable method than
LDA. The topical dispersion of representative key words does not hinder
segmentation of the document, on the contrary, it enhances segmentation not isolating
segment from each other but connecting them. The disadvantage of NMF is its
computing complexity. But this is the problem of computing technologies, not
of the method itself.
5</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion And Future Work</title>
      <p>In this paper we considered semantic map as a tool of smart document’s
segmentation and organization. We presented an approach to automatic creation
of semantic maps and showed how this process can benefit from a new
interpretation of NMF based on the concept of support terms. Also we performed a
little case study to illustrate proposed approach. However, more work should be
done to evaluate advantages and disadvantages of this approach, to substantiate
choice of the NMF parameters (e.g. the number of reasonable topics, or start
number of units, or threshold for topic validation). Eforts must also be devoted
to a complete comparison of proposed approach with LDA and NMF.
Acknowledgments. This work was supported in part under grant of
Foundation of Ministry of Education and Science of the Republic of Kazakhstan
"Development of intellectual high-performance information-analytical search system
of processing of semi-structured data" (2015-2017).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Caracciolo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Van Hage</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Rijke</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Towards topic driven access to full text documents</article-title>
          .
          <source>In: Research and Advanced Technology for Digital Libraries. LNCS, 8th European Conference, ECDL</source>
          <year>2004</year>
          ,
          <article-title>Bath</article-title>
          , UK,
          <source>September 12-17</source>
          ,
          <year>2004</year>
          , Proceedings. pp.
          <fpage>495</fpage>
          -
          <lpage>500</lpage>
          . Springer, Heidelberg (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Xing</surname>
            <given-names>K.</given-names>
          </string-name>
          et al.:
          <article-title>Adi: Towards a framework of app developer inspection</article-title>
          .
          <source>In: International Conference on Database Systems for Advanced Applications</source>
          . Springer International Publishing. pp.
          <fpage>266</fpage>
          -
          <lpage>280</lpage>
          . (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Blei D. M.</surname>
          </string-name>
          <article-title>Probabilistic topic models</article-title>
          .
          <source>J. Communications of the ACM</source>
          . Vol.
          <volume>55</volume>
          , issue 4, pp.
          <fpage>77</fpage>
          -
          <lpage>84</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Riedl</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biemann</surname>
            <given-names>C</given-names>
          </string-name>
          .
          <article-title>Text segmentation with topic models</article-title>
          .
          <source>Journal for Language Technology and Computational Linguistics</source>
          . Vol.
          <volume>27</volume>
          , issue 1, pp.
          <fpage>47</fpage>
          -
          <lpage>69</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Misra H</surname>
          </string-name>
          . et al.:
          <article-title>Text segmentation via topic modeling: an analytical study</article-title>
          .
          <source>In: 18th ACM conference on Information and knowledge management, ACM</source>
          , pp.
          <fpage>1553</fpage>
          -
          <lpage>1556</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Du</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Buntine</surname>
            <given-names>W. L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Topic Segmentation with a Structured Topic Model</article-title>
          . In:
          <article-title>Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies HLT-NAACL</article-title>
          , pp.
          <fpage>190</fpage>
          -
          <lpage>200</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kazantseva</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Szpakowicz</surname>
            <given-names>S.</given-names>
          </string-name>
          : Measuring Lexical Cohesion:
          <article-title>Beyond Word Repetition</article-title>
          .
          <source>In: International Conference on Computational Linguistics COLING</source>
          , pp.
          <fpage>476</fpage>
          -
          <lpage>485</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Du</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pate</surname>
            <given-names>J. K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Topic models with topic ordering regularities for topic segmentation</article-title>
          .
          <source>In:2014 IEEE International Conference on Data Mining. IEEE</source>
          , pp.
          <fpage>803</fpage>
          -
          <lpage>808</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sun</surname>
            <given-names>Q.</given-names>
          </string-name>
          et al.:
          <article-title>Text segmentation with LDA-based Fisher kernel</article-title>
          .
          <source>In: Proceedings of the 46th Annual Meeting of the Association for Computational Linguistics on Human Language Technologies: Short Papers. Association for Computational Linguistics</source>
          , pp.
          <fpage>269</fpage>
          -
          <lpage>272</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Frigyesi</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoglund</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Non-negative matrix factorization for the analysis of complex gene expression data: identification of clinically relevant tumor subtypes</article-title>
          .
          <source>J. Cancer informatics</source>
          , Vol.
          <volume>6</volume>
          , pp.
          <fpage>275</fpage>
          -
          <lpage>Џ292</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Hearst</surname>
            <given-names>M. A.</given-names>
          </string-name>
          :
          <article-title>TextTiling: Segmenting text into multi-paragraph subtopic passages</article-title>
          .
          <source>J. Computational linguistics</source>
          . Vol.
          <volume>23</volume>
          , issue
          <volume>1</volume>
          .
          <fpage>33</fpage>
          -
          <lpage>64</lpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. Choi F.:
          <article-title>Advances in domain independent linear text segmentation. In: 1st North American chapter of the Association for Computational Linguistics conference</article-title>
          .
          <source>Association for Computational Linguistics</source>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>33</lpage>
          (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Choi</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiemer-Hastings</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moore</surname>
            <given-names>J.:</given-names>
          </string-name>
          <article-title>Latent semantic analysis for text segmentation</article-title>
          .
          <source>In: EMNLP</source>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Stevens</surname>
            <given-names>K.</given-names>
          </string-name>
          et al.:
          <article-title>Exploring topic coherence over many models and many topics</article-title>
          .
          <source>In: Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Association for Computational Linguistics</source>
          , pp.
          <fpage>952</fpage>
          -
          <lpage>961</lpage>
          . (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Choo</surname>
            <given-names>J</given-names>
          </string-name>
          . et al.:
          <article-title>Utopian: User-driven topic modeling based on interactive nonnegative matrix factorization</article-title>
          .
          <source>J. Visualization and Computer Graphics</source>
          , IEEE Transactions on. Vol.
          <volume>19</volume>
          , issue 12, pp.
          <fpage>1992</fpage>
          -
          <lpage>2001</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Hu</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hao</surname>
            <given-names>Q</given-names>
          </string-name>
          . (ed.):
          <article-title>Intelligent sensor networks: the integration of sensor networks, signal processing and machine learning</article-title>
          . CRC Press (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>General</surname>
          </string-name>
          <article-title>Geology: in 2 vols</article-title>
          . Edited by
          <string-name>
            <surname>Professor L.K. Sokolovsky</surname>
          </string-name>
          . Moscow: KDU (in Russian) (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>