<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Word Sense Disambiguation in the Uzbek Language</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Axmedova Xolisxon Ilxomovna</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gulyamova Shakhnoza Qakhramonovna</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Murtazayeva Umida Isakulovna</string-name>
          <email>murtazayeva198202@gmail.com</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mavlonov Bokhodir Biloljonovich</string-name>
          <email>mavlonov@moti.uz</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alimova Shahnoza Maqsudovna</string-name>
          <email>shaxnozaalimova103@urdu.uz</email>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>International Islamic Academy of Uzbekistan</institution>
          ,
          <addr-line>Tashkent</addr-line>
          ,
          <country country="UZ">Uzbekistan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Mahalla and Family Research Institute</institution>
          ,
          <addr-line>Tashkent</addr-line>
          ,
          <country country="UZ">Uzbekistan</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Tashkent University of Applied Sciences</institution>
          ,
          <addr-line>Tashkent</addr-line>
          ,
          <country country="UZ">Uzbekistan</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Tashkent University of Information Technologies</institution>
          ,
          <addr-line>Samarkand</addr-line>
          ,
          <country country="UZ">Uzbekistan</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Urgench State University</institution>
          ,
          <addr-line>Urgench</addr-line>
          ,
          <country country="UZ">Uzbekistan</country>
        </aff>
      </contrib-group>
      <fpage>224</fpage>
      <lpage>232</lpage>
      <abstract>
        <p>In the field of Natural Language Processing (NLP), semantic analysis remains one of the most important and relevant challenges. Word Sense Disambiguation (WSD) is a key task within semantic analysis, and the automatic identification of homonymous words requires the application of modern algorithms. Machine learning (ML) algorithms and transformer-based models are among such approaches. In this study, the K-Nearest Neighbors (K-NN), Random Forest, and SenseBERT algorithms were applied to disambiguate homonymous words in the Uzbek language. A dataset containing sentences for diferent senses of Uzbek homonymous words was compiled. The collected data set was subjected to initial pre-processing procedures. After cleaning, the models were trained using Random Forest and Sense BERT algorithms. The trained models were tested, achieving an average accuracy of 92 %. To further improve accuracy, it is recommended to expand the size of the dataset and train the models on high-memory computing resources.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Word sense disambiguation</kwd>
        <kwd>Machine learning approaches</kwd>
        <kwd>K-NN algorithm</kwd>
        <kwd>Random forest algorithm</kwd>
        <kwd>Sense Bert model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Semantic analysis is itself divided into two categories: sentiment analysis and word sense
disambiguation. While sentiment analysis evaluates the emotional tone of a text, word sense disambiguation
helps to determine the intended meaning of a word within its context. The task of word-sense
disambiguation is considered one of the core areas of NLP and plays a crucial role in the field. Solving the
problem of identifying the correct meaning of a word in a natural language is achieved by semantically
distinguishing between diferent types of words.</p>
      <p>
        In numerous natural languages around the world, this problem has been explored and addressed
efectively. The commonly adopted approaches to word-sense disambiguation are presented below:
• Knowledge-based approaches
• Machine learning-based approaches
– Supervised learning algorithms
– Unsupervised learning algorithms
– Semi-supervised learning algorithms
• Neural network models
Numerous studies have been conducted across a wide range of the world’s languages to address the
problem of word sense disambiguation. In recent years, modern approaches have been applied to tackle
this task. For example, in Russian, neural network models have been utilized to identify homonyms [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ],
while in Kazakh [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], Arabic [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], Hindi [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], English [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], and again in Russian [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], various machine
learning algorithms and deep learning techniques have been employed to detect homonymous and
polysemous words.
      </p>
      <p>
        Similarly, in the Uzbek language, a number of studies have been conducted to determine the meaning
of words. The problem of word-sense disambiguation can be addressed using knowledge-based
approaches, though their accuracy is relatively limited. Such approaches are typically applied to determine
the contextual meaning of ambiguous words within a specific domain [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. However, to disambiguate
the meaning of a polysemous word in any given context, the use of machine learning-based algorithms
has proven to be more efective [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Supervised learning algorithms require a dataset consisting of semantically tagged data that belongs
to specific categories (1...). Each data point within the dataset is represented by a set of features
(1...2). The goal of the algorithm is to learn and identify the relationships between these features and
their corresponding categories, enabling it to accurately classify new, unseen data based on the learned
patterns.</p>
      <p>
        To determine word meaning in context, supervised learning requires a semantically tagged dataset
or corpus. For instance, when disambiguating polysemous words in English, the Sem-Cor corpus
is commonly used. Sem-Cor is a subset of the Brown Corpus, containing 234,000 words, where the
lexical units in each sentence are annotated with their corresponding WordNet senses. Naturally, these
WordNet sense annotations for the lexical units were manually labeled by human annotators [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>When applying these algorithms, two main types of features are typically used, often combined in
various ways: collocational features and co-occurrence features.</p>
      <p>Collocational features refer to specific words (along with their part-of-speech (POS) tags) that appear
at certain positions immediately to the left or right of the target word.
"Traktor matorining shovqini suruvdagi otlarni hurkitib yubordi"</p>
      <p>If the target word in this sentence is "ot+larni", its feature vector - consisting of the two words to the
left and two words to the right — would be as follows.
[shovqin, N, suruv, N, hurkitmoq, V, yubormoq, V].</p>
      <p>The feature vector is made up of the lemmas of the two preceding and two following words relative
to the current ambiguous word, along with their corresponding part-of-speech (POS) tags.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Main Part</title>
      <p>Co-occurrence features rely on the words surrounding the target word. In this approach, the features
are the surrounding words themselves, without considering their part-of-speech (POS) categories. The
value of each characteristic indicates how frequently these words appear in the context of the target
word.</p>
      <p>To apply this method efectively, a small number of meaningful words that frequently occur near
the target word are typically selected as features. For example, for the word "ot", the most frequently
occurring surrounding words from sentences containing its diferent senses might include the following:
tog’, dara, chipor, toy, poyga, dala, ...
"Zotdor otlar orasida poyga uchun qabul davom etmoqda. . . "</p>
      <p>If the window size is set to 10, this sentence can be represented as the following vector.
2.1. Unsupervised learning algorithm.</p>
      <p>In this approach, we start with an untagged training dataset — meaning we do not know which class
each data point belongs to. Only the features are available, and the algorithm must independently
identify which data points belong to the same class based on the patterns it discovers in the data.</p>
      <p>Machine learning approaches and transformer-based models ofer significant capabilities for
identifying the contextual meaning of homonymous words. In this study, we utilize machine learning
algorithms to semantically distinguish homonymous words within sentences in the Uzbek language.</p>
      <sec id="sec-2-1">
        <title>2.1.1. K-NN algorithm</title>
        <p>
          K-NN is a supervised machine learning algorithm that applies a classification method based on semantic
similarity [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. For an ambiguous word, it identifies the k nearest similar instances and determines the
word’s meaning based on these neighbors. The sequence of this process is illustrated in Figure 5.
        </p>
        <p>
          Feature extraction refers to the process known in English as Feature Extraction [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. In developing a
WSD (Word Sense Disambiguation) system, correctly extracting relevant features is crucial. The set of
features typically includes the following types:
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.1.2. Term Frequency (TF)</title>
        <p>
          In this method, the most frequently occurring words in the text are selected. Based on TF, the words
that appear most often in the context of a particular sense of an ambiguous word are identified, as they
help determine the intended meaning of the ambiguous word [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. When using the K-Nearest Neighbors
(K-NN) algorithm, the following steps are performed:
1. Data Preparation: The available data consists of items separated by other words (for example,
words in a text).
2. Distance Measurement: The similarity between words can be measured using a distance metric.
        </p>
        <p>
          In the K-NN algorithm, Euclidean distance is commonly used as a distance measure
3. Finding the k Nearest Neighbors: For a new word, the  closest neighboring words are selected
4. Learning and Classification: Based on these  neighbors, the meaning of the new word is
determined. In other words, the sense of an ambiguous word in a newly provided sentence is
identified by relating it to the overall meaning of its  neighboring words [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>In word sense disambiguation, the position and frequency of words within a text play an important
role. The weight of each word can be calculated based on the following formula:</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.1.3. Weighting Strategy</title>
        <p>The significance (weight) of a word is determined according to its frequency in the context and its
relative position within the sentence. This weighting approach helps prioritize more contextually
relevant words when identifying the meaning of an ambiguous word.</p>
        <p>(, ) =  (, ) (1)
 ()</p>
        <p>Here:  (, ) – the number of feature words  that co-occur with the word in sense ,  () – the
total number of samples for sense .
Average accuracy: 76.1</p>
        <p>The proposed method outperformed previously developed approaches.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.1.4. Random forest</title>
        <p>This algorithm uses multiple decision trees and yields good results in identifying complex homonyms.
Let’s review the process of determining the meanings of homonymous words in a text using the Random
Forest algorithm. This process consists of the following main stages:
• Homonyms and their context-based sentences are identified through a dataset (Excel file).
• The model is trained and then used to determine word senses in new texts
• The model’s accuracy is evaluated through testing.</p>
        <p>To build a model using the Random Forest algorithm, sentences are first collected for each meaning
of the homonymous words. The dataset for this algorithm was compiled in the following format.</p>
        <p>Libraries required for model development (Figure 8).</p>
        <p>Model Training and Evaluation</p>
        <p>Context Tag_sense
Kasr sonlarni o‘nli kasrga aylan- Siniq, parcha. Matematika termini;
tirish bo‘yicha dars o‘tdik miqdor birligining qismi
Bir litr sutning yarm kasri qay- Siniq, parcha. Matematika termini;
natildi miqdor birligining qismi.</p>
        <p>Bog‘ning uchdan bir kasri sabzavot Siniq, parcha. Matematika termini;
ekish uchun ishlatildi. miqdor birligining qismi
U kasr tufayli yugurish mu- Shikastalik, nuqsonlilik
sobaqasida qatnasha olmadi
Kun bulutlarni yorib, qishloqqa is- Yerga issiqlik va nur taratib
turusiqlik berdi vchi planeta, quyosh, oftob
U kun chiqishini suratga olish Yerga issiqlik va nur taratib
turuuchun dron ishlatdi vchi planeta, quyosh, oftob.
1. Sentences and their corresponding senses were extracted from an Excel file.
2. The sentences were converted into vector representations using the CountVectorizer method.
3. The dataset was divided into training and testing subsets in an 80/20 ratio using the train-test-split
function.</p>
        <p>4. A RandomForestClassifier() was employed to train the model.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>Evaluation of the Trained Model</title>
        <p>During the model development process, a total of 54,000 semantically annotated sentences were
collected. Of these, 43,200 sentences were used to train the model. The resulting model was then tested
on 10,800 instances. The test dataset contains 20 sentences for each meaning of Uzbek noun homonyms.
In the test results, each homonym was analyzed separately. The results of the analysis can be seen in
Table 3.</p>
        <p>In this article, the  1 scores are calculated for the words presented in Table 2. Based on the results
of the testing phase, the  1 score was calculated to evaluate the performance of the model. The test
results are summarized below:
• Total phrases recognized as valid expressions: 10800
• Correctly identified synonyms: 9950
• Incorrectly matched synonyms: 50
To assess the overall performance of the model, the  1 score was calculated based on these results.</p>
        <p>* 
 1 = 2 *   + 
  = 1000 (3)</p>
        <p>1000 + 50
 = 1000 (4)</p>
        <p>1000 + 9950
Based on the calculated precision values (3) and recall (4), (2) was found to be 0.664.</p>
        <p>The data was extracted from a corpus containing 460 million words, which includes news articles,
scientific abstracts, spoken-language transcripts, and literary texts. In terms of size, this represents
one of the largest training and testing datasets used in the field of word-sense disambiguation. The
compiled dataset was used to train a modern transformer-based model, specifically the Sense BERT
algorithm, and the resulting model was subsequently evaluated. The model trained with Sense BERT
demonstrated significantly higher accuracy compared to traditional approaches. In particular, the same
data set size was used for both the Random Forest and Sense BERT algorithms. The Sense BERT model
achieved an accuracy score of 92 %.
(2)</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>Modern approaches to word-sense disambiguation rely on several empirically observed linguistic
properties, particularly the principle that a word typically conveys a single meaning within a given
collocational and discourse context. Algorithms aim to exploit these properties by modeling the diversity
of collocational relationships as efectively as possible. In this study, the problem of identifying the
contextual meanings of homonymous words in the Uzbek language was addressed using Random
Forest, K-Nearest Neighbors (K-NN) and Sense BERT algorithms. Given that homonyms carry diferent
meanings depending on the context in which they occur, automatically determining their intended
sense is one of the essential challenges in natural language processing. The primary goal was to identify
the meaning of a homonymous word in a user-provided text based on its context.</p>
      <p>During the model development phase:
• A dataset in Excel format was compiled, containing columns for the homonym, the sentence, and
its corresponding meaning
• The texts were cleaned and lemmatized.
• Sentences were converted into numerical representations using vectorization techniques such as</p>
      <p>TF-IDF.
• The Random Forest algorithm was selected for model training because:
– It achieves high accuracy in classification tasks.
– It is resistant to overfitting.</p>
      <p>– It provides interpretable results.</p>
      <p>The final model successfully identified the contextual meaning of homonymous words in user-submitted
texts.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <sec id="sec-5-1">
        <title>The authors have not employed any Generative AI tools.</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Elov</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdurakhmonova</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Axmedova</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xusainova</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iskandarova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Fattaxova</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Designing Processes and Models of Semantical Diferentiation for Polyfunctional Words in the Uzbek Contexts</article-title>
          . In: Nguyen,
          <string-name>
            <surname>N.T.</surname>
          </string-name>
          , et al.
          <source>Recent Challenges in Intelligent Information and Database Systems. ACIIDS</source>
          <year>2024</year>
          , Communications in Computer and Information Science, vol.
          <volume>2145</volume>
          . Springer, Singapore (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Boltayevich</surname>
            ,
            <given-names>E.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirdjonovna</surname>
            ,
            <given-names>H.S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ilxomovna</surname>
            ,
            <given-names>A.X.</given-names>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Methods for Creating a Morphological Analyzer</article-title>
          . In: Zaynidinov,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Tiwary</surname>
          </string-name>
          , U.S., and
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>D</given-names>
          </string-name>
          . (Eds.),
          <source>Intelligent Human Computer Interaction. IHCI 2022, Lecture Notes in Computer Science</source>
          , vol.
          <volume>13741</volume>
          . Springer, Cham.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Boltayevich</surname>
            ,
            <given-names>E.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ilxomovna</surname>
            ,
            <given-names>A.X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hakim</surname>
            <given-names>Qizi</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>P.M.</given-names>
            , and
            <surname>Uktambay O'g'li</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.N.</surname>
          </string-name>
          (
          <year>2023</year>
          ).
          <article-title>Semantic Diferentiation of Uzbek Homonyms Using the Lesk Algorithm</article-title>
          .
          <year>2023</year>
          8th International Conference on Computer Science and Engineering (UBMK), Burdur, Turkiye, pp.
          <fpage>137</fpage>
          -
          <lpage>140</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>[4] Comparable Corpora to improve SMT performance Sadaf. 12th Conference of the European Chapter of the ACL</article-title>
          , pages
          <fpage>16</fpage>
          -
          <lpage>23</lpage>
          , Athens, Greece,
          <volume>30</volume>
          <fpage>March</fpage>
          - 3
          <source>April</source>
          <year>2009</year>
          .pp
          <fpage>16</fpage>
          -
          <lpage>23</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Gale</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Church</surname>
            , and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Yarowsky</surname>
          </string-name>
          .
          <article-title>: A Method for Disambiguating Word Senses in a Large Corpus</article-title>
          .
          <source>Computers and Humanities</source>
          , vol.
          <volume>26</volume>
          , pp.
          <fpage>415</fpage>
          -
          <lpage>439</lpage>
          (
          <year>1992</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , et al. (
          <year>2011</year>
          ).
          <article-title>Scikit-learn: Machine Learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>12</volume>
          , pp.
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Mihalcea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Moldovan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Word sense disambiguation with pattern learning and automatic feature selection</article-title>
          .
          <source>Natural Language Engineering</source>
          ,
          <volume>8</volume>
          (
          <issue>4</issue>
          ),
          <fpage>349</fpage>
          -
          <lpage>370</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Oqila</given-names>
            <surname>Abdullayeva</surname>
          </string-name>
          .
          <article-title>O'zbek tili matnlarida sintaktik teg va teglash masalasi. O'zbekiston: til va madaniyat</article-title>
          .
          <source>Kompyuter lingvistikasi</source>
          .
          <source>2024</source>
          Vol.
          <volume>1</volume>
          (
          <issue>6</issue>
          ) B.
          <fpage>46</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Ayazbayev</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bogdanchikov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Orynbekova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Varlamis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <article-title>Defining Semantically Close Words of Kazakh Language with Distributed System Apache Spark</article-title>
          .
          <source>Big Data Cogn. Comput</source>
          .
          <year>2023</year>
          ,
          <volume>7</volume>
          ,
          <fpage>160</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>El-Razzaz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fakhr</surname>
            ,
            <given-names>M. W.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Maghraby</surname>
            ,
            <given-names>F. A.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Arabic Gloss WSD Using BERT</article-title>
          .
          <source>Applied Sciences</source>
          ,
          <volume>11</volume>
          (
          <issue>6</issue>
          ),
          <fpage>2567</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Zhong</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
          </string-name>
          , H.T.:
          <article-title>It makes sense: a wide-coverage word sense disambiguation system for free text</article-title>
          .
          <source>In: ACL</source>
          <year>2010</year>
          -
          <article-title>48th Annual Meeting of the Association for Computational Linguistics</article-title>
          ,
          <source>Proceedings of the Conference</source>
          ,
          <volume>78</volume>
          -
          <fpage>83</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Zaripova</surname>
            <given-names>D.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loukachevitch</surname>
            <given-names>N.V.</given-names>
          </string-name>
          (
          <year>2023</year>
          )
          <article-title>Approaches to Automatic Word Sense Disambiguation Based on Uneven Distribution of Word Senses in Corpus</article-title>
          .
          <source>Lomonosov Philology Journal. Series 9</source>
          . Philology, no.
          <issue>6</issue>
          , pp.
          <fpage>40</fpage>
          -
          <lpage>51</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <article-title>Graph and Word Similarity for Word Sense Disambiguation, 2020 13th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)</article-title>
          , Chengdu, China,
          <year>2020</year>
          , pp.
          <fpage>1114</fpage>
          -
          <lpage>1118</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Sumanathilaka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Micallef</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Hough</surname>
          </string-name>
          ,
          <source>Exploring the Impact of Temperature on Large Language Models: A Case Study for Classification Task Based on Word Sense Disambiguation</source>
          ,
          <source>2025 7th International Conference on Natural Language Processing (ICNLP)</source>
          , Guangzhou, China,
          <year>2025</year>
          , pp.
          <fpage>178</fpage>
          -
          <lpage>182</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>U. R.</given-names>
            <surname>Dhungana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shakya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Baral</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>Word Sense Disambiguation using WSD specific WordNet of polysemy words</article-title>
          ,
          <source>Proceedings of the 2015 IEEE 9th International Conference on Semantic Computing (IEEE ICSC</source>
          <year>2015</year>
          ), Anaheim, CA, USA,
          <year>2015</year>
          , pp.
          <fpage>148</fpage>
          -
          <lpage>152</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>