<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Crypto social media analysis: Deep learning framework for Text Classification and Question Answering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jay Vaghela</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Siba Sankar Sahu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, Sardar Vallabhbhai National Institute of Technology</institution>
          ,
          <addr-line>Surat</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>In the current era, people frequently use social media platforms such as Facebook, Twitter, and Reddit to express their opinions. The opinion contains cryptocurrency-related content such as posts, tweets, and discussions. However, user-generated content contains frequent informal discussions. As part of the FIRE 2025 CryptoQA shared task, we investigate diferent deep learning models, where the goal is classification and question answering of cryptocurrency-related social media content. In this study, our team SVNIT_CSE investigated diferent deep learning models for text classification and Q&amp;A. Among the models evaluated, the GRU architecture provides an average Macro-F1 score of 0.52 on the Reddit, Twitter, and YouTube datasets. Similarly, GRU with Fasttext embedding provides optimal performance and ofers a Macro-F 1 score of 0.75 in Q&amp;A. The deep learning models evaluated provide competitive performance in cryptocurrency social media classification and Q&amp;A.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Cryptocurrency</kwd>
        <kwd>multiclass classification</kwd>
        <kwd>opinion extraction</kwd>
        <kwd>question answering</kwd>
        <kwd>social media</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The advent of cryptocurrencies has revolutionized financial systems and introduced decentralized
digital assets that work independently of old banking infrastructures. The rapid rise of cryptocurrencies
has transformed global financial landscapes, creating opportunities and uncertainties for investors,
researchers, and policymakers. The cryptocurrency ecosystem is heavily shaped by user-generated
content on social media platforms such as Twitter, Reddit, and YouTube. The presence of on-line content
provides a good source of information, but it also presents challenges. Social media texts are often noisy,
unstructured, and context-dependent, making it a challenging task to extract meaningful opinions and
reasonable information. Posts may include slang, abbreviations, emojis, sarcasm, or multimedia elements,
complicating automated analysis. In addition, content can range from factual reports and objective
analyses to highly subjective opinions, promotional spam, or outright misinformation, necessitating
sophisticated methods to discern valuable insights.</p>
      <p>In the rapid growth of digitization, the cryptocurrency data on social networks have grown
tremendously. Initially, the research relied on basic word matching techniques, using tools such as VADER
or SentiWordNet to tag posts as positive or negative. These approaches were simple and struggled to
understand context, informal language, and crypto-specific terms. So, few researchers explored machine
learning techniques to overcome limitations. They used diferent machine learning models such as
support vector machine (SVM), naive bayes, or random forest [9]. These models mainly focused on
text features, such as word frequency or short phrases, to classify sentiments or predict market trends.
These models are efective in some cases; however, the model provides poor performance in handling
noisy and informal data. The deep learning model provides better performance by capturing semantic
meaning of sentences and makes the model more suitable for sentiment analysis in social media data.</p>
      <p>
        In this study, we explore various deep learning models, including RNN, LSTM and GRU along with
diferent embedding techniques like Word2Vec [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], Glove [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and FastText [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to classify a cryptocurrency
post on social media into various categories. In addition, the model addresses various challenges by
focusing on opinion extraction and question answering from cryptocurrency-related social media
content. We found that the proposed deep learning models improve the performance in cryptocurrency
text classification and question answering.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem Statement</title>
      <p>FIRE1 (Forum for Information Retrieval Evaluation) organised a shared task called CryptoQA2 on text
classification and question answering from cryptocurrency-related social media posts [1]. The shared
task includes two diferent tasks. The objective of task 1 is to classify the social media post into multiple
classes. Similarly, the objective of task 2 is the question-answering from the QnA dataset, given a set of
question-comments pair, the task is to identify the relevance of the comment with respect to a question.
Together, these tasks aim to benchmark and advance the development of robust natural language
processing (NLP) systems that can understand, classify, and retrieve meaningful information from
noisy large-scale cryptocurrency data. Building an eficient NLP model can be directly implemented
for financial decision-making, market research, and the development of reliable information retrieval
systems in the domain of cryptocurrencies.</p>
      <p>The remainder of the paper is organized as follows. In Section 3 we present the state-of-the-art work
in the domain of cryptocurrency text classification and question &amp; answering. Section 4 describes
the statistics of the data set used in the experiment. Section 5 presents the methods used in the text
classification and QA system. The experimental results and their analysis are presented in Section 6.
Finally, we conclude with the direction of future work in Section 7.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>The analysis of social media content for cryptocurrency insights has gained significant attention
as researchers seek to understand public sentiment, predict market trends, and extract actionable
knowledge from diferent social media platforms such as YouTube, Twitter, and Reddit. In this section,
we present some of the existing work on opinion extraction and question answering in the context of
social networks.</p>
      <p>
        Bakliwal et al.[1] focused on the analysis of political sentiment on Twitter, specifically during the Irish
general elections. They classify tweets as positive, negative, or neutral sentiment. The authors explored
lexicon-based approaches and supervised machine learning methods. They found that lexicon-based
methods improved performance. However, the support vector machine provides the best efectiveness.
Rajadesingan &amp; Liu [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] introduced ReLP, a semi-supervised framework for detecting user opinions on
Twitter data. The framework combines a retweet-based label propagation algorithm with a supervised
classifier. ReLP demonstrates that leveraging retweet cooccurrence patterns for semi-supervised label
propagation can drastically reduce labeling efort while outperforming traditional supervised and
clustering baselines in identifying opposing stances on contentious political debates.
      </p>
      <p>Balikas, Moura &amp; Amini et al. [6] addressed the fine-grained Twitter sentiment classification by
leveraging the relationship between multiple sentiment analysis tasks. The author proposed a multitask
learning framework based on a bidirectional LSTM (biLSTM) network, which jointly learns both tasks to
improve classification performance. Bagdori &amp; Oard [5] explored cross-platform question answering by
linking unanswered questions on Twitter to potential answers from Yahoo! answers. They demonstrated
that a significant portion of twitter questions could be successfully answered using Yahoo! answers.
Askari et al. [11] introduced LegalQA, a new benchmark dataset for the legal domain. The data set
comprises 9,846 questions and 33,670 lawyer-verified answers. The author addressed two challenges:
the knowledge gap between lawyers and common people and the mixed formal or informal style of legal
QA forums. Nguyen et al. [8], introduces BERTweet, the first large-scale pre-trained language model
specifically designed for English Tweets. Built with the same architecture as BERTbase but trained
using the RoBERTa procedure, BERTweet addresses the challenges of informal and noisy Twitter text.
1https://fire.irsi.org.in/fire/2025/home
2https://sites.google.com/view/cryptoqa-2025/task-description
It was pre-trained on a massive corpus of 850M Tweets. The model outperformed RoBERTabase and
XLM-Rbase in all tasks.</p>
      <p>
        Roumeliotis, Tselikas &amp; Nasiopoulos [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] explored the capabilities of diferent large language models
(LLMs) to analyze the sentiment of cryptocurrency-related news articles. They fine-tuned models such
as GPT-4, BERT, and FinBERT. All models achieved strong performance with the Adam optimizer, with
GPT-4 showing the best results, closely followed by FinBERT and BERT. In another similar approach
employed by Kulakowski &amp; Frasincar [10] focuses on developing two new approaches to the sentiment
analysis of cryptocurrency-related social media posts. CryptoBERT - a Twitter oriented fined-tuned
BERTweet model improved performance, and the Language-Universal Cryptocurrency Emoji (LUKE)
sentiment lexicon, using Support Vector Machine, also enhanced performance. Deroy &amp; Maity [12]
investigated prompt-based learning with GPT-4 Turbo using zero-shot prompting for text classification
and few-shot learning for Q&amp;A. The model improved performance in diferent text classification and
Q&amp;A domain. Sarkar et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] created a multilevel dataset CryptOpiQA with inclusion of gold standard
(manually annotated) and silver standard (inferred from the gold standard) labels. SVM with BERT
and GPT-3.5 Turbo improve both the performance of Twitter and Reddit platforms, and GPT-3.5 ofers
optimal results.
      </p>
      <p>From the above study, we found that the various methods have been explored in the
cryptocurrencyrelated social media data. In this study, we investigated GRU-based models with FastText and Word2Vec
embeddings to address text classification and question answering in the social media domain.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Dataset</title>
      <p>The CryptoQA shared task is part of the FIRE (Forum for Information Retrieval Evaluation) evaluation
campaign comprising two subtasks. Task 1 is about text classification, and task 2 is about question and
answer. In task 1, three training data sets were provided from diferent social media platforms such as
Reddit, Twitter, and YouTube. Each data set consists of posts related to cryptocurrency and annotated
at following three hierarchical levels. The statistics of the training and testing data set are shown in
Tables 1 and 11. In task-2, a question and answering (Q&amp;A) dataset was provided. The statistics of
Q&amp;A dataset is shown in Table 6. The Q&amp;A dataset comprises queries corresponding to their comments
as columns. Moreover, the training data comments have a binary relevance. The statistic of relevance
and non-relevance comments is shown in Table 7.</p>
      <p>• Level 1: 0: NOISE, 1: OBJECTIVE, 2: SUBJECTIVE
• Level 2: Only for SUBJECTIVE posts → 0: NEUTRAL, 1: NEGATIVE, 2: POSITIVE
• Level 3: Only for NEUTRAL posts → 0: NEUTRAL SENTIMENT, 1: QUESTION, 2:
ADVERTISE</p>
      <p>MENT, 3: MISCELLANEOUS
The class distribution for the three training datasets is mentioned in Tables 3, 4 and 5.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Methodology</title>
      <p>We implement three steps to perform text classification and question-answering tasks on social media.
The three steps include data pre-processing, model design, and evaluation. In the preprocessing step,
the raw text is tokenized, segmented, and presented in a structured format. The raw text data are
cleaned and transformed into numerical sequences in the following ways.
• Data cleaning: The raw data contains missing values, null values, and the data set is imbalanced.</p>
      <p>Hence, we need necessary cleaning steps to provide consistency in the data. We remove missing
and null value data and balance the data set by implementing the class weights 3 method.
• Tokenization: A Keras tokenizer 4 was used to convert text into numerical sequences, limiting the
vocabulary to 10,000 words, and handling out-of-vocabulary words with a special token. Texts
were tokenized according to word frequency, and sequences were padded to a fixed length of 100
tokens to ensure uniform input for models.</p>
      <p>In the preprocessing step, the data set is cleaned and presented in a structured format. We used deep
learning models such as LSTM and GRU for text classification. Moreover, the QA model is designed
using deep learning methods with modern embedding-based approaches. Finally, we evaluate the text
classification and QA system using standard evaluation metrics.</p>
      <sec id="sec-5-1">
        <title>5.1. Opinion classification</title>
        <p>We built a unified pipeline model that classified social media content into diferent categories. In text
classification, we used long-short-term memory (LSTM) unit 5 and gated recurrent unit (GRU)6. A single
cell LSTM and GRU unit is shown in Figs. 1 and 2. The architecture of the LSTM unit comprises two
layers. First, the embedding layer with 128 dimensional dense vectors then two LSTM layers with 64
neurons and tanh activation function each, both having dropout of 0.3 to reduce overfitting. Ending
with a dense layer of 32 neurons with tanh activation and an output layer comprising the softmax
activation function. The GRU architecture contained three layers. First, the embedding layer with 128
dimensional dense vectors then three GRU layers with 128 neurons and tanh activation function each,
and uses dropout rate of 0.3 to reduce overfitting. Ending with a dense layer of 32 neurons with tanh
activation and an output layer comprising the softmax activation function. The hyperparameter used
in the training process is shown in Table 8.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Question Answering</title>
        <p>The main objective of the Q&amp;A is to check if the comments for the corresponding question in the data
set are relevant or not. The Q&amp;A model comprises RNN, LSTM, and GRU in conjunction with TF-IDF 7,
Word2Vec, GloVe and FastText embeddings. The hyperparameters used in the training process are
shown in table 9. To efectively represent textual data in the cryptocurrency Q&amp;A task, distributed word
embeddings were employed as word embeddings efectively keep semantic and syntactic relationships
3https://scikit-learn.org/stable/modules/generated/sklearn.utils.class_weight.compute_class_weight.html
4https://www.tensorflow.org/api_docs/python/tf/keras/preprocessing/text/Tokenizer
5https://shorturl.at/JyzA8
6https://shorturl.at/kZSwP
7https://www.learndatasci.com/glossary/tf-idf-term-frequency-inverse-document-frequency/
between words, so models such as LSTM and GRU to better understand the underlying contextual
meaning in social media posts.</p>
        <p>• Word2Vec Embedding: It is a widely used word embedding technique that uses a fixed-dimension
dense vector to represent words. It has two main architectures, that is, the continuous
bagof-words (CBOW) and skip-gram approach. In the CBOW approach, it predicts a word based
on context and the skip-gram approach predicts the surrounding word provided a target word.
Word2Vec capture the semantically similar words by putting words with similar meanings closer
together in vector space. It helps the model understand the domain-specific relationships between
words that frequently co-occur in cryptocurrency Q&amp;A task.
• FastText Embedding: It is introduced by Facebook, extends the Word2Vec model by representing
each word as a bag of character n-grams. Instead of learning a single vector for each word, the
model learns embeddings for character substrings, and the final representation of words is derived
from the sum of its n-gram vectors. The model generates embeddings for out-of-vocabulary
(OOV) words, which is highly useful for noisy social media data (misspellings, hashtags, slang,
abbreviations, etc.). Cryptocurrency discussions often contain informal expressions like holding
presented as ‘hodl’, bitcoin presented as ‘btc’, or ethereum as ‘eth2.0’. FastText’s subword modeling
captures these variations and makes more robust for domain-specific contexts. The representation
of words in FastText and Word2Vec 8 is shown in Figure 3.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results &amp; Discussion</title>
      <p>In this study, we explore various deep learning models and embedding techniques for multiclass
classification and question-answering for social media texts. To evaluate the efectiveness of the models,
we used accuracy and macro-F1 score. The evaluation was carried out on a masked test set of Reddit
(500 instances), Twitter (500 instances), Youtube (500 instances), and Q&amp;A (6323 instances). Table 10
presents the models submitted for text classification (Task 1) and Q&amp;A (Task 2). The performance of
the submitted models on diferent test data is shown in Table 11. From the evaluation results, we found
that the GRU architecture and GRU with FastText embedding techniques provides optimal performance
and outperformed other models in text classification and question answering. The evaluated model
provides competitive performance in both text classification and Q&amp;A task.</p>
      <p>In text classification, we found that the LSTM model provides the best performance on YouTube
(Macro-F1 = 0.69) and Reddit (Macro-F1 = 0.47), while the GRU architecture ofers the best performance
on Twitter (Macro-F1 = 0.49) and attained the best average score (Macro-F1 = 0.52). A similar observation
was also found in the validation data set, as shown in Table 12. From the evaluation results, we found
that both the models, i.e., LSTM and GRU provide a competitive score in text classification.</p>
      <p>In Q&amp;A, the GRU architecture with FastText embeddings achieved the highest Macro-F1 score (0.75)
on the test set, as shown in Table 11. GRU with FastText embeddings emerged as the most balanced
configuration, highlighting the importance of combining advanced recurrent models with subword-level
embedding methods for handling cryptocurrency-related social media text. A similar observation was
also found in the validation data set, as shown in Table 13. From the validation results, we found
that word embedding models provide better accuracy than traditional TF-IDF based methods. The
primary reason is that the word embedding model captures contextual meaning between words in the
embedding space. Diferent word embedding models provide a similar performance in diferent social
media data. In general, the findings indicate that the word embedding method with a deep learning
architecture provides optimal performance in Q&amp;A domain.</p>
      <p>According to the shared task evaluation policy, we can submit up to two runs for each task. Hence,
we evaluated the remaining models by ourselves based on the accuracy evaluation metric. Table 12
shows the accuracy achieved by diferent models in the validation data on text classification. Similarly,
Table 13 shows the accuracy achieved by diferent embeddings on the RNN, LSTM and GRU models in
the validation data on Q&amp;A. The efectiveness of diferent models is presented graphically in Figures 4
and 5.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>Text classification and Q&amp;A is an important downstream task in the text analysis domain. In the
cryptocurrency-related data, it presents unique challenges due to the noisy nature of user-generated
content. In this study, we explore diferent deep learning models with embedding techniques to handle
noisy content on social networks. From the evaluation results, we found that deep learning models are
efective at capturing the nuances of crypto-related discourse in both classification and QA tasks. Models
such as LSTM &amp; GRU provide similar performance in opinion classification. GRU with the FastText
model provides optimal performance and outperforms other models in Q&amp;A. Although deep learning
models provide promising results, several challenges remain. The performance of the models varied
considerably in diferent datasets and evaluation metrics, indicating that no single architecture provides
optimal performance in all datasets. In the future, we can explore the applicability of transformer
models in cryptocurrency-related data.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT, QuillBot in order to: Spelling Check,
Paraphrase and reword. After using this tool/service, the author(s) reviewed and edited the content as
needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Bakliwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Foster</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. van der Puil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. O</given-names>
            <surname>'Brien</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tounsi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Hughes</surname>
          </string-name>
          , “
          <article-title>Sentiment analysis of political tweets: Towards an accurate classifier</article-title>
          ,”
          <source>in Proceedings of the 2013 Conference of the Association for Computational Linguistics</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          , “
          <article-title>Eficient estimation of word representations in vector space</article-title>
          ,
          <source>” arXiv preprint arXiv:1301.3781</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pennington</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , “Glove:
          <article-title>Global vectors for word representation</article-title>
          ,”
          <source>in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4] [5] [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Rajadesingan</surname>
          </string-name>
          and H. Liu, “
          <article-title>Identifying users with opposing opinions in twitter debates,” in International conference on social computing, behavioral-cultural modeling</article-title>
          , and prediction, Springer,
          <year>2014</year>
          , pp.
          <fpage>153</fpage>
          -
          <lpage>160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Bagdouri</surname>
          </string-name>
          and
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Oard</surname>
          </string-name>
          , “
          <article-title>Building bridges across social platforms: Answering twitter questions with yahoo! answers,”</article-title>
          <source>in Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1181</fpage>
          -
          <lpage>1184</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>G.</given-names>
            <surname>Balikas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Moura</surname>
          </string-name>
          , and
          <string-name>
            <surname>M.-R. Amini</surname>
          </string-name>
          , “
          <article-title>Multitask learning for fine-grained twitter sentiment analysis</article-title>
          ,
          <source>” in Proceedings of the 40th International ACM SIGIR conference on research and development in information retrieval</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1005</fpage>
          -
          <lpage>1008</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bojanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Grave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joulin</surname>
          </string-name>
          , and T. Mikolov, “
          <article-title>Enriching word vectors with subword information,” Transactions of the Association for Computational Linguistics</article-title>
          , vol.
          <volume>5</volume>
          , pp.
          <fpage>135</fpage>
          -
          <lpage>146</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>D. Q.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Vu</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A. T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          , “
          <article-title>Bertweet: A pre-trained language model for english tweets</article-title>
          ,
          <source>” EMNLP</source>
          <year>2020</year>
          , p.
          <fpage>9</fpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Pabuçcu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ongan</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Ongan</surname>
          </string-name>
          , “
          <article-title>Forecasting the movements of bitcoin prices: An application of machine learning algorithms</article-title>
          ,
          <source>” Quantitative Finance and Economics</source>
          , vol.
          <volume>4</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>679</fpage>
          -
          <lpage>692</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Kulakowski</surname>
          </string-name>
          and
          <string-name>
            <given-names>F.</given-names>
            <surname>Frasincar</surname>
          </string-name>
          , “
          <article-title>Sentiment classification of cryptocurrency-related social media posts,” IEEE Intelligent Systems</article-title>
          , vol.
          <volume>38</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>5</fpage>
          -
          <lpage>9</lpage>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Askari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ren</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Verberne</surname>
          </string-name>
          , “
          <article-title>Answer retrieval in legal community question answering</article-title>
          ,
          <source>” in European Conference on Information Retrieval</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>477</fpage>
          -
          <lpage>485</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>A.</given-names>
            <surname>Deroy</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Maity</surname>
          </string-name>
          , “
          <article-title>Cryptollm: Unleashing the power of prompted llms for smartqna and classification of crypto posts</article-title>
          ,
          <source>” arXiv preprint arXiv:2411.07917</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>K. I.</given-names>
            <surname>Roumeliotis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Tselikas</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Nasiopoulos</surname>
          </string-name>
          , “
          <article-title>Llms and nlp models in cryptocurrency sentiment analysis: A comparative classification study</article-title>
          ,
          <source>” Big Data and Cognitive Computing</source>
          , vol.
          <volume>8</volume>
          , no.
          <issue>6</issue>
          , p.
          <fpage>63</fpage>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sarkar</surname>
          </string-name>
          et al., “Cryptopiqa:
          <article-title>A new opinion and question answering dataset on cryptocurrency,”</article-title>
          <source>in Proceedings of the 31st International Conference on Computational Linguistics</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>