<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Dublin, Ireland, March</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>A Framework for Constructing Concept Maps from E-Books Using Large Language Models: Challenges and Future Directions</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Boxuan Ma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Li Chen</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Art and Science, Kyushu University</institution>
          ,
          <country country="JP">Japan</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Information Science and Electrical Engineering, Kyushu University</institution>
          ,
          <country country="JP">Japan</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>0</volume>
      <fpage>3</fpage>
      <lpage>07</lpage>
      <abstract>
        <p>Concept maps have been widely used in education to organize and represent information hierarchically. However, traditional methods for constructing concept maps often depend on human experts, which can be costly and time-consuming. The emergence of large language models (LLMs), such as GPT-4, has transformed concept construction and reasoning tasks by offering automated and scalable solutions. This paper introduces a novel framework for generating concept maps of e-books with three key components: section segmentation, key concept extraction, and relationship identification. Additionally, the paper highlights challenges and future opportunities to enhance LLM-driven concept map generation for educational applications.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;E-book</kwd>
        <kwd>LLM</kwd>
        <kwd>ChatGPT</kwd>
        <kwd>Concept Map</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the digital age, e-books have become widely used for education. Their portability, accessibility, and
search functionality have made them popular resources across diverse domains [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2,3,4</xref>
        ]. However,
their inherently linear structure often falls short of meeting the needs of learners who benefit from
a more interconnected and structured view of complex content [
        <xref ref-type="bibr" rid="ref2 ref3 ref5">2,3,5</xref>
        ].
      </p>
      <p>
        Concept maps are useful tools to address these challenges. They are visual representations that
organize and represent knowledge by highlighting relationships between key concepts. These maps
have been widely recognized for their ability to enhance comprehension, retention, and critical
thinking skills by providing learners with a structured and interconnected view of content [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. For
instance, they enable learners to identify overarching themes and gain insights into the hierarchical
structure of information. Such cognitive aids are particularly valuable in domains requiring a
systemslevel understanding. Additionally, concept maps are utilized as navigational tools in e-books, allowing
learners to interact with the material more effectively. In these methods, concept maps are displayed
as an interactive interface, where learners can click on individual nodes to directly access the
corresponding pages or sections related to each concept [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Traditionally, the construction of concept maps has relied on manual processes, which are
timeconsuming and resource-intensive. This has limited the scalability of concept maps, especially in
large-scale educational contexts. While researchers have made strides in automatically extracting
concepts from teaching materials using NLP methods [
        <xref ref-type="bibr" rid="ref7 ref8">7,8</xref>
        ], these methods still rely on human labeling
and lack interactivity generation capabilities.
      </p>
      <p>The advent of large language models (LLMs) has introduced a transformative potential for
automating concept map generation. Leveraging their advanced contextual understanding, LLMs can
produce meaningful content with minimal human oversight, and existing studies have demonstrated
that LLMs are capable of a wide range of tasks beyond text summarization, translation, and
refinement. Such versatility highlights their potential for automatically generating concept maps.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Concept maps have long been employed in the education area as a means to organize information
and improve students' comprehension hierarchically. Constructing a concept map typically involves
multiple tasks, such as concept extraction and relation identification [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Researchers have recently
made strides in automatically completing these tasks from textual data using NLP methods.
      </p>
      <p>Traditional concept extraction methods rely on Term Frequency-Inverse Document Frequency
(TF-IDF), Latent Dirichlet allocation (LDA) topic model, co-occurrence statistics, and neighboring
document analysis. For example, TextRank [17] is a well-known method that uses a co-occurrence
graph to rank key concepts. ExpandRank [18] leverages neighborhood documents to improve key
concept extraction. In addition, many approaches incorporate external knowledge sources to enrich
concept extraction, such as Wikipedia and Knowledge Base [14, 15].</p>
      <p>
        While concept extraction focuses on identifying standalone terms or phrases, constructing
concept maps involves defining relationships between concepts. Early work in this area used textbook
structures to organize extracted concepts and construct prerequisite relationships using features
from Wikipedia [15]. [19] derived prerequisite relations from ontologies, translating interactions
among instances into relationships. Similarly, Liang et al. proposed an optimization-based framework
to uncover concept prerequisites from course dependencies [21]. Gordon et al. introduced a
crossentropy approach and an information-flow approach for discovering concept dependency relations
automatically from a text corpus [20]. Pan et al. proposed a representation learning-based method
to learn the concepts and focused on generating prerequisite relations among concepts on a Massive
Open Online Courses (MOOCs) corpus [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
      <p>The release of large language models like GPT-4, recognized for their remarkable general
capabilities, has been considered by researchers as the spark of artificial general intelligence and has
introduced transformative possibilities for concept map construction. LLMs excel in understanding
and processing language, making them ideal for generating concept maps [10]. LLMs have been
evaluated for their adaptability in creating knowledge representations for education [13, 16, 23]. For
instance, de Paiva et al. used ChatGPT to extract Category Theory concepts from academic papers,
showcasing its ability to identify mathematical entities [22]. Recently, Li et al. [11] and Chen et al. [12]
utilized LLMs to analyze learning material texts to identify knowledge concepts and their
interrelationships. However, their work uses knowledge concepts to predict student performance or
assist collaborative problem-solving. And none of them evaluate the effectiveness of LLM-generated
concepts and relationships.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Framework</title>
      <p>As shown in Figure 1, we propose a framework for LLM-based e-book concept map construction.
Specifically, the process involves e-book section segmentation, key concept extraction from each
section, and relationship identification between concepts for each section. Finally, it merges and
refines the results of each section to create a comprehensive concept map for the e-book. The
prompts we used can be seen in Figure 2.</p>
      <p>The first step in constructing concept maps involves segmenting the e-book content into distinct
sections using LLM. The objective is to break down the material into units that facilitate concept
extraction. This structured segmentation ensures that each section contains logically connected
content, setting the stage for accurate concept extraction and relationship identification. E-book
lectures typically have inherent structure, such as chapters and headings, which is helpful for
segmentation tasks. After the section segmentation, the LLM is prompted to identify the main
concepts within each specified segment. The process involves carefully designed prompts that direct
the LLM to identify key concepts within the section and obtain a set of knowledge concept entities K
= {k1, k2, ..., kn}. After identifying the main concepts, the LLM proceeds to map relationships between
them. These relationships are fundamental in constructing a concept map, as they provide the
connective structure that illustrates how concepts interrelate. The LLM's ability to contextualize
relationships is enhanced through example-based prompting and iterative refinement. This step is
crucial for ensuring that the generated concept map accurately represents the material's underlying
structure. With the concepts and relationships identified for each section, the next step is to construct
the concept maps. These maps are visualized as directed graphs, where nodes represent concepts
and edges represent relationships. Once all section-level concept maps are generated, they are
aggregated into a unified concept map representing the entire e-book content. During this
aggregation, cross-section relationships are identified to ensure coherence and connectivity across
the material.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>4.1.</p>
      <sec id="sec-4-1">
        <title>Dataset</title>
        <p>We evaluated LLMs based on our framework to have an initial understanding of their capabilities by
evaluating their performance on concept map construction. We collected lecture files from an
Introductory Python Programming course at our university. A total of 12 lecture files were used in
this study. The content covers the basics of Python programming, including functions, variables,
strings, lists, branches, loops, and so on.
4.2.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Preliminary Results</title>
        <p>We tested GPT4o because of its affordability and widespread adoption. We compare the results
generated by GPT and those generated by the instructor.
The evaluation results show that GPT-4o exhibited exceptional accuracy in section segmentation.
Across 12 lectures, it divided the content into 57 sections, averaging 4.75 sections per lecture. Notably,
the segmentation of 10 out of 12 lectures was perfectly aligned with the instructor's structure. This
high level of accuracy can be attributed to the inherent structure of e-books, such as outlines and
chapter headings provided on the page, which provided clear guidance for GPT-4o’s section
segmentation. Minor discrepancies occurred in two lectures: in one case, GPT added an extra section,
and in another, it omitted a section. These deviations were related to the course's pedagogical design.
4.2.2.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Concept Extraction</title>
        <p>GPT generated a diverse range of relationships, which we carefully reviewed. Since no constraints
were placed on the generation process, the resulting relationships sometimes lacked clear logical
structure or coherence. It generated a total of relationships, including 154 hierarchical relationships
and 70 other relationships. For hierarchical relationships, GPT performed well in most cases,
accurately identifying connections that aligned with the internal structure of the chapters or content.
However, GPT also produced several relationships that, while seemingly plausible at first glance,
offered little meaningful insight. These extraneous relationships, though not entirely incorrect, did
not contribute significantly to enhancing the conceptual understanding and may have added
unnecessary complexity to the overall map. To refine this process, future improvements could involve
introducing constraints or guiding rules to ensure the relationships generated are more focused,
meaningful, and aligned with the specific needs of the concept map. This would help balance creative
exploration with logical coherence, maximizing the utility of the extracted relationships.
4.2.4.</p>
      </sec>
      <sec id="sec-4-4">
        <title>Concept Map Construction</title>
        <p>The final step involved having GPT use its generated concepts and relationships to construct a concept
map. We observed that the resulting map primarily employs hierarchical relationships, with most of
the content being accurate and well-organized. Figure 3 showcases an example of a generated
concept map. Interestingly, we noticed that the structure of the concept map generated by GPT
differed from the original e-book structure designed by the instructor. For example, in Lecture 1, the
e-book's structure was organized as follows: (1) Program, (2) Programming Language, (3) Introduction
to Python, and (4) Function. In contrast, GPT's concept map adhered more closely to the logical and
content-based relationships within the knowledge rather than strictly following the e-book’s
predefined order. This divergence suggests that GPT is capable of reinterpreting content to create
maps that emphasize conceptual connections and hierarchy. While the instructor’s structure is
pedagogically motivated, the GPT-generated map offers an alternative perspective that focuses on
content flow and interrelationships, potentially enhancing understanding by presenting concepts in
a way that mirrors their logical and thematic linkages.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The evaluation of large language models in concept map construction underscores their significant
potential for advancing educational content creation and content analysis. Across the four stages of
the framework, LLMs demonstrated notable accuracy, producing well-structured and coherent
concept maps that closely aligned with the source e-book. Their ability to integrate these distinct
tasks highlights their versatility and efficiency. While many NLP-based approaches can perform tasks
such as concept extraction or relationship identification, they are often task-specific, requiring
separate models for each step of the concept map construction process. This fragmented approach
limits their scalability and efficiency. Furthermore, traditional methods frequently lack context
awareness, which can lead to semantic drift—a phenomenon where knowledge concepts with similar
semantics but from unrelated domains are incorrectly generated or linked [14]. LLMs, however,
effectively address this challenge by leveraging their advanced contextual understanding to maintain
semantic consistency and ensure the relevance of extracted concepts within a given domain. The
unique strengths of LLMs to optimize resource utilization make them outperform smaller,
taskspecific models in adaptability for diverse application domains and data-limited settings. By unifying
multiple tasks under one framework, LLMs reduce the complexity of the process and improve overall
coherence, positioning themselves as indispensable tools for concept map generation and reasoning.</p>
      <p>Despite these advantages, several challenges and areas for improvement remain. In the following
section, we delve deeper into these challenges and propose actionable insights to guide future
developments in leveraging LLMs for concept map construction.
5.1.
5.1.1.</p>
      <sec id="sec-5-1">
        <title>Challenges</title>
      </sec>
      <sec id="sec-5-2">
        <title>Misalignment with Educational Goals</title>
        <p>Based on our observation, one significant challenge is the misalignment between LLM-generated
results and pedagogical intentions. Although the concepts generated by GPT are often drawn directly
from e-book content, some of these concepts may not align with the educational priorities or the
core objectives of the lesson. For instance, GPT might include concepts that are mentioned in passing
within the material but are not critical to the intended teaching focus. This divergence can dilute the
map's effectiveness as a teaching tool, highlighting the need for mechanisms to guide LLMs toward
emphasizing key educational priorities while minimizing peripheral or non-essential concepts.
5.1.2.</p>
      </sec>
      <sec id="sec-5-3">
        <title>Balancing Human Oversight and Automation</title>
        <p>Another key issue is the balance between human oversight and full automation. While GPT is
powerful, it is not yet capable of consistently producing flawless results. Therefore, generating
accurate concepts and relationships often requires iterative refinement, which calls for human
interaction at various stages. A human-in-the-loop approach is essential, not only to correct errors
but also to ensure that the generated content aligns with teaching objectives. Although such an
approach can increase time and labor costs if every step of the process requires human intervention,
on the flip side, GPT can serve as a valuable assistant, enabling teachers to refine and enhance the
educational design. Striking the right balance between GPT autonomy and human involvement
remains a critical area for improvement.
5.1.3.</p>
      </sec>
      <sec id="sec-5-4">
        <title>The Hallucination Problem in LLMs</title>
        <p>Hallucination, the tendency of LLMs to generate inaccurate or nonsensical information, poses
another challenge. While LLMs perform with high accuracy in concept generation and rarely produce
irrelevant or false concepts, the issue becomes more pronounced when identifying relationships.
Hierarchical relationships, such as part-of relationships, are often reliable due to their grounding in
the structure of the content. However, when tasked with generating other types of relationships
without explicit constraints, LLMs may produce plausible but unhelpful or unreliable connections.
This highlights the need for strict-designed prompts when generating complex relationships. Existing
research suggests that GPT performs better when constrained to choose from predefined options
rather than generating relationships freely [10]. Despite these limitations, ChatGPT’s ability to
generate hierarchical relationships is reliable. For purposes such as creating navigation-oriented
concept maps or tools to aid reading comprehension, it already shows strong potential and could
become a powerful asset in educational content creation with further refinement.
5.2.</p>
      </sec>
      <sec id="sec-5-5">
        <title>Future Directions</title>
        <p>Looking ahead, the application of LLMs for generating e-book concept maps presents exciting
opportunities. Below, we outline three key directions to enhance the development and integration
of LLM-generated concept maps for e-books.
5.2.1.</p>
      </sec>
      <sec id="sec-5-6">
        <title>Fine-Tuning and Domain Adaptation</title>
        <p>Integrating educational design principles into the LLM generation process is important. This
alignment will ensure that generated concept maps better reflect instructional goals and pedagogical
priorities. Fine-tuning LLMs with domain-specific datasets offers a promising avenue, enabling
models to extract nuanced concepts and accurately map relationships. Incorporating specialized
corpora, such as annotated academic texts or field-specific materials, can assist in generating
concepts that are not only relevant to the content but also central to the instructional goals.
Moreover, domain-specific fine-tuning tailored to particular subjects can enhance the precision and
relevance of generated concept maps [16].
5.2.2.</p>
      </sec>
      <sec id="sec-5-7">
        <title>Concept Expansion and Categorization</title>
        <p>Beyond the concepts explicitly written in e-books, supplementary concepts can provide valuable
context and open avenues for deeper exploration [14]. For example, in Figure 3, when introducing
Built-in Functions, while the core concept is print() function, it is valuable to introduce other built-in
functions, such as input(). These supplementary concepts connect the core concept to broader
contexts, encouraging curiosity and fostering a more comprehensive understanding. LLMs’ ability to
expand on core concepts offers a transformative approach to enhancing student understanding and
engagement. By leveraging its vast knowledge base and contextual reasoning, GPT bridges the gap
between narrowly focused course content and the broader spectrum of relevant knowledge, enabling
students to explore beyond the standard curriculum.</p>
        <p>To further enhance the utility of concept expansion, LLMs can categorize concepts into distinct
groups, ensuring clarity and prioritization. Core concepts represent the foundational knowledge that
aligns directly with course objectives, while supplementary concepts provide additional insights or
alternative perspectives to deepen understanding. Advanced concepts cater to more inquisitive
students, offering pathways for further exploration beyond the standard syllabus. By organizing
concepts into these categories, LLMs can empower students to focus on what is essential while also
providing opportunities for enrichment. For educators, this structured approach offers flexibility to
tailor content based on student needs, ensuring that both the core curriculum and optional
extensions are effectively addressed. Such categorization not only supports personalized learning but
also ensures that the expanded knowledge remains manageable and relevant.</p>
      </sec>
      <sec id="sec-5-8">
        <title>Communicative Intelligent Agents for Concept Map Construction</title>
        <p>The efficacy of LLMs heavily leans on human engagement in dialogue generation. Further refining
model responses necessitates intricate user task descriptions and enriched interaction contexts, a
process that remains demanding and time-intensive in the development lifecycle. Consequently,
research efforts concerning intelligent agents can independently generate prompts and carry out
tasks. Inspired by Zhu et al. [10], we introduced a framework for concept map construction using
communicative intelligent agents, each assigned specific roles. As shown in Figure 4, we can assign
specific roles for agents, such as a Concept map administrator, a Domain expert, and a Data Engineer,
to collaborate to complete tasks iteratively.</p>
        <p>The administrator acts as a coordinator, clarifying ambiguities and aligning the agents’ work with
instructional goals. The domain expert agent generates concepts and structures, ensuring coverage
of core material and verifying the accuracy and relevance of expanded concepts. Also, the data
engineer agent acts as a web searcher. It retrieves additional information for supplementary concepts
from external sources, such as Wikipedia and Stack Overflow, to help expand concepts with
realworld examples or related threads. The iterative process allows agents to consult, update, and refine
their outputs based on feedback, ensuring higher accuracy and alignment with user needs.
5.2.4.</p>
      </sec>
      <sec id="sec-5-9">
        <title>Integrating with E-book Systems</title>
        <p>
          Integrating concept maps into e-book systems has the potential to revolutionize digital learning. By
providing learners with interactive navigation aids, it can transform static e-books into adaptive
learning environments [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. Users could click on concepts to access definitions, explore related topics,
or view hierarchical structures dynamically. This interactivity would help students navigate complex
material more effectively, fostering deeper understanding [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Additionally, incorporating user
feedback loops into the system could significantly improve the quality of LLM-generated concept
maps [14]. Educators and students could add missing concepts, refine incorrect relationships, or
suggest alternative interpretations, creating a collaborative and iterative refinement process [
          <xref ref-type="bibr" rid="ref6">6, 14</xref>
          ].
Furthermore, tracking student interactions with the e-book system could help identify relationships
between concepts that might otherwise go unnoticed, offering valuable insights for improving both
the maps and the learning experience. For instance, recently, Lu et al. [24] integrated LLM-generated
concept maps and student-concept-page interaction data to generate a reading path dashboard. By
seamlessly integrating LLMs and e-book systems, the instructional content and student interaction
data could create a more comprehensive visualization of the concept map for e-books.
        </p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this paper, we explored the capabilities of large language models in constructing e-book concept
maps and introduced a framework encompassing section segmentation, key concept extraction, and
relationship identification. We evaluated the framework’s performance using a real-world dataset,
demonstrating the significant potential of LLMs in this task. Furthermore, we discussed the challenges
and opportunities associated with leveraging LLMs for e-book concept map construction and
corresponding applications. For future work, we plan to evaluate the framework on larger and more
diverse datasets. Additionally, extending this study to include other LLMs could provide valuable
insights into their comparative performance and generalizability.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This work was supported by JSPS KAKENHI Grant Number JP24K20903.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used GPT-4o in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and
take full responsibility for the publication’s content.
[10] Yuqi Zhu, Xiaohan Wang, Jing Chen, Shuofei Qiao, Yixin Ou, Yunzhi Yao, Shumin Deng, Huajun
Chen, and Ningyu Zhang. "LLMs for knowledge graph construction and reasoning: Recent
capabilities and future opportunities." World Wide Web 27, no. 5 (2024): 58.
[11] Gen Li, Cheng Tang, Li Chen, Daisuke Deguchi, Takayoshi Yamashita, and Atsushi Shimada.
"LLMDriven Ontology Learning to Augment Student Performance Analysis in Higher Education." In
International Conference on Knowledge Science, Engineering and Management, pp. 57-68.
Singapore: Springer Nature Singapore, 2024.
[12] Li Chen, Gen Li, Boxuan Ma, Cheng Tang, Masanori Yamada. A Three-Step Knowledge Graph
Approach Using LLMs In Collaborative Problem Solving-based Stem Education. In 21st International
Conference on Cognition and Exploratory Learning in Digital Age, CELDA 2024. IADIS Press, 51–
58.
[13] Siwi, Satria, Kemas Wiharja, and Raihan Hafizh. "Transforming Indonesian Geography Education
Books Into Knowledge Graphs Using ChatGPT LLMs." In 2024 12th International Conference on
Information and Communication Technology (ICoICT), pp. 50-56. IEEE, 2024.
[14] Jifan Yu, Chenyu Wang, Gan Luo, Lei Hou, Juanzi Li, Zhiyuan Liu, and Jie Tang. "Course Concept
Expansion in MOOCs with External Knowledge and Interactive Game." In Proceedings of the 57th
Annual Meeting of the Association for Computational Linguistics, pp. 4292-4302. 2019.
[15] Shuting Wang, Alexander Ororbia, Zhaohui Wu, Kyle Williams, Chen Liang, Bart Pursel, and C.</p>
      <p>Lee Giles. "Using prerequisites to extract concept maps fromtextbooks." In Proceedings of the
25th acm international on conference on information and knowledge management, pp. 317-326.
2016.
[16] Reales, Daniel, Rubén Manrique, and Christian Grévisse. "Core Concept Identification in
Educational Resources via Knowledge Graphs and Large Language Models." SN Computer Science 5, no.
8 (2024): 1029.
[17] Rada Mihalcea, and Paul Tarau. "Textrank: Bringing order into text." In Proceedings of the 2004
conference on empirical methods in natural language processing, pp. 404-411. 2004.
[18] Xiaojun Wan, and Jianguo Xiao. "Single document keyphrase extraction using neighborhood
knowledge." In AAAI, vol. 8, pp. 855-860. 2008.
[19] Shian-Shyong Tseng, Pei-Chi Sue, Jun-Ming Su, Jui-Feng Weng, and Wen-Nung Tsai. "A new
approach for constructing the concept map." Computers &amp; Education 49, no. 3 (2007): 691-707.
[20] Jonathan Gordon, Linhong Zhu, Aram Galstyan, Prem Natarajan, and Gully Burns. "Modeling
concept dependencies in a scientific corpus." In Proceedings of the 54th Annual Meeting of the
Association for Computational Linguistics (ACL), pp. 866-875. 2016.
[21] Chen Liang, Jianbo Ye, Zhaohui Wu, Bart Pursel, and C. Giles. "Recovering concept prerequisite
relations from university course dependencies." In Proceedings of the AAAI Conference on
Artificial Intelligence, vol. 31, no. 1. 2017.
[22] Valeria de Paiva, Qiyue Gao, Pavel Kovalev, and Lawrence S. Moss. "Extracting Mathematical</p>
      <p>Concepts with Large Language Models." arXiv preprint arXiv:2309.00642 (2023).
[23] Tianyuan Yang, Baofeng Ren, Chenghao Gu, Boxuan Ma, and Shin’ichi Konomi. “Leveraging
Chatgpt for Automated Knowledge Concept Generation.” In Proceedings of International
Conference on Cognition and Exploratory Learning in Digital Age (CELDA 2024), 2024.
[24] Min Lu, Boxuan Ma, Xuewang Geng, and Masanori Yamada. Enhancing E-Book Learning
Dashboards with GPT-Assisted Page Grouping and Adaptive Navigation Link Visualization. In
Companion Proceedings of the 15th International Conference on Learning Analytics &amp; Knowledge (LAK25),
2025.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Liang-Yi</surname>
            <given-names>Li</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gwo-Dong Chen</surname>
          </string-name>
          , and
          <string-name>
            <surname>Sheng-Jie Yang</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Construction of cognitive maps to improve e-book reading and navigation</article-title>
          .
          <source>Computers &amp; Education</source>
          <volume>60</volume>
          ,
          <issue>1</issue>
          (
          <year>2013</year>
          ),
          <fpage>32</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Boxuan</given-names>
            <surname>Ma</surname>
          </string-name>
          , Min Lu, Yuta Taniguchi, and
          <string-name>
            <surname>Shin'ichi Konomi</surname>
          </string-name>
          .
          <year>2022</year>
          .
          <article-title>Exploring jump back behavior patterns and reasons in e-book system</article-title>
          .
          <source>Smart Learning Environments 9</source>
          ,
          <issue>1</issue>
          (
          <year>2022</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Boxuan</given-names>
            <surname>Ma</surname>
          </string-name>
          , Jiadong Chen,
          <string-name>
            <given-names>Chenhao</given-names>
            <surname>Li</surname>
          </string-name>
          , Likun Liu, Min Lu, Yuta Taniguchi, and
          <string-name>
            <surname>Shin'ichi Konomi</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Understanding jump back behaviors in e-book system</article-title>
          .
          <source>In Comanion Proceedings of the 10th International Conference on Learning Analytics &amp; Knowledge</source>
          .
          <fpage>623</fpage>
          -
          <lpage>631</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Boxuan</given-names>
            <surname>Ma</surname>
          </string-name>
          , Min Lu, and
          <string-name>
            <surname>Shin'ichi Konomi</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Understanding student slide reading patterns during the pandemic</article-title>
          .
          <source>In 18th International Conference on Cognition and Exploratory Learning in Digital Age</source>
          ,
          <string-name>
            <surname>CELDA</surname>
          </string-name>
          <year>2021</year>
          . IADIS Press,
          <fpage>87</fpage>
          -
          <lpage>94</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Boxuan</given-names>
            <surname>Ma</surname>
          </string-name>
          , Li Chen, and
          <string-name>
            <given-names>Min</given-names>
            <surname>Lu</surname>
          </string-name>
          .
          <article-title>"Personalized Navigation Recommendation for E-book Page Jump."</article-title>
          <source>In CEUR Workshop Proceedings</source>
          , vol.
          <volume>3667</volume>
          , pp.
          <fpage>32</fpage>
          -
          <lpage>41</lpage>
          . CEUR-WS,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Shang</given-names>
            <surname>Wang</surname>
          </string-name>
          and
          <string-name>
            <given-names>Erin</given-names>
            <surname>Walker</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Providing adaptive feedback in concept mapping to improve reading comprehension</article-title>
          .
          <source>In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems</source>
          .
          <volume>1</volume>
          -
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Liangming</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Chengjiang</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Juanzi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Jie</given-names>
            <surname>Tang</surname>
          </string-name>
          .
          <article-title>"Prerequisite relation learning for concepts in MOOCs." In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL</article-title>
          ), pp.
          <fpage>1447</fpage>
          -
          <lpage>1456</lpage>
          .
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Liangming</given-names>
            <surname>Pan</surname>
          </string-name>
          , Xiaochen Wang,
          <string-name>
            <given-names>Chengjiang</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Juanzi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Jie</given-names>
            <surname>Tang</surname>
          </string-name>
          .
          <article-title>"Course concept extraction in MOOCs via embedding-based graph propagation."</article-title>
          <source>In Proceedings of the Eighth International Joint Conference on Natural Language Processing (ACL)</source>
          , pp.
          <fpage>875</fpage>
          -
          <lpage>884</lpage>
          .
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Irene</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <surname>Alexander R. Fabbri</surname>
          </string-name>
          ,
          <string-name>
            <surname>Robert R. Tung</surname>
          </string-name>
          , and
          <string-name>
            <surname>Dragomir</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Radev</surname>
          </string-name>
          .
          <article-title>"What should I learn first: Introducing lecturebank for NLP education and prerequisite chain learning."</article-title>
          <source>In Proceedings of the AAAI Conference on Artificial Intelligence</source>
          , vol.
          <volume>33</volume>
          , no.
          <issue>01</issue>
          , pp.
          <fpage>6674</fpage>
          -
          <lpage>6681</lpage>
          .
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>