<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Augmenting Patent Summarization using Large Language Model with Knowledge Graph</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Shoon</forename><forename type="middle">Lei</forename><surname>Phyu</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Tokyo International University</orgName>
								<address>
									<addrLine>4-42-31 Higashi-Ikebukuro</addrLine>
									<postCode>170-0013</postCode>
									<settlement>Toshima, Tokyo</settlement>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Murataly</forename><surname>Uchkempirov</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Tokyo International University</orgName>
								<address>
									<addrLine>4-42-31 Higashi-Ikebukuro</addrLine>
									<postCode>170-0013</postCode>
									<settlement>Toshima, Tokyo</settlement>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mayesha</forename><forename type="middle">Maliha</forename><surname>Proma</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Tokyo International University</orgName>
								<address>
									<addrLine>4-42-31 Higashi-Ikebukuro</addrLine>
									<postCode>170-0013</postCode>
									<settlement>Toshima, Tokyo</settlement>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Parag</forename><surname>Kulkarni</surname></persName>
							<email>pkulkarni@al.tiu.ac.jp</email>
							<affiliation key="aff0">
								<orgName type="institution">Tokyo International University</orgName>
								<address>
									<addrLine>4-42-31 Higashi-Ikebukuro</addrLine>
									<postCode>170-0013</postCode>
									<settlement>Toshima, Tokyo</settlement>
									<country key="JP">Japan</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Augmenting Patent Summarization using Large Language Model with Knowledge Graph</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">58B7489EE99782BC9C1D344450641035</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:28+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Patent Summarization</term>
					<term>Natural Language Processing</term>
					<term>Knowledge Graph</term>
					<term>Knowledge Graph Summarization</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The increasing quantity and intricacy of patent papers need the creation of effective and precise summary techniques. The scientific and comprehensive nature of patents is often excessive for traditional summary techniques and machine learning-based approaches. This paper explores the integration of large language models (LLMs) with knowledge graphs (KGs) to improve patent document summarization. LLMs such as GPT-4 provide advanced language understanding and generation capabilities, but can cause problems with domainspecific content. Knowledge graphs offer a structured representation of knowledge that allows domain-specific information to be included in the summarization process. The goal of this integration is to enhance the readability and informativeness of patent summaries. We then propose a framework for combining LLM with KG and evaluate its performance compared to traditional baseline summarization technique using both ROUGE scores and human evaluations. The results demonstrate a significant improvement in the quality of patent summary, highlighting the potential of this approach to produce more informative and accurate reviews of complex technical documents. This research contributes to the development of more robust artificial intelligence summarization system capable of performing complex data interpretation and decision making in specialized areas.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Patent documents are a rich source of technical and legal information, but their complexity often makes them difficult to understand quickly. Traditional summarization methods struggle with the specificity and technicality of patent language. Recent advancements in large language models (LLMs) and knowledge graphs (KGs) offer new opportunities for improving patent summarization. This paper explores the augmentation of LLM-based patent summarization using knowledge graphs to generate more concise and knowledgeable summaries. The primary focus is on leveraging Ollama3 for entity and relationship extraction to build a patent knowledge graph and subsequently using it to generate summaries. This approach aims to enhance the readability and informativeness of patent summaries compared to traditional methods.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Literature Review</head><p>Automatic patent document summarization is a critical task that has received significant research attention. Several methods have been proposed to solve this problem. For example, <ref type="bibr" target="#b0">[1]</ref> presented a non-ontology-based document summarization method that uses keyword recognition and significant information density to extract paragraphs that best represent the patent text. <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref> presented a new approach called PQPS, a state-of-the-art query-based patent summarizer that uses RBM and Bi-LSTM models for patent summarization. <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref> used surface mining, NER, and machine learning techniques to extract concepts and conceptual relationships from the claims section of patent documents. Moreover, <ref type="bibr" target="#b3">[4]</ref> described a summarization method for extracting surrogate patents to effectively support patent mapping. <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b3">4]</ref> examined the identification of conflicting proposals in patents using the BERT summarization method. Additionally, <ref type="bibr" target="#b5">[6]</ref> discussed clustering patent generalizations into homogeneous groups for knowledge generation. Moreover, <ref type="bibr" target="#b6">[7]</ref> studied feature term extraction from patent documents. <ref type="bibr" target="#b7">[8]</ref> developed a multilingual patent text mining approach to compute relatedness scores of patent documents. <ref type="bibr" target="#b8">[9]</ref> used graph representations to improve abstract summarization of multiple documents, demonstrating the potential of graph-based methods in summarization tasks.</p><p>Despite progress in this area, several shortcomings in patent document summarization methods have been identified, highlighting the need for more sophisticated and efficient methods to address them. The literature has noted the shortcomings of methods for summarizing patent documentation. According to <ref type="bibr" target="#b3">[4]</ref> there is a growing demand for automated tools to support patent engineers and decision makers in patent analysis. Traditional methods face problems such as difficulties in natural language translation and the lack of parallel sets of bilingual documents <ref type="bibr" target="#b9">[10]</ref>. Existing patent summarization strategies often fail due to the unique characteristics of patents and the emphasis on specific sections, such as claims, while ignoring important details from other sections <ref type="bibr" target="#b11">[11]</ref>. In addition, the huge number of patent documents leads to information overload, which requires more effective summarization and analysis methods <ref type="bibr" target="#b12">[12]</ref>. <ref type="bibr" target="#b13">[13]</ref> provided a survey on Graph-based representation for text analysis and found out that the graph shows superiority over traditional methods. The author also suggested the novel approach in construction of semi-graph and sentence extraction, which helps finding linear and non-linear relationship between features <ref type="bibr" target="#b14">[14]</ref>.</p><p>Despite advances in text abstracting, annotations generated for patent documents still do not match human annotations <ref type="bibr" target="#b1">[2]</ref>. In addition, the complexity and length of patents, as well as the requirement for domain knowledge, contribute to their under-utilization <ref type="bibr" target="#b4">[5]</ref>. Moreover, limitations of existing patent analysis tools hinder the creation of impact technology patent maps, highlighting the need for improved methods for efficiently extracting information from patent documents <ref type="bibr" target="#b15">[15]</ref>. Problems associated with cryptic descriptions in patent texts limit the effectiveness and comprehensiveness of patent analysis <ref type="bibr" target="#b16">[16]</ref>.</p><p>Several studies show that integrating LLMs into patent document summarization provides significant benefits in terms of accuracy and efficiency. Large language models (LLMs) have significantly impacted the field of automatic abstracting, especially in the context of patent document abstracting. The use of LLMs in summarizing patent documents has various advantages and disadvantages, as follows:</p><p>• LLMs are excellent at capturing the semantic meaning of text, which allows them to create contextually rich and coherent summaries <ref type="bibr" target="#b17">[17]</ref>, • LLMs are adept at handling the intricacies of patent documents, which often contain technical terminology and detailed information, resulting in more accurate and informative summaries <ref type="bibr" target="#b18">[18]</ref>, • LLMs demonstrate versatility in summarizing different types of documents, such as scientific literature, radiology reports, and patient problem lists <ref type="bibr" target="#b18">[18]</ref>, • LLMs have the ability to process voluminous input data, making them suitable for summarizing extensive patent documents without leaving out important details <ref type="bibr" target="#b17">[17]</ref>, • LLMs can be combined with knowledge graphs to improve generalization by integrating domainspecific knowledge and relationships extracted from the graph At the same time, the readability of summaries created using LLM may vary, making it difficult to ensure that all users, including non-specialists, can easily understand <ref type="bibr">[Decker, 2023]</ref>. According to Decker [2023], graduates may have difficulty ensuring that their resumes are complete, especially when gathering all the relevant details from lengthy and detailed patent documents. Bush <ref type="bibr">[2024]</ref> points out that LLMs can reflect biases present in the training data, potentially affecting the quality and objectivity of the summaries produced. The comprehension of LLM-based summaries can be a concern because understanding the rationale behind a particular summary generated by the model can be challenging <ref type="bibr" target="#b19">[19]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Integration of LLM and KG in patent document summarization</head><p>As a solution to these problems, this paper proposes an integrated approach to summarizing patent documents using LLM with knowledge graphs. Knowledge graphs have become integral to various applications, gaining popularity particularly after the comprehensive representation by <ref type="bibr" target="#b20">[20]</ref>. The concept of "Knowledge Graph" dates back to 1972 with the work of Schneider, E. W. Today, knowledge graphs are widely utilized in implementations such as the Google Knowledge Graph, Airbnb, and other social application analyses. Despite their broad applications, there has been limited research focusing specifically on patent datasets. By combining the power of understanding of the LLM with the structured representation of knowledge in the form of graphs, you can create more accurate, informative, and contextually rich summaries of patent documents. Large language models, such as pre-trained encoders <ref type="bibr" target="#b21">[21]</ref>, have shown promise in text summarization tasks. These models can effectively capture the semantic meaning of text and generate coherent summaries. Knowledge graphs play a critical role in organizing and presenting structured information. Research has shown that incorporating knowledge graphs into language models can improve the representation of domain information <ref type="bibr" target="#b22">[22]</ref>. Research on question-answering systems has demonstrated the effectiveness of combining large language models with knowledge graphs to improve answer selection <ref type="bibr" target="#b23">[23]</ref>. By using semantic relationships encoded in knowledge graphs, the system can better understand and extract relevant information to answer questions. This approach can be extended to patent document summarization to extract key ideas and relationships embedded in patent texts. <ref type="bibr" target="#b24">[24]</ref> proposes the framework for predictive summarization highlighting the possible areas of patent applications in the future by analyzing the past data in the legal domain.</p><p>Moreover, leveraging external knowledge from open source knowledge graphs has been shown to improve the performance of document summarization models <ref type="bibr" target="#b25">[25]</ref>. By incorporating external knowledge into the summarization process, the model gains access to a broader context and can generate more informative and complete summaries. Several studies have explored the integration of knowledge graphs into patent-related tasks, demonstrating their potential benefits. <ref type="bibr" target="#b26">[26]</ref> provided a comprehensive overview of knowledge graph embedding approaches and applications, highlighting the importance of structured knowledge representation. <ref type="bibr" target="#b27">[27]</ref> discusses the role of knowledge graphs in integrating health information and their application in healthcare, highlighting the benefits of incorporating knowledge graphs into data analytics. <ref type="bibr" target="#b28">[28]</ref> implement knowledge proximity in a US patent database using a knowledge graph called "PatNet" built from patent metadata. <ref type="bibr">Chen and Deng [2023]</ref> propose an accountable patent recommendation method using a knowledge graph to demonstrate the practical application of structured knowledge in patent-related problems. <ref type="bibr" target="#b29">[29]</ref> extract knowledge graphs from patent applications, filling the gap in knowledge extraction from patent documents. <ref type="bibr" target="#b30">[30]</ref> use text mining and machine learning techniques to visualize processes in utility chemical patents, demonstrating the utility of knowledge graphs in patent analysis. Additionally, <ref type="bibr" target="#b31">[31]</ref> extend patent searching to include text embedding and knowledge graphs, emphasizing the importance of capturing various aspects of knowledge in patent documents.</p><p>In <ref type="bibr" target="#b2">[3]</ref>, the researchers highlighted that patent summarizations could be significantly enhanced using conceptual knowledge graphs. Similarly, <ref type="bibr" target="#b32">[32]</ref> demonstrated the implementation of knowledge graphs using Open-IE to produce summaries through Cloze Reward. This paper builds on these studies by integrating language models to generate human-readable summaries from patent documents. LAUREN, as detailed by <ref type="bibr" target="#b33">[33]</ref>, focuses on summarizing knowledge graphs specifically for question answering systems. This approach leverages graph-based techniques to resize knowledge graphs, making them more manageable and tailored for answering user queries effectively. The summarization process ensures that the key information necessary for answering questions is retained while extraneous details are minimized. <ref type="bibr" target="#b34">[34]</ref> addresses the summarization of knowledge graphs for search purposes. Their work proposes a method to compute a summarization set that provides maximal descriptive power with minimal redundancy. This approach enhances the efficiency of knowledge graph searches by presenting users with the most relevant information. <ref type="bibr" target="#b35">[35]</ref> discuss the concept of adaptive summarization to create personalized knowledge graphs. This method adapts the summarization process to the user's specific information needs, ensuring that the resulting knowledge graph is both concise and highly relevant. The adaptability of this method makes it particularly useful for applications requiring tailored knowledge delivery. The GATES framework, presented by <ref type="bibr" target="#b36">[36]</ref>, utilizes graph attention networks for entity summarization within knowledge graphs. This technique addresses the challenge of summarizing entities by focusing on the most significant relationships and attributes. The use of attention mechanisms allows for the generation of summaries that are both comprehensive and focused on the most pertinent aspects of the knowledge graph.</p><p>These studies collectively advance the field of knowledge graph summarization by addressing various challenges such as customization, scalability, and efficiency. They provide a solid foundation for further research and development in creating concise, informative, and user-tailored knowledge graph summaries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Building the Patent Knowledge Graph</head><p>The initial step involves constructing the knowledge graph by creating triplets that represent entities and relationships extracted from patent documents. This requires identifying subjects, actions, and objects within the text. Patent documents, characterized by complex sentence structures and nounheavy content, present a challenge for entity extraction. Preliminary attempts using BERT models proved inadequate for the patent dataset.</p><p>To address this, we employ llama-3 pre-trained language model<ref type="foot" target="#foot_0">1</ref> with the parameter of 8.0B, which effectively identifies entities and relationships from the text, specifically from the description and claims sections of patents. The description section provides key terms, while the claims section offers detailed information about the invention, both crucial for generating a comprehensive summary. By excluding other sections, we aim to minimize the hallucination effect of the language model. The extracted data is then used to construct a knowledge graph representing the core elements of the patent. The knowledge graph is created in the Neo4j database using Cypher queries with the extracted relationships and entities on the previous step, ensuring a structured and accessible representation of the patent as shown in Figure <ref type="figure" target="#fig_1">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Generating Summaries from the Knowledge Graph</head><p>With the knowledge graph established, the next step involves generating summaries using the language model (llama-3:8.0B). This involves transforming the structured data from the knowledge graph into a human-readable format that emphasizes conciseness and informativeness. To obtain an effective abstractive summary, it is essential to instruct the language model with a custom prompt. This step ensures the generation of a coherent summary rather than a disjointed question-and-answer format. Additionally, Cypher queries are utilized to retrieve all relevant nodes and relationships from the knowledge graph, facilitating the creation of a comprehensive summary.</p><p>This methodology leverages the strengths of both knowledge graphs and advanced language models, aiming to produce high-quality, informative summaries of patent documents, thereby enhancing the accessibility and understanding of patent information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Baseline Method: Language Model Summarization</head><p>The performance of the proposed method is compared against a baseline summarizing technique that solely utilizes a language model for document summarization. The baseline approach is straightforward: the language model is employed to summarize the content of the patent documents to generate a relevant summarization. We defined this baseline to suit the summarization of one document rather than multiple documents. In this step, we come to notice that we can use multiple document summarization using RAG architecture, which we aims to do the further research. Nevertheless, in this current baseline, the user input is the patent document itself, which is then processed to retrieve relevant sections within the document (such as the description and claims). These sections are fed into the language model, which generates a cohesive summary of the patent.</p><p>Our adaptation involves ensuring that the retrieval mechanism focuses on internal sections of the patent document, effectively treating it as a self-contained database. This approach leverages the structure of the patent document to enhance the quality of the generated summary. This summary is intended to capture the essence of the invention described in the patent document. The simplicity of this approach allows for a clear comparison with more complex methods.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Algorithm 1 Summarizing Patent Documents Using Knowledge RAG and Llama3</head><p>Input: HUPD dataset, Neo4j credentials, patent number.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Output:</head><p>Generation of the structured summary of the patent document. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Dataset</head><p>We utilized the HUPD (Harvard University Patent Dataset), which provides an extensive range of raw patent claim texts. This dataset was chosen over other prominent datasets such as BIGPATENT<ref type="foot" target="#foot_1">2</ref> , CLEF-IP 2011, and USPTO-2M for several key reasons. Firstly, HUPD houses approximately 4.5 million patent documents, significantly more than BIGPATENT's 1.3 million <ref type="bibr" target="#b37">[37]</ref>. This expansive volume offers a richer and more varied base for analysis. Secondly, HUPD encompasses 34 distinct sections, including claims, offering a more comprehensive view compared to BIGPATENT's limitation to just 4 sections. Lastly, unlike the pre-tokenized texts in BIGPATENT, HUPD's provision of raw texts affords greater flexibility and depth in our analysis, allowing for a more tailored approach to data processing <ref type="bibr" target="#b38">[38]</ref>.</p><p>To test our proposed method, we randomly selected approximately 10 documents from the 'G' category of the CPC code (Physics) within the HUPD dataset as representative samples for our study. We generated both baseline summaries and knowledge graph summaries for each patent document and used ROUGE scores to track the performance metrics. Additionally, we employed human evaluation to assess the real-world clarity of the generated summaries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Evaluation</head><p>This section presents the results of our proposed method compared to the baseline summarization technique.The proposed method was evaluated using ROUGE scores to measure the quality of the generated summaries. The evaluation metrics included ROUGE-1, ROUGE-2, and ROUGE-L, which assess the overlap of n-grams and longest common subsequences between the generated and reference summaries. Moreover, we evaluated the performance with human evaluations to assess the quality and understandability of the summaries generated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Quantitative Result</head><p>The ROUGE (Recall-Oriented Understudy for Gisting Evaluation) metric, introduced by Chin-Yew Lin in their pivotal work, is a cornerstone in the field of natural language processing, particularly for evaluating automatic summarization and machine translation <ref type="bibr" target="#b39">[39]</ref>. ROUGE assesses the quality of a summary by comparing it with reference summaries.</p><p>ROUGE scores are often reported in terms of precision, recall, and F1 score, providing a comprehensive evaluation framework:</p><p>• Precision measures the fraction of n-grams in the generated summary that are present in the reference summary. • Recall assesses the fraction of n-grams in the reference summary that appear in the generated summary. • F1 Score is the harmonic mean of precision and recall, offering a balanced metric.</p><p>The formulas for these metrics are as follows: Precision = Number of overlapping n-grams Total n-grams in the generated summary</p><p>Recall = Number of overlapping n-grams Total n-grams in the reference summary ( <ref type="formula">2</ref>)</p><formula xml:id="formula_1">𝐹 1 = 2 × Precision × Recall Precision + Recall<label>(3)</label></formula><p>Table <ref type="table" target="#tab_1">1</ref> shows the ROUGE F1 scores for both the baseline and the proposed method. The baseline method outperforms the proposed method across all ROUGE metrics, indicating a higher F1 score for summarization. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Human Evaluation Result</head><p>The human evaluation of our summarization method was a critical component of our research. We aimed to determine the effectiveness of our Knowledge Graph (KG) based summaries in terms of understandability and informativeness. To this end, we conducted an evaluation involving 12 respondents who reviewed summaries of 10 different patent documents. We selected 10 patent documents from the Harvard University Patent dataset <ref type="bibr" target="#b38">[38]</ref> to ensure a diverse and representative sample. For each patent, three types of summaries were generated: the original summary, a baseline summary, and the KG-based summary. Each respondent was presented with these three summaries for each patent document without being informed of their origins to prevent any bias.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.1.">Voting Process</head><p>Respondents were asked to vote for the summary they found easiest to understand and most informative. The summaries were presented in a randomized order to eliminate order effects. The voting was conducted using a simple voting system where respondents chose one summary out of the three options. The criteria for voting were clearly defined for understandability: how easily the summary could be comprehended, and, informativeness: how well the summary conveyed the essential information of the patent document.</p><p>To ensure the integrity of the evaluation, respondents were chosen from a pool of individuals with no prior experience in reading or writing patent documents. This choice was made to simulate the perspective of an average reader who might not have specialized knowledge. The human-evaluation results are summarized in Table <ref type="table" target="#tab_2">2</ref>. The KG-based summaries received the highest number of votes in most cases, indicating a clear preference for their comprehensibility and informativeness over the other two types of summaries.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Discussion of Results</head><p>The results from the ROUGE F1 score evaluations indicate that the baseline method outperforms the proposed Knowledge Graph (KG)-based summarization technique in terms of automated metrics. ROUGE scores measure the overlap between the generated summaries and reference summaries, focusing on precision, recall, and F1 score.</p><p>However, the positive feedback from human evaluations suggests that the proposed KG-based method generates clearer, more coherent, and more informative summaries. This contrast highlights the importance of human evaluations in assessing summarization quality, as they capture aspects of readability and informativeness that automated metrics might miss.</p><p>Notably, for Patent Number "US14990535, " the baseline method failed to produce a summary, likely due to hallucination effects commonly observed in large language models (LLMs). In contrast, the KG-based method successfully generated a precise and informative summary from the original patent document, demonstrating its robustness.</p><p>It is important to note that our primary focus is on understandability and informativeness as perceived by human readers, rather than solely relying on automated metrics like ROUGE scores. This approach ensures that the summaries effectively convey the essence of the inventions described in the patents.</p><p>The integration of knowledge graphs and the adaptation of the RAG architecture have proven effective in enhancing the readability and informativeness of patent document summarization. These findings underscore the value of human-centered evaluations in developing summarization tools.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Limitations</head><p>Despite the promising results, our research faced several challenges and limitations. The primary limitation stems from the reliance on large language models (LLMs) LLaMA3 for both entity recognition and summary generation introduces potential vulnerabilities. If the LLM fails to accurately produce entities, the subsequent knowledge graph cannot be generated correctly. Any errors in the entity recognition process can lead to inaccuracies in the final summary. This dependency on LLMs introduces a potential source of error that can affect the overall performance of the summarization method.</p><p>Additionally, the LLM used for generating summaries, specifically LLaMA3, may produce sub-optimal results if the prompts are not clearly defined or easy to understand. This can result in summaries that do not accurately reflect the content of the original document. Another limitation is the language support; currently, our method is only capable of handling English documents, which restricts its applicability to patents written in other languages.</p><p>Moreover, our approach is designed for single-document summarization. This limits its effectiveness when dealing with multiple documents, which is a common scenario in patent analysis and other legal contexts. Addressing this limitation is a key area for future research.</p><p>Nevertheless, the human evaluation results suggest that despite the lower ROUGE scores, the summaries generated by the proposed method are more understandable and informative to readers unfamiliar with patent documents. This highlights the importance of human-centered evaluation metrics in assessing the effectiveness of summarization methods, as automated metrics like ROUGE may not fully capture the nuances of human understanding.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Future Work</head><p>To address the limitations identified, future work will focus on several key areas of improvement:</p><p>Enhanced Entity Recognition: We plan to develop a dedicated entity recognition model trained on annotated patent data. By creating a custom model tailored to the specificities of patent documents, we aim to improve the accuracy of entity recognition and, consequently, the quality of the knowledge graphs and summaries.</p><p>Improved LLM Training: We will train the LLM, such as LLaMA3, on a specialized legal database to enhance its performance in summarization tasks. This targeted training will help mitigate issues arising from unclear prompts and improve the model's ability to generate accurate and informative summaries.</p><p>Multilingual Support: Expanding our model to support multiple languages is a priority. This involves training the entity recognition and summarization models on multilingual datasets to ensure accurate performance across different languages.</p><p>Multi-Document Summarization: Developing our method to handle multi-document summarization will significantly increase its applicability. We plan to extend the architecture to aggregate and summarize information from multiple documents, which is particularly important in the legal and patent analysis fields.</p><p>By addressing these areas, we aim to refine our summarization approach, making it more robust, versatile, and applicable to a wider range of documents and languages. This will enhance its utility in practical, real-world scenarios, particularly in legal and technical domains.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Conclusion</head><p>In this paper, we introduced a novel summarization approach that integrates knowledge graphs and large language models (LLMs) to enhance the readability and informativeness of patent document summaries. Our main contribution lies in generating summaries that are easier to understand while maintaining the key inventions from the patent documents. Additionally, our research contributes towards advancing the legal technology by analyzing entities in the patent domain and producing structured data through knowledge graphs combined with the contextual understanding of LLMs for automatic summarization tasks. We emphasize the importance of human-centered evaluations and contextual understandability over automated evaluations when producing summaries. Despite the limitations, our research lays a foundation for future advancements in patent summarization technology, aiming to develop a more robust tool for analyzing complex technical documents in legal domain.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The proposed system architecture for producing KG Summary</figDesc><graphic coords="4,72.00,65.61,451.26,253.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Knowledge Graph of Patent Number US14996316 after constructing in the Neo4J Database</figDesc><graphic coords="5,93.32,65.60,406.12,321.59" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>ROUGE F1 Scores Comparison of Patent Number US14996316</figDesc><table><row><cell cols="4">Method ROUGE-1 ROUGE-2 ROUGE-L</cell></row><row><cell>Baseline</cell><cell>0.461538</cell><cell>0.235294</cell><cell>0.461538</cell></row><row><cell>KG Sum</cell><cell>0.343434</cell><cell>0.136054</cell><cell>0.323232</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>The voting result of Human Evaluation</figDesc><table><row><cell>Patent</cell><cell cols="3">KGSUM Original BaselineSUM</cell><cell>Result</cell></row><row><cell>US14997116</cell><cell>9</cell><cell>0</cell><cell>3</cell><cell>KGSUM</cell></row><row><cell>US14986817</cell><cell>5</cell><cell>0</cell><cell>7</cell><cell>BaselineSUM</cell></row><row><cell>US14990669</cell><cell>8</cell><cell>1</cell><cell>0</cell><cell>KGSUM</cell></row><row><cell>US14998490</cell><cell>2</cell><cell>1</cell><cell>4</cell><cell>BaselineSUM</cell></row><row><cell>US14990535</cell><cell>11</cell><cell>1</cell><cell>0</cell><cell>KGSUM</cell></row><row><cell>US14989021</cell><cell>7</cell><cell>2</cell><cell>3</cell><cell>KGSUM</cell></row><row><cell>US15000497</cell><cell>5</cell><cell>1</cell><cell>6</cell><cell>BaselineSUM</cell></row><row><cell>US14994324</cell><cell>6</cell><cell>2</cell><cell>4</cell><cell>KGSUM</cell></row><row><cell>US15001450</cell><cell>6</cell><cell>3</cell><cell>3</cell><cell>KGSUM</cell></row><row><cell>US14996316</cell><cell>8</cell><cell>2</cell><cell>2</cell><cell>KGSUM</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">pre-trained llama3 model is available to download online at: https://ollama.com/library/llama3</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">BIGPATENT dataset is available to download online at: https://evasharma.github.io/bigpatent/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>First and foremost, we would like to express our gratitude to the creators of the Harvard University Patent Dataset (HUPD) for providing the legal dataset as publicly available, which was instrumental in our study. We also appreciate the developers of the Llama 3 model, which we utilized for testing the knowledge graph summarization on the legal dataset.</p><p>Lastly, a big token of appreciations to the testimonies who participated in the human evaluation of the summaries. Their insights and feedback were invaluable in assessing the quality and readability of our summarization methods.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Summaries for Evaluation</head><p>In this appendix section, the summaries titled "Original Summary, KG Summary (our approach), and Baseline Summary" are presented. These summaries are evaluated against each other to the testimonies to evaluate which summary is easier to understand by voting. Below are the summaries evaluated for Patent No: US14997116.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Original Summary</head><p>Systems, methods, and non-transitory computer readable media are configured to receive a resume corpus. A machine learning model is trained based on terms from the resume corpus. A job title for a user is determined based on profile information provided to the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>KG Summary</head><p>A machine learning model was trained to analyze profile information from job applicants, which was provided to the model via a computer system. The computer system received resumes as input data and used this information to train the model. This training process enables the model to learn patterns and correlations between resume data and job title. The organization recruited for by a specific job title, leveraging the trained machine learning model to improve its hiring processes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Baseline Summary</head><p>The invention relates to a system and method for determining a job title for a user based on their profile information. A machine learning model is trained using terms from a resume corpus, and then the user's profile information is provided to the model to determine their most likely job title. The model uses techniques such as vector space creation to represent terms in the resume corpus, and anchor points are used to identify the closest match for each term.</p><p>We have presented the other patent summaries, that we used in evaluation and generated from our approach, in the online resource. As our approach uses prompts in order to generate the nodes and entities from the original documents, we have fully provided the prompts we used as well as the feature processing functions before we proceed to the graph construction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Online Resources</head><p>The code for this paper is available at :</p><p>• GitHub</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Automatic patent document summarization for collaborative knowledge systems and services</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Trappey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">V</forename><surname>Trappey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1007/s11518-009-5100-7</idno>
		<ptr target="https://doi.org/10.1007/s11518-009-5100-7.doi:10.1007/s11518-009-5100-7" />
	</analytic>
	<monogr>
		<title level="j">Journal of Systems Science and Systems Engineering</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="71" to="94" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Pqps: Prior-art query-based patent summarizer using rbm and bi-lstm</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kumaravel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sankaranarayanan</surname></persName>
		</author>
		<idno type="DOI">10.1155/2021/2497770</idno>
		<ptr target="https://doi.org/10.1155/2021/2497770.doi:10.1155/2021/2497770" />
	</analytic>
	<monogr>
		<title level="j">Mobile Information Systems</title>
		<imprint>
			<biblScope unit="page">2497770</biblScope>
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Patent document summarization using conceptual graphs</title>
		<author>
			<persName><forename type="first">P</forename><surname>Rk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lalithadevi</surname></persName>
		</author>
		<idno type="DOI">10.2139/ssrn.3441645</idno>
	</analytic>
	<monogr>
		<title level="j">SSRN Electronic Journal</title>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Patent surrogate extraction and evaluation in the context of patent mapping</title>
		<author>
			<persName><forename type="first">Y.-H</forename><surname>Tseng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-I</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D.-R</forename><surname>Juang</surname></persName>
		</author>
		<idno type="DOI">10.1177/0165551507077406</idno>
		<ptr target="https://doi.org/10.1177/0165551507077406.doi:10.1177/0165551507077406" />
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="718" to="736" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Summatriz : Summarization networks for mining patent contradiction</title>
		<author>
			<persName><forename type="first">G</forename><surname>Guarino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Samet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nafi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cavallucci</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICMLA51294.2020.00159</idno>
		<ptr target="https://doi.org/10.1109/ICMLA51294.2020.00159.doi:10.1109/ICMLA51294.2020.00159" />
	</analytic>
	<monogr>
		<title level="m">19th IEEE International Conference on Machine Learning and Applications (ICMLA)</title>
				<imprint>
			<date type="published" when="2020">2020. 2020. 2020</date>
			<biblScope unit="page" from="979" to="986" />
		</imprint>
	</monogr>
	<note>19th IEEE International Conference on Machine Learning and Applications (ICMLA)</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">An r&amp;d knowledge management method for patent document summarization</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Trappey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">V</forename><surname>Trappey</surname></persName>
		</author>
		<idno type="DOI">10.1108/02635570810847608</idno>
		<ptr target="https://doi.org/10.1108/02635570810847608.doi:10.1108/02635570810847608" />
	</analytic>
	<monogr>
		<title level="j">Industrial Management &amp; Data Systems</title>
		<imprint>
			<biblScope unit="volume">108</biblScope>
			<biblScope unit="page" from="245" to="257" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Extracting characteristic terms from patent documents</title>
		<author>
			<persName><forename type="first">K</forename><surname>Takano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tanaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sakai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kitajima</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ota</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tanabe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sakaji</surname></persName>
		</author>
		<idno type="DOI">10.52731/ijscai.v4.i2.568</idno>
		<ptr target="https://doi.org/10.52731/ijscai.v4.i2.568.doi:10.52731/ijscai.v4.i2.568" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Smart Computing and Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page" from="19" to="38" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A multilingual patent text-mining approach for computing relatedness evaluation of patent documents</title>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-C</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-J</forename><surname>Li</surname></persName>
		</author>
		<idno type="DOI">10.1109/IIH-MSP.2009.162</idno>
	</analytic>
	<monogr>
		<title level="m">Fifth International Conference on Intelligent Information Hiding and Multimedia Signal Processing</title>
				<imprint>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="612" to="615" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Leveraging graph to improve abstractive multidocument summarization</title>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.555</idno>
		<ptr target="https://aclanthology.org/2020.acl-main.555.doi:10.18653/v1/2020.acl-main.555" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Chai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tetreault</surname></persName>
		</editor>
		<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="6232" to="6243" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Cross-language patent matching via an international patent classification-based concept bridge</title>
		<author>
			<persName><forename type="first">Y.-L</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-T</forename><surname>Chiu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information Science</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="page">737</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title/>
		<idno type="DOI">10.1177/0165551513494641</idno>
		<idno>arXiv:</idno>
		<ptr target="https://doi.org/10.1177/0165551513494641" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Using genre-specific features for patent summaries</title>
		<author>
			<persName><forename type="first">J</forename><surname>Codina-Filbà</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bouayad-Agha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Burga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Casamayor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mille</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Müller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Saggion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wanner</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ipm.2016.07.002</idno>
		<ptr target="https://doi.org/10.1016/j.ipm.2016.07.002" />
	</analytic>
	<monogr>
		<title level="j">Information Processing and Management</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="151" to="174" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Constructing patent maps using text mining to sustainably detect potential technological opportunities</title>
		<author>
			<persName><forename type="first">H.-F</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-H</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-Y</forename><surname>Hsin</surname></persName>
		</author>
		<idno type="DOI">10.3390/su10103729</idno>
		<ptr target="https://doi.org/10.3390/su10103729.doi:10.3390/su10103729" />
	</analytic>
	<monogr>
		<title level="j">Sustainability</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">3729</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Graph based representation and analysis of text document: A survey of techniques</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sonawane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kulkarni</surname></persName>
		</author>
		<idno type="DOI">10.5120/16899-6972</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Applications</title>
		<imprint>
			<biblScope unit="volume">96</biblScope>
			<biblScope unit="page" from="1" to="8" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Extractive summarization using semigraph (essg)</title>
		<author>
			<persName><forename type="first">S</forename><surname>Sonawane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Kulkarni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Deshpande</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Athawale</surname></persName>
		</author>
		<idno type="DOI">10.1007/s12530-018-9246-8</idno>
		<ptr target="https://doi.org/10.1007/s12530-018-9246-8.doi:10.1007/s12530-018-9246-8" />
	</analytic>
	<monogr>
		<title level="j">Evolving Systems</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="409" to="424" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Extraction of the effect and the technology terms from a patent document</title>
		<author>
			<persName><forename type="first">H</forename><surname>Nonaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kobayashi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sakaji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Suzuki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sakai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Masuyama</surname></persName>
		</author>
		<idno type="DOI">10.1109/iccie.2010.5668167</idno>
		<ptr target="https://doi.org/10.1109/iccie.2010.5668167.doi:10.1109/iccie.2010.5668167" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2010 International Conference on Computational Intelligence and Education (ICCIE)</title>
				<meeting>the 2010 International Conference on Computational Intelligence and Education (ICCIE)</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Analyzing patent topical information to identify technology pathways and potential opportunities</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Porter</surname></persName>
		</author>
		<idno type="DOI">10.1007/s11192-014-1392-6</idno>
		<ptr target="https://doi.org/10.1007/s11192-014-1392-6.doi:10.1007/s11192-014-1392-6" />
	</analytic>
	<monogr>
		<title level="j">Scientometrics</title>
		<imprint>
			<biblScope unit="volume">102</biblScope>
			<biblScope unit="page" from="811" to="827" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Chain-of-event prompting for multi-document summarization by large language models</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bao</surname></persName>
		</author>
		<idno type="DOI">10.1108/ijwis-12-2023-0249</idno>
		<ptr target="https://doi.org/10.1108/ijwis-12-2023-0249.doi:10.1108/ijwis-12-2023-0249" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Web Information Systems</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="229" to="247" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Williams</surname></persName>
		</author>
		<idno type="DOI">10.1101/2024.04.03.24305088</idno>
		<idno>doi:10.1101/2024.04.03. 24305088</idno>
		<ptr target="https://doi.org/10.1101/2024.04.03.24305088" />
		<title level="m">Evaluating large language models for drafting emergency department discharge summaries</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">O</forename><surname>Neil</surname></persName>
		</author>
		<idno type="DOI">10.1101/2024.01.31.578275</idno>
		<idno>doi:10.1101/2024.01.31.578275</idno>
		<ptr target="https://doi.org/10.1101/2024.01.31.578275" />
		<title level="m">Phenomics assistant: an interface for llm-based biomedical knowledge graph exploration</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Knowledge graphs</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hogan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Blomqvist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cochez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">D</forename><surname>Melo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gutierrez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kirrane</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E L</forename><surname>Gayo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Navigli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Neumaier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-C</forename><forename type="middle">N</forename><surname>Ngomo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Polleres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Rashid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Schmelzeisen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sequeda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Staab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zimmermann</surname></persName>
		</author>
		<idno type="DOI">10.1145/3447772</idno>
		<ptr target="http://dx.doi.org/10.1145/3447772.doi:10.1145/3447772" />
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="1" to="37" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Text summarization with pretrained encoders</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lapata</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1387</idno>
		<ptr target="https://aclanthology.org/D19-1387.doi:10.18653/v1/D19-1387" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Inui</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Ng</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">X</forename><surname>Wan</surname></persName>
		</editor>
		<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3730" to="3740" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">K-bert: enabling language representation with knowledge graph</title>
		<author>
			<persName><forename type="first">W</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1609/aaai.v34i03.5681</idno>
		<ptr target="https://doi.org/10.1609/aaai.v34i03.5681.doi:10.1609/aaai.v34i03.5681" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="2901" to="2908" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">An empirical study of pre-trained language models in simple knowledge graph question answering</title>
		<author>
			<persName><forename type="first">N</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Min</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zafar</surname></persName>
		</author>
		<idno type="DOI">10.21203/rs.3.rs-2184834/v1</idno>
		<ptr target="https://doi.org/10.21203/rs.3.rs-2184834/v1.doi:10.21203/rs.3.rs-2184834/v1" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">A survey of automatic text summarization systems, applications, challenges, and future trends</title>
		<author>
			<persName><forename type="first">M</forename><surname>Uchkempirov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shoon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Lei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Aadhya</surname></persName>
		</author>
		<author>
			<persName><surname>Kulkarni</surname></persName>
		</author>
		<idno type="DOI">10.5120/ijca2024923672</idno>
		<ptr target="https://ijcaonline.org/archives/volume186/number23/a-survey-of-automatic-text-summarization-systems-applications-challenges-and-future-trends/.doi:10.5120/ijca2024923672" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Computer Applications</title>
		<imprint>
			<biblScope unit="volume">186</biblScope>
			<biblScope unit="page" from="8" to="15" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Incorporating external knowledge into unsupervised graph model for document summarization</title>
		<author>
			<persName><forename type="first">T</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.3390/electronics9091520</idno>
		<ptr target="https://www.mdpi.com/2079-9292/9/9/1520.doi:10.3390/electronics9091520" />
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Knowledge graph embedding: a survey of approaches and applications</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
		<idno type="DOI">10.1109/tkde.2017.2754499</idno>
		<ptr target="https://doi.org/10.1109/tkde.2017.2754499.doi:10.1109/tkde.2017.2754499" />
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="2724" to="2743" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Knowledge graph: a survey</title>
		<author>
			<persName><forename type="first">A</forename><surname>Akinnubi</surname></persName>
		</author>
		<idno type="DOI">10.31219/osf.io/8xm2g</idno>
		<ptr target="https://doi.org/10.31219/osf.io/8xm2g.doi:10.31219/osf.io/8xm2g" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Embedding knowledge graph of patent metadata to measure knowledge proximity</title>
		<author>
			<persName><forename type="first">G</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Siddharth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Luo</surname></persName>
		</author>
		<idno type="DOI">10.1002/asi.24736</idno>
		<ptr target="https://doi.org/10.1002/asi.24736.doi:10.1002/asi.24736" />
	</analytic>
	<monogr>
		<title level="j">Journal of the Association for Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">74</biblScope>
			<biblScope unit="page" from="476" to="490" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Engineering knowledge graph from patent database</title>
		<author>
			<persName><forename type="first">L</forename><surname>Siddharth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Blessing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wood</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Luo</surname></persName>
		</author>
		<idno type="DOI">10.1115/1.4052293</idno>
		<idno>doi:</idno>
		<ptr target="10.1115/1.4052293" />
	</analytic>
	<monogr>
		<title level="j">Journal of Computing and Information Science in Engineering</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Ip analytics and machine learning applied to create process visualization graphs for chemical utility patents</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J C</forename><surname>Trappey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">V</forename><surname>Trappey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-W</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-S</forename><surname>Lin</surname></persName>
		</author>
		<idno type="DOI">10.3390/pr9081342</idno>
		<ptr target="https://doi.org/10.3390/pr9081342.doi:10.3390/pr9081342" />
	</analytic>
	<monogr>
		<title level="j">Processes</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page">1342</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Leveraging online behaviors for interpretable knowledgeaware patent recommendation</title>
		<author>
			<persName><forename type="first">W</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ma</surname></persName>
		</author>
		<idno type="DOI">10.1108/intr-08-2020-0473</idno>
		<ptr target="https://doi.org/10.1108/intr-08-2020-0473.doi:10.1108/intr-08-2020-0473" />
	</analytic>
	<monogr>
		<title level="j">Internet Research</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="568" to="587" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Knowledge graph-augmented abstractive summarization with semanticdriven cloze reward</title>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.457</idno>
		<ptr target="https://aclanthology.org/2020.acl-main.457.doi:10.18653/v1/2020.acl-main.457" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Chai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tetreault</surname></persName>
		</editor>
		<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5094" to="5107" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Lauren -knowledge graph summarization for question answering</title>
		<author>
			<persName><forename type="first">R</forename><surname>Jalota</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Vollmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moussallem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-C</forename><forename type="middle">N</forename><surname>Ngomo</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICSC50631.2021.00047</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE 15th International Conference on Semantic Computing (ICSC)</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="221" to="226" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Mining summaries for knowledge graph search</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">X</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.1109/TKDE.2018.2807442</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="1887" to="1900" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m" type="main">Adaptive personalized knowledge graph summarization</title>
		<author>
			<persName><forename type="first">T</forename><surname>De</surname></persName>
		</author>
		<author>
			<persName><surname>Safavi</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:52510379" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Gates: Using graph attention networks for entity summarization</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Firmansyah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moussallem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-C</forename><forename type="middle">N</forename><surname>Ngomo</surname></persName>
		</author>
		<idno type="DOI">10.1145/3460210.3493574</idno>
		<idno>doi:10.1145/3460210.3493574</idno>
		<ptr target="https://doi.org/10.1145/3460210.3493574" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 11th Knowledge Capture Conference, K-CAP &apos;21</title>
				<meeting>the 11th Knowledge Capture Conference, K-CAP &apos;21<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="73" to="80" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Bigpatent: A large-scale dataset for abstractive and coherent summarization</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Hu</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P19-1213</idno>
		<ptr target="https://doi.org/10.18653/v1/P19-1213.doi:10.18653/v1/P19-1213" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="2204" to="2213" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Suzgun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Melas-Kyriazi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Sarkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">D</forename><surname>Kominers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Shieber</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2207.04043" />
		<title level="m">The harvard uspto patent dataset: A large-scale, well-structured, and multi-purpose corpus of patent applications</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">ROUGE: A package for automatic evaluation of summaries</title>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W04-1013" />
	</analytic>
	<monogr>
		<title level="m">Text Summarization Branches Out, Association for Computational Linguistics</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="74" to="81" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
