<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Classifying Scientific Topic Relationships with SciBERT</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alessia</forename><surname>Pisu</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics and Computer Science</orgName>
								<orgName type="institution">University of Cagliari</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Livio</forename><surname>Pompianu</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics and Computer Science</orgName>
								<orgName type="institution">University of Cagliari</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Angelo</forename><surname>Salatino</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Knowledge Media Institute</orgName>
								<orgName type="institution">The Open University</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Francesco</forename><surname>Osborne</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Knowledge Media Institute</orgName>
								<orgName type="institution">The Open University</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">Department of Business and Law</orgName>
								<orgName type="institution">University of Milano Bicocca</orgName>
								<address>
									<country>IT</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Daniele</forename><surname>Riboni</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics and Computer Science</orgName>
								<orgName type="institution">University of Cagliari</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Enrico</forename><surname>Motta</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Knowledge Media Institute</orgName>
								<orgName type="institution">The Open University</orgName>
								<address>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Diego</forename><forename type="middle">Reforgiato</forename><surname>Recupero</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Mathematics and Computer Science</orgName>
								<orgName type="institution">University of Cagliari</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Classifying Scientific Topic Relationships with SciBERT</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2DAAE5A048E960F5E1FE778B322C3763</idno>
					<idno type="arXiv">arXiv:1903.10676.</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:46+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Research Topics</term>
					<term>Ontology Generation</term>
					<term>Language Models</term>
					<term>Knowledge Graph Generation</term>
					<term>SciBERT</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Current AI systems, including smart search engines and recommendation systems tools for streamlining literature reviews, and interactive question-answering platforms, are becoming indispensable for researchers to navigate and understand the vast landscape of scientific knowledge. Taxonomies and ontologies of research topics are key to this process, but manually creating them is costly and often leads to outdated results. This poster paper shows the use of SciBERT model to automatically generate research topic ontologies. Our model excels at identifying semantic relationships between research topics, outperforming traditional methods. This approach promises to streamline the creation of accurate and up-to-date ontologies, enhancing the effectiveness of AI tools for researchers.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The current generation of AI technologies, such as smart search engines, recommendation systems, and question-answering applications, significantly aids researchers in exploring and interpreting scientific literature <ref type="bibr" target="#b0">[1]</ref>. Despite this, the rapid growth of scientific publications, increasing by about 2.5 million papers annually <ref type="bibr" target="#b1">[2]</ref>, poses a substantial challenge. Although large language models have revolutionised natural language processing (NLP) <ref type="bibr" target="#b2">[3]</ref>, they still encounter limitations to process extensive text volumes and understand the broader context of a research area.</p><p>To address this, scientific knowledge graphs (SKGs) <ref type="bibr" target="#b3">[4]</ref>, such as SemOpenAlex 1 , AIDA-KG 2 , ORKG 3 , CS-KG 4 , became increasingly popular, providing structured and formal representations of research publications.</p><p>Research topics are essential for describing research concepts within SKGs, making ontologies of research topics (e.g., MeSH, UMLS, CSO, NLM) crucial for organising and querying academic information <ref type="bibr" target="#b4">[5]</ref>. Altogether, they empower intelligent systems to efficiently navigate and understand academic literature, including advanced search engines, interactive conversational agents, analytics dashboards, and academic recommender systems.</p><p>However, manually creating ontologies of research topics is costly and time-consuming, often resulting in outdated representations. To address this challenge, several approaches have been proposed, including the integration of ontology learning with crowdsourcing methods, combining statistical analysis with user feedback <ref type="bibr" target="#b5">[6]</ref>, or utilising citation-based clustering of research papers to infer research topics from the titles and abstracts of documents within clusters <ref type="bibr" target="#b6">[7]</ref>. Another approach is Klink-2 <ref type="bibr" target="#b7">[8]</ref>, which produced the Computer Science Ontology (CSO) <ref type="bibr" target="#b8">[9]</ref>, a widely adopted resource with about 14K topics and 159K semantic relationships.</p><p>In the same direction, this poster paper explores the use of SciBERT for generating research topic ontologies. Our goal is to develop a method that incorporates language model technology to update CSO and construct large-scale ontologies across scientific disciplines. We developed a model to identify four semantic relationships (supertopic, subtopic, same-as, and other) between research topics and compared its performance to traditional feature-based solutions. Preliminary results show that the transformer-based model significantly outperforms traditional models. The gold standard and code are available on a GitHub repository<ref type="foot" target="#foot_0">5</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Materials and Methods</head><p>In this section, at first we describe the addressed task and the used datasets. Then, we illustrate a traditional feature-based approach, and our transformer-based technique.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Task Definition and Datasets</head><p>In this work, we address a single-label multi-class classification problem. The task is to classify the relationship between a pair of research topics (𝑡 𝐴 , 𝑡 𝐵 ) according to four categories which are essential for ontology generation:</p><p>• supertopic: 𝑡 𝐴 is a parent topic of 𝑡 𝐵 . E.g., ontological languages is a broader area than owl • subtopic: 𝑡 𝐴 is a child topic of 𝑡 𝐵 . E.g., nosql is a specific area within databases • same-as: 𝑡 𝐴 and 𝑡 𝐵 are different labels for the same concept. E.g., haptic interface and haptic device • other: 𝑡 𝐴 and 𝑡 𝐵 do not relate according to the above categories. E.g., blockchain and user interfaces</p><p>In this context, other can refer to either negative samples or alternative semantic relationships not currently considered by our method, such as partOf, or contributesTo.</p><p>For our gold standard, we selected portions of the Computer Science Ontology <ref type="bibr" target="#b8">[9]</ref> that have been manually checked and improved. CSO is a large ontology covering 14K research topics, providing an extensive and fine-grained representation of Computer Science. It was automatically generated using the Klink-2 algorithm <ref type="bibr" target="#b7">[8]</ref> on 16 million scientific articles.</p><p>CSO comprises four primary semantic relationships. Among them, superTopicOf and related-Equivalent essentially correspond to our superTopic and same-as relationships, respectively. To construct the gold standard, we selected 4,713 superTopicOf triples from the CSO and designated them as superTopic instances. Additionally, we chose 3,034 relatedEquivalent triples to represent equivalence using the same-as relation. We also derived 4,713 subTopic relationships by reversing the superTopic relationships. Lastly, we randomly paired topics to create 5,151 other relationships, ensuring that none of these pairs shared any of the previously identified relationships within the CSO. The resulting gold standard dataset consists of 17,611 triples, divided into 15,154 triples (86%) for the training set, 2,166 triples (12.3%) for the validation set, and 291 triples (1.7%) for the test set. To prevent bias, we ensured that topic pairs in one set do not appear in another. Moreover, each test set triple includes at least one topic not present in the training set. These measures make the test set more challenging compared to those used for Klink-2 <ref type="bibr" target="#b7">[8]</ref>. In order to compute features involving the linkage of topics to relevant papers used in our feature-based method, we queried AIDA-KG <ref type="bibr" target="#b9">[10]</ref>, a KG considering 25 million publications linked to research topics in CSO.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Feature-based Method</head><p>Our classification task is commonly approached exploiting numerical features, usually measuring the frequency and common usage of the two topics <ref type="bibr" target="#b7">[8]</ref>. Extracted feature vectors are then classified through mathematical functions or machine learning algorithms <ref type="bibr" target="#b7">[8]</ref>. We devised a feature-based classification method using the following features for each pair of topics (𝑡 𝐴 , 𝑡 𝐵 ):</p><p>• occA: the frequency of 𝑡 𝐴 appearing in paper abstracts • occB: the frequency of 𝑡 𝐵 appearing in paper abstracts • cooccurrenceAB: the frequency of both 𝑡 𝐴 and 𝑡 𝐵 appearing together in abstracts • subsumption: the degree of overlap between the co-occurring topics, computed as subsumption = 𝑐𝑜𝑜𝑐𝑐𝑢𝑟𝑟𝑒𝑛𝑐𝑒𝐴𝐵 𝑜𝑐𝑐𝐴 − 𝑐𝑜𝑜𝑐𝑐𝑢𝑟𝑟𝑒𝑛𝑐𝑒𝐴𝐵</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝑜𝑐𝑐𝐵</head><p>The first two features indicate the popularity of a topic. The third feature quantifies the relatedness of two topics. The fourth feature assesses the hierarchical relationship between the topics. After normalising the features, we trained two ensemble machine learning models: Gradient Boosting (GB) and Random Forest (RF); varying the number of estimators from 10 to 3000 to determine the optimal configuration.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Language Model-based Method</head><p>Our method leveraging language models relies on SciBERT <ref type="bibr" target="#b10">[11]</ref>, an extension of BERT <ref type="bibr" target="#b11">[12]</ref>, which is a highly regarded model for its ability to effectively understand and process human language. SciBERT, trained on scientific literature from Semantic Scholar, enhances BERT's capabilities by focusing on the scientific domain.</p><p>To address our classification task we fine-tuned SciBERT using the training set described in Section 2.1. Specifically, we used the scibert-scivocab-uncased model from Huggingface. As optimiser, we selected AdamW <ref type="bibr" target="#b12">[13]</ref> to prevent overfitting in large models. For the finetuning process, we provided the model with the surface forms of the two topics, separated by a semicolon. For each couple of topics, we also provided the correct relationship class from the training set. We experimented with varying the number of epochs from 1 to 10, maintaining 50 warm-up steps. Our best-performing model was achieved when training for five epochs. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Evaluation</head><p>Using the test set described in Section 2.1, we evaluated the three methods outlined in the previous section: Gradient Boosting and Random Forest (both feature-based), and SciBERT (language model-based). We compared their performance using accuracy, precision, recall, and F-score, which are standard metrics for text classification. Table <ref type="table" target="#tab_0">1</ref> reports the experimental results. The language model-based method was far superior to the feature-based methods in all areas, achieving an impressive F1 score of 0.9129. This was over 27% higher than the other methods. Among the feature-based approaches, Random Forest performed better. The language model-based method was particularly effective in recognising superTopic and subTopic relations, where feature-based methods struggled, likely due to the presence of unfamiliar topics in the test set.</p><p>The language model-based method generally priorities precision over recall, particularly for the relations superTopic, subTopic, and same-as. However, for the other relation, it tends to miss some semantic connections, resulting in lower precision compared to recall. This suggests the model may incorrectly classify some related topics as other, an issue we intend to explore further in future research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>In this poster paper, we introduced a new method based on SciBERT to identify the relationship between research topics and conducted a comparative analysis against feature-based solutions. We fine-tuned a SciBERT model using a gold standard of triples derived from CSO. The model achieved an F1 score of 0.9129, a 27% improvement over methods using numerical features. These findings are significant given the growing demand for detailed ontologies to enhance content characterization in scientific KGs</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Experimental results. GB = Gradient Boosting, RF = Random Forest.</figDesc><table><row><cell cols="2">Classifier</cell><cell cols="3">Feature-based GB Feature-based RF Lang. Model-based</cell></row><row><cell cols="2">Accuracy</cell><cell>0.5842</cell><cell>0.6426</cell><cell>0.9141</cell></row><row><cell></cell><cell cols="2">supertopic 0.5424</cell><cell>0.5634</cell><cell>0.9143</cell></row><row><cell></cell><cell>subtopic</cell><cell>0.4815</cell><cell>0.6200</cell><cell>0.9452</cell></row><row><cell>Precision</cell><cell>same-as</cell><cell>0.5167</cell><cell>0.5804</cell><cell>0.9615</cell></row><row><cell></cell><cell>other</cell><cell>0.8621</cell><cell>0.8793</cell><cell>0.8286</cell></row><row><cell></cell><cell>average</cell><cell>0.6007</cell><cell>0.6608</cell><cell>0.9124</cell></row><row><cell></cell><cell cols="2">supertopic 0.4211</cell><cell>0.5263</cell><cell>0.8421</cell></row><row><cell></cell><cell>subtopic</cell><cell>0.3421</cell><cell>0.4079</cell><cell>0.9079</cell></row><row><cell>Recall</cell><cell>same-as</cell><cell>0.7750</cell><cell>0.8125</cell><cell>0.9375</cell></row><row><cell></cell><cell>other</cell><cell>0.8475</cell><cell>0.8644</cell><cell>0.9831</cell></row><row><cell></cell><cell>average</cell><cell>0.5964</cell><cell>0.6528</cell><cell>0.9177</cell></row><row><cell></cell><cell cols="2">supertopic 0.4740</cell><cell>0.5442</cell><cell>0.8767</cell></row><row><cell></cell><cell>subtopic</cell><cell>0.4000</cell><cell>0.4921</cell><cell>0.9262</cell></row><row><cell>F-score</cell><cell>same-as</cell><cell>0.6200</cell><cell>0.6771</cell><cell>0.9494</cell></row><row><cell></cell><cell>other</cell><cell>0.8547</cell><cell>0.8718</cell><cell>0.8992</cell></row><row><cell></cell><cell>average</cell><cell>0.5872</cell><cell>0.6463</cell><cell>0.9129</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_0">Gold standard and code -https://github.com/aleessiap/LeveragingLMforGeneratingOntologies.git</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In our future work, we aim to develop an innovative method for creating taxonomies of research topics to improve CSO and create large-scale ontologies across different scientific fields. We plan to combine language models and numerical features using knowledge injection techniques and experiment with recent large language models. We also intend to explore potential challenges when applying these techniques to other research domains and assess the impact of cross-disciplinary applications.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Bolanos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Salatino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2402.08565</idno>
		<title level="m">Artificial intelligence for literature reviews: Opportunities and challenges</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references</title>
		<author>
			<persName><forename type="first">L</forename><surname>Bornmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mutz</surname></persName>
		</author>
		<idno type="DOI">10.1002/asi.23329</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of the Association for Information Science and Technology</title>
		<imprint>
			<biblScope unit="volume">66</biblScope>
			<biblScope unit="page" from="2215" to="2222" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Performance of chatgpt on usmle: Potential for ai-assisted medical education using large language models</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">H</forename><surname>Kung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cheatham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Medenilla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Sillos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>De Leon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Elepaño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Madriaga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Aggabao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Diaz-Candido</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Maningo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PLoS digital health</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">e0000198</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Knowledge graphs: Opportunities and challenges</title>
		<author>
			<persName><forename type="first">C</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Naseriparsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="page" from="1" to="32" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Salatino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mannocci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2409.04432.arXiv:2409.04432" />
		<title level="m">A survey on knowledge organization systems of research fields: Resources and challenges</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Dynamic integration of multiple evidence sources for ontology learning</title>
		<author>
			<persName><forename type="first">G</forename><surname>Wohlgenannt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Weichselbraun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Scharl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sabou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Information and Data Management</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="243" to="254" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><surname>Openalex</surname></persName>
		</author>
		<ptr target="https://docs.google.com/document/d/1bDopkhuGieQ4F8gGNj7sEc8WSE8mvLZS/edit" />
		<title level="m">Openalex: End-to-end process for topic classification</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Klink-2: Integrating multiple web sources to generate semantic topic networks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web -ISWC 2015</title>
				<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="408" to="424" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The computer science ontology: a large-scale taxonomy of research areas</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Salatino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Thanapalasingam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mannocci</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web-ISWC 2018: 17th International Semantic Web Conference</title>
				<meeting><address><addrLine>Monterey, CA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">October 8-12, 2018. 2018</date>
			<biblScope unit="page" from="187" to="205" />
		</imprint>
	</monogr>
	<note>Proceedings, Part II 17</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Aida: A knowledge graph about research dynamics in academia and industry</title>
		<author>
			<persName><forename type="first">S</forename><surname>Angioni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Salatino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Recupero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Quantitative Science Studies</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="1356" to="1398" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Beltagy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cohan</surname></persName>
		</author>
		<title level="m">Scibert: A pretrained language model for scientific text</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Decoupled weight decay regularization</title>
		<author>
			<persName><forename type="first">I</forename><surname>Loshchilov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1711.05101</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
