<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Green AI Ontology: An Ontology for Modeling the Energy Consumption of AI Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Michael</forename><surname>Färber</surname></persName>
							<email>michael.faerber@kit.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Karlsruhe Institute of Technology (KIT)</orgName>
								<orgName type="institution">Institute AIFB</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">David</forename><surname>Lamprecht</surname></persName>
							<email>david.lamprecht@student.kit.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Karlsruhe Institute of Technology (KIT)</orgName>
								<orgName type="institution">Institute AIFB</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Green AI Ontology: An Ontology for Modeling the Energy Consumption of AI Models</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2F9FCCA0A272E7771B67F838CFAF195A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T03:13+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Machine Learning</term>
					<term>Green AI</term>
					<term>Energy Consumption</term>
					<term>Ontology Engineering (D. Lamprecht) Orcid 0000-0001-5458-8645 (M. Färber); 0000-0002-9098-5389 (D. Lamprecht)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Modeling AI systems' characteristics of energy consumption and their sustainability level as an extension of the FAIR data principles has been considered only rudimentarily. In this paper, we propose the Green AI Ontology for modeling the energy consumption and other environmental aspects of AI models. We evaluate our ontology based on competency questions. Our ontology is available at https://w3id.org/ Green-AI-Ontology and can be used in a variety of scenarios, ranging from comprehensive research data management to strategic controlling of institutions and environmental efforts in politics.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Pre-trained language models such as GPT have been commended for their artificial general intelligence capabilities and are nowadays widely used for tasks such as question answering, information extraction, and text summarization. However, in the case of GPT-3 with its 175 billion parameters, the training required 10,000 GPUs and cost 552 metric tons of carbon dioxide. 1 Thus, the question arises of how "green" AI models are. Regardless of an ethical assessment, we argue that it is useful to model AI systems' characteristics of energy consumption and sustainability (e.g., operating costs), extending the FAIR data principles <ref type="bibr" target="#b0">[1]</ref>, which focus on the availability and reuse of research data and other artifacts. Existing ontologies and knowledge graphs focus on the modeling of the research landscape, modeling publications, authors, and venues (e.g., FaBiO, ORKG, MAKG) <ref type="bibr" target="#b1">[2]</ref>. Furthermore, ontologies for modeling software and neural networks have been proposed. For instance, the Ontology for Informatics Research Artifacts (OIRA) <ref type="bibr" target="#b2">[3]</ref> provides a way to model software and datasets. In FAIRnets <ref type="bibr" target="#b3">[4]</ref>, the authors propose a schema for modeling neural networks. However, surprisingly, none of these ontologies allow the modeling of the energy consumption of AI models (e.g., runtime or CO2 footprint of pretrained language models, which can be measured via tools <ref type="bibr" target="#b4">[5]</ref>).  In this paper, we propose -to our knowledge -the first ontology for modeling the energy consumption of AI models. It is available at https://w3id.org/Green-AI-Ontology (OWL file at https://w3id.org/Green-AI-Ontology/ontology). We create a knowledge graph based on our ontology and evaluate our ontology based on competency questions. Our ontology can be used in various scenarios, ranging from improved research data management to strategic controlling of institutions and implementation of standards.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">The Green AI Ontology</head><p>Ontology Design. Figure <ref type="figure" target="#fig_1">1</ref> shows the main classes and properties of the ontology. The corresponding documentation is linked in our repository. Overall, our ontology is designed to model the following aspects:</p><p>1. Metrics and tools: This part addresses the metrics that are used to measure the energy consumption of AI models. Apart from the pure values (Energy Measure), we consider the online services (Energy Measurement Service) with which the values can be determined. Energy Measurement Service is defined as a subclass of Energy Measurement so that all relevant key figures are modeled in addition to information about the service. 2. Hardware settings, cloud service, and location: The property shows information about the hardware used. In addition to the modeling of private infrastructures, services/cloud providers are represented here. The location of the hardware (e.g., city, country), which may have an impact on the environmental balance, can also be taken into account. 3. Software settings: The property hasSoftwareSettings shows information about the software used, including software packages and modules. 4. Linking to scholarly linked data: This part of the ontology is designed to integrate the modeled energy consumption information into the modeling of the scientific landscape.</p><p>Since the AI models are closely linked to further computer science artifacts (e.g., datasets, software), we reuse the Ontology for Informatics Research Artefacts <ref type="bibr" target="#b2">[3]</ref> considering best practice for reusing existing ontologies. As a result, the modeled information is not a silo but is closely linked to papers, data sets, and researchers. In this way, novel queries and strategic controlling are possible (e.g., answering: What is the average energy consumption of AI models developed and trained at my institution over the last five years? ).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Knowledge Graph Construction.</head><p>To create a knowledge graph based on our ontology, we first applied 10 regex patterns (an extension of <ref type="bibr" target="#b5">[6]</ref>) on all 217,000 arXive computer science papers as of July 31, 2020 (from http://unarxive.org). In this way, we obtained 3,016 energy information units. However, we noticed that the accuracy of the matched patterns is insufficient due to the low precision of the information extraction approach. For instance, a large portion of the extracted energy information refers to non-AI models, such as mobile phones and e-cars.</p><p>Thus, we refrained from this approach and instead asked AI researchers via a questionnaire to report the energy consumption of AI models published in papers. In this way, we obtained a proof-of-concept knowledge graph, modeling 40 AI models and 1,975 statements.</p><p>Ontology Evaluation. Following the best practices of ontology engineering (e.g., the NeOn ontology engineering methodology), we identified 15 competency questions (see our repository; based on 79 Green AI-related papers that are listed in our repository) that our ontology should be able to answer. Based on our created knowledge graph and created SPARQL queries (see our repository and Listing 1), we were able to answer all competence questions.</p><p>Use Cases. In the following, we outline several potential use cases of our ontology.</p><p>Research Data Management. The FAIR principles <ref type="bibr" target="#b0">[1]</ref> have been proposed to ensure that resources are findable, accessible, interoperable, and reusable. Our ontology can be considered SELECT * WHERE { ?AIModel a gai:AIModel .</p><p>?AIModel gai:hasEnergyMetrics ?EnergyMetrics . ?EnergyMetrics gai:hasFPO ?FloatingPointOperations . } Listing 1: SPARQL query answering "How many floating point operations (FPO) do the AI models have?"</p><p>an extension of these principles, allowing the modeling usage information next to existing ontologies and knowledge graphs. AI Systems. Engineers training and deploying AI models as well as end users may be increasingly interested in knowing the environmental background of given AI models <ref type="bibr" target="#b4">[5]</ref> in order to assess them more thoroughly than merely for their effectiveness. Our modeling of the energy consumption of AI models is not restricted to one metric (e.g., CO2, run time); instead, our ontology allows the modeling of several measurements for each AI model.</p><p>Society. From the perspective of popular sciences and politicians, our ontology complies with the rising public awareness of Green AI and environmental studies. The ontology enables energy consumption to be put into perspective (e.g., comparing energy consumption of language models and bitcoin mining).</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Main classes and properties of the Green AI Ontology.</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Conclusion</head><p>In this paper, we proposed the Green AI Ontology for modeling the energy consumption of AI models. It can be used to extend academic knowledge graphs, to encourage researchers to provide information on the energy consumption of their AI models, and to ensure that the community appreciates this information.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>CEUR Workshop Proceedings (CEUR-WS.org) 1 https://fortune.com/2021/04/21/ai-carbon-footprint-reduce-environmental-impact-of-tech-google-research-study/</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The FAIR Guiding Principles for scientific data management and stewardship</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Wilkinson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific Data</title>
		<imprint>
			<biblScope unit="page" from="2052" to="4463" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Ontologies Supporting Research-Related Information Foraging Using Knowledge Graphs: Literature Survey and Holistic Model Mapping</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">B</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Svátek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Rabby</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ó</forename><surname>Corcho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of EKAW</title>
				<meeting>of EKAW</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="88" to="103" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Ontology for informatics research artifacts</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">B</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Svátek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th Extended Semantic Web Conference, ESWC&apos;21</title>
				<meeting>the 18th Extended Semantic Web Conference, ESWC&apos;21</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="126" to="130" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Making neural networks FAIR</title>
		<author>
			<persName><forename type="first">A</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Weller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Färber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sure</surname></persName>
		</author>
		<author>
			<persName><surname>Vetter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second Iberoamerican Conference and First Indo-American Conference on Knowledge Graphs and Semantic Web, KGSWC&apos;20</title>
				<meeting>the Second Iberoamerican Conference and First Indo-American Conference on Knowledge Graphs and Semantic Web, KGSWC&apos;20</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="29" to="44" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Quantifying the Carbon Emissions of Machine Learning</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lacoste</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Luccioni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Dandres</surname></persName>
		</author>
		<idno>CoRR abs/1910.09700</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Towards the systematic reporting of the energy and carbon footprints of machine learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Henderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Romoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Brunskill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pineau</surname></persName>
		</author>
		<idno>CoRR abs/2002.05651</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
