<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Estimating the Performance of Ab Initio Calculation by VASP on Openpower High Performance System</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Vyacheslav</forename><forename type="middle">E</forename><surname>Lozhnikov</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Theoretical Physics</orgName>
								<orgName type="institution">Omsk State University</orgName>
								<address>
									<settlement>Omsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alexander</forename><forename type="middle">V</forename><surname>Mamonov</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Theoretical Physics</orgName>
								<orgName type="institution">Omsk State University</orgName>
								<address>
									<settlement>Omsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vadim</forename><forename type="middle">O</forename><surname>Borzilov</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Theoretical Physics</orgName>
								<orgName type="institution">Omsk State University</orgName>
								<address>
									<settlement>Omsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marina</forename><forename type="middle">V</forename><surname>Mamonova</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Theoretical Physics</orgName>
								<orgName type="institution">Omsk State University</orgName>
								<address>
									<settlement>Omsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Pavel</forename><forename type="middle">V</forename><surname>Prudnikov</surname></persName>
							<email>prudnikovpv@omsu.ru</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Theoretical Physics</orgName>
								<orgName type="institution">Omsk State University</orgName>
								<address>
									<settlement>Omsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Aleksei</forename><forename type="middle">A</forename><surname>Sorokin</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Computing Center of Far-Eastern Branch</orgName>
								<orgName type="institution">Russian Academy of Sciences</orgName>
								<address>
									<settlement>Khabarovsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Georgy</forename><forename type="middle">G</forename><surname>Baksheev</surname></persName>
							<email>g.baksheev@g.nsu.ru</email>
							<affiliation key="aff2">
								<orgName type="institution">Novosibirsk State University</orgName>
								<address>
									<settlement>Novosibirsk</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Estimating the Performance of Ab Initio Calculation by VASP on Openpower High Performance System</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">7AF992B561A5E4BB1376A31EFCBF6BBF</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:30+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this work we compare the performance of Pascal P100 GPUs vs POWER8 CPU on OpenPOWER HPC system by VASP calculation of energy and magnetic characteristics of Fe/Cu(111)/Fe and Co/Cu(100)/Co multilayer magnetic nanostructures. We revealed that the VASP code demonstrates a maximum performance on OpenPOWER System with the GPUs.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The behavior of multilayer magnetic structures has become of great technological importance due to the applications in magnetic storage devices. The ab initio calculations are widely used to calculate some characteristics of solids <ref type="bibr" target="#b0">[1]</ref> and multilayer magnetic structures <ref type="bibr" target="#b1">[2]</ref>. The main advantage of ab initio approach is independence from experimental data. Unlike the case of semi-empirical methods, there is no need for calibration or fitting parameters. Thus, ab initio methods can also be used to calculate the characteristics of perspective systems, i.e., for prediction of properties of materials that have not yet been developed. Mainly used packages that can perform ab initio calculations are VASP <ref type="bibr" target="#b2">[3]</ref><ref type="bibr" target="#b3">[4]</ref><ref type="bibr" target="#b4">[5]</ref>, Quantum Espresso <ref type="bibr" target="#b5">[6]</ref>, ABINIT <ref type="bibr" target="#b6">[7]</ref>, Wien2K <ref type="bibr">[8]</ref>.</p><p>Effective application of ab initio calculations requires the scalability of the code for novel high performance systems (HPS) with different hardware architectures. In this work we focused on the Vienna Ab initio Simulation Package (VASP). VASP is a complex package for performing ab initio quantum-mechanical molecular dynamics simulations using pseudopotentials or the projector-augmented wave method and a plane wave basis set <ref type="bibr" target="#b3">[4]</ref>. Now it is one of the most popular parallel code for quantum chemistry and solid-state calculations of electronic structure. So the estimation of VASP code performance for HPS is an actual and non-trivial task. In this paper, we compare time of execution on POWER8 CPU and PASCAL P100 GPU with NVLink interconnection. We apply adjusting parameters in VASP INCAR file to increase performance on GPU.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>2</head><p>The Basics of the Density Functional Method</p><p>We calculate energy and magnetic characteristics of Fe/Cu(111)/Fe and Co/Cu(100)/Co multilayer magnetic nanostructures. The central idea of density functional theory (DFT) <ref type="bibr" target="#b2">[3]</ref> is to consider the electron density n(r) instead of the full many-body wave functions ) ,..., ( 1 N r r  . To ensure the possibility of calculating the magnetic properties, the energy of the system is written in the form of a functional not only of the electron density n(r), but also of the magnetization density m(r), see formula <ref type="bibr" target="#b0">(1)</ref>. The Kohn-Sham wave functions are replaced by two-component Pauli wave functions Ψν (r), capable of representing both the electron density and the magnetization density. Index ν here denotes spin states.</p><formula xml:id="formula_0">, ) ( ) ( 1  −  =  = N i i r r r n   =   = N v v v r r r m 1 * ) ( ) ( ) (  (1)</formula><p>From the variational principle, the Kohn-Sham equations are obtained:</p><formula xml:id="formula_1">, 0 ) ( ) ( 2 2 =        −  + +  − r r B V m v v xc eff      ) ( ) ( ), ( r m r m r E B xc xc   =  (2)</formula><p>where Bxc is the effective magnetic field arising from the exchange-correlation energy.</p><p>The main problem associated with the density functional theory method is that exact analytical expressions for exchange and correlation functionals are known only for the particular case of a gas of free electrons. Nevertheless, the existing approximations allow us to calculate a number of physical quantities with sufficient accuracy.</p><p>In this work we used GGA (generalized gradient approximations) approximations in terms of Perdew-Burke-Ernzerhof (PBE) <ref type="bibr" target="#b9">[9]</ref>:</p><formula xml:id="formula_2">    r d r n r n r n r n r n r n E GGA xc                 = ) ( ), ( ), ( ), ( ) ( ), ( <label>(3)</label></formula><p>The essence of the projection augmented wave (PAW) method is to transform the pseudowave functions, obtained in the pseudopotential method into all-electron wave functions, thereby restoring the information lost when considering pseudowave functions. The number of plane wave components is limited by the Cut-off Energy. To describe the first Brillouin zone we used standard method Monkhorst-Pack with the parameter K-points characterizing regular grid in k-space <ref type="bibr" target="#b10">[10]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Compiling the Parallel Version of VASP Code for Openpower and Intel Architectures</head><p>Official support of the GPU calculations appeared in VASP from version 5. <ref type="bibr" target="#b3">4</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Hardware Information</head><p>IBM Power System S822LC is two-socket HPC system with two POWER8 CPUs with 20 cores running at 4 GHz and interconnected with two Nvidia Pascal P100 GPUs with a high bandwidth (80GByte in and 80GByte out) NVLink 1.0 interface (Fig. <ref type="figure" target="#fig_0">1</ref>). It is very important for exchange data between multiple GPUs and fast load data from CPU. The major goal of this system is efficiently use GPU units and accelerating calculations. A large part of HPC resources installed during the last decade are based on Intel CPUs. Novel generations of Intel CPUs present a wide spectrum of multicore processors <ref type="bibr" target="#b4">[5]</ref>. Intel Core i7 4770 is desktop processor, but it is "tock" model in Intel extensive strategy of microprocessor developmentit is mostly complete 22 nm architecture. We compared IBM POWER8 with Intel Haswell because both architectures introduced in 2013 year and had 22 nm technical processes. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Model and Simulation Parameters</head><p>In this work, the results of numerical first-principles calculations of the energy and magnetic characteristics for cobalt and iron films on a copper surface obtained by using VASP software package by means of the Projector Augmented Wave (PAW) method are presented. The values of the total energy of collinear spin configuration, the total magnetic moment and the magnetic moments of Co and Fe atoms are calculated. We investigated a system consisting of a copper slab and adsorbed on it from both sides by a ferromagnetic film with the thickness of the films in three monoatomic layers. The multilayer structure was simulated using a periodic 2×2 36-atom supercell with the lattice constant corresponding to the copper substrate a = 3.6367 (5) Å, which we obtained as a result of calculations taking into account the optimization of the lattice parameters. The surface face orientation is (100) for Co/Cu system and (111) for Fe/Cu system. There is an opinion that for GPU NSIM parameter needs to be increased while we have free memory on the GPU. − LPLANE is a useful parameter for optimization which can reduce intercommunication time, but it is actual firstly for massively parallel systems, according to VASP documentation.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Estimation of Performance and Accuracy of Calculations</head><p>For CoCu system we used three configurations to compare VASP calculation times with similar INCAR parameters for IBM POWER8 CPU and Intel Haswell CPU. We set LREAL=.TRUE. as described at VASP official documentation and use NCORE=1 with one MPI thread for GPU calculations. We do not use NVIDIA MPS system and do not set NSIM parameter clearly, but we know that it is important for large tasks especially. Core i7 has only 4 real cores and we run VASP with 4 processes only. On POWER8 we run VASP on 8 cores for using most of one CPU. The results of magnetization and free energy calculations (Table <ref type="table" target="#tab_2">3</ref>) are well correlated. The calculations with the GPU provide less accuracy, but the value of the error is not so sufficient. The times of calculations are different for POWER8 system and Intel Core i7 (Table <ref type="table" target="#tab_3">4</ref>). The one POWER8 thread was more efficient than one Intel thread for the VASP calculations. If we use GPU only with one MPI thread we have much better performance ( (Fig. <ref type="figure" target="#fig_3">4</ref>). As we can see in Table <ref type="table" target="#tab_5">5</ref> antiferromagnet spin configuration need sufficient more memory than a ferromagnet. Execution time on GPU highly depends on spin configuration too. It makes sense to note, that GPU utilization is not full and floating from about 20% to 70% during calculation for both systems.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions</head><p>VASP is widely used by researchers to get characteristics of solids and multilayer magnetic structures. NCORE and NSIM parameters can be very useful to maximize performance on GPU. The values of acquired quantities and accuracy of GPU calculations are in a good agreement with CPU results. To use VASP efficiently with GPUs more memory and calculation time is required in comparing with calculations on CPU, especially for antiferromagnet spin configurations. </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: NVLink communications protocol in IBM Power System S822LC</figDesc><graphic coords="3,168.10,70.90,304.01,200.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Representations of Co/Cu/Co and Fe/Cu/Fe multilayer nanostructures For Fe/Cu system the calculations of the total energy were realized for ferromagnetic and two different antiferromagnetic spin configurations. The antiferromagnetic spin configurations for which the calculation was carried out are shown in Fig. 3. The magnetic moment of the atoms is directed along the z axis. VASP INCAR file has several adjusting parameters that can increase GPU performance. The main of those are NCORE, NPAR, NSIM, LPLANE. − NCORE determines how many cores work on individual orbital; − NPAR depend on NCORE as NCORE = number of the cores / NPAR;</figDesc><graphic coords="3,218.90,450.79,77.20,216.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Fe/Cu/Fe antiferromagnetic spin configurations</figDesc><graphic coords="4,320.95,180.76,120.05,135.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Dependence of the execution times on used architecture and different tasks</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0"><head></head><label></label><figDesc></figDesc><graphic coords="1,0.00,190.95,595.32,460.02" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>.1 and in the our work we used version 5.4.4. VASP has one precompiled configuration file, named makefile.include, with a lot of parameters. Showing all parameters is redundancy and we present the main part of it in table 1. We use Intel Parallel Studio XE C/C++ with Intel MKL library to compile the VASP package on X86_64 architecture. The optimization flags were choosen -O1 and -O2 because compilation with harder optimization was not complete successfully. It was used Ubuntu 16.04 with 4.4.0-137 kernel. We used XlC 13.1.5 and Xlf 15.1.5 with including ESSL library on CentOS 7 with 3.10.0-514 kernel to compile VASP on IBM Power System S822LC. Compilers options</figDesc><table><row><cell>Compiler</cell><cell>Version</cell><cell>Flags</cell></row><row><cell></cell><cell></cell><cell>-g -q64 -O3 -qarch=pwr8</cell></row><row><cell>IBM XL C/C++</cell><cell>13.1.5</cell><cell>-qtune=pwr8:st</cell></row><row><cell></cell><cell></cell><cell>-qfullpath -qsaveopt</cell></row><row><cell></cell><cell></cell><cell>-g -q64 -O3 -qarch=pwr8</cell></row><row><cell>IBM XL Fortran</cell><cell>15.1.5</cell><cell>-qtune=pwr8:st -qfullpath -qsaveopt</cell></row><row><cell></cell><cell></cell><cell>-qflag=i:e -qsuppress=cmpmsg</cell></row><row><cell></cell><cell></cell><cell>-DCUDA_GPU -DRPROMU_CPROJ</cell></row><row><cell></cell><cell></cell><cell>OVERLAP</cell></row><row><cell>NVIDIA CUDA compilation tools</cell><cell>8.0.61</cell><cell>-DCUFFT_MIN=28 -UscaLAPACK -fPIC -DADD -DMAGMA_WITH_MKL -DMAGMA_SETAFFINITY -</cell></row><row><cell></cell><cell></cell><cell>DGPUSHMEM=300</cell></row><row><cell></cell><cell></cell><cell>-DHAVE_CUBLAS</cell></row><row><cell>Intel Parallel Studio XE C/C++</cell><cell>2017</cell><cell>-O2 -f_com=no -free -w0</cell></row><row><cell>Intel Parallel Studio XE Fortran</cell><cell>2017</cell><cell>-O1 -mkl=sequential -lstdc++</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Parameters of modeling structures</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 :</head><label>3</label><figDesc>Comparison of the accuracy of the calculations performed on the CPU and GPU in VASP for CoCu</figDesc><table><row><cell>Architecture</cell><cell>Total</cell><cell>Free energy (eV)</cell></row><row><cell></cell><cell>magnetization</cell><cell></cell></row><row><cell>Intel Core i7-4770 Haswell</cell><cell>40.646</cell><cell>-202.20538224</cell></row><row><cell>IBM POWER8</cell><cell>40.646</cell><cell>-202.20538199</cell></row><row><cell>IBM POWER8+Nvidia P100</cell><cell>40.658</cell><cell>-202.04101161</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 )</head><label>4</label><figDesc>than Intel or POWER CPUs even without optimizations of VASP parameters in INCAR file.For FeCu systems we used NCORE = 4 and NSIM = 32 to get better performance for GPU calculations We perform simulation of the ferromagnet FeCu system on ten POWER8 cores to compare execution times with GPU</figDesc><table><row><cell></cell><cell>Co/Cu</cell><cell>Fe/Cu</cell><cell>Fe/Cu AF1</cell><cell>Fe/Cu AF2</cell></row><row><cell>Cutoff energy</cell><cell>500 eV</cell><cell>350 eV</cell><cell>350 eV</cell><cell>350 eV</cell></row><row><cell>K-points</cell><cell>12</cell><cell>10</cell><cell>10</cell><cell>10</cell></row><row><cell>Number of atoms</cell><cell>36</cell><cell>36</cell><cell>36</cell><cell>36</cell></row><row><cell>Thickness of vacuum layers</cell><cell>5 Å</cell><cell>4 Å</cell><cell>4 Å</cell><cell>4 Å</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 :</head><label>4</label><figDesc>Execution times for different computing systems (in hours)</figDesc><table><row><cell>Architecture</cell><cell>Co/Cu</cell><cell>Fe/Cu</cell><cell cols="2">Fe/Cu AF1 Fe/Cu AF2</cell></row><row><cell>Intel Core i7-4770 Haswell</cell><cell>38</cell><cell></cell><cell></cell><cell></cell></row><row><cell>IBM POWER8</cell><cell>14.4</cell><cell>9.55</cell><cell></cell><cell></cell></row><row><cell>IBM POWER8+Nvidia P100</cell><cell>8</cell><cell>4.04</cell><cell>12.15</cell><cell>4.89</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5 :</head><label>5</label><figDesc>Used memory for different computing systems (in Gb)</figDesc><table><row><cell>Architecture</cell><cell>Co/Cu</cell><cell>Fe/Cu</cell><cell cols="2">Fe/Cu AF1 Fe/Cu AF2</cell></row><row><cell>Intel Core i7-4770 Haswell</cell><cell>6.91</cell><cell></cell><cell></cell><cell></cell></row><row><cell>POWER8</cell><cell>6.67</cell><cell>4.87</cell><cell></cell><cell></cell></row><row><cell>IBM POWER8+Nvidia P100</cell><cell>29.64</cell><cell>15.9</cell><cell>25.88</cell><cell>20.84</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>We would like to thank the IBM experts, who help us to optimize the VASP package for IBM Power Systems S822LC. This research was supported by the grants 17-02-00279, 18-42-550003 of Russian Foundation of Basic Research and by the grant MD-6868.2018.2 of the President of the Russian Federation. The simulations were supported by the computational resources of Shared Facility Center "Data Center of FEB RAS" (Khabarovsk) <ref type="bibr" target="#b10">[10]</ref>. Computations were performed with the methods and techniques which had been developed under the RFBR scientific project number 18-29-03196.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Reproducibility in density functional theory calculations of solids</title>
		<author>
			<persName><forename type="first">K</forename><surname>Lejaeghere</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bihlmayer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Björkman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Science</title>
		<imprint>
			<biblScope unit="volume">351</biblScope>
			<biblScope unit="page">3000</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">First-principles investigations of the atomic structure and magnetic properties of Ni and Co films on Cu substrate</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Kondrashov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Mamonova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">S</forename><surname>Povoroznuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">V</forename><surname>Prudnikov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lobachevskii Journal of Mathematics</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page">940</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Efficient iterative schemes for ab initio total-energy calculations using a plane-wave basis set</title>
		<author>
			<persName><forename type="first">G</forename><surname>Kresse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Furthmüller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Phys. Rev. B</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page">11169</biblScope>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Kresse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Marsman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Furthmüller</surname></persName>
		</author>
		<ptr target="https://cms.mpi.univie.ac.at/vasp/vasp/vasp.html" />
		<title level="m">VASP THE GUIDE</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Efficiency Analysis of Intel and AMD x86 64 Architectures for Ab Initio Calculations: A Case Study of VASP</title>
		<author>
			<persName><forename type="first">V</forename><surname>Stegailov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Vecher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Supercomputing RuSCDays 2017</title>
		<title level="s">Communications in Computer and Information Science</title>
		<editor>
			<persName><forename type="first">V</forename><surname>Voevodin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Sobolev</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">793</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">QUANTUM ESPRESSO: a modular and open-source software project for quantum simulations of materials</title>
		<author>
			<persName><forename type="first">P</forename><surname>Giannozzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Baroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bonini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Physics: Condensed Matter</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page">395502</biblScope>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">ABINIT: First-principles approach to material and nanosystem properties</title>
		<author>
			<persName><forename type="first">X</forename><surname>Gonze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Amadon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">M</forename><surname>Anglade</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Comput. Phys. Commun</title>
		<imprint>
			<biblScope unit="volume">180</biblScope>
			<biblScope unit="page">2582</biblScope>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Solid state calculations using WIEN2k</title>
		<author>
			<persName><forename type="first">K</forename><surname>Schwarz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Blaha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Materials Science</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="259" to="273" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Generalized Gradient Approximation Made Simple</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Perdew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Burke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ernzerhof</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Phys. Rev. Lett</title>
		<imprint>
			<biblScope unit="volume">77</biblScope>
			<biblScope unit="page">3865</biblScope>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Special points for Brillouin-zone integrations</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Monkhorst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Pack</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Phys. Rev. B</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">5188</biblScope>
			<date type="published" when="1976">1976</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title/>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Sorokin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">I</forename><surname>Makogonov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Korolev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific and Technical Information Processing</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page">302</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
