<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Low-Latency Privacy-Preserving Deep Learning Design via Secure MPC</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Ke</forename><surname>Lin</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Tsinghua University</orgName>
								<address>
									<addrLine>30 Shuangqing Rd., Haidian District</addrLine>
									<postCode>100084</postCode>
									<settlement>Beijing</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yasir</forename><surname>Glani</surname></persName>
							<email>yasirglani@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Tsinghua University</orgName>
								<address>
									<addrLine>30 Shuangqing Rd., Haidian District</addrLine>
									<postCode>100084</postCode>
									<settlement>Beijing</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ping</forename><surname>Luo</surname></persName>
							<email>luop@tsinghua.edu.cn</email>
							<affiliation key="aff0">
								<orgName type="institution">Tsinghua University</orgName>
								<address>
									<addrLine>30 Shuangqing Rd., Haidian District</addrLine>
									<postCode>100084</postCode>
									<settlement>Beijing</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Low-Latency Privacy-Preserving Deep Learning Design via Secure MPC</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">13F124455D8F40F6B75A9EE4BC935051</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Multi-party computation</term>
					<term>deep learning</term>
					<term>privacy-preserving</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Secure multi-party computation (MPC) facilitates privacy-preserving computation between multiple parties without leaking private information. While most secure deep learning techniques utilize MPC operations to achieve feasible privacy-preserving machine learning on downstream tasks, the overhead of the computation and communication still hampers their practical application. This work proposes a low-latency secret-sharing-based MPC design that reduces unnecessary communication rounds during the execution of MPC protocols. We also present a method for improving the computation of commonly used nonlinear functions in deep learning by integrating multivariate multiplication and coalescing different packets into one to maximize network utilization. Our experimental results indicate that our method is effective in a variety of settings, with a speedup in communication latency of 10 ∼ 20%.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Secure multi-party computation (MPC) <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref> enables parties to compute securely over their private data without revealing the data to each other. Secure MPC offers privacypreserving property, which makes it suitable for most privacy-sensitive domains, such as medical research and finance. Upon the development of deep learning techniques, the ability to capture important information from large datasets of neural models raises concerns regarding the surveillance of individuals <ref type="bibr" target="#b2">[3]</ref>. In this case, the prospects of secure MPC demonstrate its application to secure machine learning and deep learning. While MPC-based deep learning frameworks have achieved significant performance in general scenarios, most works suffer from the limitations caused by 1. network communication due to the nature of exchanging intermediate information during MPC execution, 2. excessive computation introduced by complex MPC protocols. Since the computation of MPC protocols is largely determined by their sophisticated design, optimizing the protocol itself would seem to be difficult and infeasible. Thus, some studies <ref type="bibr" target="#b3">[4]</ref> are concerned with improving the communication stage of MPC protocols to make them more practical. In this paper, we present an approach to reduce the communication latency of the MPC protocol through optimized multivariate multiplication.</p><p>In general, privacy-preserving deep learning frameworks usually adopt secret-sharing-based techniques to avoid extensive computational overheads <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8]</ref>. Consequently, secret-sharing-based methods require multiple exchanges of intermediate results to achieve collaborative MPC operations. As these MPC techniques are based on linear computations, such as addition and multiplication, modern deep learning techniques that inherently rely on linear algebra benefit significantly from them. Considering the heavy dependency on linear operations, our research aims to reduce unnecessary communication rounds following <ref type="bibr" target="#b8">[9]</ref> during the execution of MPC protocols.</p><p>Our main contributions are as follows:</p><p>• We improve the computation of nonlinear functions by integrating the proposed multivariate multiplication and coalescing different packets into one single packet to maximize network utilization.</p><p>• We conducted experiments to evaluate the effectiveness of our method in the context of models with varying sizes, networks with different latency and bandwidth, the accuracy of downstream classification tasks, and the number of participants involved. The results indicate an overall improvement of 10 ∼ 20% in communication latency.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Arithmetic Secret Sharing Based Scheme</head><p>Our setup is primarily focused on arithmetic operations, so we represent all inputs and intermediate results in terms of linear secret sharing between 𝑛 parties, especially in the context of additive secret-sharing schemes.</p><p>Apart from the general (𝑛, 𝑡)-Shamir secret sharing scheme <ref type="bibr" target="#b9">[10]</ref>, which relies on the degree-𝑡 polynomials over 𝑛 parties, we adopt the simple arithmetic secret sharing scheme based on (𝑛, 0)-Shamir secret sharing. In other words, we share a scalar value 𝑥 ∈ Z/𝑄Z across 𝑛 parties 𝒫, where Z/𝑄Z denotes a ring with 𝑄 elements, following the notations of <ref type="bibr" target="#b4">[5]</ref>. The sharing of 𝑥 is defined as [𝑥] = {[𝑥]𝑝}𝑝∈𝒫 , where [𝑥]𝑝 is the party 𝑝's share of 𝑥. The ground-truth value 𝑥 could be reconstructed from the sum of the shares of each party, i.e. 𝑥 = ∑︀ 𝑝∈𝒫 [𝑥]𝑝. When parties wish to share a value 𝑥, they generate a pseudorandom zero-share that sums to 0. The party that possesses the value adds 𝑥 to their share in secret. To represent floating-point numbers, we adopt a fixed-point encoding to encode any floating-point number 𝑥𝐹 into a fixed-point representation, 𝑥. Alternatively, we consider that each 𝑥 is the result of multiplying a floating-point number 𝑥𝐹 by a scaling factor 𝐵 = 2 𝐿 and rounding to the nearest integer, i.e. 𝑥 = ⌊𝐵𝑥𝐹 ⌉. Here 𝐿 is the precision of the fixed-point encoding. To decode a ground-truth floating-point value 𝑥𝐹 from 𝑥, we compute as follows: 𝑥𝐹 ≈ 𝑥/𝐵.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Arithmetic Secret Sharing Based MPC</head><p>It is noteworthy that arithmetic secret shares are homomorphic and can be used to implement secure MPC, especially in the context of linear computation in most cases. Linear functions. It is possible to implement functions that consist of linear operations by combining additions and multiplications. Common operations in deep learning, such as element-wise product and convolution, are allowed in a linear paradigm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Addition</head><p>Nonlinear functions. Due to the inherent infeasibility of nonlinear functions in the standard arithmetic secretsharing scheme, most works use approximation methods to simulate the outcome of nonlinear functions. In particular, Taylor Expansion, Newton-Rhapson, and Householder methods are commonly used to approximate nonlinear functions using only linear operations. All reciprocal functions, exponential functions, loss functions, kernel functions, and other useful functions in deep learning are calculated this way, for example. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Algorithm 1 Beaver Multiplication Mul([𝑥], [𝑦])</head><formula xml:id="formula_0">Input: Secret-shared inputs [𝑥], [𝑦], Beaver triple ([𝑎], [𝑏], [𝑎𝑏]). Output: [𝑥𝑦]. 1: ◁ Compute masked values 2: [𝜖] ← [𝑥 − 𝑎] = [𝑥] − [𝑎] 3: [𝛿] ← [𝑦 − 𝑏] = [𝑦] − [𝑏]</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Notations</head><p>This section summarizes the notations used throughout this work. We denote [𝑥] as a secret sharing of 𝑥. Reveal([𝑥]) means that the ground-truth value 𝑥 is revealed to every party involved in the computation through one-round communications. Since most linear operations are also applicable to element-wise operations and matrix operations, 𝑥 can also represent a vector, matrix, or even a tensor if there is no confusion and ambiguity.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Related Work</head><p>To achieve communication-efficient MPC, various approaches have been developed to optimize the communication rounds and the throughput of communication. Ishai and Kushilevitz <ref type="bibr" target="#b11">[12]</ref> proposes a new representation of polynomials for round-efficient secure computation, dividing high-degree polynomials into multiple low-degree polynomials that are easy to solve. Mohassel and Franklin <ref type="bibr" target="#b12">[13]</ref> performs operations directly on polynomials, such as polynomial multiplication and division. Dachman-Soled et al. <ref type="bibr" target="#b13">[14]</ref> improves the evaluation of multivariate polynomials with different variables being held as private inputs by each party. Then, Lu et al. <ref type="bibr" target="#b3">[4]</ref> proposes an efficient method for evaluating high-degree polynomials with arbitrary numbers of variables. While the current research has focused on improving the calculation of polynomials, our study aims to develop a communication-efficient and effective MPC system for use in modern deep learning frameworks by leveraging arithmetic tuples computation from Krips et al. <ref type="bibr" target="#b8">[9]</ref>. This system is not confined to only computing polynomials within finite rings, as seen in previous studies.</p><p>In recent years, several deep learning frameworks that preserve privacy have emerged to enable the secure inference of neural network models. Wagh et al. <ref type="bibr" target="#b7">[8]</ref> implements a maliciously secure 3-party MPC protocol from Se-cureNN <ref type="bibr" target="#b14">[15]</ref> and ABY 3 <ref type="bibr" target="#b15">[16]</ref>. Knott et al. <ref type="bibr" target="#b4">[5]</ref> provides flexible machine-learning APIs with a rich set of functions for secure deep learning. Li et al. <ref type="bibr" target="#b6">[7]</ref> presents a fast and performant MPC Transformer inference framework designed to be privacy-preserving. Our low-latency linear MPC implementation is built on top of Knott et al.'s CrypTen framework and provides a significant improvement in communication latency.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Multivariate Multiplication</head><p>Since Beaver triples illustrate how to multiply two variables with pre-shared triplets, a classic multiplication between multiple variables, such as [𝑥𝑦𝑧], requires several rounds of binary multiplication, i.e. Mul(Mul([𝑥], [𝑦]), <ref type="bibr">[𝑧]</ref>). This naive implementation, however, introduces additional communication rounds during the on-the-fly Reveal process. In general, a 𝑛-ary multiplication requires 𝑛 − 1 rounds of communication.</p><p>To reduce the communication rounds involved in the multivariate multiplications, the basic binary Beaver triple is extended into a general 𝑛-ary Beaver triple. This results in only one round of communication required throughout the entire process.</p><p>Assume the 𝑛 inputs could be represented as {[𝑥𝑖]} 𝑛 𝑖=1 . The precomputation and preshared information required by the extended protocol is {𝒜𝑖} 𝑛 𝑖=1 . Here 𝒜1 := {[𝑎𝑗]} 𝑛 𝑗=1 is defined as the set of 𝑛 auxiliary shared values used to blind the the inputs {[𝑥𝑖]} 𝑛 𝑖=1 , which is also similar to the Beaver's idea. Then 𝒜𝑖(𝑖 ≥ 2) is defined as the set of shared degree-𝑖 cross-terms consisting of the variables in 𝒜1. For example, 𝒜2 could be defined as </p><formula xml:id="formula_1">𝒜2 := {[𝑎𝑖𝑎𝑗] | 𝑖 ̸ = 𝑗 ∧ 1 ≤ 𝑖, 𝑗 ≤ 𝑛}, and 𝒜3 := {[𝑎𝑖𝑎𝑗𝑎 𝑘 ] | 𝑖 ̸ = 𝑗 ̸ = 𝑘 ∧ 1 ≤ 𝑖, 𝑗, 𝑘 ≤ 𝑛},</formula><formula xml:id="formula_2">= ∏︁ 𝛿𝑖 + ∑︁ 𝑖 𝛿𝑖 ∏︀ 𝑎𝑚 𝑎𝑖 + ∑︁ 𝑖,𝑗,𝑖̸ =𝑗 𝛿𝑖𝛿𝑗 ∏︀ 𝑎𝑚 𝑎𝑖𝑎𝑗 + • • • + ∏︁ 𝑎𝑖.</formula><p>(</p><formula xml:id="formula_3">)<label>1</label></formula><p>Here we informally use the fractional representation, such as</p><formula xml:id="formula_4">∏︀ 𝑎𝑚</formula><p>𝑎 𝑖 , to denote the products of all the terms except for certain ones. Note that this fractional form does not involve any actual division. Also, each secret-shared term of [ ∏︀ 𝑎𝑚 𝑎 𝑖 ...𝑎 𝑗 ] could be found in the auxiliary sets {𝒜𝑖} 𝑛 𝑖=1 , which is preshared across all parties.</p><p>Adaptation of Equation 1 in secret-sharing scheme is as follows:</p><formula xml:id="formula_5">[ 𝑛 ∏︁ 𝑖=1 𝑥𝑖] = ∏︁ 𝛿𝑖 + ∑︁ 𝑖 𝛿𝑖[ ∏︀ 𝑎𝑚 𝑎𝑖 ] + ∑︁ 𝑖,𝑗,𝑖̸ =𝑗 𝛿𝑖𝛿𝑗[ ∏︀ 𝑎𝑚 𝑎𝑖𝑎𝑗 ] + • • • + [ ∏︁ 𝑎𝑖].</formula><p>(2) Since Equation 2 is linear to the secret-sharing terms, all communications could be conducted in parallel, i.e. in a single round of communications. In this case, we could simply reveal all the secret-sharing terms in {𝒜𝑖} 𝑛 𝑖=1 and compute the sharing of final results in constant complexity. The protocol is formally described in Algorithm 2. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Algorithm 2 Multivariate Beaver Multiplication of 𝑛 inputs</head><formula xml:id="formula_6">Mul([𝑥1], [𝑥2], . . . , [𝑥𝑛]) Input: Secret-shared inputs {[𝑥𝑖]} 𝑛 𝑖=1 , auxiliary sets {𝒜𝑖} 𝑛 𝑖=1 . Output: [ ∏︀ 𝑥𝑖]. 1: ◁ Compute masked values 2: for 𝑖 ∈ [1, 𝑛] do 3: [𝛿𝑖] ← [𝑥𝑖 − 𝑎𝑖] = [𝑥𝑖] − [𝑎𝑖]</formula><formula xml:id="formula_7">∏︀ 𝑎𝑚 𝑎 𝑖 𝑎 𝑗 ] + • • • + [ ∏︀ 𝑎𝑖]</formula><p>The total rounds of communications are indeed reduced from 𝑛−1 to constant 1 when a regular 𝑛-ary multiplication is performed, but the overall size of communication data increases from linear to exponential. In a naïve implementation, the data size of 𝑛-ary multiplication is only 3(𝑛 − 1) for a total transmission of 𝑛 − 1 Beaver triples. As opposed to a multivariate implementation, it is 2 𝑛 − 1 to transmit the auxiliary sets {𝒜𝑖} 𝑛 𝑖=1 . Therefore, in practice, there is a trade-off between communication latency and throughput.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Univariate Polynomials</head><p>The formal form of univariate polynomials is defined as 𝑃 (𝑥) = ∑︀ 𝑛 𝑖=0 𝑏𝑖𝑥 𝑖 , where 𝑏𝑖 refers to the coefficients of the degree-𝑖 term. The use of univariate polynomials enables efficient evaluation and manipulation of polynomial expressions. According to Damgård et al. <ref type="bibr" target="#b16">[17]</ref>, we can compute all required [𝑥 𝑖 ] in parallel using multivariate multiplications, then multiply them with corresponding plaintext coefficients. Despite its benefits, this trick has the disadvantage of exponentially increasing the size of transmitted data, which becomes unbearable when the exponent exceeds 5.</p><p>This method can be implemented in practice by computing a tuple of base terms and then multiplying the tuple by a certain term iteratively, as in the exponentiating by squaring method or the fast modulo algorithm. In other words, a tuple 𝑔 = (1, 𝑥, . . . , 𝑥 𝑚−1 ) of size ‖𝑔‖ = 𝑚 could be multiplied by 𝑥 ‖𝑔‖ repeatedly to iterate all the 𝑥 𝑖 terms. The overview of the implementation of univariate polynomials is described in Algorithm 3. Note that 𝑏𝑠:𝑒 is the subvector of 𝑏 from position 𝑠 to 𝑒. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Nonlinear Approximations</head><p>In this section, we take one step further to optimize the commonly used nonlinear functions by leveraging the property of parallelization of our proposed multivariate multiplication.</p><p>Exponentiation. Since exponential functions grow in geometrical speed, approximations based on series expansion generally suffer from a significant reduction in accuracy since we do not know the exact value of the input. Consequently, we resort to the naive iterative approximation, which is capable of utilizing multivariate multiplication effectively:</p><formula xml:id="formula_8">𝑒 [𝑥] = lim 𝑛→∞ (︂ 1 + [𝑥] 𝑑 𝑛 )︂ 𝑑 𝑛 .</formula><p>During each iteration, the 𝑑-th power of the previous result is calculated. With increasing iteration rounds 𝑛, the answer would become closer to the actual results.</p><p>Logarithm. The calculation of logarithms relies on the higher-order iterative methods for a better convergence, i.e. Householder methods on 𝑦 = ln 𝑥:</p><formula xml:id="formula_9">[ℎ𝑛] = 1 − [𝑥]𝑒 −[𝑦𝑛] [𝑦𝑛+1] = [𝑦𝑛] − ∞ ∑︁ 𝑘=1 1 𝑘 [ℎ 𝑘 𝑛 ]</formula><p>Note that the implementation of logarithm is the combination of exponentiation and univariate polynomials. The degree of the polynomials determines the precision of the output.</p><p>Reciprocal. The reciprocal function 𝑦 = 1 𝑥 is calculated using the Newton-Raphson method with an initial guess 𝑦0:</p><formula xml:id="formula_10">[𝑦𝑛+1] = [𝑦𝑛](2 − [𝑥][𝑦𝑛]) = 2[𝑦𝑛] − [𝑥][𝑦𝑛][𝑦𝑛]</formula><p>Trigonometry. Trigonometric functions could be treated as the special case of exponentiation with 𝑑 = 2. The sine and cosine functions are calculated in the field of complex numbers:</p><formula xml:id="formula_11">[sin 𝑥] = Im([𝑒 𝑖𝑥 ]) [cos 𝑥] = Re([𝑒 𝑖𝑥 ])</formula><p>Using the above-mentioned nonlinear functions, we can calculate most of the existing loss functions in deep learning, such as the sigmoid, tanh, and cross-entropy functions. Various other common nonlinear functions, such as the softmax function and kernel function, can also be calculated using exponential and reciprocal functions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Communication Coalescing</head><p>The key to achieving low-latency secret-sharing computation is to reduce the total number of rounds of communications among different parties. While we introduce a latency-friendly implementation of basic math operations, other kinds of communications, such as precision checking, still require an additional but independent communication round.</p><p>In general, the communication involved in multiple math operations could be abstracted as a communication graph, or strictly, as a communication tree. Accordingly, we observe some independent communications that do not affect downstream results can be deferred and combined into one single round of communication. This process is referred to as communication coalescing, and it eliminates unnecessary rounds of communication and improves the utilization of network bandwidth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Security Analysis</head><p>The correctness of the multivariate multiplication is trivial based on the observation in Equation 1 and 2. As univariate polynomials are implemented using the same method as the extended fast modulo algorithm, their effectiveness could also be demonstrated by the correctness and security of multivariate multiplication. Coalescing mechanisms only alter the order of communication rounds without modifying the payload, which is also reliable and secure.</p><p>Multivariate computations are similarly secure as traditional Beaver multiplications under semi-honest conditions. It is intuitively obvious that since 𝑎𝑖 is chosen at random by TTP, the 𝛿𝑖 = 𝑥𝑖 − 𝑎𝑖 value is indistinguishable from a random number. Consequently, the disclosure of [𝛿𝑖] does not reveal any critical information regarding 𝑥𝑖. This assumption holds even if multiple parties, except for the TTP, collude.</p><p>To clarify the security of multivariate multiplication formally, we denote [𝑥]𝑝 as the secret share of 𝑥 for party 𝑝 ∈ 𝒫. The global equations of the multivariate system are as follows:</p><p>∑︁</p><formula xml:id="formula_12">𝑝∈𝒫 [𝑥𝑖]𝑝 = 𝑥𝑖 ∑︁ 𝑝∈𝒫 [𝑎𝑖]𝑝 = 𝑎𝑖 ∑︁ 𝑝∈𝒫 [𝑎𝑖𝑎𝑗]𝑝 = 𝑎𝑖𝑎𝑗 . . . . . . ∑︁ 𝑝∈𝒫 [𝑎1 . . . 𝑎𝑛] = 𝑎1 . . . 𝑎𝑛 [𝑥𝑖]𝑝 − [𝑎𝑖]𝑝 = [𝛿𝑖]𝑝<label>(3)</label></formula><p>with known [𝛿𝑖]𝑝 for every 𝑝 ∈ 𝒫 to each party. From each party's view, these 2 𝑛 + 2𝑛 − 1 equations have Θ(2 𝑛 ‖𝒫‖) unknown variables. This indicates the difficulty in determining the exact value of 𝑥𝑖, as shown in <ref type="bibr" target="#b17">[18]</ref>. A party's view represents all the values it can obtain during its execution. Then the following theorem holds: This guarantees the security of multivariate multiplication by ensuring the indistinguishability between the random distribution and the view's distribution.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Experiments</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Experimental Setup</head><p>As part of our proposed methodology, we use CrypTen <ref type="bibr" target="#b4">[5]</ref> as the basic MPC deep learning framework, which has already provided naïve implementations of secret-sharingbased computations. In most of our experiments, we use 3-party MPC on CPUs. Additionally, we allow a maximum of 4-ary multiplication as stated in Section 4.1, and we set 𝑑 = 3 for exponentiation and 𝑘 = 8 for logarithm as described in Section 4.3.</p><p>To measure the performance, we perform several experiments with deep learning models with different sizes: (a) Linear Support Vector Classification (LinearSVC) with L2 penalty; (b) LeNet <ref type="bibr" target="#b18">[19]</ref> with shallow convolutional and linear layers along with ReLU activation functions; (c) ResNet-18 model <ref type="bibr" target="#b19">[20]</ref> with multiple convolutional, linear, pooling, and activation layers; (d) Transformer Encoder model <ref type="bibr" target="#b20">[21]</ref> with a single multi-head attention layer and BatchNorm <ref type="bibr" target="#b21">[22]</ref> in place of LayerNorm <ref type="bibr" target="#b22">[23]</ref>. We employ several datasets for classification tasks with appropriate adaption to specific models, including MNIST <ref type="bibr" target="#b18">[19]</ref>, CIFAR-10 <ref type="bibr" target="#b23">[24]</ref>, ImageNet <ref type="bibr" target="#b24">[25]</ref>, and Sentiment140 <ref type="bibr" target="#b25">[26]</ref> datasets. Each of our experiments is conducted in a simulated multinode environment using Docker. TTP is conducted in an independent environment separate from the normal parties. To manually simulate different network environments concerning bandwidth and latency, we utilize the docker-tc tool to adjust the docker network settings accordingly.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Metrics</head><p>To provide a comprehensive evaluation of our proposed method, we adopt metrics from a variety of perspectives.</p><p>• 𝑡comp: The computational time cost for evaluating a single data sample in one round.</p><p>• 𝑡comm: The time cost of communication associated with evaluating a single data sample in one round.</p><p>• Size of Transmission Data: The size of transmitted network packets when evaluating a single data sample in one round.</p><p>• Accuracy: The classification accuracy when evaluated on a particular dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3.">Latency &amp; Throughput</head><p>To assess the efficiency of our proposed method, we simulate networks with different network latencies: (a) network 𝑁 low with 0.1ms latency, (b) network 𝑁 med with 5ms latency, (c) and network 𝑁 high with 40ms latency. All of these networks have a bandwidth of 1Gbps. Our simulated multinode settings include 3 nodes with an additional TTP by default.</p><p>As shown in Table <ref type="table" target="#tab_2">1</ref>, the computation cost of each model is negligible in medium and high latency network settings in comparison to the communication cost. Therefore, we will focus only on the communication costs associated with our proposed method.</p><p>Compared to the naïve method implemented by CrypTen, our method illustrated in Section 4.1 remains close since it does not introduce a substantial amount of additional communication payload if the maximum number of input variables is set appropriately. For instance, a 3-ary or 4-ary multiplication would not produce a significant increase in the total size of communications.</p><p>It is noteworthy that our proposed method reduces the communication cost in every network setting as compared to the naïve implementation of MPC. Overall, we achieve an improvement of 10 ∼ 20%, which shows significant enhancement in the performance of high-latency environments for practical purposes.</p><p>Furthermore, we observe that our proposed method behaves differently with neural models with different architectures. Figure <ref type="figure" target="#fig_3">1</ref> illustrates the communication occupation  As can be seen, the attention mechanism is constrained by its communication bottleneck in Softmax operation, while CNN is constrained by its communication via convolutional operations. Considering that our method makes an improved optimization for nonlinear functions, attentionbased models show a significant improvement in latency, with almost a 25% improvement. Additionally, this explains the limited improvement of only 8 ∼ 15% in traditional machine-learning models and CNN-based models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.4.">Evaluation</head><p>In this section, we examine the side effects and factors associated with the basic settings, such as the downstream tasks' accuracy, the number of parties involved, and the trade-off between network latency and bandwidth.</p><p>To evaluate the drop in accuracy, we compare our method with both the original baseline and the naïve implementation without a low-latency design. Figure <ref type="figure" target="#fig_5">2</ref> shows that, in relatively small scenarios, both the naïve implementation and our methods are capable of achieving perfect performance as the baselines. Nevertheless, both MPC-based implementations obtain lower accuracy in complex scenarios than the baseline, while our methods perform slightly worse than the naïve implementation. We hypothesize that the multivariate multiplication introduces additional precision requirements, which in turn reduces accuracy.</p><p>The throughput and latency of MPC-based methods are also affected by the number of parties involved in the computation. From Figure <ref type="figure" target="#fig_5">2</ref>, it can be seen that the communication data size of both methods increases linearly as the number of parties involved increases. There is, however, a tendency for the latency to be worse when there are more parties involved. Moreover, Figure <ref type="figure" target="#fig_6">3</ref> illustrates how network bandwidth affects communication costs. When sufficient bandwidth is available, our method can still optimize the network latency. It is important to note, however, that when the bandwidth becomes the bottleneck, our method would not be any more effective in reducing the overall costs of communication. This indicates that bandwidth remains an important factor in a multi-node MPC setting, especially as the number of nodes in use grows.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Discussion</head><p>Since the proposed multivariate multiplication is based on a finite ring, it is likely to have precision issues that lead to incorrect results. Fortunately, a loss in precision would not significantly affect the overall performance of deep learning, since the loss could be interpreted as random noise and Moreover, our proposed method is only applicable to functions that are based on linear MPC operation. To avoid heavy communications, a modern MPC-based deep learning framework would also involve other protocols, such as Homomorphic Encryption <ref type="bibr" target="#b26">[27]</ref>, Garbled Circuit <ref type="bibr" target="#b27">[28]</ref>, and Function Secret Sharing <ref type="bibr" target="#b28">[29]</ref>. Though these works may have less communication, our approach could still be seamlessly integrated with the current secret-sharing framework and achieve a latency improvement of ~20% without adding excessive computational workload.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Conclusion</head><p>This study proposes a secret-sharing-based MPC method for enhancing the linear computation required in deep learning through increased communication utilization. By utilizing the multivariate multiplication and communication coalescing mechanisms, we can reduce the number of unnecessary communication rounds during the execution of both linear and nonlinear deep learning functions. In our experiments, we demonstrate that our proposed methods achieve an overall improvement in latency of 10 ∼ 20% when compared to the naïve MPC implementation. Additionally, it indicates that throughput and downstream task performance are comparable to naïve implementations, which demonstrate the method's validity and efficiency. We hope that this work will inspire future improvements in privacy-preserving deep learning techniques and lead to more practical MPC applications.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>.</head><label></label><figDesc>The sum of two secret shared values [𝑥] and [𝑦] could be directly computed as [𝑧] = [𝑥] + [𝑦], where each party 𝑝 ∈ 𝒫 computes [𝑧]𝑝 = [𝑥]𝑝 + [𝑦]𝑝 without multi-party communications. Multiplication. Two secret shared values [𝑥] and [𝑦] are multiplied using a random Beaver triple [11] generated by the Trusted Third Party (TTP): a triplet ([𝑎], [𝑏], [𝑎𝑏]). It should be noted that the Beaver triple could be shared in advance by each party. The parties first calculate [𝜖] = [𝑥] − [𝑎] and [𝛿] = [𝑦] − [𝑏]. In this way, the [𝜖] and [𝛿] are then revealed to all parties (denoted as Reveal(•)) without compromising information since the ground-truth values 𝑎, 𝑏 remain unknown to each party except for the TTP. The final results could be computed as [𝑥𝑦] = [𝑐] + 𝜖[𝑏] + [𝑎]𝛿 + 𝜖𝛿. Algorithm 1 illustrates the multiplication using Beaver triples.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>and so on. Similar to the construction of [𝜖] and [𝛿] in Section 2.2, we define the difference between the inputs and the masks as [𝛿𝑖] := [𝑥𝑖] − [𝑎𝑖]. The secretshared [𝛿𝑖] is then made public across all parties without leakage to the ground-truth value of both [𝑥𝑖] and [𝑎𝑖]. The improvement of our method originates from the following equation:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Algorithm 3 ◁ Vectorized Beaver Multiplication 13 :</head><label>313</label><figDesc>Univariate Polynomial Poly([𝑥], 𝑏) Input: Secret-shared input [𝑥], coefficients 𝑏 = (𝑏0, 𝑏1, . . . , 𝑏𝑛), base terms size ‖𝑔‖. Output: ∑︀ 𝑛 𝑖=0 𝑏𝑖[𝑥 𝑖 ]. 1: ◁ Construct base terms 2: parallel for 𝑖 ∈ [1, ‖𝑔‖] do 3: [𝑥 𝑖 ] ← Mul([𝑥], . . . , [𝑥]) ◁ multiplied by [𝑥] of 𝑖 times 4: end parallel for 5: 𝑔 ← (1, [𝑥], . . . , [𝑥 ‖𝑔‖−1 ]) 6: ◁ Iteratively exponentiating 7: 𝑡 ← 0 8: for 𝑖 ∈ [0, ⌊ 𝑛 ‖𝑔‖ ⌋ − 1] do 9: s ← 𝑖 • ‖𝑔‖ 10: e ← (𝑖 + 1) • ‖𝑔‖ 11: 𝑡 ← 𝑡 + 𝑏𝑠:𝑒 • 𝑔 12: 𝑔 ← [𝑥 ‖𝑔‖ ] • 𝑔 14: end for 15: return 𝑡</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Theorem 1 .</head><label>1</label><figDesc>Let {𝑥 ′ 𝑖 } and {𝑥 ′′ 𝑖 } be random values. The distribution of the view of each party is identical when 𝑥𝑖 = 𝑥 ′ 𝑖 or 𝑥𝑖 = 𝑥 ′′ 𝑖 .</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Communication percentage of different models.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Transimission data and latency of naïve and our proposed methods when a different number of parties are involved. The experiment is conducted using the LeNet model on CIFAR-10 using a medium latency network.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Latency of naïve and our proposed methods concerning different network bandwidth. The experiment is conducted using the LeNet model on CIFAR-10 using a medium latency network.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 1</head><label>1</label><figDesc>Communication Latency and Data with Different Network Settings. Latency and data are measured in milliseconds and MiBs.</figDesc><table><row><cell>Model</cell><cell>Dataset</cell><cell>𝑡comp</cell><cell cols="2">Data Size</cell><cell cols="2">𝑡comm (𝑁 low )</cell><cell cols="2">𝑡comm (𝑁 med )</cell><cell cols="2">𝑡comm (𝑁 high )</cell></row><row><cell></cell><cell></cell><cell></cell><cell>Naïve</cell><cell>Ours</cell><cell>Naïve</cell><cell>Ours</cell><cell>Naïve</cell><cell>Ours</cell><cell>Naïve</cell><cell>Ours</cell></row><row><cell>LinearSVC</cell><cell>MNIST</cell><cell>0.032</cell><cell>0.022</cell><cell>0.024</cell><cell>0.082</cell><cell>0.059</cell><cell>0.604</cell><cell>0.568</cell><cell>4.339</cell><cell>4.026</cell></row><row><cell>LeNet</cell><cell>CIFAR-10</cell><cell>0.900</cell><cell>38.562</cell><cell>41.647</cell><cell>1.373</cell><cell>1.182</cell><cell>7.673</cell><cell>6.983</cell><cell>53.010</cell><cell>47.868</cell></row><row><cell>ResNet-18</cell><cell>ImageNet</cell><cell>110.447</cell><cell cols="2">11571.294 12612.711</cell><cell cols="5">219.541 185.293 1681.685 1407.570 ~11 760</cell><cell>~9 870</cell></row><row><cell cols="2">Transformer Sentiment140</cell><cell>7.824</cell><cell>771.787</cell><cell>893.421</cell><cell>12.190</cell><cell>9.715</cell><cell>78.624</cell><cell>61.877</cell><cell cols="2">559.645 421.413</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Classification Accuracy using Different Methods.</figDesc><table><row><cell>Model</cell><cell>Dataset</cell><cell cols="2">Accuracy (%)</cell><cell></cell></row><row><cell></cell><cell></cell><cell cols="2">Origin Naïve</cell><cell>Ours</cell></row><row><cell>LinearSVC</cell><cell>MNIST</cell><cell>100.00</cell><cell>100.00</cell><cell>100.00</cell></row><row><cell>LeNet</cell><cell>CIFAR-10</cell><cell>100.00</cell><cell>100.00</cell><cell>100.00</cell></row><row><cell>ResNet-18</cell><cell>ImageNet</cell><cell>69.30</cell><cell>61.58</cell><cell>60.10</cell></row><row><cell cols="2">Transformer Sentiment140</cell><cell>59.87</cell><cell>58.55</cell><cell>57.74</cell></row><row><cell>Conv2D (49.8%)</cell><cell></cell><cell>Softmax (60.5%)</cell><cell></cell><cell></cell></row><row><cell></cell><cell>Other (6.1%)</cell><cell></cell><cell></cell><cell>Other</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>(2.0%)</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>MatMul</cell></row><row><cell>(19.1%) ReLU</cell><cell>(25.0%) BatchNorm</cell><cell>GeLU</cell><cell></cell><cell>(14.4%)</cell></row><row><cell></cell><cell></cell><cell>(23.1%)</cell><cell></cell><cell></cell></row><row><cell cols="2">(a) ResNet Basic Block</cell><cell cols="3">(b) Attention Block</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is supported by the National Key R&amp;D Program of China under grant (No. 2022YFB2703001).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Multiparty computation from somewhat homomorphic encryption</title>
		<author>
			<persName><forename type="first">I</forename><surname>Damgård</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pastro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Smart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zakarias</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-32009-5_38</idno>
	</analytic>
	<monogr>
		<title level="m">CRYPTO 2012</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Safavi-Naini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Canetti</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="643" to="662" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Mp-spdz: A versatile framework for multiparty computation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Keller</surname></persName>
		</author>
		<idno type="DOI">10.1145/3372297.3417872</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, CCS &apos;20</title>
				<meeting>the 2020 ACM SIGSAC Conference on Computer and Communications Security, CCS &apos;20<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1575" to="1590" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Privacy and security issues in deep learning: A survey</title>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Vasilakos</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2020.3045078</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="4566" to="4593" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Polymath: Low-latency mpc via secure polynomial evaluations and its applications</title>
		<author>
			<persName><forename type="first">D</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kate</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Maji</surname></persName>
		</author>
		<idno type="DOI">10.2478/popets-2022-0020</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings on Privacy Enhancing Technologies</title>
				<meeting>on Privacy Enhancing Technologies</meeting>
		<imprint>
			<date type="published" when="2021">2022. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Crypten: Secure multiparty computation meets machine learning</title>
		<author>
			<persName><forename type="first">B</forename><surname>Knott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Venkataraman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hannun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sengupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ibrahim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Van Der Maaten</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2109.00984</idno>
	</analytic>
	<monogr>
		<title level="j">NIPS</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="4961" to="4973" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">CryptGPU: Fast privacy-preserving machine learning on the gpu</title>
		<author>
			<persName><forename type="first">S</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Knott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1109/SP40001.2021.00098</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE S&amp;P</title>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Mpcformer: fast, performant and private transformer inference with mpc</title>
		<author>
			<persName><forename type="first">D</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Shao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Xing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2211.01452</idno>
	</analytic>
	<monogr>
		<title level="m">The 11th International Conference on Learning Representations</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Falcon: Honest-majority maliciously secure framework for private deep learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wagh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tople</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Benhamouda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kushilevitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mittal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rabin</surname></persName>
		</author>
		<idno type="DOI">10.2478/popets-2021-0011</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings on Privacy Enhancing Technologies</title>
				<meeting>on Privacy Enhancing Technologies</meeting>
		<imprint>
			<date type="published" when="2020">2021. 2020</date>
			<biblScope unit="page" from="188" to="208" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Arithmetic tuples for mpc</title>
		<author>
			<persName><forename type="first">T</forename><surname>Krips</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Küsters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Reisert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rivinius</surname></persName>
		</author>
		<ptr target="https://eprint.iacr.org/2022/667" />
	</analytic>
	<monogr>
		<title level="j">Cryptology ePrint Archive</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">How to share a secret</title>
		<author>
			<persName><forename type="first">A</forename><surname>Shamir</surname></persName>
		</author>
		<idno type="DOI">10.1145/359168.359176</idno>
	</analytic>
	<monogr>
		<title level="j">Commun. ACM</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="612" to="613" />
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Efficient multiparty protocols using circuit randomization</title>
		<author>
			<persName><forename type="first">D</forename><surname>Beaver</surname></persName>
		</author>
		<idno type="DOI">10.1007/3-540-46766-1_34</idno>
	</analytic>
	<monogr>
		<title level="m">CRYPTO 1991</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Feigenbaum</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="420" to="432" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Randomizing polynomials: A new representation with applications to roundefficient secure computation</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ishai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kushilevitz</surname></persName>
		</author>
		<idno type="DOI">10.1109/SFCS.2000.892118</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings 41st Annual Symposium on Foundations of Computer Science</title>
				<meeting>41st Annual Symposium on Foundations of Computer Science</meeting>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="page" from="294" to="304" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Efficient polynomial operations in the shared-coefficients setting</title>
		<author>
			<persName><forename type="first">P</forename><surname>Mohassel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Franklin</surname></persName>
		</author>
		<idno type="DOI">10.1007/11745853_4</idno>
	</analytic>
	<monogr>
		<title level="m">PKC 2006</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Yung</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Dodis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Kiayias</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Malkin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="44" to="57" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Secure efficient multiparty computing of multivariate polynomials and applications</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dachman-Soled</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Malkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Raykova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Yung</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-642-21554-4_8</idno>
	</analytic>
	<monogr>
		<title level="m">Applied Cryptography and Network Security</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Lopez</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Tsudik</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="130" to="146" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<author>
			<persName><forename type="first">S</forename><surname>Wagh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Chandran</surname></persName>
		</author>
		<ptr target="https://eprint.iacr.org/2018/442" />
	</analytic>
	<monogr>
		<title level="m">Securenn: Efficient and private neural network training</title>
		<title level="s">Cryptology ePrint Archive</title>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Aby3: A mixed protocol framework for machine learning</title>
		<author>
			<persName><forename type="first">P</forename><surname>Mohassel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename></persName>
		</author>
		<idno type="DOI">10.1145/3243734.3243760</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, CCS &apos;18</title>
				<meeting>the 2018 ACM SIGSAC Conference on Computer and Communications Security, CCS &apos;18<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="35" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Unconditionally secure constant-rounds multi-party computation for equality, comparison, bits and exponentiation</title>
		<author>
			<persName><forename type="first">I</forename><surname>Damgård</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fitzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kiltz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Nielsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Toft</surname></persName>
		</author>
		<idno type="DOI">10.1007/11681878_15</idno>
	</analytic>
	<monogr>
		<title level="m">Theory of Cryptography</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Halevi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Rabin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="285" to="304" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">A note on the communication complexity of multiparty computation in the correlated randomness model</title>
		<author>
			<persName><forename type="first">G</forename><surname>Couteau</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-17656-3_17</idno>
	</analytic>
	<monogr>
		<title level="m">EUROCRYPT 2019</title>
				<meeting><address><addrLine>Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="473" to="503" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Gradientbased learning applied to document recognition</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bottou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haffner</surname></persName>
		</author>
		<idno type="DOI">10.1109/5.726791</idno>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the IEEE</title>
		<imprint>
			<biblScope unit="volume">86</biblScope>
			<biblScope unit="page" from="2278" to="2324" />
			<date type="published" when="1998">1998</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Deep residual learning for image recognition</title>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR.2016.90</idno>
	</analytic>
	<monogr>
		<title level="m">CVPR</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="770" to="778" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">U</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="DOI">10.5555/3295222.3295349</idno>
		<editor>I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, R. Garnett</editor>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Curran Associates, Inc</publisher>
			<biblScope unit="volume">30</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Batch normalization: accelerating deep network training by reducing internal covariate shift</title>
		<author>
			<persName><forename type="first">S</forename><surname>Ioffe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Szegedy</surname></persName>
		</author>
		<idno type="DOI">10.5555/3045118.3045167</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. 32nd Int. Conf. Machine Learning -Volume</title>
				<meeting>32nd Int. Conf. Machine Learning -Volume</meeting>
		<imprint>
			<publisher>JMLR</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="448" to="456" />
		</imprint>
	</monogr>
	<note>ICML&apos;15</note>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Ba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Kiros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1607.06450</idno>
		<title level="m">Layer normalization</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Krizhevsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
		<title level="m">Learning multiple layers of features from tiny images</title>
				<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Imagenet: A large-scale hierarchical image database</title>
		<author>
			<persName><forename type="first">J</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fei-Fei</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR.2009.5206848</idno>
	</analytic>
	<monogr>
		<title level="m">CVPR</title>
				<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="248" to="255" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Twitter sentiment classification using distant supervision, CS224N project report</title>
		<author>
			<persName><forename type="first">A</forename><surname>Go</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bhayani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Stanford</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<date type="published" when="2009">2009. 2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Fully homomorphic encryption using ideal lattices</title>
		<author>
			<persName><forename type="first">C</forename><surname>Gentry</surname></persName>
		</author>
		<idno type="DOI">10.1145/1536414.1536440</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Forty-First Annual ACM Symposium on Theory of Computing, STOC &apos;09</title>
				<meeting>the Forty-First Annual ACM Symposium on Theory of Computing, STOC &apos;09<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="169" to="178" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Protocols for secure computations</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Yao</surname></persName>
		</author>
		<idno type="DOI">10.1109/SFCS.1982.38</idno>
	</analytic>
	<monogr>
		<title level="m">23rd Annual Symposium on Foundations of Computer Science</title>
				<meeting><address><addrLine>sfcs</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1982">1982. 1982</date>
			<biblScope unit="page" from="160" to="164" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Function secret sharing</title>
		<author>
			<persName><forename type="first">E</forename><surname>Boyle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Gilboa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ishai</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-662-46803-6_12</idno>
	</analytic>
	<monogr>
		<title level="m">EUROCRYPT 2015</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Oswald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Fischlin</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="337" to="367" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
