<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">48115</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2024.048115</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Differentially Private Support Vector Machines with Knowledge Aggregation</article-title>
<alt-title alt-title-type="left-running-head">Differentially Private Support Vector Machines with Knowledge Aggregation</alt-title>
<alt-title alt-title-type="right-running-head">Differentially Private Support Vector Machines with Knowledge Aggregation</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Wang</surname><given-names>Teng</given-names></name></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Zhang</surname><given-names>Yao</given-names></name></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Liang</surname><given-names>Jiangguo</given-names></name></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Wang</surname><given-names>Shuai</given-names></name></contrib>
<contrib id="author-5" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Liu</surname><given-names>Shuanggen</given-names></name><email>liushuanggen201@xupt.edu.cn</email></contrib>
<aff><institution>School of Cyberspace Security, Xi&#x2019;an University of Posts and Telecommunications</institution>, <addr-line>Xi&#x2019;an, 710121</addr-line>, <country>China</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Shuanggen Liu. Email: <email>liushuanggen201@xupt.edu.cn</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2024</year></pub-date>
<pub-date date-type="pub" publication-format="electronic"><day>26</day>
<month>3</month>
<year>2024</year></pub-date>
<volume>78</volume>
<issue>3</issue>
<fpage>3891</fpage>
<lpage>3907</lpage>
<history>
<date date-type="received">
<day>28</day>
<month>11</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>1</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2024 Wang et al.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Wang et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_48115.pdf"></self-uri>
<abstract>
<p>With the widespread data collection and processing, privacy-preserving machine learning has become increasingly important in addressing privacy risks related to individuals. Support vector machine (SVM) is one of the most elementary learning models of machine learning. Privacy issues surrounding SVM classifier training have attracted increasing attention. In this paper, we investigate Differential Privacy-compliant Federated Machine Learning with Dimensionality Reduction, called Fed<sub>DPDR-DPML</sub>, which greatly improves data utility while providing strong privacy guarantees. Considering in distributed learning scenarios, multiple participants usually hold unbalanced or small amounts of data. Therefore, Fed<sub>DPDR-DPML</sub> enables multiple participants to collaboratively learn a global model based on weighted model averaging and knowledge aggregation and then the server distributes the global model to each participant to improve local data utility. Aiming at high-dimensional data, we adopt differential privacy in both the principal component analysis (PCA)-based dimensionality reduction phase and SVM classifiers training phase, which improves model accuracy while achieving strict differential privacy protection. Besides, we train Differential privacy (DP)-compliant SVM classifiers by adding noise to the objective function itself, thus leading to better data utility. Extensive experiments on three high-dimensional datasets demonstrate that Fed<sub>DPDR-DPML</sub> can achieve high accuracy while ensuring strong privacy protection.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Differential privacy</kwd>
<kwd>support vector machine</kwd>
<kwd>knowledge aggregation</kwd>
<kwd>data utility</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>National Natural Science Foundation of China</funding-source>
<award-id>62102311</award-id>
<award-id>62202377</award-id>
<award-id>62272385</award-id>
</award-group>
<award-group id="awg2">
<funding-source>Natural Science Basic Research Program of Shaanxi</funding-source>
<award-id>2022JQ-600</award-id>
<award-id>2022JM-353</award-id>
<award-id>2023-JC-QN-0327</award-id>
</award-group>
<award-group id="awg3">
<funding-source>Shaanxi Distinguished Youth Project</funding-source>
<award-id>2022JC-47</award-id>
</award-group>
<award-group id="awg4">
<funding-source>Shaanxi Provincial Education Department</funding-source>
<award-id>22JK0560</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>The rapid development of generative artificial intelligence and large language models (LLMs) is accelerating changes in our production and living habits [<xref ref-type="bibr" rid="ref-1">1</xref>,<xref ref-type="bibr" rid="ref-2">2</xref>]. As a subfield of artificial intelligence (AI), machine learning (ML) algorithms such as support vector machines and logistic regression can play important roles in text classification, sentiment analysis, information extraction, etc. [<xref ref-type="bibr" rid="ref-3">3</xref>]. However, the proliferation of data collection and training leads to increasing privacy concerns [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-5">5</xref>]. The adversary may snoop on users&#x2019; sensitive information through membership inference attacks, attribute inference attacks, or model inversion attacks [<xref ref-type="bibr" rid="ref-6">6</xref>,<xref ref-type="bibr" rid="ref-7">7</xref>], which leads to privacy breaches, identity theft, or other malicious activities.</p>
<p>Privacy-preserving machine learning (PPML) [<xref ref-type="bibr" rid="ref-4">4</xref>] addresses these concerns by allowing the training and inference processes to be performed without exposing the raw data. Support vector machine (SVM) [<xref ref-type="bibr" rid="ref-8">8</xref>] is one of the most elementary learning models. Therefore, there is a huge demand for studying privacy-preserving SVM algorithms. Differential privacy (DP) [<xref ref-type="bibr" rid="ref-9">9</xref>,<xref ref-type="bibr" rid="ref-10">10</xref>] is a rigorous privacy paradigm nowadays and is widely adopted in AI and ML. DP has a formal mathematical foundation and therefore prevents the disclosure of any information about the presence or absence of any individual from any statistical operations.</p>
<p>Several approaches have been proposed to train SVM models with differential privacy [<xref ref-type="bibr" rid="ref-11">11</xref>&#x2013;<xref ref-type="bibr" rid="ref-13">13</xref>]. These methods typically add noise or perturbation to the training data or model parameters to limit the amount of information that can be learned about any individual data point. Due to the lack of dimensionality reduction considerations, both computation overhead and accuracy are restricted by the curse of dimensionality. Dwork et al. [<xref ref-type="bibr" rid="ref-14">14</xref>] first studied the problem of privacy-preserving principal component analysis (PCA) and proved the optimal bounds of DP-compliant PCA, which lays the foundation for applying PCA in PPML.</p>
<p>Hereafter, Huang et al. [<xref ref-type="bibr" rid="ref-15">15</xref>] leveraged the Laplace mechanism into PCA-SVM algorithms to achieve differential privacy protection. Sun et al. [<xref ref-type="bibr" rid="ref-16">16</xref>] proposed DPSVD which is a differentially private singular value decomposition (SVD) algorithm to provide privacy guarantees for SVM training. To sum end, these methods all consider achieving dimensionality reduction by using PCA, so the algorithms are usually divided into two stages: the PCA phase and the SVM phase. However, the DPPCA-SVM, PCA-DPSVM in [<xref ref-type="bibr" rid="ref-15">15</xref>], and DPSVD in [<xref ref-type="bibr" rid="ref-16">16</xref>] all apply differential privacy in only one stage of PCA or SVM, resulting in an insufficient degree of privacy protection. A strict differential privacy protection mechanism should satisfy that DP must be applied whenever the train data is accessed in the algorithm [<xref ref-type="bibr" rid="ref-10">10</xref>]. Therefore, a DP-compliant SVM training mechanism with dimensionality reduction should be further studied.</p>
<p>Besides, when considering distributed learning scenarios, a common challenge is that multiple parties often have unbalanced or small amounts of data, resulting in inaccurate model accuracy. Hopefully, federated learning [<xref ref-type="bibr" rid="ref-17">17</xref>,<xref ref-type="bibr" rid="ref-18">18</xref>] is proposed to solve this problem through federated averaging [<xref ref-type="bibr" rid="ref-19">19</xref>] (specifically, including model averaging [<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>] and gradient averaging [<xref ref-type="bibr" rid="ref-22">22</xref>,<xref ref-type="bibr" rid="ref-23">23</xref>]). The existing DP-based SVM training mechanisms almost focus on centralized settings and do not take federated learning into account. Some SVM training frameworks based on federated learning [<xref ref-type="bibr" rid="ref-24">24</xref>,<xref ref-type="bibr" rid="ref-25">25</xref>] are mainly based on encryption technology rather than differential privacy, resulting in high computational overhead. Intuitively, we can directly adopt model averaging to obtain global information in distributed training settings. However, this ignores the fact that different data owners have different contributions to the global model.</p>
<p>To this end, this paper studies a strict differentially private SVM algorithm with dimensionality reduction and knowledge aggregation, which aims to maintain high data utility while providing strong privacy protection. Furthermore, considering that data among participants may be small and uneven, this paper focuses on the collaborative training of a global machine learning model by multiple participants. Our main contributions are summarized as follows:
<list list-type="bullet">
<list-item>
<p>We propose Fed<sub>DPDR-DPML</sub>, a federated machine learning framework incorporating dimensionality reduction and knowledge aggregation, which greatly improves data utility while providing strong privacy guarantees. Fed<sub>DPDR-DPML</sub> enables multiple participants to collaboratively learn a global model based on weighted model averaging and then the server distributes the global model to each participant to improve local data utility.</p></list-item>
<list-item>
<p>We design a strict privacy-preserving machine learning mechanism DPDR-DPML which introduces DP in both the dimensionality reduction phase and SVM training phase to provide strict and strong privacy guarantees. Specifically, we leverage a DP-based principal component analysis (PCA) method to extract the key low-dimensional features from high-dimensional data, which reduces computation costs and improves model accuracy.</p></list-item>
<list-item>
<p>By leveraging the empirical risk minimization approximations, we train DP-compliant SVM classifiers by adding noise to the objective function itself, leading to better data utility.</p></list-item>
<list-item>
<p>We conduct extensive experiments on three high-dimensional datasets. The experimental results demonstrate that our mechanisms achieve high accuracy while ensuring strong privacyprotection.</p></list-item>
</list></p>
<p>The remainder of the paper is organized as follows. A literature review is provided in <xref ref-type="sec" rid="s2">Section 2</xref>. <xref ref-type="sec" rid="s3">Section 3</xref> introduces preliminaries and research problems. We present our solution Fed<sub>DPDR-DPML</sub> in <xref ref-type="sec" rid="s4">Section 4</xref>. <xref ref-type="sec" rid="s5">Section 5</xref> shows the experimental results and <xref ref-type="sec" rid="s6">Section 6</xref> concludes the paper.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>Privacy-preserving machine learning (PPML) [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-26">26</xref>,<xref ref-type="bibr" rid="ref-27">27</xref>] enables data-driven decision-making and the development of intelligent systems while protecting individuals&#x2019; sensitive information. Since the introduction of differential privacy (DP) [<xref ref-type="bibr" rid="ref-9">9</xref>,<xref ref-type="bibr" rid="ref-10">10</xref>], DP-based PPML [<xref ref-type="bibr" rid="ref-28">28</xref>] has gained significant attention as a means to ensure privacy while training models on sensitive data. Support vector machine (SVM) [<xref ref-type="bibr" rid="ref-8">8</xref>,<xref ref-type="bibr" rid="ref-29">29</xref>] is a popular class of machine learning algorithm used for classification, regression, and outlier detection tasks. Differential privacy (DP) is widely adopted in SVM to provide privacy guarantees for sensitive data.</p>
<p>However, a serious challenge facing SVM model learning under DP is how to achieve a good trade-off between privacy and utility. To this end, Chaudhuri et al. [<xref ref-type="bibr" rid="ref-30">30</xref>] proposed to produce privacy-preserving approximations of classifiers learned via (regularized) empirical risk minimization (ERM). They also analyzed the accuracy of proposed mechanisms and the upper bound of the number of training samples, laying the foundation for subsequent research. Zhang et al. [<xref ref-type="bibr" rid="ref-11">11</xref>] first proposed a dual variable perturbation scheme for differentially private SVM classifier training, which improves prediction accuracy. Farokhi [<xref ref-type="bibr" rid="ref-12">12</xref>] introduced additive privacy-preserving noise when conducting DP-based SVM training, which is proved as the optimal privacy-preserving noise distribution. Besides, Chen et al. [<xref ref-type="bibr" rid="ref-13">13</xref>] focused on privacy-preserving multi-class SVM training on medical diagnosis, which can deal with both linearly separable data and nonlinear data. However, these works do not consider dimensionality reduction, which will lead to higher computational overhead and lower classification accuracy when directly applied to high-dimensional data.</p>
<p>To address this, Dwork et al. [<xref ref-type="bibr" rid="ref-14">14</xref>] first studied the problem of differential privacy-based principal component analysis (PCA) and proved the optimal bounds of DP-compliant PCA, which lays the foundation for applying PCA in DP-based SVM model learning. They proposed to perturb the matrix of covariance with Gaussian noise. In contrast, Jiang et al. [<xref ref-type="bibr" rid="ref-31">31</xref>] perturbed the matrix of covariance with Wishart noise, which was able to output a perturbed positive semi-definite matrix. Besides, Xu et al. [<xref ref-type="bibr" rid="ref-32">32</xref>] applied the Laplace mechanism to introduce perturbation and proposed the Laplace input perturbation and Laplace output perturbation. These studies focus on DP-based dimensionality reduction, which provides an important research foundation for DP-based SVM training with dimensionality reduction.</p>
<p>Therefore next, Huang et al. [<xref ref-type="bibr" rid="ref-15">15</xref>] proposed DPPCA-SVM and PCA-DPSVM for privacy-preserving SVM learning with dimensionality reduction, which perturbed the matrix of covariance with symmetric Laplace noise. However, the DPPCA-SVM and PCA-DPSVM mechanisms only apply differential privacy at one stage in PCA or SVM, resulting in an insufficient degree of privacy protection. It should be claimed that a strict differential privacy protection mechanism should satisfy that DP must be applied whenever the train data is accessed in the algorithm. Besides, Sun et al. [<xref ref-type="bibr" rid="ref-16">16</xref>] proposed DPSVD which uses singular value decomposition (SVD) to project the training instances into the low-dimensional singular subspace. They first added the noise to the raw data <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:mi>D</mml:mi></mml:math></inline-formula> and then obtained the singular values by applying SVD on the perturbed data <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>. However, the original training dataset is accessed again when computing low-dimensional singular subspace, thus resulting in insufficient privacy protection.</p>
<p>Federated learning [<xref ref-type="bibr" rid="ref-17">17</xref>,<xref ref-type="bibr" rid="ref-18">18</xref>] is a distributed machine learning framework designed to allow dispersed participants to collaborate on machine learning without disclosing private data to other participants. Tavara et al. [<xref ref-type="bibr" rid="ref-33">33</xref>] used alternating direction method of multipliers to efficiently learn a global SVM model with differential privacy in a distributed manner. Moreover, Truex et al. [<xref ref-type="bibr" rid="ref-24">24</xref>] used an encryption-based federated learning framework to generate a new SVM model based on the received local parameters from different data parties. Meanwhile, they also discussed introducing Gaussian noise to the gradients to achieve differential privacy. However, the article does not consider dimensionality reduction and lacks clear derivation and proof. Xu et al. [<xref ref-type="bibr" rid="ref-25">25</xref>] also studied privacy-preserving federated learning over vertically partitioned data, which can be applied to SVM training. Like [<xref ref-type="bibr" rid="ref-24">24</xref>], Xu et al. also used secure gradient computation to compute the global model, but the difference is that it targets the vertical setting and uses encryption for privacy protection. These studies all achieve differential privacy by adding noise to gradients.</p>
<p>Furthermore, the above studies all adopt federated averaging [<xref ref-type="bibr" rid="ref-19">19</xref>] (including model averaging [<xref ref-type="bibr" rid="ref-21">21</xref>,<xref ref-type="bibr" rid="ref-22">22</xref>] and gradient averaging [<xref ref-type="bibr" rid="ref-22">22</xref>,<xref ref-type="bibr" rid="ref-23">23</xref>]) to obtain the global model parameters in many scenarios. However, the classical federated averaging schemes ignore the contribution degrees of different participants. Thus, this paper investigates and proposes to utilize a weighted model averaging mechanism for collaborative machine learning while satisfying strict differential privacy.</p>
</sec>
<sec id="s3">
<label>3</label>
<title>Preliminaries</title>
<sec id="s3_1">
<label>3.1</label>
<title>System Model and Safety Model</title>
<p>The system model considered in this article is a distributed machine-learning scenario, which contains a central server and multiple participants. This paper considers that the central server obeys the semi-honest (honest but curious) adversary model. That is, the server adheres to the agreement but also tries to learn more from the received information than the output was unexpected. In addition, this paper assumes that the multiple participants adhere to the agreement but do not trust each other.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Differential Privacy</title>
<p>Differential privacy (DP) [<xref ref-type="bibr" rid="ref-9">9</xref>,<xref ref-type="bibr" rid="ref-10">10</xref>] is a strict privacy protection model that gives rigorous and quantified proof of privacy disclosure risk. Since differential privacy was proposed ten years ago, hundreds of papers based on differential privacy technology have been proposed in security, database, machine learning, and statistical computing applications.</p>
<p><bold>Definition 3.1</bold> (<inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-Differential Privacy (<inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP)). A randomized mechanism <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow></mml:math></inline-formula> satisfies <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP if and only if for any neighboring datasets <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:mi>D</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>, and for any possible output <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:mi>O</mml:mi><mml:mo>&#x2286;</mml:mo><mml:mrow><mml:mtext>Range</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, it holds
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:mrow><mml:mi mathvariant="double-struck">P</mml:mi></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>D</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mi>O</mml:mi><mml:mo>]</mml:mo></mml:mrow><mml:mo>&#x2264;</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mi mathvariant="double-struck">P</mml:mi></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>&#x02133;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mi>O</mml:mi><mml:mo>]</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:mrow><mml:mi mathvariant="double-struck">P</mml:mi></mml:mrow></mml:math></inline-formula> denotes probability.</p>
<p><inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP is also called approximated DP. When <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP becomes <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>-DP, that is, pure differential privacy. The neighboring datasets <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mi>D</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula> are considered to be neighboring if they differ by a single record.</p>
<p>Differential privacy provides a mathematical guarantee of privacy by introducing controlled randomness (i.e., noise) into the data or results of computations. This paper adopts the Gaussian mechanism [<xref ref-type="bibr" rid="ref-10">10</xref>] to achieve differential privacy, which is defined as follows.</p>
<p><bold>Theorem 3.1</bold> (Gaussian Mechanism). The Gaussian mechanism achieves <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP by adding Gaussian noise with standard deviation <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi>&#x03C3;</mml:mi><mml:mo>=</mml:mo><mml:msqrt><mml:mn>2</mml:mn><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1.25</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msqrt><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>, where <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow></mml:math></inline-formula> is <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>-sensitivity and is computed as the maximal <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:msub><mml:mi>&#x2113;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>-norm difference of two neighboring datasets <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:mi>D</mml:mi></mml:math></inline-formula> and <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msup><mml:mi>D</mml:mi><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Problem Formulation</title>
<p><bold><italic>Data model</italic>.</bold> Let <italic>N</italic> denote the number of participants. Each participant <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>N</mml:mi><mml:mo>}</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>owns a local dataset <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4B4;</mml:mi></mml:mrow></mml:mrow><mml:mo>&#x003A;</mml:mo><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>n</mml:mi><mml:mo>}</mml:mo></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula> with <italic>n</italic> samples, where <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> in each sample <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denote the data space and label set, respectively. As for binary classification in ML, the data space is <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:mrow><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> and the label set is <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mrow><mml:mrow><mml:mi>&#x1D4B4;</mml:mi></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>. That is, each <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> is a <italic>d</italic>-dimensional vector, and each <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> or <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>. Besides, it assumes <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2264;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> which facilitates the efficient calculation of sensitivity in the following [<xref ref-type="bibr" rid="ref-34">34</xref>,<xref ref-type="bibr" rid="ref-35">35</xref>]. For convenience, let <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> denote the data space of dataset <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and let <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>;</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula> denote the label space of dataset <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. That is, <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
<p><bold><italic>Empirical Risk Minimization (ERM)</italic>.</bold> In this paper, we build machine learning models that are expressed as empirical risk minimization. We would like to train a predictor <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x003A;</mml:mo><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi>y</mml:mi></mml:math></inline-formula>. As for machine learning algorithms with empirical risk minimization, the predictor <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi></mml:math></inline-formula> minimizes the regularized empirical loss. For each participant <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> owning dataset <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the ERM can be formulated as
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula>where <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the loss function, <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is a <italic>d</italic>-dimensional parameter vector.</p>
<p>Moreover, we further introduce structure risk on <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref> as follows:
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mfrac><mml:mi>&#x03BB;</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:math></disp-formula>where <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mi>&#x03BB;</mml:mi><mml:mo>&#x003E;</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula> is a regularization parameter. Here, introducing regularization terms can effectively reduce the risk of overfitting.</p>
<p>Based on <xref ref-type="disp-formula" rid="eqn-3">Eq. (3)</xref>, we aim to compute a <italic>d</italic>-dimensional parameter vector <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> such that
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mfrac><mml:mi>&#x03BB;</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p><bold><italic>Problem Statement</italic>.</bold> For each participant <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, we aim to privately train a machine learning model (i.e., private predictor <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula>) based on ERM on the client side. For the service provider, we aim to privately aggregate all the local models of <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mi>N</mml:mi></mml:math></inline-formula> participants and compute a global ML model <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> on the server side. Besides, we will also integrate dimensionality reduction into all training processes to improve model accuracy and reduce computing costs.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Our Solution</title>
<p>Insufficient data samples and high-dimensional features are some of the key factors restricting small data owners from training high-performance models. Therefore, this paper considers a scenario in which multiple participants collaborate to train a global machine learning model in a privacy-preserving way, which can improve accuracy while providing privacy guarantees. To this end, we propose a differential privacy-compliant federated machine learning framework with dimensionality reduction, called Fed<sub>DPDR-DPML</sub>.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Overview of Fed<sub>DPDR-DPML</sub></title>
<p>The high-level overview of Fed<sub>DPDR-DPML</sub> is shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>. The Fed<sub>DPDR-DPML</sub> mainly includes two phases: the first phase aims to obtain the global low-dimensional features of high-dimensional data, and the second phase aims to obtain the global machine learning model. Specifically, the Fed<sub>DPDR-DPML</sub> adopts three design rationales as follows. 1) To overcome the high-dimensional features of data, we conduct dimensionality reduction before training by using the principal component analysis (PCA) method, which can improve model accuracy and reduce computation overhead. 2) To provide strict privacy guarantees, we introduce differential privacy in both dimensionality reduction and machine learning. 3) To solve the challenges of unbalanced data size among data owners, we leverage weighted averaging for both dimensionality reduction and machine learning procedures to improve the model accuracy.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>The framework of our Fed<sub>DPDR-DPML</sub> mechanism</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-1.tif"/>
</fig>
<p>However, the traditional model averaging [<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>] method can improve the performance of participants who own a small amount of data, but will reduce the performance of participants who own a large amount of data. That is, participants with different amounts of data contribute differently to the global model. Therefore, we propose a weighted model averaging scheme that computes the global information through a weighted average method, in which the weight of each participant depends on the data size it possesses. Let <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> be the data size of the participant <inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Then, the weight of participant <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>n</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<p>Algorithm 1 presents a high-level description of the proposed Fed<sub>DPDR-DPML</sub>. The two phases are described in detail as follows:
<list list-type="bullet">
<list-item>
<p>In the first phase, each participant <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> locally employs a DP-compliant dimensionality reduction (DPDR) algorithm to generate private <italic>k</italic>-dimensional features <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>and sends <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> to the server. The server computes the weighted average of the private <italic>k</italic>-dimensional features as <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac></mml:mstyle><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and returns the global low-dimensional features <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> to each participant. The DPDR satisfies <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP.</p></list-item>
<list-item>
<p>In the second phase, each participant <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> locally executes algorithm DPDR-DPML to get local ML model parameters <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and sends <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> to the server. Next, the server computes the weighted average of the private ML predictor as <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup><mml:mo stretchy="false">&#x2190;</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac></mml:mstyle><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and returns the global machine-learning parameters <inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> to each participant. In addition, as shown in the 5-th line, the raw dataset <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> of each participant will be used when executing DPDR-DPML. To achieve privacy protection, the algorithm DPDR-DPML involves differential privacy again when training local ML models and satisfies <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>-DP.</p></list-item>
</list></p>
<fig id="fig-6">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-6.tif"/>
</fig>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>DP-Compliant Dimensionality Reduction</title>
<p>We utilize principal component analysis (PCA) to achieve dimensionality reduction under DP. For <italic>d</italic>-dimensional dataset <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> of participant <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:mi>d</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>d</mml:mi></mml:math></inline-formula> covariance matrix is defined as
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>Thus, we can achieve DP-compliant PCA by applying the Gaussian mechanism to <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Then, the <italic>k</italic>-principle features of the original dataset are computed by choosing the top-<italic>k</italic> singular subspace of the noised covariance matrix based on singular value decomposition (SVD).</p>
<p>Algorithm 2 shows the pseudo-code of PCA-based dimensionality reduction while satisfying DP. We simply formalize Algorithm 2 as <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:mrow><mml:mtext>DPDR</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Given dataset <inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> of each participant <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, we add Gaussian noise to the covariance matrix to achieve DP. For the function <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, the sensitivity of <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>, as shown in Lemma 4.1. Thus, the Gaussian noise matrix <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:msub><mml:mi>R</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> is generated from <inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:mrow><mml:mrow><mml:mi>&#x1D4A9;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1.25</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:msubsup><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and is processed to be a symmetric matrix by each lower triangle entry copied from its upper triangle counterpart. Next, we apply SVD to the noisy covariance matrix <inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:msub><mml:mrow><mml:mover><mml:mi>M</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and thereby grab the top-<italic>k</italic> singular subspace of <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msub><mml:mrow><mml:mover><mml:mi>M</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, as shown in Lines 5&#x2013;6. Then, <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the private <italic>k</italic>-dimensional features of dataset <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<fig id="fig-7">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-7.tif"/>
</fig>
<p><bold>Lemma 4.1.</bold> In Algorithm 2 (i.e., DPDR), for input dataset <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, the sensitivity of function <inline-formula id="ieqn-114"><mml:math id="mml-ieqn-114"><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is at most one.</p>
<p><bold>Proof.</bold> Let <inline-formula id="ieqn-115"><mml:math id="mml-ieqn-115"><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> denote the neighboring dataset of <inline-formula id="ieqn-116"><mml:math id="mml-ieqn-116"><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Assuming <inline-formula id="ieqn-117"><mml:math id="mml-ieqn-117"><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula id="ieqn-118"><mml:math id="mml-ieqn-118"><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> differ in the <italic>t-</italic>th row. Then, based on the definition of DP, the sensitivity can be computed as
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>M</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msubsup><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mtd><mml:mtd><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2032;</mml:mo><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2032;</mml:mo></mml:mrow></mml:msubsup><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mtable columnalign="center center center center" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22F1;</mml:mo></mml:mtd><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:mo>&#x22EF;</mml:mo></mml:mtd><mml:mtd><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup></mml:mtd></mml:mtr></mml:mtable><mml:mo>]</mml:mo></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:msqrt><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.047em" minsize="2.047em">[</mml:mo></mml:mrow></mml:mstyle><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.047em" minsize="2.047em">]</mml:mo></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mrow><mml:mo>[</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>]</mml:mo></mml:mrow></mml:msqrt></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:msqrt><mml:msup><mml:mrow><mml:mo>[</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>+</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>]</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:msqrt></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:msqrt><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msubsup><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup></mml:msqrt><mml:mo>&#x2264;</mml:mo><mml:mn>1</mml:mn></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>where the step of &#x201C;<inline-formula id="ieqn-119"><mml:math id="mml-ieqn-119"><mml:mo>&#x2264;</mml:mo></mml:math></inline-formula>&#x201D; is achieved since <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:msub><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2264;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula><inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:mrow><mml:mo>(</mml:mo><mml:mi>j</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>n</mml:mi><mml:mo>}</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>DP-Compliant Machine Learning with Dimensionality Reduction</title>
<p>This part presents the DP-compliant machine learning with dimensionality reduction. As a representative, we consider building support vector machine (SVM) models from multiple participants. Specifically, the SVM model is trained based on empirical risk minimization. Moreover, the loss function of SVM is defined as <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:msub><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mtext>SVM</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>,</mml:mo><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mo>,</mml:mo><mml:mi>y</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>y</mml:mi><mml:msup><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msup><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>.</p>
<p>To improve model accuracy, we first apply dimensionality reduction on the original high-dimensional dataset. Besides, to achieve privacy protection, we perturb the objective function to produce the minimizer of the noisy objective function.</p>
<fig id="fig-8">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-8.tif"/>
</fig>
<p>Algorithm 3 shows the pseudo-code of our proposed machine-learning training process under DP. We formalize Algorithm 3 as <inline-formula id="ieqn-142"><mml:math id="mml-ieqn-142"><mml:mrow><mml:mtext>DPDR-DPML</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03BB;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Given the dataset <inline-formula id="ieqn-143"><mml:math id="mml-ieqn-143"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> of participant <inline-formula id="ieqn-144"><mml:math id="mml-ieqn-144"><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, we first project the data into <italic>k</italic>-dimensional space based on the private <italic>k</italic>-dimensional features <inline-formula id="ieqn-145"><mml:math id="mml-ieqn-145"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. The <inline-formula id="ieqn-146"><mml:math id="mml-ieqn-146"><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> can be obtained from Algorithm 2. Therefore, the input dataset for machine learning is <inline-formula id="ieqn-147"><mml:math id="mml-ieqn-147"><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>X</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Next, we compute the privacy parameter which will be used to generate noise for objective function perturbation, as shown in Lines 3&#x2013;9. The <inline-formula id="ieqn-148"><mml:math id="mml-ieqn-148"><mml:mi>h</mml:mi></mml:math></inline-formula> is the Huber loss function parameter and is picked as <inline-formula id="ieqn-149"><mml:math id="mml-ieqn-149"><mml:mi>h</mml:mi><mml:mo>=</mml:mo><mml:mn>0.5</mml:mn></mml:math></inline-formula> for Huber SVM, a typical value [<xref ref-type="bibr" rid="ref-36">36</xref>].</p>
<p>Based on the privacy parameter <inline-formula id="ieqn-150"><mml:math id="mml-ieqn-150"><mml:mi>p</mml:mi></mml:math></inline-formula>, the noise vector <inline-formula id="ieqn-151"><mml:math id="mml-ieqn-151"><mml:msub><mml:mi>R</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> can be drawn based on the probability density function <inline-formula id="ieqn-152"><mml:math id="mml-ieqn-152"><mml:msup><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi>R</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula>. Then, we can perturb the objective function as
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mfrac><mml:mrow><mml:mtext>1</mml:mtext></mml:mrow><mml:mi>n</mml:mi></mml:mfrac><mml:msubsup><mml:mi>R</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>At last, we can produce the minimizer of noisy <inline-formula id="ieqn-153"><mml:math id="mml-ieqn-153"><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> by</p>
<p><disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mfrac><mml:mi>&#x03B8;</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-154"><mml:math id="mml-ieqn-154"><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> is the optimal parameters of <inline-formula id="ieqn-155"><mml:math id="mml-ieqn-155"><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></inline-formula></p>
<p>Based on <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref> in <xref ref-type="sec" rid="s3_3">Subsection 3.3</xref>, the minimizer of <inline-formula id="ieqn-156"><mml:math id="mml-ieqn-156"><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></inline-formula> is computed as</p>
<p><disp-formula id="eqn-9"><label>(9)</label><mml:math id="mml-eqn-9" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mover><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:mtd><mml:mtd><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>priv</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mi>&#x03B8;</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x2131;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>D</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac></mml:mstyle><mml:msubsup><mml:mi>R</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mi>&#x03B8;</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>=</mml:mo><mml:mi>arg</mml:mi><mml:mo>&#x2061;</mml:mo><mml:munder><mml:mo form="prefix">min</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:mo>[</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac></mml:mstyle><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi>&#x2113;</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mi>&#x03BB;</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:mi>n</mml:mi></mml:mfrac></mml:mstyle><mml:msubsup><mml:mi>R</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x22A4;</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mi>&#x03B8;</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mstyle><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03B2;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>where <inline-formula id="ieqn-157"><mml:math id="mml-ieqn-157"><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> denotes the private <italic>k</italic>-dimensional data space of <inline-formula id="ieqn-158"><mml:math id="mml-ieqn-158"><mml:msubsup><mml:mrow><mml:mover><mml:mrow><mml:mtext mathvariant="bold">x</mml:mtext></mml:mrow><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Theoretical Analysis</title>
<sec id="s4_4_1">
<label>4.4.1</label>
<title>Privacy Analysis</title>
<p><bold>Theorem 4.1.</bold> Algorithm 2 (i.e., DPDR) satisfies <inline-formula id="ieqn-159"><mml:math id="mml-ieqn-159"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-differential privacy.</p>
<p><bold>Proof.</bold> As shown in the 4-th line of Algorithm 2, the Gaussian noise <inline-formula id="ieqn-160"><mml:math id="mml-ieqn-160"><mml:msub><mml:mi>R</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> is drawn from <inline-formula id="ieqn-161"><mml:math id="mml-ieqn-161"><mml:mrow><mml:mrow><mml:mi>&#x1D4A9;</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1.25</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:msubsup><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:msubsup><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, that is, the deviation <inline-formula id="ieqn-162"><mml:math id="mml-ieqn-162"><mml:mi>&#x03C3;</mml:mi><mml:mo>=</mml:mo><mml:msqrt><mml:mn>2</mml:mn><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1.25</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msqrt><mml:mo>&#x22C5;</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>, based on Theorem 3.1, Algorithm 2 (i.e., DPDR) satisfies <inline-formula id="ieqn-163"><mml:math id="mml-ieqn-163"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-differential privacy.</p>
<p><bold>Theorem 4.2.</bold> Algorithm 3 (i.e., DPDR-DPML) satisfies <inline-formula id="ieqn-164"><mml:math id="mml-ieqn-164"><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>-differential privacy.</p>
<p><bold>Proof.</bold> The privacy guarantee of objective perturbation is shown in lines 3&#x2013;10, which uses privacy parameter <inline-formula id="ieqn-165"><mml:math id="mml-ieqn-165"><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>. This can be proved to satisfy <inline-formula id="ieqn-166"><mml:math id="mml-ieqn-166"><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>-differential privacy by Theorem 9 in [<xref ref-type="bibr" rid="ref-30">30</xref>]. We omit the details due to space limitations.</p>
<p><bold>Theorem 4.3.</bold> Algorithm 1 (i.e., Fed<sub>DPDR-DPML</sub>) satisfies <inline-formula id="ieqn-167"><mml:math id="mml-ieqn-167"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-differential privacy, where <inline-formula id="ieqn-168"><mml:math id="mml-ieqn-168"><mml:mi>&#x03B5;</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<p><bold>Proof.</bold> As shown in Algorithm 1, Fed<sub>DPDR-DPML</sub> sequentially executes <inline-formula id="ieqn-169"><mml:math id="mml-ieqn-169"><mml:mrow><mml:mtext>DPDR</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-170"><mml:math id="mml-ieqn-170"><mml:mrow><mml:mtext>DPDR-DPML</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>U</mml:mi><mml:mo>&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Global</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03BB;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. Thus, based on Theorem 4.1 and Theorem 4.2, Algorithm 1 (i.e., Fed<sub>DPDR-DPML</sub>) satisfies <inline-formula id="ieqn-171"><mml:math id="mml-ieqn-171"><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>-DP according to the sequential composition theorems [<xref ref-type="bibr" rid="ref-10">10</xref>].</p>
</sec>
<sec id="s4_4_2">
<label>4.4.2</label>
<title>Noise Scale Comparisons</title>
<p><xref ref-type="table" rid="table-1">Table 1</xref> shows the comparisons between our proposed algorithms and other state-of-the-art mechanisms from different perspectives. At first, this paper considers a distributed scenario in which multiple participants jointly train a model, each of which has a different amount of data. In terms of privacy guarantees, our proposed Fed<sub>DPDR-DPML</sub> insists that DP must be applied whenever the train data is accessed in an algorithm. Thus, compared to existing methods, Fed<sub>DPDR-DPML</sub> involves noise addition in both the dimensionality reduction phase (i.e., PCA) and training phase (i.e., SVM), which provides strict and strong privacy protection.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Comparisons between different mechanisms</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Mechanism</th>
<th>System model</th>
<th>Noise addition phase</th>
<th>Noise scale</th>
<th>Noise mechanism</th>
<th>Privacy level</th>
</tr>
</thead>
<tbody>
<tr>
<td>AG [<xref ref-type="bibr" rid="ref-14">14</xref>]</td>
<td>Centralized</td>
<td>PCA</td>
<td><inline-formula id="ieqn-173"><mml:math id="mml-ieqn-173"><mml:mi>O</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mi>d</mml:mi></mml:msqrt><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>&#x03B5;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>Gaussian mechanism</td>
<td><inline-formula id="ieqn-174"><mml:math id="mml-ieqn-174"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td>DPPCA-SVM [<xref ref-type="bibr" rid="ref-15">15</xref>]</td>
<td>Centralized</td>
<td>PCA</td>
<td><inline-formula id="ieqn-175"><mml:math id="mml-ieqn-175"><mml:mi>O</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>d</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>&#x03B5;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>Laplace mechanism</td>
<td><inline-formula id="ieqn-176"><mml:math id="mml-ieqn-176"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td>PCA-DPSVM [<xref ref-type="bibr" rid="ref-15">15</xref>]</td>
<td>Centralized</td>
<td>SVM</td>
<td><inline-formula id="ieqn-177"><mml:math id="mml-ieqn-177"><mml:mi>O</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>d</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi>&#x03B5;</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula></td>
<td>Laplace mechanism</td>
<td><inline-formula id="ieqn-178"><mml:math id="mml-ieqn-178"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td>DPSVD [<xref ref-type="bibr" rid="ref-16">16</xref>]</td>
<td>Centralized</td>
<td>PCA</td>
<td><inline-formula id="ieqn-179"><mml:math id="mml-ieqn-179"><mml:mi>O</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mi>d</mml:mi></mml:msqrt><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>&#x03B5;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>Gaussian mechanism</td>
<td><inline-formula id="ieqn-180"><mml:math id="mml-ieqn-180"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td>Truex et al. [<xref ref-type="bibr" rid="ref-24">24</xref>]</td>
<td>Distributed</td>
<td>SVM</td>
<td>/</td>
<td>Gaussian mechanism</td>
<td><inline-formula id="ieqn-181"><mml:math id="mml-ieqn-181"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td>Fed<sub>DPDR-DPML</sub></td>
<td>Distributed</td>
<td>PCA&#x002B;SVM</td>
<td><inline-formula id="ieqn-182"><mml:math id="mml-ieqn-182"><mml:mi>O</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>d</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>n</mml:mi><mml:mi>&#x03B5;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>Gaussian mechanism</td>
<td><inline-formula id="ieqn-183"><mml:math id="mml-ieqn-183"><mml:mrow><mml:mo>(</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In addition, the noise scales of AG [<xref ref-type="bibr" rid="ref-14">14</xref>] and DPSVD [<xref ref-type="bibr" rid="ref-16">16</xref>] are both <inline-formula id="ieqn-172"><mml:math id="mml-ieqn-172"><mml:mi>O</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msqrt><mml:mi>d</mml:mi></mml:msqrt><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mi>&#x03B5;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> since they only perturb the PCA procedure. DPPCA-SVM [<xref ref-type="bibr" rid="ref-15">15</xref>] and PCA-DPSVM [<xref ref-type="bibr" rid="ref-15">15</xref>] adopt the output perturbation in the noise addition phase, thus the noise scale is relatively large. Although our proposed Fed<sub>DPDR-DPML</sub> introduces noise in both PCA and SVM phases, still maintains a small noise scale when compared to DPPCA-SVM and PCA-DPSVM. Besides, Fed<sub>DPDR-DPML</sub> also has a relatively acceptable noise level compared to AG and DPSVD, while providing stronger privacy guarantees than AG and DPSVD.</p>
</sec>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Experiments</title>
<sec id="s5_1">
<label>5.1</label>
<title>Experiment Setup</title>
<p><bold>Dataset.</bold> As we know, image datasets usually have higher dimensions compared to general tabular data. Therefore, we select three image datasets with high dimensions and different characteristics to verify the performance of the mechanism proposed in this paper. MNIST and Fashion-MNIST share the same external characteristics, namely data size and dimension. But Fashion-MNIST is no longer the abstract number symbols, but more concrete clothing images. In contrast, the size of CIFAR-10 is similar to MNIST and Fashion-MNIST in magnitude, but the dimension of CIFAR-10 is much larger than the other two. The details of the three datasets (as shown in <xref ref-type="table" rid="table-2">Table 2</xref>) are as follows.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Datasets used in the experiment</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Dataset</th>
<th>Data size</th>
<th>Dimension</th>
<th>Target dimension <italic>k</italic></th>
</tr>
</thead>
<tbody>
<tr>
<td>MNIST</td>
<td>70,000</td>
<td>784 (<inline-formula id="ieqn-187"><mml:math id="mml-ieqn-187"><mml:mn>28</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula> pixels)</td>
<td>{5,10,20,50,100}</td>
</tr>
<tr>
<td>Fashion-MNIST</td>
<td>70,000</td>
<td>784 (<inline-formula id="ieqn-188"><mml:math id="mml-ieqn-188"><mml:mn>28</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula> pixels)</td>
<td>{5,10,20,50,100}</td>
</tr>
<tr>
<td>CIFAR-10</td>
<td>60,000</td>
<td>3,072 (<inline-formula id="ieqn-189"><mml:math id="mml-ieqn-189"><mml:mn>32</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>32</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula> pixels)</td>
<td>{5,10,20,50,100}</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><list list-type="bullet">
<list-item>
<p>MNIST dataset [<xref ref-type="bibr" rid="ref-37">37</xref>] consists of 60,000 training examples and 10,000 testing examples. Each example is a handwritten gray image with <inline-formula id="ieqn-184"><mml:math id="mml-ieqn-184"><mml:mn>28</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula> pixels, associated with a label from 10 classes (i.e., numbers 0 to 9).</p></list-item>
<list-item>
<p>Fashion-MNIST [<xref ref-type="bibr" rid="ref-38">38</xref>] is a dataset of Zalando&#x2019;s article images, which consists of a training set of 60,000 examples and a test set of 10,000 examples. Each example is a <inline-formula id="ieqn-185"><mml:math id="mml-ieqn-185"><mml:mn>28</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>28</mml:mn></mml:math></inline-formula> gray-scale image, associated with a label from 10 classes (e.g., coat, dress, bag, etc.).</p></list-item>
<list-item>
<p>CIFAR-10 dataset [<xref ref-type="bibr" rid="ref-39">39</xref>] a computer vision dataset for universal object recognition, which consists of 50,000 training examples and 10,000 testing examples. Each example is a <inline-formula id="ieqn-186"><mml:math id="mml-ieqn-186"><mml:mn>32</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>32</mml:mn></mml:math></inline-formula> color image, associated with a label from 10 classes (e.g., bird, cat, deer, etc.).</p></list-item>
</list></p>
<p><bold>Competitors.</bold> Non-Priv conducts machine learning with dimensionality reduction but without privacy protection. DPML conducts machine learning under differential privacy protection but without dimensionality reduction. DPDR-DPML and Fed<sub>DPDR-DPML</sub> are our proposed methods. As shown in <xref ref-type="table" rid="table-1">Table 1</xref>, the existing mechanisms, such as AG [<xref ref-type="bibr" rid="ref-14">14</xref>], DPPCA-SVM [<xref ref-type="bibr" rid="ref-15">15</xref>], PCA-DPSVM [<xref ref-type="bibr" rid="ref-15">15</xref>], DPSVD [<xref ref-type="bibr" rid="ref-16">16</xref>], and Truex et al.&#x2019;s method [<xref ref-type="bibr" rid="ref-24">24</xref>] all introduce perturbation to only one phase (i.e., PCA or SVM). In contrast, our proposed Fed<sub>DPDR-DPML</sub> involves noise addition in both the dimensionality reduction phase (i.e., PCA) and training phase (i.e., SVM), which provides strict and strong privacy protection. Therefore, such existing mechanisms theoretically provide insufficient privacy protection, thus not comparable to our paper.</p>

</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Experimental Results</title>
<p>This section presents our experimental results, including evaluations of accuracy and running time on SVM. By default, we set the parameters as <inline-formula id="ieqn-190"><mml:math id="mml-ieqn-190"><mml:mi>&#x03B5;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.1</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-191"><mml:math id="mml-ieqn-191"><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>4</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>, <inline-formula id="ieqn-192"><mml:math id="mml-ieqn-192"><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>20</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-193"><mml:math id="mml-ieqn-193"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula>, <inline-formula id="ieqn-194"><mml:math id="mml-ieqn-194"><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>, and <inline-formula id="ieqn-195"><mml:math id="mml-ieqn-195"><mml:mi>&#x03BB;</mml:mi><mml:mo>=</mml:mo><mml:mn>0.01</mml:mn></mml:math></inline-formula> in all experiments, where <inline-formula id="ieqn-196"><mml:math id="mml-ieqn-196"><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B5;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0.5</mml:mn><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> are used for DP-compliant dimensionality reduction and DP-compliant machine learning, respectively. We will show the accuracy and run time of different methods varying from parameters <inline-formula id="ieqn-197"><mml:math id="mml-ieqn-197"><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi><mml:mo>,</mml:mo><mml:mi>n</mml:mi></mml:math></inline-formula>.</p>
<sec id="s5_2_1">
<label>5.2.1</label>
<title>Evaluation of Accuracy</title>
<p>We first validate the performance of dimensionality reduction on SVM classification varying from the target dimension <italic>k</italic> on three high-dimensional datasets, as shown in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. We can see that the SVM classification accuracy of all mechanisms continuously increases with the dimension <italic>k</italic> increasing from 5 to 100 for all datasets. And, the accuracy does not change much when <italic>k</italic> is greater than 20. Therefore, we choose the target dimension as <inline-formula id="ieqn-198"><mml:math id="mml-ieqn-198"><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>20</mml:mn></mml:math></inline-formula> by default in the following experiments. Besides, it can be observed from three datasets that the accuracy of our proposed Fed<sub>DPDR-DPML</sub> and DPDR-DPML is much better than that of DPML and is close to Non-Priv when <italic>k</italic> is large. This demonstrates that DPDR-DPML can improve accuracy when dealing with high-dimensional data and can ensure superior data utility while providing strong privacy protection. Besides, Fed<sub>DPDR-DPML</sub> has the best accuracy on all datasets. It shows that knowledge aggregation can surely improve the data utility of machine learning.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Accuracy <italic>vs</italic>. target dimension <italic>k</italic> on SVM classification (<inline-formula id="ieqn-203"><mml:math id="mml-ieqn-203"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> &#x003D; 0.1, <italic>n</italic> &#x003D; 10,000, <italic>N</italic> &#x003D; 5)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-2.tif"/>
</fig>
<p>As for the CIFAR-10 dataset that has much higher dimensions (i.e., <inline-formula id="ieqn-199"><mml:math id="mml-ieqn-199"><mml:mi>d</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mn>3,072</mml:mn></mml:mrow></mml:math></inline-formula>), we also utilize the histogram of oriented gradient (HOG) in the experiment to improve accuracy, where the HOG parameters are used as follows: cell size is 4 pixels, number of bins is 9, block size is 2 cell, sliding step is 4 pixels. Nonetheless, the accuracy is not very high compared to MNIST and Fashion-MNIST. Because the SVM used in this paper is a linear model (using hinge loss strategy), and no kernel function is introduced to build a nonlinear model, nor is a convolutional network used. In the follow-up, we will further study the privacy-preserving SVM under the nonlinear model and the convolutional network.</p>
<p>Moreover, <xref ref-type="fig" rid="fig-3">Fig. 3</xref> shows the high accuracy of our proposed mechanisms on three datasets with the privacy parameter <inline-formula id="ieqn-200"><mml:math id="mml-ieqn-200"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> varying from 0.01 to 2.0, where <inline-formula id="ieqn-201"><mml:math id="mml-ieqn-201"><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>20</mml:mn><mml:mo>,</mml:mo><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mn>4</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>4</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. Specifically, we consider <inline-formula id="ieqn-202"><mml:math id="mml-ieqn-202"><mml:mi>&#x03B5;</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>0.01</mml:mn><mml:mo>,</mml:mo><mml:mn>0.05</mml:mn><mml:mo>,</mml:mo><mml:mn>0.1</mml:mn><mml:mo>,</mml:mo><mml:mn>0.5</mml:mn><mml:mo>,</mml:mo><mml:mn>1.0</mml:mn><mml:mo>,</mml:mo><mml:mn>2.0</mml:mn><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>. It can be seen from the three figures in <xref ref-type="fig" rid="fig-3">Fig. 3</xref> that the accuracy of Fed<sub>DPDR-DPML</sub> is much closer to Non-Priv which has no privacy protection. Thus, this demonstrates again that our proposed Fed<sub>DPDR-DPML</sub> can achieve better accuracy in distributed training tasks while keeping strong privacy protection. What&#x2019;s more, <xref ref-type="fig" rid="fig-3">Fig. 3</xref> shows that the accuracy of DPDR-DPML is much superior to DPSVM when applying the same level of privacy protection, which indicates DPDR-DPML holds better data utility while keeping the same privacy guarantees.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Accuracy <italic>vs</italic>. privacy parameter <inline-formula id="ieqn-204"><mml:math id="mml-ieqn-204"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> on SVM classification (<italic>k</italic> &#x003D; 20, <italic>n</italic> &#x003D; 10,000, <italic>N</italic> &#x003D; 5)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-3.tif"/>
</fig>
<p>Furthermore, <xref ref-type="fig" rid="fig-4">Fig. 4</xref> shows the comparisons of the impact of data size <inline-formula id="ieqn-205"><mml:math id="mml-ieqn-205"><mml:mi>n</mml:mi></mml:math></inline-formula> on accuracy, where <inline-formula id="ieqn-206"><mml:math id="mml-ieqn-206"><mml:mi>n</mml:mi></mml:math></inline-formula> is set as <inline-formula id="ieqn-207"><mml:math id="mml-ieqn-207"><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mn>100</mml:mn><mml:mo>,</mml:mo><mml:mn>500</mml:mn><mml:mo>,</mml:mo><mml:mn>1000</mml:mn><mml:mo>,</mml:mo><mml:mn>5000</mml:mn><mml:mo>,</mml:mo><mml:mn>10000</mml:mn><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>. It can be seen from <xref ref-type="fig" rid="fig-4">Fig. 4</xref> that the accuracy of the three mechanisms will increase with the increase of data size for three datasets. With different data sizes, our proposed Fed<sub>DPDR-DPML</sub> always outperforms DPML under the same privacy protection level. This is because Fed<sub>DPDR-DPML</sub> involves the knowledge aggregation to learn the global information, thus leading to a better data utility. Besides, we can also observe that DPDR-DPML has a higher accuracy than DPML. This demonstrates that the DP-compliant dimensionality reduction in DPDR-DPML can surely extract the key feature of high-dimensional data, thus leading to a higher accuracy than DPML. This also demonstrates that Fed<sub>DPDR-DPML</sub> can also improve the data utility in practice even when dealing with high-dimensional data.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Accuracy <italic>vs</italic>. data size <italic>n</italic> on SVM classification (<inline-formula id="ieqn-208"><mml:math id="mml-ieqn-208"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> &#x003D; 0.1, <italic>k</italic> &#x003D; 20, <italic>N</italic> &#x003D; 5)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-4.tif"/>
</fig>
<p>We also conduct experiments on uneven datasets to evaluate the performance of Fed<sub>DPDR-DPML</sub>, as shown in <xref ref-type="fig" rid="fig-5">Fig. 5</xref>. The number of participants is <inline-formula id="ieqn-209"><mml:math id="mml-ieqn-209"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn>5</mml:mn></mml:math></inline-formula>. We used three sets of uneven data in the experiment, where the sizes of the three sets of uneven data are <inline-formula id="ieqn-210"><mml:math id="mml-ieqn-210"><mml:mrow><mml:mo>(</mml:mo><mml:mn>0.05</mml:mn><mml:mo>,</mml:mo><mml:mn>0.1</mml:mn><mml:mo>,</mml:mo><mml:mn>0.5</mml:mn><mml:mo>,</mml:mo><mml:mn>1.0</mml:mn><mml:mo>,</mml:mo><mml:mn>2.0</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>, <inline-formula id="ieqn-211"><mml:math id="mml-ieqn-211"><mml:mrow><mml:mo>(</mml:mo><mml:mn>0.1</mml:mn><mml:mo>,</mml:mo><mml:mn>0.5</mml:mn><mml:mo>,</mml:mo><mml:mn>1.0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>,</mml:mo><mml:mn>10.0</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>, and <inline-formula id="ieqn-212"><mml:math id="mml-ieqn-212"><mml:mrow><mml:mo>(</mml:mo><mml:mn>0.1</mml:mn><mml:mo>,</mml:mo><mml:mn>1.0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>,</mml:mo><mml:mn>8.0</mml:mn><mml:mo>,</mml:mo><mml:mn>10.0</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:msup><mml:mn>10</mml:mn><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>. That is, each set of uneven data contains five different data sizes, corresponding to the uneven data of five participants.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Accuracy <italic>vs</italic>. uneven data on SVM classification (<inline-formula id="ieqn-213"><mml:math id="mml-ieqn-213"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula> &#x003D; 0.1, <italic>k</italic> &#x003D; 20, <italic>N</italic> &#x003D; 5)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_48115-fig-5.tif"/>
</fig>
<p>As we can see from <xref ref-type="fig" rid="fig-5">Fig. 5</xref>, our proposed Fed<sub>DPDR-DPML</sub> has a superior performance in dealing with uneven data. Fed<sub>DPDR-DPML</sub> is almost guaranteed to be as accurate as the Non-Priv method, so it is more suitable for scenarios with imbalanced data. This is because Fed<sub>DPDR-DPML</sub> can learn the global information of the training process through knowledge aggregation, thus performing well in handling imbalanced data. Compared to DPML and DPDR-DPML, our proposed Fed<sub>DPDR-DPML</sub> will surely improve the data utility while providing strong privacy guarantees.</p>
</sec>
<sec id="s5_2_2">
<label>5.2.2</label>
<title>Evaluation of Running Time</title>
<p>We also compared the running time of different mechanisms on SVM, as shown in <xref ref-type="table" rid="table-3">Table 3</xref>. Here, we set the data size as 10,000 and the target dimension as 20. It can be observed that the running time of Non-Priv, DPDR-DPML, and Fed<sub>DPDR-DPML</sub> is much lower than DPML, especially when the dataset (i.e., CIFAR-10) is very large. This proves that privacy-preserving dimensionality reduction can surely improve the efficiency of SVM training while providing privacy protection. Besides, compared with Non-Priv and DPML, our proposed Fed<sub>DPDR-DPML</sub> can maintain relatively excellent performance under the premise of providing strong privacy protection.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Running time of different mechanisms on SVM classification</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th></th>
<th align="center" colspan="4">Mechanism</th>
</tr>
<tr>
<th>Dataset</th>
<th>Non-priv</th>
<th>DPML</th>
<th>DPDR-DPML</th>
<th>Fed<sub>DPDR-DPML</sub></th>
</tr>
</thead>
<tbody>
<tr>
<td>MNIST</td>
<td>3.68 s</td>
<td>7,873.23 s</td>
<td>111.70 s</td>
<td>113.20 s</td>
</tr>
<tr>
<td>Fashion-MNIST</td>
<td>5.02 s</td>
<td>7,988.60 s</td>
<td>112.70 s</td>
<td>123.50 s</td>
</tr>
<tr>
<td>CIFAR-10</td>
<td>63.45 s</td>
<td>36,960.70 s</td>
<td>147.96 s</td>
<td>150.30 s</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Conclusion</title>
<p>Support vector machine (SVM) training inevitably faces severe privacy leakage issues when dealing with sensitive or private high-dimensional data. Therefore, this paper proposes a differential privacy-compliant support vector machine algorithm called Fed<sub>DPDR-DPML</sub>. Specifically, considering multi-party joint training with uneven data, Fed<sub>DPDR-DPML</sub> is a distributed framework that incorporates dimensionality reduction and knowledge aggregation to obtain global learning information, which greatly improves the data utility while providing strong privacy guarantees. We conduct extensive experiments on three high-dimensional data with different characteristics. The experimental results show that our proposed algorithm can maintain good data utility while providing strong privacy guarantees.</p>
<p>Furthermore, the privacy paradigm and the framework of Fed<sub>DPDR-DPML</sub> can be easily extended to other machine learning models, such as logistic regression, Bayesian classification, or decision trees. Based on Fed<sub>DPDR-DPML</sub>, we will consider investigating distributed deep learning with differential privacy. Moreover, personalized, dynamic, and efficient privacy-preserving machine learning frameworks require further research in the future.</p>
</sec>
</body>
<back>
<ack><p>The authors wish to express their appreciation to the reviewers for their helpful suggestions which greatly improved the presentation of this paper.</p>
</ack>
<sec><title>Funding Statement</title>
<p>This work was supported in part by National Natural Science Foundation of China (Nos. 62102311, 62202377, 62272385), in part by Natural Science Basic Research Program of Shaanxi (Nos. 2022JQ-600, 2022JM-353, 2023-JC-QN-0327), in part by Shaanxi Distinguished Youth Project (No. 2022JC-47), in part by Scientific Research Program Funded by Shaanxi Provincial Education Department (No. 22JK0560), in part by Distinguished Youth Talents of Shaanxi Universities, and in part by Youth Innovation Team of Shaanxi Universities.</p>
</sec>
<sec><title>Author Contributions</title>
<p>The authors confirm their contribution to the paper as follows: study conception and design: T. Wang, Y. Zhang; validation: J. Liang, S. Wang; analysis and interpretation of results: T. Wang, Y. Zhang, S. Liu; draft manuscript preparation: T. Wang. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability"><title>Availability of Data and Materials</title>
<p>The three data used in this study can be found in references [<xref ref-type="bibr" rid="ref-37">37</xref>, <xref ref-type="bibr" rid="ref-38">38</xref>], and [<xref ref-type="bibr" rid="ref-39">39</xref>], respectively.</p>
</sec>
<sec sec-type="COI-statement"><title>Conflicts of Interest</title>
<p>The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Kocon</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>ChatGPT: Jack of all trades, master of none</article-title>,&#x201D; <source>Inf. Fusion</source>, vol. <volume>99</volume>, pp. <fpage>101861</fpage>, <year>2023</year>. doi: <pub-id pub-id-type="doi">10.1016/j.inffus.2023.101861</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Brown</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Language models are few-shot learners</article-title>,&#x201D; in <conf-name>Adv. Neural Inf. Proc. Syst. (NeurIPS)</conf-name>, <year>Dec. 2020</year>, pp. <fpage>1877</fpage>&#x2013;<lpage>1901</lpage>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Wu</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Farokhi</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Smith</surname></string-name>, and <string-name><given-names>M. A.</given-names> <surname>Kaafar</surname></string-name></person-group>, &#x201C;<chapter-title>The value of collaboration in convex machine learning with differential privacy</chapter-title>,&#x201D; in <source>IEEE S&#x0026;P</source>, <publisher-loc>San Francisco, CA, USA</publisher-loc>, <year>May 2020</year>, pp. <fpage>304</fpage>&#x2013;<lpage>317</lpage>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Al-Rubaie</surname></string-name> and <string-name><given-names>J. M.</given-names> <surname>Chang</surname></string-name></person-group>, &#x201C;<article-title>Privacy-preserving machine learning: Threats and solutions</article-title>,&#x201D; <source>IEEE Secur. Privacy</source>, vol. <volume>17</volume>, no. <issue>2</issue>, pp. <fpage>49</fpage>&#x2013;<lpage>58</lpage>, <year>Mar. 2019</year>. doi: <pub-id pub-id-type="doi">10.1109/MSEC.2018.2888775</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H. C.</given-names> <surname>Tanuwidjaja</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Choi</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Baek</surname></string-name>, and <string-name><given-names>K.</given-names> <surname>Kim</surname></string-name></person-group>, &#x201C;<article-title>Privacy-preserving deep learning on machine learning as a service-a comprehensive survey</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>167425</fpage>&#x2013;<lpage>167447</lpage>, <year>Sep. 2020</year>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2020.3023084</pub-id>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Fredrikson</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Jha</surname></string-name>, and <string-name><given-names>T.</given-names> <surname>Ristenpart</surname></string-name></person-group>, &#x201C;<article-title>Model inversion attacks that exploit confidence information and basic countermeasures</article-title>,&#x201D; in <conf-name>Proc. ACM SIGSAC Conf. on Comput. and Communica. Securi.</conf-name>, <publisher-loc>Denver, USA</publisher-loc>, <year>Oct. 2015</year>, pp. <fpage>1322</fpage>&#x2013;<lpage>1333</lpage>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Xie</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Xiang</surname></string-name></person-group>, &#x201C;<article-title>A survey on privacy inference attacks and defenses in cloud-based deep neural network</article-title>,&#x201D; <source>Comput. Stand. Interfaces</source>, vol. <volume>83</volume>, pp. <fpage>103672</fpage>, <year>Jan. 2023</year>. doi: <pub-id pub-id-type="doi">10.1016/j.csi.2022.103672</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Jakkula</surname></string-name></person-group>, &#x201C;<article-title>Tutorial on support vector machine (SVM)</article-title>,&#x201D; <publisher-name>Sch. EECS, Washington State Univ.</publisher-name>, <year>2006</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Dwork</surname></string-name></person-group>, &#x201C;<article-title>Differential privacy</article-title>,&#x201D; in <conf-name> Int. Conf. ICALP</conf-name>, <publisher-loc>Venice, Italy</publisher-loc>, <year>Jul. 2006</year>, pp. <fpage>1</fpage>&#x2013;<lpage>12</lpage>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Dwork</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Roth</surname></string-name></person-group>, &#x201C;<article-title>The algorithmic foundations of differential privacy</article-title>,&#x201D; <source>Found. Trends&#x00AE; Theor. Comput. Sci.</source>, vol. <volume>9</volume>, no. <issue>3&#x2013;4</issue>, pp. <fpage>211</fpage>&#x2013;<lpage>407</lpage>, <year>Aug. 2014</year>. doi: <pub-id pub-id-type="doi">10.1561/0400000042</pub-id>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Hao</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>A differential privacy support vector machine classifier based on dual variable perturbation</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>7</volume>, pp. <fpage>98238</fpage>&#x2013;<lpage>98251</lpage>, <year>Jul. 2019</year>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2019.2929680</pub-id>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Farokhi</surname></string-name></person-group>, &#x201C;<article-title>Privacy-preserving public release of datasets for support vector machine classification</article-title>,&#x201D; <source>IEEE Trans. Big Data</source>, vol. <volume>7</volume>, no. <issue>5</issue>, pp. <fpage>893</fpage>&#x2013;<lpage>899</lpage>, <year>Jan. 2020</year>. doi: <pub-id pub-id-type="doi">10.1109/TBDATA.2019.2963391</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Mao</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Duan</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Zhang</surname></string-name> and <string-name><given-names>Z.</given-names> <surname>Hong</surname></string-name></person-group>, &#x201C;<article-title>Privacy-preserving multi-class support vector machine model on medical diagnosis</article-title>,&#x201D; <source>IEEE J. Biomed. Health Inform.</source>, vol. <volume>26</volume>, no. <issue>7</issue>, pp. <fpage>3342</fpage>&#x2013;<lpage>3353</lpage>, <year>Mar. 2022</year>. doi: <pub-id pub-id-type="doi">10.1109/JBHI.2022.3157592</pub-id>; <pub-id pub-id-type="pmid">35259122</pub-id></mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Dwork</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Talwar</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Thakurta</surname></string-name>, and <string-name><given-names>L.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Analyze gauss: Optimal bounds for privacy-preserving principal component analysis</article-title>,&#x201D; in <conf-name>Proc. ACM Symp. on Theory of Computi.</conf-name>, <publisher-loc>New York, USA</publisher-loc>, <year>May 2014</year>, pp. <fpage>11</fpage>&#x2013;<lpage>20</lpage>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Huang</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Yang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Xu</surname></string-name>, and <string-name><given-names>H.</given-names> <surname>Zhou</surname></string-name></person-group>, &#x201C;<article-title>Differential privacy principal component analysis for support vector machines</article-title>,&#x201D; <source>Secur. Commun. Netw.</source>, vol. <volume>2021</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>12</lpage>, <year>Jul. 2021</year>. doi: <pub-id pub-id-type="doi">10.1155/2021/5542283</pub-id>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yang</surname></string-name>, and <string-name><given-names>X.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Differentially private singular value decomposition for training support vector machines</article-title>,&#x201D; <source>Comput. Intell. Neurosci.</source>, vol. <volume>2022</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>11</lpage>, <year>Mar. 2022</year>. doi: <pub-id pub-id-type="doi">10.1155/2022/2935975</pub-id>; <pub-id pub-id-type="pmid">35378802</pub-id></mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>P.</given-names> <surname>Kairouz</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Advances and open problems in federated learning</article-title>,&#x201D; <source>Found. Trends&#x00AE; Mach. Learn.</source>, vol. <volume>14</volume>, no. <issue>1&#x2013;2</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>210</lpage>, <year>Jun. 2021</year>. doi: <pub-id pub-id-type="doi">10.1561/2200000083</pub-id>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Wu</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Zhou</surname></string-name>, and <string-name><given-names>Y.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>Vertical federated learning based on consortium blockchain for data sharing in mobile edge computing</article-title>,&#x201D; <source>Comp. Model. Eng.</source>, vol. <volume>137</volume>, no. <issue>1</issue>, pp. <fpage>345</fpage>&#x2013;<lpage>361</lpage>, <year>2023</year>. doi: <pub-id pub-id-type="doi">10.32604/cmes.2023.026920</pub-id>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>McMahan</surname></string-name>, <string-name><given-names>E.</given-names> <surname>Moore</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Ramage</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Hampson</surname></string-name>, and <string-name><given-names>B. A. Y</given-names> <surname>Arcas</surname></string-name></person-group>, &#x201C;<article-title>Communication-efficient learning of deep networks from decentralized data</article-title>,&#x201D; in <conf-name>Int. Conf. Artif. Intell. Stat. (AISTATS)</conf-name>, <publisher-loc>Fort Lauderdale, USA</publisher-loc>, <year>Apr. 2017</year>, pp. <fpage>1273</fpage>&#x2013;<lpage>1282</lpage>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Yang</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Zhu</surname></string-name></person-group>, &#x201C;<article-title>Parallel restarted SGD with faster convergence and less communication: Demystifying why model averaging works for deep learning</article-title>,&#x201D; in <conf-name>Proc. AAAI Conf. Artif. Intell.</conf-name>, <publisher-loc>Honolulu, Hawaii, USA</publisher-loc>, <year>Jan. 2019</year>, pp. <fpage>5693</fpage>&#x2013;<lpage>5700</lpage>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>L. T.</given-names> <surname>Phong</surname></string-name> and <string-name><given-names>T. T.</given-names> <surname>Phuong</surname></string-name></person-group>, &#x201C;<article-title>Privacy-preserving deep learning via weight transmission</article-title>,&#x201D; <source>IEEE Trans. Inf. Forens. Secur.</source>, vol. <volume>14</volume>, no. <issue>11</issue>, pp. <fpage>3003</fpage>&#x2013;<lpage>3015</lpage>, <year>Apr. 2019</year>. doi: <pub-id pub-id-type="doi">10.1109/TIFS.2019.2911169</pub-id>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Yu</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Distributed learning over unreliable networks</article-title>,&#x201D; in <conf-name>Proc. Int. Conf. Mach. Learn. (ICML)</conf-name>, <publisher-loc>Long Beach, California, USA</publisher-loc>, <year>Jun. 2019</year>, pp. <fpage>7202</fpage>&#x2013;<lpage>7212</lpage>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Zhao</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Local differential privacy-based federated learning for internet of things</article-title>,&#x201D; <source>IEEE Internet Things J.</source>, vol. <volume>8</volume>, no. <issue>11</issue>, pp. <fpage>8836</fpage>&#x2013;<lpage>8853</lpage>, <year>Nov. 2020</year>. doi: <pub-id pub-id-type="doi">10.1109/JIOT.2020.3037194</pub-id>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Truex</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>A hybrid approach to privacy-preserving federated learning</article-title>,&#x201D; in <conf-name>Proc. ACM Workshop Artif. Intell. Secur. (AISec)</conf-name>, <publisher-loc>London, UK</publisher-loc>, <year>Nov. 2019</year>, pp. <fpage>1</fpage>&#x2013;<lpage>11</lpage>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Baracaldo</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhou</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Anwar</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Joshi</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Ludwig</surname></string-name></person-group>, &#x201C;<article-title>FedV: Privacy-preserving federated learning over vertically partitioned data</article-title>,&#x201D; in <conf-name>Proc. ACM Workshop Artif. Intell. Secur. (AISec)</conf-name>, <publisher-loc>Korea</publisher-loc>, <year>Nov. 2021</year>, pp. <fpage>181</fpage>&#x2013;<lpage>192</lpage>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Ye</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Zhou</surname></string-name>, and <string-name><given-names>S. Y.</given-names> <surname>Philip</surname></string-name></person-group>, &#x201C;<article-title>More than privacy: Applying differential privacy in key areas of artificial intelligence</article-title>,&#x201D; <source>IEEE Trans. Knowl. Data Eng.</source>, vol. <volume>34</volume>, no. <issue>6</issue>, pp. <fpage>2824</fpage>&#x2013;<lpage>2843</lpage>, <year>Aug. 2020</year>. doi: <pub-id pub-id-type="doi">10.1109/TKDE.2020.3014246</pub-id>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Ponomareva</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>How to DP-fy ML: A practical guide to machine learning with differential privacy</article-title>,&#x201D; <source>J. Artif. Intell. Res.</source>, vol. <volume>77</volume>, pp. <fpage>1113</fpage>&#x2013;<lpage>1201</lpage>, <year>Jul. 2023</year>. doi: <pub-id pub-id-type="doi">10.1613/jair.1.14649</pub-id>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Ji</surname></string-name>, <string-name><given-names>Z. C.</given-names> <surname>Lipton</surname></string-name>, and <string-name><given-names>C.</given-names> <surname>Elkan</surname></string-name></person-group>, &#x201C;<article-title>Differential privacy and machine learning: A survey and review</article-title>,&#x201D; <comment>arXiv preprint arXiv:1412.7584</comment>, <year>Dec. 2014</year>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1412.7584</pub-id>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>D. A.</given-names> <surname>Pisner</surname></string-name> and <string-name><given-names>D. M.</given-names> <surname>Schnyer</surname></string-name></person-group>, &#x201C;<chapter-title>Support vector machine</chapter-title>,&#x201D; in <source>Machine Learning</source>, <publisher-loc>San Diego, USA</publisher-loc>, <publisher-name>Academic Press</publisher-name>, <year>2020</year>, pp. <fpage>101</fpage>&#x2013;<lpage>121</lpage>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>Chaudhuri</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Monteleoni</surname></string-name>, and <string-name><given-names>A. D.</given-names> <surname>Sarwate</surname></string-name></person-group>, &#x201C;<article-title>Differentially private empirical risk minimization</article-title>,&#x201D; <source>J. Mach. Learn. Res.</source>, vol. <volume>12</volume>, no. <issue>3</issue>, pp. <fpage>1069</fpage>&#x2013;<lpage>1109</lpage>, <year>Mar. 2011</year>; <pub-id pub-id-type="pmid">21892342</pub-id></mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Jiang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Xie</surname></string-name>, and <string-name><given-names>Z.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Wishart mechanism for differentially private principal components analysis</article-title>,&#x201D; in <conf-name>Proc. AAAI Conf. Artif. Intell.</conf-name>, <publisher-loc>Phoenix, Arizona, USA</publisher-loc>, <year>Feb. 2016</year>, pp. <fpage>1730</fpage>&#x2013;<lpage>1736</lpage>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Yang</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Bai</surname></string-name></person-group>, &#x201C;<article-title>Laplace input and output perturbation for differentially private principal components analysis</article-title>,&#x201D; <source>Secur. Commun. Netw.</source>, vol. <volume>2019</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>10</lpage>, <year>Nov. 2019</year>. doi: <pub-id pub-id-type="doi">10.1155/2019/9169802</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Tavara</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Schliep</surname></string-name>, and <string-name><given-names>D.</given-names> <surname>Basu</surname></string-name></person-group>, &#x201C;<chapter-title>Federated learning of oligonucleotide drug molecule thermodynamics with differentially private ADMM-based SVM</chapter-title>,&#x201D; in <source>ECML PKDD</source>, <publisher-loc>Bilbao, Spain</publisher-loc>, <year>Sep. 2021</year>, pp. <fpage>459</fpage>&#x2013;<lpage>467</lpage>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Abadi</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Deep learning with differential privacy</article-title>,&#x201D; in <conf-name>Proc. ACM SIGSAC Conf. on Comput. and Communica. Securi.</conf-name>, <publisher-loc>Vienna, Austria</publisher-loc>, <year>Oct. 2016</year>, pp. <fpage>308</fpage>&#x2013;<lpage>318</lpage>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Yue</surname></string-name>, <string-name><given-names>S. S.</given-names> <surname>Chow</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Huang</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Sun</surname></string-name></person-group>, &#x201C;<article-title>DP-Forward: Fine-tuning and inference on language models with differential privacy in forward pass</article-title>,&#x201D; in <conf-name>Proc. ACM SIGSAC Conf. on Comput. and Communica. Securi.</conf-name>, <publisher-loc>Copenhagen, Denmark</publisher-loc>, <year>Nov. 2023</year>, pp. <fpage>2665</fpage>&#x2013;<lpage>2679</lpage>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Chapelle</surname></string-name></person-group>, &#x201C;<article-title>Training a support vector machine in the primal</article-title>,&#x201D; <source>Neural. Comput.</source>, vol. <volume>19</volume>, no. <issue>5</issue>, pp. <fpage>1155</fpage>&#x2013;<lpage>1178</lpage>, <year>May 2007</year>. doi: <pub-id pub-id-type="doi">10.1162/neco.2007.19.5.1155</pub-id>; <pub-id pub-id-type="pmid">17381263</pub-id></mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>LeCun</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Bottou</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Bengio</surname></string-name>, and <string-name><given-names>P.</given-names> <surname>Haffner</surname></string-name></person-group>, &#x201C;<article-title>Gradient-based learning applied to document recognition</article-title>,&#x201D; in <source>Proc. IEEE</source>, vol. <volume>86</volume>, no. <issue>11</issue>, pp. <fpage>2278</fpage>&#x2013;<lpage>2324</lpage>, <year>Nov. 1998</year>. doi: <pub-id pub-id-type="doi">10.1109/5.726791</pub-id>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Xiao</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Rasul</surname></string-name>, and <string-name><given-names>R.</given-names> <surname>Vollgraf</surname></string-name></person-group>, &#x201C;<article-title>Fashion-MNIST: A novel image dataset for benchmarking machine learning algorithms</article-title>,&#x201D; <comment>arXiv preprint arXiv:1708.07747</comment>, <year>Sep. 2017</year>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1708.07747</pub-id>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Krizhevsky</surname></string-name> and <string-name><given-names>G.</given-names> <surname>Hinton</surname></string-name></person-group>, &#x201C;<article-title>Learning multiple layers of features from tiny images</article-title>,&#x201D; in <source>Technical Report (CIFAR)</source>, <publisher-loc>Toronto, Canada</publisher-loc>, <publisher-name>University of Toronto</publisher-name>, <year>2009</year>, pp. <fpage>1</fpage>&#x2013;<lpage>58</lpage>. </mixed-citation></ref>
</ref-list>
</back></article>