<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">47275</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2024.047275</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>LDAS&#x0026;ET-AD: Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation</article-title>
<alt-title alt-title-type="left-running-head">LDAS&#x0026;ET-AD: Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation</alt-title>
<alt-title alt-title-type="right-running-head">LDAS&#x0026;ET-AD: Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Li</surname><given-names>Shuyi</given-names></name></contrib>
<contrib id="author-2" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Hu</surname><given-names>Hongchao</given-names></name><email>hhc19820523@163.com</email></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Yang</surname><given-names>Xiaohan</given-names></name></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Cheng</surname><given-names>Guozhen</given-names></name></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Liu</surname><given-names>Wenyan</given-names></name></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Guo</surname><given-names>Wei</given-names></name></contrib>
<aff><institution>National Digital Switching System Engineering &#x0026; Technological R&#x0026;D Center, The PLA Information Engineering University</institution>, <addr-line>Zhengzhou, 450000</addr-line>, <country>China</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Hongchao Hu. Email: <email>hhc19820523@163.com</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2024</year></pub-date>
<pub-date date-type="pub" publication-format="electronic"><day>15</day>
<month>5</month>
<year>2024</year></pub-date>
<volume>79</volume>
<issue>2</issue>
<fpage>2331</fpage>
<lpage>2359</lpage>
<history>
<date date-type="received">
<day>31</day>
<month>10</month>
<year>2023</year>
</date>
<date date-type="accepted">
<day>27</day>
<month>3</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2024 Li et al.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Li et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_47275.pdf"></self-uri>
<abstract>
<p>Adversarial distillation (AD) has emerged as a potential solution to tackle the challenging optimization problem of loss with hard labels in adversarial training. However, fixed sample-agnostic and student-egocentric attack strategies are unsuitable for distillation. Additionally, the reliability of guidance from static teachers diminishes as target models become more robust. This paper proposes an AD method called Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation (LDAS&#x0026;ET-AD). Firstly, a learnable distillation attack strategies generating mechanism is developed to automatically generate sample-dependent attack strategies tailored for distillation. A strategy model is introduced to produce attack strategies that enable adversarial examples (AEs) to be created in areas where the target model significantly diverges from the teachers by competing with the target model in minimizing or maximizing the AD loss. Secondly, a teacher evolution strategy is introduced to enhance the reliability and effectiveness of knowledge in improving the generalization performance of the target model. By calculating the experimentally updated target model&#x2019;s validation performance on both clean samples and AEs, the impact of distillation from each training sample and AE on the target model&#x2019;s generalization and robustness abilities is assessed to serve as feedback to fine-tune standard and robust teachers accordingly. Experiments evaluate the performance of LDAS&#x0026;ET-AD against different adversarial attacks on the CIFAR-10 and CIFAR-100 datasets. The experimental results demonstrate that the proposed method achieves a robust precision of 45.39% and 42.63% against AutoAttack (AA) on the CIFAR-10 dataset for ResNet-18 and MobileNet-V2, respectively, marking an improvement of 2.31% and 3.49% over the baseline method. In comparison to state-of-the-art adversarial defense techniques, our method surpasses Introspective Adversarial Distillation, the top-performing method in terms of robustness under AA attack for the CIFAR-10 dataset, with enhancements of 1.40% and 1.43% for ResNet-18 and MobileNet-V2, respectively. These findings demonstrate the effectiveness of our proposed method in enhancing the robustness of deep learning networks (DNNs) against prevalent adversarial attacks when compared to other competing methods. In conclusion, LDAS&#x0026;ET-AD provides reliable and informative soft labels to one of the most promising defense methods, AT, alleviating the limitations of untrusted teachers and unsuitable AEs in existing AD techniques. We hope this paper promotes the development of DNNs in real-world trust-sensitive fields and helps ensure a more secure and dependable future for artificial intelligence systems.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Adversarial training</kwd>
<kwd>adversarial distillation</kwd>
<kwd>learnable distillation attack strategies</kwd>
<kwd>teacher evolution strategy</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>National Key Research and Development Program of China</funding-source>
<award-id>2021YFB1006200</award-id>
</award-group>
<award-group id="awg2">
<funding-source>Major Science and Technology Project of Henan Province in China</funding-source>
<award-id>221100211200</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>In recent years, deep neural networks (DNNs) have become increasingly popular for solving complex real-world problems, including computer vision [<xref ref-type="bibr" rid="ref-1">1</xref>], natural language processing [<xref ref-type="bibr" rid="ref-2">2</xref>], and other fields [<xref ref-type="bibr" rid="ref-3">3</xref>]. However, Szegedy et al. [<xref ref-type="bibr" rid="ref-4">4</xref>] have revealed that DNNs are susceptible to adversarial examples (AEs), which involve imperceptible perturbations on input. These perturbations can easily mislead the prediction model, posing a challenge to the development of DNNs in trust-sensitive fields like autonomous driving [<xref ref-type="bibr" rid="ref-5">5</xref>], facial authentication [<xref ref-type="bibr" rid="ref-6">6</xref>], and healthcare [<xref ref-type="bibr" rid="ref-1">1</xref>].</p>
<p>To combat adversarial attacks, various defense strategies have emerged, including input preprocessing [<xref ref-type="bibr" rid="ref-7">7</xref>&#x2013;<xref ref-type="bibr" rid="ref-9">9</xref>], adversarial training (AT) [<xref ref-type="bibr" rid="ref-10">10</xref>&#x2013;<xref ref-type="bibr" rid="ref-13">13</xref>], and certified defense [<xref ref-type="bibr" rid="ref-14">14</xref>&#x2013;<xref ref-type="bibr" rid="ref-17">17</xref>]. Among them, AT is considered one of the most effective methods for improving the robustness of DNNs. It achieves this by incorporating AEs into the training procedure through a minimax formulation [<xref ref-type="bibr" rid="ref-13">13</xref>]. However, learning directly from AEs is challenging due to the difficult optimization of loss with hard labels [<xref ref-type="bibr" rid="ref-18">18</xref>], hindering improvements in both clean accuracy and adversarial robustness.</p>
<p>Recent studies have shown that knowledge distillation (KD) can enhance AT by providing data-driven soft labels to smooth the hard labels. Adversarial distillation (AD) methods aim to have target models to mimic the outputs or features of either a single adversarially pre-trained teacher [<xref ref-type="bibr" rid="ref-19">19</xref>&#x2013;<xref ref-type="bibr" rid="ref-21">21</xref>] or both an adversarially pre-trained teacher and a standard pre-trained teacher [<xref ref-type="bibr" rid="ref-22">22</xref>&#x2013;<xref ref-type="bibr" rid="ref-24">24</xref>]. By utilizing the guidance of these teachers, the target model can learn the ability to identify AEs and clean samples simultaneously. In the aforementioned methods, the target models fully trust teacher models. Zhu et al. [<xref ref-type="bibr" rid="ref-25">25</xref>] noted that the knowledge from static teacher models becomes less reliable over time, as they become progressively less accurate in predicting stronger AEs. To enhance the reliability of guidance received by the target model, Introspective Adversarial Distillation (IAD) was introduced to encourage the target model to partially trust the teacher model and gradually trust itself more. However, the parameters of the teacher models remain constant, hindering the target model from acquiring increasingly reliable knowledge from the teachers.</p>
<p>Additionally, the fixed sample-agnostic and student-egocentric attack strategies used to generate AEs may not be suitable for distillation, limiting the target model&#x2019;s generalization performance improvement.</p>
<p>To address the reliability reduction of teacher knowledge in KD, the emerging field of learning to teach (L2T) distillation algorithms [<xref ref-type="bibr" rid="ref-26">26</xref>] has made significant progress. Existing L2T distillation techniques involve fine-tuning teachers to enforce similarity between the outputs of teacher and student models on the training set [<xref ref-type="bibr" rid="ref-27">27</xref>&#x2013;<xref ref-type="bibr" rid="ref-30">30</xref>], maximizing the student model&#x2019;s generalization ability on a held-out dataset [<xref ref-type="bibr" rid="ref-31">31</xref>&#x2013;<xref ref-type="bibr" rid="ref-34">34</xref>], and incorporating distillation influence to estimate the impacts of each training sample on the student&#x2019;s validation performance [<xref ref-type="bibr" rid="ref-35">35</xref>]. By incorporating distillation influence and self-evolution into the teacher&#x2019;s learning process, Reference [<xref ref-type="bibr" rid="ref-35">35</xref>] prioritized samples likely to enhance the student&#x2019;s generalization ability, resulting in superior performance when updating the teacher model. However, existing L2T distillation techniques only utilize the clean accuracy of the student model to update the standard teacher, without considering updating the robust teacher to enhance the target model&#x2019;s robustness.</p>
<p>To solve the issue of limited generalization performance caused by fixed attack strategies, some works [<xref ref-type="bibr" rid="ref-12">12</xref>,<xref ref-type="bibr" rid="ref-36">36</xref>&#x2013;<xref ref-type="bibr" rid="ref-38">38</xref>] have improved AT by exploiting different attack strategies at different training stages. Reference [<xref ref-type="bibr" rid="ref-12">12</xref>] proposed a novel AT framework by introducing a learnable attack strategy (LAS-AT), which consists of a target network trained with AEs to improve robustness and a strategy network that automatically produces attack strategies based on the target model&#x2019;s robustness and the given sample. This framework requires less domain expertise. However, directly extending it into the AD framework makes the generated AEs independent of the teacher model and unsuitable for distillation, hindering the closer matching between teacher and target models.</p>
<p>In this paper, an adversarial defense method called Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation (LDAS&#x0026;ET-AD) is proposed, which aims to improve the performance of AD by enhancing the quality of AEs and the reliability of teacher knowledge. Our contributions are summarized as follows:
<list list-type="order">
<list-item>
<p>A learnable distillation attack strategies generating mechanism is proposed to automatically generate sample-dependent attack strategies tailored for distillation. A strategy model is introduced to generate attack strategies capable of misleading the target model and creating maximum divergence between the target and teacher models by competing with the target model in minimizing or maximizing the AD loss. AEs are produced by perturbing clean samples in the direction of the gradient of the difference between the target and teacher models, causing a closer match between them.</p></list-item>
<list-item>
<p>A teacher evolution strategy is devised to enhance the reliability and effectiveness of knowledge in improving the target mode&#x2019;s generalization performance on both clean samples and AEs. The adversarial distillation influence, which estimates the impact of distillation from each training sample and AE on the target model&#x2019;s performance on the validation set and AEs, is introduced to assign loss weights of the training samples and AEs. The standard and robust teachers are fine-tuned on prioritized samples that are likely to enhance the target model&#x2019;s clean and robust generalization abilities, respectively.</p></list-item>
</list></p>
<p>To evaluate the effectiveness of the LDAS&#x0026;ET-AD method, we construct two typical DNNs, namely ResNet-18 and MobileNet-V2, and test them against various adversarial attacks on the CIFAR-10 and CIFAR-100 datasets. In comparison to state-of-the-art adversarial defense techniques, our method demonstrates robustness enhancements ranging from 0.80% to 1.47% for the CIFAR-10 dataset and 1.43% to 2.11% for the CIFAR-100 dataset when applied to ResNet-18. When implemented on MobileNet-V2, our method showcases improvements ranging from 1.20% to 2.55% for the CIFAR-10 dataset and 1.23% to 2.30% for the CIFAR-100 dataset.</p>
<p>The remainder of the paper is organized as follows: <xref ref-type="sec" rid="s2">Section 2</xref> reviews related background and recent research. <xref ref-type="sec" rid="s3">Section 3</xref> describes the proposed LDAS&#x0026;ET-AD method in detail. <xref ref-type="sec" rid="s4">Section 4</xref> presents experimental results and comparisons. <xref ref-type="sec" rid="s5">Section 5</xref> gives discussions. <xref ref-type="sec" rid="s6">Section 6</xref> concludes the paper and <xref ref-type="sec" rid="s7">Section 7</xref> provides limitations.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<sec id="s2_1">
<label>2.1</label>
<title>Adversarial Attacks and Adversarial Training</title>
<p>Since the identification of DNNs&#x2019; vulnerability to adversarial attacks, several effective attack algorithms have been proposed [<xref ref-type="bibr" rid="ref-13">13</xref>,<xref ref-type="bibr" rid="ref-39">39</xref>&#x2013;<xref ref-type="bibr" rid="ref-41">41</xref>]. These methods can be categorized as white-box attacks and black-box attacks based on the adversary&#x2019;s knowledge. White-box attacks such as the fast gradient sign method (FGSM) [<xref ref-type="bibr" rid="ref-39">39</xref>], projection gradient descent method (PGD) [<xref ref-type="bibr" rid="ref-13">13</xref>], and Carlini Wagner Attack (CW) [<xref ref-type="bibr" rid="ref-40">40</xref>], have full access to all the parameter information of the attacked model. To comprehensively evaluate the effectiveness of the proposed defense method, we employ PGD [<xref ref-type="bibr" rid="ref-13">13</xref>], FGSM [<xref ref-type="bibr" rid="ref-39">39</xref>], CW [<xref ref-type="bibr" rid="ref-40">40</xref>], and AutoAttack (AA) [<xref ref-type="bibr" rid="ref-41">41</xref>].</p>
<p>To mitigate the threat of adversarial attacks, various defense methods have been proposed [<xref ref-type="bibr" rid="ref-5">5</xref>,<xref ref-type="bibr" rid="ref-10">10</xref>,<xref ref-type="bibr" rid="ref-14">14</xref>]. AT [<xref ref-type="bibr" rid="ref-10">10</xref>&#x2013;<xref ref-type="bibr" rid="ref-13">13</xref>], which adds adversarial perturbations to the inputs during training, has proven to be one of the most effective approaches for enhancing the DNNs&#x2019; adversarial robustness. Madry et al. [<xref ref-type="bibr" rid="ref-13">13</xref>] formulated standard AT (SAT) as a minimax optimization problem, where the inner maximization represents the attack strategy guiding AE generation. Solving the inner maximization problem in SAT is achieved using the PGD attack.</p>
<p>Several studies have proposed methods to improve the performance of SAT. Zhang et al. [<xref ref-type="bibr" rid="ref-10">10</xref>] introduced TRadeoff-inspired Adversarial DEfense via Surrogate-loss minimization (TRADES) to balance adversarial robustness and clean accuracy. Wang et al. [<xref ref-type="bibr" rid="ref-42">42</xref>] further improved performance by Misclassification-Aware Adversarial Training (MART). While these methods employed fixed attack strategies, other studies [<xref ref-type="bibr" rid="ref-12">12</xref>,<xref ref-type="bibr" rid="ref-36">36</xref>&#x2013;<xref ref-type="bibr" rid="ref-38">38</xref>] demonstrated that employing different attack strategies at different training phases can further improve AT. Cai et al. [<xref ref-type="bibr" rid="ref-36">36</xref>] introduced curriculum adversarial training (CAT), which employs AEs generated by attacks with varying strengths in a curriculum manner. Zhang et al. [<xref ref-type="bibr" rid="ref-37">37</xref>] proposed friendly adversarial training (FAT) that trains a DNN using both wrongly-predicted and correctly-predicted AEs. Wang et al. [<xref ref-type="bibr" rid="ref-38">38</xref>] introduced First-Order Stationary Condition for constrained optimization (FOSC) as a quantitative indicator for assessing AE convergence quality. However, these methods rely on manually designed metrics to evaluate the AE difficulty and still use a single strategy at each stage, thus limiting the robustness improvement and requiring domain expertise. Jia et al. [<xref ref-type="bibr" rid="ref-12">12</xref>] proposed a learnable attack strategy that allows the strategy model to automatically produce sample-dependent attack strategies using a gaming mechanism. However, when directly applied to AD, this method generated the AEs that are independent of the teacher model and not applicable for distillation, thus limiting the closer match between teacher and student models.</p>
<p>To address this limitation and generate sample-dependent attack strategies advantageous to distillation, reference [<xref ref-type="bibr" rid="ref-12">12</xref>] is improved and introduced into the AD framework. This improvement considers the differences in output between the target and teacher models, resulting in a closer match between them.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Adversarial Distillation</title>
<p>Recently, there has been a growing body of research highlighting the potential for improving AT through the integration of KD. KD offers data-driven soft labels to smooth the hard labels. In Adversarially Robust Distillation (ARD) [<xref ref-type="bibr" rid="ref-19">19</xref>], the target model was encouraged to mimic the softmax output of an adversarially pre-trained teacher model on clean input when facing an adversary. In Robust Soft Label Adversarial Distillation (RSLAD) [<xref ref-type="bibr" rid="ref-21">21</xref>], the generation of AEs and the training of target models were guided by the Robust Soft Labels (RSLs) derived from adversarially pre-trained teachers. Adversarial Knowledge Distillation (AKD) [<xref ref-type="bibr" rid="ref-20">20</xref>] leveraged a linear combination of the AEs&#x2019; predictions from the teacher model and the original labels, effectively guiding the student model&#x2019;s predictions on AE.</p>
<p>However, these methods only utilize the knowledge of adversarially pre-trained teachers to enhance the adversarial robustness of the target model, overlooking considerations related to clean accuracy. Chen et al. [<xref ref-type="bibr" rid="ref-23">23</xref>] imposed the adversarial predictions of the target model to mimic those of standard teachers and robust teachers, hereinafter referred to as self-teacher training (STS). This method notably improves accuracy on both clean samples and AEs, yet it heavily relies on trust in teacher models. IAD [<xref ref-type="bibr" rid="ref-25">25</xref>] highlighted the diminishing reliability of teacher guidance, advocating for a gradual development of confidence in the student model&#x2019;s adversarial robustness while partially trusting the teacher model. The methods mentioned earlier predominantly focus on distilling logit knowledge from the teacher model. Vanilla Feature Distillation Adversarial Training (VFD-Adv) [<xref ref-type="bibr" rid="ref-22">22</xref>] distilled feature knowledge from the teacher&#x2019;s intermediate layer, aligning features of clean examples from the teacher model with those from the student model in the feature space. We utilize logit distillation since it requires less computational and storage costs and logits are at a higher semantic level than deep features.</p>
<p>The baseline in our paper is STS presented in [<xref ref-type="bibr" rid="ref-23">23</xref>], and we use the same AD framework. The adversarial robustness and clean accuracy of the target model are simultaneously improved by leveraging the standard and robust teachers to provide clean and robust knowledge, respectively. Recognizing the decreasing reliability of teacher knowledge during training [<xref ref-type="bibr" rid="ref-24">24</xref>], we update the parameters of teacher models by incorporating supervision from the training set and AEs, as well as feedback from the target model&#x2019;s performances on the validation set and AEs.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Learning to Teach Distillation</title>
<p>Current AD techniques employ the conventional two-stage offline KD technique, where the teacher model&#x2019;s parameters remain unchanged during the distillation process. However, this technique cannot guarantee a match between the teacher and student models, especially when there is a significant difference in predictive performance between them. Additionally, two-stage offline KD cannot adjust the knowledge transfer process in real time based on the learning status of the student model. To address these issues, L2T distillation has been proposed [<xref ref-type="bibr" rid="ref-26">26</xref>], which involves training the student model and fine-tuning the teacher model simultaneously, allowing the teacher model to adjust its behavior based on the feedback from the student model.</p>
<p>Online distillation [<xref ref-type="bibr" rid="ref-27">27</xref>&#x2013;<xref ref-type="bibr" rid="ref-30">30</xref>] is a commonly used L2T algorithm, which involves simultaneously training the student and teacher models and ensuring similarity between their outputs on the training set by minimizing the Kullback-Leibler (KL) divergence between them. However, this only considers the knowledge transfer on the training set without considering the validation performance of the student model. Meta distillation [<xref ref-type="bibr" rid="ref-31">31</xref>&#x2013;<xref ref-type="bibr" rid="ref-34">34</xref>] addresses this issue by fine-tuning the teacher model to minimize the loss of the updated student on the validation set. However, the teacher model only receives supervision from the student model, which can result in performance degradation.</p>
<p>Recently, Ren et al. [<xref ref-type="bibr" rid="ref-35">35</xref>] proposed a novel L2T distillation framework called Learning Good Teacher Matters (LGTM), which introduced the distillation influence to assign a loss weight to each training sample based on the student model&#x2019;s performance on the validation set. However, this method does not consider the accuracy of the target model on AEs as feedback to fine-tune the robust teacher.</p>
<p>To improve the reliability and effectiveness of the standard and robust teachers&#x2019; knowledge in the generalization ability of the target model on both clean samples and AEs, LGTM [<xref ref-type="bibr" rid="ref-35">35</xref>] is extended and incorporated into the AD framework. We use feedback from the target model on the validation data and AEs to update both standard and robust teachers. Fine-tuning the teachers narrows the capacity gap between the teacher and target models and makes teacher models more adaptable to the stronger AEs, increasing their reliability. Additionally, due to the involvement of teacher knowledge in the AE generation in our method, more reliable teachers can also improve the quality of AEs.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Method</title>
<sec id="s3_1">
<label>3.1</label>
<title>Method Overview</title>
<p>Existing AD techniques employ fixed and sample-agnostic attack strategies that are centered around the target model, which leads to AEs being irrelevant to the teacher models and unsuitable for AD. Besides, static teachers face challenges in accurately predicting stronger AEs generated by the increased robustness of the target model. Distilling unreliable knowledge can hurt the performance of the target model. To enhance the suitability of AEs for distillation and improve the reliability and effectiveness of teachers&#x2019; knowledge in promoting the generalization performance of the target model, LDAS&#x0026;ET-AD is proposed to generate AEs by leveraging a learnable distillation attack strategies generating mechanism that considers prediction differences between the teacher and target models, as well as update teachers by using a teacher evolution strategy that takes into account the performance of the target model on validation set and AEs. The proposed AD framework, depicted in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>, comprises a target model, a strategy model, and standard and adversarially pre-trained teacher models.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>The framework of proposed LDAS&#x0026;ET-AD. Given a clean training image <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, the strategy model generates an attack strategy <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mi>a</mml:mi></mml:math></inline-formula>. The target model utilizes <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mi>a</mml:mi></mml:math></inline-formula> to generate an AE <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>. The update of the target model consists of experimental updates and actual updates. In each training step, we first obtain a copy of the target model and experimentally update it using the AD loss. Then, we sample <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> from the validation set and generate AEs <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, and calculate the losses of the copied model on these samples. The losses provide feedback signals to fine-tune the teachers by calculating adversarial distillation influence. The losses of the teachers on the training set and AEs are also utilized to update teachers. Finally, we discard the copied target model and employ the updated teachers to guide the training of the target model on the same training batches and AEs</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-1.tif"/>
</fig>
<p>The training process consists of two stages: Generating AEs and fine-tuning teachers.
<list list-type="order">
<list-item>
<p>In the stage of generating AEs, the target model and the strategy model compete with each other in minimizing or maximizing the same objective function. The strategy model is trained to automatically generate attack strategies that produce AEs capable of misleading the target model and inducing maximum divergence between the target and teacher models. The target model is trained to defend against AEs generated by the attack strategies while receiving guidance from both standard and adversarially pre-trained teachers to minimize the prediction distance with them.</p></list-item>
<list-item>
<p>In the stage of fine-tuning teachers, a temporary copy of the target model first performs experimental AD and provides feedback for fine-tuning teachers based on its accuracy on the validation set and AEs. The standard and adversarially pre-trained teachers are then fine-tuned based on their performances on the training set and AEs, respectively, as well as the feedback provided by the temporary copy of the target model. Finally, the parameters of the target model are actually updated under the guidance of fine-tuned teachers&#x2019; knowledge.</p></list-item>
</list></p>
<p>In the subsequent section, we provide a detailed description of the learnable distillation attack strategies generating mechanism that considers prediction differences, as well as the teacher evolution strategy that takes into account the validation performance of the target model. The equation symbols and abbreviations used throughout this paper are summarized in <xref ref-type="table" rid="table-1">Tables 1</xref> and <xref ref-type="table" rid="table-2">2</xref>, respectively.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>The symbols of the equations used in this paper</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Symbols</th>
<th>Meanings</th>
<th>Symbols</th>
<th>Meanings</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:mi>a</mml:mi></mml:math></inline-formula></td>
<td>An attack strategy. It is determined by the values chosen for attack parameters, such as the maximal perturbation strength <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>, the attack step size <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula>, and the attack iteration <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:mi>I</mml:mi></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula><break/>The adversarial example of a given clean image <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> by an attack strategy <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mi>a</mml:mi></mml:math></inline-formula> at the m<sup>th</sup> training step, <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the PGD attack</td>
</tr>
<tr>
<td><inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula></td>
<td>The adversarial perturbation</td>
<td><inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula></td>
<td>The adversarial example of <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> at step <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mi>n</mml:mi></mml:math></inline-formula></td>
</tr>
<tr>
<td><inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:munder><mml:mo>&#x220F;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0212C;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>The projection function that projects the AEs back into the <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>-ball centered at <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>A distribution of the clean training examples <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and the ground truth labels <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
</tr>
<tr>
<td><inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></inline-formula></td>
<td>The cross-entropy loss between <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>The target model at the m<sup>th</sup> training step. <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the model parameters</td>
</tr>
<tr>
<td><inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula><break/>The adversarial examples of training samples <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></inline-formula></td>
<td>The Kullback-Leibler divergence between <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
</tr>
<tr>
<td><inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mi>f</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mi>v</mml:mi></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:mi>&#x03B3;</mml:mi></mml:math></inline-formula></td>
<td>A small scalar</td>
</tr>
<tr>
<td><inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula><inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>The static standard and adversarially pre-trained teachers</td>
<td><inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula>, <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td>The parameters of the standard and robust teachers after fine-tuning</td>
</tr>
<tr>
<td><inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td>The parameters of the standard and robust teachers before fine-tuning</td>
<td><inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>Two hyperparameters to control the guidance ratio of standard and robust teachers</td>
</tr>
<tr>
<td><inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td>The parameters of the target model after the update</td>
<td><inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>The parameters of the strategy model</td>
</tr>
<tr>
<td><inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula></td>
<td>The batch size of a training batch</td>
<td><inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:msub><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:msub><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>&#x003A;</mml:mo><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x2264;</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>. A bound</td>
</tr>
<tr>
<td><inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:mi>k</mml:mi></mml:math></inline-formula></td>
<td>A hyperparameter to control alternative updates of <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. We update <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> every <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:mi>k</mml:mi></mml:math></inline-formula> times of updating <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>The loss ratios to control the self-evolution of standard teacher and robust teacher, respectively</td>
</tr>
<tr>
<td><inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> The AEs of validation samples <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> are their ground truth labels</td>
<td><inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:msubsup><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula></td>
<td><inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:msubsup><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>a</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> The adversarial examples generated by another attack strategy <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mrow><mml:mover><mml:mi>a</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula> which is used to evaluate the robustness of the one-step updated target model <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula></td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>The abbreviations used in this paper</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Abbreviations</th>
<th>Symbols</th>
<th>Abbreviations</th>
<th>Symbols</th>
</tr>
</thead>
<tbody>
<tr>
<td>Adversarial examples</td>
<td>AEs</td>
<td>Adversarial training</td>
<td>AT</td>
</tr>
<tr>
<td>Knowledge distillation</td>
<td>KD</td>
<td>Carlini wagner attack</td>
<td>CW</td>
</tr>
<tr>
<td>Learnable Distillation Attack Strategies and Evolvable Teachers Adversarial Distillation</td>
<td>LDAS&#x0026;ET-AD</td>
<td>TRadeoff-inspired adversarial defense via Surrogate-loss minimization</td>
<td>TRADES</td>
</tr>
<tr>
<td>Learning to teach</td>
<td>L2T</td>
<td>Fast gradient sign method</td>
<td>FGSM</td>
</tr>
<tr>
<td>Deep neural networks</td>
<td>DNNs</td>
<td>Projection gradient descent method</td>
<td>PGD</td>
</tr>
<tr>
<td>AutoAttack</td>
<td>AA</td>
<td>Standard AT</td>
<td>SAT</td>
</tr>
<tr>
<td>An adversarial training framework by introducing a learnable attack strategy</td>
<td>LAS-AT</td>
<td>Curriculum distillation attack strategy and evolvable teachers adversarial distillation</td>
<td>CDAS&#x0026;ET-AD</td>
</tr>
<tr>
<td>Curriculum adversarial training</td>
<td>CAT</td>
<td>Friendly adversarial training</td>
<td>FAT</td>
</tr>
<tr>
<td>First-order stationary condition</td>
<td>FOSC</td>
<td>Adversarially robust distillation</td>
<td>ARD</td>
</tr>
<tr>
<td>Robust soft label adversarial distillation</td>
<td>RSLAD</td>
<td>Multi-teacher adversarial robustness distillation</td>
<td>MTARD</td>
</tr>
<tr>
<td>Adversarial knowledge distillation</td>
<td>AKD</td>
<td>Self-teachers training</td>
<td>STS</td>
</tr>
<tr>
<td>Robust soft labels</td>
<td>RSLs</td>
<td>Introspective adversarial distillation</td>
<td>IAD</td>
</tr>
<tr>
<td>Vanilla feature distillation adversarial training</td>
<td>VFD-Adv</td>
<td>Kullback-Leibler</td>
<td>KL</td>
</tr>
<tr>
<td>Learning good teacher matters</td>
<td>LGTM</td>
<td>Ground truth</td>
<td>GT</td>
</tr>
<tr>
<td>Cross-entropy</td>
<td>CE</td>
<td>Standard training</td>
<td>ST</td>
</tr>
<tr>
<td>Stochastic gradient descent</td>
<td>SGD</td>
<td>N-step PGD</td>
<td>PGD-N</td>
</tr>
<tr>
<td>MobileNet-V2</td>
<td>MN-V2</td>
<td>ResNet-18</td>
<td>RN-18</td>
</tr>
<tr>
<td>Fixed distillation attack strategy and evolvable teachers adversarial distillation</td>
<td>FDAS&#x0026;ET-AD</td>
<td>Learnable attack strategy and evolvable teachers adversarial distillation</td>
<td>LAS&#x0026;ET-AD</td>
</tr>
<tr>
<td>Learnable distillation attack strategies and evolvable robust teachers adversarial distillation</td>
<td>LDAS&#x0026;ERoT-AD</td>
<td>Learnable distillation attack strategies and evolvable standard teachers adversarial distillation</td>
<td>LDAS&#x0026;EStT-AD</td>
</tr>
<tr>
<td>Learnable distillation attack strategies adversarial distillation</td>
<td>LDAS-AD</td>
<td>Misclassification-aware adversarial training</td>
<td>MART</td>
</tr>
<tr>
<td>First-order stationary condition distillation attack strategy and evolvable teachers adversarial distillation</td>
<td>FOCSDAS&#x0026;ET-AD</td>
<td>Friendly distillation attack strategy and evolvable teachers adversarial distillation</td>
<td>FriDAS&#x0026;ET-AD</td>
</tr>
<tr>
<td>Learnable distillation attack strategy and mate adversarial distillation</td>
<td>LDAS&#x0026;meta-AD</td>
<td>Learnable distillation attack strategy and online adversarial distillation</td>
<td>LDAS&#x0026;OL-AD</td>
</tr>
<tr>
<td>Adversarial distillation</td>
<td>AD</td>
<td></td>
<td></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Learnable Distillation Attack Strategies Generating Mechanism Considering Prediction Differen- ces between Teacher and Target Models</title>
<p>An attack strategy is determined by the values chosen for attack parameters, such as the maximal perturbation strength <inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>, attack step size <inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula>, and attack iteration <inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:mi>I</mml:mi></mml:math></inline-formula>. These parameters play a crucial role in the inner optimization problem of AT, significantly impacting performance. Given a clean image <inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> and its ground truth (GT) label <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, the generation of AE <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> using an attack strategy <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mi>a</mml:mi></mml:math></inline-formula> at the m<sup>th</sup> training step can be defined as follows:
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> represents the adversarial perturbation, <inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is the parameters of the target model at the m<sup>th</sup> training step, <inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denotes the PGD attack employed in our method following [<xref ref-type="bibr" rid="ref-13">13</xref>]. Concretely, PGD recursively searches:
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mo movablelimits="false">&#x220F;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0212C;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>g</mml:mi><mml:mi>n</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>until a stopping criterion is met. <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula> are the AEs at step <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:mi>n</mml:mi></mml:math></inline-formula>, and <inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:munder><mml:mo>&#x220F;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x0212C;</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow></mml:munder><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the projection function that projects the AEs back into the <inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:mi>&#x03B5;</mml:mi></mml:math></inline-formula>-ball centered at <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula>, <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is the cross-entropy (CE) loss between <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>n</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:msup><mml:mrow><mml:msub><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>. For simplicity, we annotate <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mi>f</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mi mathvariant="normal">&#x2202;</mml:mi><mml:mi>v</mml:mi></mml:math></inline-formula> as <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<p>Current AD techniques still rely on fixed sample-agnostic and student-egocentric attack strategies, where the attack parameters are artificially set and remain unchanged during training. The loss function for current AD at the m<sup>th</sup> training step can be expressed as:
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:munder><mml:mrow><mml:mo form="prefix">min</mml:mo></mml:mrow><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x223C;</mml:mo><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:munder><mml:mrow><mml:mo form="prefix">max</mml:mo></mml:mrow><mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.047em" minsize="2.047em">]</mml:mo></mml:mrow></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> denotes the target model at the m<sup>th</sup> training step, <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> represent the static standard and adversarially pre-trained teachers, respectively. <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> denotes the distribution of the clean training examples <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and their GT labels <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. <inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:msub><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> represents a bound defined as <inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:msub><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>&#x003A;</mml:mo><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mi>&#x1D4AB;</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>&#x2264;</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>}</mml:mo></mml:mrow></mml:math></inline-formula>. <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> are the AEs of <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> represents the CE loss of the target model between <inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-101"><mml:math id="mml-ieqn-101"><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>. <inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> are the KL divergence between <inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and between <inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-109"><mml:math id="mml-ieqn-109"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are hyperparameters that control the guidance ratio of the standard and robust teachers, respectively. The target model is trained to minimize both the AT loss and the prediction distance with standard and adversarial pre-trained teachers on AEs. The process of AE generation in existing AD methods is illustrated in <xref ref-type="fig" rid="fig-2">Fig. 2a</xref>, which results in AEs unsuitable for distillation, limiting the closeness between the teacher and target models.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Comparison of the attack strategies of vanilla AD and our LDAS&#x0026;ET-AD</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-2.tif"/>
</fig>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>The AD Loss of Target Model</title>
<p>To enhance the suitability of AEs for distillation, a learnable distillation attack strategies generating mechanism that takes into account the prediction disparities between the teacher and target models is introduced into the AD framework. A strategy model is utilized to automatically produce sample-dependent attack strategies by competing with the target model in minimizing or maximizing the AD loss. Consequently, the generated AEs not only mislead the target model but also maximize the difference in predictions between the target and teacher models. In this worst-case scenario of AD, updating the parameters of the target model towards correctly classifying and minimizing the difference makes the AEs more suitable for distillation and brings the target and teacher models closer together. The attack strategies are based on the given samples <inline-formula id="ieqn-110"><mml:math id="mml-ieqn-110"><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, where the attack parameters are related to the strategy model&#x2019;s parameters <inline-formula id="ieqn-111"><mml:math id="mml-ieqn-111"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and the samples <inline-formula id="ieqn-112"><mml:math id="mml-ieqn-112"><mml:mi>a</mml:mi><mml:mo>&#x223C;</mml:mo><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>a</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>;</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. The loss function of AD with a learnable distillation attack strategies generating mechanism can be written as follows:
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:munder><mml:mrow><mml:mo form="prefix">min</mml:mo></mml:mrow><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x223C;</mml:mo><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:munder><mml:mrow><mml:mo form="prefix">max</mml:mo></mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mo>&#x223C;</mml:mo><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>a</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mstyle scriptlevel="0"><mml:mrow><mml:mo maxsize="2.047em" minsize="2.047em">]</mml:mo></mml:mrow></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>The Evaluating Loss of Strategy Model</title>
<p>The evaluation metric proposed in [<xref ref-type="bibr" rid="ref-12">12</xref>] serves as a guiding principle for the training of the strategy model in our approach. First, an attack strategy <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:mi>a</mml:mi></mml:math></inline-formula> is employed to create AEs and then the target model is updated based on these samples using first-order gradient descent for one step, as described in <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref>. If the updated target model can effectively defend against the AEs generated by another attack strategy <inline-formula id="ieqn-114"><mml:math id="mml-ieqn-114"><mml:mrow><mml:mover><mml:mi>a</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-115"><mml:math id="mml-ieqn-115"><mml:mi>a</mml:mi></mml:math></inline-formula> can be considered effective. The evaluation metric of robustness can be defined as follows:
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-116"><mml:math id="mml-ieqn-116"><mml:msubsup><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>a</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> presents the AEs generated by another attack strategy <inline-formula id="ieqn-117"><mml:math id="mml-ieqn-117"><mml:mrow><mml:mover><mml:mi>a</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula>, which is used to evaluate the robustness of the one-step updated target model <inline-formula id="ieqn-118"><mml:math id="mml-ieqn-118"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<p>Furthermore, an effective attack strategy should ensure good performance in predicting clean samples. Thus, we also consider the performance of the one-step updated target model in predicting clean samples for training the strategy model. The evaluation metric of clean accuracy can be defined as follows:
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
</sec>
<sec id="s3_2_3">
<label>3.2.3</label>
<title>The AD Process with Learnable Distillation Attack Strategies</title>
<p>During the initial training stage, the target model is susceptible to attacks and there are significant differences in predictions between the target and pre-trained teacher models. Therefore, effective attack strategies can be easily generated by the strategy model. As the training process progresses, the target model becomes more robust, and the prediction differences decrease. Consequently, the strategy model needs to learn how to generate attack strategies that can produce stronger AEs.</p>
<p>The game formulation between the target and teacher models can be defined as follows:
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:munder><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:munder><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x223C;</mml:mo><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>[</mml:mo><mml:munder><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:msub><mml:mi>E</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mo>&#x223C;</mml:mo><mml:mi>p</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>a</mml:mi><mml:mrow><mml:mo stretchy="false">|</mml:mo></mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>;</mml:mo><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-119"><mml:math id="mml-ieqn-119"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mspace width="thinmathspace" /><mml:mo>+</mml:mo><mml:mspace width="thinmathspace" /><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is a function of the parameters of the target model, strategy model, and two teacher models. <inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> involve the parameters of the strategy model. <inline-formula id="ieqn-123"><mml:math id="mml-ieqn-123"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-124"><mml:math id="mml-ieqn-124"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> are the trade-off hyperparameters of the two loss terms. The target model and strategy model are alternatively optimized using the REINFORCE algorithm [<xref ref-type="bibr" rid="ref-12">12</xref>]. The alternative update is controlled by a hyperparameter <inline-formula id="ieqn-125"><mml:math id="mml-ieqn-125"><mml:mi>k</mml:mi></mml:math></inline-formula>, where we update <inline-formula id="ieqn-126"><mml:math id="mml-ieqn-126"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> every <inline-formula id="ieqn-127"><mml:math id="mml-ieqn-127"><mml:mi>k</mml:mi></mml:math></inline-formula> times of updating <inline-formula id="ieqn-128"><mml:math id="mml-ieqn-128"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. <xref ref-type="fig" rid="fig-2">Fig. 2b</xref> illustrates the generation process of AEs in our proposed LDAS&#x0026;ET-AD.</p>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Teacher Evolution Strategy Considering the Validation Performance of the Target Model</title>
<p>As the robustness of the target model increases and AEs become stronger, the reliability of static teachers&#x2019; knowledge diminishes. This unreliable guidance not only negatively impacts the performance of the target model, but also affects the quality of AEs that rely on the knowledge of teacher models. To enhance the reliability and effectiveness of teachers&#x2019; knowledge in promoting the generalization performance of the target model, a teacher evolution strategy is introduced in our AD framework. This strategy takes into consideration the validation performance of the target model. The feedback for fine-tuning teachers is determined by the adversarial distillation influence, which extends the distillation influence proposed in [<xref ref-type="bibr" rid="ref-35">35</xref>].</p>
<sec id="s3_3_1">
<label>3.3.1</label>
<title>Adversarial Distillation Influence</title>
<p>To ensure both clean accuracy and adversarial robustness of the target model, it is necessary to update both standard and adversarially pre-trained teachers. Therefore, we expand the distillation influence and difference approximation method [<xref ref-type="bibr" rid="ref-35">35</xref>], which does not consider adversarial robustness. The adversarial distillation influence measures the change in clean accuracy and adversarial robustness of the target model on validation data and AEs when the AE of a training sample is included in the AD process. Specifically, the adversarial distillation influence of the standard teacher is determined by calculating the similarity of gradients between the AE of the training sample <inline-formula id="ieqn-129"><mml:math id="mml-ieqn-129"><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> before updating the target model parameters and the validation batch <inline-formula id="ieqn-130"><mml:math id="mml-ieqn-130"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> after updating (<xref ref-type="disp-formula" rid="eqn-8">Eq. (8)</xref>). The adversarial distillation influence of the robust teacher is obtained by calculating the similarity of gradients between the AE of the training sample <inline-formula id="ieqn-131"><mml:math id="mml-ieqn-131"><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> before updating and the AE of the validation batch <inline-formula id="ieqn-132"><mml:math id="mml-ieqn-132"><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> after updating (<xref ref-type="disp-formula" rid="eqn-9">Eq. (9)</xref>).
<disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula>
<disp-formula id="eqn-9"><label>(9)</label><mml:math id="mml-eqn-9" display="block"><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mi mathvariant="normal">&#x005F;</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mi mathvariant="normal">&#x2207;</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mi>y</mml:mi><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-133"><mml:math id="mml-ieqn-133"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula id="ieqn-134"><mml:math id="mml-ieqn-134"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:math></inline-formula> are the parameters of the target model before and after the update, respectively. <inline-formula id="ieqn-135"><mml:math id="mml-ieqn-135"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula id="ieqn-136"><mml:math id="mml-ieqn-136"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> are the parameters of the standard and robust teachers before fine-tuning.</p>
</sec>
<sec id="s3_3_2">
<label>3.3.2</label>
<title>The Fine-Tuning Loss of Teacher Models</title>
<p>The adversarial distillation influence highlights the importance of each training sample&#x2019;s AE in improving the target model&#x2019;s generalization performance. Therefore, we consider it as feedback from the target model&#x2019;s performance on the verification set and use it to assign a weight to each AE for fine-tuning the teacher models. This fine-tuning process enhances the teachers&#x2019; teaching abilities. The weighted fine-tuning losses can be defined as <xref ref-type="disp-formula" rid="eqn-10">Eq. (10)</xref> for the standard teacher and <xref ref-type="disp-formula" rid="eqn-11">Eq. (11)</xref> for the robust teacher:
<disp-formula id="eqn-10"><label>(10)</label><mml:math id="mml-eqn-10" display="block"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:munderover><mml:msup><mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula>
<disp-formula id="eqn-11"><label>(11)</label><mml:math id="mml-eqn-11" display="block"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:munderover><mml:msup><mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-137"><mml:math id="mml-ieqn-137"><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> is the batch size of a training batch, <inline-formula id="ieqn-138"><mml:math id="mml-ieqn-138"><mml:msup><mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-139"><mml:math id="mml-ieqn-139"><mml:msup><mml:mrow><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msubsup><mml:mi>z</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>. And we approximate them by:
<disp-formula id="eqn-12"><label>(12)</label><mml:math id="mml-eqn-12" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2248;</mml:mo><mml:msub><mml:mover><mml:mi>L</mml:mi><mml:mo>&#x223C;</mml:mo></mml:mover><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mspace width="thinmathspace" /></mml:mtd><mml:mtd><mml:mfrac><mml:mn>1</mml:mn><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mfrac><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:munderover><mml:mrow><mml:mo>(</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mo>+</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x03B3;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mo>&#x2212;</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x03B3;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-13"><label>(13)</label><mml:math id="mml-eqn-13" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2248;</mml:mo><mml:msub><mml:mover><mml:mi>L</mml:mi><mml:mo>&#x223C;</mml:mo></mml:mover><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mspace width="thinmathspace" /></mml:mtd><mml:mtd><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mfrac><mml:mn>1</mml:mn><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mfrac></mml:mstyle><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msup><mml:mi>B</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:munderover><mml:mrow><mml:mo>(</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mo>+</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x03B3;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mo>&#x2212;</mml:mo></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msup><mml:mrow><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mn>2</mml:mn><mml:mi>&#x03B3;</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>where <inline-formula id="ieqn-140"><mml:math id="mml-ieqn-140"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mo>&#x00B1;</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x00B1;</mml:mo><mml:mi>&#x03B3;</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-141"><mml:math id="mml-ieqn-141"><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mo>&#x00B1;</mml:mo></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x00B1;</mml:mo><mml:mi>&#x03B3;</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-142"><mml:math id="mml-ieqn-142"><mml:mi>&#x03B3;</mml:mi></mml:math></inline-formula> is a small scalar.</p>
<p>In addition to improving teaching abilities, teacher models should also focus on minimizing CE loss related to GT labels (clean accuracy for the standard teacher and adversarial robustness for the robust teacher). This is crucial for optimizing their reasoning performance. The overall losses for fine-tuning the standard teacher and robust teacher can be defined as <xref ref-type="disp-formula" rid="eqn-14">Eqs. (14)</xref> and <xref ref-type="disp-formula" rid="eqn-15">(15)</xref>, respectively.
<disp-formula id="eqn-14"><label>(14)</label><mml:math id="mml-eqn-14" display="block"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>L</mml:mi><mml:mo stretchy="false">&#x007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p><disp-formula id="eqn-15"><label>(15)</label><mml:math id="mml-eqn-15" display="block"><mml:mtable columnalign="left" rowspacing="4pt" columnspacing="1em"><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mover><mml:mi>L</mml:mi><mml:mo stretchy="false">&#x007E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>a</mml:mi><mml:mi>u</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>y</mml:mi><mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>l</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x22C5;</mml:mo><mml:mi>K</mml:mi><mml:mi>L</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msubsup><mml:mi>x</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>where the hyperparameters <inline-formula id="ieqn-143"><mml:math id="mml-ieqn-143"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-144"><mml:math id="mml-ieqn-144"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> control the self-evolution of standard teacher and robust teacher, respectively.</p>
</sec>
<sec id="s3_3_3">
<label>3.3.3</label>
<title>The Fine-Tuning Process</title>
<p>To obtain adversarial distillation influence involving gradients before and after updating the target model parameters, an experimental update mechanism is introduced as shown in <xref ref-type="fig" rid="fig-3">Fig. 3a</xref>. First, a temporary copy of the current target model <inline-formula id="ieqn-145"><mml:math id="mml-ieqn-145"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> is created. This copy is then experimentally updated by applying the AD loss on the AEs generated using the learnable distillation attack strategies generating mechanism, as proposed in <xref ref-type="sec" rid="s3_2">Section 3.2</xref>. The losses of the updated copy <inline-formula id="ieqn-146"><mml:math id="mml-ieqn-146"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> on the validation set and their AEs are calculated to obtain the adversarial distillation influence.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>The workflow of teacher evolution strategy in our LDAS&#x0026;ET-AD</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-3.tif"/>
</fig>
<p>The adversarial distillation influence serves as feedback from the target model on the validation set for fine-tuning the teachers <inline-formula id="ieqn-147"><mml:math id="mml-ieqn-147"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-148"><mml:math id="mml-ieqn-148"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> to improve their teaching abilities. Their training performance is also taken into account to achieve self-evolution as described in <xref ref-type="fig" rid="fig-3">Fig. 3b</xref>.</p>
<p>After fine-tuning the teachers, the real target model <inline-formula id="ieqn-149"><mml:math id="mml-ieqn-149"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> is actually updated through the AD of the fine-tuned teachers <inline-formula id="ieqn-150"><mml:math id="mml-ieqn-150"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-151"><mml:math id="mml-ieqn-151"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:msubsup><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup></mml:mrow></mml:msub></mml:math></inline-formula> by <xref ref-type="disp-formula" rid="eqn-7">Eq. (7)</xref>, as depicted in <xref ref-type="fig" rid="fig-3">Fig. 3c</xref>. The entire process of our LDAS&#x0026;ET-AD is presented in Algorithm 1.</p>
<fig id="fig-6">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-6.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experiments</title>
<sec id="s4_1">
<label>4.1</label>
<title>Experiment Setup</title>
<sec id="s4_1_1">
<label>4.1.1</label>
<title>Datasets and Competitive Methods</title>
<p>We conducted experiments on various benchmark datasets, including CIFAR-10 and CIFAR-100 [<xref ref-type="bibr" rid="ref-43">43</xref>]. All models were implemented in PyTorch and trained on a single RTX 2080 Ti GPU. We compared our LDAS&#x0026;ET-AD with baseline STS [<xref ref-type="bibr" rid="ref-23">23</xref>]. Besides, standard training (ST) method and four state-of-the-art adversarial defense methods (SAT [<xref ref-type="bibr" rid="ref-13">13</xref>], TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>], LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>], and IAD [<xref ref-type="bibr" rid="ref-25">25</xref>]) were considered for comparison.</p>
</sec>
<sec id="s4_1_2">
<label>4.1.2</label>
<title>Student, Teacher, and Strategy Models</title>
<p>We considered ResNet-18 [<xref ref-type="bibr" rid="ref-44">44</xref>] and MobileNet-V2 [<xref ref-type="bibr" rid="ref-45">45</xref>] as the target models. Their structures are described in <xref ref-type="table" rid="table-3">Table 3</xref>. The pre-trained models with the same architectures were utilized as self-teachers, following previous work [<xref ref-type="bibr" rid="ref-23">23</xref>]. One model could be trained using either AT or ST way, resulting in two self-teachers: Adversarial and standard pre-trained self-teachers. The models with the same architectures were chosen as the strategy models.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>The architecture of the target networks ResNet-18 and MobileNet-V2</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>ResNet-18</th>
<th>MobileNet-V2</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td>Conv2D (64, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU</td>
<td>Conv2D (32, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU</td>
</tr>
<tr>
<td>[Conv2D (64, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 2</td>
<td>DepthwiseConv2D (16, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (16, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU</td>
</tr>
<tr>
<td>[Conv2D (128, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 2</td>
<td>[DepthwiseConv2D (24, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (24, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 2</td>
</tr>
<tr>
<td>[Conv2D (256, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 2</td>
<td>[DepthwiseConv2D (32, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (32, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 3</td>
</tr>
<tr>
<td>[Conv2D (512, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 2</td>
<td>[DepthwiseConv2D (64, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (64, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 4</td>
</tr>
<tr>
<td>AvgPooling2D ((2, 2))</td>
<td>[DepthwiseConv2D (96, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (96, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 3</td>
</tr>
<tr>
<td>Linear ()</td>
<td>[DepthwiseConv2D (160, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (160, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU] &#x00D7; 3</td>
</tr>
<tr>
<td></td>
<td>DepthwiseConv2D (320, 3 &#x00D7; 3) &#x002B; BatchNorm2D &#x002B; ReLU<break/>PointwiseConv2D (320, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU</td>
</tr>
<tr>
<td></td>
<td>Conv2D (1280, 1 &#x00D7; 1) &#x002B; BatchNorm2D &#x002B; ReLU</td>
</tr>
<tr>
<td></td>
<td>AvgPooling2D ((2, 2))</td>
</tr>
<tr>
<td></td>
<td>Linear ()</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_1_3">
<label>4.1.3</label>
<title>Training Settings</title>
<p>We trained the target models and the pre-trained teachers using the Stochastic Gradient Descent (SGD) optimizer with a momentum of 0.9 and weight decay 5e-4. The training process consisted of 200 epochs with a batch size of 128. The learning rate started from 0.1 for ResNet-18 and 0.01 for MobileNet-V2 and decayed to one-tenth at epochs 50 and 150, respectively. The strategy model in our method employed an SGD momentum optimizer with a learning rate of 0.001 for ResNet-18 and 0.0001 for MobileNet-V2. The pre-trained teachers were fine-tuned using an SGD momentum optimizer with a learning rate of 0.01. For ST, we trained the models for 100 epochs on clean images with standard data augmentations. The learning rate was divided by 10 at the 75<sup>th</sup> and 90<sup>th</sup> epochs. We strictly followed the original settings of SAT [<xref ref-type="bibr" rid="ref-13">13</xref>], TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>], and LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]. For STS [<xref ref-type="bibr" rid="ref-23">23</xref>] and IAD [<xref ref-type="bibr" rid="ref-25">25</xref>], we used the same self-teachers as our LDAS&#x0026;ET-AD. A 10-step PGD (PGD-10) with a random start size of 0.001, step size 2/255 was employed to solve the inner maximization.</p>
<p>In our method, we actually updated the target model every <inline-formula id="ieqn-199"><mml:math id="mml-ieqn-199"><mml:mi>k</mml:mi></mml:math></inline-formula>&#x003D; 30 times updating the strategy model. The hyperparameters <inline-formula id="ieqn-200"><mml:math id="mml-ieqn-200"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-201"><mml:math id="mml-ieqn-201"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> related to the balance between self-evolution and knowledge transfer were set to 0.8 and 0.7, respectively. The trade-off hyperparameters <inline-formula id="ieqn-202"><mml:math id="mml-ieqn-202"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-203"><mml:math id="mml-ieqn-203"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> were set to 2.0 and 4.0. These selections were based on the results of ablation studies in <xref ref-type="sec" rid="s5">Section 5</xref>. The selection of the hyperparameters of attack strategies was followed by [<xref ref-type="bibr" rid="ref-12">12</xref>]. Specifically, the maximum perturbation strength ranged from 3 to 15, the attack step ranged from 1 to 6, and the attack iteration ranged from 3 to 15. We set <inline-formula id="ieqn-204"><mml:math id="mml-ieqn-204"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>S</mml:mi><mml:mi>T</mml:mi><mml:mi>S</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> to 0.25 and <inline-formula id="ieqn-205"><mml:math id="mml-ieqn-205"><mml:msub><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:mi>S</mml:mi><mml:mi>T</mml:mi><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:math></inline-formula> to 0.5, as recommended in [<xref ref-type="bibr" rid="ref-23">23</xref>].</p>
</sec>
<sec id="s4_1_4">
<label>4.1.4</label>
<title>Evaluation Attacks</title>
<p>After training, we evaluated the models against four commonly used adversarial attacks: FGSM [<xref ref-type="bibr" rid="ref-39">39</xref>], PGD [<xref ref-type="bibr" rid="ref-13">13</xref>], CW<sub>&#x221E;</sub> [<xref ref-type="bibr" rid="ref-40">40</xref>], and AA [<xref ref-type="bibr" rid="ref-41">41</xref>]. The maximum perturbation allowed for evaluation was set to 8/255 for both datasets. The perturbation steps for PGD and CW<sub>&#x221E;</sub> were both set to 20. We calculated the natural accuracy (&#x2018;Natural&#x2019; in Tables) on the natural test data and the robust accuracy on the adversarial test data generated by FGSM, PGD, CW<sub>&#x221E;</sub>, and AA attacks, following [<xref ref-type="bibr" rid="ref-24">24</xref>].</p>
</sec>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Adversarial Robustness Evaluation</title>
<p>In accordance with previous studies [<xref ref-type="bibr" rid="ref-24">24</xref>], we reported the test accuracy at both the best checkpoint and the last checkpoint. The best checkpoint of ST is chosen based on its performance on clean test examples, while the best checkpoints of SAT [<xref ref-type="bibr" rid="ref-13">13</xref>], TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>], LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>], STS [<xref ref-type="bibr" rid="ref-23">23</xref>], IAD [<xref ref-type="bibr" rid="ref-25">25</xref>], and our LDAS&#x0026;ET-AD are selected based on their robustness against the PGD attack.</p>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Comparison with Baseline</title>
<p>The test accuracy of our LDAS&#x0026;ET-AD and the baseline STS [<xref ref-type="bibr" rid="ref-23">23</xref>] are presented in <xref ref-type="table" rid="table-4">Table 4</xref> for CIFAR-10 and <xref ref-type="table" rid="table-5">Table 5</xref> for CIFAR-100.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Test accuracy (%) on the CIFAR-10 dataset using our proposed LDAS&#x0026;ET-AD and baseline: STS [<xref ref-type="bibr" rid="ref-23">23</xref>]. MN-V2 and RN-18 are abbreviations of MobileNet-V2 and ResNet-18, respectively. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<th>Model</th>
<th>Method</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="2">RN-18</td>
<td>STS [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>83.15</td>
<td>63.97</td>
<td>51.30</td>
<td>50.61</td>
<td>43.08</td>
<td>83.77</td>
<td>62.71</td>
<td>48.41</td>
<td>48.71</td>
<td>42.72</td>
</tr>
<tr>
<td>Ours</td>
<td><bold>85.20</bold></td>
<td><bold>64.92</bold></td>
<td><bold>53.90</bold></td>
<td><bold>52.14</bold></td>
<td><bold>45.39</bold></td>
<td><bold>85.44</bold></td>
<td><bold>64.44</bold></td>
<td><bold>50.71</bold></td>
<td><bold>50.11</bold></td>
<td><bold>44.24</bold></td>
</tr>
<tr>
<td rowspan="2">MN-V2</td>
<td>STS [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>81.15</td>
<td>62.65</td>
<td>50.10</td>
<td>48.75</td>
<td>39.14</td>
<td>82.27</td>
<td>62.36</td>
<td>48.41</td>
<td>46.43</td>
<td>38.20</td>
</tr>
<tr>
<td>Ours</td>
<td><bold>84.72</bold></td>
<td><bold>66.40</bold></td>
<td><bold>54.32</bold></td>
<td><bold>52.78</bold></td>
<td><bold>42.63</bold></td>
<td><bold>84.96</bold></td>
<td><bold>64.76</bold></td>
<td><bold>51.49</bold></td>
<td><bold>49.52</bold></td>
<td><bold>41.28</bold></td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Test accuracy (%) on the CIFAR-100 dataset using our proposed LDAS&#x0026;ET-AD and baseline: STS [<xref ref-type="bibr" rid="ref-23">23</xref>]. MN-V2 and RN-18 are abbreviations of MobileNet-V2 and ResNet-18, respectively. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<th>Model</th>
<th>Method</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="2">RN-18</td>
<td>STS [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>58.02</td>
<td>35.87</td>
<td>26.83</td>
<td>25.35</td>
<td>25.23</td>
<td>56.75</td>
<td>32.28</td>
<td>23.15</td>
<td>22.19</td>
<td>22.55</td>
</tr>
<tr>
<td>Ours</td>
<td><bold>60.41</bold></td>
<td><bold>39.29</bold></td>
<td><bold>30.52</bold></td>
<td><bold>29.62</bold></td>
<td><bold>29.21</bold></td>
<td><bold>60.22</bold></td>
<td><bold>37.06</bold></td>
<td><bold>27.10</bold></td>
<td><bold>25.49</bold></td>
<td><bold>25.11</bold></td>
</tr>
<tr>
<td rowspan="2">MN-V2</td>
<td>STS [<xref ref-type="bibr" rid="ref-23">23</xref>]</td>
<td>54.28</td>
<td>33.33</td>
<td>24.74</td>
<td>23.07</td>
<td>23.89</td>
<td>53.87</td>
<td>32.48</td>
<td>23.14</td>
<td>21.96</td>
<td>22.35</td>
</tr>
<tr>
<td>Ours</td>
<td><bold>56.87</bold></td>
<td><bold>36.34</bold></td>
<td><bold>27.07</bold></td>
<td><bold>26.39</bold></td>
<td><bold>27.22</bold></td>
<td><bold>56.17</bold></td>
<td><bold>34.64</bold></td>
<td><bold>25.22</bold></td>
<td><bold>23.10</bold></td>
<td><bold>24.74</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Our LDAS&#x0026;ET-AD builds upon the AD framework proposed in [<xref ref-type="bibr" rid="ref-23">23</xref>] which applies a robust teacher and a clean teacher to guide robustness and clean accuracy simultaneously. We have made improvements in two aspects: AE generation and teacher knowledge.</p>
<p>Firstly, instead of using hand-crafted strategies, sample-dependent attack strategies are automatically generated by the strategy network, which takes into account the prediction distance between target and teacher models. This results in more suitable AEs for AD and a closer match of target and teacher models. Secondly, the model parameters of the teachers are fine-tuned based on the validation performance of the target model, rather than being static, making teacher knowledge more helpful in improving the generalization performance of the target model and the quality of AEs involving teacher knowledge.</p>
<p>As shown in <xref ref-type="table" rid="table-4">Tables 4</xref> and <xref ref-type="table" rid="table-5">5</xref>, our LDAS&#x0026;ET-AD outperforms the baseline on both CIFAR-10 and CIFAR-100 datasets, at either the best or the last checkpoints. Specifically, for ResNet-18, LDAS&#x0026;ET-AD improves accuracy by 2.05%, 0.95%, 2.60%, 1.53%, and 2.31% under clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA attacks on CIFAR-10 dataset, and by 2.39%, 3.42%, 3.69%, 4.27%, and 3.98% on CIFAR-100 dataset compared to benchmark results. For MobileNet-V2, LDAS&#x0026;ET-AD brings 3.57%, 3.75%, 4.22%, 4.03%, and 3.49% improvements on CIFAR-10 dataset and 2.59%, 3.01%, 2.33%, 3.32%, and 3.33% improvements on CIFAR-100 dataset.</p>
<p>In conclusion, our LDAS&#x0026;ET-AD consistently improves clean and adversarial accuracy on two commonly used datasets against four attacks when applied to two target models compared to the baseline. This indicates the effectiveness of (I) considering the prediction differences of teacher and target models in the generation of sample-dependent AEs, and (II) fine-tuning the teacher models based on the accuracy of the target model on the validation set and AEs in improving AD.</p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Comparison with State-of-the-Art Adversarial Defense Methods</title>
<p>We present the test results of our LDAS&#x0026;ET-AD framework applied to ResNet-18 and MobileNet-V2 target models in comparison to state-of-the-art adversarial defense methods on CIFAR-10 and CIFAR-100 datasets in <xref ref-type="table" rid="table-6">Tables 6</xref> and <xref ref-type="table" rid="table-7">7</xref>, respectively.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Test accuracy (%) on the CIFAR-10 dataset using our proposed LDAS&#x0026;ET-AD, current commonly used and state-of-the-art defense methods. MN-V2 and RN-18 are abbreviations of MobileNet-V2 and ResNet-18, respectively. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<td>Model</td>
<td>Method</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="6">RN-18</td>
<td>ST</td>
<td><bold>94.95</bold></td>
<td>29.38</td>
<td>0</td>
<td>0</td>
<td>0</td>
<td><bold>94.79</bold></td>
<td>31.62</td>
<td>0</td>
<td>0</td>
<td>0</td>
</tr>
<tr>
<td>SAT [<xref ref-type="bibr" rid="ref-13">13</xref>]</td>
<td>83.96</td>
<td>63.43</td>
<td>49.13</td>
<td>48.83</td>
<td>37.14</td>
<td>84.39</td>
<td>60.23</td>
<td>43.76</td>
<td>44.49</td>
<td>35.00</td>
</tr>
<tr>
<td>TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>]</td>
<td>81.70</td>
<td>64.12</td>
<td>51.22</td>
<td>50.03</td>
<td>41.69</td>
<td>82.62</td>
<td>61.87</td>
<td>46.90</td>
<td>46.62</td>
<td>40.09</td>
</tr>
<tr>
<td>LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]</td>
<td>81.43</td>
<td>63.34</td>
<td>52.62</td>
<td>50.67</td>
<td>42.78</td>
<td>82.93</td>
<td>62.01</td>
<td>49.00</td>
<td>48.91</td>
<td>41.29</td>
</tr>
<tr>
<td>IAD [<xref ref-type="bibr" rid="ref-25">25</xref>]</td>
<td>83.40</td>
<td>63.95</td>
<td>51.32</td>
<td>50.32</td>
<td>43.99</td>
<td>83.45</td>
<td>62.77</td>
<td>48.42</td>
<td>48.37</td>
<td>42.34</td>
</tr>
<tr>
<td>Ours</td>
<td>85.20</td>
<td><bold>64.92</bold></td>
<td><bold>53.90</bold></td>
<td><bold>52.14</bold></td>
<td><bold>45.39</bold></td>
<td>85.44</td>
<td><bold>64.44</bold></td>
<td><bold>50.71</bold></td>
<td><bold>50.11</bold></td>
<td><bold>44.24</bold></td>
</tr>
<tr>
<td rowspan="6">MN-V2</td>
<td>ST</td>
<td><bold>93.04</bold></td>
<td>19.55</td>
<td>0</td>
<td>0</td>
<td>0</td>
<td><bold>92.84</bold></td>
<td>20.04</td>
<td>0</td>
<td>0</td>
<td>0</td>
</tr>
<tr>
<td>SAT [<xref ref-type="bibr" rid="ref-13">13</xref>]</td>
<td>82.07</td>
<td>63.38</td>
<td>48.71</td>
<td>48.01</td>
<td>37.62</td>
<td>82.52</td>
<td>60.96</td>
<td>45.27</td>
<td>45.40</td>
<td>34.05</td>
</tr>
<tr>
<td>TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>]</td>
<td>81.00</td>
<td>64.40</td>
<td>50.15</td>
<td>48.84</td>
<td>39.44</td>
<td>81.34</td>
<td>61.39</td>
<td>47.86</td>
<td>46.68</td>
<td>35.30</td>
</tr>
<tr>
<td>LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]</td>
<td>82.34</td>
<td>64.82</td>
<td>52.46</td>
<td>50.67</td>
<td>40.45</td>
<td>82.93</td>
<td>63.80</td>
<td>49.45</td>
<td>48.63</td>
<td>39.18</td>
</tr>
<tr>
<td>IAD [<xref ref-type="bibr" rid="ref-25">25</xref>]</td>
<td>80.49</td>
<td>62.70</td>
<td>50.75</td>
<td>48.95</td>
<td>41.20</td>
<td>81.41</td>
<td>61.83</td>
<td>48.63</td>
<td>47.46</td>
<td>40.64</td>
</tr>
<tr>
<td>Ours</td>
<td>84.72</td>
<td><bold>66.40</bold></td>
<td><bold>54.32</bold></td>
<td><bold>52.78</bold></td>
<td><bold>42.63</bold></td>
<td>84.96</td>
<td><bold>64.76</bold></td>
<td><bold>51.49</bold></td>
<td><bold>49.52</bold></td>
<td><bold>41.28</bold></td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Test accuracy (%) on the CIFAR-100 dataset using our proposed LDAS&#x0026;ET-AD, current commonly used and state-of-the-art defense methods. MN-V2 and RN-18 are abbreviations of MobileNet-V2 and ResNet-18, respectively. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<td>Model</td>
<td>Method</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="6">RN-18</td>
<td>ST</td>
<td><bold>76.11</bold></td>
<td>3.89</td>
<td>0</td>
<td>0</td>
<td>0</td>
<td><bold>75.91</bold></td>
<td>3.91</td>
<td>0</td>
<td>0</td>
<td>0</td>
</tr>
<tr>
<td>SAT [<xref ref-type="bibr" rid="ref-13">13</xref>]</td>
<td>56.89</td>
<td>33.94</td>
<td>24.01</td>
<td>23.68</td>
<td>22.91</td>
<td>56.11</td>
<td>30.27</td>
<td>19.68</td>
<td>20.09</td>
<td>20.12</td>
</tr>
<tr>
<td>TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>]</td>
<td>55.10</td>
<td>35.12</td>
<td>26.16</td>
<td>25.16</td>
<td>24.36</td>
<td>54.75</td>
<td>32.08</td>
<td>21.44</td>
<td>22.60</td>
<td>21.31</td>
</tr>
<tr>
<td>LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]</td>
<td>55.95</td>
<td>37.25</td>
<td>29.00</td>
<td>27.07</td>
<td>28.01</td>
<td>55.20</td>
<td>31.96</td>
<td>22.52</td>
<td>22.60</td>
<td>20.74</td>
</tr>
<tr>
<td>IAD [<xref ref-type="bibr" rid="ref-25">25</xref>]</td>
<td>56.46</td>
<td>35.75</td>
<td>27.02</td>
<td>25.52</td>
<td>25.75</td>
<td>56.13</td>
<td>32.39</td>
<td>22.45</td>
<td>21.65</td>
<td>21.52</td>
</tr>
<tr>
<td>Ours</td>
<td>60.41</td>
<td><bold>39.29</bold></td>
<td><bold>30.52</bold></td>
<td><bold>29.62</bold></td>
<td><bold>29.21</bold></td>
<td>60.22</td>
<td><bold>37.06</bold></td>
<td><bold>27.10</bold></td>
<td><bold>25.49</bold></td>
<td><bold>25.11</bold></td>
</tr>
<tr>
<td rowspan="6">MN-V2</td>
<td>ST</td>
<td><bold>71.62</bold></td>
<td>3.19</td>
<td>0</td>
<td>0</td>
<td>0</td>
<td><bold>71.44</bold></td>
<td>3.19</td>
<td>0</td>
<td>0</td>
<td>0</td>
</tr>
<tr>
<td>SAT [<xref ref-type="bibr" rid="ref-13">13</xref>]</td>
<td>53.62</td>
<td>31.42</td>
<td>22.82</td>
<td>21.58</td>
<td>20.98</td>
<td>53.82</td>
<td>29.30</td>
<td>19.88</td>
<td>19.74</td>
<td>18.60</td>
</tr>
<tr>
<td>TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>]</td>
<td>52.29</td>
<td>32.41</td>
<td>23.88</td>
<td>22.68</td>
<td>22.86</td>
<td>52.01</td>
<td>30.62</td>
<td>20.16</td>
<td>21.58</td>
<td>20.13</td>
</tr>
<tr>
<td>LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]</td>
<td>54.67</td>
<td>34.36</td>
<td>25.61</td>
<td>24.09</td>
<td>22.22</td>
<td>54.20</td>
<td>31.14</td>
<td>21.86</td>
<td>21.22</td>
<td>19.39</td>
</tr>
<tr>
<td>IAD [<xref ref-type="bibr" rid="ref-25">25</xref>]</td>
<td>53.56</td>
<td>33.32</td>
<td>25.16</td>
<td>23.16</td>
<td>25.99</td>
<td>54.06</td>
<td>33.22</td>
<td>23.60</td>
<td>21.91</td>
<td>23.99</td>
</tr>
<tr>
<td>Ours</td>
<td>56.87</td>
<td><bold>36.34</bold></td>
<td><bold>27.07</bold></td>
<td><bold>26.39</bold></td>
<td><bold>27.22</bold></td>
<td>56.17</td>
<td><bold>34.64</bold></td>
<td><bold>25.22</bold></td>
<td><bold>23.10</bold></td>
<td><bold>24.74</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As shown in the tables, LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>], an AT framework incorporating learnable attack strategies, outperforms SAT [<xref ref-type="bibr" rid="ref-13">13</xref>] and TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>] in terms of adversarial robustness due to the automatic generation of sample-dependent attack strategies. IAD [<xref ref-type="bibr" rid="ref-25">25</xref>] solve the problem of reduced reliability of teacher guidance in AD is alleviated by partially instead of fully trusting the teacher model. These observations highlight the effectiveness of KD, learnable attack strategies, and reliable teachers in enhancing AT on both CIFAR-10 and CIFAR-100 datasets. Our LDAS&#x0026;ET-AD introduces a learnable distillation attack strategies generating mechanism and a teacher evolution strategy into the AD framework to integrate their benefits of them.</p>
<p>Compared to state-of-the-art AT methods (SAT [<xref ref-type="bibr" rid="ref-13">13</xref>], TRADES [<xref ref-type="bibr" rid="ref-10">10</xref>], and LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>]), our proposed method introduces the AD of evolvable teachers, which can provide more reliable soft labels to better smooth hard labels in AT. In addition, maximizing the prediction distance between teacher and target models is introduced to automatically generate attack strategies by the strategy model, making AEs more suitable for distillation and leading to a closer match between the teacher and target models. The results in <xref ref-type="table" rid="table-6">Tables 6</xref> and <xref ref-type="table" rid="table-7">7</xref> demonstrate superior clean accuracy and robustness against four different attacks on both CIFAR-10 and CIFAR-100 datasets. Specifically, for ResNet-18, our LDAS&#x0026;ET-AD outperforms the best AT method with improvements of 1.24%, 0.80%, 1.28%, 1.47%, and 2.61% in clean, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy on CIFAR-10 dataset, and 3.52%, 2.04%, 1.52%, 2.55%, and 1.20% on CIFAR-100 dataset. For MobileNet-V2, our proposed method improves accuracy by 2.38%, 1.58%, 1.86%, 2.11%, and 2.18% on CIFAR-10 dataset, and 2.20%, 1.98%, 1.46%, 2.30%, and 4.36% on CIFAR-100 dataset.</p>
<p>IAD [<xref ref-type="bibr" rid="ref-25">25</xref>] encourages the target model to partially trust the teacher models and gradually trust itself more as the teacher models become progressively unreliable. The teacher knowledge in our proposed method has a more significant effect on improving the generalization performance of the target model since the teacher models in our method are updated based on the validation performance of the target model. Besides, the generation of sample-dependent attack strategies that consider teacher knowledge enhances the quality of AEs. The results highlight the superior performance of our LDAS&#x0026;ET-AD on both CIFAR-10 and CIFAR-100 datasets. Specifically, our LDAS&#x0026;ET-AD improves the accuracy of ResNet-18 by 1.80%, 0.97%, 2.58%, 1.82%, and 1.40% in terms of clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy on CIFAR-10 dataset, and 3.95%, 3.54%, 3.50%, 4.10%, and 3.46% on CIFAR-100 dataset. For MobileNet-V2, our LDAS&#x0026;ET-AD shows improvements of 4.23%, 3.70%, 3.57%, 3.83%, and 1.43% on CIFAR-10 dataset, and 3.31%, 3.02%, 1.91%, 3.23%, and 1.23% on CIFAR-100 dataset.</p>
<p>Overall, our LDAS&#x0026;ET-AD surpasses state-of-the-art adversarial defense methods against various attacks using different models due to the more reliable teachers and more suitable AEs for distillation by introducing the learnable distillation attack strategies generating mechanism that considers prediction differences between the teacher and target models, as well as the teacher evolution strategy that takes into account the validation performance of target model in the AD framework.</p>
</sec>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Analysis and Discussion</title>
<p>To comprehensively understand our LDAS&#x0026;ET-AD, we conducted a series of experiments on the CIFAR-10 dataset. These experiments encompassed ablation studies of each component, utilization of diverse dynamic attack strategies generating methods, adoption of distinct teacher fine-tuning methods based on L2T distillation, exploration of different <inline-formula id="ieqn-206"><mml:math id="mml-ieqn-206"><mml:mi>k</mml:mi></mml:math></inline-formula> concerning the optimized frequency of the strategy model, examination of different <inline-formula id="ieqn-207"><mml:math id="mml-ieqn-207"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-208"><mml:math id="mml-ieqn-208"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> associated with the self-evolution of the teachers, and investigation of different <inline-formula id="ieqn-209"><mml:math id="mml-ieqn-209"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-210"><mml:math id="mml-ieqn-210"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> related to the trade-off between robustness and clean accuracy. Subsequently, we delve into the training and inference complexity of our LDAS&#x0026;ET-AD. The ResNet-18 model was selected as the backbone model.</p>
<sec id="s5_1">
<label>5.1</label>
<title>Ablation of LDAS&#x0026;ET-AD</title>
<p>We conducted a set of ablation studies to better grasp the impact of each component in our LDAS&#x0026;ET-AD.</p>
<p>Firstly, the learnable distillation attack strategies generating mechanism in our LDAS&#x0026;ET-AD was replaced with the fixed distillation attack strategy in STS [<xref ref-type="bibr" rid="ref-23">23</xref>] considering prediction differences between student and teacher, denoted as Fixed Distillation Attack Strategy and Evolvable Teachers Adversarial Distillation (FDAS&#x0026;ET-AD), to verify the effectiveness of the introduction of learnable attack strategies. Besides, this mechanism was replaced with the learnable attack strategies in LAS-AT [<xref ref-type="bibr" rid="ref-12">12</xref>], denoted as Learnable Attack Strategy and Evolvable Teachers Adversarial Distillation (LAS&#x0026;ET-AD), to demonstrate the importance of the consideration of the prediction differences.</p>
<p>Secondly, we fine-tuned the model parameters of one, denoted as Learnable Distillation Attack Strategies and Evolvable Robust Teachers Adversarial Distillation (LDAS&#x0026;ERoT-AD) and Learnable Distillation Attack Strategies and Evolvable Standard Teachers Adversarial Distillation (LDAS&#x0026;EStT-AD), or none, denoted as Learnable Distillation Attack Strategies Adversarial Distillation (LDAS-AD), of the two pre-trained teachers in our LDAS&#x0026;ET-AD. The purpose was to illustrate the different effects of each teacher&#x2019;s update on performance improvement. Subsequently, the test clean and adversarial accuracy of the trained target models were evaluated. The results of the ablation studies are presented in <xref ref-type="table" rid="table-8">Table 8</xref>.</p>
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>Ablation studies on CIFAR-10 with ResNet-18. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<th></th>
<th>Method</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
<th>Natural</th>
<th>FGSM</th>
<th>PGD-20</th>
<th>CW<sub>&#x221E;</sub></th>
<th>AA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="2">Attack strategy</td>
<td>FDAS&#x0026;ET-AD</td>
<td>83.39</td>
<td>64.25</td>
<td>51.63</td>
<td>51.18</td>
<td>43.21</td>
<td>83.96</td>
<td>62.95</td>
<td>49.75</td>
<td>48.89</td>
<td>43.05</td>
</tr>
<tr>
<td>LAS&#x0026;ET-AD</td>
<td>84.19</td>
<td>64.63</td>
<td>52.10</td>
<td>52.08</td>
<td>44.28</td>
<td>84.41</td>
<td>63.98</td>
<td>49.80</td>
<td>50.25</td>
<td>43.60</td>
</tr>
<tr>
<td rowspan="3">Teacher evolution strategy</td>
<td>LDAS-AD</td>
<td>83.71</td>
<td>63.90</td>
<td>52.87</td>
<td>50.85</td>
<td>42.90</td>
<td>83.18</td>
<td>63.76</td>
<td>50.12</td>
<td>49.39</td>
<td>42.90</td>
</tr>
<tr>
<td>LDAS&#x0026;ERoT-AD</td>
<td>83.55</td>
<td>64.71</td>
<td>53.51</td>
<td>51.83</td>
<td>44.53</td>
<td>84.38</td>
<td>64.08</td>
<td>50.47</td>
<td>49.82</td>
<td>43.87</td>
</tr>
<tr>
<td>LDAS&#x0026;EStT-AD</td>
<td>84.49</td>
<td>63.85</td>
<td>52.69</td>
<td>50.55</td>
<td>43.69</td>
<td>84.41</td>
<td>63.22</td>
<td>49.28</td>
<td>49.04</td>
<td>43.25</td>
</tr>
<tr>
<td></td>
<td>Ours</td>
<td><bold>85.20</bold></td>
<td><bold>64.92</bold></td>
<td><bold>53.90</bold></td>
<td><bold>52.14</bold></td>
<td><bold>45.39</bold></td>
<td><bold>85.44</bold></td>
<td><bold>64.44</bold></td>
<td><bold>50.71</bold></td>
<td><bold>50.11</bold></td>
<td><bold>44.24</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>As shown in <xref ref-type="table" rid="table-8">Table 8</xref>, our LDAS&#x0026;ET-AD outperforms all five variants against all four attacks. Firstly, our LDAS&#x0026;ET-AD automatically generates sample-dependent and increasingly stronger attack strategies, enabling the creation of AEs that can adapt to more robust target models. Consequently, our LDAS&#x0026;ET-AD outperforms FDAS&#x0026;ET-AD, resulting in improvements of 1.81%, 0.67%, 2.27%, 0.96%, and 2.18% in clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy, respectively. Furthermore, by incorporating prediction differences into the learnable attack strategies, AEs are not only able to mislead the target model but also maximize the prediction discrepancy between the target and teacher models, achieving a closer match between them. Therefore, our LDAS&#x0026;ET-AD outperforms LAS&#x0026;ET-AD in terms of clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy by 1.01%, 0.29%, 1.80%, 0.06%, and 1.11%, respectively. These findings highlight the superiority of introducing the learnable attack strategies and prediction differences into the AD framework due to the generation of AEs that are more suitable for AD and a closer match between the teacher and target models.</p>
<p>Secondly, fine-tuning only the adversarially pre-trained teacher in LDAS&#x0026;ERoT-AD ensures the reliability and effectiveness of adversarial knowledge which aims to guide the target model in accurately classifying AEs. Therefore, LDAS&#x0026;ERoT-AD outperforms LDAS-AD solely in terms of adversarial robustness. LDAS&#x0026;EStT-AD, on the other hand, only updates the standard pre-trained teacher to enhance the quality of clean knowledge, which is designed to specifically enhance the clean accuracy of the target model. LDAS&#x0026;EStT-AD achieves higher accuracy on clean samples compared to LDAS-AD. Our LDAS&#x0026;ET-AD, which fine-tunes both teacher models, shows improved clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy by 0.71%, 0.21%, 0.39%, 0.31%, and 0.86%, respectively, compared to the methods that either do not update or only update one teacher. The experimental results indicate that fine-tuning both robust and standard teachers has positive effects on improving both clean accuracy and adversarial robustness of the target model, highlighting the potential of evolvable standard and robust teachers.</p>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Comparison of Different Dynamic Attack Strategies Generating Methods</title>
<p>To verify the superiority of the learnable distillation attack strategies generating mechanism in our LDAS&#x0026;ET-AD over other dynamic hand-crafted attack strategies generating methods, we replaced it with CAT [<xref ref-type="bibr" rid="ref-36">36</xref>], FOCS [<xref ref-type="bibr" rid="ref-38">38</xref>], and FAT [<xref ref-type="bibr" rid="ref-37">37</xref>] and considered prediction differences, denoted as Curriculum Distillation Attack Strategy and Evolvable Teachers Adversarial Distillation (CDAS&#x0026;ET-AD), First-Order Stationary Condition Distillation Attack Strategy and Evolvable Teachers Adversarial Distillation (FOCSDAS&#x0026;ET-AD), and Friendly Distillation Attack Strategy and Evolvable Teachers Adversarial Distillation (FriDAS&#x0026;ET-AD), respectively. The results are shown in <xref ref-type="table" rid="table-9">Table 9</xref>.</p>
<table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>Test accuracy (%) on CIFAR-10 dataset of ResNet-18 target model trained using our LDAS&#x0026;ET-AD with four types of attack strategies generating mechanisms (CAT [<xref ref-type="bibr" rid="ref-36">36</xref>], FOCS [<xref ref-type="bibr" rid="ref-38">38</xref>], FAT [<xref ref-type="bibr" rid="ref-37">37</xref>], and ours). The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<td>Method</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td>CDAS&#x0026;ET-AD</td>
<td>80.56</td>
<td>63.68</td>
<td>51.43</td>
<td>51.32</td>
<td>40.83</td>
<td>81.29</td>
<td>62.41</td>
<td>49.28</td>
<td>48.96</td>
<td>39.20</td>
</tr>
<tr>
<td>FOCSDAS&#x0026;ET-AD</td>
<td>82.15</td>
<td>64.24</td>
<td>51.73</td>
<td>51.56</td>
<td>43.54</td>
<td>83.97</td>
<td>62.85</td>
<td>49.53</td>
<td>49.14</td>
<td>42.34</td>
</tr>
<tr>
<td>FriDAS&#x0026;ET-AD</td>
<td>84.24</td>
<td>64.60</td>
<td>52.86</td>
<td>51.91</td>
<td>44.25</td>
<td>84.81</td>
<td>63.31</td>
<td>50.21</td>
<td>49.83</td>
<td>43.07</td>
</tr>
<tr>
<td>Ours</td>
<td><bold>85.20</bold></td>
<td><bold>64.92</bold></td>
<td><bold>53.90</bold></td>
<td><bold>52.14</bold></td>
<td><bold>45.39</bold></td>
<td><bold>85.44</bold></td>
<td><bold>64.44</bold></td>
<td><bold>50.71</bold></td>
<td><bold>50.11</bold></td>
<td><bold>44.24</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The obtained results demonstrate that the learnable distillation attack strategies generating mechanism in our LDAS&#x0026;ET-AD outperforms all three variants. This improvement can be attributed to the AEs being more suitable for AD of the increasingly robust target model. Specifically, compared to the best variant, our LDAS&#x0026;ET-AD achieves higher accuracy in clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA attacks by 0.96%, 0.32%, 1.04%, 0.23%, and 1.14%, respectively. These findings emphasize the advantages of introducing learnable attack strategies in the proposed LDAS&#x0026;ET-AD method for generating AEs suitable for AD when compared to other dynamic hand-crafted attack strategy methods.</p>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Comparison of Different Teacher Fine-Tuning Methods Based on L2T Distillation</title>
<p>To assess the superiority of the teacher fine-tuning strategy in our LDAS&#x0026;ET-AD over other teacher fine-tuning methods based on L2T distillation, we replace it with (1) meta distillation [<xref ref-type="bibr" rid="ref-31">31</xref>], which considers feedback from the target model on the validation set while all training samples equally and solely receiving supervision from the target model, referred to as Learnable Distillation Attack Strategy and Mate Adversarial Distillation (LDAS&#x0026;meta-AD) and (2) online distillation [<xref ref-type="bibr" rid="ref-27">27</xref>], which enforces similarity between the outputs of the target and teacher models on the training set without considering the target model&#x2019;s performance on the validation set, denoted as Learnable Distillation Attack Strategy and Online Adversarial Distillation (LDAS&#x0026;OL-AD). The results are presented in <xref ref-type="table" rid="table-10">Table 10</xref>.</p>
<table-wrap id="table-10">
<label>Table 10</label>
<caption>
<title>Test accuracy (%) on CIFAR-10 dataset of ResNet-18 target model trained using our LDAS&#x0026;ET-AD with three types of teacher fine-tuning methods (meta distillation [<xref ref-type="bibr" rid="ref-31">31</xref>], online distillation [<xref ref-type="bibr" rid="ref-27">27</xref>], and ours). The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th></th>
<th colspan="5" align="center">Best checkpoint</th>
<th colspan="5" align="center">Last checkpoint</th>
</tr>
<tr>
<td>Method</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
<td>Natural</td>
<td>FGSM</td>
<td>PGD-20</td>
<td>CW<sub>&#x221E;</sub></td>
<td>AA</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td>LDAS&#x0026;meta-AD</td>
<td>84.16</td>
<td>64.57</td>
<td>53.36</td>
<td>51.62</td>
<td>43.60</td>
<td>84.43</td>
<td>64.29</td>
<td>50.37</td>
<td>49.72</td>
<td>43.88</td>
</tr>
<tr>
<td>LDAS&#x0026;OL-AD</td>
<td>84.71</td>
<td>64.13</td>
<td>52.98</td>
<td>51.17</td>
<td>43.11</td>
<td>84.92</td>
<td>63.84</td>
<td>49.90</td>
<td>49.54</td>
<td>43.26</td>
</tr>
<tr>
<td>Ours</td>
<td>85.20</td>
<td><bold>64.92</bold></td>
<td><bold>53.90</bold></td>
<td><bold>52.14</bold></td>
<td><bold>45.39</bold></td>
<td><bold>85.44</bold></td>
<td><bold>64.44</bold></td>
<td><bold>50.71</bold></td>
<td><bold>50.11</bold></td>
<td><bold>44.24</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="table-10">Table 10</xref> demonstrates that our LDAS&#x0026;ET-AD outperforms all two variants, achieving the highest test accuracy on both clean samples and AEs. Our LDAS&#x0026;ET-AD uses the target model&#x2019;s performance on the verification set as feedback to assign the loss weight of each training sample for fine-tuning of teacher models, enhancing the effectiveness of the teachers&#x2019; knowledge in the generalization ability of the target model on both clean samples and AEs. This improvement is achieved by introducing adversarial distillation influence. Additionally, the training of teacher models is also supervised by the AEs of the training set, improving the reliability of their knowledge. Our LDAS&#x0026;ET-AD demonstrates significant improvements compared to the best variant, achieving enhancements of 0.96%, 0.32%, 1.04%, 0.23%, and 1.14% on clean, FGSM, PGD-20, CW<sub>&#x221E;</sub>, and AA accuracy, respectively. These results validate the effectiveness of the teacher fine-tuning teacher strategy in our proposed LDAS&#x0026;ET-AD, surpassing other teacher fine-tuning methods.</p>
</sec>
<sec id="s5_4">
<label>5.4</label>
<title>Comparison of Different <inline-formula id="ieqn-211"><mml:math id="mml-ieqn-211"><mml:mi mathvariant="bold-italic">k</mml:mi></mml:math></inline-formula> Values</title>
<p>The hyperparameter <inline-formula id="ieqn-212"><mml:math id="mml-ieqn-212"><mml:mi>k</mml:mi></mml:math></inline-formula> controls the alternating update of <inline-formula id="ieqn-213"><mml:math id="mml-ieqn-213"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-214"><mml:math id="mml-ieqn-214"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Every <inline-formula id="ieqn-215"><mml:math id="mml-ieqn-215"><mml:mi>k</mml:mi></mml:math></inline-formula> times <inline-formula id="ieqn-216"><mml:math id="mml-ieqn-216"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are updated, <inline-formula id="ieqn-217"><mml:math id="mml-ieqn-217"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are updated once. It affects not only performance but also training efficiency. Firstly, the efficiency of the proposed method decreases with the increase of <inline-formula id="ieqn-218"><mml:math id="mml-ieqn-218"><mml:mi>k</mml:mi></mml:math></inline-formula>. Smaller <inline-formula id="ieqn-219"><mml:math id="mml-ieqn-219"><mml:mi>k</mml:mi></mml:math></inline-formula> results in more frequent updates of <inline-formula id="ieqn-220"><mml:math id="mml-ieqn-220"><mml:msub><mml:mi>&#x03B8;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, thus requiring more training time. Secondly, selecting an appropriate <inline-formula id="ieqn-221"><mml:math id="mml-ieqn-221"><mml:mi>k</mml:mi></mml:math></inline-formula> is crucial for the adversarial robustness of the target model. If <inline-formula id="ieqn-222"><mml:math id="mml-ieqn-222"><mml:mi>k</mml:mi></mml:math></inline-formula> is too small, the target model&#x2019;s discrimination ability towards attack strategies generated by the strategy model may be impaired. This, in turn, affects the diversity of attack strategies and the update stability of the teacher models. On the other hand, if <inline-formula id="ieqn-223"><mml:math id="mml-ieqn-223"><mml:mi>k</mml:mi></mml:math></inline-formula> is excessively large, the generation ability of the strategy model may be compromised, resulting in insufficiently effective AEs for updating the teacher and target models. To determine the optimal <inline-formula id="ieqn-224"><mml:math id="mml-ieqn-224"><mml:mi>k</mml:mi></mml:math></inline-formula>, we conducted experiments on hyperparameter selection. The performance results are depicted in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>, and the efficiency results are listed in <xref ref-type="table" rid="table-11">Table 11</xref>.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>The accuracy on the CIFAR-10 dataset with the ResNet-18 target model trained using our LDAS&#x0026;ET-AD about different values of <inline-formula id="ieqn-225"><mml:math id="mml-ieqn-225"><mml:mi>k</mml:mi></mml:math></inline-formula></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-4.tif"/>
</fig><table-wrap id="table-11">
<label>Table 11</label>
<caption>
<title>Training time (s) on CIFAR-10 dataset with ResNet-18 target model trained using our LDAS&#x0026;ET-AD about different values of <inline-formula id="ieqn-226"><mml:math id="mml-ieqn-226"><mml:mi>k</mml:mi></mml:math></inline-formula>. The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Values</th>
<th>SAT [<xref ref-type="bibr" rid="ref-13">13</xref>]</th>
<th>1</th>
<th>10</th>
<th>20</th>
<th>30</th>
<th>40</th>
<th>50</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td>Time (Avg. Epoch)</td>
<td>674</td>
<td>5166</td>
<td>1954</td>
<td>1705</td>
<td>1482</td>
<td>1269</td>
<td><bold>910</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="fig" rid="fig-4">Fig. 4</xref> shows how the selection of <inline-formula id="ieqn-227"><mml:math id="mml-ieqn-227"><mml:mi>k</mml:mi></mml:math></inline-formula> impacts the clean and PGD-20 accuracy of the target model and there is a trade-off between these two metrics. Specifically, when <inline-formula id="ieqn-228"><mml:math id="mml-ieqn-228"><mml:mi>k</mml:mi></mml:math></inline-formula> is too small, the target model exhibits poor discrimination ability against attack strategies generated by the frequently updated strategy model. Consequently, the diversity of attack strategies diminishes, leading to low-quality AEs, lower adversarial robustness, and higher clean accuracy of the target model. Conversely, when <inline-formula id="ieqn-229"><mml:math id="mml-ieqn-229"><mml:mi>k</mml:mi></mml:math></inline-formula> is too large, the attack strategies generated by the strategy model with low generation ability become approximately fixed, preventing the target model from achieving optimal adversarial robustness. The results in <xref ref-type="fig" rid="fig-4">Fig. 4</xref> indicate that the proposed LDAS&#x0026;ET-AD achieves the best adversarial robustness when <inline-formula id="ieqn-230"><mml:math id="mml-ieqn-230"><mml:mi>k</mml:mi></mml:math></inline-formula> is set to 30.</p>
<p><xref ref-type="table" rid="table-11">Table 11</xref> demonstrates that the training time of the proposed LDAS&#x0026;ET-AD decreases with the increase of <inline-formula id="ieqn-231"><mml:math id="mml-ieqn-231"><mml:mi>k</mml:mi></mml:math></inline-formula>. As <inline-formula id="ieqn-232"><mml:math id="mml-ieqn-232"><mml:mi>k</mml:mi></mml:math></inline-formula> increases, the update frequency of the strategy model decreases, resulting in a decrease in the overall training time.</p>
<p>Considering both efficiency and adversarial robustness, we set <inline-formula id="ieqn-233"><mml:math id="mml-ieqn-233"><mml:mi>k</mml:mi></mml:math></inline-formula> to 30.</p>
</sec>
<sec id="s5_5">
<label>5.5</label>
<title>Comparison of Different <inline-formula id="ieqn-234"><mml:math id="mml-ieqn-234"><mml:msub><mml:mi mathvariant="bold-italic">&#x03B1;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mi mathvariant="bold-italic">d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-235"><mml:math id="mml-ieqn-235"><mml:msub><mml:mi mathvariant="bold-italic">&#x03B1;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">d</mml:mi><mml:mi mathvariant="bold-italic">v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> Values</title>
<p>The hyperparameters <inline-formula id="ieqn-236"><mml:math id="mml-ieqn-236"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-237"><mml:math id="mml-ieqn-237"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> play vital roles in controlling the self-evolution of standard and robust teachers and impact the guidance quality for classifying clean samples and AEs. Specifically, an excessive focus on self-evolution may lead to neglecting the feedback provided by the target model, resulting in guidance not meeting the target model&#x2019;s needs. Besides, a lack of focus on self-evolution may hinder teachers from enhancing their abilities, reducing the reliability of teacher knowledge. The controlled experiments were conducted to analyze the impact of self-evolution on the target model&#x2019;s performance.</p>
<p>We fix <inline-formula id="ieqn-238"><mml:math id="mml-ieqn-238"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> at 0.6 and vary <inline-formula id="ieqn-239"><mml:math id="mml-ieqn-239"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from {1.0, 0.9, 0.8, 0.7, 0.6, 0.5} to evaluate the adversarial robustness of ResNet-18 on CIFAR-10 against PGD-20 attack since <inline-formula id="ieqn-240"><mml:math id="mml-ieqn-240"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> controls fine-tuning of the robust teacher and mainly affects the robustness of the target model. <xref ref-type="fig" rid="fig-5">Fig. 5a</xref> demonstrates that when <inline-formula id="ieqn-241"><mml:math id="mml-ieqn-241"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is too large, the feedback from the target model has little influence on the robust teacher&#x2019;s update. Consequently, the fine-tuned robust teacher does not significantly improve the generalization performance of the target model on AEs. Conversely, when <inline-formula id="ieqn-242"><mml:math id="mml-ieqn-242"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is too small, neglect of self-evolution causes performance degradation. The guidance from low-performance teachers impairs the robustness of the target model. Both situations result in suboptimal adversarial robustness. The results indicate that LDAS&#x0026;ET-AD achieves the best robustness when <inline-formula id="ieqn-243"><mml:math id="mml-ieqn-243"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is set to 0.7.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Test accuracy on CIFAR-10 dataset with ResNet-18 target model trained using our LDAS&#x0026;ET-AD about different values of <inline-formula id="ieqn-244"><mml:math id="mml-ieqn-244"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-245"><mml:math id="mml-ieqn-245"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_47275-fig-5.tif"/>
</fig>
<p>Next, we fix <inline-formula id="ieqn-246"><mml:math id="mml-ieqn-246"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> at 0.7 and vary <inline-formula id="ieqn-247"><mml:math id="mml-ieqn-247"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> from {1.0, 0.9, 0.8, 0.7, 0.6, 0.5} to evaluate the clean accuracy of ResNet-18 on CIFAR-10 since <inline-formula id="ieqn-248"><mml:math id="mml-ieqn-248"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> controls fine-tuning of the standard teacher and mainly affects the clean accuracy of the target model. Consistent with the analysis of <inline-formula id="ieqn-249"><mml:math id="mml-ieqn-249"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, both excessively large and small <inline-formula id="ieqn-250"><mml:math id="mml-ieqn-250"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> hamper the fine-tuning of the standard teacher, preventing the target model from achieving optimal generalization on clean samples. <xref ref-type="fig" rid="fig-5">Fig. 5b</xref> illustrates that LDAS&#x0026;ET-AD performs well when <inline-formula id="ieqn-251"><mml:math id="mml-ieqn-251"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is set to 0.8.</p>
<p>In conclusion, we set <inline-formula id="ieqn-252"><mml:math id="mml-ieqn-252"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> to 0.8 and <inline-formula id="ieqn-253"><mml:math id="mml-ieqn-253"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>v</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> to 0.7 to strike a balance between self-evolution and target model feedback, ensuring the best performance of LDAS&#x0026;ET-AD.</p>
</sec>
<sec id="s5_6">
<label>5.6</label>
<title>Comparison of Different <inline-formula id="ieqn-254"><mml:math id="mml-ieqn-254"><mml:msub><mml:mi mathvariant="bold-italic">&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mn mathvariant="bold">2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-255"><mml:math id="mml-ieqn-255"><mml:msub><mml:mi mathvariant="bold-italic">&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mn mathvariant="bold">3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> Values</title>
<p>The hyperparameters <inline-formula id="ieqn-256"><mml:math id="mml-ieqn-256"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-257"><mml:math id="mml-ieqn-257"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> balance the trade-off between evaluating robustness loss term and predicting clean samples loss term in attack strategies generating mechanism. When <inline-formula id="ieqn-258"><mml:math id="mml-ieqn-258"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is relatively large compared to <inline-formula id="ieqn-259"><mml:math id="mml-ieqn-259"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula>, it may result in lower clean accuracy. Conversely, an excessively large <inline-formula id="ieqn-260"><mml:math id="mml-ieqn-260"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> can lead to insufficient attention to robustness, resulting in low adversarial robustness. We present the performance of our proposed LDAS&#x0026;ET-AD with various <inline-formula id="ieqn-261"><mml:math id="mml-ieqn-261"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-262"><mml:math id="mml-ieqn-262"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> pairs on CIFAR-10 using the ResNet-18 target model in <xref ref-type="table" rid="table-12">Table 12</xref>.</p>
<table-wrap id="table-12">
<label>Table 12</label>
<caption>
<title>Test accuracy (%) on CIFAR-10 dataset with ResNet-18 target model trained using our LDAS&#x0026;ET-AD about different values of <inline-formula id="ieqn-263"><mml:math id="mml-ieqn-263"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-264"><mml:math id="mml-ieqn-264"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mspace width="negativethinmathspace" /><mml:mo>.</mml:mo></mml:math></inline-formula> The best results are <bold>boldfaced</bold></title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th colspan="2" align="center">Values</th>
<th>Natural</th>
<th>PGD-20</th>
<th>AA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="3"><inline-formula id="ieqn-265"><mml:math id="mml-ieqn-265"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 2</td>
<td><inline-formula id="ieqn-266"><mml:math id="mml-ieqn-266"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 2</td>
<td>85.05</td>
<td>52.49</td>
<td>44.11</td>
</tr>
<tr>
<td><inline-formula id="ieqn-267"><mml:math id="mml-ieqn-267"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 4</td>
<td><bold>85.20</bold></td>
<td><bold>53.40</bold></td>
<td><bold>45.39</bold></td>
</tr>
<tr>
<td><inline-formula id="ieqn-268"><mml:math id="mml-ieqn-268"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 6</td>
<td>84.39</td>
<td>53.26</td>
<td>45.21</td>
</tr>
<tr>
<td rowspan="3"><inline-formula id="ieqn-269"><mml:math id="mml-ieqn-269"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 4</td>
<td><inline-formula id="ieqn-270"><mml:math id="mml-ieqn-270"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 2</td>
<td><bold>85.20</bold></td>
<td><bold>53.40</bold></td>
<td><bold>45.39</bold></td>
</tr>
<tr>
<td><inline-formula id="ieqn-271"><mml:math id="mml-ieqn-271"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 4</td>
<td>84.98</td>
<td>52.67</td>
<td>44.52</td>
</tr>
<tr>
<td><inline-formula id="ieqn-272"><mml:math id="mml-ieqn-272"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> &#x003D; 6</td>
<td>85.06</td>
<td>52.14</td>
<td>43.87</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Firstly, we fix <inline-formula id="ieqn-273"><mml:math id="mml-ieqn-273"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> at 2 and vary <inline-formula id="ieqn-274"><mml:math id="mml-ieqn-274"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> from {2, 4, 6} to evaluate performance. As <inline-formula id="ieqn-275"><mml:math id="mml-ieqn-275"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> increases, the total loss function places more emphasis on the robustness evaluation loss term. The results in <xref ref-type="table" rid="table-12">Table 12</xref> demonstrate that the clean accuracy shows a downward trend. Although the best robustness is achieved when <inline-formula id="ieqn-276"><mml:math id="mml-ieqn-276"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is 4, there is only a small improvement compared to when <inline-formula id="ieqn-277"><mml:math id="mml-ieqn-277"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is 2. The robustness continues to increase, and the optimal value is still achieved when <inline-formula id="ieqn-278"><mml:math id="mml-ieqn-278"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is 4, but the difference is minimal compared to when <inline-formula id="ieqn-279"><mml:math id="mml-ieqn-279"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is 2.</p>
<p>Secondly, we fix <inline-formula id="ieqn-280"><mml:math id="mml-ieqn-280"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> at 4 and vary <inline-formula id="ieqn-281"><mml:math id="mml-ieqn-281"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> from {2, 4, 6}. As <inline-formula id="ieqn-282"><mml:math id="mml-ieqn-282"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> increases, the total loss function focuses more on the clean accuracy loss term. It can be observed from the results in <xref ref-type="table" rid="table-12">Table 12</xref> that increasing <inline-formula id="ieqn-283"><mml:math id="mml-ieqn-283"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> leads to robustness decreasing and clean accuracy with little change. This indicates that when <inline-formula id="ieqn-284"><mml:math id="mml-ieqn-284"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> is fixed at 4, the clean accuracy is not sensitive to the change of <inline-formula id="ieqn-285"><mml:math id="mml-ieqn-285"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula>.</p>
<p>Although the performance of the target model is affected by <inline-formula id="ieqn-286"><mml:math id="mml-ieqn-286"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-287"><mml:math id="mml-ieqn-287"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula>, the changes do not occur within a large range. Therefore, the proposed method is not highly sensitive to these two hyperparameters, which aligns with the observation in [<xref ref-type="bibr" rid="ref-12">12</xref>]. Overall, we set <inline-formula id="ieqn-288"><mml:math id="mml-ieqn-288"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> to 2 and <inline-formula id="ieqn-289"><mml:math id="mml-ieqn-289"><mml:msub><mml:mi>&#x03B1;</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:math></inline-formula> to 4.</p>
</sec>
<sec id="s5_7">
<label>5.7</label>
<title>Training and Inference Complexity</title>
<p>The proposed method entails a higher training complexity than the baseline, primarily due to the training of the strategy model parameters and the fine-tuning of the teacher models. However, our LDAS&#x0026;ET-AD offers pronounced improvements over state-of-the-art adversarial defense methods. Specifically, the sample-dependent attack strategies generated by the strategy model in the game with the target model are highly effective, as are the more reliable teacher models fine-tuned according to the validation performance of the target model. In contrast, fixed hand-crafted attack strategies and static teacher models are far less effective.</p>
<p>Besides, to ensure a suitable trade-off between efficiency and robustness during the training of the strategy model, we have considered various factors, including the frequency of updating parameters. We have also introduced a finite difference approximation [<xref ref-type="bibr" rid="ref-35">35</xref>] to address the slowness of computing per-sample gradients and improve computational efficiency. Importantly, no additional complexity is introduced in the inference stage. However, we acknowledge that further work is necessary to reduce the training complexity of our approach.</p>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Conclusion</title>
<p>To enhance the quality of AEs and the reliability of teacher knowledge in existing AD techniques, an AD method LDAS&#x0026;ET-AD is proposed. Firstly, a learnable distillation attack strategies generating mechanism is developed to automatically create sample-dependent AEs well-suited for AD. A strategy model is introduced to produce attack strategies by competing with the target model in minimizing or maximizing the AD loss. Secondly, a teacher evolution strategy is devised to enhance the reliability and effectiveness of knowledge in improving the target model&#x2019;s generalization performance. The model parameters of the standard and robust teachers are dynamically adjusted based on the target model&#x2019;s performance on the validation set and AEs. We evaluate the method using ResNet-18 and MobileNet-V2 on the CIFAR-10 and CIFAR-100 datasets. Experiments demonstrate the superiority of our proposed LDAS&#x0026;ET-AD method over state-of-the-art adversarial defense techniques in improving robustness against various adversarial attacks. The results confirm that introducing teacher knowledge to enhance the applicability of AEs and considering the target model&#x2019;s validation performance to improve the reliability of the teacher knowledge are effective in promoting robustness.</p>
</sec>
<sec id="s7">
<label>7</label>
<title>Limitations and Prospects</title>
<p>While the proposed LDAS&#x0026;ET-AD method demonstrates superiority over existing AD methods, it is essential to recognize its limitations. Firstly, the reliance on a separate validation set is crucial for obtaining feedback to fine-tune the teachers. However, this approach results in a reduction of training samples, which may impact performance, particularly in datasets of limited or moderate size. Exploring an alternative approach that leverages all data samples for both training and validation holds the potential for extracting more comprehensive information from the dataset. This avenue warrants further exploration in future research. Secondly, the proposed method involves various hyperparameters that significantly influence performance, necessitating manual configuration based on experimental results. This trial-and-error method demands additional time. To address this challenge, future endeavors will encompass the introduction of automatic hyperparameter optimization methods such as Random Search and Bayesian Optimization to identify the optimal combination of hyperparameters. Lastly, while our experiments have primarily focused on image classification tasks, which are relatively straightforward for current deep learning models, it is imperative for future work to extend the application of LDAS&#x0026;ET-AD to more complex computer version tasks, and other domains such as natural language processing, and beyond. Such expansion will provide a more comprehensive evaluation of the method&#x2019;s efficacy across diverse applications.</p>
</sec>
</body>
<back>
<ack>
<p>The authors are very grateful to the editors and all anonymous reviewers for their insightful comments.</p>
</ack>
<sec><title>Funding Statement</title>
<p>This study was funded by the National Key Research and Development Program of China (2021YFB1006200); Major Science and Technology Project of Henan Province in China (221100211200). Grant was received by S. Li.</p>
</sec>
<sec><title>Author Contributions</title>
<p>The authors confirm contribution to the paper as follows: study conception and design: S. Li, X. Yang; data collection: G. Cheng; analysis and interpretation of results: S. Li, W. Liu, W. Guo; draft manuscript preparation: S. Li, H. Hu. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability"><title>Availability of Data and Materials</title>
<p>The data underlying this article will be shared on reasonable request to the corresponding author.</p>
</sec>
<sec sec-type="COI-statement"><title>Conflicts of Interest</title>
<p>The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>K. S.</given-names> <surname>Kumar</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Rajendran</surname></string-name></person-group>, &#x201C;<article-title>Deep convolutional neural network for brain tumor segmentation</article-title>,&#x201D; <source>J. Electr. Eng. Technol.</source>, vol. <volume>18</volume>, no. <issue>5</issue>, pp. <fpage>3925</fpage>&#x2013;<lpage>3932</lpage>, <year>2023</year>. doi: <pub-id pub-id-type="doi">10.1007/s42835-023-01479-y</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Ji</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Survey of hallucination in natural language generation</article-title>,&#x201D; <source>ACM Comput. Surv.</source>, vol. <volume>55</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>38</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>L.</given-names> <surname>Chai</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Liu</surname></string-name>, and <string-name><given-names>C.</given-names> <surname>Lee</surname></string-name></person-group>, &#x201C;<article-title>A cross-entropy-guided measure (CEGM) for assessing speech recognition performance and optimizing DNN-based speech enhancement</article-title>,&#x201D; <source>IEEE/ACM Trans. Audio Speech Lang. Process.</source>, vol. <volume>29</volume>, pp. <fpage>106</fpage>&#x2013;<lpage>117</lpage>, <year>2021</year>. doi: <pub-id pub-id-type="doi">10.1109/TASLP.2020.3036783</pub-id>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Szegedy</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Intriguing properties of neural networks</article-title>,&#x201D; in <conf-name>2014 Int. Conf. Learn. Rep.</conf-name>, <publisher-loc>Alberta, USA</publisher-loc>, <year>2014</year>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Xiong</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Li</surname></string-name>, and <string-name><given-names>Z.</given-names> <surname>Cai</surname></string-name></person-group>, &#x201C;<article-title>Multi-source adversarial sample attack on autonomous vehicles</article-title>,&#x201D; <source>IEEE Trans. Vehicular Technol.</source>, vol. <volume>70</volume>, no. <issue>3</issue>, pp. <fpage>2822</fpage>&#x2013;<lpage>2835</lpage>, <year>2021</year>. doi: <pub-id pub-id-type="doi">10.1109/TVT.2021.3061065</pub-id>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Shen</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Li</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Hu</surname></string-name></person-group>, &#x201C;<article-title>Effective and robust physical-world attacks on deep learning face recognition systems</article-title>,&#x201D; <source>IEEE Trans. Inf. Forensics Secur.</source>, vol. <volume>16</volume>, pp. <fpage>4063</fpage>&#x2013;<lpage>4077</lpage>, <year>2021</year>. doi: <pub-id pub-id-type="doi">10.1109/TIFS.2021.3102492</pub-id>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Kherchouche</surname></string-name>, <string-name><given-names>S. A.</given-names> <surname>Fezza</surname></string-name>, and <string-name><given-names>W.</given-names> <surname>Hamidouche</surname></string-name></person-group>, &#x201C;<article-title>Detect and defense against adversarial examples in deep learning using natural scene statistics and adaptive denoising</article-title>,&#x201D; <source>Neural Comput. Appl.</source>, vol. <volume>34</volume>, no. <issue>24</issue>, pp. <fpage>21567</fpage>&#x2013;<lpage>21582</lpage>, <year>2021</year>. doi: <pub-id pub-id-type="doi">10.1007/s00521-021-06330-x</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Singh</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Kumar Awasthi</surname></string-name>, <string-name><surname>Urvashi</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Shorfuzzaman</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Alsufyani</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Uddin</surname></string-name></person-group>, &#x201C;<article-title>Chained dual-generative adversarial network: A generalized defense against adversarial attacks</article-title>,&#x201D; <source>Comput. Mater. Contin.</source>, vol. <volume>74</volume>, no. <issue>2</issue>, pp. <fpage>2541</fpage>&#x2013;<lpage>2555</lpage>, <year>2023</year>. doi: <pub-id pub-id-type="doi">10.32604/cmc.2023.032795</pub-id>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Jia</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Wei</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Cao</surname></string-name>, and <string-name><given-names>H.</given-names> <surname>Foroosh</surname></string-name></person-group>, &#x201C;<article-title>ComDefend: An efficient image compression model to defend adversarial examples</article-title>,&#x201D; in <conf-name>Proc. 2022 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>New Orleans, USA</publisher-loc>, <year>2018</year>, pp. <fpage>6077</fpage>&#x2013;<lpage>6085</lpage>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Jiao</surname></string-name>, <string-name><given-names>E. P.</given-names> <surname>Xing</surname></string-name>, <string-name><given-names>L. E.</given-names> <surname>Ghaoui</surname></string-name> and <string-name><given-names>M. I.</given-names> <surname>Jordan</surname></string-name></person-group>, &#x201C;<article-title>Theoretically principled trade-off between robustness and accuracy</article-title>,&#x201D; in <conf-name>36th Int. Conf. Mach. Learn.</conf-name>, <publisher-loc>California, USA</publisher-loc>, <year>2019</year>, pp. <fpage>7472</fpage>&#x2013;<lpage>7482</lpage>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Haroon</surname></string-name> and <string-name><given-names>H.</given-names> <surname>Ali</surname></string-name></person-group>, &#x201C;<article-title>Adversarial training against adversarial attacks for machine learning-based intrusion detection systems</article-title>,&#x201D; <source>Comput. Mater. Contin.</source>, vol. <volume>73</volume>, no. <issue>2</issue>, pp. <fpage>3513</fpage>&#x2013;<lpage>3527</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Jia</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Wu</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Ma</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Wang</surname></string-name> and <string-name><given-names>X.</given-names> <surname>Cao</surname></string-name></person-group>, &#x201C;<article-title>LAS-AT: Adversarial training with learnable attack strategy</article-title>,&#x201D; in <conf-name>Proc. 2022 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Louisiana, USA</publisher-loc>, <year>2022</year>, pp. <fpage>13388</fpage>&#x2013;<lpage>13398</lpage>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Madry</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Makelov</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Schmidt</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Tsipras</surname></string-name>, and <string-name><given-names>A.</given-names> <surname>Vladu</surname></string-name></person-group>, &#x201C;<article-title>Towards deep learning models resistant to adversarial attacks</article-title>,&#x201D; in <conf-name>5th Int. Conf. Learn. Rep.</conf-name>, <publisher-loc>Toulon, France</publisher-loc>, <year>2017</year>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Nandi</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Addepalli</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Rangwani</surname></string-name>, and <string-name><given-names>R. V.</given-names> <surname>Babu</surname></string-name></person-group>, &#x201C;<article-title>Certified adversarial robustness within multiple perturbation bounds</article-title>,&#x201D; in <conf-name>Proc. 2023 IEEE/CVF Conf. Comput. Vis. Pattern Recognit. Workshops</conf-name>, <publisher-loc>Vancouver, Canada</publisher-loc>, <year>2023</year>, pp. <fpage>2298</fpage>&#x2013;<lpage>2305</lpage>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Zhang</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Boosting verified training for robust image classifications via abstraction</article-title>,&#x201D; in <conf-name>Proc. 2023 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Vancouver, Canada</publisher-loc>, <year>2023</year>, pp. <fpage>16251</fpage>&#x2013;<lpage>16260</lpage>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Xiao</surname></string-name>, and <string-name><given-names>B.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>DiffSmooth: Certifiably robust learning via diffusion models and local smoothing</article-title>,&#x201D; in <conf-name>32nd USENIX Secur. Symp.</conf-name>, <publisher-loc>California, USA</publisher-loc>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Vor&#x00E1;&#x010D;ek</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Hein</surname></string-name></person-group>, &#x201C;<article-title>Improving l1-certified robustness via randomized smoothing by leveraging box constraints</article-title>,&#x201D; in <conf-name>Int. Conf. Mach. Learn.</conf-name>, <publisher-loc>Hawaii, USA</publisher-loc>, <year>2023</year>, pp. <fpage>35198</fpage>&#x2013;<lpage>35222</lpage>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Salzmann</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Lin</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Tomioka</surname></string-name>, and <string-name><given-names>S. E.</given-names> <surname>Susstrunk</surname></string-name></person-group>, &#x201C;<article-title>On the loss landscape of adversarial training: Identifying challenges and how to overcome them</article-title>,&#x201D; in <conf-name>Neural Inf. Process. Syst.</conf-name>, <publisher-loc>Vancouver, Canada</publisher-loc>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Goldblum</surname></string-name>, <string-name><given-names>L. H.</given-names> <surname>Fowl</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Feizi</surname></string-name>, and <string-name><given-names>T.</given-names> <surname>Goldstein</surname></string-name></person-group>, &#x201C;<article-title>Adversarially robust distillation</article-title>,&#x201D; in <conf-name>Assoc. Advan. Artif. Intell.</conf-name>, <publisher-loc>Hawaii, USA</publisher-loc>, vol. <volume>34</volume>, no. <issue>2</issue>, pp. <fpage>3996</fpage>&#x2013;<lpage>4003</lpage>, <year>2019</year>. </mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Maroto</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Ortiz-Jim&#x00E9;nez</surname></string-name>, and <string-name><given-names>P.</given-names> <surname>Frossard</surname></string-name></person-group>, &#x201C;<article-title>On the benefits of knowledge distillation for adversarial robustness</article-title>,&#x201D; <comment>arXiv preprint arXiv:2203.07159</comment>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Zi</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Zhao</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Ma</surname></string-name>, and <string-name><given-names>Y.</given-names> <surname>Jiang</surname></string-name></person-group>, &#x201C;<article-title>Revisiting adversarial robustness distillation: Robust soft labels make student better</article-title>,&#x201D; in <conf-name>Proc. 2021 IEEE/CVF Int. Conf. Comput. Vis.</conf-name>, <publisher-loc>Montreal, Canada</publisher-loc>, <year>2021</year>, pp. <fpage>16423</fpage>&#x2013;<lpage>16432</lpage>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Cao</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Vanilla feature distillation for improving the accuracy-robustness trade-off in adversarial training</article-title>,&#x201D; <comment>arXiv preprint arXiv:2206.02158</comment>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>T.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>Z. A.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Chang</surname></string-name>, and <string-name><given-names>Z.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>Robust overfitting may be mitigated by properly learned smoothening</article-title>,&#x201D; in <conf-name>2021 Int. Conf. Learn. Rep.</conf-name>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Zhao</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Sun</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Zhang</surname></string-name>, and <string-name><given-names>X.</given-names> <surname>Wei</surname></string-name></person-group>, &#x201C;<article-title>Enhanced accuracy and robustness via multi-teacher adversarial distillation</article-title>,&#x201D; in <conf-name>Eur. Conf. Comput. Vis.</conf-name>, <year>2022</year>, vol. <volume>13664</volume>, pp. <fpage>585</fpage>&#x2013;<lpage>602</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-3-031-19772-7</pub-id>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Zhu</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Reliable adversarial distillation with unreliable teachers</article-title>,&#x201D; in <conf-name>Int. Conf. Learn. Rep.</conf-name>, <publisher-loc>Vienna, Austria</publisher-loc>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Fan</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Tian</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Qin</surname></string-name>, <string-name><given-names>X. Y.</given-names> <surname>Li</surname></string-name>, and <string-name><given-names>T. Y.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Learning to teach</article-title>,&#x201D; in <conf-name>Int. Conf. Learn. Rep.</conf-name>, <publisher-loc>Vancouver, Canada</publisher-loc>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Xiang</surname></string-name>, <string-name><given-names>T. M.</given-names> <surname>Hospedales</surname></string-name>, and <string-name><given-names>H.</given-names> <surname>Lu</surname></string-name></person-group>, &#x201C;<article-title>Deep mutual learning</article-title>,&#x201D; in <conf-name>2018 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Utah, USA</publisher-loc>, <year>2017</year>, pp. <fpage>4320</fpage>&#x2013;<lpage>4328</lpage>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Lan</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhu</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Gong</surname></string-name></person-group>, &#x201C;<article-title>Knowledge distillation by On-the-Fly native ensemble</article-title>,&#x201D; in <conf-name>Neural Inf. Process. Syst.</conf-name>, <publisher-loc>Montreal, Canada</publisher-loc>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Zhang</surname></string-name>, and <string-name><given-names>D.</given-names> <surname>Ji</surname></string-name></person-group>, &#x201C;<article-title>Embedded mutual learning: A novel online distillation method integrating diverse knowledge sources</article-title>,&#x201D; <source>Appl. Intell.</source>, vol. <volume>53</volume>, no. <issue>10</issue>, pp. <fpage>11524</fpage>&#x2013;<lpage>11537</lpage>, <year>2022</year>. doi: <pub-id pub-id-type="doi">10.1007/s10489-022-03974-7</pub-id>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Qian</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Yin</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Hong</surname></string-name>, and <string-name><given-names>M.</given-names> <surname>Wang</surname></string-name></person-group>, &#x201C;<article-title>Switchable online knowledge distillation</article-title>,&#x201D; in <conf-name>Eur. Conf. Comput. Vis.</conf-name>, <publisher-loc> Tel Aviv, Israel</publisher-loc>, <year>2022</year>, pp. <fpage>449</fpage>&#x2013;<lpage>466</lpage>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Zhou</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Xu</surname></string-name>, and <string-name><given-names>J.</given-names> <surname>Mcauley</surname></string-name></person-group>, &#x201C;<article-title>BERT learns to teach: Knowledge distillation with meta learning</article-title>,&#x201D; in <conf-name>Annual Meet. Assoc. Comput. Lingist.</conf-name>, <publisher-loc>Bangkok, Thailand</publisher-loc>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Chen</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Liu</surname></string-name></person-group>, &#x201C;<article-title>Learning knowledge representation with meta knowledge distillation for single image super-resolution</article-title>,&#x201D; <source>J. Vis. Commun. Image Represent.</source>, vol. <volume>95</volume>, no. <issue>9</issue>, pp. <fpage>103874</fpage>, <year>2022</year>. doi: <pub-id pub-id-type="doi">10.1016/j.jvcir.2023.103874</pub-id>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Pham</surname></string-name>, <string-name><given-names>Q.</given-names> <surname>Xie</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Dai</surname></string-name>, and <string-name><given-names>Q. V.</given-names> <surname>Le</surname></string-name></person-group>, &#x201C;<article-title>Meta pseudo labels</article-title>,&#x201D; in <conf-name>Proc. 2021 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Nashville, USA</publisher-loc>, <year>2020</year>, pp. <fpage>11552</fpage>&#x2013;<lpage>11563</lpage>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Abu</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Abdukarimov</surname></string-name>, <string-name><given-names>N. A.</given-names> <surname>Tu</surname></string-name>, and <string-name><given-names>M.</given-names> <surname>Lee</surname></string-name></person-group>, &#x201C;<article-title>Meta pseudo labels for chest x-ray image classification</article-title>,&#x201D; in <conf-name>Proc. 2022 IEEE Int. Conf. Syst., Man, Cybernet. </conf-name>, <publisher-loc>Prague, Czech Republic</publisher-loc>, <year>2022</year>, pp. <fpage>2735</fpage>&#x2013;<lpage>2739</lpage>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y. X.</given-names> <surname>Ren</surname></string-name>, <string-name><given-names>Z. H.</given-names> <surname>Zhong</surname></string-name>, <string-name><given-names>X. J.</given-names> <surname>Shi</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Yuan</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>Tailoring instructions to studentss learning levels boosts knowledge distillation</article-title>,&#x201D; in <conf-name>Annual Meet. Assoc. Comput. Linguist.</conf-name>, <publisher-loc>Toronto, Canada</publisher-loc>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Cai</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Liu</surname></string-name>, and <string-name><given-names>D. X.</given-names> <surname>Song</surname></string-name></person-group>, &#x201C;<article-title>Curriculum adversarial training</article-title>,&#x201D; in <conf-name>27th Int. Joint Conf. Artif. Intell.</conf-name>, <publisher-loc>Stockholm, Sweden</publisher-loc>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Zhang</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Attacks which do not kill training make adversarial learning stronger</article-title>,&#x201D; in <conf-name>37th Int. Conf. Mach. Learn.</conf-name>, <publisher-loc>Florida, USA</publisher-loc>, <year>2002</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Ma</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Bailey</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yi</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Zhou</surname></string-name> and <string-name><given-names>Q.</given-names> <surname>Gu</surname></string-name></person-group>, &#x201C;<article-title>On the convergence and robustness of adversarial training</article-title>,&#x201D; in <conf-name>38th Int. Conf. Mach. Learn.</conf-name>, <publisher-loc>Vienna, Austria</publisher-loc>, <year>2021</year>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>I. J.</given-names> <surname>Goodfellow</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Shlens</surname></string-name>, and <string-name><given-names>C.</given-names> <surname>Szegedy</surname></string-name></person-group>, &#x201C;<article-title>Explaining and harnessing adversarial examples</article-title>,&#x201D; in <conf-name>3rd Int.Conf. Learn. Represent.</conf-name>, <publisher-loc>San Diego, CA, USA</publisher-loc>, <year>2015</year>, pp. <fpage>1</fpage>&#x2013;<lpage>11</lpage>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Carlini</surname></string-name> and <string-name><given-names>D. A.</given-names> <surname>Wagner</surname></string-name></person-group>, &#x201C;<article-title>Towards evaluating the robustness of neural networks</article-title>,&#x201D; in <conf-name>2017 IEEE Symp. Secur. Priv.</conf-name>, <publisher-loc>California, USA</publisher-loc>, <year>2016</year>, pp. <fpage>39</fpage>&#x2013;<lpage>57</lpage>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Croce</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Hein</surname></string-name></person-group>, &#x201C;<article-title>Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks</article-title>,&#x201D; in <conf-name>Int. Conf. Mach. Learn.</conf-name>, <publisher-loc>Maryland, USA</publisher-loc>, <year>2020</year>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Wang</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Zou</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Yi</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Bailey</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Ma</surname></string-name> and <string-name><given-names>Q.</given-names> <surname>Gu</surname></string-name></person-group>, &#x201C;<article-title>Improving adversarial robustness requires revisiting misclassified examples</article-title>,&#x201D; in <conf-name>7th Int. Conf. Learn. Represent.</conf-name>, <publisher-loc>New Orleans, USA</publisher-loc>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Krizhevsky</surname></string-name></person-group>, &#x201C;<article-title>Learning multiple layers of features from tiny images</article-title>,&#x201D; <comment>M.S. dissertation, Univ. of Toronto, Canada</comment>, <year>2009</year>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>K.</given-names> <surname>He</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ren</surname></string-name>, and <string-name><given-names>J.</given-names> <surname>Sun</surname></string-name></person-group>, &#x201C;<article-title>Deep residual learning for image recognition</article-title>,&#x201D; in <conf-name>2016 IEEE Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Massachusetts, USA</publisher-loc>, <year>2015</year>, pp. <fpage>770</fpage>&#x2013;<lpage>778</lpage>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Sandler</surname></string-name>, <string-name><given-names>A. G.</given-names> <surname>Howard</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Zhu</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Zhmoginov</surname></string-name>, and <string-name><given-names>L.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>MobileNetV2: Inverted residuals and linear bottlenecks</article-title>,&#x201D; in <conf-name>2018 IEEE/CVF Conf. Comput. Vis. Pattern Recognit.</conf-name>, <publisher-loc>Utah, USA</publisher-loc>, <year>2018</year>, pp. <fpage>4510</fpage>&#x2013;<lpage>4520</lpage>.</mixed-citation></ref>
</ref-list>
</back></article>