<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">67211</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2025.067211</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Adapting Convolutional Autoencoder for DDoS Attack Detection via Joint Reconstruction Learning and Refined Anomaly Scoring</article-title>
<alt-title alt-title-type="left-running-head">Adapting Convolutional Autoencoder for DDoS Attack Detection via Joint Reconstruction Learning and Refined Anomaly Scoring</alt-title>
<alt-title alt-title-type="right-running-head">Adapting Convolutional Autoencoder for DDoS Attack Detection via Joint Reconstruction Learning and Refined Anomaly Scoring</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Han</surname><given-names>Seulki</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Son</surname><given-names>Sangho</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Sakong</surname><given-names>Won</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-4" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Jung</surname><given-names>Haemin</given-names></name><xref ref-type="aff" rid="aff-3">3</xref><email>hmjung@ut.ac.kr</email></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Digital Analytics, Yonsei University</institution>, <addr-line>Seoul, 03722</addr-line>, <country>Republic of Korea</country></aff>
<aff id="aff-2"><label>2</label><institution>Department of Industrial Engineering, Yonsei University</institution>, <addr-line>Seoul, 03722</addr-line>, <country>Republic of Korea</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Industrial &#x0026; Management Engineering, Korea National University of Transportation</institution>, <addr-line>Chungju, 27469</addr-line>, <country>Republic of Korea</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Haemin Jung. Email: <email>hmjung@ut.ac.kr</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2025</year></pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>23</day><month>09</month><year>2025</year></pub-date>
<volume>85</volume>
<issue>2</issue>
<fpage>2893</fpage>
<lpage>2912</lpage>
<history>
<date date-type="received">
<day>27</day>
<month>4</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>20</day>
<month>8</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2025 The Authors.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_67211.pdf"></self-uri>
<abstract>
<p>As cyber threats become increasingly sophisticated, Distributed Denial-of-Service (DDoS) attacks continue to pose a serious threat to network infrastructure, often disrupting critical services through overwhelming traffic. Although unsupervised anomaly detection using convolutional autoencoders (CAEs) has gained attention for its ability to model normal network behavior without requiring labeled data, conventional CAEs struggle to effectively distinguish between normal and attack traffic due to over-generalized reconstructions and naive anomaly scoring. To address these limitations, we propose CA-CAE, a novel anomaly detection framework designed to improve DDoS detection through asymmetric joint reconstruction learning and refined anomaly scoring. Our architecture connects two CAEs sequentially with asymmetric filter allocation, which amplifies reconstruction errors for anomalous data while preserving low errors for normal traffic. Additionally, we introduce a scoring mechanism that incorporates exponential decay weighting to emphasize recent anomalies and relative traffic volume adjustment to highlight high-risk instances, enabling more accurate and timely detection. We evaluate CA-CAE on a real-world network traffic dataset collected using Cisco NetFlow, containing over 190,000 normal instances and only 78 anomalous instances&#x2014;an extremely imbalanced scenario (0.0004% anomalies). We validate the proposed framework through extensive experiments, including statistical tests and comparisons with baseline models. Despite this challenge, our method achieves significant improvement, increasing the F1-score from 0.515 obtained by the baseline CAE to 0.934, and outperforming other models. These results demonstrate the effectiveness, scalability, and practicality of CA-CAE for unsupervised DDoS detection in realistic network environments. By combining lightweight model architecture with a domain-aware scoring strategy, our framework provides a robust solution for early detection of DDoS attacks without relying on labeled attack data.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Anomaly detection</kwd>
<kwd>DDoS attack detection</kwd>
<kwd>convolutional autoencoder</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>Korea National University</funding-source>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>With the rapid expansion of digital infrastructure, cyber threats are evolving at an unprecedented pace. Among these, Distributed Denial-of-Service (DDoS) attacks remain a critical challenge, overwhelming targeted systems by flooding them with a massive amount of malicious traffic [<xref ref-type="bibr" rid="ref-1">1</xref>]. Detecting DDoS attacks early and accurately is crucial to mitigating impact on network availability and security.</p>
<p>Accordingly, anomaly detection techniques are widely employed for DDoS attack detection, as such attacks inherently deviate from normal network behavior. Recently, deep learning-based models, particularly autoencoder architectures [<xref ref-type="bibr" rid="ref-2">2</xref>], have gained traction due to their ability to learn normal traffic patterns in an unsupervised manner [<xref ref-type="bibr" rid="ref-3">3</xref>]. Autoencoders achieve this by learning a compact representation of normal network traffic through an encoder-decoder structure [<xref ref-type="bibr" rid="ref-4">4</xref>]. Once trained on normal data, an AE can reconstruct normal traffic with high accuracy, whereas anomalous traffic, which deviates from learned patterns, tends to exhibit higher reconstruction errors. By using reconstruction error as an anomaly score, normal and abnormal traffic can be effectively distinguished [<xref ref-type="bibr" rid="ref-5">5</xref>].</p>
<p>Among these, convolutional autoencoders (CAEs) have demonstrated strong performance in analyzing network traffic sequences by capturing local dependencies [<xref ref-type="bibr" rid="ref-6">6</xref>]. However, existing CAE-based models still suffer from two fundamental limitations. First, CAEs tend to over-generalize during reconstruction, making it difficult to distinguish between normal and anomalous traffic [<xref ref-type="bibr" rid="ref-7">7</xref>]. Because these models are highly expressive, they can sometimes reconstruct even anomalous data with low error, leading to a high false negative rate. For an effective detection mechanism, the model should accurately reconstruct normal traffic while struggling to reconstruct anomalous samples. Second, most CAE-based methods rely on average reconstruction error as an anomaly score, which presents two key issues. As the sequence length increases, this score becomes less representative of the most recent anomalies, reducing detection sensitivity. Additionally, the reconstruction error alone does not capture the severity or potential harm of an anomaly. This is particularly problematic in DDoS detection, where distinguishing between harmful and benign anomalies is essential to avoid false alarms.</p>
<p>To address these challenges, this paper proposes a novel CAE-based anomaly detection framework that enhances DDoS detection through joint reconstruction learning and a refined anomaly scoring method. The core component of this framework is a two-stage CAE model in which two consecutive autoencoders, each with a different filter allocation, perform a joint reconstruction task. This process amplifies reconstruction errors for anomalous data while maintaining high accuracy for normal traffic, improving anomaly separability. In addition to architectural enhancements, we introduce an improved anomaly scoring mechanism that refines detection by incorporating exponential decay weighting and relative traffic volume adjustments. The exponential decay weighting emphasizes recent anomalies, preventing older, irrelevant data from distorting the detection process, while the relative traffic volume adjustment helps capture the risk level of each anomaly, ensuring that high-risk DDoS attacks are prioritized.</p>
<p>To validate the effectiveness of the proposed framework, we conduct extensive experiments using real-world network traffic data. The results demonstrate that our framework outperforms existing CAE-based and supervised models in detecting DDoS attacks, achieving higher precision and recall. This study highlights the importance of both model architecture and anomaly scoring mechanisms in improving anomaly detection performance, particularly in highly imbalanced datasets such as DDoS detection scenarios.</p>
<p>In summary, the main challenges in anomaly-based DDoS detection include the over-generalization of CAE models and the limitations of reconstruction-based scoring in capturing temporal urgency and risk severity. To overcome these, we propose a joint reconstruction framework with asymmetric architecture and a risk-aware anomaly scoring method that together improve both anomaly separability and detection accuracy. The main contributions of this study are summarized as follows:
<list list-type="bullet">
<list-item>
<p>We propose a novel CAE-based anomaly detection framework CA-CAE that performs joint reconstruction via two consecutive autoencoders with asymmetric filter allocation, effectively amplifying reconstruction errors for anomalies.</p></list-item>
<list-item>
<p>We introduce a refined anomaly scoring mechanism that integrates exponential decay and relative traffic volume to emphasize recent and high-risk anomalies.</p></list-item>
<list-item>
<p>We conduct extensive experiments on real-world network traffic, demonstrating that CA-CAE achieves an F1-score of 0.934, significantly outperforming baselines.</p></list-item>
<list-item>
<p>We validate the individual contributions of each component&#x2014;model architecture, exponential decay, and risk-aware scoring&#x2014;through ablation studies and hyperparameter sensitivity analysis.</p></list-item>
</list></p>
<p>The remainder of this paper is structured as follows. <xref ref-type="sec" rid="s2">Section 2</xref> provides background on DDoS detection and autoencoder-based methods. <xref ref-type="sec" rid="s3">Section 3</xref> presents the proposed framework, detailing its architecture and anomaly scoring mechanism. <xref ref-type="sec" rid="s4">Section 4</xref> discusses experimental setup and results, comparing our framework-based model with baseline models. Finally, <xref ref-type="sec" rid="s5">Section 5</xref> concludes with key takeaways and future research directions.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Background</title>
<sec id="s2_1">
<label>2.1</label>
<title>Autoencoder-Based Anomaly Detection</title>
<p>Autoencoder is an artificial neural network trained in an unsupervised manner so that does not require class labels for learning [<xref ref-type="bibr" rid="ref-2">2</xref>]. It consists of an encoder part, which compresses input data into a low-dimensional vector, and a decoder part, which reconstructs the compressed vector back to a higher dimension. Namely, autoencoder can be viewed as a composite function of an encoding function <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> that projects an input instance to the latent feature space and a decoding function <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> that operates in the reverse direction [<xref ref-type="bibr" rid="ref-8">8</xref>]. <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref> is a general notation for autoencoder.
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mi>z</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow><mml:mo>=</mml:mo><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mi>z</mml:mi><mml:mo>;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>Here, <inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mi>z</mml:mi></mml:math></inline-formula> is the compressed vector from the encoder, and <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:math></inline-formula> is output of the decoder. Two functions <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:msub><mml:mi>f</mml:mi><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are parameterized by <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:msub><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, respectively.</p>
<p>Training objective for autoencoder is to minimize the distance between the input and the output by generating reduced but important feature sets from the original features. Therefore, optimal parameter sets <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:msubsup><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:msubsup><mml:mrow><mml:mi mathvariant="normal">&#x0398;</mml:mi></mml:mrow><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:msubsup></mml:math></inline-formula> are learned in the process of minimizing the difference or distance between the input and the output, typically the mean squared error. This can be formally expressed as <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>.
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:mtable displaystyle="true" columnalign="right left" columnspacing="0em" rowspacing="3pt"><mml:mtr><mml:mtd></mml:mtd><mml:mtd><mml:mstyle><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msubsup><mml:mtext>&#x0398;</mml:mtext><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mtext>&#x0398;</mml:mtext><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>&#x2217;</mml:mo></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo>}</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:munder><mml:mrow><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>g</mml:mi><mml:mrow><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>i</mml:mi><mml:mo>&#x2062;</mml:mo><mml:mi>n</mml:mi></mml:mrow></mml:mrow></mml:mrow><mml:mrow><mml:msub><mml:mtext>&#x0398;</mml:mtext><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mtext>&#x0398;</mml:mtext><mml:mrow><mml:mi>d</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:msup><mml:mrow><mml:mo>&#x2225;</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mover><mml:mi>x</mml:mi><mml:mo stretchy="false">^</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mo>&#x2225;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:mrow><mml:mo>.</mml:mo></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>The distance between the input and the output is called reconstruction error.</p>
<p>Autoencoders are widely used in anomaly detection due to their ability to learn compact representations of normal data. By training exclusively on normal traffic, autoencoder-based models can detect anomalies based on reconstruction errors, as anomalous instances tend to exhibit higher errors due to their deviation from learned normal patterns. Since reconstruction error quantifies the degree of deviation, it is commonly used as an anomaly score in detection models. A data instance is classified as anomalous when its anomaly score exceeds a predefined threshold, which is determined based on the characteristics of the dataset and the specific requirements of the detection process.</p>
<p>One of the key advantages of autoencoder-based detection models is their ability to handle data imbalance issues, a common challenge in real-world environments. Since these models do not require labeled anomalous samples, they are particularly effective in scenarios where anomalies are rarely present, such as DDoS attack detection, where attack instances constitute only a small fraction of overall network traffic. Furthermore, because the model learns the distribution of normal data rather than specific attack patterns, it can detect previously unseen anomalies without requiring additional retraining. This adaptability makes autoencoder-based approaches highly effective for various anomaly detection tasks.</p>
<p>Among the various autoencoder models, CAEs are particularly effective for processing high-dimensional time-series data like network traffic, as they leverage convolutional operations to capture local dependencies within each sequence or window [<xref ref-type="bibr" rid="ref-9">9</xref>]. Despite these advantages, CAEs still face critical limitations in DDoS detection, such as over-generalization and ineffective anomaly scoring.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Strategies for Autoencoder-Based Anomaly Detection</title>
<p>To improve anomaly detection performance, autoencoder-based models typically adopt one of two strategies. One approach focuses on enhancing the model&#x2019;s ability to accurately reconstruct normal data, often by modifying the model architecture. A common technique is stacking, where multiple single-layer autoencoders are combined to create a stacked autoencoder, which has shown strong performance in detecting outliers [<xref ref-type="bibr" rid="ref-10">10</xref>]. Another study proposed an asymmetric stacked autoencoder, consisting of three encoders and a single decoder, further improving feature extraction for anomaly detection [<xref ref-type="bibr" rid="ref-11">11</xref>].</p>
<p>While these architectures can enhance feature extraction, they do not fundamentally resolve a key limitation of autoencoder-based anomaly detection: the reconstruction error overlap between normal and anomalous data. This overlap can cause models to inadvertently reconstruct anomalies too well, reducing their effectiveness in distinguishing between normal and attack traffic. Recent studies have also pointed out that vanilla AEs may suffer from excessive generalization, potentially reconstructing anomalous inputs as well as normal ones [<xref ref-type="bibr" rid="ref-12">12</xref>].</p>
<p>Alternatively, modifying the anomaly scoring method can also improve detection performance. One example is the Reconstruction along Projection Pathway (RaPP) model, which achieved high performance by adjusting the anomaly score without altering the learning objective [<xref ref-type="bibr" rid="ref-13">13</xref>]. Unlike conventional approaches that rely solely on the final reconstruction error, RaPP computes an alternative anomaly score by utilizing intermediate representations from both the encoder and decoder. By incorporating these additional layers, the model captures more nuanced deviations in data.</p>
<p>However, despite its improved scoring method, RaPP still relies on average reconstruction error, which has notable limitations. Specifically, it fails to prioritize recent anomalies and does not explicitly account for traffic volume and intensity, both of which are crucial for identifying DDoS attacks. As a result, while RaPP improves certain aspects of anomaly detection, it remains insufficient for scenarios where anomalies evolve dynamically over time or occur in bursts, as seen in large-scale DDoS attacks.</p>
<p>Given these limitations, there is a need for more effective architecture and a refined anomaly scoring mechanism specifically designed for DDoS detection. While increasingly complex models such as vision transformer variants have been explored in recent anomaly detection literature [<xref ref-type="bibr" rid="ref-14">14</xref>], we intentionally pursue performance gains through a lightweight convolutional autoencoder with minimal architectural changes. Our proposed framework addresses these challenges by introducing a joint reconstruction learning approach that amplifies the distinction between normal and anomalous traffic while leveraging an improved anomaly detection score that incorporates temporal importance and relative traffic volume.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Methodology</title>
<p>This section provides a detailed explanation of the framework for adapting a CAE for DDoS attack detection through joint reconstruction learning and refined anomaly scoring. The overall structure of the proposed methodology is illustrated in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Overview of our proposed framework</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-1.tif"/>
</fig>
<sec id="s3_1">
<label>3.1</label>
<title>Preprocessing</title>
<p>Network traffic data is sequential data where each traffic record contains multiple features collected by network devices. Since normal traffic patterns can vary across different devices, we preprocess the data by computing the difference between consecutive time steps instead of using absolute values. This transformation, commonly used in time-series analysis, helps stabilize the data and reduce device-specific variations [<xref ref-type="bibr" rid="ref-15">15</xref>]. The overall preprocessing flow is illustrated in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Preprocessing</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-2.tif"/>
</fig>
<p>Next, the computed differences are standardized to follow a normal distribution, ensuring consistency across features. Finally, we segment the traffic sequence into overlapping windows of fixed length (or window size) <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:mi>w</mml:mi></mml:math></inline-formula>, where each window <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> consists of <inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mi>w</mml:mi></mml:math></inline-formula> consecutive standardized difference values. This window serves as the unit of classification. Each data point <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is represented by a feature vector of size <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mrow><mml:mo>|</mml:mo><mml:mi>F</mml:mi><mml:mo>|</mml:mo></mml:mrow></mml:math></inline-formula>, where <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mi>F</mml:mi></mml:math></inline-formula> denotes a set of features extracted from network traffic. The label of a given window <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, denoted as <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi>l</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, is determined by the label of its last data point. This labeling strategy ensures that detection focuses on the most recent network state, which is crucial for real-time DDoS attack detection. The above preprocessing procedure can be mathematically formulated as follows (<xref ref-type="disp-formula" rid="eqn-3">Eqs. (3)</xref>&#x2013;<xref ref-type="disp-formula" rid="eqn-5">(5)</xref>).
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>+</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>}</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:msubsup><mml:mi>f</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msubsup><mml:mi>f</mml:mi><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>F</mml:mi><mml:mo>|</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msubsup><mml:mo>}</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mi>l</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>w</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>l</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>+</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:mtable columnalign="left left" rowspacing=".2em" columnspacing="1em" displaystyle="false"><mml:mtr><mml:mtd><mml:mn>0</mml:mn></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mn>1</mml:mn></mml:mtd></mml:mtr></mml:mtable><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Consecutive Asymmetric Convolutional Autoencoder</title>
<p>Single CAEs often struggle to effectively separate normal and anomalous data, as they tend to reconstruct both with similar accuracy. To address this limitation, we propose Consecutive Asymmetric CAE (CA-CAE), a model featuring joint reconstruction task and asymmetric architecture. <xref ref-type="fig" rid="fig-3">Fig. 3</xref> illustrates CA-CAE, which consists of two autoencoders connected sequentially. Specifically, the output of the first autoencoder (CAE<sub>1</sub>) serves as the input to the second autoencoder (CAE<sub>2</sub>), leading to two consecutive reconstruction processes. This sequential reconstruction helps amplify the distinction between normal and anomalous data, making anomalies more detectable.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Architecture of CA-CAE</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-3.tif"/>
</fig>
<p>CA-CAE shares similarities with stacked autoencoders, but it differs in two key aspects: (1) its loss function, which is designed to enhance anomaly separation, and (2) its asymmetric architecture, where each autoencoder has a different number of filters. These modifications improve the model&#x2019;s ability to distinguish between normal and anomalous traffic more effectively.</p>
<p>Unlike conventional stacked architectures, CA-CAE deliberately employs an asymmetric configuration where CAE<sub>2</sub> has a greater capacity (i.e., more convolutional filters) than CAE<sub>1</sub>. This design allows CAE<sub>2</sub> to perform a more complex and finer reconstruction, placing a higher burden on the model when processing inputs. For normal traffic, the learned feature representation remains sufficiently robust to undergo this more difficult reconstruction, resulting in low reconstruction error. In contrast, anomalous traffic, which deviates from learned normal patterns, struggles through the second reconstruction stage, resulting in amplified reconstruction errors. This asymmetric setup thus increases the separation between normal and anomalous data, enhancing detection performance.</p>
<p>The sample-wise loss function for the joint reconstruction task is defined as follows (<xref ref-type="disp-formula" rid="eqn-6">Eq. (6)</xref>).
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:msub><mml:mrow><mml:mi>&#x02112;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03B1;</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo>,</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> denotes the mean squared error. The first term, <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, measures the reconstruction error of the original input from the CAE<sub>1</sub>, while the second term, <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, represents the reconstruction error between the original input and the final output of the entire CA-CAE model. The two terms are weighted by the parameter <inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula>, balancing the contributions of CAE<sub>1</sub> and CAE<sub>2</sub> during training.</p>
<p>The role of CAE<sub>1</sub> is not merely to reconstruct the input <inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, but also to generate an intermediate representation <inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> that facilitates the learning process of CAE<sub>2</sub>. Since CAE<sub>2</sub> reconstructs the original <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> instead of <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, CAE<sub>1</sub> must learn to preserve essential features that enable better reconstruction in the next step. This additional constraint makes training more challenging, encouraging CAE<sub>1</sub> to generate a more informative representation that captures key patterns in the data.</p>
<p>From another perspective, the transformation of <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> into <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> by CAE<sub>1</sub> resembles the corruption step in denoising autoencoders. In other words, CAE<sub>2</sub> learns to refine and recover the original data from a partially transformed version, which helps mitigate overfitting and enhances detection performance. By forcing CAE<sub>2</sub> to reconstruct from a modified version of the input, the model becomes more robust to variations in normal data while amplifying differences in anomalous data.</p>
<p>We hypothesize that this joint reconstruction task increases the separation between normal and anomalous sequences in terms of reconstruction error distribution. Since reconstructing the original data through two consecutive steps is a more complex task than single-step reconstruction, we expect that normal sequences will be reconstructed with higher precision, while anomalous sequences will retain higher reconstruction errors. This assumption is validated in <xref ref-type="sec" rid="s4_2_2">Section 4.2.2</xref>, where we demonstrate that CA-CAE significantly enhances the contrast between normal and anomalous reconstruction errors, improving DDoS detection performance.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Refined Anomaly Scoring for DDoS Attack Detection</title>
<p>In CAE-based anomaly detection, anomalies are typically identified when the anomaly score exceeds a pre-determined threshold. To enhance DDoS attack detection, we propose a refined detection score, which extends the conventional anomaly score by replacing the average reconstruction error with a more adaptive metric. Our detection score incorporates two key factors: temporal importance and relative traffic volume, ensuring a more accurate assessment of anomalous traffic.</p>
<p>First, we apply an exponential decay to weigh recent time points more heavily when computing the average reconstruction error, ensuring that the detection score is more influenced by recent anomalies. Next, we introduce a risk factor, which quantifies the potential harm of network traffic based on its volume, and adjust the detection score accordingly. By integrating these factors, our detection score provides a more effective and context-aware measure for distinguishing DDoS attack traffic from normal network behavior.</p>
<sec id="s3_3_1">
<label>3.3.1</label>
<title>Exponential Decay</title>
<p>Average reconstruction error is one of the most widely used anomaly scores for CAE-based detection. However, as the window size (or sequence length) increases, it becomes less effective because it fails to emphasize recent anomalies. For example, if a DDoS attack has just occurred but the sequence contains a large proportion of earlier normal data, the average reconstruction error remains low despite a high error at recent time steps. This highlights a trade-off between incorporating historical information and maintaining high detection sensitivity.</p>
<p>To address this issue, we adopt the concept of exponential decaying, commonly used in time-series forecasting, to assign higher weights to recent observations while gradually reducing the influence of older data points [<xref ref-type="bibr" rid="ref-16">16</xref>]. By applying exponential decay, the anomaly score could be more responsive to recent anomalies. The decayed (or weighted) anomaly score is defined as follows (<xref ref-type="disp-formula" rid="eqn-7">Eq. (7)</xref>).
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>w</mml:mi></mml:mfrac><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msubsup><mml:msup><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msup><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> is the decay constant (<inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:mn>0</mml:mn><mml:mo>&#x003C;</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>&#x2264;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula>) that controls the rate at which older reconstruction errors decay, and <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msub><mml:mi>e</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> represents the reconstruction error at the <inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:mi>i</mml:mi></mml:math></inline-formula>-th time step. A higher <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> gives more weight to recent anomalies, while a lower <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> retains more influence from past reconstruction errors.</p>
<p>As a result, the weighting sequence follows <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:mo stretchy="false">[</mml:mo><mml:msup><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mn>0</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msup><mml:mi>&#x03B4;</mml:mi><mml:mrow><mml:mi>w</mml:mi></mml:mrow></mml:msup><mml:mo stretchy="false">]</mml:mo></mml:math></inline-formula>, where the weights gradually decrease over time. This formulation ensures that the score remains highly sensitive to recent anomalies, improving its effectiveness in identifying ongoing DDoS attacks.</p>
</sec>
<sec id="s3_3_2">
<label>3.3.2</label>
<title>Relative Traffic Volume</title>
<p>A major limitation of using average reconstruction error as an anomaly score is its inability to reliably distinguish DDoS attacks from other benign anomalies, leading to a high false positive rate. In particular, sequences containing abnormal traffic patterns that are unrelated to an actual attack may be misclassified as DDoS incidents.</p>
<p>To address this issue, we refined the detection score by incorporating relative traffic volume, which represents the potential risk posed by the traffic. Anomalies with high reconstruction error but low traffic volume are unlikely to be DDoS attacks, even when their anomaly scores exceed the threshold. Therefore, adjusting the detection score based on traffic intensity helps reduce false positives while maintaining high sensitivity to actual attacks.</p>
<p>To quantify relative traffic volume, we extract features that directly represent network load. Specifically, we use the total number of transmitted bytes and packets, as these features provide a direct measure of network activity (described further in <xref ref-type="sec" rid="s4_1_2">Section 4.1.2</xref>). The norms of these features are computed, and min-max scaling is applied to normalize them within the sequence, ensuring comparability across different time steps.</p>
<p>Since DDoS attacks manifest as sudden spikes in traffic, relative traffic volume is computed only at the final timestep of each sequence, rather than over the entire sequence. This ensures that the detection score reflects the most recent traffic state, improving its responsiveness to real-time attack patterns.</p>
<p>We refer to the relative traffic volume value as the risk coefficient (<inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:mi>r</mml:mi></mml:math></inline-formula>), while the safety coefficient is defined as <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:mi>s</mml:mi><mml:mtext>&#x00A0;</mml:mtext><mml:mo stretchy="false">(</mml:mo><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. The final detection score is then computed as follows (<xref ref-type="disp-formula" rid="eqn-8">Eq. (8)</xref>).
<disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>s</mml:mi><mml:mi>g</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:mi>g</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> is an interpolating function that adjusts the final detection score based on the safety level.</p>
<p>We consider the following scenario for our scoring: when the average reconstruction error is already low enough, <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> should remain close to <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Conversely, when the reconstruction error is high, the score should be adjusted based on the safety coefficient <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mi>s</mml:mi></mml:math></inline-formula>. If the safety level is low (<inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:mi>s</mml:mi></mml:math></inline-formula> is small), the high error value should be maintained. Otherwise, if <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:mi>s</mml:mi></mml:math></inline-formula> is large, the error should be interpolated downward to reduce false alarms.</p>
<p>To satisfy these conditions, the interpolating function <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mi>g</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> must meet the following constraints (<xref ref-type="disp-formula" rid="eqn-9">Eq. (9)</xref>).
<disp-formula id="eqn-9"><label>(9)</label><mml:math id="mml-eqn-9" display="block"><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mtable columnspacing="1em" rowspacing=".2em" columnalign="left left"><mml:mtr><mml:mtd><mml:mrow><mml:mrow><mml:munder><mml:mo movablelimits="true">lim</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:mrow></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:msup><mml:mi>g</mml:mi><mml:mrow><mml:mi>&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mrow><mml:munder><mml:mo movablelimits="true">lim</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo stretchy="false">&#x2192;</mml:mo><mml:mi mathvariant="normal">&#x221E;</mml:mi></mml:mrow></mml:munder><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:msup><mml:mi>g</mml:mi><mml:mrow><mml:mi>&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mrow><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mrow><mml:mrow><mml:mrow><mml:msup><mml:mi>g</mml:mi><mml:mrow><mml:mi>&#x2032;</mml:mi><mml:mi>&#x2032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>&lt;</mml:mo><mml:mn>0</mml:mn></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mi mathvariant="normal">&#x2200;</mml:mi><mml:mi>x</mml:mi></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:mrow><mml:mo fence="true" stretchy="true" symmetric="true"></mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>We adopt the Softplus function <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:mi>&#x03B6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> as <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mi>g</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, defined as follows:
<disp-formula id="eqn-10"><label>(10)</label><mml:math id="mml-eqn-10" display="block"><mml:mi>&#x03B6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>ln</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msup><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>Incorporating the safety coefficient <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:mi>s</mml:mi></mml:math></inline-formula>, the final detection score is given by:
<disp-formula id="eqn-11"><label>(11)</label><mml:math id="mml-eqn-11" display="block"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>s</mml:mi><mml:mi>&#x03B6;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>This formulation ensures smooth interpolation based on the safety level, effectively adjusting the detection score according to the traffic risk assessment while ensuring non-negative values.</p>
<p><xref ref-type="fig" rid="fig-4">Fig. 4</xref> illustrates the behavior of the detection score function with respect to different safety coefficients. As the safety level increases, the detection score is reduced, preventing low-risk anomalies from being misclassified as DDoS attacks. When the safety coefficient <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:mi>s</mml:mi></mml:math></inline-formula> is 1.0 (indicating the lowest traffic volume), the detection score remains zero for all values of <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>, ensuring that low-volume traffic anomalies are not falsely flagged as attacks.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Effect of safety coefficient on detection score</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-4.tif"/>
</fig>
<p>To further illustrate how the Softplus function adjusts the final detection score based on the safety level, consider a numerical example. Let&#x2019;s assume a decayed anomaly score <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> of 5.0.
<disp-formula id="eqn-12"><label>(12)</label><mml:math id="mml-eqn-12" display="block"><mml:mi>&#x03B6;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>&#x03B6;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>5.0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>l</mml:mi><mml:mi>n</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mn>5</mml:mn></mml:mrow></mml:msup><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2248;</mml:mo><mml:mn>5.0067</mml:mn></mml:math></disp-formula></p>
<p>If the safety coefficient <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mi>s</mml:mi></mml:math></inline-formula> is 0.8 (indicating low traffic volume, hence safer context), detection score can be calculated as follows:
<disp-formula id="eqn-13"><label>(13)</label><mml:math id="mml-eqn-13" display="block"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mn>0.8</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>5.0067</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mn>4.00536</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2248;</mml:mo><mml:mn>0.9946</mml:mn></mml:math></disp-formula></p>
<p>Here, the detection score is significantly reduced, mitigating potential false positives for low-risk anomalies. Otherwise, if the safety coefficient <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:mi>s</mml:mi></mml:math></inline-formula> is 0.2 (indicating higher traffic volume), the detection score is slightly reduced from the decayed anomaly score.
<disp-formula id="eqn-14"><label>(14)</label><mml:math id="mml-eqn-14" display="block"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mn>0.2</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mn>5.0067</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo movablelimits="true" form="prefix">max</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>5.0</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mn>1.00134</mml:mn><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2248;</mml:mo><mml:mn>3.9987</mml:mn></mml:math></disp-formula></p>
<p>This example highlights how the safety coefficient adaptively lowers the detection score for less critical anomalies, ensuring that high-risk, high-volume DDoS attacks are prioritized while effectively managing false alarms.</p>
</sec>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Training and Inference Workflow</title>
<p>We summarize the overall CA-CAE training and inference procedure in the form of a simplified pseudo code in Algorithm 1, which outlines model training on normal traffic, threshold selection on validation data, and test-time detection on unseen traffic.</p>
<fig id="fig-6">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-6.tif"/>
</fig>
</sec>
<sec id="s3_5">
<label>3.5</label>
<title>Hyperparameter Tuning Considerations</title>
<p>Our framework introduces several hyperparameters, including window size <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:mi>w</mml:mi></mml:math></inline-formula>, <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula> in <xref ref-type="disp-formula" rid="eqn-6">Eq. (6)</xref>, and <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> in <xref ref-type="disp-formula" rid="eqn-7">Eq. (7)</xref>. Each parameter serves a specific role in enhancing detection performance. <xref ref-type="table" rid="table-1">Table 1</xref> summarizes the key hyperparameters along with their roles and recommended settings.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Summary of hyperparameters with functional role and recommended ranges</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Hyperparameter</th>
<th>Description</th>
<th>Default (Suggested range)</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:mi>w</mml:mi></mml:math></inline-formula></td>
<td>Window size (sequence length)</td>
<td>Data-dependent</td>
</tr>
<tr>
<td><inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula></td>
<td>Loss weighting between CAE<sub>1</sub> and CAE<sub>2</sub></td>
<td>0.5 ([0.3, 0.7])</td>
</tr>
<tr>
<td><inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula></td>
<td>Decay constant emphasizing recent time steps</td>
<td>0.7 ([0.5, 0.9])</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The window size <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:mi>w</mml:mi></mml:math></inline-formula> determines how many previous time steps are used for each DDoS detection decision. A larger <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:mi>w</mml:mi></mml:math></inline-formula> incorporates more historical context, which may improve detection stability but also risks diluting recent anomalies. A smaller <inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:mi>w</mml:mi></mml:math></inline-formula> focuses on immediate patterns but may become too sensitive to noise. Thus, selecting <inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:mi>w</mml:mi></mml:math></inline-formula> reflects a trade-off between historical awareness and real-time responsiveness. The loss balancing coefficient <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:mi>&#x03B1;</mml:mi></mml:math></inline-formula> balances the reconstruction errors from CAE<sub>1</sub> and CAE<sub>2</sub>, affecting how much the model relies on intermediate vs. final reconstruction output. The decay constant <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> governs the temporal weighting of reconstruction errors within a window, allowing the model to prioritize recent time steps, which is critical in bursty attack scenarios.</p>
<p>These values can be selected using simple tuning strategies such as grid or randomized search, guided by the operational context and model objectives.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experiment</title>
<sec id="s4_1">
<label>4.1</label>
<title>Experimental Setup</title>
<sec id="s4_1_1">
<label>4.1.1</label>
<title>Dataset</title>
<p>We conducted experiments on a dataset provided by a software company. The traffic data was collected from the company&#x2019;s network router using Cisco NetFlow, a protocol for network traffic monitoring. The dataset spans two weeks, from August 1st to August 14th, 2020, with traffic data collected at 1-min intervals. We used the first week&#x2019;s traffic as training data and the second week&#x2019;s traffic as test data.</p>
<p>The dataset consists of 16 Internet Protocols (IPs), among which 5 IPs include attack traffic, while 11 IPs contain only normal traffic. During training, only normal sequences were used to train the CA-CAE model, whereas both normal and anomalous sequences were included in the test phase.</p>
<p>Due to the rarity of DDoS attacks, the dataset exhibits a high degree of class imbalance, with 99.9% normal data and only 0.0004% anomalous data. Specifically, the dataset contains 190,817 normal instances and 78 anomalous instances, meaning that 78 windows are defined as anomalous.</p>
<p>Each traffic record includes 12 features, such as bytes, packets, and source IP entropy, which provide key indicators of network behavior.</p>
</sec>
<sec id="s4_1_2">
<label>4.1.2</label>
<title>Features</title>
<p>We extracted the following key features from the dataset for DDoS detection:
<list list-type="bullet">
<list-item>
<p><bold>Entropy of IP addresses and port numbers.</bold> During a bandwidth DDoS attack, the number of unique source IP addresses and source port numbers increases significantly, leading to higher entropy [<xref ref-type="bibr" rid="ref-17">17</xref>]. Conversely, the entropy of destination IP addresses and destination port numbers decreases, as a large volume of traffic is directed towards a single target.</p></list-item>
<list-item>
<p><bold>Total size of bytes and packets.</bold> The total volume of bytes and packets in incoming traffic is a useful indicator for DDoS detection, as attack traffic tends to be significantly larger than normal traffic [<xref ref-type="bibr" rid="ref-18">18</xref>].</p></list-item>
<list-item>
<p><bold>Ratio and entropy of protocols.</bold> DDoS attacks often involve botnets flooding the target server with traffic using a single protocol. This causes a spike in the ratio of a specific protocol compared to normal conditions. The dominant protocol also varies depending on the attack type&#x2014;User Datagram Protocol (UDP) flooding increases the ratio of UDP packets, while Internet Control Message Protocol (ICMP) flooding increases the ratio of ICMP packets [<xref ref-type="bibr" rid="ref-19">19</xref>]. As a result, protocol entropy decreases during an attack.</p></list-item>
<list-item>
<p><bold>Entropy of TCP flags.</bold> In TCP SYN flooding attacks, the proportion of TCP packets increases. However, this alone is insufficient for detection. Instead, we analyze the entropy of TCP flags to determine whether the TCP three-way handshake is functioning normally.</p></list-item>
</list></p>
<p>Among these features, entropy-based metrics are computed using the following formula:
<disp-formula id="eqn-15"><label>(15)</label><mml:math id="mml-eqn-15" display="block"><mml:mi>E</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>o</mml:mi><mml:mi>p</mml:mi><mml:mi>y</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>X</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:msubsup><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mi>log</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mi>P</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>The effectiveness of combining traffic volume indicators such as total byte and packet count with entropy-based metrics has also been demonstrated in recent studies [<xref ref-type="bibr" rid="ref-20">20</xref>], reinforcing the relevance of our selected feature set.</p>
<p>As described in <xref ref-type="sec" rid="s3_1">Section 3.1</xref>, for each variable, the difference between time steps <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:mi>t</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:mi>t</mml:mi></mml:math></inline-formula> is used as the data for time <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:mi>t</mml:mi></mml:math></inline-formula>. These values are then standardized to follow a standard normal distribution before being used as model inputs. Finally, we apply a sliding window technique with a window size of 10. For each window, the label of the last data point is assigned as the window&#x2019;s label. In other words, we aim to detect anomalies at the latest timestep by leveraging traffic data from the previous 10 min.</p>
</sec>
<sec id="s4_1_3">
<label>4.1.3</label>
<title>Evaluation</title>
<p>For evaluation, we computed the precision, recall, and F1-score as defined in <xref ref-type="disp-formula" rid="eqn-16">Eqs. (16)</xref>&#x2013;<xref ref-type="disp-formula" rid="eqn-18">(18)</xref>. Since this is a binary classification task, our evaluation metrics are derived from the confusion matrix. Specifically, a true positive denotes a case where an anomalous instance is correctly detected. Given the severe class imbalance in our network traffic dataset, these metrics specifically reflect the model&#x2019;s performance in detecting the anomaly class.
<disp-formula id="eqn-16"><label>(16)</label><mml:math id="mml-eqn-16" display="block"><mml:mrow><mml:mtext>Precision</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>True Positives</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mtext>True Positives</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mtext>False Positives</mml:mtext></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:math></disp-formula>
<disp-formula id="eqn-17"><label>(17)</label><mml:math id="mml-eqn-17" display="block"><mml:mrow><mml:mtext>Recall</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>True Positives</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mtext>True Positives</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mtext>False Negatives</mml:mtext></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:math></disp-formula>
<disp-formula id="eqn-18"><label>(18)</label><mml:math id="mml-eqn-18" display="block"><mml:mrow><mml:mtext>F1</mml:mtext></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mtext>score</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mfrac><mml:mrow><mml:mrow><mml:mtext>Precision</mml:mtext></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mtext>Recall</mml:mtext></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mtext>Precision</mml:mtext></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mtext>Recall</mml:mtext></mml:mrow></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula></p>
</sec>
<sec id="s4_1_4">
<label>4.1.4</label>
<title>Threshold Selection</title>
<p>To determine the decision threshold for anomaly detection, we adopted a validation-based quantile thresholding strategy using only normal data. After training the CA-CAE model, we computed the final anomaly scores on a separate hold-out validation set composed exclusively of normal sequences (20% of the training data). This validation set, which was not used during training or testing, serves to assess the model&#x2019;s generalization performance under normal conditions.</p>
<p>Based on the distribution of validation scores, we considered several high-quantile candidates&#x2014;specifically the 99.90th, 99.95th, and 99.99th percentiles&#x2014;as potential thresholds. These candidate thresholds were evaluated on the test set, and the threshold corresponding to the 99.95th percentile (i.e., top 0.05%) consistently yielded the best F1-score across multiple runs. We therefore selected it as the final threshold for anomaly detection.</p>
<p>This procedure ensures that the decision boundary aligns with realistic deployment conditions, where labeled attack data may not be available. An overly conservative threshold (i.e., set too high) would reduce recall by failing to detect actual attacks, while an overly lenient threshold (i.e., set too low) would result in a high false positive rate. The proposed quantile-based thresholding offers a principled and reproducible way to balance sensitivity and precision, thereby enabling robust DDoS detection in highly imbalanced scenarios.</p>
</sec>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Performance Gains through Joint Reconstruction and Refined Scoring</title>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Detection Performance by Scoring Strategy</title>
<p><xref ref-type="table" rid="table-2">Table 2</xref> presents the detection performance of CAE and CA-CAE using different detection scores. To ensure the statistical robustness of our findings, we represented the mean and standard deviation of the performance metrics from 10 independent runs for each model. Across all types of detection scores, CA-CAE consistently outperformed CAE. Additionally, our proposed scoring techniques significantly improved detection performance across both models. By applying exponential decay to emphasize recent reconstruction errors, the F1-score improved by over 20%, suggesting that historical traffic information can introduce noise, and that mitigating its influence leads to significant gains in detection performance. Furthermore, incorporating the risk factor based on traffic volume further enhanced the accuracy of DDoS attack detection.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Detection performance of CAE and CA-CAE with respect to detection scores</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Detection score</th>
<th>Model</th>
<th>Precision</th>
<th>Recall</th>
<th>F1 score</th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="2">Original score</td>
<td>CAE</td>
<td>0.622 &#x00B1; 0.038</td>
<td>0.440 &#x00B1; 0.047</td>
<td>0.515 &#x00B1; 0.043</td>
</tr>
<tr>
<td>CA-CAE</td>
<td>0.714 &#x00B1; 0.016</td>
<td>0.662 &#x00B1; 0.038</td>
<td>0.687 &#x00B1; 0.016</td>
</tr>
<tr>
<td rowspan="2"><inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>y</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>CAE</td>
<td>0.769 &#x00B1; 0.053</td>
<td>0.829 &#x00B1; 0.042</td>
<td>0.797 &#x00B1; 0.026</td>
</tr>
<tr>
<td>CA-CAE</td>
<td>0.856 &#x00B1; 0.030</td>
<td>0.931 &#x00B1; 0.031</td>
<td>0.892 &#x00B1; 0.029</td>
</tr>
<tr>
<td rowspan="2"><inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>t</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula></td>
<td>CAE</td>
<td>0.867 &#x00B1; 0.036</td>
<td>0.864 &#x00B1; 0.048</td>
<td>0.865 &#x00B1; 0.023</td>
</tr>
<tr>
<td>CA-CAE</td>
<td>0.907 &#x00B1; 0.021</td>
<td>0.964 &#x00B1; 0.025</td>
<td>0.934 &#x00B1; 0.016</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>To statistically validate the significant performance enhancements of our proposed CA-CAE over the baseline CAE model, we conducted Wilcoxon Signed-Rank Tests on the F1-scores for each of the three scoring methods. The results consistently demonstrated that CA-CAE achieved a statistically significant improvement in F1-score over CAE across all three scoring methods (<italic>p</italic> &#x003C; 0.05). This robust statistical significance, observed uniformly across different scoring methods, confirms that the architectural enhancements through joint reconstruction learning and the refined anomaly scoring mechanisms lead to performance gains in DDoS attack detection compared to CAE-based detection.</p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Reconstruction Behavior in Joint Autoencoders</title>
<p>To evaluate the effect of joint reconstruction and asymmetric architecture in CA-CAE, we compared two mean squared errors: <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> for each input window <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:mi>&#x03C7;</mml:mi></mml:math></inline-formula>. If <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003E;</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, it indicates that the joint reconstruction task improved the final reconstruction quality. Conversely, if <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003C;</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, the final reconstruction quality worsened after passing through the second autoencoder.</p>
<p><xref ref-type="table" rid="table-3">Table 3</xref> summarizes the results when the total number of filters across the two CAEs is 36. The notation (M, N) indicates the number of filters assigned to CAE<sub>1</sub> and CAE<sub>2</sub>, respectively. In the symmetric configuration (18, 18), 89.0% of normal windows showed improved reconstruction after joint reconstruction, while only 17.3% of anomalous windows did. In contrast, 82.7% of anomalous windows exhibited an increased reconstruction error. This indicates that the joint reconstruction task strengthens the model&#x2019;s ability to reconstruct normal traffic while making anomalous traffic more difficult to reconstruct. This discrepancy increases the separation between the reconstruction error distributions of normal and anomalous sequences, thereby enhancing anomaly detectability.</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Comparison of mean squared error</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Structure</th>
<th>Window</th>
<th><inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="bold">&#x2032;</mml:mi></mml:mrow></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003E;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="bold">&#x2032;</mml:mi><mml:mi mathvariant="bold">&#x2032;</mml:mi></mml:mrow></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></th>
<th><inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="bold">&#x2032;</mml:mi></mml:mrow></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x003C;</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi mathvariant="bold-italic">&#x03C7;</mml:mi><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="bold">&#x2032;</mml:mi><mml:mi mathvariant="bold">&#x2032;</mml:mi></mml:mrow></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></th>
</tr>
</thead>
<tbody>
<tr>
<td rowspan="2">Symmetric (18, 18)</td>
<td>Normal</td>
<td>0.890</td>
<td>0.110</td>
</tr>
<tr>
<td>Anomalous</td>
<td>0.173</td>
<td>0.827</td>
</tr>
<tr>
<td rowspan="2">Asymmetric (12, 24)</td>
<td>Normal</td>
<td>0.931</td>
<td>0.069</td>
</tr>
<tr>
<td>Anomalous</td>
<td>0.146</td>
<td>0.854</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>This effect is further amplified in the asymmetric configuration (12, 24), where more filters are allocated to CAE<sub>2</sub>, which is responsible for the more challenging reconstruction task. As shown in <xref ref-type="table" rid="table-3">Table 3</xref>, in the symmetric setting, the gap between the normal and the anomalous was 0.717 (0.890<inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:mo>&#x2212;</mml:mo></mml:math></inline-formula>0.173), whereas in the asymmetric setting, it increased to 0.785 (0.931<inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:mo>&#x2212;</mml:mo></mml:math></inline-formula>0.146).</p>

<p>Since CAE<sub>2</sub> reconstructs an already compressed representation from CAE<sub>1</sub>, allocating more filters allows it to extract richer features for normal traffic. However, for anomalous data, the increased model complexity amplifies reconstruction errors rather than improving reconstruction quality, thereby further widening the gap between normal and anomalous reconstruction errors.</p>
<p>While the absolute numerical increase in the reconstruction error gap between normal and anomalous samples may appear modest, its functional impact is meaningful in the context of highly imbalanced anomaly detection. Even a slight increase in this gap enhances the model&#x2019;s ability to distinguish anomalous samples with higher confidence, thereby contributing to improved anomaly separability and more reliable detection performance.</p>
<p>Importantly, this improvement was achieved through a simple yet effective architectural adjustment: reallocating a fixed total number of filters asymmetrically across the two reconstruction stages. This lightweight design change avoids increasing model complexity while still producing a measurable benefit in detection quality.</p>
</sec>
<sec id="s4_2_3">
<label>4.2.3</label>
<title>Computational Cost and Model Efficiency</title>
<p>To evaluate the computational cost of our proposed CA-CAE model, we compared its parameter size and inference latency with the baseline CAE. For a representative configuration, the CAE model consisted of 18 convolutional filters, while the CA-CAE allocated 12 filters in CAE<sub>1</sub> and 24 in CAE<sub>2</sub>.</p>
<p>The number of trainable parameters was 2934 for CAE and 6084 for CA-CAE, representing a 207% increase. We further measured the inference time using a single window of size 10 (with 12 input features) on an NVIDIA T4 GPU. Each experiment was repeated 10 times with 1000 inference iterations per trial to obtain statistically stable measurements.</p>
<p>The baseline CAE required an average of 0.495 ms per forward pass, with a standard deviation of 0.011 ms, whereas the proposed CA-CAE required 0.987 ms on average, with a standard deviation of 0.015 ms.</p>
<p>While the CA-CAE architecture introduces a moderate increase in computational cost&#x2014;both in terms of parameter size and inference time&#x2014;this overhead is accompanied by a substantial improvement in detection performance. The average inference latency remains within a range compatible with near-real-time network monitoring.</p>
<p>In this study, we placed greater emphasis on achieving reliable anomaly detection under severe class imbalance, even at the cost of slight increases in latency and model size. We view this design as a pragmatic trade-off between computational efficiency and model robustness. Furthermore, the architecture remains amenable to deployment-oriented optimizations such as pruning and quantization, which can mitigate computational burden in practice.</p>
</sec>
<sec id="s4_2_4">
<label>4.2.4</label>
<title>Effect of Detection Score Definitions</title>
<p>In addition, we examined the impact of different detection score definitions based on reconstruction errors. <xref ref-type="table" rid="table-4">Table 4</xref> presents the detection performance when using <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula>, and their combination as anomaly scores. Using <inline-formula id="ieqn-101"><mml:math id="mml-ieqn-101"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> alone yielded a precision of 0.884 and a recall of 0.913. Using <inline-formula id="ieqn-102"><mml:math id="mml-ieqn-102"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula> improved recall to 0.970 but resulted in a reduced precision to 0.841. Combining the two errors yielded the best performance, achieving an F1-score of 0.934.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Detection performance depending on loss as anomaly score</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Anomaly score</th>
<th>Precision</th>
<th>Recall</th>
<th>F1 score</th>
</tr>
</thead>
<tbody>
<tr>
<td><inline-formula id="ieqn-103"><mml:math id="mml-ieqn-103"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>0.884</td>
<td>0.913</td>
<td>0.898</td>
</tr>
<tr>
<td><inline-formula id="ieqn-104"><mml:math id="mml-ieqn-104"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>0.841</td>
<td>0.970</td>
<td>0.901</td>
</tr>
<tr>
<td><inline-formula id="ieqn-105"><mml:math id="mml-ieqn-105"><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>L</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mi>&#x03C7;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x2032;</mml:mi><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td>0.907</td>
<td>0.964</td>
<td>0.934</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>These findings suggest that CAE<sub>1</sub> effectively reconstructs normal traffic while introducing distortions in anomalous traffic, thereby enhancing the distinguishability between normal and anomalous instances. As a result, CAE<sub>2</sub> struggles more with anomalies, further amplifying the reconstruction error gap. This sequential difficulty in reconstruction ultimately improves recall by making anomalies more easily detectable.</p>
</sec>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Impact of Window Size and Exponential Decay</title>
<p>As discussed in <xref ref-type="sec" rid="s3_3_1">Section 3.3.1</xref>, anomaly detection in sequential data requires balancing historical context with real-time adaptability. To analyze this trade-off, we evaluated the impact of two hyperparameters: window size <inline-formula id="ieqn-106"><mml:math id="mml-ieqn-106"><mml:mi>w</mml:mi></mml:math></inline-formula> and decay constant <inline-formula id="ieqn-107"><mml:math id="mml-ieqn-107"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> on detection performance. Window size determines how much historical data is considered for each decision [<xref ref-type="bibr" rid="ref-21">21</xref>], whereas the decay constant <inline-formula id="ieqn-108"><mml:math id="mml-ieqn-108"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> controls the relative importance of recent observations. If <inline-formula id="ieqn-109"><mml:math id="mml-ieqn-109"><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>, only the most recent observation influences the anomaly score, whereas <inline-formula id="ieqn-110"><mml:math id="mml-ieqn-110"><mml:mi>&#x03B4;</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:math></inline-formula> assigns equal importance to all observations within the window.</p>
<p><xref ref-type="fig" rid="fig-5">Fig. 5</xref> illustrates how the F1-score varies with difference values of <inline-formula id="ieqn-111"><mml:math id="mml-ieqn-111"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> across multiple window sizes. Solid lines indicate the average F1-score over 10 runs, and shaded regions represent standard deviation, highlighting variability in detection sensitivity. Several trends can be observed.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Detection performance across different decay constants and window sizes</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_67211-fig-5.tif"/>
</fig>
<p>For smaller window sizes (<italic>w</italic> &#x003D; 5), the decay constant has minimal effect, as the sequence already consists mostly of recent data. However, for larger window sizes (<italic>w</italic> &#x003D; 20 and <italic>w</italic> &#x003D; 30), performance drops significantly as <inline-formula id="ieqn-112"><mml:math id="mml-ieqn-112"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> approaches 1, indicating that assigning equal weight to all historical data introduces noise and degrades detection capability. This highlights the necessity of using exponential decay to prioritize recent observations, especially when longer historical windows are employed.</p>
<p>The best performance was achieved when the window size <italic>w</italic> was 10 and the decay constant <inline-formula id="ieqn-113"><mml:math id="mml-ieqn-113"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> was 0.7, achieving the highest F1-score. This suggests that incorporating a moderate amount of historical context while emphasizing recent traffic leads to optimal detection performance.</p>
<p>These results validate the effectiveness of the exponential decay mechanism in balancing the benefits of historical information and the need for real-time responsiveness. Furthermore, applying exponential decay reduces the model&#x2019;s sensitivity to the choice of window size, as even suboptimal windows maintain stable performance by focusing primarily on recent observations.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Comparison with Baseline Models</title>
<p>While the primary focus of this study is on improving CAE-based anomaly detection through joint reconstruction learning and refined scoring, we also compared our proposed CA-CAE model against several representative baseline algorithms to evaluate its relative performance in a broader context.</p>
<p>Specifically, we included one supervised learning model and two unsupervised anomaly detection models for comparison.</p>
<p>The supervised baseline is the LSTM Classifier [<xref ref-type="bibr" rid="ref-22">22</xref>], a recurrent neural network trained on labeled attack data. Although it can effectively capture sequential traffic patterns, its reliance on labeled datasets limits its ability to generalize to novel or evolving attack types.</p>
<p>The unsupervised baselines are:
<list list-type="bullet">
<list-item>
<p>Isolation Forest [<xref ref-type="bibr" rid="ref-23">23</xref>]: A tree-based anomaly detection algorithm that isolates anomalies by recursively partitioning the feature space. It assumes that anomalies are easier to separate (i.e., require fewer splits) than normal data.</p></list-item>
<list-item>
<p>RaPP: An enhanced autoencoder model that calculates anomaly scores not only based on the final reconstruction error but also by aggregating deviations across multiple layers in the encoder and decoder pathways.</p></list-item>
<list-item>
<p>LSTM-VAE [<xref ref-type="bibr" rid="ref-24">24</xref>]: A recurrent variational autoencoder that detects anomalies in time-series data by modeling normal temporal patterns and identifying low-probability sequences based on reconstruction likelihood.</p></list-item>
</list></p>
<p>Finally, CAE serves as our direct baseline to verify the effectiveness of joint reconstruction and refined anomaly scoring.</p>
<p><xref ref-type="table" rid="table-5">Table 5</xref> presents the precision, recall, and F1-score for each model. As shown, CA-CAE consistently achieves superior performance, with an F1-score of 0.934&#x2014;significantly higher than the best-performing baseline, the LSTM-VAE, which attained 0.811.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Performance comparison among anomaly detection models</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr>
<th>Method</th>
<th>Precision</th>
<th>Recall</th>
<th>F1 score</th>
</tr>
</thead>
<tbody>
<tr>
<td>LSTM Classifier</td>
<td>0.711</td>
<td>0.723</td>
<td>0.717</td>
</tr>
<tr>
<td>Isolation forest</td>
<td>0.030</td>
<td>0.633</td>
<td>0.057</td>
</tr>
<tr>
<td>RaPP</td>
<td>0.552</td>
<td>0.510</td>
<td>0.530</td>
</tr>
<tr>
<td>LSTM-VAE</td>
<td>0.793</td>
<td>0.830</td>
<td>0.811</td>
</tr>
<tr>
<td>CAE</td>
<td>0.622</td>
<td>0.440</td>
<td>0.515</td>
</tr>
<tr>
<td><bold>CA-CAE (ours)</bold></td>
<td><bold>0.907</bold></td>
<td><bold>0.964</bold></td>
<td><bold>0.934</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>From these results, several important observations emerge. While supervised models such as the LSTM Classifier perform moderately well, their reliance on labeled data limits their scalability to unseen attack patterns. Notably, Isolation Forest exhibited extremely low precision despite a moderate recall, likely due to its sensitivity to threshold selection and its inability to effectively discriminate rare anomalies in highly imbalanced traffic datasets. RaPP, while improving upon standard autoencoders by leveraging intermediate representations, still falls short in fully capturing temporal importance and traffic-specific risks essential for DDoS detection. LSTM-VAE achieves relatively strong performance; however, it may still exhibit limitations in distinguishing subtle anomalies or adapting to sudden shifts in traffic behavior.</p>
<p>In contrast, our proposed CA-CAE model significantly outperforms all baselines by integrating architectural improvements with a domain-aware anomaly scoring method, resulting in superior performance in real-time DDoS attack detection.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusion</title>
<p>This study introduced CA-CAE, an anomaly detection framework for DDoS detection that integrates asymmetric joint reconstruction learning with a refined anomaly scoring method. Our experimental results demonstrate that the proposed approach significantly improves anomaly separability and detection accuracy.</p>
<p>A key finding is that asymmetric joint reconstruction enhances the distinction between normal and anomalous data. By allocating more filters to the second CAE, which handles a more challenging reconstruction task, the model reconstructs normal data with lower error while amplifying reconstruction errors for anomalous data. This sequential reconstruction difficulty increased the gap between normal and abnormal reconstruction errors, with the gap widening from 0.717 to 0.785 in the asymmetric setting, thereby making anomalies more easily distinguishable.</p>
<p>In addition to architectural improvements, our refined anomaly scoring method further boosts detection performance. The use of an exponential decay filter prioritizes recent traffic patterns, while adjusting scores based on relative traffic volume ensures that high-risk anomalies are appropriately emphasized. These refinements resulted in notable improvements across all evaluation metrics, reinforcing the critical role of scoring mechanisms in anomaly detection.</p>
<p>Compared to baseline models, CA-CAE consistently outperformed both supervised and unsupervised approaches. While the LSTM classifier achieved moderate performance, its reliance on labeled data limits its adaptability to unseen attacks. Traditional autoencoder-based methods, such as CAE and RaPP, also struggled to effectively distinguish between normal and abnormal traffic. In contrast, CA-CAE achieved an F1-score of 0.934, significantly surpassing the best-performing baseline model.</p>
<p>Despite these promising results, several challenges remain. The model&#x2019;s sensitivity to hyperparameters&#x2014;particularly the decay factor and window size&#x2014;could affect performance stability. Although exponential decay mitigates this issue to some extent, adaptive parameter tuning strategies would further improve robustness. Additionally, while this study focused on DDoS attack detection, applying CA-CAE to broader cybersecurity threats such as data exfiltration and advanced persistent threats remain an important avenue for future work.</p>
<p>In addition, recent advances in distributed and semantic anomaly detection, such as federated learning and the use of large language models, offer new research opportunities. For instance, LLM-AE-MP demonstrates the potential of integrating LLMs with autoencoder-based architectures for more contextualized detection of web-based threats [<xref ref-type="bibr" rid="ref-25">25</xref>]. Exploring such directions could extend CA-CAE to more intelligent detection frameworks.</p>
<p>Finally, real-time deployment considerations require further investigation. Although CA-CAE demonstrated strong offline performance, optimizing computational efficiency for high-speed, real-world network environments will be critical for practical adoption. Addressing these challenges will further enhance the applicability and impact of CA-CAE in real-world cybersecurity systems.</p>
</sec>
</body>
<back>
<ack>
<p>The authors would like to express sincere gratitude to Smart Systems Lab for their invaluable support and guidance throughout this research.</p>
</ack>
<sec>
<title>Funding Statement</title>
<p>This research was supported by Korea National University of Transportation Industry-Academy Cooperation Foundation in 2024.</p>
</sec>
<sec>
<title>Author Contributions</title>
<p>Conceptualization, Seulki Han; Methodology, Seulki Han; Validation, Seulki Han, Sangho Son and Won Sakong; Investigation, Seulki Han, Sangho Son and Won Sakong; Writing&#x2014;original draft preparation, Seulki Han, Sangho Son, Won Sakong and Haemin Jung; Writing&#x2014;review and editing, Seulki Han, Sangho Son, Won Sakong and Haemin Jung; Supervision, Haemin Jung; Funding acquisition, Haemin Jung. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>The datasets used and analyzed during the current study are available from the corresponding author on reasonable request.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>Not applicable.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Yaar</surname> <given-names>A</given-names></string-name>, <string-name><surname>Perrig</surname> <given-names>A</given-names></string-name>, <string-name><surname>Song</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Pi: a path identification mechanism to defend against DDoS attacks</article-title>. In: <conf-name>2003 Symposium on Security and Privacy; 2003 May 11&#x2013;14</conf-name>; <publisher-loc>Berkeley, CA, USA</publisher-loc>. p. <fpage>93</fpage>&#x2013;<lpage>107</lpage>. doi:<pub-id pub-id-type="doi">10.1109/SECPRI.2003.1199330</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hinton</surname> <given-names>GE</given-names></string-name>, <string-name><surname>Salakhutdinov</surname> <given-names>RR</given-names></string-name></person-group>. <article-title>Reducing the dimensionality of data with neural networks</article-title>. <source>Science</source>. <year>2006</year>;<volume>313</volume>(<issue>5786</issue>):<fpage>504</fpage>&#x2013;<lpage>7</lpage>. doi:<pub-id pub-id-type="doi">10.1126/science.1127647</pub-id>; <pub-id pub-id-type="pmid">16873662</pub-id></mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Vaiyapuri</surname> <given-names>T</given-names></string-name>, <string-name><surname>Binbusayyis</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Enhanced deep autoencoder based feature representation learning for intelligent intrusion detection system</article-title>. <source>Comput Mater Contin</source>. <year>2021</year>;<volume>68</volume>(<issue>3</issue>):<fpage>3271</fpage>&#x2013;<lpage>88</lpage>. doi:<pub-id pub-id-type="doi">10.32604/cmc.2021.017665</pub-id>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Charte</surname> <given-names>D</given-names></string-name>, <string-name><surname>Charte</surname> <given-names>F</given-names></string-name>, <string-name><surname>del Jesus</surname> <given-names>MJ</given-names></string-name>, <string-name><surname>Herrera</surname> <given-names>F</given-names></string-name></person-group>. <article-title>An analysis on the use of autoencoders for representation learning: fundamentals, learning task case studies, explainability and challenges</article-title>. <source>Neurocomputing</source>. <year>2020</year>;<volume>404</volume>:<fpage>93</fpage>&#x2013;<lpage>107</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.neucom.2020.04.057</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Sakurada</surname> <given-names>M</given-names></string-name>, <string-name><surname>Yairi</surname> <given-names>T</given-names></string-name></person-group>. <article-title>Anomaly detection using autoencoders with nonlinear dimensionality reduction</article-title>. In: <conf-name>Proceedings of the MLSDA 2014 2nd Workshop on Machine Learning for Sensory Data Analysis. 2014 Dec 2</conf-name>; <publisher-loc>Gold Coast Australia, QLD, Australia</publisher-loc>. p. <fpage>4</fpage>&#x2013;<lpage>11</lpage>. doi:<pub-id pub-id-type="doi">10.1145/2689746.2689747</pub-id>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Park</surname> <given-names>S</given-names></string-name>, <string-name><surname>Kim</surname> <given-names>M</given-names></string-name>, <string-name><surname>Lee</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Anomaly detection for HTTP using convolutional autoencoders</article-title>. <source>IEEE Access</source>. <year>2018</year>;<volume>6</volume>:<fpage>70884</fpage>&#x2013;<lpage>901</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ACCESS.2018.2881003</pub-id>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Cheng</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>S</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Zhu</surname> <given-names>E</given-names></string-name></person-group>. <article-title>Improved autoencoder for unsupervised anomaly detection</article-title>. <source>Int J Intell Syst</source>. <year>2021</year>;<volume>36</volume>(<issue>12</issue>):<fpage>7103</fpage>&#x2013;<lpage>25</lpage>. doi:<pub-id pub-id-type="doi">10.1002/int.22582</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Charte</surname> <given-names>D</given-names></string-name>, <string-name><surname>Charte</surname> <given-names>F</given-names></string-name>, <string-name><surname>Garc&#x00ED;a</surname> <given-names>S</given-names></string-name>, <string-name><surname>del Jesus</surname> <given-names>MJ</given-names></string-name>, <string-name><surname>Herrera</surname> <given-names>F</given-names></string-name></person-group>. <article-title>A practical tutorial on autoencoders for nonlinear feature fusion: taxonomy, models, software and guidelines</article-title>. <source>Inf Fusion</source>. <year>2018</year>;<volume>44</volume>:<fpage>78</fpage>&#x2013;<lpage>96</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.inffus.2017.12.007</pub-id>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Andresini</surname> <given-names>G</given-names></string-name>, <string-name><surname>Appice</surname> <given-names>A</given-names></string-name>, <string-name><surname>Di Mauro</surname> <given-names>N</given-names></string-name>, <string-name><surname>Loglisci</surname> <given-names>C</given-names></string-name>, <string-name><surname>Malerba</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Multi-channel deep feature learning for intrusion detection</article-title>. <source>IEEE Access</source>. <year>2020</year>;<volume>8</volume>:<fpage>53346</fpage>&#x2013;<lpage>59</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ACCESS.2020.2980937</pub-id>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Vincent</surname> <given-names>P</given-names></string-name>, <string-name><surname>Larochelle</surname> <given-names>H</given-names></string-name>, <string-name><surname>Lajoie</surname> <given-names>I</given-names></string-name>, <string-name><surname>Bengio</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Manzagol</surname> <given-names>A</given-names></string-name>, <string-name><surname>Bottou</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion</article-title>. <source>J Mach Learn Res</source>. <year>2010</year>;<volume>11</volume>(<issue>12</issue>):<fpage>3371</fpage>&#x2013;<lpage>408</lpage>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Majumdar</surname> <given-names>A</given-names></string-name>, <string-name><surname>Tripathi</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Asymmetric stacked autoencoder</article-title>. In: <conf-name>2017 International Joint Conference on Neural Networks (IJCNN); 2017 May 14&#x2013;19</conf-name>; <publisher-loc>Anchorage, AK, USA</publisher-loc>. p. <fpage>911</fpage>&#x2013;<lpage>8</lpage>. doi:<pub-id pub-id-type="doi">10.1109/IJCNN.2017.7965949</pub-id>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Lunardi</surname> <given-names>WT</given-names></string-name>, <string-name><surname>Lopez</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Giacalone</surname> <given-names>JP</given-names></string-name></person-group>. <article-title>ARCADE: adversarially regularized convolutional autoencoder for network anomaly detection</article-title>. <source>IEEE Trans Netw Serv Manage</source>. <year>2023</year>;<volume>20</volume>(<issue>2</issue>):<fpage>1305</fpage>&#x2013;<lpage>18</lpage>. doi:<pub-id pub-id-type="doi">10.1109/TNSM.2022.3229706</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kim</surname> <given-names>KH</given-names></string-name>, <string-name><surname>Shim</surname> <given-names>S</given-names></string-name>, <string-name><surname>Lim</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Jeon</surname> <given-names>J</given-names></string-name>, <string-name><surname>Choi</surname> <given-names>J</given-names></string-name>, <string-name><surname>Kim</surname> <given-names>B</given-names></string-name>, <etal>et al.</etal></person-group> <article-title>Rapp: novelty detection with reconstruction along projection pathway</article-title>. In: <conf-name>International Conference on Learning Representations; 2020 Apr 30</conf-name>; <publisher-loc>Addis Ababa, Ethiopia</publisher-loc>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Sana</surname> <given-names>L</given-names></string-name>, <string-name><surname>Nazir</surname> <given-names>MM</given-names></string-name>, <string-name><surname>Yang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Hussain</surname> <given-names>L</given-names></string-name>, <string-name><surname>Chen</surname> <given-names>YL</given-names></string-name>, <string-name><surname>Ku</surname> <given-names>CS</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Securing the IoT cyber environment: enhancing intrusion anomaly detection with vision transformers</article-title>. <source>IEEE Access</source>. <year>2024</year>;<volume>12</volume>:<fpage>82443</fpage>&#x2013;<lpage>68</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ACCESS.2024.3404778</pub-id>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="book"><person-group person-group-type="author"><string-name><surname>Hyndman</surname> <given-names>RJ</given-names></string-name>, <string-name><surname>Athanasopoulos</surname> <given-names>G</given-names></string-name></person-group>. <source>Forecasting: principles and practice</source>. <edition>3rd ed</edition>. <publisher-loc>Melbourne, VIC, Australia</publisher-loc>: <publisher-name>OTexts</publisher-name>; <year>2018</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Ostertagov&#x00E1;</surname> <given-names>E</given-names></string-name>, <string-name><surname>Ostertag</surname> <given-names>O</given-names></string-name></person-group>. <article-title>The simple exponential smoothing model</article-title>. In: <conf-name>The 4th International Conference on Modelling of Mechanical and Mechatronic Systems; 2011 Sep 20&#x2013;22</conf-name>. <publisher-loc>Herlany, Slovak Republic</publisher-loc>. p. <fpage>380</fpage>&#x2013;<lpage>84</lpage>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Wang</surname> <given-names>R</given-names></string-name>, <string-name><surname>Jia</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Ju</surname> <given-names>L</given-names></string-name></person-group>. <article-title>An entropy-based distributed DDoS detection mechanism in software-defined networking</article-title>. In: <conf-name>2015 IEEE Trustcom/BigDataSE/ISPA; 2015 Aug 20&#x2013;22</conf-name>. <publisher-loc>Helsinki, Finland</publisher-loc>. p. <fpage>310</fpage>&#x2013;<lpage>7</lpage>. doi:<pub-id pub-id-type="doi">10.1109/Trustcom.2015.389</pub-id>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Malialis</surname> <given-names>K</given-names></string-name>, <string-name><surname>Devlin</surname> <given-names>S</given-names></string-name>, <string-name><surname>Kudenko</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Distributed reinforcement learning for adaptive and robust network intrusion response</article-title>. <source>Connect Sci</source>. <year>2015</year>;<volume>27</volume>(<issue>3</issue>):<fpage>234</fpage>&#x2013;<lpage>52</lpage>. doi:<pub-id pub-id-type="doi">10.1080/09540091.2015.1031082</pub-id>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Zhou</surname> <given-names>B</given-names></string-name>, <string-name><surname>Li</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Guo</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Li</surname> <given-names>Z</given-names></string-name></person-group>. <article-title>Machine-learning-based online distributed denial-of-service attack detection using spark streaming</article-title>. In: <conf-name>2018 IEEE International Conference on Communications (ICC); 2018 May 20&#x2013;24</conf-name>; <publisher-loc>Kansas City, MO, USA</publisher-loc>. p. <fpage>1</fpage>&#x2013;<lpage>6</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ICC.2018.8422327</pub-id>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>AlSaleh</surname> <given-names>I</given-names></string-name>, <string-name><surname>Al-Samawi</surname> <given-names>A</given-names></string-name>, <string-name><surname>Nissirat</surname> <given-names>L</given-names></string-name></person-group>. <article-title>Novel machine learning approach for DDoS cloud detection: bayesian-based CNN and data fusion enhancements</article-title>. <source>Sensors</source>. <year>2024</year>;<volume>24</volume>(<issue>5</issue>):<fpage>1418</fpage>. doi:<pub-id pub-id-type="doi">10.3390/s24051418.doi:</pub-id>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Salahuddin</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Pourahmadi</surname> <given-names>V</given-names></string-name>, <string-name><surname>Alameddine</surname> <given-names>HA</given-names></string-name>, <string-name><surname>Bari</surname> <given-names>MF</given-names></string-name>, <string-name><surname>Boutaba</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Chronos: DDOS attack detection using time-based autoencoder</article-title>. <source>IEEE Trans Netw Serv Manag</source>. <year>2022</year>;<volume>19</volume>(<issue>1</issue>):<fpage>627</fpage>&#x2013;<lpage>41</lpage>. doi:<pub-id pub-id-type="doi">10.1109/TNSM.2021.3088326</pub-id>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Malhotra</surname> <given-names>P</given-names></string-name>, <string-name><surname>Vig</surname> <given-names>L</given-names></string-name>, <string-name><surname>Shroff</surname> <given-names>G</given-names></string-name>, <string-name><surname>Agarwal</surname> <given-names>P</given-names></string-name></person-group>. <article-title>Long short term memory networks for anomaly detection in time series</article-title>. In: <conf-name>ESANN 2015: European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning; 2015 Apr 22&#x2013;24</conf-name>. <publisher-loc>Bruges, Belgium</publisher-loc>. p. <fpage>89</fpage>&#x2013;<lpage>94</lpage>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Liu</surname> <given-names>FT</given-names></string-name>, <string-name><surname>Ting</surname> <given-names>KM</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>ZH</given-names></string-name></person-group>. <article-title>Isolation forest</article-title>. In: <conf-name>2008 Eighth IEEE International Conference on Data Mining; 2008 Dec 15&#x2013;19</conf-name>; <publisher-loc>Pisa, Italy</publisher-loc>. p. <fpage>413</fpage>&#x2013;<lpage>22</lpage>. doi:<pub-id pub-id-type="doi">10.1109/ICDM.2008.17</pub-id>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Park</surname> <given-names>D</given-names></string-name>, <string-name><surname>Hoshi</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Kemp</surname> <given-names>CC</given-names></string-name></person-group>. <article-title>A multimodal anomaly detector for robot-assisted feeding using an LSTM-based variational autoencoder</article-title>. <source>IEEE Robot Autom Lett</source>. <year>2018</year>;<volume>3</volume>(<issue>3</issue>):<fpage>1544</fpage>&#x2013;<lpage>51</lpage>. doi:<pub-id pub-id-type="doi">10.1109/LRA.2018.2801475</pub-id>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Yang</surname> <given-names>J</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yuan</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Xue</surname> <given-names>H</given-names></string-name>, <string-name><surname>Bourouis</surname> <given-names>S</given-names></string-name>, <string-name><surname>Abdel-Salam</surname> <given-names>M</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>LLM-AE-MP: web attack detection using a large language model with autoencoder and multilayer perceptron</article-title>. <source>Expert Syst Appl</source>. <year>2025</year>;<volume>274</volume>:<fpage>126982</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.eswa.2025.126982</pub-id>.</mixed-citation></ref>
</ref-list>
</back></article>