<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">61426</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2025.061426</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>TIDS: Tensor Based Intrusion Detection System (IDS) and Its Application in Large Scale DDoS Attack Detection</article-title>
<alt-title alt-title-type="left-running-head">TIDS: Tensor Based Intrusion Detection System (IDS) and Its Application in Large Scale DDoS Attack Detection</alt-title>
<alt-title alt-title-type="right-running-head">TIDS: Tensor Based Intrusion Detection System (IDS) and Its Application in Large Scale DDoS Attack Detection</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Sun</surname><given-names>Hanqing</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Li</surname><given-names>Xue</given-names></name><xref ref-type="aff" rid="aff-2">2</xref><email>lixue@hait.edu.cn</email></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Fan</surname><given-names>Qiyuan</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Wang</surname><given-names>Puming</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<aff id="aff-1"><label>1</label><institution>School of Information Engineering, Henan University of Animal Husbandry and Economy</institution>, <addr-line>Zhengzhou, 450046</addr-line>, <country>China</country></aff>
<aff id="aff-2"><label>2</label><institution>School of Electronic Information Engineering, Henan Institute of Technology</institution>, <addr-line>Xinxiang, 453002</addr-line>, <country>China</country></aff>
<aff id="aff-3"><label>3</label><institution>School of Software, Yunnan University</institution>, <addr-line>Kunming, 650500</addr-line>, <country>China</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Xue Li. Email: <email>lixue@hait.edu.cn</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2025</year>
</pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>09</day><month>06</month><year>2025</year>
</pub-date>
<volume>84</volume>
<issue>1</issue>
<fpage>1659</fpage>
<lpage>1679</lpage>
<history>
<date date-type="received">
<day>24</day>
<month>11</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>06</day>
<month>5</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2025 The Authors.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Published by Tech Science Press.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_61426.pdf"></self-uri>
<abstract>
<p>The era of big data brings new challenges for information network systems (INS), simultaneously offering unprecedented opportunities for advancing intelligent intrusion detection systems. In this work, we propose a data-driven intrusion detection system for Distributed Denial of Service (DDoS) attack detection. The system focuses on intrusion detection from a big data perceptive. As intelligent information processing methods, big data and artificial intelligence have been widely used in information systems. The INS system is an important information system in cyberspace. In advanced INS systems, the network architectures have become more complex. And the smart devices in INS systems collect a large scale of network data. How to improve the performance of a complex intrusion detection system with big data and artificial intelligence is a big challenge. To address the problem, we design a novel intrusion detection system (IDS) from a big data perspective. The IDS system uses tensors to represent large-scale and complex multi-source network data in a unified tensor. Then, a novel tensor decomposition (TD) method is developed to complete big data mining. The TD method seamlessly collaborates with the XGBoost (eXtreme Gradient Boosting) method to complete the intrusion detection. To verify the proposed IDS system, a series of experiments is conducted on two real network datasets. The results revealed that the proposed IDS system attained an impressive accuracy rate over 98%. Additionally, by altering the scale of the datasets, the proposed IDS system still maintains excellent detection performance, which demonstrates the proposed IDS system&#x2019;s robustness.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Intrusion detection system</kwd>
<kwd>big data</kwd>
<kwd>tensor decomposition</kwd>
<kwd>multi-modal feature</kwd>
<kwd>DDoS</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>National Nature Science Foundation of China</funding-source>
<award-id>62166047</award-id>
</award-group>
<award-group id="awg2">
<funding-source>Yunnan International Joint Laboratory of Natural Rubber Intelligent Monitor and Digital Applications</funding-source>
<award-id>202403AP140001</award-id>
</award-group>
<award-group id="awg3">
<funding-source>Xingdian Talent Support Program</funding-source>
<award-id>YNWR-QNBJ-2019-270</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>The rapid expansion of network infrastructure, coupled with the increasing complexity of network structures, has posed significant challenges in intrusion detection systems. On one hand, smart devices in intrusion detection systems (IDS) collect a large scale of data, the data have complex structures and are from diverse sources, which exhibit the characteristics of big data. On the other hand, the intrusion attack&#x2019;s technologies are becoming increasingly advanced, and the attack&#x2019;s methods are becoming more covert [<xref ref-type="bibr" rid="ref-1">1</xref>]. These attacks exploit network vulnerabilities to overwhelm systems, leading to congestion, paralysis, and potentially severe information leakage [<xref ref-type="bibr" rid="ref-2">2</xref>]. Therefore, it is urgent to design effective intrusion detection systems (IDS) to detect abnormal attacks in large-scale complex networks using big data and artificial intelligence technologies.</p>
<p>Traditional intrusion detection systems mainly rely on matrix-based techniques such as Principal Component Analysis (PCA) and Singular Value Decomposition (SVD), and traditional intrusion detection methods have shown effectiveness in simple contexts [<xref ref-type="bibr" rid="ref-3">3</xref>]. However, traditional intrusion detection methods are difficult to cope with the large-scale and multi-modal network data. Consequently, traditional intrusion detection systems often fall short of accurately identifying attacks in real-world big data scenarios. There are heterogeneous and huge volume data streams in real-world big data scenarios. This inadequacy underscores the pressing need for advanced intrusion detection systems that can not only enhance detection accuracy but also improve data quality through effective denoising techniques.</p>
<p>To address these challenges, this paper proposes a novel tensor-based intrusion detection system (IDS) with big data. By integrating state-of-the-art tensor decomposition techniques with advanced machine learning algorithms, the intrusion detection system aims to provide a more accurate and scalable solution for identifying Distributed Denial of Service (DDoS) attacks with big data. Tensor decomposition allows for keeping multidimensional relationships inherent in network data, facilitating a more nuanced understanding of traffic patterns [<xref ref-type="bibr" rid="ref-4">4</xref>]. This method enables the detection system to better distinguish between normal and malicious traffic, thereby enhancing the robustness and efficiency of DDoS attack detection systems. Through this approach, we seek to improve detection rates and contribute to resilient network security measures in an increasingly interconnected digital landscape. The proposed system brings several contributions to the area of network intrusion detection, including the ability to:
<list list-type="bullet">
<list-item>
<p>Proposing to use tensors to model large-scale heterogeneous network big data in intrusion detection systems. The method integrates features from different modalities in a unified format to obtain a more comprehensive representation.</p></list-item>
<list-item>
<p>Developing Tucker-2 Decomposition to propose HOBISVD method by employing Minimum Description Length Principle (MDLP) for feature extraction. We fuse two modalitie&#x2019;s features through tensor computation to obtain eigentensors. The eigentensors (factor matrices) from the tensor decomposition reveal significant interactions and anomalies within multi-modal network big data.</p></list-item>
<list-item>
<p>Proposing a novel intrusion detection system from a perspective of big data, which utilizes tensor algebra to model and analyze the multi-model network data for capturing intricate dependencies and patterns. The system represents large-scale network data in a unified tensor and then combines the HOBISVD and XGBoost classification to effectively detect intrusion.</p></list-item>
</list></p>
<p>The organization of the remaining parts of the paper is as follows. <xref ref-type="sec" rid="s2">Section 2</xref> introduces the related work of IDS detection. Preliminary is introduced in <xref ref-type="sec" rid="s3">Section 3</xref>. In <xref ref-type="sec" rid="s4">Section 4</xref>, a big data-driven INS system is proposed for DDoS detection. And we illustrate a case study to verify the proposed system in <xref ref-type="sec" rid="s5">Section 5</xref>. In <xref ref-type="sec" rid="s6">Section 6</xref>, we summarize the paper.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>We analyze the existing body of work related to IDS systems, providing a comprehensive review of the current state of research in the field, including big data-driven methods and tensor decomposition methods.</p>
<sec id="s2_1">
<label>2.1</label>
<title>Big Data Driven Method</title>
<p>Big data technology extracts knowledge and values from data by statistical analysis. Big data has three features: 1) objectivity, 2) accuracy, and 3) testability. Big data-based methods have become an important approach to network security. Statistical algorithms are typical big data-driven methods. The methods represent a well-established approach to anomaly detection. This type of anomaly detection identifies DDoS attacks by calculating thresholds to flag unusual behaviors. Analyzing daily traffic distributions can pinpoint activities that significantly deviate from the norm as potential anomalies. Hamdi and Boudriga [<xref ref-type="bibr" rid="ref-5">5</xref>] introduced a novel method for the statistical analysis of DDoS attacks utilizing wavelet analysis techniques. By transforming traffic data into the frequency domain through wavelet transformation, they were able to identify specific patterns and features associated with DDoS attacks. This method effectively detects and characterizes these malicious activities by extracting frequency-domain features and incorporating statistical analysis. Tao and Yu [<xref ref-type="bibr" rid="ref-6">6</xref>] proposed a DDoS attack detection method based on information entropy within local area network environments. Under typical conditions, the IP entropy tends to be relatively high. However, during a DDoS attack, the volume of packets directed at the target IP increases significantly, leading to a decrease in IP entropy. This reduction serves as an early warning sign of potential DDoS activity. Fortunati et al. [<xref ref-type="bibr" rid="ref-7">7</xref>] proposed an enhanced anomaly detection method based on covariance analysis. This approach involves constructing a covariance matrix from network traffic data to establish a normal distribution profile. By analyzing this covariance matrix, the method can identify deviations from the expected traffic patterns. Abnormal traffic is detected by setting specific thresholds, allowing for the effective identification of anomalies indicative of potential DDoS attacks or other malicious activities.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Tensor Decomposition and Intrusion Detection System</title>
<p>Tensors are multidimensional arrays. Tensor models with strong expressive ability can mine abundant intrinsic information contained in massive data, and it is a promising method to solve security problems in a big data environment. In terms of the multimodal data problem in large-scale networks, researchers have presented a considerable amount of work based on tensor models for anomaly detection. In [<xref ref-type="bibr" rid="ref-8">8</xref>], a novel approach is proposed for efficient tracking of intrusion in the normal subspace arising from the decomposition of the Parallel Factor Analysis tensor. The method is based on the extraction of a normal subspace obtained by the tensor decomposition technique, considering the correlation between different metrics. Aiming to address dynamic detection issues, some researchers have proposed a series of online detection methods. In [<xref ref-type="bibr" rid="ref-9">9</xref>], the network data is first represented as a unified tensor. Then, an incremental tensor decomposition is proposed for tensor data dimensionality reduction and denoising. In the end, by combining machine learning algorithms, intrusion detection is completed. The work [<xref ref-type="bibr" rid="ref-10">10</xref>] presents an online anomaly detection system capable of handling operational network traffic of large networks.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Preliminary</title>
<p>In this section, some preliminaries will be described. The preliminary mainly includes the mathematical theories and operations used in this paper. Tensors are an important tool used in the proposed intrusion detection system, and we will focus on discussing the theory and operations related to tensors.</p>
<p><monospace><bold>Define1: Eigentensor</bold></monospace> The eigentensor is the extension of the eigenvector, which is defined as follows. Given a tensor <inline-formula id="ieqn-1"><mml:math id="mml-ieqn-1"><mml:mi>A</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mi>M</mml:mi></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-2"><mml:math id="mml-ieqn-2"><mml:mi>X</mml:mi><mml:mo>=</mml:mo><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mn>11</mml:mn><mml:mo>&#x22EF;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mn>11</mml:mn><mml:mo>&#x22EF;</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x22EF;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">]</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, if a pair (<inline-formula id="ieqn-3"><mml:math id="mml-ieqn-3"><mml:mi>&#x03BB;</mml:mi></mml:math></inline-formula>-X) satisfies <inline-formula id="ieqn-4"><mml:math id="mml-ieqn-4"><mml:mi>A</mml:mi><mml:mrow><mml:msub><mml:mo>&#x2217;</mml:mo><mml:mi>N</mml:mi></mml:msub></mml:mrow><mml:mi>X</mml:mi><mml:mo>=</mml:mo><mml:mi>&#x03BB;</mml:mi><mml:mrow><mml:msub><mml:mo>&#x2217;</mml:mo><mml:mi>N</mml:mi></mml:msub></mml:mrow><mml:mi>X</mml:mi></mml:math></inline-formula>, we call <inline-formula id="ieqn-5"><mml:math id="mml-ieqn-5"><mml:mi>&#x03BB;</mml:mi></mml:math></inline-formula> an eigenvalue and <italic>X</italic> an eigentensor related to <inline-formula id="ieqn-6"><mml:math id="mml-ieqn-6"><mml:mi>&#x03BB;</mml:mi></mml:math></inline-formula>. Here <inline-formula id="ieqn-7"><mml:math id="mml-ieqn-7"><mml:mrow><mml:msub><mml:mo>&#x2217;</mml:mo><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> denotes the multimodal product.</p>
<p><monospace><bold>Define2: Matricization</bold></monospace> Matrixization involves unfolding the tensor <inline-formula id="ieqn-8"><mml:math id="mml-ieqn-8"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>I</mml:mi><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> along the <inline-formula id="ieqn-9"><mml:math id="mml-ieqn-9"><mml:mi>n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi></mml:math></inline-formula> and representing it as a matrix. Specially, the mapping transforms tensor elements <inline-formula id="ieqn-10"><mml:math id="mml-ieqn-10"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">N</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> into matrix elements <inline-formula id="ieqn-11"><mml:math id="mml-ieqn-11"><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mrow><mml:mi mathvariant="normal">j</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> as <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref>:
<disp-formula id="eqn-1"><label>(1)</label><mml:math id="mml-eqn-1" display="block"><mml:mi mathvariant="normal">j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mover><mml:mrow><mml:mi mathvariant="normal">k</mml:mi><mml:mo>&#x2260;</mml:mo><mml:mi mathvariant="normal">n</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mover></mml:mrow><mml:mi mathvariant="normal">N</mml:mi></mml:munderover><mml:mtext>&#xA0;</mml:mtext><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">k</mml:mi></mml:msub><mml:mtext>&#xA0;</mml:mtext><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mtext>&#xA0;</mml:mtext><mml:msub><mml:mi mathvariant="normal">J</mml:mi><mml:mi mathvariant="normal">k</mml:mi></mml:msub><mml:mstyle scriptlevel="0"><mml:mspace width="1em"></mml:mspace></mml:mstyle><mml:mi mathvariant="normal">w</mml:mi><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">t</mml:mi><mml:mi mathvariant="normal">h</mml:mi><mml:mstyle scriptlevel="0"><mml:mspace width="1em"></mml:mspace></mml:mstyle><mml:msub><mml:mi mathvariant="normal">J</mml:mi><mml:mi mathvariant="normal">k</mml:mi></mml:msub><mml:mtext>&#xA0;</mml:mtext><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x220F;</mml:mo><mml:mrow><mml:mover><mml:mrow><mml:mi mathvariant="normal">m</mml:mi><mml:mo>&#x2260;</mml:mo><mml:mi mathvariant="normal">n</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">m</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mover></mml:mrow><mml:mrow><mml:mi mathvariant="normal">k</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:munderover><mml:mtext>&#xA0;</mml:mtext><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">m</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p><monospace><bold>Define3: Singular Value Decomposition</bold></monospace> SVD is very significant in our study. Based on the literature, the SVD formula for matrix <inline-formula id="ieqn-12"><mml:math id="mml-ieqn-12"><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow></mml:math></inline-formula> is given as <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>:
<disp-formula id="eqn-2"><label>(2)</label><mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:mi mathvariant="normal">&#x03A3;</mml:mi><mml:msup><mml:mrow><mml:mtext mathvariant="bold">V</mml:mtext></mml:mrow><mml:mi>T</mml:mi></mml:msup><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>r</mml:mi></mml:munderover><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mrow><mml:mtext mathvariant="bold">u</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:msubsup><mml:mrow><mml:mtext mathvariant="bold">v</mml:mtext></mml:mrow><mml:mi>k</mml:mi><mml:mi>T</mml:mi></mml:msubsup><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>r</mml:mi></mml:munderover><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>k</mml:mi></mml:msub><mml:msub><mml:mrow><mml:mtext mathvariant="bold">u</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:mo>&#x2297;</mml:mo><mml:msub><mml:mrow><mml:mtext mathvariant="bold">v</mml:mtext></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p><inline-formula id="ieqn-13"><mml:math id="mml-ieqn-13"><mml:mo>&#x2297;</mml:mo></mml:math></inline-formula> represents the tensor product, which is defined as <inline-formula id="ieqn-14"><mml:math id="mml-ieqn-14"><mml:mi>x</mml:mi><mml:mo>&#x2297;</mml:mo><mml:mi>y</mml:mi><mml:mrow><mml:mover><mml:mrow><mml:mo>=</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi></mml:mrow></mml:mover></mml:mrow><mml:mi>x</mml:mi><mml:msup><mml:mi>y</mml:mi><mml:mi>T</mml:mi></mml:msup><mml:mo>.</mml:mo></mml:math></inline-formula> The rank of <inline-formula id="ieqn-15"><mml:math id="mml-ieqn-15"><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow></mml:math></inline-formula>, denoted as <inline-formula id="ieqn-16"><mml:math id="mml-ieqn-16"><mml:mi>r</mml:mi><mml:mo>&#x2264;</mml:mo><mml:mo movablelimits="true" form="prefix">min</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mi>m</mml:mi><mml:mo>,</mml:mo><mml:mi>n</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>, corresponds to the dimension of space spanned by the columns or rows of <inline-formula id="ieqn-17"><mml:math id="mml-ieqn-17"><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow></mml:math></inline-formula>. <inline-formula id="ieqn-18"><mml:math id="mml-ieqn-18"><mml:mi mathvariant="normal">&#x03A3;</mml:mi></mml:math></inline-formula> is a diagonal <inline-formula id="ieqn-19"><mml:math id="mml-ieqn-19"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>r</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> matrix that includes the nonzero singular values of <inline-formula id="ieqn-20"><mml:math id="mml-ieqn-20"><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow></mml:math></inline-formula>. The singular values are real, non-negative, and follow the convention where <inline-formula id="ieqn-21"><mml:math id="mml-ieqn-21"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo></mml:math></inline-formula><inline-formula id="ieqn-22"><mml:math id="mml-ieqn-22"><mml:mo>&#x22EF;</mml:mo></mml:math></inline-formula><inline-formula id="ieqn-23"><mml:math id="mml-ieqn-23"><mml:mo>&gt;</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo><mml:mn>0</mml:mn><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo></mml:math></inline-formula><inline-formula id="ieqn-24"><mml:math id="mml-ieqn-24"><mml:mo>&#x22EF;</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:math></inline-formula> The vectors <inline-formula id="ieqn-25"><mml:math id="mml-ieqn-25"><mml:msub><mml:mi>u</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-26"><mml:math id="mml-ieqn-26"><mml:msub><mml:mi>v</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> are the orthonormal columns of matrices <inline-formula id="ieqn-27"><mml:math id="mml-ieqn-27"><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>m</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>r</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and <inline-formula id="ieqn-28"><mml:math id="mml-ieqn-28"><mml:mrow><mml:mtext mathvariant="bold">V</mml:mtext></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>r</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula>. Specially, <inline-formula id="ieqn-29"><mml:math id="mml-ieqn-29"><mml:msub><mml:mi>v</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:math></inline-formula> are the eigenvectors of <inline-formula id="ieqn-30"><mml:math id="mml-ieqn-30"><mml:msup><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-31"><mml:math id="mml-ieqn-31"><mml:msub><mml:mtext>u</mml:mtext><mml:mi>k</mml:mi></mml:msub><mml:mo>=</mml:mo></mml:math></inline-formula>A<inline-formula id="ieqn-32"><mml:math id="mml-ieqn-32"><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mi>k</mml:mi></mml:msub><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:math></inline-formula> Both <inline-formula id="ieqn-33"><mml:math id="mml-ieqn-33"><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-34"><mml:math id="mml-ieqn-34"><mml:mrow><mml:mtext mathvariant="bold">V</mml:mtext></mml:mrow></mml:math></inline-formula> could be expanded with additional columns to form square and orthogonal matrices of dimensions <inline-formula id="ieqn-35"><mml:math id="mml-ieqn-35"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>m</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>m</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> and <inline-formula id="ieqn-36"><mml:math id="mml-ieqn-36"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>n</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>n</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> matrices.</p>
<p><monospace><bold>Define4: Tensor Norm</bold></monospace> The norm of the tensor <inline-formula id="ieqn-37"><mml:math id="mml-ieqn-37"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula> is defined as the square root of the sum of the squares of all its elements. Mathematically, this can be expressed as:
<disp-formula id="eqn-3"><label>(3)</label><mml:math id="mml-eqn-3" display="block"><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo fence="false" stretchy="false">&#x2016;</mml:mo><mml:mo>=</mml:mo><mml:msqrt><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:munderover><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:munderover><mml:mo>&#x22EF;</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mi>N</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mtext>I</mml:mtext></mml:mrow><mml:mi>N</mml:mi></mml:msub></mml:mrow></mml:munderover><mml:msubsup><mml:mrow><mml:mtext>x</mml:mtext></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mn>1</mml:mn></mml:msub><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x22EF;</mml:mo><mml:msub><mml:mrow><mml:mtext>i</mml:mtext></mml:mrow><mml:mi>N</mml:mi></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup></mml:msqrt><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p><monospace><bold>Define5: Tensor Multiplication</bold></monospace> The n-mode product of the tensor <inline-formula id="ieqn-38"><mml:math id="mml-ieqn-38"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:msub></mml:mrow></mml:mrow></mml:msup></mml:math></inline-formula> and the matrix <inline-formula id="ieqn-39"><mml:math id="mml-ieqn-39"><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">J</mml:mi></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="normal">I</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula> can be donated as <inline-formula id="ieqn-40"><mml:math id="mml-ieqn-40"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow></mml:math></inline-formula>, whose size is <inline-formula id="ieqn-41"><mml:math id="mml-ieqn-41"><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mrow><mml:mi mathvariant="normal">n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mi mathvariant="normal">J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mrow><mml:mi mathvariant="normal">n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula>. Therefore, we can obtain the calculating formula:
<disp-formula id="eqn-4"><label>(4)</label><mml:math id="mml-eqn-4" display="block"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mrow><mml:mtext>n</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:msub><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x22EF;</mml:mo><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mrow><mml:mi mathvariant="normal">n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mi mathvariant="normal">j</mml:mi><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mrow><mml:mi mathvariant="normal">n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x22EF;</mml:mo><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">n</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi mathvariant="normal">I</mml:mi><mml:mi mathvariant="normal">n</mml:mi></mml:msub></mml:mrow></mml:munderover><mml:msub><mml:mi mathvariant="normal">x</mml:mi><mml:mrow><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x22EF;</mml:mo><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mtext>&#xA0;</mml:mtext><mml:msub><mml:mi mathvariant="normal">u</mml:mi><mml:mrow><mml:mi mathvariant="normal">j</mml:mi><mml:msub><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">n</mml:mi></mml:msub></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>In addition, the formula can also transform matrices and tensors:
<disp-formula id="eqn-5"><label>(5)</label><mml:math id="mml-eqn-5" display="block"><mml:mrow><mml:mi>&#x1D4B4;</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mrow><mml:mtext>n</mml:mtext></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:mspace width="1em" /><mml:mo stretchy="false">&#x21D4;</mml:mo><mml:mspace width="1em" /><mml:msub><mml:mrow><mml:mtext mathvariant="bold">Y</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi mathvariant="normal">n</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mrow><mml:mtext mathvariant="bold">U</mml:mtext></mml:mrow><mml:msub><mml:mrow><mml:mtext mathvariant="bold">X</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi mathvariant="normal">n</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msub></mml:math></disp-formula></p>
</sec>
<sec id="s4">
<label>4</label>
<title>The Big Data-Driven Intrusion Detection System</title>
<p>The era of big data brings new challenges for network security, while also providing rich resources for IDS innovation. In this work, we propose a data-driven intrusion detection system for DDoS detection. The system specializes in big data-driven intrusion detection methodologies. <xref ref-type="fig" rid="fig-1">Fig. 1</xref> shows the big data-driven intrusion detection system, which consists of four layers, namely i) the data layer, ii) the big data processing layer, iii) the big data rule layer, and iv) the big data application layer. Below, we will describe each layer&#x2019;s function, respectively.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>The proposed big data-driven Intrusion detection system. This big data-based intrusion detection system adopts a data-centric security paradigm, specifically designed for large-scale network threat analytics. The Intrusion Detection System consists of four layers, namely i) the data layer, ii) the big data processing layer, iii) the rule layer, and iv) the big data application layer</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-1.tif"/>
</fig>
<sec id="s4_1">
<label>4.1</label>
<title>Big Data Sensing Layer</title>
<p>In the real world, there is a huge amount of multi-source and heterogeneous data in cyberspace, including network traffic data, network topological structure data, time-related information, and so on [<xref ref-type="bibr" rid="ref-11">11</xref>]. This is the basics of designing a big data-driven intrusion detection system. The data sensing plane is the collecting plane of the intrusion detection system. Some data stream collection tools (Sniffer, NetFlow, probe, and Flow tools) are embedded in the switches, which are a set of distributed monitors. The monitors capture the network data day and night. These data are source IP, port, destination IP, protocol type, data packet length, the number of bytes of traffic between two hosts in a fixed time, the number of data streams, flow entropy, etc. These data have various sources and formats, and the network traffic flow is related to multiple factors. For example, it is influenced by the time. From 9:00 PM to 11:00 PM, the network traffic volume exceeds typical baseline levels, but at 2:00 AM, it is low. Also, the network traffic is influenced by weather. Hence, a holistic integration of these multi-source datasets is essential for robust network security analysis. How to fuse these data and the relationships between them is a huge challenge [<xref ref-type="bibr" rid="ref-12">12</xref>]. To cope with the problem, we use the network unified tensor (NUT) to fuse the multi-source and heterogeneous data as <xref ref-type="fig" rid="fig-2">Fig. 2</xref>. Firstly, these data are represented as local tensors in the data <xref ref-type="fig" rid="fig-3"> </xref>collection layer. With the transmission plane, distributed local tensors are propagated to the big data processing plane, where they undergo tensor fusion operations to form a unified global tensor representation.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>Initially, the raw data is transformed into distributed tensor representations within the big data infrastructure layer. Subsequently, these localized tensors are propagated to the big data processing plane through the dedicated transmission plane, where they will be fused as a global large tensor</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-2.tif"/>
</fig><fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>The HOBISVD of 3th-order tensor <inline-formula id="ieqn-49"><mml:math id="mml-ieqn-49"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow></mml:math></inline-formula>. The tensor <inline-formula id="ieqn-50"><mml:math id="mml-ieqn-50"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow></mml:math></inline-formula> is decomposed into a core tensor <italic>S</italic> and U and V, which represent the features on the two modalities, respectively. The matrices of U and V are orthogonal, and their column vectors are the orthogonal basis of the space corresponding to each modality</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-3.tif"/>
</fig>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>The Proposed Method</title>
<p>From the viewpoint of big data, how to model these multi-source heterogeneous data is a big challenge [<xref ref-type="bibr" rid="ref-13">13</xref>]. The data models are always complex and need powerful computing methods [<xref ref-type="bibr" rid="ref-9">9</xref>]. Considering the tensor&#x2019;s multi-model features, in this work, we use the tensor to fuse the data in <xref ref-type="sec" rid="s4_2">Section 4.2</xref>. Aiming at the data model, a novel tensor decomposition method is proposed. Specifically, the tensor decomposition method effectively reduces noise in network data and performs feature cutting on each mode through the Minimum Description Length Principle (MDLP) rule. Different from previous tensor decomposition algorithms such as HOSVD(Higher-Order Singular Value Decomposition), this algorithm further integrates features at the feature level through tensor mode multiplication, seamlessly achieving feature dimension reduction and denoising.</p>
<p><bold><italic>1) Definition:</italic></bold> Inspired by the insights from [<xref ref-type="bibr" rid="ref-14">14</xref>], we developed a multi-modal feature extraction method (HOBISVD) shown in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>. HOBISVD is a variation of the Tucker-2 decomposition. It decomposes a tensor into a core tensor multiplied (or transformed) by a matrix along each mode except the first mode. The HOBISVD of a third-order tensor sets the first-factor matrices as the identity matrix. For instance, a HOBISVD can be described as <xref ref-type="disp-formula" rid="eqn-6">Eq. (6)</xref>:
<disp-formula id="eqn-6"><label>(6)</label><mml:math id="mml-eqn-6" display="block"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mi>B</mml:mi><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mspace width="-0.15em" /><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>:</mml:mo><mml:mi>I</mml:mi><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mspace width="-0.15em" /><mml:mo>]</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-42"><mml:math id="mml-ieqn-42"><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>Q</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> with <italic>R</italic> &#x003D; <italic>K</italic> and <inline-formula id="ieqn-43"><mml:math id="mml-ieqn-43"><mml:mrow><mml:mtext mathvariant="bold">C</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mtext mathvariant="bold">I</mml:mtext></mml:mrow></mml:math></inline-formula>, the <inline-formula id="ieqn-44"><mml:math id="mml-ieqn-44"><mml:mi>K</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:math></inline-formula> identity matrix. These concepts extend easily to the <italic>N</italic>-way case, we can set any subset of the factor matrices to the identity matrix.</p>
<p>Feature truncation is a method to optimize the obtained decomposition, thus obtaining the high-value principal components of each modality, and the result of the truncated decomposition is the approximate solution. In the work, the feature cutting on each mode is based on Minimum Description Length Principle (MDLP) rule. The truncated HOBISVD decomposition is a form of high-order PCA. Thus, in the three-way case where <inline-formula id="ieqn-45"><mml:math id="mml-ieqn-45"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, the <xref ref-type="disp-formula" rid="eqn-7">Eq. (7)</xref> is got,
<disp-formula id="eqn-7"><label>(7)</label><mml:math id="mml-eqn-7" display="block"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2248;</mml:mo><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mi>B</mml:mi><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mi>C</mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>Q</mml:mi></mml:munderover><mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>r</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>R</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>g</mml:mi><mml:mrow><mml:mo>:</mml:mo><mml:mo>,</mml:mo><mml:mi>q</mml:mi><mml:mi>r</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>b</mml:mi><mml:mi>q</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mi>r</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>here <inline-formula id="ieqn-46"><mml:math id="mml-ieqn-46"><mml:mrow><mml:mtext mathvariant="bold">B</mml:mtext></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>Q</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, and <inline-formula id="ieqn-47"><mml:math id="mml-ieqn-47"><mml:mrow><mml:mtext mathvariant="bold">C</mml:mtext></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>K</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula> are truncated factor matrices in each mode, and their column vectors are orthogonal, which are the principal components (PC) related various mode. The tensor <inline-formula id="ieqn-48"><mml:math id="mml-ieqn-48"><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>Q</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>R</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> is the core tensor.</p>
<p>Assuming that <inline-formula id="ieqn-51"><mml:math id="mml-ieqn-51"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, the tensor decomposition can be formulated as an optimization problem as <xref ref-type="disp-formula" rid="eqn-8">Eq. (8)</xref>:
<disp-formula id="eqn-8"><label>(8)</label><mml:math id="mml-eqn-8" display="block"><mml:munder><mml:mrow><mml:munder><mml:mo form="prefix">min</mml:mo><mml:mo>&#x23DF;</mml:mo></mml:munder></mml:mrow><mml:mrow><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:munder><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mspace width="-0.15em" /><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>;</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mspace width="-0.15em" /><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mtext>F</mml:mtext></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>,</mml:mo></mml:math></disp-formula>subject to
<disp-formula id="ueqn-9"><mml:math id="mml-ueqn-9" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd /><mml:mtd><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>and <inline-formula id="ieqn-52"><mml:math id="mml-ieqn-52"><mml:mi>B</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>J</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, <inline-formula id="ieqn-53"><mml:math id="mml-ieqn-53"><mml:mi>C</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>K</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> are columnwise orthogonal.</p>
<p>Next, HOBISVD is described in detail. HOBISVD involves three key stages, i) the tensor is unfolded in two modes I2 and I3; ii) perform orthogonality-constrained matrix factorization on the unfolded matrix to obtain the eigenvector for each independent modality; iii) by using tensor N-model multiplication, feature-level fusion and dimensionality reduction are performed on two modes to obtain a serious of feature tensors.</p>
<p><bold><italic>2) Unfolding of Tensor <inline-formula id="ieqn-54"><mml:math id="mml-ieqn-54"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow></mml:math></inline-formula>:</italic></bold> Unfolding a tensor <inline-formula id="ieqn-55"><mml:math id="mml-ieqn-55"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> in two modes <italic>J</italic> and <italic>K</italic>. Thus, two unfolded matrix <inline-formula id="ieqn-56"><mml:math id="mml-ieqn-56"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula>, <inline-formula id="ieqn-57"><mml:math id="mml-ieqn-57"><mml:msub><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula> can be got as follows:
<disp-formula id="eqn-9"><label>(9)</label><mml:math id="mml-eqn-9" display="block"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>J</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>K</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>I</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>
<disp-formula id="eqn-10"><label>(10)</label><mml:math id="mml-eqn-10" display="block"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>K</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>I</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>J</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p><xref ref-type="fig" rid="fig-4">Fig. 4</xref> illustrates the matrixization process of a 3th-order tensor <inline-formula id="ieqn-58"><mml:math id="mml-ieqn-58"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula> in two modes, in which <inline-formula id="ieqn-59"><mml:math id="mml-ieqn-59"><mml:msub><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msub><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula>, <inline-formula id="ieqn-60"><mml:math id="mml-ieqn-60"><mml:msub><mml:mrow><mml:mtext mathvariant="bold">X</mml:mtext></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msub><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>n</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>n</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:math></inline-formula> are the matrices of the two modes.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>The unfolded processing of the tensor <inline-formula id="ieqn-61"><mml:math id="mml-ieqn-61"><mml:mrow><mml:mtext mathvariant="bold">X</mml:mtext></mml:mrow></mml:math></inline-formula> in two modes</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-4.tif"/>
</fig>
<p><bold><italic>3) Applying Orthogonality-Constrained Matrix Factorization on the Unfolded Matrix <inline-formula id="ieqn-62"><mml:math id="mml-ieqn-62"><mml:msub><mml:mi>A</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-63"><mml:math id="mml-ieqn-63"><mml:msub><mml:mi>A</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula>:</italic></bold> To achieve the optimization target illustrated in the <xref ref-type="disp-formula" rid="eqn-11">Eq. (11)</xref> and ensure the resulting space remains orthogonal,
<disp-formula id="eqn-11"><label>(11)</label><mml:math id="mml-eqn-11" display="block"><mml:munder><mml:mrow><mml:munder><mml:mo form="prefix">min</mml:mo><mml:mo>&#x23DF;</mml:mo></mml:munder></mml:mrow><mml:mrow><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>,</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow></mml:munder><mml:msubsup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mspace width="-0.15em" /><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mrow><mml:mi>&#x1D4A2;</mml:mi></mml:mrow><mml:mo>;</mml:mo><mml:mi>B</mml:mi><mml:mo>,</mml:mo><mml:mi>C</mml:mi></mml:mrow><mml:mo>]</mml:mo></mml:mrow><mml:mspace width="-0.15em" /><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mtext>F</mml:mtext></mml:mrow></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>,</mml:mo></mml:math></disp-formula>various methods have been devised by researchers. This article employs the singular value decomposition approach for this purpose. The singular value decomposition enables to calculation of the eigenvalues associated with each mode, which in turn allows for the determination of the optimal truncation strategy for each mode based on these eigenvalues, which will be described in the following. We apply orthogonality-constrained matrix factorization on the two unfolded tensors <inline-formula id="ieqn-64"><mml:math id="mml-ieqn-64"><mml:msub><mml:mi>A</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula>. Three new matrix <inline-formula id="ieqn-65"><mml:math id="mml-ieqn-65"><mml:mrow><mml:msub><mml:mi>U</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x03A3;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>V</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> are got as <xref ref-type="disp-formula" rid="eqn-12">Eq. (12)</xref>:
<disp-formula id="eqn-12"><label>(12)</label><mml:math id="mml-eqn-12" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mtd><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:msub><mml:mi>U</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x03A3;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>V</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mi>j</mml:mi><mml:mi>J</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mi>j</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo></mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>J</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p>Similarly, the same method is applied to matrix <inline-formula id="ieqn-66"><mml:math id="mml-ieqn-66"><mml:msub><mml:mi>A</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula>, resulting in three matrices <inline-formula id="ieqn-67"><mml:math id="mml-ieqn-67"><mml:mrow><mml:msub><mml:mi>U</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x03A3;</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>V</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> as <xref ref-type="disp-formula" rid="eqn-13">Eq. (13)</xref>:
<disp-formula id="eqn-13"><label>(13)</label><mml:math id="mml-eqn-13" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mtd><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:msub><mml:mi>U</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi mathvariant="normal">&#x03A3;</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:msub><mml:mi>V</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mi>k</mml:mi><mml:mi>K</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mi>k</mml:mi><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>+</mml:mo></mml:mrow><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>J</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo>&#x2218;</mml:mo><mml:mrow><mml:msub><mml:mi>v</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula></p>
<p><xref ref-type="fig" rid="fig-5">Fig. 5</xref> presents an example of orthogonal matrix decomposition applied to an unfolded matrix derived from a 3rd-order tensor. In the context of tensor dimensionality reduction, several parameters need to be specified. Among these, <inline-formula id="ieqn-68"><mml:math id="mml-ieqn-68"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-69"><mml:math id="mml-ieqn-69"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> are particularly important. These parameters represent the number of leading singular vectors that are retained in the tensor bases <inline-formula id="ieqn-70"><mml:math id="mml-ieqn-70"><mml:msub><mml:mi>U</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-71"><mml:math id="mml-ieqn-71"><mml:msub><mml:mi>U</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula>. The values of <inline-formula id="ieqn-72"><mml:math id="mml-ieqn-72"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-73"><mml:math id="mml-ieqn-73"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> directly influence the final dimensionality of the eigentensor <italic>S</italic>, which is a crucial component in the tensor decomposition process. The selection of the eigenvector parameters <inline-formula id="ieqn-74"><mml:math id="mml-ieqn-74"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> and <inline-formula id="ieqn-75"><mml:math id="mml-ieqn-75"><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:math></inline-formula> is based on retaining a specified portion of the original data from tensors <inline-formula id="ieqn-76"><mml:math id="mml-ieqn-76"><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-77"><mml:math id="mml-ieqn-77"><mml:msub><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula>. In the following, the Minimum Description Length Principle (MDLP) based truncated method will be described [<xref ref-type="bibr" rid="ref-15">15</xref>].</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>The process of ALS-based matrix decomposition for the unfolded matrix</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-5.tif"/>
</fig>
<p><bold><italic>4) MDLP Based Truncated Method:</italic></bold> Enhancing data quality without significantly compromising its inherent characteristics is essential for effective noise reduction. We opt for the Minimum Description Length Principle (MDLP) as our model order selection (MOS) strategy, which aids in ascertaining the necessary rank for Singular Value Decomposition (SVD) based noise elimination. The MDLP is encapsulated by the following principle:
<disp-formula id="eqn-14"><label>(14)</label><mml:math id="mml-eqn-14" display="block"><mml:mtable columnalign="right left right left right left right left right left right left" rowspacing="3pt" columnspacing="0em 2em 0em 2em 0em 2em 0em 2em 0em 2em 0em" displaystyle="true"><mml:mtr><mml:mtd><mml:mi>M</mml:mi><mml:mi>D</mml:mi><mml:mi>L</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mtd><mml:mtd><mml:mi></mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x220F;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:msubsup><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>p</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mi>p</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:mfrac><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mi>k</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msubsup><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>p</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mi>N</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr><mml:mtr><mml:mtd /><mml:mtd><mml:mi></mml:mi><mml:mo>+</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:mfrac><mml:mi>k</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mi>p</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>g</mml:mi><mml:mi>N</mml:mi><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>here <inline-formula id="ieqn-78"><mml:math id="mml-ieqn-78"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> are eigenvalues which are from SVD and <inline-formula id="ieqn-79"><mml:math id="mml-ieqn-79"><mml:mi>i</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>p</mml:mi><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>. The eigenvalues are convergent and sorted in descending order, namely <inline-formula id="ieqn-80"><mml:math id="mml-ieqn-80"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&gt;</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo></mml:math></inline-formula><inline-formula id="ieqn-81"><mml:math id="mml-ieqn-81"><mml:mo>&#x22EF;</mml:mo><mml:mo>&gt;</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:math></inline-formula>. <inline-formula id="ieqn-82"><mml:math id="mml-ieqn-82"><mml:mi>p</mml:mi></mml:math></inline-formula> denotes the number of eigenvalues, and <italic>N</italic> denotes the sample&#x2019;s number in the dataset. When <italic>MDL(k)</italic> gets the minimum value as <xref ref-type="disp-formula" rid="eqn-14">Eq. (14)</xref>, the best rank of the matrix A can be ascertained by <xref ref-type="disp-formula" rid="eqn-15">Eq. (15)</xref>:
<disp-formula id="eqn-15"><label>(15)</label><mml:math id="mml-eqn-15" display="block"><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mi>g</mml:mi><mml:mi>m</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>D</mml:mi><mml:mi>L</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>k</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo></mml:math></disp-formula>where <inline-formula id="ieqn-83"><mml:math id="mml-ieqn-83"><mml:mi>k</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mo fence="false" stretchy="false">{</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>p</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo fence="false" stretchy="false">}</mml:mo></mml:math></inline-formula>. Indeed, as the variable <inline-formula id="ieqn-84"><mml:math id="mml-ieqn-84"><mml:mi>k</mml:mi></mml:math></inline-formula> commences at 0, the aforementioned equation must incorporate an additional 1. Subsequently, only the initial <inline-formula id="ieqn-85"><mml:math id="mml-ieqn-85"><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> eigenvalue is preserved, while all remaining eigenvalues are assigned a value of 0, denoted as <inline-formula id="ieqn-86"><mml:math id="mml-ieqn-86"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>&gt;</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>&gt;</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula id="ieqn-87"><mml:math id="mml-ieqn-87"><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext mathvariant="bold">A</mml:mtext></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x03C3;</mml:mi><mml:mrow><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:math></inline-formula>.</p>
<p>We construct a canonical sinusoidal signal <inline-formula id="ieqn-88"><mml:math id="mml-ieqn-88"><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>x</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:math></inline-formula> as a baseline waveform and superimpose zero-mean Gaussian white noise onto the pristine signal. We then apply singular value decomposition (SVD) for noise reduction. <xref ref-type="fig" rid="fig-6">Fig. 6</xref> illustrates the progression of singular values as a rank function. Numerical analysis demonstrates that when the matrix rank satisfies <inline-formula id="ieqn-89"><mml:math id="mml-ieqn-89"><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo>&#x2265;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>, the singular values are diminutive when <inline-formula id="ieqn-90"><mml:math id="mml-ieqn-90"><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo>&#x2265;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula>. The data encompassed within the eigenvectors associated with these singular values is noise-related. Hence, we maintain the initial three singular values.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>The progression of singular values as a function of rank</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-6.tif"/>
</fig>
<p><bold><italic>5) Tensor Multiplication Based Feature-Level Fusion and Dimensionality Reduction:</italic></bold> According to the above step, we have obtained the two reduced <inline-formula id="ieqn-91"><mml:math id="mml-ieqn-91"><mml:msub><mml:mi>U</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula>, and <inline-formula id="ieqn-92"><mml:math id="mml-ieqn-92"><mml:msub><mml:mi>U</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula> in <inline-formula id="ieqn-93"><mml:math id="mml-ieqn-93"><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> mode and <inline-formula id="ieqn-94"><mml:math id="mml-ieqn-94"><mml:mrow><mml:msub><mml:mi>I</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> mode, which are directly isolated from each other. How to integrate the information of the two modalities is a big challenge. To address this issue, the tensor N-mode multiplication between tensors and matrices is proposed to solve the problem, as shown in <xref ref-type="disp-formula" rid="eqn-16">Eq. (16)</xref>. First, the characteristic information on modality 2 is integrated, and then the feature information on the integrated modality is fused. Since a principal component truncation operation is used, dimensionality reduction and denoising are also achieved in the process.
<disp-formula id="eqn-16"><label>(16)</label><mml:math id="mml-eqn-16" display="block"><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>U</mml:mi><mml:mn>2</mml:mn><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>T</mml:mi></mml:msup></mml:mrow><mml:mrow><mml:msub><mml:mo>&#x00D7;</mml:mo><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>U</mml:mi><mml:mn>3</mml:mn><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>T</mml:mi></mml:msup></mml:mrow><mml:mo>,</mml:mo></mml:math></disp-formula>where <italic>X</italic> is the initial tensor, <inline-formula id="ieqn-95"><mml:math id="mml-ieqn-95"><mml:mrow><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>U</mml:mi><mml:mn>2</mml:mn><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>T</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> is the transpose of the <inline-formula id="ieqn-96"><mml:math id="mml-ieqn-96"><mml:msub><mml:mi>r</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula>-dimensionally reduced <inline-formula id="ieqn-97"><mml:math id="mml-ieqn-97"><mml:msub><mml:mi>U</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula> tensor, <inline-formula id="ieqn-98"><mml:math id="mml-ieqn-98"><mml:mrow><mml:msup><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msubsup><mml:mi>U</mml:mi><mml:mn>3</mml:mn><mml:mrow><mml:mrow><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mi>T</mml:mi></mml:msup></mml:mrow></mml:math></inline-formula> is the transpose of the <inline-formula id="ieqn-99"><mml:math id="mml-ieqn-99"><mml:msub><mml:mi>r</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula>-dimensionally reduced <inline-formula id="ieqn-100"><mml:math id="mml-ieqn-100"><mml:msub><mml:mi>U</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:math></inline-formula> tensor. The Algorithm 1 shows the process of HOBISVD. In nature, HOBISVD is a mathematical method used for tensor decomposition in high-order dimensions. It can decompose a high-dimensional data tensor into the product of multiple low-rank matrices, thereby extracting important features from the data. Essentially, the method can also be applied to deep learning to reduce dimensionality and extract useful information from input data. By performing HOBISVD decomposition on input features, we can obtain feature subspaces at different levels and directions, each containing strongly correlated features to some extent in the original data.</p>
<fig id="fig-12">
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-12.tif"/>
</fig>
<p><bold><italic>6) XGBoost Classification Method:</italic></bold> Aiming to deal with large-scale data, XGBoost (eXtreme Gradient Boosting) was proposed to improve the stability and accuracy of the large-scale model. It is an ensemble learning algorithm based on gradient-boosted trees (GBT). The key to XGBoost&#x2019;s success lies in its adaptability across diverse scenarios. XGBoost is a powerful and versatile machine-learning system known for its tree-boosting capabilities. Its influence has been acknowledged in various machine learning and data mining competitions. With tensor mode multiplication, feature-level fusion and dimensionality reduction are performed on two modes to obtain a series of feature tensors. In nature, it is a large-scale feature dataset. Aiming at the feature dataset. Aiming at the series of feature tensors, the XGBoost method is used for DDoS detection through classification. The input is the serious of feature tensors <inline-formula id="ieqn-120"><mml:math id="mml-ieqn-120"><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>S</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mo>,</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mo>,</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mo>&#x22EF;</mml:mo><mml:mo>,</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>I</mml:mi><mml:mo>,</mml:mo><mml:mo>,</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:math></inline-formula>, the output is <xref ref-type="disp-formula" rid="eqn-17">Eq. (17)</xref>:
<disp-formula id="eqn-17"><label>(17)</label><mml:math id="mml-eqn-17" display="block"><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mover><mml:mi>Y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:mrow><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>S</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mo>,</mml:mo><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:mi>F</mml:mi><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>The object is <xref ref-type="disp-formula" rid="eqn-18">Eq. (18)</xref>:
<disp-formula id="eqn-18"><label>(18)</label><mml:math id="mml-eqn-18" display="block"><mml:mi>O</mml:mi><mml:mi>b</mml:mi><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>I</mml:mi></mml:munderover><mml:mrow><mml:mi>l</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mover><mml:mi>Y</mml:mi><mml:mo stretchy="false">&#x005E;</mml:mo></mml:mover></mml:mrow></mml:mrow><mml:mi>i</mml:mi></mml:msub></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mi>i</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>+</mml:mo><mml:munderover><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>K</mml:mi></mml:munderover><mml:mrow><mml:mi mathvariant="normal">&#x03A9;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mi>k</mml:mi></mml:msub></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>.</mml:mo></mml:math></disp-formula></p>
<p>The primary objective of XGBoost is to push the boundaries of machine learning limitations to offer scalable, portable, and precise solutions. When dealing with huge network data, the distributed version of XGBoost demonstrates exceptional portability, effectively addressing the challenges associated with large-scale datasets [<xref ref-type="bibr" rid="ref-16">16</xref>].</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>The Case Study</title>
<p>As depicted in <xref ref-type="fig" rid="fig-7">Fig. 7</xref>, a case study is conducted to validate the proposed system. In our experiment, we used an SDN environment to simulate the experimental setup, where the OpenFlow switch collected data, which was then sent to the SDN(Software Defined Network) controller, and fused into a unified tensor model. Through HOBISVD decomposition, multimodal features were extracted, clustering was completed using XGBoost, and DDoS attack detection was achieved through clustering.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>The case study. Firstly, the intrusion detection system collects data from the flow tables and constructs a unified tensor. Through big data processing, the tensor rule is mined and then provides intrusion detection services</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-7.tif"/>
</fig>
<p>This section is structured into three distinct subsections. <xref ref-type="sec" rid="s5_1">Section 5.1</xref> provides an overview of the experimental dataset used. <xref ref-type="sec" rid="s5_2">Section 5.2</xref> focuses on detailing the pertinent evaluation metrics. Finally, <xref ref-type="sec" rid="s5_3">Section 5.3</xref> presents the results of the comparative experiment.</p>
<sec id="s5_1">
<label>5.1</label>
<title>Experimental Dataset</title>
<p>Dataset CICDDOS 2019, curated by the Canadian Institute for Cyber Security (CIC), includes a vast array of network data with 87 features and millions of traffic instances, encompassing various types of DDoS attacks [<xref ref-type="bibr" rid="ref-17">17</xref>]. In our experimental setup, we constructed a representative subset by randomly sampling 40,000 data points from the original dataset, with a stratified distribution of 32,000 normal traffic instances and 8000 malicious attack samples, as specified in <xref ref-type="table" rid="table-1">Table 1</xref>. Additionally, we eliminated several features that were deemed insignificant for the classification process from the initial set of 87. Consequently, we narrowed it down to 64 key features, as outlined in <xref ref-type="table" rid="table-2">Table 2</xref>. The approximate distribution of the used data is shown in <xref ref-type="fig" rid="fig-8">Fig. 8</xref>.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>CICDDoS2019 subset construction</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr align="center">
<th>Order</th>
<th>Traffic type</th>
<th>Total</th>
<th>Order</th>
<th>Traffic type</th>
<th>Total</th>
</tr>
</thead>
<tbody>
<tr align="center">
<td>(1)</td>
<td>DNS</td>
<td>800</td>
<td>(7)</td>
<td>NTP</td>
<td>800</td>
</tr>
<tr align="center">
<td>(2)</td>
<td>LDAP</td>
<td>800</td>
<td>(8)</td>
<td>SSDP</td>
<td>800</td>
</tr>
<tr align="center">
<td>(3)</td>
<td>NETBIOS</td>
<td>800</td>
<td>(9)</td>
<td>UDP-Lag</td>
<td>800</td>
</tr>
<tr align="center">
<td>(4)</td>
<td>SNMP</td>
<td>800</td>
<td>(10)</td>
<td>SYN</td>
<td>800</td>
</tr>
<tr align="center">
<td>(5)</td>
<td>UDP</td>
<td>800</td>
<td>(11)</td>
<td>BENIGN</td>
<td>32,000</td>
</tr>
<tr align="center">
<td>(6)</td>
<td>MSSQL</td>
<td>800</td>
<td></td>
<td></td>
<td></td>
</tr>
</tbody>
</table>
</table-wrap><table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>CICDDoS2019 network data set features used</title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr align="center">
<th>Order</th>
<th>Feature</th>
<th>Order</th>
<th>Feature</th>
</tr>
</thead>
<tbody>
<tr align="center">
<td>(1)</td>
<td>Source-Port</td>
<td>(33)</td>
<td>Packet-Length-Min</td>
</tr>
<tr align="center">
<td>(2)</td>
<td>Destination-Port</td>
<td>(34)</td>
<td>Packet-Length-Max</td>
</tr>
<tr align="center">
<td>(3)</td>
<td>Flow-Duration</td>
<td>(35)</td>
<td>Packet-Length-Avg</td>
</tr>
<tr align="center">
<td>(4)</td>
<td>Total-Fwd-Packet</td>
<td>(36)</td>
<td>Packet-Length-Std-Dev</td>
</tr>
<tr align="center">
<td>(5)</td>
<td>Total-Bwd-Packet</td>
<td>(37)</td>
<td>Packet-Length-Var</td>
</tr>
<tr align="center">
<td>(6)</td>
<td>Total-Length-Fwd-Packet</td>
<td>(38)</td>
<td>FIN-Flag-Count</td>
</tr>
<tr align="center">
<td>(7)</td>
<td>Total-Length-Bwd-Packet</td>
<td>(39)</td>
<td>SYN-Flag-Count</td>
</tr>
<tr align="center">
<td>(8)</td>
<td>Fwd-Packet-Length-Max</td>
<td>(40)</td>
<td>RST-Flag-Count</td>
</tr>
<tr align="center">
<td>(9)</td>
<td>Fwd-Packet-Length-Min</td>
<td>(41)</td>
<td>PUSH-Flag-Count</td>
</tr>
<tr align="center">
<td>(10)</td>
<td>Fwd-Packet-Length-Avg</td>
<td>(42)</td>
<td>ACK-Flag-Count</td>
</tr>
<tr align="center">
<td>(11)</td>
<td>Fwd-Packet-Length-Std-Dev</td>
<td>(43)</td>
<td>URG-Flag-Count</td>
</tr>
<tr align="center">
<td>(12)</td>
<td>Bwd-Packet-Length-Max</td>
<td>(44)</td>
<td>CWE-Flag-Count</td>
</tr>
<tr align="center">
<td>(13)</td>
<td>Bwd-Packet-Length-Min</td>
<td>(45)</td>
<td>ECE-Flag-Count</td>
</tr>
<tr align="center">
<td>(14)</td>
<td>Bwd-Packet-Length-Avg</td>
<td>(46)</td>
<td>Download/Upload-Ratio</td>
</tr>
<tr align="center">
<td>(15)</td>
<td>Bwd-Packet-Length-Std-Dev</td>
<td>(47)</td>
<td>Avg-Packet-Size</td>
</tr>
<tr align="center">
<td>(16)</td>
<td>Flow-Bytes/s</td>
<td>(48)</td>
<td>Avg-Fwd-Segment-Size</td>
</tr>
<tr align="center">
<td>(17)</td>
<td>Flow-Packets/s</td>
<td>(49)</td>
<td>Avg-Bwd-Segment-Size</td>
</tr>
<tr align="center">
<td>(18)</td>
<td>Flow-IAT-Avg</td>
<td>(50)</td>
<td>Subflow-Fwd-Packets</td>
</tr>
<tr align="center">
<td>(19)</td>
<td>Flow-IAT-Max</td>
<td>(51)</td>
<td>Subflow-Fwd-Bytes</td>
</tr>
<tr align="center">
<td>(20)</td>
<td>Flow-IAT-Min</td>
<td>(52)</td>
<td>Subflow-Bwd-Packets</td>
</tr>
<tr align="center">
<td>(21)</td>
<td>Fwd-IAT-Total</td>
<td>(53)</td>
<td>Subflow-Bwd-Bytes</td>
</tr>
<tr align="center">
<td>(22)</td>
<td>Fwd-IAT-Avg</td>
<td>(54)</td>
<td>Fwd-Win-Bytes</td>
</tr>
<tr align="center">
<td>(23)</td>
<td>Fwd-IAT-Max</td>
<td>(55)</td>
<td>Bwd-Win-Bytes</td>
</tr>
<tr align="center">
<td>(24)</td>
<td>Fwd-IAT-Min</td>
<td>(56)</td>
<td>Fwd-Active-Data-Packet</td>
</tr>
<tr align="center">
<td>(25)</td>
<td>Bwd-IAT-Total</td>
<td>(57)</td>
<td>Fwd-Min-Segment-Size</td>
</tr>
<tr align="center">
<td>(26)</td>
<td>Bwd-IAT-Avg</td>
<td>(58)</td>
<td>Avg-Time-Active-Flow</td>
</tr>
<tr align="center">
<td>(27)</td>
<td>Bwd-IAT-Max</td>
<td>(59)</td>
<td>Std-Dev-Time-Active-Flow</td>
</tr>
<tr align="center">
<td>(28)</td>
<td>Bwd-IAT-Min</td>
<td>(60)</td>
<td>Max-Time-Active-Flow</td>
</tr>
<tr align="center">
<td>(29)</td>
<td>Fwd-Header-Length</td>
<td>(61)</td>
<td>Min-Time-Active-Flow</td>
</tr>
<tr align="center">
<td>(30)</td>
<td>Bwd-Header-Length</td>
<td>(62)</td>
<td>Avg-Time-Idle-Flow</td>
</tr>
<tr align="center">
<td>(31)</td>
<td>Fwd-Packet/s</td>
<td>(63)</td>
<td>Std-Dev-Time-Idle-Flow</td>
</tr>
<tr align="center">
<td>(32)</td>
<td>Bwd-Packet/s</td>
<td>(64)</td>
<td>Min-Time-Idle-Flow</td>
</tr>
</tbody>
</table>
</table-wrap><fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Data distribution of CIC-DDoS2019</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-8.tif"/>
</fig>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Evaluation Metrics</title>
<p>According to the confusion matrix, we use TP, FP, TN, and FN to represent True Positive, False Positive, True Negative, and False Negative.</p>
<sec id="s5_2_1">
<label>5.2.1</label>
<title>Accuracy (Acc)</title>
<p>Accuracy is a metric that quantifies the classifier&#x2019;s ability to correctly predict the entire sample, indicating the degree to which the true values align with the predicted values. A higher Accuracy value signifies better classification performance, as it suggests that a larger proportion of the predictions are accurate. Accuracy is defined in <xref ref-type="disp-formula" rid="eqn-19">Eq. (19)</xref>:
<disp-formula id="eqn-19"><label>(19)</label><mml:math id="mml-eqn-19" display="block"><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mi>c</mml:mi><mml:mi>u</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula></p>
</sec>
<sec id="s5_2_2">
<label>5.2.2</label>
<title>Precision (Pre)</title>
<p>Precision is a metric that assesses the classifier&#x2019;s ability to correctly predict the accuracy of positive samples. It measures the proportion of positive predictions that are positive, indicating how many of the predicted positive samples are true positives. A higher Precision value indicates better classification performance in terms of identifying positive instances correctly. It is defined in <xref ref-type="disp-formula" rid="eqn-20">Eq. (20)</xref>:
<disp-formula id="eqn-20"><label>(20)</label><mml:math id="mml-eqn-20" display="block"><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula></p>
</sec>
<sec id="s5_2_3">
<label>5.2.3</label>
<title>Recall (Rec)</title>
<p>Recall can measure the proportion of actual positive samples that are correctly identified by the classifier. A higher Recall value indicates better classification performance in terms of capturing positive instances within the dataset, and is is defined in <xref ref-type="disp-formula" rid="eqn-21">Eq. (21)</xref>:
<disp-formula id="eqn-21"><label>(21)</label><mml:math id="mml-eqn-21" display="block"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula></p>
</sec>
<sec id="s5_2_4">
<label>5.2.4</label>
<title>F1-Score</title>
<p>The F1-score is a pivotal metric that encapsulates the performance of a model by harmonizing the critical aspects of Recall and Precision. It is particularly adept at providing a balanced evaluation, especially in scenarios where the dataset may be unevenly distributed. The F1-score is calculated using a formula that employs the harmonic mean, which effectively weighs both Precision and Recall, ensuring that neither metric can overshadow the other without affecting the overall score. Fi-Score is defined in <xref ref-type="disp-formula" rid="eqn-22">Eq. (22)</xref>:
<disp-formula id="eqn-22"><label>(22)</label><mml:math id="mml-eqn-22" display="block"><mml:mrow><mml:mtext mathvariant="italic">F1 - Score</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x22C5;</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:mfrac><mml:mo>.</mml:mo></mml:math></disp-formula></p>
</sec>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Contrast Experiments</title>
<p>In this subsection, a series of comparative experiments will be carried out. Focusing on the CICDDoS2019 dataset detailed in <xref ref-type="table" rid="table-1">Table 1</xref>, which comprises 64 features, the data matrix is denoted as <inline-formula id="ieqn-121"><mml:math id="mml-ieqn-121"><mml:mrow><mml:mtext mathvariant="bold">X</mml:mtext></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>N</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, where M represents the total number of data points and <inline-formula id="ieqn-122"><mml:math id="mml-ieqn-122"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:mn>64</mml:mn></mml:math></inline-formula> signifies the number of features. Within the context of the HOBISVD-based denoising algorithm, the matrix <inline-formula id="ieqn-123"><mml:math id="mml-ieqn-123"><mml:mrow><mml:mtext mathvariant="bold">X</mml:mtext></mml:mrow></mml:math></inline-formula> is transformed into a three-way tensor <inline-formula id="ieqn-124"><mml:math id="mml-ieqn-124"><mml:mrow><mml:mi>&#x1D4B3;</mml:mi></mml:mrow><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="double-struck">R</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>, with <inline-formula id="ieqn-125"><mml:math id="mml-ieqn-125"><mml:msub><mml:mi>N</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn>8</mml:mn></mml:math></inline-formula> and <inline-formula id="ieqn-126"><mml:math id="mml-ieqn-126"><mml:mi>N</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x00D7;</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:math></inline-formula>.</p>

<sec id="s5_3_1">
<label>5.3.1</label>
<title>Different Denoising Algorithm</title>
<p>In <xref ref-type="table" rid="table-2">Table 2</xref>, various datasets of different sizes are randomly chosen to assess the impact of various denoising techniques on the classification capabilities of XGBoost. The results of these experiments are depicted in <xref ref-type="fig" rid="fig-9">Fig. 9</xref> and <xref ref-type="table" rid="table-3">Table 3</xref>. The findings indicate that denoising enhances detection accuracy compared to RAW (nondenoised) data, and SVD, HOSVD, and HOOI (Higher Order Orthogonal Iteration of Tensors) denoising algorithms yield lower classification efficiency and performance than HOBISVD.</p>
<fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>The effect of different denoising algorithms is compared under the truncated size <inline-formula id="ieqn-127"><mml:math id="mml-ieqn-127"><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-9.tif"/>
</fig><table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Comparison of different denoising algorithms with the truncation sizes <inline-formula id="ieqn-128"><mml:math id="mml-ieqn-128"><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula><break/></title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr align="center">
<th>Model</th>
<th>Accuracy (%)</th>
<th>Precision (%)</th>
<th>Recall (%)</th>
<th>F1-Score (%)</th>
</tr>
</thead>
<tbody>
<tr align="center">
<td><bold>HOBISVD (Proposed detection system)</bold></td>
<td><bold>98.18</bold></td>
<td><bold>96.65</bold></td>
<td><bold>96.85</bold></td>
<td><bold>97.10</bold></td>
</tr>
<tr align="center">
<td>RAW</td>
<td>89.30</td>
<td>85.33</td>
<td>86.81</td>
<td>88.37</td>
</tr>
<tr align="center">
<td>SVD [<xref ref-type="bibr" rid="ref-18">18</xref>]</td>
<td>96.16</td>
<td>94.26</td>
<td>94.55</td>
<td>95.82</td>
</tr>
<tr align="center">
<td>HOSVD [<xref ref-type="bibr" rid="ref-19">19</xref>]</td>
<td>97.17</td>
<td>95.92</td>
<td>96.21</td>
<td>96.57</td>
</tr>
<tr align="center">
<td>HOOI [<xref ref-type="bibr" rid="ref-19">19</xref>]</td>
<td>96.62</td>
<td>95.01</td>
<td>95.87</td>
<td>96.19</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Furthermore, to validate the efficacy of HOBISVD, experiments are carried out with truncation sizes ranging from <inline-formula id="ieqn-129"><mml:math id="mml-ieqn-129"><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> to <inline-formula id="ieqn-130"><mml:math id="mml-ieqn-130"><mml:mn>3</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula>. Results in <xref ref-type="fig" rid="fig-10">Fig. 10</xref> and <xref ref-type="table" rid="table-4">Table 4</xref> show that the truncation sizes <inline-formula id="ieqn-131"><mml:math id="mml-ieqn-131"><mml:mn>3</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula> are worse than the truncation sizes <inline-formula id="ieqn-132"><mml:math id="mml-ieqn-132"><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> in both the performance of numerical results and the algorithm&#x2019;s stability. This is because, after high-order decomposition and reconstruction, the truncation sizes <inline-formula id="ieqn-133"><mml:math id="mml-ieqn-133"><mml:mn>2</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> condense nearly all the required properties, which makes it an optimal choice without any redundancy. This also proves the correctness of the selection of <inline-formula id="ieqn-134"><mml:math id="mml-ieqn-134"><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>2</mml:mn></mml:math></inline-formula> in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>.</p>
<fig id="fig-10">
<label>Figure 10</label>
<caption>
<title>The effect of different denoising algorithms is compared under the truncated size <inline-formula id="ieqn-135"><mml:math id="mml-ieqn-135"><mml:mn>3</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula></title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-10.tif"/>
</fig><table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Comparison of different denoising algorithms with the truncation sizes <inline-formula id="ieqn-136"><mml:math id="mml-ieqn-136"><mml:mn>3</mml:mn><mml:mo>&#x2217;</mml:mo><mml:mn>3</mml:mn></mml:math></inline-formula><break/></title>
</caption>
<table>
<colgroup>
<col/>
<col/>
<col/>
<col/>
<col/>
</colgroup>
<thead>
<tr align="center">
<th>Model</th>
<th>Accuracy (%)</th>
<th>Precision (%)</th>
<th>Recall (%)</th>
<th>F1-Score (%)</th>
</tr>
</thead>
<tbody>
<tr align="center">
<td><bold>HOBISVD (Proposed detection system)</bold></td>
<td><bold>95.52</bold></td>
<td><bold>94.38</bold></td>
<td><bold>94.15</bold></td>
<td><bold>93.96</bold></td>
</tr>
<tr align="center">
<td>RAW</td>
<td>85.66</td>
<td>88.89</td>
<td>80.12</td>
<td>83.33</td>
</tr>
<tr align="center">
<td>SVD [<xref ref-type="bibr" rid="ref-18">18</xref>]</td>
<td>91.34</td>
<td>90.26</td>
<td>91.25</td>
<td>91.24</td>
</tr>
<tr align="center">
<td>HOSVD [<xref ref-type="bibr" rid="ref-20">20</xref>]</td>
<td>93.78</td>
<td><bold>94.38</bold></td>
<td>93.27</td>
<td>93.82</td>
</tr>
<tr align="center">
<td>HOOI [<xref ref-type="bibr" rid="ref-20">20</xref>]</td>
<td>93.81</td>
<td>93.33</td>
<td><bold>94.15</bold></td>
<td>93.72</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s5_3_2">
<label>5.3.2</label>
<title>Different Classification Algorithm</title>
<p>A classification algorithm is another important part of the intrusion detection system. Therefore, we set the fixed denoising algorithm as the proposed HOBISVD, and evaluate the performance of different classification algorithms by varying the size of the datasets. The compared algorithms include Linear Discriminant Analysis (LDA), Logistic Regression (LR), Random Forest (RF), and Support Vector Machine (SVM), and XGBoost. The results of these comparisons are illustrated in <xref ref-type="fig" rid="fig-11">Fig. 11</xref>, leading to the following conclusions:</p>
<p><list list-type="bullet">
<list-item>
<p>As the size of the data set increases, the classification gets better because more and more features are available for classification;</p></list-item>
<list-item>
<p>The application of HOBISVD for denoising datasets leads to superior detection outcomes across various classification algorithms;</p></list-item>
<list-item>
<p>XGBoost demonstrates both rapid processing speed and effective detection capabilities across datasets of varying sizes;</p></list-item>
<list-item>
<p>The proposed intrusion detection system exhibits significant robustness across datasets of different sizes, consistently delivering high classification performance.</p></list-item>
</list></p>
<fig id="fig-11">
<label>Figure 11</label>
<caption>
<title>Classification performance of different classification algorithms</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_61426-fig-11.tif"/>
</fig>
</sec>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Summary</title>
<p>The exponential growth of internet users and the proliferation of smart devices have led to an exponential increase in the volume of data, forming massive data collections. The sources and formats of network data are complex and varied. This poses significant challenges in data modeling and analysis for cyber attack detection. To address the problem, this article uses a unified tensor to construct a DDoS attack detection model. Aiming at the model, a novel data analysis method is proposed for reducing the dimensionality and denoising multi-modal data through tensor decomposition. Then we seamlessly integrate the XGBoost classification algorithm to solve the DDoS attack detection problem.</p>
<p>Future research will focus on advancing tensor decomposition techniques, such as tensor train decomposition, to better capture the intricate relationships within network big data. Integrating tensors with advanced machine learning algorithms such as deep learning or ensemble methods could lead to more robust and accurate network attack detection systems. Real-time tensor processing algorithms are also crucial; the co-development of (1) real-time tensor processing frameworks, (2) parallel-optimized tensor operators, and (3) streaming-enabled tensor architectures emerges as a critical triad in modern intrusion detection systems.</p>
</sec>
</body>
<back>
<ack>
<p>The authors acknowledge Jing Xu and Sizhang Li for the helpful discussions for this paper.</p>
</ack>
<sec>
<title>Funding Statement</title>
<p>This work was supported in part by the National Nature Science Foundation of China under Project 62166047; in part by the Yunnan International Joint Laboratory of Natural Rubber Intelligent Monitor and Digital Applications under Grant 202403AP140001; in part by the Xingdian Talent Support Program under Grant YNWR-QNBJ-2019-270.</p>
</sec>
<sec>
<title>Author Contributions</title>
<p>The authors confirm contribution to the paper as follows: Study conception and design: Hanqing Sun, Xue Li, Puming Wang; Data collection: Hanqing Sun, Xue Li, Qiyuan Fan; Analysis and interpretation of results: Hanqing Sun, Xue Li; Draft manuscript preparation: Hanqing Sun, Xue Li. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability">
<title>Availability of Data and Materials</title>
<p>Not applicable.</p>
</sec>
<sec>
<title>Ethics Approval</title>
<p>Not applicable.</p>
</sec>
<sec sec-type="COI-statement">
<title>Conflicts of Interest</title>
<p>The authors declare no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Srivastava</surname> <given-names>A</given-names></string-name>, <string-name><surname>Sinha</surname> <given-names>D</given-names></string-name></person-group>. <article-title>FP-growth-based signature extraction and unknown variants of DoS/DDoS attack detection on real-time data stream</article-title>. <source>J Inf Secur Appl</source>. <year>2025</year>;<volume>89</volume>:<fpage>103996</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jisa.2025.103996</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Habib</surname> <given-names>AA</given-names></string-name>, <string-name><surname>Imtiaz</surname> <given-names>A</given-names></string-name>, <string-name><surname>Tripura</surname> <given-names>D</given-names></string-name>, <string-name><surname>Faruk</surname> <given-names>MO</given-names></string-name>, <string-name><surname>Hossain</surname> <given-names>MA</given-names></string-name>, <string-name><surname>Ara</surname> <given-names>I</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Distributed denial-of-service attack detection short review: issues, challenges, and recommendations</article-title>. <source>Bulletin Electr Eng Inform</source>. <year>2025</year>;<volume>14</volume>(<issue>1</issue>):<fpage>438</fpage>&#x2013;<lpage>46</lpage>. doi:<pub-id pub-id-type="doi">10.11591/eei.v14i1.8377</pub-id>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>X</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhang</surname> <given-names>B</given-names></string-name>, <string-name><surname>Tang</surname> <given-names>X</given-names></string-name>, <string-name><surname>Sun</surname> <given-names>M</given-names></string-name></person-group>. <article-title>An adaptive DDoS detection and classification method in blockchain using an integrated multi-models</article-title>. <source>Comput Mater Contin</source>. <year>2023</year>;<volume>77</volume>(<issue>3</issue>):<fpage>3265</fpage>&#x2013;<lpage>88</lpage>. doi:<pub-id pub-id-type="doi">10.32604/cmc.2023.045588</pub-id>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Alabdulatif</surname> <given-names>A</given-names></string-name>, <string-name><surname>Thilakarathne</surname> <given-names>NN</given-names></string-name>, <string-name><surname>Aashiq</surname> <given-names>M</given-names></string-name></person-group>. <article-title>Machine learning enabled novel real-time iot targeted DoS/DDoS cyber attack detection system</article-title>. <source>Comput Mater Contin</source>. <year>2024</year>;<volume>80</volume>(<issue>3</issue>):<fpage>3655</fpage>&#x2013;<lpage>83</lpage>. doi:<pub-id pub-id-type="doi">10.32604/cmc.2024.054610</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hamdi</surname> <given-names>M</given-names></string-name>, <string-name><surname>Boudriga</surname> <given-names>N</given-names></string-name></person-group>. <article-title>Detecting Denial-of-Service attacks using the wavelet transform</article-title>. <source>Comput Commun</source>. <year>2007</year>;<volume>30</volume>(<issue>16</issue>):<fpage>3203</fpage>&#x2013;<lpage>13</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.comcom.2007.05.061</pub-id>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Tao</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Yu</surname> <given-names>S</given-names></string-name></person-group>. <article-title>DDoS attack detection at local area networks using information theoretical metrics</article-title>. In: <conf-name>12th IEEE International Conference on Trust, Security and Privacy in Computing and Communications</conf-name>; <year>2013</year>; <publisher-loc>Melbourne, Australia</publisher-loc>. p. <fpage>233</fpage>&#x2013;<lpage>40</lpage>.</mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fortunati</surname> <given-names>S</given-names></string-name>, <string-name><surname>Gini</surname> <given-names>F</given-names></string-name>, <string-name><surname>Greco</surname> <given-names>MS</given-names></string-name>, <string-name><surname>Farina</surname> <given-names>A</given-names></string-name>, <string-name><surname>Graziano</surname> <given-names>A</given-names></string-name>, <string-name><surname>Giompapa</surname> <given-names>S</given-names></string-name></person-group>. <article-title>An improvement of the state-of-the-art covariance-based methods for statistical anomaly detection algorithms</article-title>. <source>Signal Image Video Process</source>. <year>2016</year>;<volume>10</volume>:<fpage>687</fpage>&#x2013;<lpage>94</lpage>. doi:<pub-id pub-id-type="doi">10.1007/s11760-015-0796-y</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Streit</surname> <given-names>A</given-names></string-name>, <string-name><surname>Santos</surname> <given-names>GH</given-names></string-name>, <string-name><surname>Le&#x00E3;o</surname> <given-names>RM</given-names></string-name>, <string-name><surname>e Silva</surname> <given-names>EdS</given-names></string-name>, <string-name><surname>Menasch&#x00E9;</surname> <given-names>D</given-names></string-name>, <string-name><surname>Towsley</surname> <given-names>D</given-names></string-name></person-group>. <article-title>Network anomaly detection based on tensor decomposition</article-title>. <source>Comput Netw</source>. <year>2021</year>;<volume>200</volume>:<fpage>108503</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.comnet.2021.108503</pub-id>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fan</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Li</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yao</surname> <given-names>S</given-names></string-name>, <string-name><surname>Miao</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>IDAD: an improved tensor train based distributed DDoS attack detection framework and its application in complex networks</article-title>. <source>Future Gener Comput Syst</source>. <year>2025</year>;<volume>162</volume>:<fpage>107471</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.future.2024.07.049</pub-id>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Shajari</surname> <given-names>M</given-names></string-name>, <string-name><surname>Geng</surname> <given-names>H</given-names></string-name>, <string-name><surname>Hu</surname> <given-names>K</given-names></string-name>, <string-name><surname>Leon-Garcia</surname> <given-names>A</given-names></string-name></person-group>. <article-title>Tensor-based online network anomaly detection and diagnosis</article-title>. <source>IEEE Access</source>. <year>2022</year>;<volume>10</volume>:<fpage>85792</fpage>&#x2013;<lpage>817</lpage>. doi:<pub-id pub-id-type="doi">10.1109/access.2022.3197651</pub-id>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Ye</surname> <given-names>Z</given-names></string-name>, <string-name><surname>Luo</surname> <given-names>J</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>W</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>M</given-names></string-name>, <string-name><surname>He</surname> <given-names>Q</given-names></string-name></person-group>. <article-title>An ensemble framework with improved hybrid breeding optimization-based feature selection for intrusion detection</article-title>. <source>Future Gener Comput Syst</source>. <year>2024</year>;<volume>151</volume>:<fpage>124</fpage>&#x2013;<lpage>36</lpage>. doi:<pub-id pub-id-type="doi">10.1016/j.future.2023.09.035</pub-id>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Hu</surname> <given-names>X</given-names></string-name>, <string-name><surname>Gao</surname> <given-names>W</given-names></string-name>, <string-name><surname>Cheng</surname> <given-names>G</given-names></string-name>, <string-name><surname>Li</surname> <given-names>R</given-names></string-name>, <string-name><surname>Zhou</surname> <given-names>Y</given-names></string-name>, <string-name><surname>Wu</surname> <given-names>H</given-names></string-name></person-group>. <article-title>Toward early and accurate network intrusion detection using graph embedding</article-title>. <source>IEEE Trans Inform Forensics Secur</source>. <year>2023</year>;<volume>18</volume>:<fpage>5817</fpage>&#x2013;<lpage>31</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tifs.2023.3318960</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Kumar</surname> <given-names>GA</given-names></string-name>, <string-name><surname>Katiyar</surname> <given-names>A</given-names></string-name>, <string-name><surname>Srinivasan</surname> <given-names>K</given-names></string-name></person-group>. <article-title>Elevating IDS capabilities: the convergence of SVM, Deep learning, and RFECV in network security</article-title>. In: <conf-name>2024 Second International Conference on Emerging Trends in Information Technology and Engineering (ICETITE)</conf-name>. <publisher-loc>Vellore, India</publisher-loc>: <publisher-name>IEEE</publisher-name>; 2024. p. <fpage>1</fpage>&#x2013;<lpage>16</lpage>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Kolda</surname> <given-names>TG</given-names></string-name>, <string-name><surname>Bader</surname> <given-names>BW</given-names></string-name></person-group>. <article-title>Tensor decompositions and applications</article-title>. <source>SIAM Review</source>. <year>2009</year>;<volume>51</volume>(<issue>3</issue>):<fpage>455</fpage>&#x2013;<lpage>500</lpage>. doi:<pub-id pub-id-type="doi">10.1137/07070111x</pub-id>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Fan</surname> <given-names>Q</given-names></string-name>, <string-name><surname>Li</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yao</surname> <given-names>S</given-names></string-name>, <string-name><surname>Miao</surname> <given-names>S</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>BDIP: an efficient big data-driven information processing framework and its application in DDoS attack detection</article-title>. <source>IEEE Trans Netw Serv Manag</source>. <year>2025</year>;<volume>22</volume>(<issue>1</issue>):<fpage>284</fpage>&#x2013;<lpage>98</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tnsm.2024.3464729</pub-id>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Chen</surname> <given-names>T</given-names></string-name>, <string-name><surname>Guestrin</surname> <given-names>C</given-names></string-name></person-group>. <article-title>XGBoost: a scalable tree boosting system</article-title>. In: <conf-name>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>; <year>2016</year>; <publisher-loc>San Francisco, CA, USA</publisher-loc>. p. <fpage>785</fpage>&#x2013;<lpage>94</lpage>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><surname>Subrmanian</surname> <given-names>M</given-names></string-name>, <string-name><surname>Shanmugavadivel</surname> <given-names>K</given-names></string-name>, <string-name><surname>Nandhini</surname> <given-names>P</given-names></string-name>, <string-name><surname>Sowmya</surname> <given-names>R</given-names></string-name></person-group>. <article-title>Evaluating the performance of LSTM and GRU in detection of distributed denial of service attacks using CICDDoS2019 dataset</article-title>. In: <conf-name>Proceedings of 7th International Conference on Harmony Search, Soft Computing and Applications: ICHSA 2022</conf-name>. <publisher-loc>Seoul, Republic of Korea</publisher-loc>: <publisher-name>Springer</publisher-name>; <year>2022</year>. p. <fpage>395</fpage>&#x2013;<lpage>406</lpage>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Maranh&#x00E3;o</surname> <given-names>JPA</given-names></string-name>, <string-name><surname>da Costa</surname> <given-names>JPC</given-names></string-name>, <string-name><surname>Javidi</surname> <given-names>E</given-names></string-name>, <string-name><surname>de Andrade</surname> <given-names>CAB</given-names></string-name>, <string-name><surname>de Sousa</surname> <given-names>Jr RT</given-names></string-name></person-group>. <article-title>Tensor based framework for distributed denial of Service attack detection</article-title>. <source>J Netw Comput Appl</source>. <year>2021</year>;<volume>174</volume>:<fpage>102894</fpage>. doi:<pub-id pub-id-type="doi">10.1016/j.jnca.2020.102894</pub-id>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Xu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Li</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>X</given-names></string-name>, <string-name><surname>Yao</surname> <given-names>S</given-names></string-name></person-group>. <article-title>Multi-modal noise-robust DDoS attack detection architecture in large-scale networks based on tensor SVD</article-title>. <source>IEEE Trans Netw Sci Eng</source>. <year>2022</year>;<volume>10</volume>(<issue>1</issue>):<fpage>152</fpage>&#x2013;<lpage>65</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tnse.2022.3205708</pub-id>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><surname>Li</surname> <given-names>S</given-names></string-name>, <string-name><surname>Xu</surname> <given-names>J</given-names></string-name>, <string-name><surname>Liu</surname> <given-names>P</given-names></string-name>, <string-name><surname>Li</surname> <given-names>X</given-names></string-name>, <string-name><surname>Wang</surname> <given-names>P</given-names></string-name>, <string-name><surname>Jin</surname> <given-names>X</given-names></string-name>, <etal>et al</etal></person-group>. <article-title>Truncated lanczos-TSVD: an effective dimensionality reduction algorithm for detecting DDoS attacks in large-scale networks</article-title>. <source>IEEE Trans Netw Sci Eng</source>. <year>2024</year>;<volume>11</volume>(<issue>5</issue>):<fpage>4689</fpage>&#x2013;<lpage>703</lpage>. doi:<pub-id pub-id-type="doi">10.1109/tnse.2024.3368048</pub-id>.</mixed-citation></ref>
</ref-list>
</back></article>