<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CSSE</journal-id>
<journal-id journal-id-type="nlm-ta">CSSE</journal-id>
<journal-id journal-id-type="publisher-id">CSSE</journal-id>
<journal-title-group>
<journal-title>Computer Systems Science &#x0026; Engineering</journal-title>
</journal-title-group>
<issn pub-type="ppub">0267-6192</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">39550</article-id>
<article-id pub-id-type="doi">10.32604/csse.2023.039550</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A Model Training Method for DDoS Detection Using CTGAN under 5GC Traffic</article-title>
<alt-title alt-title-type="left-running-head">A Model Training Method for DDoS Detection Using CTGAN under 5GC Traffic</alt-title>
<alt-title alt-title-type="right-running-head">A Model Training Method for DDoS Detection Using CTGAN under 5GC Traffic</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author">
<name name-style="western"><surname>Kim</surname><given-names>Yea-Sul</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Kim</surname><given-names>Ye-Eun</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-3" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Kim</surname><given-names>Hwankuk</given-names></name><xref ref-type="aff" rid="aff-2">2</xref><email>rinyfeel@smu.ac.kr</email></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Electronics Information and System Engineering, Sangmyung University</institution>, <addr-line>Cheonan, 31066, Korea</addr-line></aff>
<aff id="aff-2"><label>2</label><institution>Department of Information Security Engineering, Sangmyung University</institution>, <addr-line>Cheonan, 31066, Korea</addr-line></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Hwankuk Kim. Email: <email>rinyfeel@smu.ac.kr</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic"><year>2023</year></pub-date>
<pub-date date-type="pub" publication-format="electronic"><day>26</day><month>5</month><year>2023</year></pub-date>
<volume>47</volume>
<issue>1</issue>
<fpage>1125</fpage>
<lpage>1147</lpage>
<history>
<date date-type="received"><day>04</day><month>2</month><year>2023</year>
</date>
<date date-type="accepted"><day>10</day><month>4</month><year>2023</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2023 Kim, Kim and Kim</copyright-statement>
<copyright-year>2023</copyright-year>
<copyright-holder>Kim, Kim and Kim</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CSSE_39550.pdf"></self-uri>
<abstract>
<p>With the commercialization of 5th-generation mobile communications (5G) networks, a large-scale internet of things (IoT) environment is being built. Security is becoming increasingly crucial in 5G network environments due to the growing risk of various distributed denial of service (DDoS) attacks across vast IoT devices. Recently, research on automated intrusion detection using machine learning (ML) for 5G environments has been actively conducted. However, 5G traffic has insufficient data due to privacy protection problems and imbalance problems with significantly fewer attack data. If this data is used to train an ML model, it will likely suffer from generalization errors due to not training enough different features on the attack data. Therefore, this paper aims to study a training method to mitigate the generalization error problem of the ML model that classifies IoT DDoS attacks even under conditions of insufficient and imbalanced 5G traffic. We built a 5G testbed to construct a 5G dataset for training to solve the problem of insufficient data. To solve the imbalance problem, synthetic minority oversampling technique (SMOTE) and generative adversarial network (GAN)-based conditional tabular GAN (CTGAN) of data augmentation were used. The performance of the trained ML models was compared and meaningfully analyzed regarding the generalization error problem. The experimental results showed that CTGAN decreased the accuracy and f1-score compared to the Baseline. Still, regarding the generalization error, the difference between the validation and test results was reduced by at least 1.7 and up to 22.88 times, indicating an improvement in the problem. This result suggests that the ML model training method that utilizes CTGANs to augment attack data for training data in the 5G environment mitigates the generalization error problem.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>5G core traffic</kwd>
<kwd>machine learning</kwd>
<kwd>SMOTE</kwd>
<kwd>GAN-CTGAN</kwd>
<kwd>IoT DDoS detection</kwd>
<kwd>tabular form</kwd>
<kwd>cyber security</kwd>
<kwd>B5G</kwd>
<kwd>mobile network security</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>Institute of Information &#x0026; communications Technology Planning &#x0026; Evaluation (IITP)</funding-source>
<award-id>2021-0-00796</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>The main characteristics of 5G networks are hyper-connectivity, ultra-high speed, and ultra-low latency. This study focuses on the hyper-connectivity of 5G networks because it is the basis for building a large-scale IoT environment [<xref ref-type="bibr" rid="ref-1">1</xref>]. In addition, as of 2021, 5G networks had over 580 million 5G subscribers worldwide. The traffic on 5G networks generated by these large IoT environments and the increasing number of users is approximately 26,331 MB per person. With this amount of traffic, IoT devices have serious security vulnerabilities, and DDoS attacks targeting these devices are increasing. Therefore, research on automated intrusion detection technology, in which ML is applied to respond to IoT DDoS attacks in the 5G network environment, is being actively conducted.</p>
<p>The 3rd generation partnership project (3GPP), a 5G standardization organization, defined the Network Analytics Function (NWDAF) to provide network data analysis functions by utilizing artificial intelligence (AI) in the network core [<xref ref-type="bibr" rid="ref-2">2</xref>]. This approach serves to store and utilize the trained ML model and share the analysis results in the core network. This study aims to internalize ML models trained for IoT DDoS attack detection in 5G core (5GC) as a reference model for NWDAF. To this end, this study is concerned with 5GC traffic (among several other types of traffic) when considering the 5G environment. However, studying 5GC traffic is problematic in that insufficient datasets are publicly available for training ML models because of privacy concerns. Therefore, open-source software was used to directly construct a testbed, which is a virtual environment, for the collection of 5GC traffic by conducting simulation experiments using a dataset of a wired environment. In addition, feature extraction was performed based on protocol headers to detect DDoS attacks by ML models from the collected 5GC traffic. Therefore, the experimental data were processed and used in the experiment as 5GC traffic in a tabular form with both continuous and discrete data. Additionally, as the type of GAN used to augment the data, CTGAN, which can more accurately model features of discrete data such as those in tabular form, was selected and utilized [<xref ref-type="bibr" rid="ref-3">3</xref>,<xref ref-type="bibr" rid="ref-4">4</xref>].</p>
<p>Because of the network traffic characteristics, significantly fewer attack data are generated than normal. This characteristic means that the dataset of the wired network, which was used to train the ML model, also has a data imbalance problem. A model trained using a dataset with these characteristics would increase the possibility of misclassification when real data that differ from the training data are entered. In this case, a generalization error problem would occur [<xref ref-type="bibr" rid="ref-5">5</xref>]. However, attack data are becoming more complex and sophisticated over time. To this end, training methods with fewer generalization errors are important when training ML models. Therefore, using the available imbalanced attack data, we created synthesized attack data using CTGAN, and an experiment was conducted using these data to train the ML model. Then we compared the corresponding results with SMOTE of the oversampling technique to confirm the performance of each trained ML model.</p>
<p>In this study, we conducted comparative and analytical research to develop an effective method for training an ML model to reduce the generalization error by solving the problem of insufficient and imbalanced data consisting of tabular 5GC traffic. The contributions of this study are as follows. 5GC traffic was considered because the research objective is to internalize a training model that detects IoT DDoS attacks in 5GC. This required a 5G testbed, a virtual environment, to be constructed, which was used to conduct a simulation experiment to collect 5GC traffic. Features based on protocol headers were extracted using the collected traffic to detect DDoS attacks in network traffic. Therefore, we used 5GC traffic in tabular form for the experiment. Oversampling (SMOTE) and data augmentation (GAN-CTGAN) were used to solve the insufficient and imbalanced problem of 5GC traffic. The classification performance of the trained ML models was compared and analyzed in terms of the evaluation metrics and generalization error obtained for three comparable models (Baseline, SMOTE, and GAN-CTGAN). Because the attack data of 5GC traffic is imbalanced, the attack data were augmented using GAN-CTGAN. The experiment confirmed that the ML model training method using GAN-CTGAN, compared to Baseline and SMOTE, tended to alleviate the generalization error problem.</p>
<p>The rest of this paper is organized as follows. Section 2 introduces relevant existing studies. Section 3 provides details of the background of this study, the 5G environment, and two techniques (SMOTE, GAN-CTGAN) to solve the data imbalance problem. Section 4 describes the datasets and procedures for collecting 5GC traffic used in the study. Section 5 reviews the experimental results and differences from previous studies. Finally, Section 6 concludes the study and suggests future research.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Research</title>
<p>This section discusses research related to this study. First, the latest research on network security using AI is presented. Second, related work in which the data imbalance problem was solved using SMOTE is discussed. At the same time, we explore research that has utilized SMOTE in various research areas beyond network security. Finally, studies using GANs in the field of network security are introduced. A GAN is used to solve the data imbalance problem and various other problems.</p>
<sec id="s2_1">
<label>2.1</label>
<title>AI-Related Network Security Research</title>
<p>Saheed et al. (2022) [<xref ref-type="bibr" rid="ref-6">6</xref>] proposed support vector machine (SVM)-based k-means and genetic algorithm (GA) models to improve the efficiency of the intrusion detection system (IDS). The dataset they used was NSL-KDD, and the experimental results showed a low false acceptance rate (FAR) and high accuracy of 99%.</p>
<p>Haq et al. (2022) [<xref ref-type="bibr" rid="ref-7">7</xref>] proposed two models, DNNBoT1 and DNNBoT2, which utilized the principal component analysis (PCA) process, model optimization, and a callback function in the learning process to detect IoT-based botnet attacks. Their dataset was N-BaIoT, and the experimental results confirmed excellent training and verification accuracy.</p>
<p>Rizvi et al. (2023) [<xref ref-type="bibr" rid="ref-8">8</xref>] proposed a model using the 1D-dilated causal neural network (1D-DCNN) to solve the problem of IDSs requiring significant computational resources and processing time. They used two datasets (CIC-IDS2017, CSE-CIC-IDS2018) to evaluate their proposed model and showed the accuracy of CIC-IDS2017 to be 99.7% and CSE-CIC-IDS2018 to be 99.98%.</p>
<p>The AI-related research described above aims to improve IDS efficiency. In addition, it can be seen that NSL-KDD, N-BaIoT, CIC-IDS2017, and CSE-CIC-IDS2018, which are various datasets for each study, are being used to propose and evaluate AI-related models.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>SMOTE-Related Research to Address Data Imbalance</title>
<p>Ma et al. (2020) [<xref ref-type="bibr" rid="ref-9">9</xref>] combined the reinforcement learning of the adapted SMOTE and adversarial environment using reinforcement learning (AE-RL) to develop the AESMOTE algorithm for anomaly detection. The corresponding solution was obtained using the NSL-KDD dataset, and the performance of the proposed model surpassed that of the previous AE-RL.</p>
<p>Won et al. (2020) [<xref ref-type="bibr" rid="ref-10">10</xref>] compared oversampling (SMOTE and its variants) and data augmentation using the two-side-channel protected advanced encryption standard (AES). Their experimental results indicated that adapting SMOTE variants could increase the attack&#x2019;s efficiency in general. Additionally, the proposed model achieved key recovery in the AS-CAD dataset (desync &#x003D; 100) with 180 traces. This result was a 50% improvement from the latest technology.</p>
<p>Karatas et al. (2020) [<xref ref-type="bibr" rid="ref-11">11</xref>] used SMOTE in an ML-based IDS and proposed a model for the synthesis data of the minor class. These increased the data in the minor class to the average amount of data, and a significant increase in the detection rate for seldom-occurring trespassing was observed.</p>
<p>Zhang et al. (2020) [<xref ref-type="bibr" rid="ref-12">12</xref>] proposed a new SGM model. They tested the model performance and training rate by designing a mixed Gaussian model comprising SMOTE, undersampling, and convolutional neural network (CNN). The proposed model was validated using UNSW-NB15 and CICIDS2017. They achieved detection rates of 99.74% (binary) and 96.54% (multi) for UNSW-NB15 and a detection rate of 99.85% (15-classification) for CIC-IDS2017.</p>
<p>Haq et al. (2022) [<xref ref-type="bibr" rid="ref-13">13</xref>] proposed SMOTEDNN to classify and predict air pollution. The method was developed to solve the problem of data imbalance and non-optimization in classification models. The proposed model was developed using a dataset of monitored air pollutant levels in India released by NAMP, and the accuracy was 99.9%.</p>
<p>Dablain et al. (2022) [<xref ref-type="bibr" rid="ref-14">14</xref>] proposed DeepSMOTE, a SMOTE-based encoder/decoder framework for oversampling and artificial instance generation for images with imbalanced and high-dimensional data. They utilized five datasets (MNIST, Fashion-MNIST, CIFAR-10, SVHNs, and CelebA), and the performance results showed that their proposed method could generate artificial images of superior quality, outperforming GAN-based oversampling.</p>
<p>Joloudari et al. (2022) [<xref ref-type="bibr" rid="ref-15">15</xref>] conducted a deep learning (DL)-based sampling comparison experiment to solve the imbalanced data problem. They used three datasets (KEEL, breast cancer, Z-Alizadeh Sani). The SMOTE-based CNN model delivered the best performance, which achieved 99.08% accuracy on the imbalanced dataset.</p>
<p>As such, seven studies [<xref ref-type="bibr" rid="ref-9">9</xref>&#x2013;<xref ref-type="bibr" rid="ref-15">15</xref>] related to SMOTE addressed the issue of data imbalance. The scope of this research is not limited to network security [<xref ref-type="bibr" rid="ref-9">9</xref>&#x2013;<xref ref-type="bibr" rid="ref-12">12</xref>], which is associated with this work but also covers a wide range of studies, such as air pollution and imagery [<xref ref-type="bibr" rid="ref-13">13</xref>&#x2013;<xref ref-type="bibr" rid="ref-15">15</xref>].</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>GAN-Related Network Security Research</title>
<p>Hu et al. (2023) [<xref ref-type="bibr" rid="ref-16">16</xref>] proposed a GAN-based algorithm called MalGAN, consisting of a substitute detector and generative network, to generate adversarial malware cases that bypass the black-box ML-based detection model. Furthermore, the experiment trained the generative network to minimize the probability of generating adversarial malware predicted by the substitute detector. Consequently, they confirmed that the performance of MalGAN reduced the detection rate to near zero and could render retraining-based defense methods against adversarial examples challenging to operate.</p>
<p>Ahmed et al. (2020) [<xref ref-type="bibr" rid="ref-17">17</xref>] proposed a model for generating uniform resource locator (URL) based phishing cases using GANs. The GAN structure used in this approach consisted of a generator network, discriminator network, and black-box phishing detector. The performance results demonstrated that the GAN network could effectively create adversarial phishing examples to thwart simple and sophisticated ML phishing detection models. Therefore, their study confirmed the need to create countermeasures in response to adversarial examples.</p>
<p>Abdelaty et al. (2021) [<xref ref-type="bibr" rid="ref-18">18</xref>] proposed an adversarial training approach called GADoT, which utilizes GANs to generate adversarial DDoS samples for robust model training. The dataset was detected using DDoS samples (SYN and HTTP Flood), and the classifier used was LUCID, an ML-based model. The model trained with GADoT detected all DDoS attacks with high accuracy and achieved an f1-score of at least 98% and a false negative rate (FNR) of up to 1.8% as evaluation indicators.</p>
<p>Guo et al. (2021) [<xref ref-type="bibr" rid="ref-19">19</xref>] proposed a GAN-based traffic augmentation (TA-GAN) framework for unbalanced traffic classifications. Their approach involved pre-training and tripartite learning processes and used a 1D-convolutional neural networks (1D-CNN) model as a classifier. The binary classification performance of their framework yielded an f1-score of up to 14.64% higher for the minority class owing to the influence of traffic imbalance.</p>
<p>Nan et al. (2022) [<xref ref-type="bibr" rid="ref-20">20</xref>] proposed an effective anomaly detection framework using GAN and long short-term memory (LSTM) neural networks in 5G environments. This framework analyzes the correlation between user actions to scale datasets and improves the performance without altering the data distribution. The experimental results confirmed an accuracy of 97.16% and a false positive rate (FPR) of 2.30% by utilizing the correlation between user actions and dataset expansion.</p>
<p>Kim et al. (2021) [<xref ref-type="bibr" rid="ref-21">21</xref>] proposed a network anomaly detection method that solves the problem of acquiring labeled network traffic data in 5G networks. Their approach was based on a GAN with scaling properties that adopt unsupervised learning. UNSW-NB15 was used as the dataset to evaluate the framework performance and utilized a CNN as a classifier. The performance evaluations confirmed a maximum accuracy of 99.30% and an f1-score of 0.99.</p>
<p>Park et al. (2022) [<xref ref-type="bibr" rid="ref-22">22</xref>] used boundary equilibrium GAN (BEGAN) and the autoencoder-driven DL model to develop an AI-based network intrusion detection system (NIDS) with performance reinforced by creating new synthesis data. They used four datasets (NSL-KDD, UNSW-NB15, IoT, and Real Data) to evaluate the performance of the proposed model. The classification of experimental results was superior to that of previous methods.</p>
<p>The above studies are ML and DL-based security-related studies using GAN in 5G and wired network environments. Currently, researchers are focusing on GANs rather than SMOTE, in which GANs are used for various purposes as well as data imbalance processing. In addition, previous research has shown that different types of GANs are being utilized depending on the type of data being augmented. This study used GAN to solve the imbalanced nature of malicious data. In addition, since the data types used are tabular, CTGAN, which generates attack network data that are effectively synthesized, was utilized [<xref ref-type="bibr" rid="ref-4">4</xref>].</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Background</title>
<sec id="s3_1">
<label>3.1</label>
<title>5G Network Environment</title>
<p>5G wireless network technology succeeded 4th generation mobile communication (4G), and its commercialization proceeded from 2018. The visions and objectives of 5G were presented by the international telecommunication union (ITU), and the development of a technological standard is underway at the international standardization organization 3GPP [<xref ref-type="bibr" rid="ref-23">23</xref>]. 3GPP defines the 5G network as a service-based architecture (SBA), which provides service by dispersing and virtualizing at the network function (NF) unit and by connecting at the service-based interface (SBI) [<xref ref-type="bibr" rid="ref-24">24</xref>]. The following sections introduce the composition of 5G in three parts and the User Plane Function (UPF) and NWDAF as two NFs of 5G that are significantly relevant to the work presented in this paper.</p>
<sec id="s3_1_1">
<label>3.1.1</label>
<title>3 Parts of 5G: RAN, DN, and Core</title>
<p><xref ref-type="fig" rid="fig-1">Fig. 1</xref> shows a diagram of the 5G network, with the core inside and the base station and data network (DN) outside. 5G network has three major components: NG-RAN (RAN), DN, and core. The RAN further comprises the base station gNodeB (gNB) and user equipment (UE), i.e., the smartphone. The user is provided network service by being connected to the core using the N3 interface of gNB. DN is connected to UE using the N6 interface of the core and delivers the data to the user. The core can be divided into the user plane (UP), responsible for data transfer, and the control plane (CP), which controls the user end. The core provides services to all the NFs connected to the SBI.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Three major components of 5G: RAN, DN, and core</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-1.tif"/>
</fig>
</sec>
<sec id="s3_1_2">
<label>3.1.2</label>
<title>UPF and NWDAF</title>
<p>Among the NFs used in this research, the UPF and NWDAF are the most relevant. The UPF was the NF that mainly comprised the 5G dataset in our research. In addition, NWDAF is an NF for internalizing security analysis functions using ML in 5GC, which we are studying and analyzing.</p>
<p>The 5G dataset was established using the N3 interface connected to UPF. The UPF is an NF that handles UP, which is responsible for packet routing and forwarding, packet inspection, quality of service (QoS) handling, and external protocol data unit (PDU) sessions and also interconnects DN [<xref ref-type="bibr" rid="ref-25">25</xref>]. UPF has three reference points, as shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>: N6, N4, and N3 [<xref ref-type="bibr" rid="ref-24">24</xref>]. N4 is the interface between Session Management Function (SMF) and UPF, N6 is DN and UPF, and N3 is gNB and UPF. The N3 interface undergoes encapsulation that expands the GPRS tunneling protocol (GTP) header in the previous internet protocol (IP) header through the GTP. GTP is a protocol for encapsulating and tunneling the IP packet data provided and received by the user in DN. The experiments to construct the 5G dataset utilized the N3 interface. As a result, <xref ref-type="fig" rid="fig-2">Fig. 2</xref> shows a structure where the UPF and RAN are connected via the N3 interface. In addition, as an experiment, IP-based packets were dumped to UEs in the RAN and changed to GTP-based packets in the 5G Network (5GC) as they passed through the N3 interface. We utilized packets from this modified 5G core in our experiments.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>5G dataset composition method for the experiment&#x2014;5G network (5GC) GTP-based packet capture using N3 interface (gNB-UPF)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-2.tif"/>
</fig>
<p>To provide security analysis functions to the 5GC, NWDAF, which has the ML function among the 5GC NFs, was studied and analyzed. The current 3GPP standardization defines NWDAF. The major function of NWDAF is to use ML and provide analysis results of the network data [<xref ref-type="bibr" rid="ref-2">2</xref>]. <xref ref-type="fig" rid="fig-3">Fig. 3</xref> shows the three steps NWDAF takes to deliver the derived network analysis results: (1) collect network data, (2) analyze network data using ML, and (3) provide the analysis results. In addition, the NWDAF structure consists of the Model Training Logical Function (MTLF) and the Analytics Logical Function (AnLF). MTLF trains the ML models and provides the trained ML models. AnLF performs inferences, derives analytics information, and exposes analytics services. As such, NWDAF of 5GC trains the ML models and uses the trained models to provide the derived network data analysis services. Therefore, we aimed to study the ML training method for detecting IoT DDoS attacks by using and collecting the 5GC traffic to provide the security analysis function using the ML model that was trained in 5GC, similar to NWDAF.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>Procedure used by NWDAFs for providing derived network data analysis results</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-3.tif"/>
</fig>
</sec>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Techniques for Resolving Data Imbalance Problems</title>
<p>Several techniques exist to resolve the dataset imbalance issue, which degrades the model performance when training the ML model. In this research, oversampling and data augmentation were selected as two methods for solving the problem of imbalanced 5GC network traffic data [<xref ref-type="bibr" rid="ref-26">26</xref>,<xref ref-type="bibr" rid="ref-27">27</xref>]. Specifically, the experiment used SMOTE for oversampling and GAN-CTGAN for attack data augmentation.</p>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>SMOTE for Oversampling</title>
<p>Oversampling is the most widely used technique for overcoming network traffic imbalance due to the intuitive principle that rewards imbalance by increasing the number of traffic samples that belong to the minority class [<xref ref-type="bibr" rid="ref-9">9</xref>&#x2013;<xref ref-type="bibr" rid="ref-15">15</xref>]. This study used SMOTE, which is the most usual oversampling method. SMOTE generates new data by copying virtual data in the minority class through the distance-based rule. Among previously reported oversampling methods, SMOTE improved overfitting as the disadvantage of random oversampling (ROS) [<xref ref-type="bibr" rid="ref-19">19</xref>]. However, SMOTE does not consider the majority class because new data is generated among the minority class data. Thus, the generated data only reflect the properties of data in the minority class. As a result, they may be prone to noise and unable to augment high-dimensional data efficiently. This research employed scikit-learn, an ML open-source library for SMOTE [<xref ref-type="bibr" rid="ref-28">28</xref>].</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>GAN-CTGAN for Attack Data Augmentation</title>
<p>Data augmentation is a technique that increases the number of data by applying various transformations to the original data to create new data [<xref ref-type="bibr" rid="ref-29">29</xref>]. This study used GAN, a popular method for data augmentation. GAN, which uses DL for augmenting data, comprises a generator that generates data and a discriminator that classifies the generated data. The two have a feedback relationship; the generator generates data that are improved by learning through feedback provided by the discriminator. GAN is being actively researched and is used for data diversification. Data can be distinguished mainly into sequential and tabular forms. This study used tabular 5GC traffic data to train the ML model, and a CTGAN validated for generating tabular network data was employed [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-30">30</xref>,<xref ref-type="bibr" rid="ref-31">31</xref>]. In particular, this study attempted to solve the data imbalance problem by generating attack data. To use CTGAN, we leveraged a related paper and open-source published on GitHub [<xref ref-type="bibr" rid="ref-32">32</xref>,<xref ref-type="bibr" rid="ref-33">33</xref>].</p>
</sec>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Experimental Design</title>
<p>In this work, SMOTE and GAN were used as methods of training the ML model to detect IoT DDoS attacks in the 5GC environment. Experiments were designed to evaluate the performance of the trained model. First, the experimental procedures and the methods used for collecting 5GC traffic for experimental use are introduced.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Experimental Dataset for 5GC Traffic</title>
<p>To collect 5GC traffic for the experiments, we need (1) a dataset from the wired network environment and (2) a 5G testbed. We used the GPRS tunneling experiment of the wired network environment dataset because the 5G testbed using this dataset could generate and gather 5GC traffic data.</p>
<sec id="s4_1_1">
<label>4.1.1</label>
<title>Kitsune Dataset (Wired Network)</title>
<p>The kitsune dataset, which is openly offered in Kaggle, was used as the wired network environment dataset to generate and gather 5GC traffic [<xref ref-type="bibr" rid="ref-34">34</xref>,<xref ref-type="bibr" rid="ref-35">35</xref>]. Kitsune consists of nine DDoS attack datasets generated by IoT equipment in a wired network environment. The nine types of DDoS attacks are as follows: OS Scan, Fuzzing, Video Injection, ARP MitM, Active Wiretap, SSDP Flood, SYN DoS, SSL Renegotiation, and Mirai botnet. These attack types were used, excluding two (Video Injection and Mirai botnet). The exclusion was because Video Injection was the only type that included the Ethernet class&#x2019;s logical-link control (LLC) protocol. The Mirai botnet had 3.4 times fewer attack packets compared with the average of the eight other types.</p>
<p>Another reason for selecting kitsune for this study is that it was suitable for conducting the GPRS Tunneling experiment using the 5G testbed, as the provided data formats were diverse. Kitsune offers the following three data formats for the nine types of attacks: the dataset with the original packet pre-processed (.csv), a labeled dataset indicating the malignity or normality of the packet (.csv), and captured dataset of the original network (.pcap). The original network dataset (.pcap) and labeled dataset (.csv) were used for gathering 5GC traffic.</p>
</sec>
<sec id="s4_1_2">
<label>4.1.2</label>
<title>5G Testbed</title>
<p><bold>Building the 5G testbed.</bold> The 5G testbed was constructed in a Linux 18.04 environment. To build the 5G testbed, we leveraged two open-source software programs. The open-source utilized UERANSIM for RAN and Open5GS for the core. Finally, we built a testbed with 5G RAN (UERANSIM) &#x002B; 5G Core (Open5GS). UERANSIM is the simulator of 5G UE and gNB. UERANSIM is open-source software available on GitHub and can be used to test the core and study the 5G system. In addition, it is compatible with Open5GS among the 5GS open projects [<xref ref-type="bibr" rid="ref-36">36</xref>]. Open5GS implements core according to 3GPP release 16, and it was developed in the C language. Open5GS is available on GitHub, offering 10 NFs as of September 22, 2022. Active development updates have been underway for the corresponding project since September 22 [<xref ref-type="bibr" rid="ref-37">37</xref>].</p>
<p><bold>Experiments on 5G testbed.</bold> Using the constructed 5G testbed, an experiment for collecting the 5GC traffic was performed. First, the packet was dumped with the UE, and the GTP packet that comprised GPRS Tunneling was captured as it passed through the gNB-UPF (N3 interface). Thus, in the experiment, we transformed the IP-based packet of the wired network into the GTP-based packet of the 5G network. Then, to capture the corresponding packet, Linux tcpdump was used.</p>
<p><bold>Feature extraction.</bold> We extracted GTP-based packets, which are 5GC traffic obtained from our experiments, and converted them to CSV for use in the ML model. <xref ref-type="table" rid="table-1">Table 1</xref> summarizes information about all 57 features used in this experiment. The features are extracted based on the six network protocols used in the collected dataset. The six protocols are Internet Control Message Protocol (ICMP), Transmission Control Protocol (TCP), User Datagram Protocol (UDP), Internet Group Management Protocol (IGMP), IP, and GTP. Wireshark&#x0027;s Tshark was used for feature extraction. The six protocols of the gathered packet were used as the criteria for extracting the packet according to the header. Considering the headers of these protocols, 93 pitchers were extracted, and the pitchers that did not have values were removed after checking the presence of values for all data from the extracted headers.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Extracted feature type according to the protocol header</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Protocol</th>
<th>Count</th>
<th>Feature</th>
</tr>
</thead>
<tbody>
<tr>
<td>ICMP</td>
<td>4</td>
<td>icmp.type, icmp.code, icmp.checksum, icmp.checksum.status</td>
</tr>
<tr>
<td>TCP</td>
<td>17</td>
<td>tcp.port, tcp.seq, tcp.seq_raw, tcp.ack, tcp.ack.nonzero, tcp.ack_raw</td>
</tr>
<tr>
<td/>
<td/>
<td>tcp.flag.res, tcp.flags.urg, tcp.flags.ack, tcp.flags.syn, tcp.flags.fin</td>
</tr>
<tr>
<td/>
<td/>
<td>tcp.flags, tcp.flags.push, tcp.window_size, tcp.checksum</td>
</tr>
<tr>
<td/>
<td/>
<td>tcp.urgent_pointer, tcp.checksum.status</td>
</tr>
<tr>
<td>UDP</td>
<td>5</td>
<td>udp.srcport, udp.dstport, udp.checksum, udp.checksum.status, udp.length</td>
</tr>
<tr>
<td>IGMP</td>
<td>5</td>
<td>igmp.type, igmp.max_resp, igmp.checksum, igmp.checksum.status</td>
</tr>
<tr>
<td/>
<td/>
<td>igmp.num_grp_recs</td>
</tr>
<tr>
<td>IP</td>
<td>10</td>
<td>ip.version, ip.len, ip.id, ip.flags, ip.frag_offset, ip.ttl, ip.proto, ip.checksum</td>
</tr>
<tr>
<td/>
<td/>
<td>ip.src, ip.dst</td>
</tr>
<tr>
<td>GTP</td>
<td>16</td>
<td>gtp.ip.src, gtp.ip.dst, gtp.flags, gtp.flags.version, gtp.flags.payload</td>
</tr>
<tr>
<td/>
<td/>
<td>gtp.flags.reserved, gtp.flags.e, gtp.flags.s, gtp.glags.pn, gtp.message</td>
</tr>
<tr>
<td/>
<td/>
<td>gtp.length, gtp.teid, gtp.ext_hdr.next, gtp.ext_hdr.length</td>
</tr>
<tr>
<td/>
<td/>
<td>gtp.ext_hdr.pdu_ses_con.pdu_type, gtp.ext_her.pdu_ses_con.qos_flow_id</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Experimental Procedure</title>
<p>This section describes the experimental procedure. In the experimental procedure, we first review the workflow to understand the flow and procedure of the experiment and then describe the procedure in detail based on that workflow.</p>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Workflow</title>
<p><xref ref-type="fig" rid="fig-4">Fig. 4</xref> is the workflow for this experiment, showing the flow and procedure. It had five procedures, and each procedure is explained as follows.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>Workflow of experiment</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-4.tif"/>
</fig>
<p>(1)&#x007E;(2). After pre-processing the dataset, divide it into training-testing sets.</p>
<p>(3). Configure the training dataset to create three comparison models (Baseline, SMOTE, and GAN-CTGAN).</p>
<p>(3)-a. Baseline is used as the training set.</p>
<p>(3)-b. SMOTE is used for training by oversampling in the training set. The reason for excluding the testing set from oversampling is that the sample data replicated by SMOTE is not included in the testing set for model performance verification.</p>
<p>(3)-c. GAN-CTGAN is used for training by augmenting malicious data by utilizing the training-testing set and including the synthesized malicious data in the training set. The reason for using the testing set for malicious data augmentation is that we have a small dataset to experiment with, so we want to utilize as much data as possible to generate plausible synthetic malicious data.</p>
<p>(4). Train the ML model using the training set for the three comparison models.</p>
<p>(5). For the trained ML model, validation and test results are derived depending on the presence or absence of the testing set.</p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Detailed Procedure</title>
<p>This time, we describe the procedures in the workflow in more detail, such as the methods used, data counts, and algorithms. <xref ref-type="fig" rid="fig-5">Fig. 5</xref> showed five detailed experimental procedures: (1) pre-processing, (2) splitting the training and testing sets, (3) three comparative models (Baseline, SMOTE, and GAN-CTGAN), (4) binary and multi-classification, and (5) model evaluation.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Experimental detailed procedure</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-5.tif"/>
</fig>
<p><bold>Pre-processing.</bold> Pre-processing involves data cleaning and feature scaling. Data cleaning entailed the elimination of missing values by inserting the value 0 where no value existed. Additionally, we transformed hexadecimal numbers into decimal numbers and floats into integers. The feature scaling then went through standardization and normalization to bring the values to a consistent level.</p>
<p><bold>Splitting the training and testing sets.</bold> To evaluate the performance of the ML training model in this experiment, we separate the training and testing sets. The total size of the dataset for the experiment was 65,000, and the kitsune dataset rate (7:3) for the (benign/malicious) ratio was maintained (45,500/19,500). Finally, the (benign/malicious) rate was maintained while the data were divided into the training and testing sets (52,000 and 13,000 data, respectively). To ensure that the ML model underwent effective training on the imbalanced class dataset, the training and testing set both maintained their (benign/malicious) ratios.</p>
<p><bold>Three comparative models: Baseline, SMOTE, and GAN-CTGAN.</bold> This study analyzed and compared the performance of the three models. The models can be distinguished by their ML model training methods. First, the ML model was trained with the Baseline training set (52,000). Second, SMOTE was oversampled in the training set (52,000), and the ML model was trained. Lastly, the 26,005 attack data were augmented for GAN-CTGAN using the seven types of attack data, with 3,715 data per session. Thus, we added 26,005 data were added via augmentation to the original training set (52,000) and used 78,005 data to train the ML model.</p>
<p><bold>Binary and multi-classification.</bold> The kitsune dataset was used with multiple attacks to experiment on binary and multi-classification. First, the data were divided into benign (label 0) and maligned (label 1) for binary classification. Subsequently, multi-classification was performed by dividing the data into benign (label 0) and seven types of attacks (labels 1&#x2013;7). In addition, ML and the ensemble of machine learning (EML) were used as classifiers. First, ML classifiers were used for SVM, k-nearest neighbor (KNN), and decision tree (DT). Then, EML classifiers were used for random forest (RF), AdaBoost, and XGBoost.</p>
<p><bold>Model evaluation.</bold> The stratified k-fold (cv &#x003D; 5) was used to train the model and evaluate the performance because the dataset for training the ML model was insufficient and had imbalanced classes. In addition, these performed fivefold cross-validation and then generated five trained models. Finally, we utilized the validation set for the five trained models to get five validation results and averaged them to get the final validation result. We also got the same test results.</p>
<p>The confusion matrices used in this study were accuracy and f1-score. Accuracy intuitively refers to model prediction performance. Its definition is expressed as <xref ref-type="disp-formula" rid="eqn-1">Eq. (1)</xref>. However, it is unsuitable for use as a metric for imbalanced datasets. For this study, we also selected the f1-score, as it is highly prone to data bias due to the imbalanced classification problem. F1-score is the harmonic mean of recall and precision. Recall is a metric showing the percentage of positives correctly classified as true positives. Its definition is expressed as <xref ref-type="disp-formula" rid="eqn-2">Eq. (2)</xref>. precision is a metric showing the percentage of negatives correctly classified as true negatives. Its definition is expressed as <xref ref-type="disp-formula" rid="eqn-3">Eq. (3)</xref>. Therefore, the f1-score can cancel out differences even if the ratio of positives to negatives is imbalanced. Its definition is expressed as <xref ref-type="disp-formula" rid="eqn-4">Eq. (4)</xref>.</p>
<p><disp-formula id="eqn-1">
<label>(1)</label>
<mml:math id="mml-eqn-1" display="block"><mml:mrow><mml:mtext>Accuracy</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p><disp-formula id="eqn-2">
<label>(2)</label>
<mml:math id="mml-eqn-2" display="block"><mml:mrow><mml:mtext>Precision</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p><disp-formula id="eqn-3">
<label>(3)</label>
<mml:math id="mml-eqn-3" display="block"><mml:mrow><mml:mtext>Recall</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
<p><disp-formula id="eqn-4">
<label>(4)</label>
<mml:math id="mml-eqn-4" display="block"><mml:mrow><mml:mi mathvariant="normal">F</mml:mi></mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mtext>score</mml:mtext></mml:mrow><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mo>&#x00D7;</mml:mo><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:mrow></mml:mfrac></mml:math></disp-formula></p>
</sec>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Results and Discussions</title>
<p>This section presents the results of the experiments. First, the performance of the binary and multi-classification of the Baseline is presented for problem definition and verification. Next, the performance results for the three trained models with binary classification are compared. The results for the three models are evaluated from the perspective of the generalization error. A comparison with previous research revealed differences.</p>
<sec id="s5_1">
<label>5.1</label>
<title>Experiment Results</title>
<sec id="s5_1_1">
<label>5.1.1</label>
<title>Problem Definition and Proof with Baseline</title>
<p>A model that does not use either SMOTE as the Baseline oversampling technique or GAN-CTGAN as the data augmentation technique was referenced for comparison. We compared the Baseline&#x2019;s binary and multi-classification performance results to define and prove the problem. <xref ref-type="table" rid="table-2">Table 2</xref> shows the validation results, test results, and two value differences for binary and multi-classification. In addition, the results are shown as accuracy and f1-score of six ML classification models. For binary classification, the accuracy and f1-score decreased. The accuracy had a minimum reduction of 5.95% (DT) and a maximum value of 14.89% (AdaBoost), and the f1-score had a minimum value of 16.08% (DT) and a maximum value of 54.72% (SVM). For multi-classification, the accuracy increased, with a range of 1.74&#x2013;3.51%, excluding KNN and AdaBoost. Additionally, the f1-score increased, with a minimum value of 0.06% (RF) and a maximum value of 3.17% (DT), excluding KNN.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Acc and f1 for the validation and test results for binary and multi-classification</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th colspan="3"></th>
<th align="center" colspan="3">Binary</th>
<th align="center" colspan="3">Multi</th>
</tr>
<tr>
<th colspan="2">Classifier</th>
<th>Acc/F1</th>
<th>Validation</th>
<th>Test</th>
<th>Difference</th>
<th>Validation</th>
<th>Test</th>
<th>Difference</th>
</tr>
</thead>
<tbody>
<tr>
<td>ML</td>
<td>SVM</td>
<td>Acc</td>
<td>85.47</td>
<td>70.65</td>
<td>(&#x2212;14.82)</td>
<td>83.54</td>
<td>87.05</td>
<td>(&#x002B;3.51)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>77.11</td>
<td>22.39</td>
<td>(&#x2212;54.72)</td>
<td>74.72</td>
<td>76.51</td>
<td>(&#x002B;1.79)</td>
</tr>
<tr>
<td/>
<td>KNN</td>
<td>Acc</td>
<td>86.01</td>
<td>79.92</td>
<td>(&#x2212;6.09)</td>
<td>93.22</td>
<td>91.66</td>
<td>(&#x2212;1.56)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>71.88</td>
<td>53.83</td>
<td>(&#x2212;18.05)</td>
<td>94.29</td>
<td>91.52</td>
<td>(&#x2212;2.77)</td>
</tr>
<tr>
<td/>
<td>DT</td>
<td>Acc</td>
<td>90.17</td>
<td>84.22</td>
<td>(&#x2212;5.95)</td>
<td>91.05</td>
<td>95.85</td>
<td>(&#x002B;4.8)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>79.18</td>
<td>63.10</td>
<td>(&#x2212;16.08)</td>
<td>92.87</td>
<td>96.04</td>
<td>(&#x002B;3.17)</td>
</tr>
<tr>
<td>EML</td>
<td>RF</td>
<td>Acc</td>
<td>86.32</td>
<td>75.96</td>
<td>(&#x2212;10.36)</td>
<td>93.50</td>
<td>95.24</td>
<td>(&#x002B;1.74)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>72.51</td>
<td>49.56</td>
<td>(&#x2212;22.95)</td>
<td>95.04</td>
<td>95.10</td>
<td>(&#x002B;0.06)</td>
</tr>
<tr>
<td/>
<td>Adaboost</td>
<td>Acc</td>
<td>89.09</td>
<td>74.20</td>
<td>(&#x2212;14.89)</td>
<td>70.81</td>
<td>68.96</td>
<td>(&#x2212;1.91)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>78.37</td>
<td>43.37</td>
<td>(&#x2212;35)</td>
<td>18.84</td>
<td>18.98</td>
<td>(&#x002B;0.14)</td>
</tr>
<tr>
<td/>
<td>XGBoost</td>
<td>Acc</td>
<td>91.86</td>
<td>81.02</td>
<td>(&#x2212;10.84)</td>
<td>92.39</td>
<td>95.68</td>
<td>(&#x002B;3.29)</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>83.38</td>
<td>57.67</td>
<td>(&#x2212;25.71)</td>
<td>94.43</td>
<td>95.75</td>
<td>(&#x002B;1.32)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The kitsune dataset, which consisted of data relating to seven IoT DDoS attacks, did not experience generalization errors, as the performance of the validation and test results were not significantly different for multi-classification. However, a significant performance difference between the validation and test results occurred in the case of binary classification, and a generalization error occurred because of overfitting. Therefore, we analyzed the training methods for the ML model to reduce the generalization error for the binary classification of the kitsune dataset.</p>
</sec>
<sec id="s5_1_2">
<label>5.1.2</label>
<title>Binary Classification: Results of Three Comparative Models</title>
<p>The method of presenting the binary classification results of the three comparison models confirms and compares the trend of the results from the perspective of validation and test results. <xref ref-type="fig" rid="fig-6">Fig. 6</xref> shows the trend of accuracy and f1-score for the six classifiers. The three comparison models&#x2019; validation results (blue) and test results (orange) are shown for each classifier. Regarding validation results, SMOTE showed increased performance in both accuracy and f1-score. Regarding test results, SMOTE can confirm the improved performance in f1-score and GAN-CTGAN in accuracy. <xref ref-type="table" rid="table-3">Table 3</xref> shows the exact values of accuracy and f1-score as shown in <xref ref-type="fig" rid="fig-6">Fig. 6</xref>. The following compares the results for each of the three comparison models regarding verification and test results for the values in <xref ref-type="table" rid="table-3">Table 3</xref>.</p>
<fig id="fig-6">
<label>Figure 6</label>
<caption>
<title>Graph showing the binary&#x2014;acc and f1 trend for validation and test results of three comparative models. (a) Accuracy. (b) F1-score</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-6.tif"/>
</fig><table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Binary&#x2014;acc and f1 for the validation and test results of three comparative models</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th colspan="3"></th>
<th align="center" colspan="2">Baseline</th>
<th align="center" colspan="2">SMOTE</th>
<th align="center" colspan="2">GAN-CTGAN</th>
</tr>
<tr>
<th colspan="2">Algorithm</th>
<th></th>
<th>Validation</th>
<th>Test</th>
<th>Validation</th>
<th>Test</th>
<th>Validation</th>
<th>Test</th>
</tr>
</thead>
<tbody>
<tr>
<td>ML</td>
<td>SVM</td>
<td>Acc</td>
<td>85.47</td>
<td>70.65</td>
<td>62.42</td>
<td>41.89</td>
<td>80.31</td>
<td>71.69</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>77.11</td>
<td>22.39</td>
<td>70.37</td>
<td>47.64</td>
<td>71.67</td>
<td>20.95</td>
</tr>
<tr>
<td/>
<td>KNN</td>
<td>Acc</td>
<td>86.01</td>
<td>79.92</td>
<td>95.08</td>
<td>78.97</td>
<td>84.51</td>
<td>79.92</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>71.88</td>
<td>53.83</td>
<td>95.36</td>
<td>65.86</td>
<td>83.40</td>
<td>53.62</td>
</tr>
<tr>
<td/>
<td>DT</td>
<td>Acc</td>
<td>90.17</td>
<td>84.22</td>
<td>95.09</td>
<td>81.99</td>
<td>85.65</td>
<td>85.39</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>79.18</td>
<td>63.10</td>
<td>95.72</td>
<td>68.46</td>
<td>81.24</td>
<td>62.61</td>
</tr>
<tr>
<td>EML</td>
<td>RF</td>
<td>Acc</td>
<td>86.32</td>
<td>75.96</td>
<td>95.69</td>
<td>77.79</td>
<td>83.33</td>
<td>79.12</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>72.51</td>
<td>49.56</td>
<td>96.26</td>
<td>60.90</td>
<td>74.29</td>
<td>51.88</td>
</tr>
<tr>
<td/>
<td>Adaboost</td>
<td>Acc</td>
<td>89.09</td>
<td>74.20</td>
<td>95.02</td>
<td>78.62</td>
<td>80.62</td>
<td>75.75</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>78.37</td>
<td>43.37</td>
<td>95.40</td>
<td>53.77</td>
<td>71.85</td>
<td>42.37</td>
</tr>
<tr>
<td/>
<td>XGBoost</td>
<td>Acc</td>
<td>91.86</td>
<td>81.01</td>
<td>97.62</td>
<td>81.20</td>
<td>82.83</td>
<td>82.18</td>
</tr>
<tr>
<td/>
<td/>
<td>F1</td>
<td>83.38</td>
<td>57.67</td>
<td>97.78</td>
<td>61.31</td>
<td>73.84</td>
<td>55.49</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><bold>Performance comparison for validation result.</bold> The accuracy for the Baseline validation result was between 85.47% and 79.86%, and the f1-score was between 71.88% and 92.38%. The accuracy for the SMOTE validation result was between 95.02% and 97.62%, excluding SVM (62.42%), and the f1-score was between 75.36% and 97.78%. The accuracy for the GAN-CTGAN validation result was between 80.31% and 85.65%, and the f1-score was between 71.85% and 83.70%.</p>
<p>Compared with the Baseline, the accuracy and f1-score tended to be higher for SMOTE (excluding SVM). The lowest accuracy was 4.92% (DT), and the highest was 9.07% (KNN). The lowest f1-score was 14.4% (XGBoost), and the highest was 23.48% (KNN). Additionally, compared with the Baseline, GAN-CTGAN tended to have lower accuracy. The lowest accuracy was 1.5% (KNN), and the highest was 8.47% (AdaBoost). Compared with the Baseline, the f1-score was reduced for SVM, AdaBoost, and XGBoost and increased for KNN, DT, and RF. From the perspective of the validation result, the oversampling technique using SMOTE exhibited higher ML training model performance than the data augmentation using GAN-CTGAN.</p>
<p><bold>Performance comparison for the test result.</bold> The Baseline test result had accuracies between 70.65% and 84.22% and f1-scores between 43.37% and 63.10%, excluding SVM (22.39%). The SMOTE test result accuracy and f1-score were 77.79&#x2013;81.99% and 53.77&#x2013;68.46%, respectively, excluding SVM (47.64%). Lastly, the test result accuracy of GAN-CTGAN was between 71.69% and 85.39%, and the f1-score was between 42.37% and 62.61%, excluding SVM (20.95%).</p>
<p>In the case of SMOTE, compared with the Baseline, the accuracy was reduced for SVM, KNN, and DT and increased for RF, AdaBoost, and XGBoost. The f1-score tended to increase; the lowest value was 3.64% (XGBoost), and the highest value was 25.25% (SVM). For GAN-CTGAN, compared with the Baseline, the accuracy increased, and the F1-score tended to decrease, excluding RF. The lowest accuracy was 0% (KNN), and the highest was 3.16% (RF). The lowest f1-score was 0.21% (KNN), and the highest was 2.18% (XGBoost), excluding RF (2.35%). Therefore, from the perspective of the test result, the f1-score indicated improved performance for oversampling using SMOTE compared with data augmentation using GAN-CTGAN. However, relative to oversampling with the three classifier accuracies reduced, data augmentation generally exhibited superior performance for the six classifiers.</p>
</sec>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Perspective of Generalization Error</title>
<p>In the case of the binary classification of the Baseline, the performance difference between the validation and test results was significant. Therefore, generalization errors tended to occur because of overfitting. Oversampling based on SMOTE and data augmentation based on GAN-CTGAN were used to study the method for ML model training method. We evaluated the performance of the trained model from the perspective of the generalization error in comparison with the Baseline for SMOTE and GAN-CTGAN. Two methods were used to evaluate the generalization error for the trained model: (1) the learning curve and (2) the difference between the validation and test results.</p>
<sec id="s5_2_1">
<label>5.2.1</label>
<title>Learning Curve</title>
<p>This section examines the learning curve for the three comparative models. <xref ref-type="fig" rid="fig-7">Fig. 7</xref> shows the learning curves for the three comparative models for SVM and RF. The scikit-learn learning curves were used, and the cross-validation value was 10. The graph shows the accuracy for the training (red) and validation (green) sets for the size of the training data sample. The estimated variance was expressed by deriving the standard deviation of the average accuracy for each accuracy and utilizing the fill_between function. In the Baseline, both SVM and RF showed significant differences in accuracy and estimated variance between training and validation. Therefore, the Baseline is likely to be overfitted and exhibit generalization errors.</p>
<fig id="fig-7">
<label>Figure 7</label>
<caption>
<title>Learning curve graphs for SVM and RF of three comparative models. (a) ML-SVM algorithm. (b) EML-RF algorithm</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-7.tif"/>
</fig>
<p>For SVM, the SMOTE decreased training and validation accuracy and estimated variance compared to the Baseline. The GAN-CTGAN increased training and validation accuracy and decreased estimation variance compared to the Baseline. Compared with that Baseline, the SMOTE and GAN-CTGAN training accuracies were the same for RF, but the validation accuracy was higher, and the estimated variance was lower. Therefore, the estimated variance was reduced for both SMOTE and GAN-CTGAN, with the possibility of the generalization error having decreased by reducing the overfitting.</p>
</sec>
<sec id="s5_2_2">
<label>5.2.2</label>
<title>Performance Differences Between Validation and Test Results</title>
<p>The difference between the validation and test results is checked to assess the generalization error regarding the three comparison models. <xref ref-type="fig" rid="fig-8">Fig. 8</xref> consists of two graphs of accuracy and f1-score, which show the difference trend between the validation and test results for each of the six classifiers. We can also see in <xref ref-type="fig" rid="fig-8">Fig. 8</xref> that the model with the smallest value difference among the three comparison models mitigates the generalization error problem. <xref ref-type="table" rid="table-4">Table 4</xref> shows the exact value of the difference between the accuracy and f1-score value expressed in <xref ref-type="fig" rid="fig-8">Fig. 8</xref>. The following is a comparison of the results for each of the three comparison models for the values in <xref ref-type="table" rid="table-4">Table 4</xref>. SMOTE exhibited a performance difference: the accuracy compared with the Baseline increased to a minimum of 1.51% (AdaBoost) and a maximum of 10.02% (KNN). The performance difference was also evident from the f1-score, which increased to a minimum of 6.63% (AdaBoost) and a maximum of 12.44% (RF), excluding SVM (31.99% reduction). Additionally, a performance difference arose where the f1-score decreased to a minimum of 0.54% (RF) and a maximum of 7.26% (XGBoost), excluding KNN (11.73% increase) and DT (2.55% increase). Therefore, from the validation and test results perspective, the generalization error tended to increase for SMOTE and decrease for GAN-CTGAN.</p>
<fig id="fig-8">
<label>Figure 8</label>
<caption>
<title>Graphs showing the trend of the difference between acc and f1 of the validation and test results for the three comparison models. (a) Accuracy. (b) F1-score</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-8.tif"/>
</fig><table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>Difference in acc and f1 between validation and test results of the three comparative models</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th colspan="2"></th>
<th align="center" colspan="2">Baseline</th>
<th align="center" colspan="2">SMOTE</th>
<th align="center" colspan="2">GAN-CTGAN</th>
</tr>
<tr>
<th colspan="2">Algorithm</th>
<th>Acc diff.</th>
<th>F1 diff.</th>
<th>Acc diff.</th>
<th>F1 diff.</th>
<th>Acc diff.</th>
<th>F1 diff.</th>
</tr>
</thead>
<tbody>
<tr>
<td>ML</td>
<td>SVM</td>
<td>14.82</td>
<td>54.72</td>
<td>20.53</td>
<td>22.73</td>
<td>8.62</td>
<td>50.72</td>
</tr>
<tr>
<td/>
<td>KNN</td>
<td>6.09</td>
<td>18.05</td>
<td>16.11</td>
<td>29.5</td>
<td>4.59</td>
<td>29.78</td>
</tr>
<tr>
<td/>
<td>DT</td>
<td>5.95</td>
<td>16.08</td>
<td>13.1</td>
<td>27.26</td>
<td>0.26</td>
<td>18.63</td>
</tr>
<tr>
<td>EML</td>
<td>RF</td>
<td>10.36</td>
<td>22.95</td>
<td>17.9</td>
<td>35.36</td>
<td>4.21</td>
<td>22.41</td>
</tr>
<tr>
<td/>
<td>Adaboost</td>
<td>14.89</td>
<td>35</td>
<td>16.4</td>
<td>41.63</td>
<td>4.87</td>
<td>29.48</td>
</tr>
<tr>
<td/>
<td>XGBoost</td>
<td>10.84</td>
<td>25.71</td>
<td>16.42</td>
<td>36.47</td>
<td>0.65</td>
<td>18.45</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Computational Complexity</title>
<p>This section presents an assessment of the computational complexity of the experimental environment of SMOTE and GAN-CTGAN, the techniques that are used to solve the data imbalance of network traffic. The computational complexity was determined by measuring each technique&#x2019;s required time and RAM usage. First, the time library of Python was used to measure the required time. Next, the required time was measured by checking the current time at the start and end of the code of SMOTE and GAN-CTGAN. The following uses the psutil library to measure RAM usage. The method measures the RAM usage when the code starts and ends running and excludes the RAM usage at the start from the RAM usage at the end.</p>
<p><xref ref-type="fig" rid="fig-9">Fig. 9</xref> and <xref ref-type="table" rid="table-5">Table 5</xref> show the time required and RAM usage when processing SMOTE (Oversampling) and GAN-CTGAN (Data Augmentation). <xref ref-type="fig" rid="fig-9">Fig. 9</xref> visualizes the trend of time and RAM usage when performing SMOTE and GAN-CTGAN. In addition, <xref ref-type="table" rid="table-5">Table 5</xref> shows the exact values for the trends depicted in <xref ref-type="fig" rid="fig-9">Fig. 9</xref>. The following is a comparison of the results of SMOTE and GAN-CTGAN for <xref ref-type="table" rid="table-5">Table 5</xref>. Compared to GAN-CTGAN, the time required by SMOTE was 1158 times less and used 12 times less RAM. SMOTE uses less time and less memory than GAN-CTGAN. However, previous experimental results show that SMOTE generates simple data using less time and space resources. In addition, compared to SMOTE, GAN-CTGAN is more time-consuming and uses a large amount of memory. However, previous experimental results confirmed that sophisticated data were created using considerable time and space resources.</p>
<fig id="fig-9">
<label>Figure 9</label>
<caption>
<title>Graph of computational complexity. (a) Required time (h). (b) RAM usage (kb)</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CSSE_39550-fig-9.tif"/>
</fig><table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>Measurement of computational complexity</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Technique</th>
<th>Required time (h)</th>
<th>RAM usage (kb)</th>
</tr>
</thead>
<tbody>
<tr>
<td>SMOTE</td>
<td>0.00014586</td>
<td>321.175</td>
</tr>
<tr>
<td>GAN-CTGAN</td>
<td>0.16894671</td>
<td>3913.957</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s5_4">
<label>5.4</label>
<title>Comparison of Previous Research</title>
<p>This section compares our results to the related work reviewed in Section 2, and the differences are identified. In this study, we conducted a comparative experiment using SMOTE and GAN-CTGAN to address the problem of generalization error due to data imbalance. For each method, a comparison with related studies is performed to identify our results&#x2019; meaning, differences, or limitations. The approach involves comparing the SMOTE experiments with SMOTE-related studies, after which the GAN-CTGAN experiments are compared with GAN-related studies.</p>
<sec id="s5_4_1">
<label>5.4.1</label>
<title>SMOTE-Related Research</title>
<p>The SMOTE studies discussed in Section 2 aim to address the data imbalance problem. In addition, SMOTE is used not only in the field of network security related to this study but also in various research fields. Therefore, three studies introduced in Section 2 were selected and compared with the SMOTE experiment. The parts to be compared can be divided into four categories in <xref ref-type="table" rid="table-6">Table 6</xref>. (1) Dataset, (2) ML/DL models, (3) SMOTE description (research field/purpose/other tech.), and (4) Results. Of these categories, the main category is (3) SMOTE description. Because the research fields using SMOTE and the techniques used together differ for each corresponding category, this was analyzed and compared with a focus on these techniques. The purpose of this comparison is not to compare the performance, as the datasets are different, but to analyze the differences between the previous study and this study and the information related to SMOTE to identify the meaning, differences, or limitations of the experimental results.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>Comparison of SMOTE-related research</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Authors (year)</th>
<th>Dataset</th>
<th>ML/DL models</th>
<th>SMOTE description (research field/ purpose/<break/> other tech.)</th>
<th>Results</th>
</tr>
</thead>
<tbody>
<tr>
<td>Zhang et al. (2020) [<xref ref-type="bibr" rid="ref-12">12</xref>]</td>
<td>o UNSW-NB15<break/>o CICIDS2017</td>
<td>o ML<break/>-RF<break/>o DL<break/>-MLP, CNN</td>
<td>o (research field) Network security<break/>- (purpose) Data imbalance<break/>- (other tech.) GMM</td>
<td>Acc (96.5%&#x007E;99.7%)</td>
</tr>
<tr>
<td>Haq et al. (2022) [<xref ref-type="bibr" rid="ref-13">13</xref>]</td>
<td>o NAMP</td>
<td>o ML<break/>-SVM, KNN, XGBoost, RF<break/>o DL<break/>-DNN,</td>
<td>o (research field) Air pollution<break/>- (purpose) Data imbalance<break/>- (other tech.) Model optimization</td>
<td>Acc (99.9%)</td>
</tr>
<tr>
<td>Dablain et al. (2022) [<xref ref-type="bibr" rid="ref-14">14</xref>]</td>
<td>o MNIST<break/>o Fashion-MNIST<break/>o CIFAR-10<break/>o SVHNs<break/>o CelebA</td>
<td>o DL<break/>-CNN</td>
<td>o (research field) Image<break/>- (purpose) Data imbalance<break/>- (other tech.) Encoder/Decoder</td>
<td>Acc (96.4%&#x007E;98.1%)</td>
</tr>
<tr>
<td>In this research</td>
<td>o Kitsune</td>
<td>o ML<break/>-SVM, KNN, DT<break/>o EML<break/>-RF, Adaboost, XGBoost</td>
<td>o (research field) Network security<break/>- (purpose) Data imbalance<break/>- (other tech.) None</td>
<td>Acc<break/>(41.8%&#x007E;81.99%)<break/>F1 (47.64%&#x007E;68.46%)</td>
</tr>
</tbody>
</table>
</table-wrap>

<p>As a result of the comparison, recent SMOTE studies use both ML and DL, but it is confirmed that SMOTE performs well in DL. In addition, other techniques (GMM, Encoder, etc.) are used together rather than using SMOTE alone. As such, this SMOTE experiment utilized SMOTE only on ML, not DL, and did not combine it with other techniques. Compared to GAN-CTGAN, it was confirmed that the SMOTE experimental results were not mitigated regarding the generalization error. Therefore, the comparison with previous studies indicated that it is necessary to confirm the experimental results by using DL and other techniques (GMM, Encoder, etc.) rather than the SMOTE application method of this study.</p>
</sec>
<sec id="s5_4_2">
<label>5.4.2</label>
<title>GAN-Related Research</title>
<p>The GAN research discussed in Section 2 is from the field of network security, which is related to this research. This research used GANs to solve various problems, not just the data imbalance issue. Among the studies introduced in Section 2, three were selected for comparison with ours. The purpose of this comparison is not to compare the performance of the datasets but to analyze the differences between the previous study and this study and the GAN-related information to identify the experimental results&#x2019; meaning, differences, or limitations. The parts to be compared can be divided into the four categories in <xref ref-type="table" rid="table-7">Table 7</xref>: (1) Dataset, (2) ML/DL models, (3) GAN description (type/purpose/method), and (4) Results. In particular, compared to previous studies, the main category is (3) GAN description (type/purpose/method). The type, purpose, and method of GAN used for each study in this category differ. Focusing on these features, we analyzed and compared previous research.</p>
<table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>Comparison of GAN-related research</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Authors (year)</th>
<th>Dataset</th>
<th>ML/DL models</th>
<th>GAN description (type/purpose/method)</th>
<th>Results</th>
</tr>
</thead>
<tbody>
<tr>
<td>Ahmed et al. (2020) [<xref ref-type="bibr" rid="ref-17">17</xref>]</td>
<td>o Own dataset</td>
<td>o ML<break/>-RF, DT, LR,<break/> NN, SVM</td>
<td>o (type) GAN<break/>- (purpose) Bypassing URL classification algorithms for ML models<break/>- (method) Creating adversarial attacks</td>
<td>FPR (0.12% or less)</td>
</tr>
<tr>
<td>Park et al. (2022) [<xref ref-type="bibr" rid="ref-22">22</xref>]</td>
<td>o NSL-KDD<break/>o UNSW-NB15<break/>o IoT<break/>o Real data</td>
<td>o DL<break/>-DNN, CNN, LSTM</td>
<td>o (type) BEGAN<break/>- (purpose) Resolving data imbalance<break/>- (method) Creating synthetic data for minor attack</td>
<td>F1 (93.8%)</td>
</tr>
<tr>
<td>Abdelaty et al. (2021) [<xref ref-type="bibr" rid="ref-18">18</xref>]</td>
<td>o CICIDS2017<break/>o Adversarial SYN flood attacks</td>
<td>o ML<break/>-LUCID</td>
<td>o (type) GADoT<break/>- (purpose) Adversarial training for robust model<break/>- (method) Training set augmented adversarial DDoS attacks</td>
<td>F1 (98% or more)<break/>FNR (1.8% or less)</td>
</tr>
<tr>
<td>In this research</td>
<td>o Kitsune</td>
<td>o ML<break/>-SVM, KNN, DT<break/>o EML<break/>-RF, Adaboost, XGBoost</td>
<td>o (type) CTGAN<break/>- (purpose 1) Resolving insufficient training data of 5GC Traffic<break/>- (method 1) Creating<break/> synthetic data for 5GC traffic in tabular form<break/>- (purpose 2) Resolving data imbalance<break/>- (method 2) Training set augmented by IoT DDoS attacks</td>
<td>Acc (71.6% or more)<break/>F1 (62.6% or less)</td>
</tr>
</tbody>
</table>
</table-wrap>

<p>The comparison revealed that recent GAN research mainly focused on the field of network security, and in particular, GAN is used in various ways as well as data augmentation for data imbalance in this study. Our work was impeded by the availability of insufficient 5GC traffic in tabular form to train ML models. Therefore, CTGAN, which has been proven to generate learning data from network data in tabular form effectively, was used in the experiment [<xref ref-type="bibr" rid="ref-4">4</xref>,<xref ref-type="bibr" rid="ref-30">30</xref>,<xref ref-type="bibr" rid="ref-31">31</xref>]. In addition, owing to network traffic, the IoT DDoS attack was augmented using CTGAN and then included in the training set to train the ML model. Consequently, two perspectives, the evaluation index values and generalization errors were examined as the results of the experiment. The experimental results confirmed that CTGAN did not change significantly in the evaluation index values compared to Baseline, but the generalization error tended to be alleviated by reducing overfitting. Our investigation confirmed that various methods and applications based on GAN had been developed, as well as data augmentation methods in the field of network security.</p>
</sec>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Conclusion</title>
<p>This study conducted a comparative experiment for an ML model training method to mitigate the generalization error problem under insufficient and imbalanced data. This experiment shows that the method of augmenting attack data using GAN-CTGAN is effective as an ML model training method for classifying IoT DDoS attacks in 5G environments. First, to solve the problem of insufficient data, we built our own 5G testbed and constructed a dataset of a 5G environment. Next, we processed three training datasets (Baseline, SMOTE, and GAN-CTGAN) using SMOTE and GAN-CTGAN to solve the imbalanced data problem. Accordingly, it was confirmed that the generalization error problem occurred in the ML model trained with the Baseline. The ML model trained with GAN-CTGAN showed the tendency to alleviate the generalization error problem the most.</p>
<p>The meaning of the experimental results was analyzed from various perspectives by comparing the computational complexity and previous studies regarding SMOTE and GAN-CTGAN as the techniques used. First, we measured the computational complexity of generating data with SMOTE and GAN-CTGAN. Although GAN-CTGAN has a higher computational complexity than SMOTE, it was analyzed that the generalization error was mitigated, as it generated sophisticated data. Next, we compared SMOTE and GAN to previous studies to see the latest trends in each technique. As a recent trend, the SMOTE showed effective performance when using DL and other techniques together rather than ML alone, and it was confirmed that the GAN has various methods and applications in the field of network security.</p>
<p>In this experiment, the performance of the trained ML model was analyzed using one kitsune dataset. As a future research direction, we plan to evaluate the performance of the trained ML model in terms of higher feasibility by using multiple datasets rather than one dataset. In addition, there are not only CTGAN but also TableGAN and CopularGAN as GAN types for generating tabular data. Since the CTGAN is effective in the 5GC environment, we plan to carry out research to conduct comparative experiments on other types of GANs. In addition, this study created attack data using training and test sets due to a lack of data when augmenting attack data with CTGAN. In future research, this should be improved so that only the training set can be utilized to generate data.</p>
</sec>
</body>
<back>
<sec><title>Funding Statement</title>
<p>This work was supported by Institute of Information &#x0026; communications Technology Planning &#x0026; Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2021-0-00796, Research on Foundational Technologies for 6G Autonomous Security-by-Design to Guarantee Constant Quality of Security).</p>
</sec>
<sec sec-type="COI-statement"><title>Conflicts of Interest</title>
<p>The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A. M.</given-names> <surname>French</surname></string-name> and <string-name><given-names>J. P.</given-names> <surname>Shim</surname></string-name></person-group>, &#x201C;<article-title>The digital revolution: Internet of things, 5G, and beyond</article-title>,&#x201D; <source>Communications of the Association for Information Systems</source>, vol. <volume>38</volume>, no. <issue>1</issue>, pp. <fpage>840</fpage>&#x2013;<lpage>850</lpage>, <year>2016</year>. <pub-id pub-id-type="doi">10.17705/1CAIS.03840</pub-id></mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>3GPP&#x2014;Technical Specification Group Services and System Aspects</collab></person-group>, &#x201C;<article-title>Architecture enhancement for 5G System (5GS) to support network data analytics services (release 18)</article-title>,&#x201D; <year>2022</year>. [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.3gpp.org/ftp/Specs/archive/23_series/23.288/23288-i00.zip">https://www.3gpp.org/ftp/Specs/archive/23_series/23.288/23288-i00.zip</ext-link></mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>L.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Skoularidou</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Cuesta-Infante</surname></string-name> and <string-name><given-names>K.</given-names> <surname>Veeramachaneni</surname></string-name></person-group>, &#x201C;<article-title>Modeling tabular data using conditional GAN</article-title>,&#x201D; <source>Advances in Neural Information Processing Systems</source>, vol. <volume>32</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>15</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Bourou</surname></string-name>, <string-name><given-names>A. El</given-names> <surname>Saer</surname></string-name>, <string-name><given-names>T. H.</given-names> <surname>Velivassaki</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Voulkidis</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Zahariadis</surname></string-name></person-group>, &#x201C;<article-title>A review of tabular data synthesis using GANs on an IDS dataset</article-title>,&#x201D; <source>Information</source>, vol. <volume>12</volume>, no. <issue>9</issue>, pp. <fpage>375</fpage>&#x2013;<lpage>388</lpage>, <year>2021</year>. <pub-id pub-id-type="doi">10.3390/info12090375</pub-id></mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Nadeau</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Bengio</surname></string-name></person-group>, &#x201C;<article-title>Inference for the generalization error</article-title>,&#x201D; in <conf-name>Proc. of the Advances in Neural Information Processing Systems (NIPS)</conf-name>, <publisher-loc>London, England</publisher-loc>, vol. <volume>12</volume>, pp. <fpage>307</fpage>&#x2013;<lpage>313</lpage>, <year>1999</year>. </mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Y. K.</given-names> <surname>Saheed</surname></string-name>, <string-name><given-names>M. O.</given-names> <surname>Arowolo</surname></string-name> and <string-name><given-names>A. U.</given-names> <surname>Tosho</surname></string-name></person-group>, &#x201C;<article-title>An efficient hybridization of k-means and genetic algorithm based on support vector machine for cyber intrusion detection system</article-title>,&#x201D; <source>International Journal on Electrical Engineering and Informatics</source>, vol. <volume>14</volume>, no. <issue>2</issue>, pp. <fpage>426</fpage>&#x2013;<lpage>442</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.15676/ijeei.2022.14.2.11</pub-id></mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Haq</surname></string-name> and <string-name><given-names>M. A. R.</given-names> <surname>Khan</surname></string-name></person-group>, &#x201C;<article-title>Dnnbot: Deep neural network-based botnet detection and classification</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>71</volume>, no. <issue>1</issue>, pp. <fpage>1729</fpage>&#x2013;<lpage>1750</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.32604/cmc.2022.020938</pub-id></mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Rizvi</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Scanlon</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Macgibney</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Sheppard</surname></string-name></person-group>, &#x201C;<article-title>Deep learning based network intrusion detection system for resource-constrained environments</article-title>,&#x201D; in <conf-name>Proc. 13th Int. Conf. on Digital Forensics and Cybercrime (ICDF2C)</conf-name>, <conf-loc>Boston, MA, USA</conf-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>7</lpage>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Ma</surname></string-name> and <string-name><given-names>W.</given-names> <surname>Shi</surname></string-name></person-group>, &#x201C;<article-title>Aesmote: Adversarial reinforcement learning with SMOTE for anomaly detection</article-title>,&#x201D; <source>IEEE Transactions on Network Science and Engineering</source>, vol. <volume>8</volume>, no. <issue>2</issue>, pp. <fpage>943</fpage>&#x2013;<lpage>956</lpage>, <year>2020</year>. <pub-id pub-id-type="doi">10.1109/TNSE.2020.3004312</pub-id></mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y. S.</given-names> <surname>Won</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Jap</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Bhasin</surname></string-name></person-group>, &#x201C;<article-title>Push for more: On comparison of data augmentation and SMOTE with optimised deep learning architecture for side-channel</article-title>,&#x201D; in <conf-name>Proc. Information Security Applications: 21st Int. Conf. (WISA 2020)</conf-name>, <publisher-loc>Jeju Island, Korea</publisher-loc>, vol. <volume>21</volume>, pp. <fpage>227</fpage>&#x2013;<lpage>241</lpage>, <year>2020</year>. </mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Karatas</surname></string-name>, <string-name><given-names>O.</given-names> <surname>Demir</surname></string-name> and <string-name><given-names>O. K.</given-names> <surname>Sahingoz</surname></string-name></person-group>, &#x201C;<article-title>Increasing the performance of machine learning-based IDSs on an imbalanced and up-to-data dataset</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>32150</fpage>&#x2013;<lpage>32162</lpage>, <year>2020</year>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.2973219</pub-id></mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Huang</surname></string-name>, <string-name><given-names>C. Q.</given-names> <surname>Wu</surname></string-name> and <string-name><given-names>Z.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>An effective convolutional neural network based on SMOTE and gaussian mixture model for intrusion detection in imbalanced dataset</article-title>,&#x201D; <source>Computer Networks</source>, vol. <volume>177</volume>, no. <issue>18</issue>, pp. <fpage>107315</fpage>&#x2013;<lpage>107324</lpage>, <year>2020</year>. <pub-id pub-id-type="doi">10.1016/j.comnet.2020.107315</pub-id></mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. A.</given-names> <surname>Haq</surname></string-name></person-group>, &#x201C;<article-title>Smotednn: A novel model for air pollution forecasting and aqi classification</article-title>,&#x201D; <source>Computers, Materials &#x0026; Continua</source>, vol. <volume>71</volume>, no. <issue>1</issue>, pp. <fpage>1403</fpage>&#x2013;<lpage>1425</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.32604/cmc.2022.021968</pub-id></mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Dablain</surname></string-name>, <string-name><given-names>B.</given-names> <surname>Krawczyk</surname></string-name> and <string-name><given-names>V.</given-names> <surname>Chawla</surname></string-name></person-group>, &#x201C;<article-title>DeepSMOTE: Fusing deep learning and SMOTE for imbalanced data</article-title>,&#x201D; <source>IEEE Transactions on Neural Networks and Learning Systems</source>, pp. <fpage>1</fpage>&#x2013;<lpage>15</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.1109/TNNLS.2021.3136503</pub-id>; <pub-id pub-id-type="pmid">35085094</pub-id></mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>J. H.</given-names> <surname>Joloudari</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Marefat</surname></string-name>, <string-name><given-names>M. A.</given-names> <surname>Nematollahi</surname></string-name>, <string-name><given-names>S. S.</given-names> <surname>Oyelere</surname></string-name> and <string-name><given-names>S.</given-names> <surname>Hussain</surname></string-name></person-group>, &#x201C;<article-title>Effective class-imbalance learning based on SMOTE and convolution neural networks</article-title>,&#x201D; <comment>arXiv preprint arXiv:2209.00653</comment>, pp. <fpage>1</fpage>&#x2013;<lpage>43</lpage>, <year>2022</year>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Hu</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Tan</surname></string-name></person-group>, &#x201C;<article-title>Generating adversarial malware examples for black-box attacks based on GAN</article-title>,&#x201D; in <conf-name>Proc. Data Mining and Big Data: 7th Int. Conf. (DMBD)</conf-name>, <publisher-loc>Beijing, China</publisher-loc>, pp. <fpage>409</fpage>&#x2013;<lpage>423</lpage>, <year>2023</year>. </mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Ahmed</surname></string-name> and <string-name><given-names>G.</given-names> <surname>Karabatis</surname></string-name></person-group>, &#x201C;<article-title>Bypassing detection of URL-based phishing attacks using generative adversarial deep neural networks</article-title>,&#x201D; in <conf-name>Proc. of the 6th Int. Workshop on Security and Privacy Analytics (IWSPA)</conf-name>, <publisher-loc>LA, New Orleans, USA</publisher-loc>, pp. <fpage>53</fpage>&#x2013;<lpage>60</lpage>, <year>2020</year>. </mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Abdelaty</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Scott-Hayward</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Doriguzzi-Corin</surname></string-name> and <string-name><given-names>D.</given-names> <surname>Siracusa</surname></string-name></person-group>, &#x201C;<article-title>GADoT: GAN-based adversarial training for robust DDoS attack detection</article-title>,&#x201D; in <conf-name>Proc. 2021 IEEE Conf. on Communications and Network Security (CNS)</conf-name>, pp. <fpage>119</fpage>&#x2013;<lpage>127</lpage>, <year>2021</year>. </mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Guo</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Xiaong</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Li</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Shi</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Cui</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>TA-GAN: GAN based traffic augmentation for imbalanced network traffic classification</article-title>,&#x201D; in <conf-name>Proc. 2021 Int. Joint Conf. on Neural Networks (IJCNN)</conf-name>, <conf-loc>Shenzhen, China</conf-loc>, pp. <fpage>1</fpage>&#x2013;<lpage>8</lpage>, <year>2021</year>. </mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Nan</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhu</surname></string-name> and <string-name><given-names>J.</given-names> <surname>Ma</surname></string-name></person-group>, &#x201C;<article-title>An efficient correlation-aware anomaly detection framework in cellular network</article-title>,&#x201D; <source>China Communications</source>, vol. <volume>19</volume>, no. <issue>8</issue>, pp. <fpage>168</fpage>&#x2013;<lpage>180</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.23919/JCC.2022.08.013</pub-id></mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>H. J.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Lee</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Park</surname></string-name> and <string-name><given-names>J. G.</given-names> <surname>Park</surname></string-name></person-group>, &#x201C;<article-title>Network anomaly detection based on GAN with scaling properties</article-title>,&#x201D; in <conf-name>Proc. of the 2021 Int. Conf. on Information and Communication Technology Convergence (ICTC)</conf-name>, <publisher-loc>Jeju Island, Korea</publisher-loc>, pp. <fpage>1244</fpage>&#x2013;<lpage>1248</lpage>, <year>2021</year>. </mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Park</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Lee</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>J. G.</given-names> <surname>Park</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Kim</surname></string-name> <etal>et al.</etal></person-group><italic>,</italic> &#x201C;<article-title>An enhanced AI-based network intrusion detection system using generative adversarial networks</article-title>,&#x201D; <source>IEEE Internet of Things Journal</source>, vol. <volume>10</volume>, no. <issue>3</issue>, pp. <fpage>2330</fpage>&#x2013;<lpage>2345</lpage>, <year>2022</year>. <pub-id pub-id-type="doi">10.1109/JIOT.2022.3211346</pub-id></mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>ITU</collab></person-group>, &#x201C;<article-title>IMT vision&#x2014;Framework and overall objectives of the future development of IMT for 2020 and beyond</article-title>,&#x201D; <year>2015</year>. [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.itu.int/rec/R-REC-M.2083-0-201509-P/en">https://www.itu.int/rec/R-REC-M.2083-0-201509-P/en</ext-link></mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>3GPP&#x2014;Technical Specification Group Services and System Aspects</collab></person-group>, &#x201C;<article-title>System architecture for the 5G system (5GS); stage2 (release 18)</article-title>,&#x201D; <year>2022</year>. [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.3gpp.org/ftp/Specs/archive/23_series/23.501/23501-i00.zip">https://www.3gpp.org/ftp/Specs/archive/23_series/23.501/23501-i00.zip</ext-link></mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>Cisco Systems. Inc</collab></person-group>, &#x201C;<article-title>Ultra cloud core 5G user plane function, release 2021.01&#x2014;configuration and administration guide</article-title>,&#x201D; <year>2021</year>. [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.cisco.com/c/en/us/td/docs/wireless/ucc/upf/2021-01/b_ucc-5g-upf-config-and-admin-guide_2021-01/m_upf-overview.html">https://www.cisco.com/c/en/us/td/docs/wireless/ucc/upf/2021-01/b_ucc-5g-upf-config-and-admin-guide_2021-01/m_upf-overview.html</ext-link></mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Mohammed</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Rawashdeh</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Adbullah</surname></string-name></person-group>, &#x201C;<article-title>Machine learning with oversampling and undersampling techniques: Overview study and experimental results</article-title>,&#x201D; in <conf-name>Proc. 2020 11th Int. Conf. on Information and Communication Systems (ICICS)</conf-name>, <publisher-loc>Irbid, Jordan</publisher-loc>, pp. <fpage>243</fpage>&#x2013;<lpage>248</lpage>, <year>2020</year>. </mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Mikolajczyk</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Grochowski</surname></string-name></person-group>, &#x201C;<article-title>Data augmentation for improving deep learning in image classification problem</article-title>,&#x201D; in <conf-name>Proc. 2018 Int. Interdisciplinary PhD Workshop (IIPhDW)</conf-name>, <publisher-loc>Swinoujscie, Poland</publisher-loc>, pp. <fpage>117</fpage>&#x2013;<lpage>122</lpage>, <year>2018</year>. </mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>The imbalanced-learn developers</collab></person-group>, [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://imbalanced-learn.org/stable/references/generated/imblearn.over_sampling.SMOTE.html">https://imbalanced-learn.org/stable/references/generated/imblearn.over_sampling.SMOTE.html</ext-link></mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Shorten</surname></string-name> and <string-name><given-names>T. M.</given-names> <surname>Khoshogoftaar</surname></string-name></person-group>, &#x201C;<article-title>A survey on image data augmentation for deep learning</article-title>,&#x201D; <source>Journal of Big Data</source>, vol. <volume>6</volume>, no. <issue>1</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>48</lpage>, <year>2019</year>. <pub-id pub-id-type="doi">10.1186/s40537-019-0197-0</pub-id></mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Moon</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Jung</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Park</surname></string-name> and <string-name><given-names>A. E.</given-names> <surname>Hwang</surname></string-name></person-group>, &#x201C;<article-title>Conditional tabular GAN-based two-stage data generation scheme for short-term load forecasting</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>8</volume>, pp. <fpage>205327</fpage>&#x2013;<lpage>205339</lpage>, <year>2020</year>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.3037063</pub-id></mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Habibi</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Chemmakha</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Lazaar</surname></string-name></person-group>, &#x201C;<article-title>Imbalanced tabular data modelization using CTGAN and machine learning to improve IoT botnet attacks detection</article-title>,&#x201D; <source>Engineering Applications of Artificial Intelligence</source>, vol. <volume>118</volume>, no. <issue>9</issue>, pp. <fpage>105669</fpage>&#x2013;<lpage>105691</lpage>, <year>2023</year>. <pub-id pub-id-type="doi">10.1016/j.engappai.2022.105669</pub-id></mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>Github</collab></person-group>, &#x201C;<article-title>CTGAN-conditional GAN for generating synthetic tabular data</article-title>,&#x201D; [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://github.com/sdv-dev/CTGAN">https://github.com/sdv-dev/CTGAN</ext-link></mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>L.</given-names> <surname>Xu</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Skoularidou</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Cuesta-Infante</surname></string-name> and <string-name><given-names>K.</given-names> <surname>Veeramachaneni</surname></string-name></person-group>, &#x201C;<article-title>Modeling tabular data using conditional GAN</article-title>,&#x201D; <source>Advances in Neural Information Processing Systems</source>, vol. <volume>32</volume>, pp. <fpage>1</fpage>&#x2013;<lpage>11</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>Kaggle</collab></person-group>, &#x201C;<article-title>Kitsune network attack dataset</article-title>,&#x201D; [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://www.kaggle.com/datasets/ymirsky/network-attack-dataset-kitsune">https://www.kaggle.com/datasets/ymirsky/network-attack-dataset-kitsune</ext-link></mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>Y.</given-names> <surname>Mirsky</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Doitshman</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Elovici</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Shabtai</surname></string-name></person-group>, &#x201C;<article-title>Kitsune: An ensemble of autoencoders for online network intrusion detection</article-title>,&#x201D; <comment>arXiv preprint arXiv:1802.09089</comment>, pp. <fpage>1</fpage>&#x2013;<lpage>15</lpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>Github</collab></person-group>, &#x201C;<article-title>UERANSIM: Open source 5G UE and RAN (gNodeB) implementation</article-title>,&#x201D; [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://github.com/aligungr/UERANSIM">https://github.com/aligungr/UERANSIM</ext-link></mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><collab>Open5GS</collab></person-group>, &#x201C;<article-title>Open5GS: Quickstart</article-title>,&#x201D; [Online]. Available: <ext-link ext-link-type="uri" xlink:href="https://open5gs.org/open5gs/docs/guide/01-quickstart/">https://open5gs.org/open5gs/docs/guide/01-quickstart/</ext-link></mixed-citation></ref>
</ref-list>
</back>
</article>