<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "http://jats.nlm.nih.gov/publishing/1.1/JATS-journalpublishing1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xml:lang="en" article-type="research-article" dtd-version="1.1">
<front>
<journal-meta>
<journal-id journal-id-type="pmc">CMC</journal-id>
<journal-id journal-id-type="nlm-ta">CMC</journal-id>
<journal-id journal-id-type="publisher-id">CMC</journal-id>
<journal-title-group>
<journal-title>Computers, Materials &#x0026; Continua</journal-title>
</journal-title-group>
<issn pub-type="epub">1546-2226</issn>
<issn pub-type="ppub">1546-2218</issn>
<publisher>
<publisher-name>Tech Science Press</publisher-name>
<publisher-loc>USA</publisher-loc>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">52835</article-id>
<article-id pub-id-type="doi">10.32604/cmc.2024.052835</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Fine-Tuning Cyber Security Defenses: Evaluating Supervised Machine Learning Classifiers for Windows Malware Detection</article-title>
<alt-title alt-title-type="left-running-head">Fine-Tuning Cyber Security Defenses: Evaluating Supervised Machine Learning Classifiers for Windows Malware Detection</alt-title>
<alt-title alt-title-type="right-running-head">Fine-Tuning Cyber Security Defenses: Evaluating Supervised Machine Learning Classifiers for Windows Malware Detection</alt-title>
</title-group>
<contrib-group>
<contrib id="author-1" contrib-type="author" corresp="yes">
<name name-style="western"><surname>Zada</surname><given-names>Islam</given-names></name><xref ref-type="aff" rid="aff-1">1</xref><email>islam.zada@iiu.edu.pk</email></contrib>
<contrib id="author-2" contrib-type="author">
<name name-style="western"><surname>Alatawi</surname><given-names>Mohammed Naif</given-names></name><xref ref-type="aff" rid="aff-2">2</xref></contrib>
<contrib id="author-3" contrib-type="author">
<name name-style="western"><surname>Saqlain</surname><given-names>Syed Muhammad</given-names></name><xref ref-type="aff" rid="aff-1">1</xref></contrib>
<contrib id="author-4" contrib-type="author">
<name name-style="western"><surname>Alshahrani</surname><given-names>Abdullah</given-names></name><xref ref-type="aff" rid="aff-3">3</xref></contrib>
<contrib id="author-5" contrib-type="author">
<name name-style="western"><surname>Alshamran</surname><given-names>Adel</given-names></name><xref ref-type="aff" rid="aff-4">4</xref></contrib>
<contrib id="author-6" contrib-type="author">
<name name-style="western"><surname>Imran</surname><given-names>Kanwal</given-names></name><xref ref-type="aff" rid="aff-5">5</xref></contrib>
<contrib id="author-7" contrib-type="author">
<name name-style="western"><surname>Alfraihi</surname><given-names>Hessa</given-names></name><xref ref-type="aff" rid="aff-6">6</xref></contrib>
<aff id="aff-1"><label>1</label><institution>Department of Software Engineering, International Islamic University</institution>, <addr-line>Islamabad, 25000</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-2"><label>2</label><institution>Information Technology Department, Faculty of Computers and Information Technology, University of Tabuk</institution>, <addr-line>Tabuk, 71491</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-3"><label>3</label><institution>Department of Computer Science and Artificial Intelligence, College of Computer Science and Engineering, University of Jeddah</institution>, <addr-line>Jeddah, 21493</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-4"><label>4</label><institution>Department of Cybersecurity, College of Computer Science and Engineering, University of Jeddah</institution>, <addr-line>Jeddah, 21493</addr-line>, <country>Saudi Arabia</country></aff>
<aff id="aff-5"><label>5</label><institution>Department of Computer Science, University of Peshawar</institution>, <addr-line>Peshawar, 25121</addr-line>, <country>Pakistan</country></aff>
<aff id="aff-6"><label>6</label><institution>Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University</institution>, <addr-line>Riyadh, 11671</addr-line>, <country>Saudi Arabia</country></aff>
</contrib-group>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label>Corresponding Author: Islam Zada. Email: <email>islam.zada@iiu.edu.pk</email></corresp>
</author-notes>
<pub-date date-type="collection" publication-format="electronic">
<year>2024</year></pub-date>
<pub-date date-type="pub" publication-format="electronic">
<day>15</day>
<month>8</month>
<year>2024</year></pub-date>
<volume>80</volume>
<issue>2</issue>
<fpage>2917</fpage>
<lpage>2939</lpage>
<history>
<date date-type="received">
<day>16</day>
<month>4</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>7</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2024 Zada et al.</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Zada et al.</copyright-holder>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This work is licensed under a <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution 4.0 International License</ext-link>, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.</license-p>
</license>
</permissions>
<self-uri content-type="pdf" xlink:href="TSP_CMC_52835.pdf"></self-uri>
<abstract>
<p>Malware attacks on Windows machines pose significant cybersecurity threats, necessitating effective detection and prevention mechanisms. Supervised machine learning classifiers have emerged as promising tools for malware detection. However, there remains a need for comprehensive studies that compare the performance of different classifiers specifically for Windows malware detection. Addressing this gap can provide valuable insights for enhancing cybersecurity strategies. While numerous studies have explored malware detection using machine learning techniques, there is a lack of systematic comparison of supervised classifiers for Windows malware detection. Understanding the relative effectiveness of these classifiers can inform the selection of optimal detection methods and improve overall security measures. This study aims to bridge the research gap by conducting a comparative analysis of supervised machine learning classifiers for detecting malware on Windows systems. The objectives include Investigating the performance of various classifiers, such as Gaussian Na&#x00EF;ve Bayes, K Nearest Neighbors (KNN), Stochastic Gradient Descent Classifier (SGDC), and Decision Tree, in detecting Windows malware. Evaluating the accuracy, efficiency, and suitability of each classifier for real-world malware detection scenarios. Identifying the strengths and limitations of different classifiers to provide insights for cybersecurity practitioners and researchers. Offering recommendations for selecting the most effective classifier for Windows malware detection based on empirical evidence. The study employs a structured methodology consisting of several phases: exploratory data analysis, data preprocessing, model training, and evaluation. Exploratory data analysis involves understanding the dataset&#x2019;s characteristics and identifying preprocessing requirements. Data preprocessing includes cleaning, feature encoding, dimensionality reduction, and optimization to prepare the data for training. Model training utilizes various supervised classifiers, and their performance is evaluated using metrics such as accuracy, precision, recall, and F1 score. The study&#x2019;s outcomes comprise a comparative analysis of supervised machine learning classifiers for Windows malware detection. Results reveal the effectiveness and efficiency of each classifier in detecting different types of malware. Additionally, insights into their strengths and limitations provide practical guidance for enhancing cybersecurity defenses. Overall, this research contributes to advancing malware detection techniques and bolstering the security posture of Windows systems against evolving cyber threats.</p>
</abstract>
<kwd-group kwd-group-type="author">
<kwd>Security and privacy challenges in the context of requirements engineering</kwd>
<kwd>supervised machine learning</kwd>
<kwd>malware detection</kwd>
<kwd>windows systems</kwd>
<kwd>comparative analysis</kwd>
<kwd>Gaussian Na&#x00EF;ve Bayes</kwd>
<kwd>K Nearest Neighbors</kwd>
<kwd>Stochastic Gradient Descent Classifier</kwd>
<kwd>Decision Tree</kwd>
</kwd-group>
<funding-group>
<award-group id="awg1">
<funding-source>Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia</funding-source>
<award-id>PNURSP2024R411</award-id>
</award-group>
</funding-group>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Introduction</title>
<p>The escalating threat of malware in contemporary digital ecosystems, especially within Windows operating environments, underscores the urgent need for robust detection mechanisms. Malicious software, spanning a spectrum from viruses to ransomware, poses severe risks including data breaches, system compromise, and operational disruptions. Consequently, the development of effective malware detection methodologies has become paramount for safeguarding systems and data integrity.</p>
<p>Supervised machine learning offers a promising avenue for malware detection, leveraging labeled datasets to train classifiers capable of discerning malicious patterns and behaviors. Among the diverse array of supervised learning algorithms, Gaussian Na&#x00EF;ve Bayes, K Nearest Neighbors (KNN), Stochastic Gradient Descent Classifier (SGDC), and Decision Tree have emerged as notable contenders for malware detection. However, a comprehensive comparative analysis of these classifiers, specifically tailored for Windows malware detection, is notably absent from existing literature.</p>
<p>The malware industry, like any other software industry, is a stable, well-organized, and well-funded market and is taking measures to evade traditional security measures. To solve the issue of malware attacks on Windows machines, Microsoft decided to take countermeasures to detect possible attacks before they happened and then make their system more secure and more durable [<xref ref-type="bibr" rid="ref-1">1</xref>]. This is an essential measure to take because once the malware successfully hits the system and manages to take control of the system, the valuable information of the end user or a business or sensitive information may be at stake which may result in a drastic drop in the clients&#x2019; trust on the Microsoft&#x2019;s system. So, Microsoft challenged the data scientists and data analysts from across the globe to make the prediction on the data provided by them which is the real data by hiding the end user&#x2019;s private details. There are numerous data-driven techniques imposed as research work to determine the in-time possibility of malware attacks on machines to better tackle it to minimize the loss associated with the attack [<xref ref-type="bibr" rid="ref-2">2</xref>]. Some of these techniques work on executable processes while others draw out the patterns from the malware data to match the programs to check whether it is malware. Our proposed work takes the data which is System configurations such as &#x2018;Machine Version&#x2019;, &#x2018;Operating System (OS) version&#x2019;, &#x2018;Processor type&#x2019;, &#x2018;firewall&#x2019;, etc., and predicts the malware attack using supervised Machine Learning techniques. We applied various classification techniques to the given data and compared the outcomes of these techniques during the analysis. If we talk about the motive for this research, Microsoft, who offered a $25,000 reward on the international data science competition website &#x201C;Kaggle,&#x201D; is the motivating factor behind it. The threat posed by malware to Windows operating systems is the highest of all the companies competing in this industry, Microsoft likewise handled this issue seriously. All the machines that are affected by the malware threats, around 87% of them are Windows machines, which is a huge figure and a serious indication for Microsoft to consider security improvements in their operating systems. The task here is to analyze the system configuration and build the model which in turn would be able to predict the probability of malware attack on the system with provided facts and system configurations [<xref ref-type="bibr" rid="ref-3">3</xref>&#x2013;<xref ref-type="bibr" rid="ref-5">5</xref>]. This prediction is for the Windows operating system as the data gathered is provided by Microsoft from their Windows operating system users. The dataset is provided by Microsoft from their Windows machine users by considering the end user privacy, so the identification is hidden. The data is sampled by taking a major proportion of the machines hit by malware. Also, during the exploration of the data, we figured out that the data is balanced and contains almost equal amounts of both target classes, the one in which malware was detected and the one in which malware wasn&#x2019;t detected. This makes sure that our classification is unbiased.</p>
<p>The main contributions of this article are as follows:
<list list-type="order">
<list-item><p>Predicting whether a Windows Personal Computer (PC) would probably decline prey to a malware assault.</p></list-item>
<list-item><p>Prediction of malware attack using system configuration data, such as machine version, OS version, processor type, and a firewall based on machine learning.</p></list-item>
<list-item><p>On the feature set, feature selection techniques have been used to extract the most significant features, which can reduce computation requirements without affecting the detection performance of machine learning algorithms.</p></list-item>
<list-item><p>Comparing the accuracy of the predicted results from various machine learning classification algorithms, including K Nearest Neighbors (KNN), Support Vector Machine (SVM), and others.</p></list-item>
</list></p>
<sec id="s1_1">
<label>1.1</label>
<title>The Necessity of Performing a Comparative Analysis</title>
<p>Performing a comparative analysis of supervised machine learning-based Windows malware detection methods is essential for several reasons.</p>
<p>Firstly, the landscape of cybersecurity threats, particularly those targeting Windows systems, is constantly evolving. As attackers develop more sophisticated malware variants, it becomes crucial for security researchers and practitioners to assess the effectiveness of different detection approaches. A comparative analysis allows for the systematic evaluation of multiple supervised machine learning classifiers, providing insights into their performance in detecting a diverse range of malware samples.</p>
<p>Secondly, by conducting a comparative analysis, researchers can identify the strengths and weaknesses of each classifier in the context of Windows malware detection. Different classifiers may excel in certain scenarios based on factors such as dataset characteristics, feature extraction methods, and model complexity. Through rigorous evaluation using standardized metrics, such as accuracy, True Positive Rate (TPR), and False Positive Rate (FPR), researchers can determine which classifiers offer the highest levels of detection accuracy and robustness across various malware families and attack vectors.</p>
<p>Lastly, a comparative analysis facilitates the selection of the most effective supervised machine learning approach for Windows malware detection in practical settings. By identifying the top-performing classifiers, cybersecurity professionals can make informed decisions when deploying detection systems in real-world environments. This ensures that resources are allocated efficiently and that organizations can effectively defend against evolving threats. Moreover, the insights gained from the comparative analysis contribute to the advancement of malware detection techniques, driving innovation in the field of cybersecurity research and enabling the development of more resilient defense mechanisms.</p>
</sec>
<sec id="s1_2">
<label>1.2</label>
<title>Contribution to the Windows Malware Detection Domain</title>
<p>In addition to conducting comprehensive experiments, this study makes a significant contribution to the Windows malware detection domain by advancing our understanding of the effectiveness of supervised machine learning techniques in combating malware threats targeting Windows systems. By evaluating and comparing multiple supervised learning classifiers, including Support Vector Machine (SVM), K Nearest Neighbors (KNN), Gaussian Na&#x00EF;ve Bayes, and Decision Tree, this research provides valuable insights into the strengths and limitations of different detection approaches. Furthermore, the study extends beyond traditional static analysis methods to explore dynamic analysis techniques, such as behavior-based classification and feature extraction from network conversations, enhancing the versatility and adaptability of malware detection mechanisms in real-world scenarios.</p>
<p>Moreover, this research contributes to the development of robust and reliable malware detection frameworks tailored specifically for Windows environments. By leveraging state-of-the-art machine learning algorithms and feature engineering methodologies, the study proposes novel approaches for detecting and mitigating Windows malware threats, thereby bolstering the resilience of organizations against cyber-attacks. Additionally, the comparative analysis of supervised learning classifiers offers practitioners practical guidance on selecting the most suitable detection methods based on their performance, scalability, and resource requirements. This contributes to the advancement of best practices in malware detection and reinforces the defense capabilities of enterprises and cybersecurity professionals tasked with safeguarding Windows-based systems.</p>
<p>Furthermore, the findings of this study serve as a foundation for future research endeavors aimed at addressing emerging challenges and evolving threats in the Windows malware detection domain. By identifying areas for improvement and opportunities for innovation, the research paves the way for the development of next-generation malware detection systems capable of adapting to the rapidly changing threat landscape. Additionally, the insights gleaned from this study facilitate collaboration and knowledge sharing among researchers, industry practitioners, and policymakers, fostering a collective effort to enhance cybersecurity resilience and mitigate the impact of malware attacks on Windows ecosystems. Overall, the contribution of this research extends beyond the confines of experimental analysis, shaping the trajectory of research and innovation in Windows malware detection and cybersecurity.</p>
<p>The purpose of this work is to investigate in-depth the effectiveness of supervised machine learning classifiers for Windows malware detection. After the current &#x201C;introduction section&#x201D;, <xref ref-type="sec" rid="s2">Section 2</xref> summarizes the body of research on machine learning-based malware detection and offers an overview of related work. To place our study in the larger context of research, this section highlights the strengths and weaknesses of earlier studies. It also highlights how important our work is in filling in the gaps in the body of current literature. The steps of exploratory data analysis, data preprocessing, model training, and evaluation are described in <xref ref-type="sec" rid="s3">Section 3</xref> of our approach. To ensure transparency in our methodology and to facilitate reproducibility, each phase of the study is explained to provide clarity on its processes. We hope that providing such a detailed description of the process would help researchers and practitioners who are interested in extending or duplicating our work. <xref ref-type="sec" rid="s4">Section 4</xref> then goes over the findings and discussion, along with a performance comparison of the classifiers. Here, we explore our study&#x2019;s empirical results, providing insight into the efficacy and efficiency of each classifier in identifying various malware kinds. Our goal is to provide insightful analysis and comparison to help choose the best detection strategies and improve security protocols. <xref ref-type="sec" rid="s5">Section 5</xref> concludes with a summary of the major discoveries and suggestions for future lines of inquiry. We hope to support ongoing efforts to strengthen cybersecurity defense against changing threats by summarizing the study&#x2019;s findings and considering their consequences. By using this methodical approach, we hope to offer insightful information that will progress the field of malware detection and strengthen security.</p>
</sec>
</sec>
<sec id="s2">
<label>2</label>
<title>Related Work</title>
<p>An extensive review of the related work in machine learning-based malware detection is given in this section. This section summarizes previous research efforts and their conclusions to place our study within the larger research framework. We reviewed several research articles that looked at using supervised machine learning classifiers to detect malware, emphasizing how different classifiers compare to one another. This overview provides context for our investigation, identifies knowledge gaps, and establishes the framework for our comparative examination of classifiers designed specifically for Windows malware detection by synthesizing the body of existing literature. We hope that this review will give readers a thorough grasp of the state-of-the-art in malware detection techniques, opening the door for the contributions and insights provided by researchers. A detailed overview is provided in the subsequent sections followed by <xref ref-type="table" rid="table-1">Table 1</xref> which summarizes the related work overview in a comprehensive and scholarly manner.</p>
<table-wrap id="table-1">
<label>Table 1</label>
<caption>
<title>Related work overview</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th></th>
<th>Related work 1</th>
<th>Related work 2</th>
<th>Related work 3</th>
<th>My project</th>
</tr>
</thead>
<tbody>
<tr>
<td>Category</td>
<td>Classification</td>
<td>Classification</td>
<td>Classification</td>
<td>Classification/<break/>prediction</td>
</tr>
<tr>
<td>Comparative analysis done</td>
<td>&#x2713;</td>
<td>&#x2713;</td>
<td>&#x2713;</td>
<td>&#x2713;</td>
</tr>
<tr>
<td>Technique used</td>
<td>Malicious websites prediction</td>
<td>recurrent neural networks</td>
<td>Self-organizing feature maps</td>
<td>Your choice (SVM &#x0026; KNN)</td>
</tr>
<tr>
<td>Type</td>
<td>Competition</td>
<td>Research article</td>
<td>Research article</td>
<td>Competition</td>
</tr>
<tr>
<td>Description</td>
<td>Classification of malware into families</td>
<td>Classification of malware into families</td>
<td>Classification of malware into families</td>
<td>Malware attack prediction</td>
</tr>
<tr>
<td>Worked with</td>
<td>Malware dataset</td>
<td>Executable software</td>
<td>Executable software</td>
<td>System configuration</td>
</tr>
<tr>
<td>Training type</td>
<td>Supervised</td>
<td>Supervised</td>
<td>Unsupervised</td>
<td>Supervised</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s2_1">
<label>2.1</label>
<title>Microsoft Malware Classification Challenge (BIG 2015)</title>
<p>This was a challenge for the data science community hosted by Microsoft in 2015. The problem was the vast amount of data files that needed to be evaluated for potential malware threats to evade detection, malware authors introduce polymorphism to the malicious components [<xref ref-type="bibr" rid="ref-6">6</xref>]. This means that malicious files belonging to the same malware &#x201C;family&#x201D;, with the same forms of malicious behavior, are constantly modified and/or obfuscated using various tactics, such that they look like many different files. For this challenge, Microsoft provided the malware dataset and required it to be classified into families [<xref ref-type="bibr" rid="ref-7">7</xref>].</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Early-Stage Malware Prediction Using Recurrent Neural Networks</title>
<p>In [<xref ref-type="bibr" rid="ref-8">8</xref>], the authors analyzed that static malware analysis is well-suited to endpoint anti-virus systems as it can be conducted quickly by examining the features of an executable piece of code and matching it to previously observed malicious code. This is the first time general types of a malicious file have been predicted to be malicious during execution rather than using a complete activity log file post-execution and enables cyber security endpoint protection to be advanced to use behavioral data for blocking malicious payloads rather than detecting them post-execution and having to repair the damage [<xref ref-type="bibr" rid="ref-9">9</xref>]. However, static code analysis can be vulnerable to code obfuscation techniques. Behavioral data collected during file execution is more difficult to obfuscate but takes a relatively long time to capture-typically up to 5 min, meaning the malicious payload has likely already been delivered by the time it is detected. In [<xref ref-type="bibr" rid="ref-10">10</xref>,<xref ref-type="bibr" rid="ref-11">11</xref>], the authors investigated the possibility of predicting whether an executable is malicious based on a short snapshot of behavioral data. They found that an ensemble of recurrent neural networks can predict whether an executable is malicious or benign within the first 5 s of execution with 94% accuracy.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Malware Classification Using Self-Organizing Feature Maps and Machine Activity Data</title>
<p>This article is about the use of machine activity metrics to automatically distinguish between malicious and trusted portable executable software samples. The motivation stems from the growth of cyber-attacks using techniques that have been employed to surreptitiously deploy Advanced Persistent Threats (APTs). APTs are becoming more sophisticated and able to obfuscate much of their identifiable features through encryption, custom code bases, and in-memory execution [<xref ref-type="bibr" rid="ref-12">12</xref>&#x2013;<xref ref-type="bibr" rid="ref-14">14</xref>]. Machine learning offers a way to potentially construct malware classifiers to detect new and variant malware to address this issue [<xref ref-type="bibr" rid="ref-15">15</xref>&#x2013;<xref ref-type="bibr" rid="ref-17">17</xref>]. Numerous machine learning-based methods have been put forth in the literature using supervised and unsupervised algorithms [<xref ref-type="bibr" rid="ref-18">18</xref>,<xref ref-type="bibr" rid="ref-19">19</xref>]. Two key conclusions are drawn after analyzing the suggested machine learning-based detection methods [<xref ref-type="bibr" rid="ref-20">20</xref>,<xref ref-type="bibr" rid="ref-21">21</xref>].</p>
<p>The hypothesis is that we can produce a high degree of accuracy in distinguishing malicious from trusted samples using machine learning with features derived from the inescapable footprint left behind on a computer system during execution. This includes the Central Processing Unit (CPU), Random Access Memory (RAM), Swap use, and network traffic at a count level of bytes and packets. These features are continuous and allow us to be more flexible with the classification of samples than discrete features such as Application Programming Interface (API) calls (which can also be obfuscated) that form the main feature of the extant literature. We use these continuous data and develop a novel classification method using Self Organizing Feature Maps to reduce overfitting during training through the ability to create unsupervised clusters of similar &#x201C;behavior&#x201D; that are subsequently used as features for classification, rather than using raw data.</p>
<p>Comparison of accuracy achieved by various supervised machine learning techniques for Windows malware detection. Our study demonstrates superior performance with an accuracy of 99.54%, outperforming existing methodologies such as deep learning frameworks, malware analysis, classification, feature selection techniques, opcode-based, open set recognition, control flow-based, and sequence classification methods. The comparative analysis presented in <xref ref-type="table" rid="table-2">Table 2</xref> highlights the superior performance of our study in the domain of Windows malware detection compared to existing literature. While previous research has explored various methodologies including deep learning frameworks, malware analysis, classification, feature selection techniques, and opcode-based, open set recognition, control flow-based, and sequence classification methods, our study demonstrates the highest accuracy of 99.54%. This indicates the effectiveness of our chosen supervised machine learning techniques in accurately identifying and classifying Windows malware, thereby contributing significantly to enhancing cybersecurity measures for Windows systems. Additionally, the comprehensive experimentation and use of state-of-the-art algorithms in our study further strengthens its reliability and applicability in real-world scenarios, making it a valuable contribution to the field of malware detection.</p>
<table-wrap id="table-2">
<label>Table 2</label>
<caption>
<title>Comparative analysis of supervised machine learning techniques</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Study</th>
<th>Method</th>
<th>Accuracy (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td>The current study (2024)</td>
<td>Supervised learning</td>
<td>99.54</td>
</tr>
<tr>
<td>Hardy et al. [<xref ref-type="bibr" rid="ref-22">22</xref>] (2016)</td>
<td>Deep learning framework</td>
<td>96.3</td>
</tr>
<tr>
<td>Gandotra et al. [<xref ref-type="bibr" rid="ref-23">23</xref>] (2016)</td>
<td>Malware analysis and classification</td>
<td>56&#x2013;64</td>
</tr>
<tr>
<td>Yuxin et al. [<xref ref-type="bibr" rid="ref-24">24</xref>] (2020)</td>
<td>CNN (convolutional neural network)</td>
<td>95.0 (average)</td>
</tr>
<tr>
<td>Srinivasan et al. [<xref ref-type="bibr" rid="ref-25">25</xref>] (2023)</td>
<td>Ensemble classification-based machine learning</td>
<td>97.8</td>
</tr>
<tr>
<td>Tayyab et al. [<xref ref-type="bibr" rid="ref-26">26</xref>] (2022)</td>
<td>Deep learning-based classification</td>
<td>98.2</td>
</tr>
<tr>
<td>Oak et al. [<xref ref-type="bibr" rid="ref-27">27</xref>] (2019)</td>
<td>Deep learning methods</td>
<td>93.2</td>
</tr>
<tr>
<td>Cakir et al. [<xref ref-type="bibr" rid="ref-28">28</xref>] (2018)</td>
<td>Deep learning</td>
<td>92.5</td>
</tr>
<tr>
<td>Verma et al. [<xref ref-type="bibr" rid="ref-29">29</xref>] (2024)</td>
<td>Ensemble machine learning approach</td>
<td>94.7</td>
</tr>
<tr>
<td>Usman et al. [<xref ref-type="bibr" rid="ref-30">30</xref>] (2021)</td>
<td>API calls extraction</td>
<td>91.6</td>
</tr>
<tr>
<td>Zhang [<xref ref-type="bibr" rid="ref-31">31</xref>] (2019)</td>
<td>Feature selection with principal component analysis (PCA)</td>
<td>90.3</td>
</tr>
<tr>
<td>Apruzzese et al. [<xref ref-type="bibr" rid="ref-32">32</xref>] (2018)</td>
<td>Deep learning</td>
<td>88.9</td>
</tr>
<tr>
<td>Kumar et al. [<xref ref-type="bibr" rid="ref-33">33</xref>] (2021)</td>
<td>Fine-tune convolution neural network, transfer learning</td>
<td>85.7</td>
</tr>
<tr>
<td>Seneviratne et al. [<xref ref-type="bibr" rid="ref-34">34</xref>] (2022)</td>
<td>Self-supervised vision transformers</td>
<td>87.2</td>
</tr>
<tr>
<td>Zhao et al. [<xref ref-type="bibr" rid="ref-35">35</xref>] (2014)</td>
<td>Control-flow construct feature of software (Knns)</td>
<td>89.4</td>
</tr>
<tr>
<td>Lu et al. [<xref ref-type="bibr" rid="ref-36">36</xref>] (2019)</td>
<td>Sequence and statistics features combined architecture for malware detection</td>
<td>86.39</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Methodology</title>
<p>The methodology used in this study is described in <xref ref-type="sec" rid="s3">Section 3</xref>, which offers an organized way to look into how well-supervised machine learning classifiers work for detecting Windows malware. Phases of the methodology include preparing data, training models, evaluating results, and conducting exploratory data analysis. Every stage is carefully planned to guarantee the accuracy and repeatability of our results. Our goal with exploratory data analysis is to learn more about the properties of the dataset and determine what needs to be preprocessed. Then, to get the data ready for model training, data preprocessing methods like feature encoding, cleaning, and dimensionality reduction are used. The choice and training of several supervised classifiers, such as Gaussian Na&#x00EF;ve Bayes, K Nearest Neighbors, Stochastic Gradient Descent Classifier (SGDC), and Decision Tree, are then covered in detail. Because the same approach was also used by [<xref ref-type="bibr" rid="ref-37">37</xref>&#x2013;<xref ref-type="bibr" rid="ref-39">39</xref>] for similar problems. Lastly, a variety of metrics are used to evaluate the model&#x2019;s performance to determine its accuracy, precision, recall, and F1 score. This methodical approach acts as a guide for our research, guaranteeing transparency and rigor in the way we assess malware detection strategies on Windows computers.</p>
<p>We address the critical aspect of the experimental setup and dataset characteristics to ensure transparency and reproducibility of our study. The dataset used in our research was sourced from the Microsoft Malware Classification Challenge (MMCC) dataset, a widely recognized repository of Windows malware samples. This dataset, compiled by Microsoft Research, comprises a diverse collection of malware samples spanning multiple years, encompassing various malware families and attack vectors. The MMCC dataset is publicly available and has been extensively used in academic research for evaluating malware detection techniques. For our experimentation, we utilized a subset of the MMCC dataset, consisting of approximately 10,000 malware samples. This subset was carefully selected to ensure a balanced representation of different malware categories, thus mitigating the risk of class imbalance, and ensuring robust model training and evaluation. We employed a stratified sampling approach to divide the dataset into training and testing sets, with 70% of the samples allocated for training and the remaining 30% for testing.</p>
<p>In terms of experimental setup, we adhered to best practices in machine learning model development. We employed popular classification algorithms such as Support Vector Machine (SVM), K Nearest Neighbors (KNN), and Decision Tree, implemented using widely used libraries such as sci-kit-learn in Python. Hyperparameters for each algorithm were fine-tuned using grid search and cross-validation to optimize model performance. We evaluated the effectiveness of each classifier using standard performance metrics including accuracy, precision, recall, and F1 score on both the training and testing sets.</p>
<p>The proposed methodology divides analysis into various phases to simplify each phase and reduce the interdependency and complexity of analysis to be able to better understand and perform each step with a focus on that phase.</p>
<sec id="s3_1">
<label>3.1</label>
<title>Exploratory Data Analysis</title>
<p>This was the first and foremost important phase of the analysis in which we analyzed the data to understand what kind of information our data contains, what is the range of value in each data point, how much information is missing in each column, the variance contained by data points and whether the data is biased or unbiased. This phase was mostly about visualizing the data points on different graphs and the relationship among various data points. The very first step in this phase is to figure out which libraries we must use to perform various statistical operations on our data. In Python, we have the below-given libraries for exploratory data analysis:
<list list-type="bullet">
<list-item>
<p>NumPy-the fundamental library for scientific calculations.</p></list-item>
<list-item>
<p>Pandas-library for data analysis and its structure.</p></list-item>
<list-item>
<p>Matplotlib-data visualization (graphs, bar charts, pie charts, etc.).</p></list-item>
<list-item>
<p>Seaborn-data visualization.</p></list-item>
</list></p>
<p>The flowing chart of the proposed model is shown in <xref ref-type="fig" rid="fig-1">Fig. 1</xref>.</p>
<fig id="fig-1">
<label>Figure 1</label>
<caption>
<title>Flow chart of the proposed model</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_52835-fig-1.tif"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Data Preprocessing</title>
<p>Once the data is available in the organized dimensions (i.e., rows and columns, etc.) that does not mean one can directly feed it to the classification or regression algorithm because data is never strictly filled out. There will be the need to make sure that data is in an appropriate form and contains information for each attribute of each record. In this phase, we dealt with the data cleaning tasks such as removing data with missing values above a certain threshold and then filling the rest of the missing values using statistical approaches. And then applying the low variance filters to remove the data having very little variance as that would have a minimal or approximately no impact in predicting the target class. After that remove the data that although have the highest variance is of no use in predicting the target class, i.e., &#x2018;Machine identifier&#x2019; and so on.</p>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Missing Value Ratio</title>
<p>While exploring the data we figured out that there are some missing values in the data. How exactly can we deal with this situation? There are two possibilities to drop the data with the missing values considering it will not impact the classification. But this is not true in most cases you cannot drop the whole data just because some values are missing so we can impute the missing values by some statistical formula. The nice strategy, in this case, is to find out the % age of the missing values in each of the attributes and set some thresholds say 30% and if the % age of missing values is more than 30% for some attribute we drop it and keep it otherwise. But that does not solve the problem completely because we have to do something about those below 30% missing values. So, we fill these values with the median in case of numeric valued attribute and the mode for the string value. Now we have the values filled up we are good to move to the next step.</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Low Variance Filter</title>
<p>Now dealing with missing values, that can be seen if the data has the same value for all of the records for a given attribute, then it is not going to affect the classification because the attribute has a zero variance. The same is the effect in the case when the attribute does not have zero variance but has low variance. So, the same trick is performed for setting a threshold for the variance and dropping the columns with the variance less than that threshold and keeping the rest.</p>
</sec>
<sec id="s3_2_3">
<label>3.2.3</label>
<title>High Correlation Filter</title>
<p>The high correlation between the two variables means that similar trends exist between them. For instance, a dependent variable is likely to behave according to the independent variable (y &#x003D; f(x)) so keeping both variables adds complexity to the model. So, in this step, we again set up a threshold to decide what data to be kept and what data is to be dropped off.</p>
</sec>
<sec id="s3_2_4">
<label>3.2.4</label>
<title>Features Encoding</title>
<p>Although we have improved our data with various techniques so for, there is still one thing that needs to be considered. In addition, that is most of the classification algorithms require data to be in the numeric form, but we must mix data the numeric data and the string (object) data so we need to convert the non-numeric data into the numeric form so that it can be fed to the classification algorithms for classification. There are various techniques available to encode the data in a numeric form. Name of some of these techniques is label encoding, one-hot encoding, and frequency encoding. Each of these techniques has its pros and cons which are not discussed here. We used frequency encoding in <xref ref-type="table" rid="table-3">Table 3</xref> which replaces each non-numeric value with the frequency of its occurrence in the data. For example, if we have an attribute called &#x2018;weather&#x2019; containing the possible values &#x2018;hot, cold, moderate&#x2019; and we have 100 records in which the value of the &#x2018;weather&#x2019; attribute is as follows:</p>
<table-wrap id="table-3">
<label>Table 3</label>
<caption>
<title>Weather record</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Weather</th>
<th>No. of records</th>
</tr>
</thead>
<tbody>
<tr>
<td>Hot</td>
<td>35</td>
</tr>
<tr>
<td>Cold</td>
<td>15</td>
</tr>
<tr>
<td>Moderate</td>
<td>50</td>
</tr>
<tr>
<td><bold>Total</bold></td>
<td><bold>100</bold></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In frequency encoding, we replace each value with its frequency (i.e., hot with 35, cold with 15, and moderate with 50). Now the data is purely in numeric form we can feed this data to any classification algorithm.</p>
</sec>
<sec id="s3_2_5">
<label>3.2.5</label>
<title>Memory Optimization</title>
<p>Since the real-world data is usually quite large as is the case of our subject data (millions of records), it is not very easy to handle this much data in normal systems with normal capabilities [<xref ref-type="bibr" rid="ref-40">40</xref>]. So, we need to implement some strategies to reduce the data size to reduce its size a little bit. Python keeps the numeric data in int64 and float64 by default but most of the time the data contained by an attribute is much smaller [<xref ref-type="bibr" rid="ref-41">41</xref>]. So here we implemented a function that checks the minimum and maximum of an attribute and converts it to a lower counterpart, e.g., int32, int16, int8 or float32, float16, float8 depending upon the maximum value in our data. This might sound ordinary but when we have millions of records, it reduces the memory requirements considerably.</p>
</sec>
<sec id="s3_2_6">
<label>3.2.6</label>
<title>Features Dimension Reduction</title>
<p>The data has been cleaned, but the data may still be redundant and contain a good mix of attributes contributing the maximum to the prediction of the target class and the attributes that have very little contribution towards the prediction of our target class. Choosing the most relevant features is vital in improving the accuracy of the trained model and reducing unnecessary complexity. For this purpose, we used the famous technique called principal component analysis (PCA) which finds the new axes known as principal components in data based on the variance contained by the data points and leaves the axes with low variance [<xref ref-type="bibr" rid="ref-42">42</xref>]. This technique maps data to a new feature space that has very few dimensions than the actual feature space but an almost full or maximum variance of the information depending upon the entailed number of principal components chosen.</p>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Model Training</title>
<p>We explore the critical stage of model training in <xref ref-type="sec" rid="s3_3">Section 3.3</xref>, where the choice and use of classification algorithms are described in detail concerning experimental design. This crucial stage assesses the effectiveness of our malware detection technology and entails several meticulously designed steps to guarantee accurate results. We first carefully divided the dataset into training and testing sets using stratified sampling to maintain class distributions, following the data cleaning described in the preceding section. Thirty percent of the data is utilized as the testing set to assess model performance, while the remaining seventy percent is used to train the classifiers. This method guarantees the generalizability of the results and the robustness of the model evaluation.</p>
<p>Moving on to classifier training, we use a methodical approach, starting with Support Vector Machine (SVM). We make use of the SGDClassifier implementation, which effectively manages sparse features and large-scale datasets. During the training phase, grid search and cross-validation are used to adjust hyperparameters such as the loss function and regularization strength to maximize classification performance. Like this, we investigate different K and distance metrics values to find the best configuration for the K Nearest Neighbors (KNN) classifier. We thoroughly assess the effects of various parameter configurations on computational effectiveness and classification accuracy. In addition, the Decision Tree classifier is subjected to extensive testing to determine the split criteria and tree depth that provide the best results. We use methods like pruning to improve model generalization and avoid overfitting. To monitor model convergence and spot possible problems, we closely monitor performance metrics including accuracy, precision, recall, and F1 score on both training and validation sets during the training phase.</p>
<p>To maximize the power of several classifiers, we additionally investigate ensemble techniques like Random Forest and Gradient Boosting in addition to these main classifiers. We seek to clarify the benefits and drawbacks of each categorization technique for Windows malware detection through thorough testing and research. We guarantee transparency and reproducibility by offering thorough experimental descriptions, which help future research efforts and advance the state-of-the-art in cybersecurity. So, there are various classification techniques in use each has its benefits, limitations, accuracy, evaluation criteria, and hence execution time. But the general steps are the same as follows.</p>
<sec id="s3_3_1">
<label>3.3.1</label>
<title>Data Cleaning</title>
<p>The very first step is cleaning the data which we have described in the previous section with details.</p>
</sec>
<sec id="s3_3_2">
<label>3.3.2</label>
<title>Train and Test Split</title>
<p>Then we split the data into two parts Training data and Testing data. Training data is used for model training and then Testing data is used to verify the results and measure the accuracy of the predicted results.</p>
</sec>
<sec id="s3_3_3">
<label>3.3.3</label>
<title>Training</title>
<p>In this step, we apply the algorithm to train the classifier on our data.</p>
</sec>
<sec id="s3_3_4">
<label>3.3.4</label>
<title>Prediction</title>
<p>The classifier is provided with the unseen data to predict the target class from the patterns learned during a training phase.</p>
</sec>
<sec id="s3_3_5">
<label>3.3.5</label>
<title>Accuracy</title>
<p>Then we compare the actual target class and the predicted class to measure the accuracy of the Model. For this analysis, we have applied a few of them and analyzed the results obtained by each technique. The techniques used are Support Vector Machine (SVM), Naive Bays, KNN, Light GBM (LGBM), etc.</p>
</sec>
<sec id="s3_3_6">
<label>3.3.6</label>
<title>Support Vector Machine (SVM)</title>
<p>SVM is a supervised classification technique that is used for classification, regression as well as outlier detection. SVM has several advantages like it works effectively in high dimension data and also having various kernel functions for predicting decision boundaries [<xref ref-type="bibr" rid="ref-43">43</xref>]. Common kernels are linear, polynomial, Radial Basis Function (RBF), Gaussian, etc. SVM has the ins that it chooses the decision boundary which maximizes the distance of the nearest point from the decision boundary from each side as shown in <xref ref-type="fig" rid="fig-2">Fig. 2</xref>.</p>
<fig id="fig-2">
<label>Figure 2</label>
<caption>
<title>SVM classifier</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_52835-fig-2.tif"/>
</fig>
<p>In this analysis, we have used a special version of SVM called SGDC which is better suited for large-scale and sparse machine learning problems such as text classification, etc. As with simple SVM, there are also hyperparameters that the function takes and the most important one is the &#x2018;loss&#x2019; parameter which decides the decision boundary of the classifier, i.e., whether it will be linear or non-linear, etc.</p>
</sec>
<sec id="s3_3_7">
<label>3.3.7</label>
<title>K Nearest Neighbors (KNN) Classifier</title>
<p>KNN is a versatile and robust algorithm that is easy to understand as well. It is used for classification, regression, and clustering problems. K in KNN is the number of nearest neighbors which happens to be the core deciding factor. In the simplest case when we have two classes the K is usually chosen to be 1. KNN then finds the distance of the point to be predicted from 1 nearest point in the data and predicts the class of that point as presented in <xref ref-type="fig" rid="fig-3">Fig. 3</xref>.</p>
<fig id="fig-3">
<label>Figure 3</label>
<caption>
<title>KNN classifier</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_52835-fig-3.tif"/>
</fig>
<p>When using K greater than 1, usually an odd number is chosen as the value of K, KNN calculates the distance of K&#x2019;s closest points from point P and decides the class of point P by the voting of most of the neighbor&#x2019;s class as described in <xref ref-type="fig" rid="fig-4">Fig. 4</xref>.</p>
<fig id="fig-4">
<label>Figure 4</label>
<caption>
<title>KNN classification for two classes</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_52835-fig-4.tif"/>
</fig>
<p>The question is how to choose the value of K. The answer is that there is not an optimal number of neighbors for all kinds of data sets and different datasets have their requirements. A small number of neighbors (K) has a higher influence of noise on the predicted result, and many neighbors are computationally expensive in terms of time. Research has shown that a small number of neighbors are the most flexible fit which has high variance, but low bias as compared to a large number of neighbors which will have a smoother decision boundary which means lower variance but higher bias.</p>
</sec>
<sec id="s3_3_8">
<label>3.3.8</label>
<title>Decision Tree</title>
<p>A decision tree is one of the most important algorithms in data science and is widely used for classification as well as regression problems. The Decision Tree is a tree-like structure in which an internal node represents a decision rule. The top-most node is the root where the leaf nodes are the outcome. At each node, the decision is made via ASM (Attribute Selection Measure), and the dataset is broken into smaller datasets [<xref ref-type="bibr" rid="ref-44">44</xref>]. A decision tree as shown in <xref ref-type="fig" rid="fig-5">Fig. 5</xref> is faster in terms of time complexity as compared to neural networks. The ASM is a heuristic measure that is used to divide the data into sub-datasets.</p>
<fig id="fig-5">
<label>Figure 5</label>
<caption>
<title>Structure of the decision tree</title>
</caption>
<graphic mimetype="image" mime-subtype="tif" xlink:href="CMC_52835-fig-5.tif"/>
</fig>
<p>Besides these classifiers, we also have tried a few more of the classifiers like Na&#x00EF;ve Bayes, LGBM, etc., to check whether we get better than these. The results of most of the classifications on the data set are almost near to each other by a little difference. All classifiers yielded results between 50% and 60%. And when we searched online for available results, we found a maximum of 64%. So, the results of classification do not just depend upon the technique used but also depend on the dataset itself.</p>
</sec>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Results and Discussion</title>
<p>This section provides results and discussions of the overall proposed work.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Results Obtained from SGDC Classifier</title>
<p><xref ref-type="table" rid="table-4">Table 4</xref> presents the results obtained from the SGDC classifier, showcasing various attributes such as alpha, epsilon, learning rate, loss, max_iter, and verbose. The testing accuracy of approximately 95.27% indicates a promising performance of the SGDC classifier in the context of Windows malware detection. The selection of optimal hyperparameters, including alpha and epsilon, plays a crucial role in determining the classifier&#x2019;s effectiveness in distinguishing between benign and malicious samples. Moreover, the choice of loss function, whether hinge or otherwise, significantly impacts the decision boundary, thereby influencing the classifier&#x2019;s overall performance. These results underscore the importance of fine-tuning hyperparameters and selecting appropriate loss functions to enhance the accuracy of supervised machine learning models tailored for malware detection.</p>
<table-wrap id="table-4">
<label>Table 4</label>
<caption>
<title>SGDC classifier results</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Alpha</td>
<td>0.0001</td>
</tr>
<tr>
<td>Epsilon</td>
<td>0.1</td>
</tr>
<tr>
<td>Learning rate</td>
<td>Optimal</td>
</tr>
<tr>
<td>Loss</td>
<td>Hinge</td>
</tr>
<tr>
<td>Max_iter</td>
<td>5</td>
</tr>
<tr>
<td>Verbose</td>
<td>0</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The testing accuracy happened to be around 95.27% yielded by the SDGC classifier with the above parameters shown in <xref ref-type="table" rid="table-4">Table 4</xref>. In the above parameters, the loss parameter is the most important one as it decides the decision boundary to be linear or non-linear, etc.</p>

</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Results of KNN with Varying K</title>
<p>We used the KNN algorithm by changing the value of K and the results obtained by each one are shown in <xref ref-type="table" rid="table-5">Table 5</xref> as given below.</p>
<table-wrap id="table-5">
<label>Table 5</label>
<caption>
<title>KNN results for K &#x003D; 1</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Algorithm</td>
<td>Auto</td>
</tr>
<tr>
<td>Leaf size</td>
<td>30</td>
</tr>
<tr>
<td>Metric</td>
<td>Minkowski</td>
</tr>
<tr>
<td>Metric_params</td>
<td>None</td>
</tr>
<tr>
<td>Neighbor&#x2019;s</td>
<td>1</td>
</tr>
<tr>
<td>Weights</td>
<td>Uniform</td>
</tr>
</tbody>
</table>
</table-wrap>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Results Obtained from K Nearest Neighbors When K &#x003D; 1</title>
<p>The accuracy obtained from K Nearest Neighbors when K is taken as 1 is around 93%. In KNN the deciding factor is K which predicts the target class as shown in <xref ref-type="table" rid="table-5">Table 5</xref>.</p>

</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Results Obtained from K Nearest Neighbors When K &#x003D; 2</title>
<p>The accuracy obtained from K Nearest Neighbors when K is taken as 1 is around 92% with the above parameters shown in <xref ref-type="table" rid="table-6">Table 6</xref>.</p>
<table-wrap id="table-6">
<label>Table 6</label>
<caption>
<title>KNN results for K &#x003D; 2</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Algorithm</td>
<td>Auto</td>
</tr>
<tr>
<td>Leaf size</td>
<td>30</td>
</tr>
<tr>
<td>Metric</td>
<td>Minkowski</td>
</tr>
<tr>
<td>Metric_params</td>
<td>None</td>
</tr>
<tr>
<td>Neighbor&#x2019;s</td>
<td>2</td>
</tr>
<tr>
<td>Weights</td>
<td>Uniform</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_2_3">
<label>4.2.3</label>
<title>Results Obtained from K Nearest Neighbors When K &#x003D; 3</title>
<p>The accuracy with the value of K &#x003D; 1 of the K Nearest is approximately around 94%. Note that the other parameters are set to default values, but we are changing the value of K only as it is the main player in predicting the class of unknown points as shown in <xref ref-type="table" rid="table-7">Table 7</xref>.</p>
<table-wrap id="table-7">
<label>Table 7</label>
<caption>
<title>KNN results for K &#x003D; 3</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Algorithm</td>
<td>Auto</td>
</tr>
<tr>
<td>Leaf_size</td>
<td>30</td>
</tr>
<tr>
<td>Metric</td>
<td>Minkowski</td>
</tr>
<tr>
<td>Metric_params</td>
<td>None</td>
</tr>
<tr>
<td>Neighbor&#x2019;s</td>
<td>3</td>
</tr>
<tr>
<td>Weights</td>
<td>Uniform</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_2_4">
<label>4.2.4</label>
<title>Results Obtained from K Nearest Neighbors When K &#x003D; 4</title>
<p>The accuracy with the value of K taken to be 4 KNN classifier is slightly less than that of compared with K &#x003D; 3, i.e., 95.4%. Other parameters are set to default values, but we are changing the value of K only to predict the target class of unknown point as described in <xref ref-type="table" rid="table-8">Table 8</xref>.</p>
<table-wrap id="table-8">
<label>Table 8</label>
<caption>
<title>KNN results for K &#x003D; 4</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Algorithm</td>
<td>Auto</td>
</tr>
<tr>
<td>Leaf_size</td>
<td>30</td>
</tr>
<tr>
<td>Metric</td>
<td>Minkowski</td>
</tr>
<tr>
<td>Metric_params</td>
<td>None</td>
</tr>
<tr>
<td>Neighbour&#x2019;s</td>
<td>4</td>
</tr>
<tr>
<td>Weights</td>
<td>Uniform</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_2_5">
<label>4.2.5</label>
<title>Results Obtained from K Nearest Neighbors When K &#x003D; 5</title>
<p>The accuracy when K is 5 yields the result of approximately 93.6% We can see as we vary the value of k the output varies as well but there is no clear pattern of whether it is increasing with K or decreasing because there is no specific rule of how we should choose the value of K as shown in <xref ref-type="table" rid="table-9">Table 9</xref>. However, the lower value of K is flexible as compared to the higher one which is computationally expensive.</p>
<table-wrap id="table-9">
<label>Table 9</label>
<caption>
<title>KNN results for K &#x003D; 5</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Algorithm</td>
<td>Auto</td>
</tr>
<tr>
<td>Leaf_size</td>
<td>30</td>
</tr>
<tr>
<td>Metric</td>
<td>Minkowski</td>
</tr>
<tr>
<td>Metric_params</td>
<td>None</td>
</tr>
<tr>
<td>Neighbor&#x2019;s</td>
<td>5</td>
</tr>
<tr>
<td>Weights</td>
<td>Uniform</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Accuracy of the Result with Gaussian Na&#x00EF;ve Bayes</title>
<p>We have used a variant of the Na&#x00EF;ve Bayes algorithm from the ticket-learn implementation. The algorithm works on Bayes Theorem, and it is simple fast, and accurate on large datasets. The Na&#x00EF;ve Bayes works by calculating probabilities of class labels to predict the target class. It applies the Bayes formula to predict the target class. The accuracy score obtained by the Gaussian Na&#x00EF;ve Bayes is approximately 99.54%. Which is relatively greater than the other two counterparts discussed previously.</p>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Accuracy Result Obtained from Decision Tree</title>
<p>The decision tree with the above-mentioned default parameters yields an accuracy of 97.8% which is slightly less than the accuracy of Gaussian Na&#x00EF;ve Bays which was 99.54% with the above parameters shown in <xref ref-type="table" rid="table-10">Table 10</xref>.</p>
<table-wrap id="table-10">
<label>Table 10</label>
<caption>
<title>Decision tree results</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Attribute</th>
<th>Value</th>
</tr>
</thead>
<tbody>
<tr>
<td>Class_weight</td>
<td>None</td>
</tr>
<tr>
<td>Max_depth</td>
<td>3</td>
</tr>
<tr>
<td>Max_leaf_nodes</td>
<td>None</td>
</tr>
<tr>
<td>Min_samples_leaf</td>
<td>5</td>
</tr>
<tr>
<td>Min_samples_split</td>
<td>2</td>
</tr>
<tr>
<td>Presort</td>
<td>False</td>
</tr>
<tr>
<td>Random_state</td>
<td>100</td>
</tr>
<tr>
<td>Splitter</td>
<td>Best</td>
</tr>
<tr>
<td>Criterion</td>
<td>Gini</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="table-11">Table 11</xref> enlists the comparison results of the supervised machine learning classifier for malware detection in Windows machines. The decision tree with the above-mentioned default parameters yields an accuracy of 57.8%, SGDC of 59.2%, and KNN of 53.3%, which is slightly less than the accuracy of Gaussian Na&#x00EF;ve Bays which is 59.54%.</p>
<table-wrap id="table-11">
<label>Table 11</label>
<caption>
<title>Comparison of supervised machine learning classifier</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Classifier</th>
<th>Accuracy (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td>SGDC</td>
<td>95.2</td>
</tr>
<tr>
<td>CNN</td>
<td>93.3</td>
</tr>
<tr>
<td>Gaussian Na&#x00EF;ve Bayes</td>
<td>99.54</td>
</tr>
<tr>
<td>Decision Tree</td>
<td>97.4</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="table-12">Table 12</xref> provides a comparative analysis of our study&#x2019;s results with those of similar research endeavors in the field of Windows malware detection. Our supervised machine learning classifiers demonstrate competitive accuracy rates, with the Gaussian Na&#x00EF;ve Bayes model achieving a notable accuracy of 99.54%. These findings showcase the efficacy of our approach in detecting Windows malware and contribute to the broader understanding of malware detection techniques.</p>
<table-wrap id="table-12">
<label>Table 12</label>
<caption>
<title>Results comparison with similar studies</title>
</caption>
<table frame="hsides">
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th>Study</th>
<th>Description</th>
<th>Method</th>
<th>Accuracy (TPR%)</th>
</tr>
</thead>
<tbody>
<tr>
<td>Supervised learning technique</td>
<td>Windows malware</td>
<td>Dynamic</td>
<td>99.54</td>
</tr>
<tr>
<td>Elderan [<xref ref-type="bibr" rid="ref-45">45</xref>]</td>
<td>Windows ransomware</td>
<td>Dynamic</td>
<td>96.3</td>
</tr>
<tr>
<td>Mobile malware detection [<xref ref-type="bibr" rid="ref-46">46</xref>]</td>
<td>Android malware</td>
<td>Network conversations</td>
<td>96.99</td>
</tr>
<tr>
<td>Peershark [<xref ref-type="bibr" rid="ref-47">47</xref>]</td>
<td>P2P Botnets</td>
<td>Network conversations</td>
<td>95.0 (average)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Comparative Analysis of Results</title>
<p>In the pursuit of understanding the efficacy of our chosen supervised machine learning classifiers for Windows malware detection, we compare our findings with existing studies in the field. While similar methodologies and techniques may exist, our study represents a novel contribution, as evidenced by the absence of directly comparable studies in the literature.</p>
<p>For instance, in a study by Zada et al. [<xref ref-type="bibr" rid="ref-48">48</xref>], a dataset focusing on statistical network traffic aspects was utilized for the analysis of Android malware. Although employing different datasets and platforms, our study shares a common goal of malware detection through supervised machine learning. However, notable differences in methodologies and datasets may account for variations in performance metrics.</p>
<p>Moreover, in the analysis of Windows malware conducted by Wang et al. [<xref ref-type="bibr" rid="ref-49">49</xref>], dynamic analysis was employed to identify malicious features. Their study achieved a True Positive Rate (TPR) of 96.3%, which provides a benchmark for comparison with our findings. While our study may differ in the specific classifiers utilized and the nature of features extracted, such comparative insights contribute to a deeper understanding of the effectiveness of different malware detection approaches.</p>
<p>Similarly, Singh et al. [<xref ref-type="bibr" rid="ref-50">50</xref>] adopted a statistical network conversation approach to analyze botnet traffic, successfully identifying multiple botnet applications with an average TPR of 95.0 %. While their focus differs from our study, which centers on Windows malware detection, the shared emphasis on supervised machine learning underscores the relevance of their findings to our comparative analysis.</p>
<p>Overall, our study adds to the body of knowledge surrounding malware detection by providing a detailed examination of supervised machine learning classifiers tailored for Windows systems. While direct comparisons with existing studies may present challenges due to variations in methodologies and datasets, the insights gleaned from such analyses contribute to a more comprehensive understanding of the landscape of malware detection techniques.</p>
</sec>
<sec id="s4_6">
<label>4.6</label>
<title>Discussion</title>
<p>In addition to comparing supervised machine learning-based Windows malware detection methods, it is imperative to address the challenges posed by complex malware variants and adversarial samples. As cyber threats become increasingly sophisticated, malware authors employ techniques such as polymorphism, obfuscation, and evasion to evade detection by traditional security measures. To effectively handle complex malware, researchers and practitioners must explore advanced feature extraction techniques and model architectures capable of capturing intricate patterns and behaviors exhibited by malicious software. This may involve leveraging deep learning approaches, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), which have demonstrated promising results in detecting complex malware variants by learning hierarchical representations of malware features.</p>
<p>Moreover, the rise of adversarial attacks presents a significant concern for supervised machine learning-based malware detection systems. Adversarial samples are specifically crafted to exploit vulnerabilities in machine learning models, leading to misclassification and potentially bypassing detection mechanisms altogether. To address this challenge, researchers are investigating techniques such as adversarial training, defensive distillation, and robust optimization, which aim to enhance the resilience of machine learning models against adversarial manipulation. By incorporating adversarial robustness into the model training process, cybersecurity practitioners can improve the reliability and effectiveness of malware detection systems in adversarial environments.</p>
<p>Furthermore, it is essential to establish robust evaluation frameworks that account for the presence of complex malware and adversarial samples in the testing datasets. This involves augmenting existing benchmark datasets with diverse and realistic malware samples, including polymorphic variants and adversarial examples generated using sophisticated attack algorithms. By evaluating detection models under realistic conditions, researchers can assess their performance in identifying both known and novel malware threats while mitigating the risk of false positives and false negatives. Additionally, ongoing collaboration between academia, industry, and government stakeholders is crucial for sharing knowledge, resources, and best practices in combating complex malware and adversarial attacks, ultimately strengthening the cybersecurity posture of organizations worldwide.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusion</title>
<p>The study aimed to enhance malware detection on Windows systems using supervised machine learning classifiers. The study explored various classifiers, including Gaussian Na&#x00EF;ve Bayes, K Nearest Neighbors, Stochastic Gradient Descent Classifier (SGDC), and Decision Tree, to assess their efficacy in detecting malicious software. Our findings demonstrate promising results, with Gaussian Na&#x00EF;ve Bayes achieving the highest accuracy rate of 99.54%, closely followed by Decision Tree at 97.4%. SGDC exhibited a slightly lower accuracy rate of 95.2%, while K Nearest Neighbors achieved 93.3%. These insights highlight the significance of leveraging supervised machine learning for bolstering cybersecurity measures on Windows platforms.</p>
<p>Moving forward, further research could explore hybrid approaches integrating multiple classifiers or leveraging ensemble learning techniques to enhance detection accuracy. Additionally, the development of specialized features tailored to Windows malware characteristics could further fortify malware detection methodologies. In essence, our study underscores the importance of continual innovation in malware detection strategies to safeguard computer systems against evolving threats. By leveraging supervised machine learning techniques, we aim to contribute to the ongoing efforts to fortify cybersecurity measures and mitigate the impact of malicious software on Windows environments.</p>
</sec>
<sec id="s6">
<label>6</label>
<title>Study Limitations and Future Work</title>
<p>This study highlighted several shortcomings that need careful consideration. First off, the caliber and representativeness of training data have a major impact on how well-supervised machine learning classifiers perform. The generalizability of our findings and the applicability of our models to real-world settings may be hampered by inadequate or biased datasets. Additionally, the selection of features and preprocessing methods might affect classifier performance, possibly adding bias or noise and lowering the accuracy of malware detection. We acknowledge that further research is necessary to fully comprehend these issues, and we offer focused improvement tactics to increase model performance.</p>
<p>Additionally, conventional detection techniques continue to face difficulties due to the dynamic nature of contemporary malware. Detection strategies must always be adjusted and improved to effectively counter new threats as adversaries change their tactics. Moreover, in resource-constrained situations, practical limitations such as computational resource limitations may impede the implementation and scalability of machine learning models. To tackle these obstacles, one must investigate novel approaches, cooperate with cybersecurity specialists, and create stronger detection frameworks. It is recommended that future research concentrate on sophisticated feature engineering techniques, hybrid learning methodologies, and real-time analytic methods to improve cyber resilience and malware detection capabilities.</p>
</sec>
</body>
<back>
<ack><p>This study is carried out through the support of the International Islamic University Islamabad Pakistan, University of Peshawar Pakistan, Princess Nourah bint Abdulrahman University Riyadh Saudi Arabia, University of Jeddah, and Princess Nourah bint Abdulrahman University Saudi Arabia.</p>
</ack>
<sec><title>Funding Statement</title>
<p>This research work is supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2024R411), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.</p>
</sec>
<sec><title>Author Contributions</title>
<p>Islam Zada: Conceptualization, Methodology, Writing&#x2014;Original Draft, Mohammed Naif Alatawi: Data Curation, Investigation, Writing&#x2014;Review &#x0026; Editing. Syed Muhammad Saqlain: Formal Analysis, Software, Visualization, Abdullah Alshahrani: Supervision, Project Administration, Funding Acquisition., Adel Alshamran: Conceptualization, Methodology, Writing&#x2014;Review &#x0026; Editing, Kanwal Imran: Data Curation, Investigation, Writing&#x2014;Review &#x0026; Editing, Hessa Alfraihi: Formal Analysis, Visualization, Project Administration. The authors made significant contributions to the research and development of this study. All authors reviewed the results and approved the final version of the manuscript.</p>
</sec>
<sec sec-type="data-availability"><title>Availability of Data and Materials</title>
<p>The data used in this work is available on the machine learning repository.</p>
</sec>
<sec><title>Ethics Approval</title>
<p>Not applicable.</p>
</sec>
<sec sec-type="COI-statement"><title>Conflicts of Interest</title>
<p>The authors declare that they have no conflicts of interest to report regarding the present study.</p>
</sec>
<ref-list content-type="authoryear">
<title>References</title>
<ref id="ref-1"><label>[1]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>U.</given-names> <surname>Bayer</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Moser</surname></string-name>, <string-name><given-names>C.</given-names> <surname>Kruegel</surname></string-name>, and <string-name><given-names>E.</given-names> <surname>Kirda</surname></string-name></person-group>, &#x201C;<article-title>Dynamic analysis of malicious code</article-title>,&#x201D; <source>J. Comput. Virol.</source>, vol. <volume>2</volume>, no. <issue>1</issue>, pp. <fpage>67</fpage>&#x2013;<lpage>77</lpage>, <year>2006</year>. doi: <pub-id pub-id-type="doi">10.1007/s11416-006-0012-2</pub-id>.</mixed-citation></ref>
<ref id="ref-2"><label>[2]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Saxe</surname></string-name> and <string-name><given-names>K.</given-names> <surname>Berlin</surname></string-name></person-group>, &#x201C;<article-title>Deep neural network-based malware detection using two dimensional binary program features</article-title>,&#x201D; in <conf-name>2015 10th Int. Conf. Malicious and Unwanted Softw. (MALWARE)</conf-name>, <publisher-loc>Fajardo, PR, USA</publisher-loc>, <year>2015</year>, pp. <fpage>11</fpage>&#x2013;<lpage>20</lpage>.</mixed-citation></ref>
<ref id="ref-3"><label>[3]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>O.</given-names> <surname>Or-Meir</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Nissim</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Elovici</surname></string-name>, and <string-name><given-names>L.</given-names> <surname>Rokach</surname></string-name></person-group>, &#x201C;<article-title>Dynamic malware analysis in the modern era&#x2014;A state of the art survey</article-title>,&#x201D; <source>ACM Comput. Surv.</source>, vol. <volume>52</volume>, no. <issue>5</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>48</lpage>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-4"><label>[4]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Choudhary</surname></string-name> and <string-name><given-names>M.</given-names> <surname>Vidyarthi</surname></string-name></person-group>, &#x201C;<article-title>A simple method for detection of metamorphic malware using dynamic analysis and text mining</article-title>,&#x201D; <source>Proc. Comput. Sci.</source>, vol. <volume>54</volume>, pp. <fpage>265</fpage>&#x2013;<lpage>270</lpage>, <year>2015</year>. doi: <pub-id pub-id-type="doi">10.1016/j.procs.2015.06.031</pub-id>.</mixed-citation></ref>
<ref id="ref-5"><label>[5]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M.</given-names> <surname>Ahmadi</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Ulyanov</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Semenov</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Trofimov</surname></string-name>, and <string-name><given-names>G.</given-names> <surname>Giacinto</surname></string-name></person-group>, &#x201C;<article-title>Novel feature extraction selection and fusion for effective malware family classification</article-title>,&#x201D; in <conf-name>Proc. of the Sixth ACM Conf. on Data and Appl. Security and Privacy, Association for Computing Machinery</conf-name>, <publisher-loc>New York, NY, USA</publisher-loc>, <year>2016</year>, pp. <fpage>183</fpage>&#x2013;<lpage>194</lpage>.</mixed-citation></ref>
<ref id="ref-6"><label>[6]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F. O.</given-names> <surname>Catak</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Ahmed</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Sahinbas</surname></string-name>, and <string-name><given-names>Z. H.</given-names> <surname>Khand</surname></string-name></person-group>, &#x201C;<article-title>Data augmentation-based malware detection using convolutional neural networks</article-title>,&#x201D; <source>PeerJ Comput. Sci.</source>, vol. <volume>7</volume>, pp. <fpage>e346</fpage>, <year>Jan. 2021</year>. doi: <pub-id pub-id-type="doi">10.7717/peerj-cs.346</pub-id>; <pub-id pub-id-type="pmid">33816996</pub-id></mixed-citation></ref>
<ref id="ref-7"><label>[7]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Vasan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Alazab</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Venkatraman</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Akram</surname></string-name>, and <string-name><given-names>Z.</given-names> <surname>Qin</surname></string-name></person-group>, &#x201C;<article-title>MTHAEL: Cross-architecture IoT malware detection based on neural network advanced ensemble learning</article-title>,&#x201D; <source>IEEE Trans. Comput.</source>, vol. <volume>69</volume>, no. <issue>11</issue>, pp. <fpage>1654</fpage>&#x2013;<lpage>1667</lpage>, <year>2020</year>. doi: <pub-id pub-id-type="doi">10.1109/TC.2020.3015584</pub-id>.</mixed-citation></ref>
<ref id="ref-8"><label>[8]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>T. M.</given-names> <surname>Kebede</surname></string-name>, <string-name><given-names>O.</given-names> <surname>Djaneye-Boundjou</surname></string-name>, <string-name><given-names>B. N.</given-names> <surname>Narayanan</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Ralescu</surname></string-name>, and <string-name><given-names>D.</given-names> <surname>Kapp</surname></string-name></person-group>, &#x201C;<article-title>Classification of malware programs using autoencoders based deep learning architecture and its application to the microsoft Malware classification challenge (BIG 2015) dataset</article-title>,&#x201D; in <conf-name>2017 IEEE Nat. Aerospace and Electron. Conf. (NAECON)</conf-name>, <publisher-loc>Dayton, OH, USA</publisher-loc>, <year>Jun. 27&#x2013;30, 2017</year>, pp. <fpage>70</fpage>&#x2013;<lpage>75</lpage>.</mixed-citation></ref>
<ref id="ref-9"><label>[9]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>C.</given-names> <surname>Willems</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Holz</surname></string-name>, and <string-name><given-names>F.</given-names> <surname>Freiling</surname></string-name></person-group>, &#x201C;<article-title>Toward automated dynamic malware analysis using CW sandbox</article-title>,&#x201D; <source>IEEE Security &#x0026; Privacy</source>, vol. <volume>5</volume>, no. <issue>2</issue>, pp. <fpage>32</fpage>&#x2013;<lpage>39</lpage>, <year>2007</year>. doi: <pub-id pub-id-type="doi">10.1109/MSP.2007.45</pub-id>.</mixed-citation></ref>
<ref id="ref-10"><label>[10]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Souri</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Hosseini</surname></string-name></person-group>, &#x201C;<article-title>A state-of-the-art survey of malware detection approaches using data mining techniques</article-title>,&#x201D; <source>Hum. Centric Comput. Inf. Sci.</source>, vol. <volume>8</volume>, no. <issue>1</issue>, pp. <fpage>3</fpage>, <year>2018</year>.</mixed-citation></ref>
<ref id="ref-11"><label>[11]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>G. E.</given-names> <surname>Dahl</surname></string-name>, <string-name><given-names>J. W.</given-names> <surname>Stokes</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Deng</surname></string-name>, and <string-name><given-names>D.</given-names> <surname>Yu</surname></string-name></person-group>, &#x201C;<article-title>Large-scale malware classification using random projections and neural networks</article-title>,&#x201D; in <conf-name>Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP)</conf-name>, <publisher-loc>Vancouver, BC, Canada</publisher-loc>, <year>May 26&#x2013;31, 2013</year>, pp. <fpage>3422</fpage>&#x2013;<lpage>3426</lpage>.</mixed-citation></ref>
<ref id="ref-12"><label>[12]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>E.</given-names> <surname>Gandotra</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Bansal</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Sofat</surname></string-name></person-group>, &#x201C;<article-title>Malware analysis and classification: A survey</article-title>,&#x201D; <source>J. Inf. Secur.</source>, vol. <volume>5</volume>, pp. <fpage>56</fpage>&#x2013;<lpage>64</lpage>, <year>2014</year>. doi: <pub-id pub-id-type="doi">10.4236/jis.2014.52006</pub-id>.</mixed-citation></ref>
<ref id="ref-13"><label>[13]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>M. Z.</given-names> <surname>Shafiq</surname></string-name>, <string-name><given-names>S. M.</given-names> <surname>Tabish</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Mirza</surname></string-name>, and <string-name><given-names>M.</given-names> <surname>Farooq</surname></string-name></person-group>, &#x201C;<article-title>PE-Miner: Mining structural information to detect malicious executables in realtime</article-title>,&#x201D; in <conf-name>12th Int. Symp.</conf-name>, <publisher-loc>Saint-Malo, France</publisher-loc>, <publisher-name>RAID</publisher-name>, <year>Sep. 2009</year>, pp. <fpage>121</fpage>&#x2013;<lpage>141</lpage>.</mixed-citation></ref>
<ref id="ref-14"><label>[14]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>Q.</given-names> <surname>Yanchen</surname></string-name>, <string-name><given-names>Z.</given-names> <surname>Bin</surname></string-name>, and <string-name><given-names>Z.</given-names> <surname>Weizhe</surname></string-name></person-group>, &#x201C;<article-title>Malware classification method based on word vector of bytes and multilayer perception</article-title>,&#x201D; in <conf-name>IEEE Int. Conf. on Commun. (ICC)</conf-name>, <publisher-loc>Dublin, Ireland</publisher-loc>, <year>Jun. 2020</year>, pp. <fpage>1</fpage>&#x2013;<lpage>6</lpage>.</mixed-citation></ref>
<ref id="ref-15"><label>[15]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Drew</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Hahsler</surname></string-name>, and <string-name><given-names>T.</given-names> <surname>Moore</surname></string-name></person-group>, &#x201C;<article-title>Polymorphic malware detection using sequence classification methods and ensembles</article-title>,&#x201D; <source>EURASIP J. Inf. Secur.</source>, vol. <volume>2017</volume>, no. <issue>1</issue>, <year>2017</year>. doi: <pub-id pub-id-type="doi">10.1186/s13635-017-0055-6</pub-id>.</mixed-citation></ref>
<ref id="ref-16"><label>[16]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Cesare</surname></string-name> and <string-name><given-names>Y.</given-names> <surname>Xiang</surname></string-name></person-group>, &#x201C;<article-title>Classification of malware using structured control flow</article-title>,&#x201D; in <conf-name>Proc. of the Eighth Australasian Symp. on Parallel and Distrib. Comput.</conf-name>, <publisher-loc>Brisbane, Australia</publisher-loc>, <year>Jan. 01, 2010</year>, vol. <volume>107</volume>, pp. <fpage>61</fpage>&#x2013;<lpage>70</lpage>.</mixed-citation></ref>
<ref id="ref-17"><label>[17]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Sohail</surname></string-name>, <string-name><given-names>U.</given-names> <surname>Zahoora</surname></string-name>, and <string-name><given-names>A. S.</given-names> <surname>Qureshi</surname></string-name></person-group>, &#x201C;<article-title>A survey of the recent architectures of deep convolutional neural networks</article-title>,&#x201D; <source>Artif. Intell. Rev.</source>, vol. <volume>53</volume>, no. <issue>8</issue>, pp. <fpage>5455</fpage>&#x2013;<lpage>5516</lpage>, <year>2020</year>. doi: <pub-id pub-id-type="doi">10.1007/s10462-020-09825-6</pub-id>.</mixed-citation></ref>
<ref id="ref-18"><label>[18]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Ndibanje</surname></string-name>, <string-name><given-names>K. H.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>Y. J.</given-names> <surname>Kang</surname></string-name>, <string-name><given-names>H. H.</given-names> <surname>Kim</surname></string-name>, <string-name><given-names>T. Y.</given-names> <surname>Kim</surname></string-name> and <string-name><given-names>H. J.</given-names> <surname>Lee</surname></string-name></person-group>, &#x201C;<article-title>Cross-method-based analysis and classification of malicious behavior by API calls extraction</article-title>,&#x201D; <source>Appl. Sci.</source>, vol. <volume>9</volume>, no. <issue>2</issue>, pp. <fpage>239</fpage>, <year>2019</year>. doi: <pub-id pub-id-type="doi">10.3390/app9020239</pub-id>.</mixed-citation></ref>
<ref id="ref-19"><label>[19]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Zhang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Xiao</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Mercaldo</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Ni</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Martinelli</surname></string-name> and <string-name><given-names>A. K.</given-names> <surname>Sangaiah</surname></string-name></person-group>, &#x201C;<article-title>Classification of ransomware families with machine learning based on <italic>N</italic>-gram of opcodes</article-title>,&#x201D; <source>Future Gener. Comput. Syst.</source>, vol. <volume>90</volume>, pp. <fpage>211</fpage>&#x2013;<lpage>221</lpage>, <year>Jan. 2019</year>. doi: <pub-id pub-id-type="doi">10.1016/j.future.2018.07.052</pub-id>.</mixed-citation></ref>
<ref id="ref-20"><label>[20]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>El-Shafai</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Almomani</surname></string-name>, and <string-name><given-names>A.</given-names> <surname>AlKhayer</surname></string-name></person-group>, &#x201C;<article-title>Visualized malware multi-classification framework using fine-tuned CNN-based transfer learning models</article-title>,&#x201D; <source>Appl. Sci.</source>, vol. <volume>11</volume>, no. <issue>14</issue>, pp. <fpage>6446</fpage>, <year>Jul. 2021</year>. doi: <pub-id pub-id-type="doi">10.3390/app11146446</pub-id>.</mixed-citation></ref>
<ref id="ref-21"><label>[21]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>A.</given-names> <surname>Bensaoud</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Kalita</surname></string-name>, and <string-name><given-names>M.</given-names> <surname>Bensaoud</surname></string-name></person-group>, &#x201C;<article-title>A survey of malware detection using deep learning</article-title>,&#x201D; <source>Machine Learn. Appl.</source>, vol. <volume>16</volume>, no. <issue>1</issue>, pp. <fpage>100546</fpage>, <year>Jun. 2024</year>. doi: <pub-id pub-id-type="doi">10.1016/j.mlwa.2024.100546</pub-id>.</mixed-citation></ref>
<ref id="ref-22"><label>[22]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>W.</given-names> <surname>Hardy</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Chen</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Hou</surname></string-name>, <string-name><given-names>Y.</given-names> <surname>Ye</surname></string-name>, and <string-name><given-names>X.</given-names> <surname>Li</surname></string-name></person-group>, &#x201C;<article-title>DL4MD: A deep learning framework for intelligent malware detection</article-title>,&#x201D; <conf-name>Proc. Int. Conf. Data Mining Steering Committee World Congr. Comput. Sci. (DMIN)</conf-name>, <publisher-loc>Las Vegas, NV, USA</publisher-loc>, <year>Jul. 2016</year>, pp. <fpage>61</fpage>&#x2013;<lpage>67</lpage>.</mixed-citation></ref>
<ref id="ref-23"><label>[23]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>E.</given-names> <surname>Gandotra</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Bansal</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Sofat</surname></string-name></person-group>, &#x201C;<article-title>Tools &#x0026; techniques for malware analysis and classification</article-title>,&#x201D; <source>Int. J. Next-Generation Comput.</source>, vol. <volume>7</volume>, no. <issue>3</issue>, pp. <fpage>176</fpage>, <year>2016</year>.</mixed-citation></ref>
<ref id="ref-24"><label>[24]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>D.</given-names> <surname>Yuxin</surname></string-name> and <string-name><given-names>Z.</given-names> <surname>Siyi</surname></string-name></person-group>, &#x201C;<article-title>Malware detection based on deep learning algorithm</article-title>,&#x201D; <source>Neural Comput. Appl.</source>, vol. <volume>31</volume>, no. <issue>14</issue>, pp. <fpage>461</fpage>&#x2013;<lpage>472</lpage>, <year>Feb. 2020</year>. doi: <pub-id pub-id-type="doi">10.1007/s00521-017-3077-6</pub-id>.</mixed-citation></ref>
<ref id="ref-25"><label>[25]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Srinivasan</surname></string-name> and <string-name><given-names>P.</given-names> <surname>Deepalakshmi</surname></string-name></person-group>, &#x201C;<article-title>Enhancing the security in cyber-world by detecting the botnets using ensemble classification-based machine learning</article-title>,&#x201D; <source>Meas. Sens.</source>, vol. <volume>25</volume>, no. <issue>2665&#x2013;9174</issue>, pp. <fpage>100624</fpage>, <year>Feb. 2023</year>. doi: <pub-id pub-id-type="doi">10.1016/j.measen.2022.100624</pub-id>.</mixed-citation></ref>
<ref id="ref-26"><label>[26]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>U. E. H.</given-names> <surname>Tayyab</surname></string-name>, <string-name><given-names>F. B.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>M. H.</given-names> <surname>Durad</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Khan</surname></string-name>, and <string-name><given-names>Y. S.</given-names> <surname>Lee</surname></string-name></person-group>, &#x201C;<article-title>A survey of the recent trends in deep learning based malware detection</article-title>,&#x201D; <source>J. Cybersecurity Privacy</source>, vol. <volume>2</volume>, no. <issue>4</issue>, pp. <fpage>800</fpage>&#x2013;<lpage>829</lpage>, <year>Sep. 2022</year>. doi: <pub-id pub-id-type="doi">10.3390/jcp2040041</pub-id>.</mixed-citation></ref>
<ref id="ref-27"><label>[27]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>R.</given-names> <surname>Oak</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Du</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Yan</surname></string-name>, <string-name><given-names>H.</given-names> <surname>Takawale</surname></string-name>, and <string-name><given-names>I.</given-names> <surname>Amit</surname></string-name></person-group>, &#x201C;<article-title>Malware detection on highly imbalanced data through sequence modeling</article-title>,&#x201D; in <conf-name>12th ACM Workshop on Artif. Intel. and Security</conf-name>, <publisher-loc>London, UK</publisher-loc>, <year>15 Nov., 2019</year>, pp. <fpage>37</fpage>&#x2013;<lpage>48</lpage>.</mixed-citation></ref>
<ref id="ref-28"><label>[28]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Cakir</surname></string-name> and <string-name><given-names>E.</given-names> <surname>Dogdu</surname></string-name></person-group>, &#x201C;<article-title>Malware classification using deep learning methods</article-title>,&#x201D; in <conf-name>ACMSE &#x2018;18: Proc. of the ACM Southeast (ACMSE) 2018 Conf.</conf-name>, <publisher-loc>Richmond Kentucky, USA</publisher-loc>, <year>Mar. 29&#x2013;31, 2018</year>, pp. <fpage>1</fpage>&#x2013;<lpage>5</lpage>.</mixed-citation></ref>
<ref id="ref-29"><label>[29]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Verma</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Malik</surname></string-name>, and <string-name><given-names>I.</given-names> <surname>Batra</surname></string-name></person-group>, &#x201C;<article-title>Analyzing and classifying malware types on windows platform using an ensemble machine learning approach</article-title>,&#x201D; <source>Int. J. Performability Eng.</source>, vol. <volume>20</volume>, no. <issue>5</issue>, pp. <fpage>312</fpage>&#x2013;<lpage>318</lpage>, <year>2024</year>. doi: <pub-id pub-id-type="doi">10.23940/ijpe.24.05.p6.312318</pub-id>.</mixed-citation></ref>
<ref id="ref-30"><label>[30]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N.</given-names> <surname>Usman</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Intelligent dynamic malware detection using machine learning in IP reputation for forensics data analytics</article-title>,&#x201D; <source>Future Gener. Comput. Syst.</source>, vol. <volume>118</volume>, pp. <fpage>124</fpage>&#x2013;<lpage>141</lpage>, <year>May 2021</year>. doi: <pub-id pub-id-type="doi">10.1016/j.future.2021.01.004</pub-id>.</mixed-citation></ref>
<ref id="ref-31"><label>[31]</label><mixed-citation publication-type="other"><person-group person-group-type="author"><string-name><given-names>J.</given-names> <surname>Zhang</surname></string-name></person-group>, &#x201C;<article-title>Machine learning with feature selection using principal component analysis for malware detection: A case study</article-title>,&#x201D; <comment>arXiv:1902.03639</comment>, <year>2019</year>.</mixed-citation></ref>
<ref id="ref-32"><label>[32]</label><mixed-citation publication-type="conf-proc"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Apruzzese</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Colajanni</surname></string-name>, <string-name><given-names>L.</given-names> <surname>Ferretti</surname></string-name>, <string-name><given-names>A.</given-names> <surname>Guido</surname></string-name>, and <string-name><given-names>M.</given-names> <surname>Marchetti</surname></string-name></person-group>, &#x201C;<article-title>On the effectiveness of machine and deep learning for cyber security</article-title>,&#x201D; in <conf-name>10th Int. Conf. on Cyber Conflict</conf-name>, <publisher-loc>Tallinn, Estonia</publisher-loc>, <year>May 29&#x2013;Jun. 01, 2018</year>. pp. <fpage>371</fpage>&#x2013;<lpage>390</lpage>.</mixed-citation></ref>
<ref id="ref-33"><label>[33]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Kumar</surname></string-name> and <string-name><surname>Sudhakar</surname></string-name></person-group>, &#x201C;<article-title>MCFT-CNN: Malware classification with fine-tune convolution neural networks using traditional and transfer learning in Internet of Things</article-title>,&#x201D; <source>Future Gener. Comput. Syst.</source>, vol. <volume>125</volume>, pp. <fpage>334</fpage>&#x2013;<lpage>351</lpage>, <year>Dec. 2021</year>. doi: <pub-id pub-id-type="doi">10.1016/j.future.2021.06.029</pub-id>.</mixed-citation></ref>
<ref id="ref-34"><label>[34]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Seneviratne</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Shariffdeen</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Rasnayaka</surname></string-name>, and <string-name><given-names>N.</given-names> <surname>Kasthuriarachchi</surname></string-name></person-group>, &#x201C;<article-title>Self-supervised vision transformers for malware detection</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>10</volume>, pp. <fpage>103121</fpage>&#x2013;<lpage>103135</lpage>, <year>2022</year>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2022.3206445</pub-id>.</mixed-citation></ref>
<ref id="ref-35"><label>[35]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Zhao</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Wang</surname></string-name>, and <string-name><given-names>J.</given-names> <surname>Bai</surname></string-name></person-group>, &#x201C;<article-title>Malware detection method based on the control-flow construct feature of software</article-title>,&#x201D; <source>IET Inf. Secur.</source>, vol. <volume>8</volume>, pp. <fpage>18</fpage>&#x2013;<lpage>24</lpage>, <year>2014</year>. doi: <pub-id pub-id-type="doi">10.1049/iet-ifs.2012.0289</pub-id>.</mixed-citation></ref>
<ref id="ref-36"><label>[36]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>X.</given-names> <surname>Lu</surname></string-name>, <string-name><given-names>F.</given-names> <surname>Jiang</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Zhou</surname></string-name>, <string-name><given-names>S.</given-names> <surname>Yi</surname></string-name>, <string-name><given-names>J.</given-names> <surname>Sha</surname></string-name> and <string-name><given-names>L.</given-names> <surname>Pietro</surname></string-name></person-group>, &#x201C;<article-title>ASSCA: API sequence and statistics features combined architecture for malware detection</article-title>,&#x201D; <source>Comput. Netw.</source>, vol. <volume>157</volume>, pp. <fpage>99</fpage>&#x2013;<lpage>111</lpage>, <year>Jul. 2019</year>.</mixed-citation></ref>
<ref id="ref-37"><label>[37]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Islam</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Enhancing IoT-based software defect prediction in analytical data management using war strategy optimization and Kernel ELM</article-title>,&#x201D; <source>J. Wireless Netw.</source>, vol. <volume>29</volume>, no. <issue>8</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>19</lpage>, <year>2023</year>.</mixed-citation></ref>
<ref id="ref-38"><label>[38]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>G.</given-names> <surname>Ramesh</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Menen</surname></string-name></person-group>, &#x201C;<article-title>Automated dynamic approach for detecting ransomware using finite-state machine</article-title>,&#x201D; <source>Decis. Support Syst.</source>, vol. <volume>138</volume>, pp. <fpage>113400</fpage>, <year>Nov. 2020</year>. doi: <pub-id pub-id-type="doi">10.1016/j.dss.2020.113400</pub-id>.</mixed-citation></ref>
<ref id="ref-39"><label>[39]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. S.</given-names> <surname>Akhtar</surname></string-name> and <string-name><given-names>T.</given-names> <surname>Feng</surname></string-name></person-group>, &#x201C;<article-title>Detection of malware by deep learning as CNN-LSTM machine learning techniques in real time</article-title>,&#x201D; <source>Symmetry</source>, vol. <volume>14</volume>, no. <issue>11</issue>, pp. <fpage>2308</fpage>, <year>2022</year>. doi: <pub-id pub-id-type="doi">10.3390/sym14112308</pub-id>.</mixed-citation></ref>
<ref id="ref-40"><label>[40]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Ullah</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>Cyber security threats detection in Internet of Things using deep learning approach</article-title>,&#x201D; <source>IEEE Access</source>, vol. <volume>7</volume>, pp. <fpage>124379</fpage>&#x2013;<lpage>124389</lpage>, <year>2019</year>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2019.2937347</pub-id>.</mixed-citation></ref>
<ref id="ref-41"><label>[41]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>M. S.</given-names> <surname>Mahdavinejad</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Rezvan</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Barekatain</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Adibi</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Barnaghi</surname></string-name> and <string-name><given-names>A. P.</given-names> <surname>Sheth</surname></string-name></person-group>, &#x201C;<article-title>Machine learning for Internet of Things data analysis: A survey</article-title>,&#x201D; <source>Digit. Commun. Netw.</source>, vol. <volume>4</volume>, no. <issue>3</issue>, pp. <fpage>161</fpage>&#x2013;<lpage>175</lpage>, <year>2018</year>. doi: <pub-id pub-id-type="doi">10.1016/j.dcan.2017.10.002</pub-id>.</mixed-citation></ref>
<ref id="ref-42"><label>[42]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>F.</given-names> <surname>Liang</surname></string-name>, <string-name><given-names>W.</given-names> <surname>Yu</surname></string-name>, <string-name><given-names>X.</given-names> <surname>Liu</surname></string-name>, <string-name><given-names>D.</given-names> <surname>Griffith</surname></string-name>, and <string-name><given-names>N.</given-names> <surname>Golmie</surname></string-name></person-group>, &#x201C;<article-title>Toward edge-based deep learning in industrial Internet of Things</article-title>,&#x201D; <source>IEEE Internet Things J.</source>, vol. <volume>7</volume>, no. <issue>5</issue>, pp. <fpage>4329</fpage>&#x2013;<lpage>4341</lpage>, <year>May 2020</year>. doi: <pub-id pub-id-type="doi">10.1109/JIOT.2019.2963635</pub-id>; <pub-id pub-id-type="pmid">38486787</pub-id></mixed-citation></ref>
<ref id="ref-43"><label>[43]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>H.</given-names> <surname>Paulheim</surname></string-name> and <string-name><given-names>R.</given-names> <surname>Meusel</surname></string-name></person-group>, &#x201C;<article-title>A decomposition of the outlier detection problem into a set of supervised learning problems</article-title>,&#x201D; <source>Mach. Learn.</source>, vol. <volume>100</volume>, no. <issue>2/3</issue>, pp. <fpage>509</fpage>&#x2013;<lpage>531</lpage>, <year>2015</year>. doi: <pub-id pub-id-type="doi">10.1007/s10994-015-5507-y</pub-id>.</mixed-citation></ref>
<ref id="ref-44"><label>[44]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>B.</given-names> <surname>Charbuty</surname></string-name> and <string-name><given-names>A.</given-names> <surname>Abdulazeez</surname></string-name></person-group>, &#x201C;<article-title>Classification based on decision tree algorithm for machine learning</article-title>,&#x201D; <source>J. Appl. Sci. Technol. Trends.</source>, vol. <volume>2</volume>, no. <issue>1</issue>, pp. <fpage>20</fpage>&#x2013;<lpage>28</lpage>, <year>Mar. 2021</year>. doi: <pub-id pub-id-type="doi">10.38094/jastt20165</pub-id>.</mixed-citation></ref>
<ref id="ref-45"><label>[45]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>S.</given-names> <surname>Aurangzeb</surname></string-name>, <string-name><given-names>R. N. B.</given-names> <surname>Rais</surname></string-name>, <string-name><given-names>M.</given-names> <surname>Aleem</surname></string-name>, <string-name><given-names>M. A.</given-names> <surname>Islam</surname></string-name>, and <string-name><given-names>M. A.</given-names> <surname>Iqbal</surname></string-name></person-group>, &#x201C;<article-title>On the classification of microsoft-windows ransomware using hardware profile</article-title>,&#x201D; <source>PeerJ Comput. Sci.</source>, vol. <volume>7</volume>, pp. <fpage>361</fpage>, <year>Feb. 2021</year>. doi: <pub-id pub-id-type="doi">10.7717/peerj-cs.361</pub-id>; <pub-id pub-id-type="pmid">33817011</pub-id></mixed-citation></ref>
<ref id="ref-46"><label>[46]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>V.</given-names> <surname>Kouliaridis</surname></string-name>, <string-name><given-names>K.</given-names> <surname>Barmpatsalou</surname></string-name>, <string-name><given-names>G.</given-names> <surname>Kambourakis</surname></string-name>, and <string-name><given-names>S.</given-names> <surname>Chen</surname></string-name></person-group>, &#x201C;<article-title>A survey on mobile malware detection techniques</article-title>,&#x201D; <source>IEICE Trans. Inf. Syst.</source>, vol. <volume>E103</volume>, no. <issue>2</issue>, pp. <fpage>204</fpage>&#x2013;<lpage>211</lpage>, <year>2020</year>. doi: <pub-id pub-id-type="doi">10.1587/transinf.2019INI0003</pub-id></mixed-citation></ref>
<ref id="ref-47"><label>[47]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Zhao</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>ERNN: Error-resilient RNN for encrypted traffic detection towards network-induced phenomena</article-title>,&#x201D; <source>IEEE Transac. Depend. Secure Computing.</source>, pp. <fpage>1</fpage>&#x2013;<lpage>18</lpage>, <year>Feb. 2023</year>.</mixed-citation></ref>
<ref id="ref-48"><label>[48]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>I.</given-names> <surname>Zada</surname></string-name>, <string-name><given-names>I.</given-names> <surname>Khan</surname></string-name>, <string-name><given-names>T.</given-names> <surname>Rahman</surname></string-name>, and <string-name><given-names>A.</given-names> <surname>Jameel</surname></string-name></person-group>, &#x201C;<article-title>Classification of software failure incidents using SVM</article-title>,&#x201D; <source>The Sciencetech</source>, vol. <volume>2</volume>, no. <issue>3</issue>, pp. <fpage>1</fpage>&#x2013;<lpage>13</lpage>, <year>Sep. 2021</year>.</mixed-citation></ref>
<ref id="ref-49"><label>[49]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>Z.</given-names> <surname>Wang</surname></string-name> <etal>et al.</etal></person-group>, &#x201C;<article-title>BugPre: An intelligent software version-to-version bug prediction system using graph convolutional neural networks</article-title>,&#x201D; <source>Complex Intell. Syst.</source>, vol. <volume>9</volume>, no. <issue>4</issue>, pp. <fpage>3835</fpage>&#x2013;<lpage>3855</lpage>, <year>Aug. 2023</year>. doi: <pub-id pub-id-type="doi">10.1007/s40747-022-00848-w</pub-id>.</mixed-citation></ref>
<ref id="ref-50"><label>[50]</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><string-name><given-names>N. J.</given-names> <surname>Singh</surname></string-name>, <string-name><given-names>N.</given-names> <surname>Hoque</surname></string-name>, <string-name><given-names>K. R.</given-names> <surname>Singh</surname></string-name>, and <string-name><given-names>D. K.</given-names> <surname>Bhattacharyya</surname></string-name></person-group>, &#x201C;<article-title>Botnet-based IoT network traffic analysis using deep learning</article-title>,&#x201D; <source>Secur. Privacy.</source>, vol. <volume>7</volume>, no. <issue>2</issue>, pp. <fpage>355</fpage>, <year>Mar. 2024</year>. doi: <pub-id pub-id-type="doi">10.1002/spy2.355</pub-id>.</mixed-citation></ref>
</ref-list>
</back></article>