Extending Network Intrusion Detection with Enhanced Particle Swarm Optimization Techniques
Surasit Songma1, Watcharakorn Netharn2* and Siriluck Lorpunmanee3
1Department of Cyber Security, Faculty of Science and Technology, Suan Dusit University, Bangkok, Thailand
2Department of Information Technology, Faculty of Science and Technology, Suan Dusit University, Bangkok, Thailand
3Department of Computer Science, Faculty of Science and Technology, Suan Dusit University, Bangkok, Thailand
ABSTRACT
The present research investigates how to improve Network Intrusion Detection Systems (NIDS) by combining Machine Learning (ML) and Deep Learning (DL) techniques, addressing the growing challenge of cybersecurity threats. A thorough process for data preparation, comprising activities like cleaning, normalization, and segmentation into training and testing sets, lays the framework for model training and evaluation. The study uses the CSE-CIC-IDS 2018 and LITNET-2020 datasets to compare ML methods (Decision Trees, Random Forest, XGBoost) and DL models (CNNs, RNNs, DNNs, MLP) against key performance metrics (Accuracy, Precision, Recall, and F1-Score). The Decision Tree model performed better across all measures after being fine-tuned with Enhanced Particle Swarm Optimization (EPSO), demonstrating the model’s ability to detect network breaches effectively. The findings highlight EPSO’s importance in improving ML classifiers for cybersecurity, proposing a strong framework for NIDS with high precision and dependability. This extensive analysis not only contributes to the cybersecurity arena by providing a road to robust intrusion detection solutions, but it also proposes future approaches for improving ML models to combat the changing landscape of network threats.
KEYWORDS
Intrusion Detection System, Machine Learning Techniques, Deep Learning, Particle Swarm Optimization, CSE-CIC-IDS 2018, LITNET-2020
1. INTRODUCTION
Network security is one of the significant challenges that network administrators and owners face, particularly given the growing number and complexity of attacks. Due to the rapid increase in those issues, various protection measures and methods must be developed. Network Intrusion Detection Systems (NIDS) scan and analyze network traffic to identify assaults and alert network administrators[1][2][3].
In recent years, the growth of digital communication networks has resulted in an unprecedented volume of data, requiring improved machine learning (ML) algorithms for successful analysis and categorization [4]. The CSE-CIC-IDS 2018 and LITNET-2020 datasets, a representative example of such extensive datasets, present challenges in multi-class classification tasks[3][4][5][6][7]. The objective of this paper is to conduct a thorough investigation into the performance of several ML classifiers, such as Random Forest (RF), Decision Tree (DT), Extreme Gradient Boosting (XGBoost)[3][6][8],Convolutional Neural Networks (CNNs)[9], Recurrent Neural Networks (RNNs), Deep Neural Networks (DNNs)[11],and Multilayer Perceptron (MLP)[6][12][13]on the CSE-CICIDS 2018 and LITNET-2020 datasets. As a result, the best model is selected for fine-tuning with Enhanced Particle Swarm Optimization (EPSO)[14][15].
The selection of classifiers for this study reflects a diverse range of approaches, from traditional clustering methods to state-of-the-art ensemble models and neural network architectures. In contrast, RF, DT, and XGBoost, renowned for their robustness and accuracy, are investigated as representative ensemble methods. Additionally, the study explores the effectiveness of Deep learning (DL) models, including CNNs, RNNs, DNNs, and MLP, which are known for their capability to automatically extract complex features from high-dimensional data.
The motivation behind this research lies in the necessity to discern the strengths and limitations of diverse ML approaches when confronted with the challenges posed by the CSE-CIC-IDS 2018 and LITNET-2020 datasets. Understanding the relative performance of these classifiers is pivotal for guiding the selection of models in real-world applications, particularly in domains where multi-class classification plays a crucial role, such as network security and anomaly detection.Through a rigorous evaluation based on accuracy, precision, recall, and F1-Score metrics[6][11][16]., this research aims to provide insights into the suitability of each classifier for the CSE-CIC-IDS 2018 and LITNET-2020 datasets, offering valuable guidance for practitioners and researchers alike in choosing the most effective approach for similar complex classification tasks.
The goal of this study was to find the most efficient classifier using preprocessing methodologies, which convert into widely used ML and DL approaches for intrusion detection systems(IDS). We investigated popular and high-performing classification techniques such as XGBoost, DT, RF,CNNs, RNNs, DNNs, and MLP. The performance evaluation was multidimensional, focusing on four essential metrics: accuracy, precision, recall, and F1-Score. Accuracy is a critical metric of a model’s effectiveness in classification tasks, expressing the proportion of correct predictions to total predictions made. A model with great accuracy can foresee outcomes that are consistent with real-world observations.The primary contributions of this work can be described in full detail as the following:
Using the CSE-CIC-IDS-2018 and LITNET-2020 datasets, which are among the most recent and include a wide range of incursion types, establishing them as standards.
• On these datasets, researchers evaluated well-known ML classifiers with DL models, that is, a combination had never been substantially examined.
• Implementation of a multi-class classification approach to assess model efficiency thoroughly.
• Verification of model adequacy across both datasets to ensure optimal fit and applicability to a variety of data scenarios.
• Model performance is improved by using EPSO for more accuracy and finer parameter selection.
• Researchers evaluated the performance of ML classifiers that they had previously classified as effective[3].
• The analysis of full datasets took into account Big Data, setting this effort apart from others that frequently use a random subset of data for research.
The structure of the paper is as follows Section 2 reviews relevant previous studies, Section 3 describes the proposed research methodology, Section 4 discusses the datasets chosen for analysis, Section 5 details the experimental setup, Section 6 presents the results and discussions of the experiments, and Section 7 concludes the paper with a summary of the findings and recommendations for future work
2. RELATED WORKS
Research into multi-class classification using ML classifiers has yielded extensive discourse, delving into the comparative strengths and weaknesses of various algorithms applied to a range of datasets. Ensemble strategies, notably RF, DT, and XGBoost, have risen as formidable contenders in the classification arena, attributed to their proficiency in curbing overfitting and enhancing predictive precision[17]. Empirical studies have corroborated the superiority of these methods in diverse sectors, warranting their inclusion for an in-depth evaluation against the CSE-CIC-IDS 2018 and LITNET-2020 datasets within this study[18][19].
The proliferation of DL has amplified the deployment of neural network models, especially CNNs, RNNs, DNNs, and MLP in sophisticated classification scenarios[4]. CNNs are lauded for their inherent ability to autonomously distill hierarchical features from structured inputs, proving formidable in image and sequence processing tasks. MLP are celebrated for their aptitude in deciphering intricate data interrelations and securing roles across various disciplines. RNNs are distinguished by their specialized design to process sequential information, adeptly capturing temporal sequences. The established adaptability of these neural network models in past research underpins their selection for this study’s analytical assessment using the CSE-CIC-IDS 2018 and LIT-NET-2020 datasets.
The literature presents extensive analyses of individual classifiers, ensemble methods, and neural networks on the intricate CSE-CIC-IDS 2018 and LITNET-2020 datasets. This study endeavors to fill this void by thoroughly comparing these classifiers, taking into account their unique attributes and appropriateness for complex multi-class classification tasks. Our goal is to improve the existing knowledge base on the ML model, particularly inside the complex frameworks of digital communication networks.
Simultaneously, there’s a burgeoning interest in developing hybrid models that synergize the descriptive power of clustering techniques with the robust predictive abilities of advanced ensemble methods and neural networks. By exploring hybrid models on the CSE-CIC-IDS 2018 and LITNET-2020 datasets, this research may uncover novel insights into effective strategies for managing the sophisticated challenges of multi-class classification. This is particularly pertinent in the field of network security and anomaly detection, where precision in classification is paramount for identifying malicious threats. Against the backdrop of rapid advancements in ML, this investigation will also consider the latest optimization algorithms, regularization methods, and architectural innovations, ensuring a state-of-the-art approach in classifier evaluation.
A review of key documents and research publications revealed that various studies used MLtechniques in conjunction with the CSE-CIC-IDS 2018 and LIT-NET-2020 datasets to detect intrusions. This can be illustrated as the following:
As a result, previous researchers have investigated different common MLalgorithms for intrusion detection
3. PROPOSED METHODOLOGY
We gained an understanding of the present difficulties and proposed solutions by examining relevant literature and scholarly papers. This informed understanding has been simplified into a complete overview of ML and DL methods used in NIDS, as shown in Figure 1
Figure 1. The methodology employs a taxonomy of machine and DL models in NIDS
Figure 1. depicts a classification of ML and DL models that are often used in NIDS. It divides models into two major branches.
3.1. Machine Learning Models
This branch includes models that use ML approaches. This category includes the following[3]
3.2. Deep Learning Models
This section contains models built on DL architectures that are intended to build hierarchical representations of data[6][22][28][29][30].
Figure 1 could be used in an academic article or technical presentation to describe the implementation of NIDS using a variety of ML and DL techniques. Figure 2 depicts the conceptual model of the operational framework produced by the researcher. This framework builds on the researcher’s earlier work, which focused on optimizing IDS utilizing three phase strategy [3], and the CSE-CIC-IDS-2018 and LITNET-2020 Datasetsand integrates aspects from that work.This flowchart from Figure 2 describes the typical process of building and evaluating ML and DL models. It outlines the following key phases.
• The method begins with the original datasets, which in this case are the CSE-CIC-IDS 2018 and LITNET-2020. These datasets provide raw data for training and testing the models.
• Data Preprocessing This is the first phase, in which raw data is prepared for modeling. This includes:
• Data Cleaning Delete or correct data that is erroneous, incomplete, irrelevant, duplicated, or incorrectly formatted.
• Exploratory Data Analysis Analyzing data to find patterns, relationships, or anomalies to guide further analysis.
• Encoding is the process of transforming categorical data into a numerical format.
• Normalization involves rescaling input data to a uniform, standard range.
• Data Splitting Divide the dataset into training and testing sets.
• Training Phase The preprocessed training dataset is supplied into a ML or DL algorithm. This method builds a model by learning from the data.
• Testing Phase The trained model is tested on a separate testing dataset. This stage is critical for evaluating the model’s performance on new, previously unseen data.
• Evaluation Model This step entails evaluating the performance of ML/DL models using metrics like accuracy, precision, recall, and F1-Score. Based on these criteria, the bestperforming model is chosen. The model’s performance is assessed using the following metrics.
• Accuracy refers to the percentage of correct predictions made by the model.
• Precision is calculated as the number of real positive results divided by the total number of positive outcomes, including those that were incorrectly detected.
• Recall is the number of true positive results divided by the total number of samples that should have been positive.
• F1-Score is the harmonic mean of precision and recall, achieving a balance between the two criteria.
• Optimization Finally, the selected model undergoes optimization using Extended Particle Swarm Optimization (EPSO) to further refine the model’s parameters and improve its performance.
Figure 1. The proposed framework.
The flowchart presents a standard approach to ML model development, showcasing the importance of data preparation, model training, and model evaluation to ensure that the developed model performs accurately and effectively on the task it’s designed for.
4. SELECTION OF DATASET
Recent advances in cybersecurity research have made new flow-based benchmark datasets available to the public, including CSE-CIC-IDS-2018 and LIT-NET-2020 [4][21].These datasets are critical to enhancing the efficacy of supervised learning systems in network intrusion detection. They provide the vital data necessary to properly train IDS, allowing these systems to identify a wide range of network threats with high accuracy and dependability. Using such datasets, IDS may be fine-tuned to detect various sorts of cyber threats, resulting in a strong defense against network intrusions.
4.1. CSE-CIC-IDS-2018 Datasets
CSE-CIC-IDS-2018 dataset is a collaboration between the Communications Security Establishment (CSE) and the Canadian Institute for Cybersecurity (CIC), and it is meant to boost intrusion detection research[3]. It has become the standard for evaluating IDSs [6][32]. This dataset was created to realistically depict the variety and complexity of cyber threats and attacks, and it includes a wide range of situations for analysis. Its usefulness stems from its capacity to replicate complicated network environments, allowing academics and practitioners to successfully test andenhance IDS technology. The information, which was collected over ten days and has eighty columns, comprises fifteen different types of assaults, such as FTP and SSH brute force, numerous DoS and DDoS attacks, web attacks, SQL injection, infiltration, and botnet activities, making it a comprehensive tool for cybersecurity research.
4.2. LITNET-2020 Datasets
The LITNET-2020 dataset represents a significant advancement in cybersecurity research, dramatically enhancing network intrusion detection across traditional computer networks, Wireless Sensor Networks (WSN), and the Internet of Things (IoT). Moving beyond thelimitations of traditional network-intrusion benchmark datasets, which struggle to accurately reflect the complexity of modern network traffic and cyberattack patterns, LITNET-2020 introduces a variety of genuine network traffic scenarios as well as meticulously annotated examples of real attacks, avoiding the reliance on synthetic attacks commonly seen in sandboxed environments. The information systematically categorizes eighty-five particular network flow variables and defines twelve distinct types of attacks, covering a wide range of network behaviors from regular operations to cyber threats. Built on a solid infrastructure that uses Cisco routers and FortiGate (FG-1500D) firewalls to generate NetFlow data, as well as a sophisticated collector system for data management, storage, and analysis, LITNET-2020 provides a detailed and authentic resource for crafting and improving security measures to combat the dynamic nature of network vulnerabilities and threats[5][20].
5. EXPERIMENTAL SETUP
This study was carried out on a 64-bit Windows 10 system equipped with an Intel® Xeon® Silver 4314 processor and 62 GB of DDR4 RAM. Due to the large volume of data, the Python 3.11 environment was utilized, which benefits from regular updates that improve language features, performance, and libraries. Numpy and Pandas were utilized for data handling and preprocessing, and Scikit Learn was used for model training, evaluation, and metric assessment. Data visualization was done with Seaborn and Matplotlib. Further information can be found inthe subsections below.
5.1. Import Full Dataset
First, ensure that the CSE-CIC-IDS-2018 and LITNET-2020 dataset is produced in a manner thatallows for easy importing into your data analysis tool, such as CSV, Excel, JSON, or via a database linkage. Then, load the dataset using relevant libraries or tools from your chosen computer language, such as Python with pandas, R, or SQL, which may include file reading, database connection, or API interaction. After successfully importing the dataset, make sure it is kept in a variable or structured data format for future analysis. Tables 1-2provide more detailed information.
Table 1. CSE-CIC-IDS-2018 Network Traffic Distribution by Labelled Attacks Type and Ratio
Table 1 depicts the distribution of data types in a dataset used for NIDS with 16,233,002 instances. It divides the data into benign and attack categories, with benign traffic accounting for 83.07% (13,484,708 instances)[21]. DDoS attacks are the most common of the near-perfect s, accounting for 7.79% of the dataset, followed by DoS, brute force, botnet, infiltration, and web attacks, which account for the remainder. The data show that normal network traffic is much more common than harmful attempts, offering a thorough perspective for analyzing and upgrading network security measures.
Table 2. LITNET-2020 Network Traffic Distribution by Labelled Attacks Type and Ratio.
Table 2 categorizes numerous types of network attacks in a dataset of 39,603,674 entries, indicating their amounts and percentages of the total. The vast majority of the data (91.9709%) is classed as ‘none’, suggesting typical, non-attack traffic[21]. Among the attack kinds, the most common is ‘TCP SYN-flood’, which accounts for 3.9896% of the data, followed by ‘Code Red Worm’ attacks at 3.1707%. Other prominent assault types include ‘Smurf’, ‘UDP-flood’, and’ICMP-flood’, among others, each accounting for a minor fraction of the dataset. The variety and proportions of these attacks illustrate the dataset’s composition, demonstrating both benign and malicious traffic, as well as extensive enumeration for analytical and modeling purposes.
5.2. Understanding the Features and Describing them
Begin by evaluating the dataset’s structure using commands such as df.head() in Python with pandas or SELECT TOP 5 * in SQL to study the first data rows. It is critical to comprehend each column’s relevance and purpose, which may require studying the dataset’s data dictionary or associated documentation. Examine the data types used throughout characteristics, such as numerical, categorical, and date-time. To understand crucial statistical factors, construct summary statistics for numerical features such as mean, median, and standard deviation. Use visualization tools likehistograms, scatter plots, and box plots to investigate numerical feature distributions. Perform frequency counts on categorical features and show the distributions with bar plots or pie charts. Finally, evaluate and address missing values in each feature using imputation or removal to prepare the data for future analysis.
5.3. Count Unique Records
After familiarizing yourself with the dataset’s features, count the unique records to see if there are any duplicates, which are indicated by rows that have identical values across all columns. Use a command or function to count unique entries, using either a primary key or a precise combination of columns that uniquely identify each item. This technique will disclose the dataset’s total number of unique records, which will aid in ensuring data integrity and accuracy for future analysis.
5.4. Clean NaN, Missing, and Infinite values from the Dataset
Identify columns in your dataset with NaN (null) values, missing data, or infinite values (positive or negative infinity). Choose how to handle missing or infinite data, such as imputation (filling gaps with mean, median, or mode), eliminating rows with missing values, or using more advanced imputation methods. For dealing with infinite values, consider replacing them with appropriate numbers or eliminating the rows containing these values, especially if they are deemed outliers. This method ensures that the dataset is appropriately prepared, with consistent and useful data for subsequent analysis.After cleansing the dataset of NaN (Null), missing, and infinite values, it was revealed that there were no such values, suggesting that the dataset was complete and consistent, with no missing or undefined entries.
5.5. Clean Data by Removing Duplicate Rows
Identify and discover duplicate rows in the dataset using a duplicate-detecting command or function, concentrating on specific columns as appropriate. Depending on the objectives of your research, decide whether to keep the first occurrence of duplication while removing the others, or vice versa.To eliminate duplicate entries, use the appropriate approach; for example, in Python with pandas, use the df.drop_duplicates() function, but in SQL, use the DELETE statement with a WHERE clause. This method guarantees that the dataset is devoid of duplicate entries, leaving it clean and ready for future analysis. After successfully cleaning the data by deleting duplicate rows, Tables 3 and 4 provide extensive information.
Table 3. Comparative Analysis of Data Before and After Duplicate Removal in the CSE-CIC-IDS-2018 Dataset by Attack Type.
Table 3 compares the original and deduplicated network attack data from the CSE-CIC-IDS-2018 dataset. It demonstrates the effectiveness of deleting duplicate entries across a wide range of threats, including benign traffic and malicious activity such as DDoS, DoS, Brute Force, Botnet, Infiltration, and Web attacks. Initially, the collection had 16,233,002 records, with benign traffic accounting for 83.07% of the data. After deduplication, the total data was decreased to 12,017,831 items, with benign traffic increasing to 88.75%. Notably, all types of assaults witnessed a decrease in data volume, with DDoS, DoS, and Brute Force attacks experiencing the greatest decreases, indicating that these categories contained a considerable quantity of duplicate data. The approach successfully reduced dataset size while changing the ratio percentages of each attack type, emphasizing the necessity of data cleaning in preparing accurate and efficient datasets for analysis.
Table 4. Comparative Analysis of Data Before and After Duplicate Removal in the LITNET-2020 Dataset.
Table 4. depicts the impact of deleting duplicate records from the LITNET-2020 dataset, which originally had 39,603,674 data points. The ‘none’ category, which indicates no attack, accounted for around 91.97% of the original dataset and reduced slightly to 91.17% after duplicates were deleted. Attack categories such as ‘Smurf’ and ‘ICMP-flood’ have much lower numbers after duplication removal, affecting their proportions within the dataset. Other categories, such as ‘UDP-flood’ and ‘TCP SYN-flood’, preserved their previous numbers, indicating no duplication. The elimination process changed the ratios for each assault type, resulting in a more realistic depiction of the dataset’s composition. Following cleanup, the dataset now contains 35,196,472 records, ensuring that subsequent studies are based on unique and non-repetitive data, which is critical for reliable cybersecurity research and modeling.
5.6. Consolidating and Removing Unnecessary Features
Consolidating and removing unnecessary features includes simplifying a dataset by combining similar qualities and discarding those that are irrelevant to the model’s performance. This procedure improves data quality, reduces model complexity, and increases computing performance, making the dataset more suitable for predictive analytics and ML applications
5.6.1. Streamlining Features of the CSE-CIC-IDS-2018 Dataset
Following an initial review of the CSE-CIC-IDS-2018 dataset, which originally contained 16,233,002 entries across 80 attributes, duplicate record elimination reduced the dataset to 12,017,831 items. Furthermore, eight attributes with constant zero values for each record were removed, including “Bwd PSH Flags,” “Bwd URG Flags,” “FwdByts/b Avg,” “FwdPkts/b Avg,” “FwdBlk Rate Avg,” “BwdByts/b Avg,” “BwdPkts/b Avg,” and “BwdBlk Rate Avg”. The “Timestamp” feature was also eliminated to avoid biasing the model toward attack prediction or detection. Further data cleaning included the removal of entries with “NaN” values, as well as two attributes with constant infinite values, “Flow Byts/s” and “Flow Pkts/s”. With these changes, the improved dataset now contains 12,017,831 records and 69 features, making it better suited to classification tasks and additional analytical research.
5.6.2. Streamlining Features of the LITNET-2020 Dataset
5.6.2.1. Convert Data with Object Type to Numeric for Algorithm Preparation
To efficiently apply ML algorithms to your dataset, identify columns that contain object or nonnumeric data types, which are frequently filled with text descriptions of various network properties. These contain categorical variables like’sa’, ‘da’, ‘pr’, ‘_flag1’, through ‘_flag6’, ‘nh’, ‘nhb’, ‘ismc’, ‘odmc’, ‘idmc’, ‘osmc’,’mpls1′ to’mpls10′, ‘ra’, ‘eng’, ‘tr’, ‘icmp_dst_ip_b’, ‘icmp_src_ip’, ‘udp_ ‘tcp_f_s’, ‘tcp_f_n_a’ to ‘tcp_f_n_u’, ‘tcp_dst_p’, ‘tcp_src_dst_f_s’, ‘tcp_src_tftp’, ‘tcp_src_kerb’, ‘tcp_src_rpc’, ‘tcp_dst_p_src’, ‘smtp_dst’, ‘udp_p_r_range’, ‘p_range_dst’, ‘udp_src_p_0’, among others. It is critical to convert these variables to a numeric format using methods such as label encoding or custom mapping. This transformation is necessary to ensure that the dataset fits the numerical criteria of ML and statistical models, enabling for the extraction of relevant, data-driven insights and analysis.
5.6.2.2. Merge Features 1-6 to Start Timestamp and Features 7-12 to End Timestamp
Improve your dataset’s analytical capabilities by identifying and combining attributes that capture an event’s start and end times. Start times range from ‘ts_year’ to ‘ts_second’, whereas end times range from ‘te_year’ to ‘te_second’. Merge these data points into two new columns,’stimestamp’ for the event’s start and ‘etimestamp’ for its end, by aggregating the date and time values from each connected feature set. Ensure that these additional columns are in datetime format to allow for timebased analysis, which will improve the practicality and precision of any temporal research while also streamlining the overall dataset structure.
5.6.2.3. Remove Feature Names with Values of Zero
Following an initial review of the LITNET-2020 dataset, 39,603,674 records spanning 85 different features were identified. The removal of duplicate entries lowered the total number of records to 35,196,472. The dataset was further streamlined by combining features 1-6 into a single ‘stimestamp’ feature and features 7-12 into ‘etimestamp’, decreasing the feature count to 75. Subsequent statistical analysis revealed that fea-tures such as ‘fwd’, ‘opkt’, ‘obyt’, ‘smk’, ‘dmk’,’dtos’, ‘_dir’, ‘nh’, ‘nhb’,’svln’, ‘dvln’, ‘ ismc’, ‘odmc’, ‘idmc’, ‘osmc’, ‘mpls1’, ‘mpls2′,’mpls3’, ‘mpls4’, ‘mpls5’, ‘mpls6’, ‘mpls7’, ‘mpls8’, ‘mpls9’ , ‘mpls10′,’cl’, ‘sl’, ‘al’, ‘ra’, eng’, ‘tr’ had zero variance andwere there-fore removed, leaving 44 features in total. Furthermore, the ‘ID’ and ‘attack_a’ attributes were removed because they were not relevant to further investigation. This updated dataset now has 35,196,472 records and 42 features, making it suitable for more advanced processing.
5.7. Convert Attack Type with Multiple Class
To prepare your dataset for ML, use one-hot encoding to convert each category within a categorical variable into several class columns, then label encoding to assign a unique integer identification to each. Identify the column that contains information about attack kinds. Multiple categorization requires that the attacktype categories remain intact, protecting the integrity of all unique classifications. This approach ensures that the data is represented comprehensively, allowing for more detailed analysis and modeling.
5.8. Perform Statistical Analysis for Min, Max, STD, and Mean for Each Feature
To gain insights into the numerical features of your dataset, calculate descriptive statistics for your dataset’s numerical properties, such as the minimum, maximum, standard deviation, and mean. Using built-in tools, like df.describe() in pandas, provides a quick overview of these fundamental statistical indicators. This stage is crucial for understanding your data’s distribution, variability, and core patterns since it lays the groundwork for future data exploration and modeling activities.
5.9. Normalize Min-Max features
Min-max scaling is an effective method for standardizing numerical features in ML algorithms that need normalized input, such as neural networks and support vector machines. This procedure adjusts the features to a certain range, usually between 0 and 1, ensuring that their magnitudes are equal and preventing any single feature from excessively influencing the learning process. The min-max scaling is accomplished by using the mathematical equation shown as (1), where X is an original value and X′ is the normalized value[3][6][33][34].
5.10. Split a Dataset into Training and Testing Sets
Once the data has been prepared, it is ready for model testing, which involves splitting it into experimental and test datasets in an 80:20 ratio. This separation ensures that a significant percentage of the data is used for training and fine-tuning the model, while another, smaller sample is saved for testing its performance under situations similar to real-world applications[35]. Detailed information about this distribution is properly organized in Table 5-6.
Table 5. Training and Testing Dataset Distribution by
Table 6. Training and Testing Dataset Distribution by Attack Type for the LITNET-2020 Dataset.
5.11. Classification Model
Classification divides data into several groups, which is especially useful in IDS for distinguishing between benign and malicious network activity. It can be classified into two types binary classification, which deals with two distinct classes, and multi-class classification, which incorporates multiple classes. The addition of classes increases computational demand and time, which may reduce algorithm efficiency. During classification, data is examined to determine whether it is normal or abnormal. This method preserves existing data structures while allowing for the inclusion of new data points. Classification aids in the detection of unexpected patterns and irregularities, while it is most typically used to identify situations of misuse. Three ML and four DL models were utilized in this investigation.
ML models are divided into two categories standard ML approaches and DL models. In the traditional category, it mentions DT, which uses a tree-like model for decision-making, RF, an ensemble method that combines multiple DT for improved accuracy, and XGBoost, a highly efficient gradient boosting implementation that aims to optimize speed and performance. In the realm of DL, it highlights CNNs, which are especially effective for visual data analysis through hierarchical feature extraction, RNNs, designed to handle sequential data with their ability to model time-dependent information, DNNs, which are distinguished by their deep structure capable of modeling complex patterns, and MLP, a basic form.
5.12. Evaluation Model
The study assesses an intrusion detection system using measures such as accuracy, precision, recall, and F1-Score[6][11][16]. Accuracy is an important indication of anML model’s performance, particularly for classification tasks, because it calculates the proportion of right predictions out of the complete dataset. A model with high accuracy may dependably predict outcomes that match the actual observed events.A confusion matrix is a specific table structure that allows for the presentation of an algorithm’s performance, which is frequently supervised learning [33]. Each row of the matrix represents examples in an actual class, but each column represents instances in a predicted class, or vice versa. Here’s a simple definition of the terms.
• True Positives (TP) These are instances where the model properly anticipated a positive outcome. It signifies that the actual class was positive, and the model expected a positive outcome.
• True Negatives (TN) These are the instances correctly predicted as negative by the model. It means that the actual class was negative, and the model also predicted it as negative.
• False Positives (FP) These are situations where the model mistakenly projected a positive outcome, also known as Type I errors. It signifies that the actual class was negative, while the model expected a positive outcome.
• False Negatives (FN) These are situations where the model mistakenly anticipated a negative value, also known as Type II error. It means that while the actual class was positive, the model projected a negative outcome.
5.12.1. Accuracy
This metric assesses the model’s overall correctness and is calculated as the ratio of correct predictions to total predictions made. It is acceptable when the class distribution is similar. The formula is[36].
5.12.2. Precision
Precision measures a model’s ability to correctly identify only relevant instances. In other words, it represents the ratio of true positive forecasts to overall positive predictions. Here’s the formula[36].
5.12.3. Recall
Recall calculates the fraction of true positives properly detected by the model. Here’s the formula[36].
5.12.4. F1-Score
F1-Score is the harmonic mean of precision and recall, resulting in a balance of the two. It is especially effective when precision and recall need to be balanced, as well as when the class distribution is unequal. Here’s the formula[36].
5.13. Particle Swarm Optimization
Kennedy and Eberhart developed Particle Swarm Optimization (PSO) in 1995, borrowing inspiration from natural processes like bird flocking and fish schooling[12]. This strategy is used to solve optimization problems ranging from simple single-objective concerns to sophisticated multi-objective scenarios. PSO functions without requiring the problem’s gradient, making it especially useful for nonlinear tasks where gradients are not available. It is frequently used in domains such as engineering, economics, and computer science for a number of applications, including neural network training, function minimization, and product optimization[14][15][37]. The core mechanism of PSO is a set of potential solutions, known as particles, that navigate across a solution space. These particles move according to mathematical rules that govern their location and velocity, taking into account both their personal best positions and the best positions discovered by the swarm. This dual impact directs the entire swarm toward optimal solutions via iterative changes. PSO’s capacity to iteratively search for superior solutions while exploiting the swarm’s collective learning makes it an effective tool for optimization problems.
6. EXPERIMENTAL RESULT AND DISCUSSIONS
This section will go over and discuss the experimental results. Previously, we used two cybersecurity datasets the CSE-CIC-IDS-2018 and LITNET-2020 datasets, and prepared the data through a series of sub-steps such as Data Cleaning, Exploratory Data Analysis, Encoding, Normalization, and Data Splitting to guarantee it was ready for classification. We separated the data into two groups training and testing. During the Training Phase, we used three ML methods for data classification DT, RF, and XGBoost, which were motivated by previous research titled “Optimizing Intrusion Detection Systems Using a Three-Phase Strategy and the CSE-CIC-IDS-2018 Dataset.” We also employed four prominent DL models CNNs, RNNs, DNNs, and MLP. Following training, we evaluated the models’ performance in the Testing Phase using conventional measures such as accuracy, precision, recall, F1-Score, and aggregated average score. These metrics are used to assess the prediction effectiveness and dependability of each classifier, with higher scores indicating better performance. The goal is to find the most efficient model for further optimization using EPSO by fine-tuning the parameters to improve results.The classifiers’ performance will be presented in detail through tables and figures, providing both visual and numerical insights into their effectiveness. We want to get significant insights into future advances in IDS through extensive research and discussion. The following table and picture summarize the outcomes of our intensive experimentation.
Table 4. Performance Metrics for ML Classifiers Using CSE-CIC-IDS-2018 Dataset.
Table 7 compares the performance of different ML classifiers on the CSE-CIC-IDS-2018 dataset using metrics including Accuracy, Precision, Recall, and F1-Score. The RF classifier performs well, with high accuracy (0.990513682), and precision (0.96084231), but somewhat lower recall (0.859125805) and F1-Score (0.893265813), showing a reasonable balance between identifying true positives and minimizing false positives. The DT model beats others in terms of consistency across accuracy (0.974639471) and recall (0.975416011), with a near-perfect F1-Score (0.975026588), indicating that it captures the majority of positive cases without severe overprediction. XGboost also performs well in terms of accuracy (0.992428337) and precisionrecall ratio, resulting in a high F1-Score (0.927667115). CNNs, DNNs, and MLP classifiers, on the other hand, outperform ensemble approaches in terms of accuracy (0.887517154), precision (0.787686741), recall (0.887517178), and F1-Score (0.834627361), with MLP showing slightly adjusted accuracy (0.887517178). This consistency shows that these networks may encounter difficulties because of the dataset’s complexity or class imbalance. The RNNs model improves precision (0.861307000) while maintaining similar accuracy and F1-Score to the CNNs, DNNs, and MLP models, demonstrating a somewhat better but still limited competence in processing sequential data within this unique dataset. Overall, while ensemble models such as RF, DT, and XGboost perform better than neural net-work-based models (CNNs, DNNs, MLP, and RNNs) on the CSE-CIC-IDS-2018 dataset, highlighting the importance of model selection based on dataset characteristics and intended use case for cybersecurity threat detection.
Table 8. Performance Metrics for ML Classifiers Using LITNET-2020 Dataset
Table 8 compares ML classifiers’ performance on the LITNET-2020 dataset using four essential metrics accuracy, precision, recall, and F1-Score. Most classifiers, including RF, DT, XGBoost, DNNs, and MLP, had perfect scores of 1.000000000 across all criteria, suggesting that they accurately predicted every case. The CNNs and RNNs classifiers performed well, but could not achieve total perfection. The CNNs had an accuracy of 0.989821970, precision of 0.983374832, recall of 0.989821992, and F1-Score of 0.986295708. The RNNs had slightly greater accuracy and recall at 0.997327268 and 0.997327289, respectively, with a precision of 0.995994388 and an F1-Score of 0.996437922. This table shows how well these classifiers can predict and categorize network traffic, with CNNs and RNNs performing somewhat lower but still outstandingly compared to the other models.
Table 9.Comparative Performance of ML Classifiers on LITNET-2020 and CSE-CIC-IDS-2018 Datasets.
Table 9 shows the performance metrics for multiple ML classifiers used on two different datasets LITNET-2020 and CSE-CIC-IDS-2018. The evaluation criteria include Accuracy, Precision, Recall, and F1-Score, which provide a complete view of each classifier’s efficacy. For the LITNET-2020 dataset, the RF, DT, XGboost, DNNs, and MLP classifiers all received perfect scores (1.000000000), showing flawless prediction ability. The CNNs and RNNs classifiers likewise performed admirably, but somewhat below perfection, with RNNs exhibiting a modest fall in metrics compared to the others. When these classifiers were applied to the CSE-CIC-IDS2018 dataset, most models showed a significant fall in performance metrics, with the exception of the DT and XGboost classifiers, which maintained high scores comparable to those attained on the LITNET-2020 dataset. On the CSE-CIC-IDS-2018 dataset, the CNNs, DNNs, MLP, and RNNs classifiers performed significantly worse than LITNET-2020 in terms of accuracy and F1- Score. This comparative analysis demonstrates the variability in classifier performance across datasets, emphasizing the importance of dataset characteristics and the potential need for model adjustment or selection based on the unique problems given by each dataset. While some classifiers remain stable across both datasets, others may require modifications to improve performance for certain types of network traffic or attack detection scenarios.
Table 10. Average Performance Metrics of ML Classifiers on the LITNET-2020 and CSE-CIC-IDS-2018 datasets.
Table 10 illustrates the average performance of several ML classifiers across two cybersecurity datasets. Accuracy, Precision, Recall, F1-Score, and an overall Average score represent the classifiers’ ability to correctly identify security threats. Table 13 shows the average performance of various ML classifiers on two cybersecurity datasets. Accuracy, Precision, Recall, F1-Score, and an overall Average score measure the classifiers’ ability to correctly identify security threats. The RF classifier performs admirably, achieving near-perfect average accuracy and good scores across all criteria. The DT and XGBoost classifiers also produce excellent results, with the DT achieving the greatest average F1-Score and Average metrics, indicating a well-balanced performance in terms of both precision and recall. In comparison, CNNs, DNNs, and MLP classifiers have lower scores, notably in precision, which has an impact on their overall F1-Score and average performance. RNNs outperforms CNNs, DNNs, and MLP in terms of balance and precision, although it still lags behind tree-based classification algorithms. The table shows a wide range of performance, with tree-based classifiers beating neural network-based classifiers on these particular datasets. To better comprehend model performance evaluation metrics, the researcher presented the data in the form of a bar chart, as illustrated in Figure 3.
Figure 3. Comparative Bar Chart of Average ML Classifier Performance on LITNET-2020 and CSE-CIC-IDS-2018 Datasets.
The analysis in Tables 7-10 shows that the DT classifier outperforms alternative models on both the CSE-CIC-IDS-2018 and LITNET-2020 datasets. Based on this discovery, the researcher decided to improve the DT model by modifying three critical parameters max_depth, min_samples_split, and min_samples_leaf. To improve the model’s performance further, EPSO was used. This optimization yielded the following optimal parameter settings for the DT: ccp_alpha=0.0,class_weight=None,
criterion=’gini’,max_depth=31,max_features=None,max_leaf_nodes=None,
in_impurity_decrease=0.0,min_samples_split=7,min_samples_leaf=1,
min_weight_fraction_leaf=0.0, and a random_state of 42, with the splitter set to ‘best’. The favorable impact of these EPSO-facilitated tweaks is detailed in the following tables 14 and figures 4, which show that the DT model performs better now.
Figure 4.EPSO Driven Accuracy Improvements for DT on CSE-CIC-IDS-2018 Dataset.
The figure depicts the performance increase of a DT classifier applied to the CSE-CIC-IDS-2018 dataset after modifying its parameters using EPSO. The X-axis displays several parameter combinations for max_depth, min_samples_split, and min_samples_leaf. The Y axis represents the classifier’s accuracy.Starting with the usual DT settings (max_depth is infinite, min_samples_split is 2, and min_samples_leaf is 1), there is a significant improvement in accuracy. This improvement reaches its pinnacle when the parameters are set to max_depth=31, min_samples_split=7, and min_samples_leaf=1, indicating the most successful parameter configuration optimized by EPSO, as indicated by a peak accuracy of around 0.997229118. Beyond this peak, accuracy levels off, meaning that making more changes to the parameters has little effect on improving accuracy. This high point is the most efficient iteration of the DT model, which was optimized using EPSO for the maximum predicted accuracy.
Table 11. Improving ML Classifier Performance with EPSO on CSE-CIC-IDS-2018 Dataset.
Table 11 compares the performance of various classifiers on the CSE-CIC-IDS-2018 dataset, focusing on the DT and its upgraded variant utilizing upgraded Particle Swarm Optimization (EPSO). The typical DT is highly efficient, with an F1-Score of 97.50% and an accuracy of nearly 99.67%. However, applying EPSO to the DT improves its metrics greatly, increasing accuracy to around 99.72% and F1-Score to 97.45%. This enhancement highlights EPSO’s importance in optimizing the classifier’s parameters, hence improving predicted accuracy and reliability in detecting network intrusions. The EPSO-enhanced DT is the most effective model, demonstrating the promise of optimization techniques for improving ML tools for cybersecurity applications.
Table 12. Average Performance Metrics for EPSO-Tuned ML Classifiers on the LITNET-2020 and CSE-CIC-IDS-2018 datasets.
Table 12 summarizes the average performance metrics of ML classifiers optimized using EPSO on the LITNET-2020 and CSE-CIC-IDS-2018 datasets. Accuracy, Precision, Recall, and F1- Score are among the metrics used, with an overall average score for each classifier calculated as well. The EPSO-tuned DT performs best, with virtually flawless scores across all criteria, followed by the regular DT and RF classifiers, demonstrating their excellent efficacy in classification tasks. XGBoost also performs well, though with slightly lower averages. DLmodels, such as CNNs, DNNs, MLP, and RNNs, perform well but with lower average scores. The findings highlight EPSO’s ability to refine models and offer precise and accurate forecast outcomes, particularly in difficult tasks like intrusion detection
Figure 5 shows a radar chart that compares the performance of numerous ML classifiers on the CSE-CIC-IDS-2018 and LITNET-2020 datasets using measures such as accuracy, precision, recall, F1-Score, and overall average score. The EPSO DT is featured as one of the classifiers, and its performance is demonstrated to be superior in nearly all criteria. This model, which was optimized using EPSO, stands out for its near-perfect accuracy of around 0.9986. Its Precision, which assesses the accuracy of positive predictions, is likewise quite high, nearly approaching 1.0. The Recall, which measures how well the model detects all relevant instances, is also quite high, implying that the EPSO DT excels at finding true positives. The F1-Score, which is the harmonic mean of Precision and Recall, is likewise approaching 0.99, indicating that the model performs well in both Precision and Recall. Finally, the Average score, which is presumably the mean of these metrics, is exceptionally high, demonstrating the classifier’s overall robustness. In summary, the EPSO DT performs remarkably well across all tested measures, illustrating the value of employing EPSO to optimize DT in ML applications, particularly intrusion detection with the CSE-CIC-IDS-2018 dataset. The optimization process resulted in a model that is both accurate and trustworthy, with few trade-offs between identifying as many true positives as feasible and keeping a low false positive rate. To increase understanding of model performance evaluation criteria, the researcher displayed the data as a radar graph, as illustrated in Figure 5.
Figure 5. Comparative Radar Chart of ML Classifier Performances on CSE-CIC-IDS-2018 and LITNET-2020 Datasets.
In the discussion of experimental results, we look at how ML classifiers performed after applying EPSO to two cybersecurity datasets CSE-CIC-IDS-2018 and LITNET-2020. The data was rigorously prepared for categorization by performing stages such as cleaning and normalization. The data was split into training and testing sets and processed with ML methods such as DT, RF, and XGBoost, as well as DL models such as CNNs, RNNs, DNNs, and MLPs. The classifiers were then compared on conventional criteria such as accuracy, precision, recall, and F1-Score, with the DT model consistently outperforming the others. The investigation found that the DT, when fine-tuned with EPSO to optimize important parameters, performed exceptionally well, achieving near-perfect accuracy and precision. This improved DT model proved to be the most effective, demonstrating EPSO’s powerful impact on classifier performance. A radar chart comparison revealed that the EPSO-optimized DT outperformed all other measures, validating its efficacy in intrusion detection tasks on the relevant datasets. The optimization resulted in a classifier that not only reliably identifies threats but also has a low false positive rate, effectively balancing accuracy with recall. The findings hint at a promising approach for future advances in IDS, with EPSO emerging as a useful tool for improving ML models to attain high levels of prediction accuracy and reliability.
Table 13. Comparing Our Intrusion Detection Model to Others Using CSE-CIC-IDS-2018 and LITNET-2020 Datasets.
Table 13 compares alternative intrusion detection models using the CSE-CIC-IDS-2018 andLITNET-2020 datasets, with a focus on key parameters including accuracy, precision, recall, and F1- Score. Our model, which uses an EPSO DT, stands out for its superior performance across both datasets. It has an excellent accuracy of around 99.86%, precision of 99.16%, recall of 98.38%, and F1-Score of 98.73%. This shows a high level of accuracy in identifying both actual and potential intrusions while avoiding false positives and negatives. The comparison highlights the efficacy of our EPSO DT model in improving the reliability and security of NIDS, establishing it as a highly competitive strategy in the field of cybersecurity.
7. CONCLUSION
In the conclusion portion of the work, we discuss the extensive testing and analysis of ML classifiers utilizing EPSO on two significant cybersecurity datasets CSE-CIC-IDS-2018 and LITNET-2020. The data was meticulously prepared, including cleaning, normalization, and partitioning into training and testing sets. The data was analyzed using a variety of ML approaches, including DT, RF, and XGBoost, as well as DL models such as CNNs, RNNs, DNNs, and MLPs. Each model was thoroughly tested with measurement metrics such as accuracy, precision, recall, and F1-Score. The DT model performed well, especially after EPSO fine-tuning, with nearly perfect scores across all criteria, suggesting its ability to appropriately identify data. The EPSO-optimized DT outperformed all other classifiers in radar chart comparisons, demonstrating its ability to identify network intrusions.
This extended research demonstrates EPSO’s effectiveness as a potent optimization tool for improving the performance of ML classifiers in cybersecurity applications. The study concluded that with thorough preparation and optimization utilizing approaches such as EPSO, ML models can attain high levels of precision and dependability, which are critical for effective IDS. The findings provide useful insights for future cybersecurity advancements, underlining the importance of continuously improving analytical models to keep up with the changing nature of network threats.
Future research should focus on improving optimization approaches beyond EPSO, validating model generalizability across several datasets, and implementing models in real-time contexts. Exploring hybrid techniques and increasing resilience to adversarial attacks will improve IDS efficacy. This study’s limitations include potential overfitting and the need for model validation across changing network conditions, necessitating further research into scalability, efficiency, and flexibility to emerging threats, particularly in IoT and edge computing contexts.
CONFLICTS OF INTEREST
The authors declare no conflict of interest.
ACKNOWLEDGMENTS
Our heartfelt gratitude goes to the team at Suan Dusit University for their invaluable insights and continuous support throughout this research. We are also appreciative to Suan Dusit University for providing crucial computersystem resources that were critical in the execution and analysis of our study.
REFERENCES
[1] A. Khraisat, I. Gondal, P. Vamplew, and J. Kamruzzaman, “Survey of intrusion detection systems: techniques, datasets and challenges,” Cybersecurity, vol. 2, no. 1, Dec. 2019, doi: 10.1186/s42400-
019-0038-7.
[2] M. Verkerken, L. D’hooge, T. Wauters, B. Volckaert, and F. De Turck, “Towards Model Generalization for Intrusion Detection: Unsupervised Machine Learning Techniques,” Journal of Network and Systems Management, vol. 30, no. 1, Jan. 2022, doi: 10.1007/s10922-021-09615-7.
[3] S. Songma, T. Sathuphan, and T. Pamutha, “Optimizing Intrusion Detection Systems in Three Phases on the CSE-CIC-IDS-2018 Dataset,” Computers, vol. 12, no. 12, Dec. 2023, doi: 10.3390/computers12120245.
[4] V. Dutta, M. Choraś, M. Pawlicki, and R. Kozik, “A deep learning ensemble for network anomaly and cyber-attack detection,” Sensors (Switzerland), vol. 20, no. 16, pp. 1–20, Aug. 2020, doi: 10.3390/s20164583.
[5] R. Damaseviciuset al., “Litnet-2020: An annotated real-world network flow dataset for network intrusion detection,” Electronics (Switzerland), vol. 9, no. 5, May 2020, doi: 10.3390/electronics9050800.
[6] S. Chimphlee and W. Chimphlee, “Machine learning to improve the performance of anomaly-based network intrusion detection in big data,” Indonesian Journal of Electrical Engineering and Computer Science, vol. 30, no. 2, pp. 1106–1119, May 2023, doi: 10.11591/ijeecs.v30.i2.pp1106-1119.
[7] A. Touzene, A. Al Farsi, and N. Al Zeidi, “HIGH PERFORMANCE NMF BASED INTRUSION DETECTION SYSTEM FOR BIG DATA IOT TRAFFIC,” International Journal of Computer Networks and Communications, vol. 16, no. 2, pp. 43–58, Mar. 2024, doi: 10.5121/ijcnc.2024.16203.
[8] W. Chimphlee and S. Chimphlee, “Hyperparameters optimization XGBoost for network intrusion detection using CSE-CIC-IDS 2018 dataset,” IAES International Journal of Artificial Intelligence (IJ-AI), vol. 13, no. 1, p. 817, Mar. 2024, doi: 10.11591/ijai.v13.i1.pp817-826.
[9] A. H. Halbouni, T. S. Gunawan, M. Halbouni, F. A. A. Assaig, M. R. Effendi, and N. Ismail, “CNN-IDS: Convolutional Neural Network for Network Intrusion Detection System,” in Proceeding of 2022 8th International Conference on Wireless and Telematics, ICWT 2022, Institute of Electrical and Electronics Engineers Inc., 2022. doi: 10.1109/ICWT55831.2022.9935478.
[10] S. Nayyar, S. Arora, and M. Singh, “Recurrent Neural Network Based Intrusion Detection System,” in 2020 International Conference on Communication and Signal Processing (ICCSP), 2020, pp. 136–140. doi: 10.1109/ICCSP48568.2020.9182099.
[11] V. Hnamte and J. Hussain, “Dependable intrusion detection system using deep convolutional neural network: A Novel framework and performance evaluation approach,” Telematics and Informatics Reports, vol. 11, Sep. 2023, doi: 10.1016/j.teler.2023.100077.
[12] A. A. Awad, A. F. Ali, and T. Gaber, “An improved long short term memory network for intrusion detection,” PLoS One, vol. 18, no. 8 August, Aug. 2023, doi: 10.1371/journal.pone.0284795.
[13] E. Jamal, M. Reza, and G. Jamal, “Intrusion Detection System Based on Multi-Layer Perceptron Neural Networks and Decision Tree,” in International Conference on Information and Knowledge Technology, 2015.
[14] I. Ariyati, S. Rosyida, K. Ramanda, V. Riyanto, S. Faizah, and Ridwansyah, “Optimization of the Decision Tree Algorithm Used Particle Swarm Optimization in the Selection of Digital Payments,” in Journal of Physics: Conference Series, IOP Publishing Ltd, Nov. 2020. doi: 10.1088/1742- 6596/1641/1/012090.
[15] S. Budilaksonoet al., “Comparison of Data Mining Algorithm: PSO-KNN, PSO-RF, and PSO-DT to Measure Attack Detection Accuracy Levels on Intrusion Detection System,” in Journal of Physics: Conference Series, Institute of Physics Publishing, Mar. 2020. doi: 10.1088/1742- 6596/1471/1/012019.
[16] L. Yang, J. Li, L. Yin, Z. Sun, Y. Zhao, and Z. Li, “Real-time intrusion detection in wireless network: A deep learning-based intelligent mechanism,” IEEE Access, vol. 8, pp. 170128–170139, 2020, doi: 10.1109/ACCESS.2020.3019973.
[17] M. A. Hossain and M. S. Islam, “Ensuring network security with a robust intrusion detection system using ensemble-based machine learning,” Array, vol. 19, Sep. 2023, doi: 10.1016/j.array.2023.100306.
[18] Canadian Institute for Cybersecurity, “CSE-CIC-IDS2018 on AWS.” Accessed: Mar. 08, 2024. [Online]. Available: https://www.unb.ca/cic/datasets/ids-2018.html
[19] J. L. Leevy and T. M. Khoshgoftaar, “A survey and analysis of intrusion detection models based on CSE-CIC-IDS2018 Big Data,” J Big Data, vol. 7, no. 1, Dec. 2020, doi: 10.1186/s40537-020- 00382-x.
[20] S. Nath, D. Pal, and S. Saha, “Detection of Cyber-attacks using Deep Convolution in Honeyed Framework,” in 2023 International Conference on Disruptive Technologies, ICDT 2023, Institute of Electrical and Electronics Engineers Inc., 2023, pp. 79–87. doi: 10.1109/ICDT57929.2023.10150916.
[21] V. Bulavas, V. Marcinkevičius, and J. Rumiński, “Study of Multi-Class Classification Algorithms’ Performance on Highly Imbalanced Network Intrusion Datasets,” Informatica (Netherlands), vol. 32, no. 3, pp. 441–475, 2021, doi: 10.15388/21-INFOR457.
[22] H. Al-Zoubi and S. Altaamneh, “A Feature Selection Technique for Network Intrusion Detection based on the Chaotic Crow Search Algorithm,” in 2022 International Conference on Intelligent Data Science Technologies and Applications, IDSTA 2022, Institute of Electrical and Electronics Engineers Inc., 2022, pp. 54–60. doi: 10.1109/IDSTA55301.2022.9923108.
[23] S. Shilpashree, S. C. Lingareddy, N. G. Bhat, and G. Sunil Kumar, “Decision tree: A machine learning for intrusion detection,” International Journal of Innovative Technology and Exploring Engineering, vol. 8, no. 6 Special Issue 4, pp. 1126–1130, Apr. 2019, doi: 10.35940/ijitee.F1234.0486S419.
[24] Institute of Electrical and Electronics Engineers, “An Anomaly-Based Intrusion Detection System for the Smart Grid Based on CART Decision Tree,”
[25] Md. A. M. Hasan, M. Nasser, S. Ahmad, and K. I. Molla, “Feature Selection for Intrusion Detection Using Random Forest,” Journal of Information Security, vol. 07, no. 03, pp. 129–140, 2016, doi: 10.4236/jis.2016.73009.
[26] N. Farnaaz and M. A. Jabbar, “Random Forest Modeling for Network Intrusion Detection System,” in Procedia Computer Science, Elsevier B.V., 2016, pp. 213–217. doi: 10.1016/j.procs.2016.06.047.
[27] S. S. Dhaliwal, A. Al Nahid, and R. Abbas, “Effective intrusion detection system using XGBoost,” Information (Switzerland), vol. 9, no. 7, Jun. 2018, doi: 10.3390/info9070149.
[28] J. Lansky et al., “Deep Learning-Based Intrusion Detection Systems: A Systematic Review,” IEEE Access, vol. 9. Institute of Electrical and Electronics Engineers Inc., pp. 101574–101599, 2021. doi: 10.1109/ACCESS.2021.3097247.
[29] M. Maithem and G. A. Al-Sultany, “Network intrusion detection system using deep neural networks,” in Journal of Physics: Conference Series, IOP Publishing Ltd, Mar. 2021. doi: 10.1088/1742-6596/1804/1/012138.
[30] W. F. Kamil and I. J. Mohammed, “Deep learning model for intrusion detection system utilizing convolution neural network,” Open Engineering, vol. 13, no. 1, Jan. 2023, doi: 10.1515/eng-2022- 0403.
[31] M. Almiani, A. AbuGhazleh, A. Al-Rahayfeh, S. Atiewi, and A. Razaque, “Deep recurrent neural network for IoT intrusion detection system,” Simul Model Pract Theory, vol. 101, May 2020, doi: 10.1016/j.simpat.2019.102031.
[32] W. Chimphlee and S. Chimphlee, “INTRUSION DETECTION SYSTEM (IDS) DEVELOPMENT USING TREE-BASED MACHINE LEARNING ALGORITHMS,” International Journal of Computer Networks and Communications, vol. 15, no. 4, pp. 93–109, Jul. 2023, doi: 10.5121/ijcnc.2023.15406.
[33] F. Abdou Vadhil, M. LemineSalihi, and M. Farouk Nanne, “Machine learning-based intrusion detection system for detecting web attacks,” IAES International Journal of Artificial Intelligence (IJ-AI), vol. 13, no. 1, p. 711, Mar. 2024, doi: 10.11591/ijai.v13.i1.pp711-721.
[34] G. Ketepalli and P. Bulla, “Data Preparation and Pre-processing of Intrusion Detection Datasets using Machine Learning,” in 6th International Conference on Inventive Computation Technologies, ICICT 2023 – Proceedings, Institute of Electrical and Electronics Engineers Inc., 2023, pp. 257–262. doi: 10.1109/ICICT57646.2023.10134025.
[35] K. K. Dobbin and R. M. Simon, “Optimally splitting cases for training and testing high dimensional classifiers,” BMC Med Genomics, vol. 4, 2011, doi: 10.1186/1755-8794-4-31.
[36] T. T. Huynh and H. T. Nguyen, “EFFECTIVE MULTI-STAGE TRAINING MODEL FOR EDGE COMPUTING DEVICES IN INTRUSION DETECTION,” International Journal of Computer Networks and Communications, vol. 16, no. 1, pp. 13–31, Jan. 2024, doi: 10.5121/ijcnc.2024.16102.
[37] Y. S. J. K. Russell C. Eberhart, Swarm Intelligence, 1st ed. Elsevier, 2001.
[38] J. Kim, Y. Shin, and E. Choi, “An Intrusion Detection Model based on a Convolutional Neural Network,” Journal of Multimedia Information System, vol. 6, no. 4, pp. 165–172, Dec. 2019, doi: 10.33851/jmis.2019.6.4.165.
[39] M. A. Khan, “HCRNNIDS: Hybrid convolutional recurrent neural network-based network intrusion detection system,” Processes, vol. 9, no. 5, 2021, doi: 10.3390/pr9050834.
AUTHORS
Surasit Songma his Ph.D. in Information Technology from Rangsit University. Currently he holds the position of Assistant Professor in the Cyber Security program at the Faculty of Science and Technology, Suan Dusit University. His current research interests include intrusion detection systems, network security, big data analytics, and machine learning. He has published several papers in peer-reviewed journals and has actively participated in various international conferences. He is a dedicated researcher with a passion for advancing the field of information technology through his work. He can contacted at email: surasit_son@dusit.ac.th
Watcharakorn Netharn* his M.Sc. in Telecommunication and Computer Networks for Rangsit University. Currently he holds the position of Assistant Professor in the Information Technology program at the Faculty of Science and Technology, Suan Dusit University. His current research interest include satellite, network and data science. He is the *corresponding author for this article and can be contacted at watcharakorn_net@dusit.ac.th.
Siriluck Lorpunmanee his PhD in Computer Science from the University Technology of Malaysia. Currently, he holds the position of Assistant Professor in the Computer Science program at Suan Dusit University in Thailand. His current research interests include Artificial Intelligence Algorithm, Simulation modelling, and Grid computing. He has published several papers in peer-reviewed journals and has actively participated in various international conferences. He is a dedicated researcher with a passion for advancing the field of computer science through his work. He can contacted at email: siriluck_lor@dusit.ac.th