Using the backpropagation algorithm, update the weight vectors between the top and bottom layers. Which may be used to calculate the \(\overrightarrow{w}\):
The effort is split into two halves. Prioritize fine-tuning the hyper-parameters first, then look at how these parameters impact \(DGN\) performance. The second comparison is between DNN and multilayer perceptrons, support vector machines, and \(k\)-NN algorithms. TensorFlow serves as the backend, while Python is used to implement all algorithms. The accuracy metric is used to gauge the effectiveness of comparing binary classification to different machine learning techniques and methods already in use. This metric is depicted in Eq. (6) and assesses the discrepancy between expected and actual classes.
6.1. Comparison
With Eq. (6), the expression is used to calculate the accuracy for Table 2 presents each classifier and the comparison outcomes. The results show that KNN, SVM, MLP, and DNN are improved in order of accuracy. It illustrates how better packet classification can be achieved using machine learning techniques. The three ACL datasets' proposed \(DGN\) model can improve classification performance throughout training and testing compared to earlier techniques. Some discussions regarding the existing works are provided below:
1) k-NN: It is executed using the scikit-library with the classifier model. The essential hyper-parameter number of k-NN is changed from 1 to 15; while the remaining parameters are considered as default.
2) SVM: It is extensively utilized for classification problems. Here, two kernels like radial bias and linear kernel is utilized. The algorithm is executed with scikit library. While optimizing SVM with using RBF and linear, the hyper-parameters like cost and gamma are concentrated. The cost is selected as \({10}^{4}, {10}^{3}, {10}^{2}, 10, 1\) and \({10}^{4}, {10}^{3}, {10}^{2}, {10}^{1}, 1, {10}^{-1}, {10}^{-2}, {10}^{-3}\). The gamma values are tested for the datasets. For algorithm optimization with RBF kernel, this work attains superior performance than linear one. The highest values is attained across the dataset using cost (\(c\)) values of 100, 1000 and 10, 000 and gamma (\(\gamma\)) value is equal to 0.01.
3) MLP: It works in a feed-forward manner and executed in scikit-learn. The classifier function is adopted and the hyper-parameters are modified. It uses hidden layer with 56 neurons and successive epochs are utilized for \(ACL1\_K,ACL1\_5K and ACL1\_10K\) where 110 hidden nodes are utilized.
4) DNN: It is executed in Keras with Tensorflow. Open source software’s is utilized to train the network model. The model is trained and the performance is measured against the testing data. With the split function learned from available sklearn package, the dataset is partitioned into testing and training. To eliminate over-fitting issue during the training process, the stopping technique is utilized. This approach stops the training process; when the loss function sets several epochs. Therefore, this approach preserves the network from over-fitting on the training samples. To model the DNN model, the weights are initialized to random values, the bias is set as 0, and the learning rate is 0.1. The relationship among the hidden layers and accuracy performance of packet classification is analyzed.
Table 2
Classifier | ACL1_1K | ACL1_5K | ACL1_10K |
Training | Testing | Training | Testing | Training | Testing |
k-NN | 89 | 87 | 86 | 85 | 88 | 86 |
SVM | 90 | 90 | 88 | 86 | 89 | 88 |
MLP | 91 | 90 | 89 | 88 | 92 | 90 |
DNN | 95 | 93 | 95 | 92 | 96 | 94 |
DGN | 96 | 95 | 96 | 95 | 97 | 95 |
When compared to KNN, SVM, and MLP, For ACL1_ 1K, DNN's training accuracy increases by 5.9%, 4.7%, and 3.9%; for ACL1_ 5K, it increases by 10%, 3%, and 4.3%; and finally, for ACL1_ 10K, it increases by 8%, 6.8%, and 3.6%. Comparable benchmarks include ACL1 _1K, ACL1 _5K, and ACL1 _10K the testing accuracy against KNN, SVM, and MLP is improved by 6.8%, 3.5%, and 2.8%, 7.6%, 5.6%, and 4%, and 7.9%, 6.4%, and 4%, respectively. The total performance of the suggested algorithm in comparison to other methods is shown in Fig. 3 to Fig. 5. \(DGN\) can enhance the performance of packet categorization due to the deep learning model's potential for extraction and recognition. Although the current searching methods are more effective at classifying packets, they are still unable to handle the vast majority of classification rules. The proposed \(DGN\) model is supposed to be able to handle this issue. Due to automated feature extraction, which significantly reduced the system's cost to identify the features, this \(DGN\) model produced the best results for inbound packets.