Research Article

MC-MLDCNN: Multichannel Multilayer Dilated Convolutional Neural Networks for Web Attack Detection

Table 1

Summary of the related work.

StudyArchitectureDatasetRemarksAttack typePerformance

Mehta et al. [28]Logistic regression, random forest, SVM, and moreSelf-collectedEvaluation of supervised/unsupervised ML algorithms (logistic regression achieves the best performance)SQL injectionAcc: 93.21, Rcll: 77.38, Prec: 100
Louk and Tama [29]Bagging ensemble of gradient boosting decision treesHIKARI-2021, NSL-KDD, UNSW-NB15Bagging ensemble of GBM performs the bestIntrusion detectionAcc: 91.57, Rcll: 86.18, Prec: 98.67, : 91.50
Althubiti et al. [30]LSTMCSIC 2010Manual feature extraction is appliedIntrusion detectionAcc: 99.97, Rcll: 99.50, Prec: 99.50
Yin et al. [32]RNNNSL-KDDFeature extraction is performed automaticallyIntrusion detectionAcc: 81.29
Xing et al. [33]LSTM + bidirectional RNNReal-world datasetsFeature extraction is performed automaticallyCyberattack rateMSE: 3,628,266, MAD: 463.2715, PMAD: 0.012, MAPE: 0.013
Kasim [34]Sparse autoencoder + principal component + light gradient boosted machineISCX-URLSparse autoencoder + principal component are applied for feature learning; light gradient boosted machine is used for feature selection and classificationPhishing attacksAcc: 99.6, : 99.58, FPR: 0.001
Dawadi et al. [35]Layered LSTMSelf-collectedManual feature extraction is performed by analyzing attack-indicator features of IDS ISCX 2012, 2019 DDoS CIC, and CISC 2010 datasetsDDoS attack, XSS and SQL injectionAccDDoS: 97.57, AccXSS/SQL: 89.34
Hao et al. [36]Bi-LSTMCSIC 2010Word2vec is applied for feature representationWeb attackAcc: 98.35, Rcll: 98.17, Prec: 99.00, : 98.58, FPR: 0.014
Alaoui and Nfaoui [38]Ensemble of LSTMsCSIC 2010Word2vec is applied for feature representationWeb attackAcc: 78.95, Rcll: 78.41, Prec: 81.54, : 77.57
Zhang et al. [39]CNNCSIC 2010Word-level embedding is appliedWeb attackAcc: 93.35, Rcll: 96.49, FPR: 0.0137
Tian et al. [40]M-ResNet + FastTextCSIC 2010, FWAF, HttpParams datasetConcatenation of Word2vec and TF-IDF is used for feature vectors. M-ResNet is applied for feature discrimination purpose. Classification is performed using a FastText classifierWeb attackAcc: 99.41, Rcll: 98.91, DRN: 99.55, : 77.57
Luo et al. [43]Ensemble of M-ResNets, LSTM, and CNNCSIC 2010 real-world datasetConcatenation of Word2vec and TF-IDF is used for feature representationWeb attackAcc: 99.47, Rcll: 99.29, Prec: 99.70, FPR: 0.0033
Rong et al. [44]CNNSelf-collectedCharacter-level embedding is appliedInjection attacksPrec: 100, Rcll: 99.7, FPR: 0.0002
Odumuyiwa and Chibueze [45]CNNECML/PKDD 2007, CSIC 2010Character-level embedding is appliedHTTP injection attacksAcc: 96.39, Prec: 98.83, Rcll: 95.00, : 97.00, FPR: 0.020
Saxe and Berlin [3]CNNSelf-collectedCharacter-level embedding is appliedMalicious URLs, paths, registry keysAUCURL: 99.30, AUCFilePath: 97.80, AUCRegistryKeys: 99.20
Gong et al. [19]CNN + LSTMCSIC 2010Character-level embedding is appliedWeb attackAcc: 97.79, Prec: 98.54, Rcll: 96.04, : 97.27
Jemal et al. [2]CNN + LSTMCSIC 2010ASCII-level embedding is appliedWeb attackAcc: 99.25, Prec: 97.73, Rcll: 99.35, : 98.53
Vinayakumar et al. [46]CNN, RNN, LSM, CNN-LSTM, and moreSelf-collectedCharacter-level embedding is applied. The most effective models are LSTM and CNN-LSTMMalicious URLAccLSTM: 99.95, AUCLSTM: 99.99, AccCNN-LSTM: 99.96, AUCCNN-LSTM: 99.99
Hung et al. [47]CNNSelf-collectedCharacter-level and word-level embeddings are usedMalicious URLAUC: 99.29
Kasim [48]Autoencoder + SVMCICIDS, NSL-KDD, virtual traffic DDoS attackAutoencoder is used for feature learning and dimensionality reduction. SVM is used for classificationDDoS attackAcc: 99.41, Prec: 99.66, Rcll: 99.67, : 99.67
Yi et al. [49]Autoencoders, restricted Boltzmann machine, deep belief networks, CNN, and moreThe datasets used in the literature are coveredReview of application of deep learning approaches. It covers feature representation, model training, model robustness enhancement techniques, and problems and challenges of developmentsNetwork attacks
Pillai and Sharma [50]Stacked autoencoder (SAE) + denoising autoencoder (DAE) + generative adversarial network (GAN) + deep Boltzmann machine + Bi-LSTMCSIC 2010v2Concatenated form of SAE and DAE outputs is fed into a GAN for feature representation. The deep Boltzmann machine is used to identify attacks. For identifying the different types of attacks, Bi-LSTM is usedWeb attackPrec: 98.78, Rcll: 98.78, : 98.78
Thajeel et al. [51]ML, deep learningFrequently used datasets are coveredLiterature review of ML and deep learning methodologies and advancementsXSS attacks
Rizvi et al. [52]Dilated CNNCSE-CIC-IDS2018, CIC-IDS2017Manual feature selection is appliedIntrusion detectionAcc: 99.98

DRN: the percentage of all normal requests that are classified as normal; MSE: mean square error; MAD: mean absolute deviation; PMAD: percent mean absolute deviation; MAPE: mean absolute percentage error. Acc, Prec, Rcll, , and FPR: accuracy, precision, recall, score, and false-positive rate, respectively. The overall performance of relative datasets is given. In studies containing multiple dataset evaluations, the provided performance is related to the bold dataset.