Analysis of Autoencoders for Network Intrusion Detection †
<p>The overall process of intrusion detection when using the dimension reduction (i.e., Encoder) and reconstruction (i.e., Decoder) methods of autoencoders.</p> "> Figure 2
<p>The process of building autoencoder models. <span class="html-italic">Train</span> denotes the data for training, and <span class="html-italic">Validation</span> denotes the data for validation. The process consists of three phases: training, model selection, and validation.</p> "> Figure 3
<p>An example autoencoder model architecture with symmetrical encoder and decoder networks. <span class="html-italic">X</span> and <math display="inline"><semantics> <mover accent="true"> <mi>X</mi> <mo>^</mo> </mover> </semantics></math> respectively represent the model’s input and its reconstructed output.</p> "> Figure 4
<p>The results of evaluation using the NSL-KDD dataset when the (<b>5</b>,<b>32</b>) model is employed.</p> "> Figure 5
<p>The results of evaluation using the NSL-KDD dataset when the (<b>5</b>,<b>64</b>) model is employed.</p> "> Figure 6
<p>The results of evaluation using the NSL-KDD dataset when the (<b>7</b>,<b>64</b>) model is employed.</p> "> Figure 7
<p>NSL-KDD Reconstruction Distance Density Plot (Normal and R2L attack data) comparing the best and the worst performance cases with the (<b>5</b>,<b>64</b>) model. Blue denotes normal data and orange denotes R2L attack data. A line depicts the kernel density estimated plot, and bar plot depicts the histogram. The dotted line indicates the threshold value, which divides the normal and the attack classes. The X-axis denotes Z-score. The Y-axis is density and limited to 2 for better visibility. (<b>a</b>) The best case (Latent size = 3, MCC = 0.712, Threshold = 2.488). (<b>b</b>) The worst case (Latent size = 1, MCC = 0.624, Threshold = 2.294).</p> "> Figure 8
<p>The results of evaluation using the IoTID20 dataset when the (<b>5</b>,<b>32</b>) model is employed.</p> "> Figure 9
<p>The results of evaluation using the IoTID20 dataset when the (<b>5</b>,<b>64</b>) model is employed.</p> "> Figure 10
<p>The results of evaluation using the IoTID20 dataset when the (<b>7</b>,<b>64</b>) model is employed.</p> "> Figure 11
<p>IoTID20 Reconstruction Distance Distribution Density Plot (Normal and attack data) comparing two similarly performing cases with the (<b>5</b>,<b>64</b>) model. Blue denotes the normal data and orange denotes the attack data. A line depicts the kernel density estimated plot, and bar plot depicts the histogram. The dotted line indicates the threshold value, which divides the normal and the attack classes. The X-axis denotes Z-score. The Y-axis is density and limited to 0.8 for better visibility. (<b>a</b>) The best case (Latent size = 28, MCC = 0.593, Threshold = 0.132). (<b>b</b>) A near-best case (Latent size = 9, MCC=0.59, Threshold = 0.132).</p> "> Figure 12
<p>MCC per device for the N-BaIoT dataset when the (<b>5</b>,<b>32</b>) model is employed.</p> "> Figure 13
<p>MCC per device for the N-BaIoT dataset when the (<b>5</b>,<b>64</b>) model is employed.</p> "> Figure 14
<p>MCC per device for the N-BaIoT dataset when the (<b>7</b>,<b>64</b>) model is employed.</p> "> Figure 15
<p>N-BaIoT Provision PT-737E Reconstruction Distance Distribution Density Plot (Normal and attack data) comparing the best and the worst performing cases when the (<b>5</b>,<b>64</b>) model was used. Blue denotes the normal data and orange denotes the attack data. A line depicts the kernel density estimated plot, and bar plot depicts the histogram. The dotted line indicates the threshold value, which divides the normal and the attack classes. The X-axis denotes Z-score. The Y-axis is density and limited to 1.6 for better visibility. (<b>a</b>) The best case (Latent Size = 29, MCC = 0.976, Threshold = 1.338). (<b>b</b>) The worst case (Latent Size = 1, MCC = 0.860, Threshold = 0.678).</p> ">
Abstract
:1. Introduction
- How many hidden layers and how many neurons for each layer are necessary? Is a bigger model better for intrusion detection?
- What dimension of the latent layer is appropriate to represent the data? For instance, the model may not represent the data well if its latent layer has only one neuron. In that case, the latent layer is likely to be too generic to characterize the data. On the other hand, the model would not benefit the advantage of generalization that occurs during the compression/reconstruction process if its latent layer contains an excessive number of neurons.
- How do we set a threshold that divides the normal and abnormal data, when labeled data are not available?
- What metrics are best to represent the difference between the input and its output (e.g., L1, L2, Cross-Entropy)?
2. Related Work
2.1. Intrusion Detection with Deep-Learning and Ensemble Learning
2.2. Autoencoders for Feature Reduction
2.3. Autoencoders for Anomaly Detection
3. Datasets
3.1. NSL-KDD Data
3.2. IoTID20 Data
3.3. N-BaIoT Data
4. Approach
4.1. Overview
- (1)
- Preprocessing: Given a training data set, we first remove attack samples from the training set so that only normal samples are used to train the autoencoder. Then we vectorize the data by converting the feature values into numerical vectors using the min-max normalization. The features are scaled using the minimum and maximum values obtained from the training set. Additional preprocessing methods applied to each particular data set are described in detail in Section 3.
- (2)
- Training: The autoencoder model is trained to reconstruct any given input as close as possible to the normal log patterns of the training set and L2 norm is used as reconstruction loss. Theoretically, the autoencoder model trained with normal logs is expected to recover any input log as close as possible to the learned normal log patterns. Due to this, if a given input in fact belongs to an attack class, it is likely that the recovered output would be significantly different from the original input. This means that the class of an input sample is determined based on the difference between the input and its output; if the difference is greater than a set threshold, it is classified as an attack. Otherwise, it is classified as normal.
- (3)
- Model Selection: At each epoch during the training phase, the Area under the Curve (AuC) score of the Receiver Operating Characteristic (ROC) curve is calculated using the validation data which contain the normal and attack logs. The AuC score enables us to measure the model’s overall performance as the model weights change. During the training phase, we record the highest AuC score and its corresponding model weights to obtain the best-performing model. When the training phase is completed, we use the model weights that resulted in the highest AuC score.
- (4)
- Threshold Selection: Since the threshold on the reconstruction distance between an input and its output divides the normal and abnormal (i.e., attack in this study) classes, selecting its value has a tremendous impact on the ML-NIDS performance. We evaluate the model’s performance using the validation set, varying the threshold value to determine the optimal figure for the trained model. We use the Z-score of standard normal distribution as the threshold metric. Using the Z-score instead of the reconstruction error values before normalization can facilitate the reproduction of the classifier even when reconstruction error values in unnormalized form change depending on a data set.To set the threshold Z-score, we first standardize reconstruction distances with the mean and variance values of reconstruction distances calculated using only the normal samples in the validation data. To find the best-performing threshold, we inspect the range of (−4,4) with 0.01 interval, which covers 99.994% of the standard normal distribution. We determine the Z-score value that best divides the normal and abnormal classes in the validation data as the threshold .
- (5)
- Evaluation: Finally, we evaluate the ML-NIDS performance using the test data. We compute the Z-score of each log’s reconstruction error using the mean and variance of normal samples in the validation set calculated in Step 4. We classify each log as an attack if its Z-score exceeds . Otherwise, it is classified as normal. By comparing these predicted labels and the test data’s original labels, we calculate the model’s evaluation metrics (i.e., accuracy, TPR, FPR, and MCC explained in Section 4.3).
4.2. Model Design
4.3. The Evaluation Metrics
5. Results and Analysis
5.1. Model Configurations
5.2. Results: NSL-KDD Data
5.3. Results: IoTID20 Data
5.4. Results: N-BaIoT Data
5.5. Discussions
5.5.1. Model Structure and Performance
5.5.2. Analysis of Reconstruction Errors and Threshold
5.5.3. Threats-to-Validity of Experimental Results
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Conflicts of Interest
References
- Barreno, M.A. The Future of Cryptography under Quantum Computers. Undergraduate Thesis, Dartmouth College, Hanover, NH, USA, 21 July 2002. [Google Scholar]
- Chen, L.; Jordan, S.; Liu, Y.K.; Moody, D.; Peralta, R.; Perlner, R.; Smith-Tone, D. Report on Post-Quantum Cryptography; US Department of Commerce, National Institute of Standards and Technology: Gaithersburg, MD, USA, 2016; Volume 12.
- Bernstein, D.J.; Lange, T. Post-quantum cryptography. Nature 2017, 549, 188–194. [Google Scholar] [CrossRef]
- Lizama-Perez, L.A.; López R, J.M. Non-Invertible Public Key Certificates. Entropy 2021, 23, 226. [Google Scholar] [CrossRef]
- Cyphort. Cyphort Data Sheet. 2017. Available online: http://go.cyphort.com/rs/181-NTN-682/images/CYPHORT_DataSheet.pdf (accessed on 21 June 2021).
- RSA. RSA Netwitness Logs and Packets. 2017. Available online: https://www.rsa.com/content/dam/en/data-sheet/rsa-netwitness-logs-and-packets.pdf (accessed on 21 June 2021).
- Sapre, S.; Ahmadi, P.; Islam, K.R. A Robust Comparison of the KDDCup99 and NSL-KDD IoT Network Intrusion Detection Datasets Through Various Machine Learning Algorithms. arXiv 2019, arXiv:abs/1912.13204. [Google Scholar]
- Siddique, K.; Akhtar, Z.; Khan, F.A.; Kim, Y. KDD Cup 99 Data Sets: A Perspective on the Role of Data Sets in Network Intrusion Detection Research. Computer 2019, 52, 41–51. [Google Scholar] [CrossRef]
- Chitrakar, R.; Huang, C. Selection of candidate support vectors in incremental SVM for network intrusion detection. Comput. Secur. 2014, 45, 231–241. [Google Scholar] [CrossRef]
- Ambusaidi, M.A.; He, X.; Nanda, P.; Tan, Z. Building an intrusion detection system using a filter-based feature selection algorithm. IEEE Trans. Comput. 2016, 65, 2986–2998. [Google Scholar] [CrossRef] [Green Version]
- Sahu, S.; Mehtre, B.M. Network intrusion detection system using J48 Decision Tree. In Proceedings of the 2015 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Kerala, India, 10–13 August 2015; pp. 2023–2026. [Google Scholar]
- Khan, S.; Kifayat, K.; Kashif Bashir, A.; Gurtov, A.; Hassan, M. Intelligent intrusion detection system in smart grid using computational intelligence and machine learning. Trans. Emerg. Telecommun. Technol. 2020, 32, e4062. [Google Scholar] [CrossRef]
- Hettich, S. KDD Cup 1999 Data. 2007. Available online: http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html (accessed on 21 June 2021).
- Tavallaee, M.; Bagheri, E.; Lu, W.; Ghorbani, A. A detailed analysis of the KDD CUP 99 data set. In Proceedings of the 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications, Ottawa, ON, Canada, 8–10 July 2009; pp. 1–6. [Google Scholar]
- Salama, M.A.; Eid, H.F.; Ramadan, R.A.; Darwish, A.; Hassanien, A.E. Hybrid intelligent intrusion detection scheme. In Soft Computing in Industrial Applications; Springer: Berlin, Germany, 2011; pp. 293–303. [Google Scholar]
- Kim, J.; Kim, J.; Thu, H.L.T.; Kim, H. Long short term memory recurrent neural network classifier for intrusion detection. In Proceedings of the 2016 International Conference on Platform Technology and Service (PlatCon), Jeju, Korea, 15–17 February 2016; pp. 1–5. [Google Scholar]
- Alom, M.Z.; Bontupalli, V.; Taha, T. Intrusion detection using deep belief networks. In Proceedings of the 2015 IEEE National Aerospace and Electronics Conference (NAECON), Piscataway, NJ, USA, 16–19 June 2015; pp. 339–344. [Google Scholar]
- Kuypers, M.A.; Maillart, T.; Paté-Cornell, E. An Empirical Analysis of Cyber Security Incidents at a Large Organization. Department of Management Science and Engineering, Stanford University, School of Information, UC Berkeley. Available online: http://fsi.stanford.edu/sites/default/files/kuypersweis_v7.pdf (accessed on 30 July 2016).
- Petersen, R. Data Mining for Network Intrusion Detection: A Comparison of Data Mining Algorithms and an Analysis of Relevant Features for Detecting Cyber-Attacks. Undergraduate Thesis, Mid Sweden University, Östersund, Sweden, 2015. [Google Scholar]
- Garcia-Teodoro, P.; Diaz-Verdejo, J.; Maciá-Fernández, G.; Vázquez, E. Anomaly-based network intrusion detection: Techniques, systems and challenges. Comput. Secur. 2009, 28, 18–28. [Google Scholar] [CrossRef]
- Kaur, H.; Singh, G.; Minhas, J. A review of machine learning based anomaly detection techniques. arXiv 2013, arXiv:1307.7286. [Google Scholar] [CrossRef]
- Buczak, A.L.; Guven, E. A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun. Surv. Tutor. 2015, 18, 1153–1176. [Google Scholar] [CrossRef]
- Song, J.; Takakura, H.; Okabe, Y.; Kwon, Y. Correlation analysis between honeypot data and IDS alerts using one-class SVM. In Intrusion Detection Systems; InTech: London, UK, 2011; pp. 173–192. [Google Scholar]
- Ishida, M.; Takakura, H.; Okabe, Y. High-performance intrusion detection using optigrid clustering and grid-based labelling. In Proceedings of the 2011 IEEE/IPSJ International Symposium on Applications and the Internet, Munich, Bavaria, Germany, 18–21 July 2011; pp. 11–19. [Google Scholar]
- Mirsky, Y.; Doitshman, T.; Elovici, Y.; Shabtai, A. Kitsune: An Ensemble of Autoencoders for Online Network Intrusion Detection. In Proceedings of the 25th Annual Network and Distributed System Security Symposium, NDSS 2018, San Diego, CA, USA, 18–21 February 2018. [Google Scholar]
- Aygun, R.C.; Yavuz, A.G. Network anomaly detection with stochastically improved autoencoder based models. In Proceedings of the 2017 IEEE 4th International Conference on Cyber Security and Cloud Computing (CSCloud), New York, NY, USA, 26–28 June 2017; pp. 193–198. [Google Scholar]
- Zavrak, S.; İskefiyeli, M. Anomaly-Based Intrusion Detection From Network Flow Features Using Variational Autoencoder. IEEE Access 2020, 8, 108346–108358. [Google Scholar] [CrossRef]
- Tang, T.A.; Mhamdi, L.; McLernon, D.; Zaidi, S.A.R.; Ghogho, M. Deep learning approach for network intrusion detection in software defined networking. In Proceedings of the International Conference on Wireless Networks and Mobile Communications (WINCOM), Fez, Morocco, 26–29 October 2016; pp. 258–263. [Google Scholar]
- Zhang, H.; Li, J.L.; Liu, X.M.; Dong, C. Multi-dimensional feature fusion and stacking ensemble mechanism for network intrusion detection. Future Gener. Comput. Syst. 2021, 122, 130–143. [Google Scholar] [CrossRef]
- Yong, B.; Wei, W.; Li, K.C.; Shen, J.; Zhou, Q.; Wozniak, M.; Połap, D.; Damaševičius, R. Ensemble machine learning approaches for webshell detection in Internet of things environments. Trans. Emerg. Telecommun. Technol. Wiley Online Libr. 2020, e4085. [Google Scholar] [CrossRef]
- Hemalatha, J.; Roseline, S.A.; Geetha, S.; Kadry, S.; Damaševičius, R. An Efficient DenseNet-Based Deep Learning Model for Malware Detection. Entropy 2021, 23, 344. [Google Scholar] [CrossRef] [PubMed]
- Li, Y.; Ma, R.; Jiao, R. A Hybrid Malicious Code Detection Method based on Deep Learning. Int. J. Softw. Eng. Appl. 2015, 9, 205–216. [Google Scholar] [CrossRef]
- Tao, X.; Kong, D.; Wei, Y.; Wang, Y. A Big Network Traffic Data Fusion Approach Based on Fisher and Deep Auto-Encoder. Information 2016, 7, 20. [Google Scholar] [CrossRef] [Green Version]
- Javaid, A.; Niyaz, Q.; Sun, W.; Alam, M. A Deep Learning Approach for Network Intrusion Detection System. In Proceedings of the 9th EAI International Conference on Bio-inspired Information and Communications Technologies (Formerly BIONETICS), New York City, NY, USA, 3–5 December 2015; pp. 21–26. [Google Scholar] [CrossRef] [Green Version]
- Shone, N.; Ngoc, T.N.; Phai, V.D.; Shi, Q. A deep learning approach to network intrusion detection. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 2, 41–50. [Google Scholar] [CrossRef] [Green Version]
- Al-Qatf, M.; Lasheng, Y.; Al-Habib, M.; Al-Sabahi, K. Deep Learning Approach Combining Sparse Autoencoder With SVM for Network Intrusion Detection. IEEE Access 2018, 6, 52843–52856. [Google Scholar] [CrossRef]
- Madani, P.; Vlajic, N. Robustness of deep autoencoder in intrusion detection under adversarial contamination. In Proceedings of the 5th Annual Symposium and Bootcamp on Hot Topics in the Science of Security, Raleigh, NC, USA, 10–11 April 2018; pp. 1–8. [Google Scholar]
- Yu, Y.; Long, J.; Cai, Z. Network intrusion detection through stacking dilated convolutional autoencoders. Secur. Commun. Netw. 2017. [Google Scholar] [CrossRef]
- Thing, V.L. IEEE 802.11 network anomaly detection and attack classification: A deep learning approach. In Proceedings of the 2017 IEEE Wireless Communications and Networking Conference (WCNC), San Francisco, CA, USA, 19–22 March 2017; pp. 1–6. [Google Scholar]
- Zhang, B.; Yu, Y.; Li, J. Network Intrusion Detection Based on Stacked Sparse Autoencoder and Binary Tree Ensemble Method. In Proceedings of the 2018 IEEE International Conference on Communications Workshops (ICC Workshops), Kansas City, MO, USA, 20–24 May 2018. [Google Scholar]
- Dutta, V.; Choraś, M.; Pawlicki, M.; Kozik, R. A Deep Learning Ensemble for Network Anomaly and Cyber-Attack Detection. Sensors 2020, 20, 4583. [Google Scholar] [CrossRef] [PubMed]
- Agustin Parmisano and Sebastian Garcia and Maria Jose Erquiaga. A Labeled Dataset with Malicious and Benign IoT Network Traffic. 2020. Available online: https://www.stratosphereips.org/datasets-iot23 (accessed on 9 March 2021).
- Damasevicius, R.; Venckauskas, A.; Grigaliunas, S.; Toldinas, J.; Morkevicius, N.; Aleliunas, T.; Smuikys, P. LITNET-2020: An annotated real-world network flow dataset for network intrusion detection. Electronics 2020, 9, 800. [Google Scholar] [CrossRef]
- NetML-2020 Dataset. 2020. Available online: https://eval.ai/web/challenges/challenge-page/526/overview (accessed on 9 March 2021).
- Martín, M.L.; Carro, B.; Sánchez-Esguevillas, A.; Lloret, J. Conditional Variational Autoencoder for Prediction and Feature Recovery Applied to Intrusion Detection in IoT. Sensors 2017, 17, 1967. [Google Scholar] [CrossRef] [Green Version]
- Meidan, Y.; Bohadana, M.; Mathov, Y.; Mirsky, Y.; Shabtai, A.; Breitenbacher, D.; Elovici, Y. N-BaIoT—Network-Based Detection of IoT Botnet Attacks Using Deep Autoencoders. IEEE Pervasive Comput. 2018, 17, 12–22. [Google Scholar] [CrossRef] [Green Version]
- Shahid, M.R.; Blanc, G.; Zhang, Z.; Debar, H. Anomalous Communications Detection in IoT Networks Using Sparse Autoencoders. In Proceedings of the 2019 IEEE 18th International Symposium on Network Computing and Applications (NCA), Cambridge, MA, USA, 26–28 September 2019. [Google Scholar] [CrossRef] [Green Version]
- Pa, Y.M.P.; Suzuki, S.; Yoshioka, K.; Matsumoto, T.; Kasama, T.; Rossow, C. IoTPOT: Analysing the rise of IoT compromises. In Proceedings of the 9th USENIX Workshop on Offensive Technologies (WOOT), Washington, DC, USA, 10–11 August 2015. [Google Scholar]
- Ullah, I.; Mahmoud, Q.H. A Scheme for Generating a Dataset for Anomalous Activity Detection in IoT Networks. In Advances in Artificial Intelligence; Goutte, C., Zhu, X., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 508–520. [Google Scholar]
- Yang, Y.; Zheng, K.; Wu, C.; Yang, Y. Improving the Classification Effectiveness of Intrusion Detection by Using Improved Conditional Variational AutoEncoder and Deep Neural Network. Sensors 2019, 19, 2528. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Liu, H.; Lang, B. Machine Learning and Deep Learning Methods for Intrusion Detection Systems: A Survey. Appl. Sci. 2019, 9, 4396. [Google Scholar] [CrossRef] [Green Version]
- Boughorbel, S.; Jarray, F.; El-Anbari, M. Optimal classifier for imbalanced data using Matthews Correlation Coefficient metric. PLoS ONE 2017, 12, e0177678. [Google Scholar] [CrossRef]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Nykvist, C.; Larsson, M.; Sodhro, A.H.; Gurtov, A. A lightweight portable intrusion detection communication system for auditing applications. Int. J. Commun. Syst. 2020, 33, e4327. [Google Scholar] [CrossRef]
Category | Training | Validation | Test | |
---|---|---|---|---|
Benign | - | 60,642 | 6701 | 9711 |
Attack | DoS | 41,325 | 4602 | 7458 |
U2R | 50 | 2 | 200 | |
R2L | 908 | 87 | 2754 | |
Probe | 10,451 | 1205 | 2421 |
Category | Train | Validation | Test | |
---|---|---|---|---|
Benign | - | 32,056 | 4001 | 4016 |
Attack | Mirai | 332,569 | 41,744 | 41,364 |
Scan | 60,249 | 7409 | 7607 | |
DoS | 47,519 | 5865 | 6007 | |
MITM | 28,233 | 3560 | 3584 |
Device | Label | Train | Validation | Test |
---|---|---|---|---|
Danmini Doorbell | Benign | 39,675 | 4895 | 4978 |
Attack | 774,963 | 96,935 | 96,852 | |
Ecobee Thermostat | Benign | 10,528 | 1303 | 1282 |
Attack | 658,173 | 82,284 | 82,306 | |
Provision PT-737E | Benign | 49,753 | 6212 | 6189 |
Attack | 612,85 | 76,614 | 76,637 | |
Philips B120N/10 Baby Monitor | Benign | 140,069 | 17,574 | 17,579 |
Attack | 738,873 | 92,293 | 92,271 | |
SimpleHome XCS7-1002-WHT | Benign | 37,302 | 4620 | 4663 |
Attack | 655,143 | 81,685 | 81,643 |
Model Structure (Depth, Size) | Number of Neurons for Each Layer |
---|---|
(5,32) | input – 32 – 16 – latent layer – 16 – 32 – output |
(5,64) | input – 64 – 32 – latent layer – 32 – 64 – output |
(7,64) | input – 64 – 32 – 16 – latent layer – 16 – 32 – 64 – output |
Model Structure | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
---|---|---|---|---|---|---|---|---|
(5,32) | 4 | 2.846 | 0.840 | 0.757 | 0.051 | 0.705 | 0.842 | 0.960 |
(5,64) | 3 | 2.488 | 0.848 | 0.782 | 0.065 | 0.712 | 0.853 | 0.960 |
(7,64) | 9 | 2.840 | 0.847 | 0.777 | 0.061 | 0.711 | 0.852 | 0.959 |
Model Structure | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
---|---|---|---|---|---|---|---|---|
(5,32) | 4 | 2.120 | 0.887 | 0.851 | 0.066 | 0.778 | 0.895 | 0.961 |
(5,64) | 3 | 2.090 | 0.885 | 0.859 | 0.081 | 0.771 | 0.894 | 0.961 |
(7,64) | 9 | 2.340 | 0.882 | 0.826 | 0.045 | 0.774 | 0.888 | 0.971 |
Model Structure | TP | FP | FN | TN |
---|---|---|---|---|
(5,32) | 10,920 | 636 | 1913 | 9075 |
(5,64) | 11,019 | 768 | 1814 | 8925 |
(7,64) | 10,604 | 428 | 2229 | 9283 |
Model Structure | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
---|---|---|---|---|---|---|---|---|
(5,32) | 14 | 0.041 | 0.945 | 0.963 | 0.313 | 0.595 | 0.971 | 0.912 |
(5,64) | 28 | −0.013 | 0.947 | 0.966 | 0.333 | 0.594 | 0.972 | 0.913 |
(7,64) | 14 | 0.060 | 0.944 | 0.961 | 0.307 | 0.590 | 0.970 | 0.911 |
Model Structure | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
---|---|---|---|---|---|---|---|---|
(5,32) | 14 | −0.350 | 0.951 | 0.971 | 0.329 | 0.614 | 0.974 | 0.909 |
(5,64) | 28 | −0.150 | 0.952 | 0.971 | 0.324 | 0.617 | 0.974 | 0.915 |
(7,64) | 14 | 0.060 | 0.952 | 0.970 | 0.321 | 0.618 | 0.974 | 0.918 |
Model Structure | TP | FP | FN | TN |
---|---|---|---|---|
(5,32) | 55,683 | 1060 | 2841 | 2956 |
(5,64) | 55,553 | 1017 | 2971 | 2999 |
(7,64) | 52,480 | 815 | 6044 | 3201 |
(a) Danmini Doorbell | ||||||||
---|---|---|---|---|---|---|---|---|
Model | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
(5,32) | 11 | 0.988 | 1.000 | 1.000 | 0.006 | 0.996 | 1.000 | 0.999 |
(5,64) | 19 | 1.108 | 1.000 | 1.000 | 0.006 | 0.996 | 1.000 | 0.999 |
(7,64) | 8 | 1.028 | 1.000 | 1.000 | 0.006 | 0.996 | 1.000 | 0.999 |
(b) Philips B120N / 10 Baby Monitor | ||||||||
Model | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
(5,32) | 10 | 3.130 | 0.996 | 0.998 | 0.013 | 0.986 | 0.998 | 0.999 |
(5,64) | 25 | 3.270 | 0.997 | 0.999 | 0.011 | 0.989 | 0.998 | 0.999 |
(7,64) | 8 | 3.344 | 0.996 | 0.998 | 0.012 | 0.987 | 0.998 | 0.999 |
(c) SimpleHome XCS7-1002-WHT Security Camera | ||||||||
Model | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
(5,32) | 11 | 0.438 | 0.995 | 0.999 | 0.074 | 0.952 | 0.997 | 0.990 |
(5,64) | 20 | 1.050 | 0.996 | 0.999 | 0.051 | 0.963 | 0.998 | 0.998 |
(7,64) | 12 | 0.846 | 0.996 | 0.999 | 0.058 | 0.959 | 0.998 | 0.996 |
(d) Provision PT-737E | ||||||||
Model | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
(5,32) | 10 | 0.854 | 0.991 | 0.994 | 0.050 | 0.937 | 0.995 | 0.992 |
(5,64) | 29 | 1.338 | 0.997 | 0.999 | 0.027 | 0.976 | 0.998 | 0.999 |
(7,64) | 9 | 1.102 | 0.995 | 0.998 | 0.040 | 0.967 | 0.998 | 0.999 |
(e) Ecobee Thermostat | ||||||||
Model | Latent Size | Threshold | Accuracy | TPR | FPR | MCC | F1 | AUC |
(5,32) | 13 | 0.802 | 0.998 | 0.999 | 0.067 | 0.928 | 0.999 | 0.997 |
(5,64) | 4 | 0.408 | 0.999 | 1.000 | 0.064 | 0.959 | 0.999 | 0.999 |
(7,64) | 4 | 0.366 | 0.999 | 1.000 | 0.079 | 0.952 | 0.999 | 0.999 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Song, Y.; Hyun, S.; Cheong, Y.-G. Analysis of Autoencoders for Network Intrusion Detection. Sensors 2021, 21, 4294. https://doi.org/10.3390/s21134294
Song Y, Hyun S, Cheong Y-G. Analysis of Autoencoders for Network Intrusion Detection. Sensors. 2021; 21(13):4294. https://doi.org/10.3390/s21134294
Chicago/Turabian StyleSong, Youngrok, Sangwon Hyun, and Yun-Gyung Cheong. 2021. "Analysis of Autoencoders for Network Intrusion Detection" Sensors 21, no. 13: 4294. https://doi.org/10.3390/s21134294