Next Article in Journal
Perioperative Chemotherapy with FLOT Scheme in Resectable Gastric Adenocarcinoma: A Preliminary Correlation between TRG and Radiomics
Next Article in Special Issue
Semantic Fusion with Deep Learning and Formal Ontologies for Evaluation of Policies and Initiatives in the Smart City Domain
Previous Article in Journal
Survey on Data Hiding Based on Block Truncation Coding
Previous Article in Special Issue
Gaze Tracking Using an Unmodified Web Camera and Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Phishing Webpage Classification via Deep Learning-Based Algorithms: An Empirical Study

1
Malaysia-Japan International Institute of Technology (MJIIT), Universiti Teknologi Malaysia Kuala Lumpur, Jalan Sultan Yahya Petra, Kuala Lumpur 54100, Malaysia
2
School of Computing, Faculty of Engineering, Universiti Teknologi Malaysia, Johor Bahru 80000, Johor, Malaysia
3
Media and Games Center of Excellence (MagicX), Universiti Teknologi Malaysia, Skudai, Johor Bahru 81310, Johor, Malaysia
4
Center for Basic and Applied Research, Faculty of Informatics and Management, University of Hradec Kralove, Rokitanskeho 62, 500 03 Hradec Kralove, Czech Republic
5
Tokyo Metropolitan College of Industrial Technology, Tokyo 140-0011, Japan
6
Andalusian Research Institute in Data Science and Computational Intelligence (DaSCI), University of Granada, 18001 Granada, Spain
7
i-SOMET Incorporated Association, Morioka 020-0000, Japan
8
Regional Research Center, Iwate Prefectural University, Iwate 028-4211, Japan
*
Author to whom correspondence should be addressed.
Submission received: 18 August 2021 / Revised: 24 September 2021 / Accepted: 29 September 2021 / Published: 3 October 2021

Abstract

:
Phishing detection with high-performance accuracy and low computational complexity has always been a topic of great interest. New technologies have been developed to improve the phishing detection rate and reduce computational constraints in recent years. However, one solution is insufficient to address all problems caused by attackers in cyberspace. Therefore, the primary objective of this paper is to analyze the performance of various deep learning algorithms in detecting phishing activities. This analysis will help organizations or individuals select and adopt the proper solution according to their technological needs and specific applications’ requirements to fight against phishing attacks. In this regard, an empirical study was conducted using four different deep learning algorithms, including deep neural network (DNN), convolutional neural network (CNN), Long Short-Term Memory (LSTM), and gated recurrent unit (GRU). To analyze the behaviors of these deep learning architectures, extensive experiments were carried out to examine the impact of parameter tuning on the performance accuracy of the deep learning models. In addition, various performance metrics were measured to evaluate the effectiveness and feasibility of DL models in detecting phishing activities. The results obtained from the experiments showed that no single DL algorithm achieved the best measures across all performance metrics. The empirical findings from this paper also manifest several issues and suggest future research directions related to deep learning in the phishing detection domain.

1. Introduction

In the past few years, deep learning (DL) techniques have proven to be an effective solution among applications across multiple disciplines, including Internet of Things (IoT), intrusion detection system (IDS), ransomware detection, etc. [1,2,3,4,5]. Numerous researchers in cyber security have shifted their attention towards DL algorithms. Notably, researchers and security experts have also recognized its significance in the phishing detection domain [6,7,8]. During the last few years, website phishing has become one of the most common phishing attacks in cyberspace. Therefore, various anti-phishing solutions have been developed to detect phishing threats early to minimize the security risks and protect the end-users. Among them, website phishing detection based on DL algorithms has caught much attention in recent studies. Security strategies based on DL mechanisms have become increasingly popular to deal with evolving phishing attacks [9,10,11]. There are numerous types of DL techniques designed to solve a specific problem or meet a system’s particular requirement; each has its advantages and disadvantages [2,12,13].
Hence, choosing the right approach best fitted to a target application is not an easy task. Especially when phishers keep changing their attacking tactics to leverage the systems’ vulnerabilities and the users’ unawareness, selecting an inappropriate algorithm would lead to unpredicted outcomes, resulting in a waste of effort and eventually affecting the model’s accuracy and efficiency [14]. Therefore, choosing an effective phishing detection model, high in performance accuracy and low in computational power, is a challenging task. The fine-tuning process of DL architectures is another issue that needs to be considered. Motivated to solve this problem, this paper adopted an empirical approach to explore the performance of several DL algorithms, such as deep neural network (DNN), convolutional neural network (CNN), Long Short-Term Memory (LSTM), and gated recurrent unit (GRU). This paper also identified the parameter settings for each DL model and investigated the effects of changing these parameters on the model’s performance accuracy. The final goal of this paper was to choose the best DL algorithm with the neural network architecture that produced the maximum accuracy with the minimum computational consumption. The findings from the empirical analysis of this paper also highlight the overlooked issues and future perspectives that encourage researchers to solve these problems.
This paper continues our previous research work that described a systematic literature review on phishing detection and machine learning [15]. One of the findings from this work suggested that DL algorithms appeared to be an effective solution for detecting phishing attacks, yet they have not been fully exploited. In this regard, an empirical analysis was conducted in this study to explore the most recent DL techniques used for phishing detection. In this paper, the following contributions were made via the empirical study:
  • We exploited the state-of-the-art DL algorithms and compared their performance using numerous evaluation metrics;
  • We identified the most common parameters and examined their influences on the performance of four DL models;
  • We highlighted several issues based on the findings from the empirical experiments and recommended possible solutions to address these issues.
The remainder of this paper is structured as follows. Section 2 provides a short description of four different DL architectures, and reviews previous studies on these algorithms in the phishing detection domain. The methodology used to conduct this research is presented in Section 3, including experiment setup, website features, DL models, and parameter optimization. Section 4 summarizes the findings, highlights the issues observed from the obtained results, and suggests possible solutions for future research directions. Finally, the conclusion and future work of this research is given in Section 5.

2. Literature Review

This section provides a general overview of four different DL algorithms, including DNN, CNN, LSTM, and GRU. Previous research on these four types of DL architectures in the phishing detection domain is also discussed. In each study, we analyzed the neural network architecture, parameter optimization, and performance metrics to achieve a comprehensive understanding of each DL model’s design, implementation, and evaluation. Finally, the novelty of this research work compared with other related studies in the same research area is also highlighted.

2.1. Deep Neural Network (DNN)

Deep neural network (DNN) is one of the most common types of DL algorithm widely used in the cybersecurity domain. DNN is well-known among DL architectures due to its success in a wide range of applications [13], its ability to express complex functions with fewer parameters, and its capability to facilitate feature extraction and representation learning [16]. However, DNN requires a substantial amount of labeled data for training. In addition, it still suffers from insufficient parameter selection techniques [17], and the learning process is time-consuming [13]. Despite its disadvantages, several research works have been conducted to examine the effectiveness of applying DNN in detecting phishing webpages. Table A1 in Appendix A summarizes previous studies related to DNN for phishing detection, in the literature.
As a single classifier, DNN was used in [18,19,20] to train the classification system for the detection of phishing websites. Instead of using DNN as a stand-alone classifier, the authors in [21,22] combined it with other DL algorithms to build a model to differentiate between malicious and benign URLs. It was observed that among these DNN-based models, parameters settings play an essential role in determining the system’s performance accuracy. Nevertheless, some studies [21,22] did not mention any of the hyper-parameters in the design of the neural network architecture, while other papers [19,20] only specified a few of them without performing parameter optimization. The authors in [18] made additional effort in fine-tuning the parameters, but not all were included.
Moreover, the performance metric is another crucial factor that needs to be considered when analyzing and evaluating a phishing detection system. Previous studies have shown a limited number of metrics were used to assess the performance of DL models in detecting phishing websites. For example, only two metrics were used in [19], and three were measured in [21]. More metrics were used in the studies [18,22], yet only the selected ones were utilized to benchmark with other machine learning classifiers.

2.2. Convolutional Neural Network (CNN)

Convolutional Neural Network (CNN) is another popular type of DL technique in the field of cybersecurity. CNN is well-fitted to multi-dimensional data and specializes in image and signal processing [1,23]. In addition, CNN can extract features from raw data more efficiently and can solve complicated tasks. It is also more scalable and requires less training time [4]. Nevertheless, CNN architecture needs high computational power and a big dataset when dealing with image data [13]. Although CNN has achieved tremendous success with computer vision, it has also been applied in the cybersecurity domain. Table A2 from Appendix A provides a summary of previous research works on CNN in the field of phishing detection.
CNN was used as a single classifier in numerous research to distinguish between phishing and legitimate websites [7,8,20,24,25,26,27,28]. It can also be used in combination with other DL techniques to form an ensemble model and to improve phishing detection accuracy [10,11,29,30,31,32,33,34,35,36]. The difference between the architectures of CNN and DNN is the use of convolutional layers and kernels. Realizing the important role of these elements in determining the performance accuracy of phishing detection models, most researchers paid more attention to specifying these parameters, not others such as learning rate, dropout rate, epoch, or batch size. While this problem was avoided in [10], details of optimizing these parameters were not provided in the paper. Similarly, the authors of [24,28,29,32] described the optimization process, but only on certain parameters, for example, the number of convolutional layers, number of kernels, and kernel size. Additionally, in terms of performance metrics, it was observed that accuracy, precision, recall, and F1-score were the most common measures [7,24,28,30,31,32,34,35,37,38]. Other evaluation metrics were training time, detection time, GPU memory requirement, etc. [7,24,28,32,33,34].

2.3. Long Short-Term Memory (LSTM)

Long Short-Term Memory (LSTM) is a type of Recurrent Neural Network (RNN) that involves a loop structure between neurons in each layer [12]. LSTM is suitable for sequential or time-series data since it can maintain the continuity of information [39]. LSTM is more popular than the original RNN because the vanishing or exploding gradient and long-term dependency problems in traditional RNN have been overcome in LSTM [1,3,4]. LSTM takes a significantly long time to train, despite these advantages, compared to other DL algorithms [1]. In addition, LSTM only considers the forward information and does not consider the backward information. This issue, however, can be resolved in Bidirectional LSTM [40]. LSTM has caught much attention among researchers, and some of their research works in the phishing detection domain are shown in Table A3 (Appendix A).
Similar to DNN and CNN, LSTM can be implemented individually [20,41,42,43,44,45], incorporated with traditional machine learning techniques [46,47], or combined with other DL algorithms in a hybrid model for an improved performance in detecting malicious websites [10,11,31,33,35,36]. Among the studies of LSTM-based phishing detection models, a majority of them specified the parameter settings for neural network architecture, number of epochs, and learning rate; but ignored the dropout rate and batch size [31,41,42,44,47]. Moreover, only certain parameters were optimized during the fine-tuning process [32,42]. To evaluate the overall performance of LSTM models, four popular metrics were used, being accuracy, precision, recall, and F1-score [31,35,41,44,45]. Other measures training time, detection time, error rate, detection cost, number of epochs per second, etc. [33,42,46,47].

2.4. Gated Recurrent Unit (GRU)

Gated Recurrent Unit (GRU) is another variant of RNN and is a lightweight version of LSTM [23]. While working on small datasets, the performance of GRU is similar to LSTM [48]. Some of the previous studies that implemented GRU in their phishing detection models are provided in Table A4 (Appendix A).
There are a limited number of studies on the implementation of GRU for phishing detection. GRU and Bidirectional GRU can be employed as a single classifier [41,48], or as a replacement to the max-pooling layer in a CNN model [34]. Similar to LSTM, in implementing GRU-based phishing detection models, only neural network architecture, learning rate, and epoch were specified, but not batch size and dropout rate [41,48]. Plus, none of the reviewed papers on GRU included parameter optimization in their experiments. Regarding the performance metrics, all three studies [34,41,48] used accuracy, precision, recall, and F1-score to assess the effectiveness of the DL algorithm. Additional metrics involved GPU memory requirement and parameter set size [34,48].

2.5. Hyper-Parameters

One of the factors that affects the performance of DL algorithms is the selection of hyper-parameters during training. Their values can be fine-tuned to optimize the performance accuracy of phishing detection models. These parameters include, but are not limited to, the number of layers in the neural networks, number of neurons (units) in each layer, learning rate, dropout rate, number of epochs, batch size, etc. [17]. A basic understanding of each parameter will assist in selecting its value in the design and implementation of various DL architectures.
Learning rate. Learning rate is one of the essential factors in the parameter settings of DL models to determine how proper the network can be trained or how fast the model can converge [22,42]. As the learning rate is associated with the convergence speed of the DL algorithm, a more significant learning rate (0.5 to 1) results in a faster convergence speed [49]. A higher learning rate guarantees not only good performance but also causes low stability. However, this only happens at the early stage; after a certain period, the model’s performance will slow down and eventually stop before reaching optimality. Meanwhile, a smaller learning rate (0.0001 to 0.01) can guarantee the model’s stability, yet it delays the speed of convergence, and hence, a longer time is needed to train the DL algorithm.
Dropout rate. Dropout is one of the regularization techniques used to avoid overfitting problems in deep neural network architectures [23]. Overfitting usually happens when the DL model performs well on the training dataset but does not perform well on the validation set. This causes the training accuracy to be much higher than the validation accuracy. The dropout rate is a probability coefficient at which neurons in a particular layer of deep neural networks are discarded during the training process [33]. For instance, when a dropout rate of 0.2 is applied to a specific layer, 20% of the total number of neurons in that particular layer will be dropped. Dropout strategy is usually used in CNN, LSTM, and GRU architectures to prevent overfitting issues [10,30,34].
Batch size. In implementing a DL algorithm, a dataset consisting of numerous samples is used and split into two parts, namely training and testing. In the training phase, instead of passing the whole set of samples to the DL model, training data is divided into batches, in which each batch contains a small amount of data. The size of this subset of data is known as batch size [50]. The normal range of batch size is from 16 [37] to 2048 [32], depending on the size of the dataset. Small datasets generally use small batch sizes, while big datasets use larger batch sizes.
Epoch. Epoch is the number of iterations for training after the DL model has been built and compiled. It is essential to select the appropriate number of training iterations since it can affect the performance accuracy of the phishing detection model [22]. The detection accuracy can increase as the number of epochs goes higher; however, it also requires longer training of the deep neural network. As a result, to determine the number of epochs that give the best performance, one can increase the number of training iterations until reaching the minimum loss. With the growing number of epochs, the model loss will continue to decrease to a specific minimum value and then fluctuate [42]. At this point, training should be stopped since the model accuracy remains stagnant for a further higher number of iterations.
Number of layers. Network layers refer to the hidden layers in DNN, the convolutional layers in CNN, the LSTM/GRU layers in LSTM/GRU architecture, the Restricted Boltzmann Machine (RBM) layers in Deep Belief Network (DBN), the number of Autoencoders (AE) in the Stacked Autoencoder (SAE) model, etc. [22]. An increase in the number of layers in neural network architecture will increase network complexity and slow down the training process. Consequently, it is advisable to slowly raise the number of layers while observing the model’s performance accuracy. Further increase in layers might result in additional processing time while the best performance cannot be guaranteed [42].
Number of neurons/units per layers. In addition to network layers, the number of neurons or units in each layer can also significantly impact the performance of the DL algorithm. Similarly, increasing the number of neurons in the hidden layers of the Deep Neural Network or the number of LSTM units in the LSTM layers might cause low detection accuracy and long training time [42]. Therefore, researchers need to fine-tune these values to ensure an effective and efficient DL model for phishing detection without compromising its performance accuracy.
Number of kernels. Instead of the number of neurons in the hidden layers of DNN, in CNN architecture, the number of kernels or filters in the convolutional layers can significantly influence the success rate at which phishing websites are detected. Kernels, or filters, are used mainly in CNN models to convolve the input data into numerous feature maps [22]. Different kernels were used in previous research works, ranging from 8 [11] to 512 [42], depending on the number of input features, size of the dataset, or the neural network architecture.
Kernel size. Kernel size, or window size, is another parameter that needs to be fine-tuned in CNN models. Kernel size is the size of a one-dimensional window, which is convolved sequentially in the convolutional layers and depends on the number of input features [42]. Different kernel sizes have been utilized in previous studies, with typical values from 1 to 10 [28,29,35]. The optimal kernel size that produces the best performance is determined based on the model’s loss and accuracy.
Optimizer. While training deep neural networks, the loss function or error is calculated to evaluate the effectiveness and efficiency of the DL algorithm. This loss function can be optimized, and the weights can be updated by using optimizers [34,42]. Popular methods of optimization include Root Mean Square Propagation (RMSProp), Gradient Decent (GD), Adam, AdaGrad, AdaDelta, etc. [26]. GD and Adam are the most common optimization techniques in classification problems to optimize the error and adjust the weights [37].
Activation function. The activation function is represented by a mathematical equation that defines an output of a neuron based on given inputs [13]. It is considered one of the important parameters in DL architectures that determines the training model’s output, accuracy, and efficiency. In the neural network, neurons of the same layer usually use the same activation function [6]. Rectified Linear Unit (ReLU), Softmax, sigmoid, and Tanh are examples of frequently-used activation functions for DL models.

2.6. Performance Metrics

To evaluate the performance of a typical DL model, several metrics can be used. The most common is the confusion matrix which comprises four basic measures: True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN), as shown in Figure 1. In addition, Precision, Recall, F1-Score, Area Under the Curve, and Accuracy are also important metrics that are required in the performance evaluation of phishing detection models. The formula of these metrics can be defined as Equations (1)–(6). In any phishing detection model, the primary purpose is to identify phishing attacks; therefore, a phishing sample is often regarded as a positive instance while a legitimate sample is considered as a negative instance [3,41,51].
  • True Positive (TP) is the number of positive instances that the model has correctly classified or the number of samples that have accurately been labeled as phishing [2,52,53].
  • True Negative (TN) represents the number of negative instances that the model has correctly predicted or the number of samples that have accurately been categorized as legitimate [2,52,53].
  • False Positive (FP) denotes the number of negative instances that have been incorrectly recognized as positive or the number of legitimate samples that have been misclassified as phishing [2,52,53].
  • False Negative (FN) refers to the number of positive instances that have been wrongly labeled as negative, or the number of malicious samples that have been misidentified as benign [2,52,53].
False Positive Rate (FPR), also known as False Alarm Rate (FAR) or Fall-Out, is defined as the ratio of FP instances to the total number of predicted negative samples. In other words, it is the number of legitimate instances incorrectly classified as phishing over the total number of legitimate samples [2,46], and is measured as follows:
FPR = FP FP + TN
False Negative Rate (FNR) is defined as the ratio of FN instances to the total number of predicted positive samples, or the percentage of phishing instances wrongly marked as legitimate in all the phishing samples [2,46], and is calculated as:
FNR = FN FN + TP
Precision (PR) is the fraction of predicted positive instances that are positive [1], or the proportion of phishing samples correctly classified as phishing over the total number of actual phishing samples [46,52]. PR indicates the confidence level of the phishing detection model, i.e., how many are malicious out of all the phishing instances detected [18]. PR can be used as a valuable measure for situations in which an unbalanced dataset is involved, in cases when accuracy fails to indicate how well the DL model performs, or in scenarios where the accuracy score alone is insufficient for security experts to make a decision [1]. High precision indicates how accurately the model can detect phishing attacks. PR can be computed using the following formula:
PR = TP TP + FP
Recall (RC), also known as True Positive Rate (TPR), Sensitivity, or Probability of Detection (PD) [54], represents the percentage of positive instances in all predicted positive samples. It shows the proportion of phishing instances accurately recognized as phishing over the total number of predicted phishing samples [46]. RC can sometimes be named as detection rate [3], which reflects the model’s ability to identify phishing activities, and is mathematically given by:
RC = TP TP + FN
F1-Score (F1), or F-Measure, is a harmonic means of precision and recall, representing the balance of both these measurements. F1-Score is a good indication of how well the model has performed [1]. A high F1 value means the model can detect malicious attacks while ensuring that FP and FN are minimized. F-Measure signifies the model’s resilience and effectiveness [32,55]. Thus, it can be used to estimate the overall performance of the DL model and is given by the following equation:
F 1 = 2 . PR . RC PR + RC
Area Under the Curve (AUC) is measured as the total area under a Receiver Operating Characteristic (ROC) curve, which is a graph plotted with the FPR as x-axis and TPR as y-axis [54]. It describes the model’s classification ability while varying the classification thresholds and typically ranges between 0.5 and 1.0 [51]. AUC closed to 1.0 is an ideal scenario with perfect classification capability, whereas AUC less than 0.5 indicates an inferior detection performance. In other words, the higher the AUC value, the better the classifier.
Accuracy (ACC) is one of the most essential and popular metrics to assess the performance of a DL algorithm [1]. In the context of phishing detection, it shows how effectively and efficiently a classifier can distinguish between phishing and legitimate. ACC measure can be used as a good indicator of how well a DL model is trained and described as the overall effectiveness of the classification model [2]. There are certain limitations with accuracy when it comes to unbalanced datasets. However, valuable insight can be derived from accuracy measures when the classes are balanced [54]. ACC is calculated as the ratio of correctly classified instances (both phishing and legitimate) to the total input samples (the whole dataset). The equation used for the measurement of accuracy is as follows:
ACC = TP + TN TP + TN + FP + FN

2.7. Research Novelty

This section highlights the novelty of this empirical study, which is viewed from three perspectives: specification of parameter settings in neural network architectures, optimization process of hyper-parameters for DL algorithms, and performance metrics for phishing detection model evaluation.
Even though DL offers various advantages, one of its major drawbacks is manual parameter tuning. There is no standard and holistic guideline for selecting these hyper-parameters to achieve the highest performance accuracy. Table 1 shows some of the previous research works utilizing DL for phishing detection. These studies are categorized into four groups according to their level of specifying the parameter settings in the neural network architecture. The four categories include Not Specified (NS), Rarely Specified (RS), Partly Specified (PS), and Fully Specified (FS). Studies belonging to the first group (NS) applied DL algorithms without mentioning any hyper-parameters, while those in the second group (RS) only specified one or two. The third group (PS) showed the highest number of studies, yet not all of the parameters were specified or described. Unlike previous research, our study specified all of the hyper-parameters in DL architectures and described the tuning process step-by-step, as provided in Section 3.4.
In addition, the previous studies were also categorized based on the number and type of DL algorithms used. Some studies employed only one technique, while others applied dual or multiple methods. The authors of [41] used four different DL algorithms, but they belonged to the same type of machine learning (unsupervised). Our research covers the highest number of DL techniques, and from various categories, such as supervised (CNN), un-supervised (LSTM, GRU), and hybrid (DNN), to provide more comprehensive research on numerous DL algorithms.
Table 2 provides a more in-depth analysis of the parameter optimization process in the related studies. Most studies focused on describing how to achieve the parameter settings for neural network architecture and learning rate, while dropout rate, batch size, and the number of epochs were neglected. Unlike the previous works from Table 1, those in Table 2 described in detail how to obtain some of the hyper-parameters, including the neural network architecture, learning rate, dropout rate, batch size, and epoch. In contrast, our study described the step-by-step process of fine-tuning all the parameters above, to further investigate how these hyper-parameters can affect the performance accuracy of a phishing detection model.
Table 3 shows a list of performance metrics frequently used by previous authors in their studies. The most common metrics adopted by researchers to evaluate the performance of DL-based phishing detection models were ACC, PR, RC, and F1-Score. Other metrics, such as FPR, FNR, training time, and testing time, were measured by only some authors. Meanwhile, GPU memory requirement, parameter size, number of URLs per second, epoch per second, detection cost, etc., were seldom used. Unlike previous studies that measured only some of the performance metrics, our study covered most of them. In addition to the four common measurements (ACC, PR, RC, F1), other metrics, including FPR, FNR, and AUC, are important indicators of the DL model’s effectiveness in detecting phishing attacks. Furthermore, time complexity (training and testing time) and memory constraints (parameter size and GPU storage) are also crucial factors that need to be considered when assessing the feasibility of DL phishing detection models. As a result, all of these metrics were included in the evaluation process of this empirical study.

3. Research Methodology

This section briefly describes the empirical experiments carried out in this study, the website features used as input vectors, the DL models, and the parameter optimization for four different DL architectures.

3.1. Experiment Setup

Figure 2 shows a theoretical workflow of the experimental setup for this empirical study. The entire process was divided into four stages: input URL (Uniform Resource Locator), data pre-processing and feature engineering, deep learning, and classification. In the first stage, a publicly available dataset was obtained from the University of California Irvine Machine Learning Repository (UCI), consisting of 11055 URLs. This dataset was comprised of both phishing and legitimate websites (4898 and 6157 URLs, respectively) [62]. In the second stage, the dataset went through a data cleaning and data transformation process in the pre-processing phase. Meanwhile, URL features were converted into feature vectors that acted as inputs to the DL model. The dataset was then split into two parts with a ratio of 8:2 (80% as training dataset and 20% as testing dataset). In the third stage, several DL algorithms were built, compiled, and evaluated. Finally, the webpage URL was classified as legitimate or phishing, and a set of performance metrics was measured to assess the performance of four DL models in detecting phishing websites.
In this research, Google Colaboratory, Python, and Tensorflow were used to build several DL models. Instead of running on a local machine and using local GPU, all four DL algorithms were trained using GPU on Google Colaboratory, with a capacity of 11.441MB. The programming language was written in Python code with the help of the TensorFlow package. The use of cloud servers allowed users to leverage the power of Google’s hardware to execute the codes and run Tensorflow operations. The dataset and source code used in the experiments are available on https://github.com/quangdn83/WebsitePhishingDetection (accessed on 21 September 2021) [63].

3.2. Website Features

In the experiment, website features were converted to feature vectors and used as inputs to DL models. Table 4 shows a list of 30 features used in this study. Each feature has three possible values: −1, 0, and 1 (−1 is phishing, 0 is suspicious, and 1 is benign). The last feature, named “class”, is the classification of the URL.
Figure 3 is the heatmap displaying the correlation matrix of these features. A standard range of correlation is from −1 to +1, where −1 is the lowest negative correlation, and +1 is the highest positive correlation. A negative correlation is displayed in the brighter color range, while a positive correlation is displayed in the darker color range. Particularly in this dataset, the mapping of two different features, named Favicon and popUpWindow, showed the darkest color, meaning they are highly or positively correlated. Positive correlations mean one feature marks the URL as phishing, and so does the other. Whereas negative correlations mean one feature marks the URL as malicious, while the other does not [6].

3.3. Deep Learning Models

This empirical study built four phishing detection models using four different DL algorithms: DNN, CNN, LSTM, and GRU. The general architecture of a typical DL-based phishing detection model consists of an input layer, one or more middle layers, and one output layer. Inputs to each DL model are website features that have already been converted to feature vectors. A total of 30 features were used in this study and hence, there were 30 neurons in the input layer of the neural network architecture. Different DL algorithms were used in the middle layers to build the phishing detection models, including DNN, CNN, LSTM, and GRU. Finally, only one neuron was used with a sigmoid activation function in the output layer to classify the web page as malicious or benign.
DNN. A DNN structure consists of an input layer, an output layer, and one or more hidden layers [18], as shown in Figure 4a. Each node or neuron in one layer is connected to the other nodes in the next layer to form a dense or fully-connected layer [19]. The number of hidden layers and neurons in each hidden layer can vary. The activation functions used in the hidden layers and the output layer are ReLU and sigmoid, respectively. Researchers need to fine-tune these parameters to find the optimal values that provide the highest detection accuracy.
CNN. The architecture of a CNN model generally consists of three basic layers: a convolutional layer, a pooling layer, and a fully connected layer [37]. Firstly, a convolutional layer is used for feature extraction and consists of multiple convolutional kernels or filters that divide the input vectors into small blocks. Then, a series of feature maps are generated by performing convolutional operations on the input vectors with the chosen kernels [10]. Secondly, a pooling layer is utilized for dimensional reduction by reducing the dimensionality of the feature maps. The pooling layer has two functions: accelerate the network operation and improve the performance of the entire convolutional network [24]. Thirdly, a fully connected (FC) layer is responsible for classification purposes. FC layer is a traditional neural network that uses extracted features from previous layers to perform the final classification task [29]. To avoid overfitting problems, batch normalization and dropout strategies are used between CNN layers (Figure 4b). ReLU is utilized as an activation function in the convolutional and FC layers, while sigmoid is implemented in the output layer.
LSTM. LSTM is a variant of RNN which involves memory cell structure. The memory cell of a typical LSTM unit is comprised of three gates: an input gate, a forget gate, and an output gate [10]. Unlike a feedforward neural network, the output of a neuron in LSTM architecture at a particular instant can become input to the same neuron. There can be more than one LSTM layer in the LSTM-based phishing detection model, in which a dropout function is used in one layer after another to prevent overfitting issues as illustrated in Figure 5a. The LSTM and dense layers use ReLU, while the output layer uses Sigmoid as their activation functions.
GRU. Similar to LSTM, GRU is constructed with gates and memory cells. Yet, it is simpler in implementation and computation [41]. Instead of a three-gate structure such as LSTM, there are only two gates in the GRU memory cell: input and forget gates. The overall architecture of GRU-based phishing detection models is similar to that of LSTM. Each GRU unit is replaced by an LSTM unit, as shown in Figure 5b.

3.4. Parameter Optimization

In designing and implementing four DL architectures, selecting a set of parameters that produce the best performance accuracy is essential. This process is called parameter tuning and was conducted through a series of experiments described as follows.
Experiment 1: Optimizing the learning rate. Firstly, a set of parameters (including the number of layers, number of neurons, number of kernels, kernel size, dropout rate, batch size, and number of epochs) was randomly selected for each DL algorithm. This set of parameters remained throughout the experiments, while the learning rate changed from 0.0001 to 0.1 to determine the value that yielded the highest detection accuracy.
Experiment 2: Optimizing the dropout rate. Using the learning rate found in the previous experiments, the impact of the dropout rate on the performance of various DL models was investigated. Different values of dropout rate were tested (from 0.1 to 0.5), and the rate with the best performance accuracy was recorded and used in the following experiments.
Experiment 3: Optimizing the neural network architecture. Since different network architectures might produce different results in detection accuracy, changing the structure of neural networks while keeping the learning rate and dropout rate constant was the next step in the parameter optimization process. Different layers, number of neurons per layer, number of kernels, and kernel sizes were examined to find the optimal set that offered the highest accuracy measure.
Experiment 4: Optimizing the batch size. With the learning rate, dropout rate, and deep neural network architecture obtained from the previous experiments, various values of batch size (from 8 to 1024) were tested, and their corresponding detection accuracies were measured. The batch size with the highest accuracy was used in the next experiment as part of parameter settings.
Experiment 5: Optimizing the number of epochs. The final step in the parameter tuning process was to optimize the number of epochs. The optimized value was determined by increasing the training iteration from 50 to 700. At this point, the optimal set of parameters that produced the best detection accuracy had been obtained. A list of parameters that affected the performance of four DL algorithms is recorded in Table 5.
Optimizer and activation functions. The parameters above vary according to the number of input features, the size of datasets, the type of DL algorithms, and the architecture of neural networks. Therefore, different authors in previous studies used different settings for their DL models. However, the common factor among them was the use of optimizer and activation functions. Most of the related studies utilized Adam as an optimizer, ReLU as an activation function for hidden or dense layer, and sigmoid as an activation function for the output layer.
Similarly, the same settings were used in the experimental setup of this empirical study without the need to fine-tune as in other hyper-parameters. Particularly, Adam was chosen because it proved to be the best optimizer among other optimization techniques. ReLU was selected as an activation function in the hidden layers of DNN, convolutional and fully connected layers in CNN, or LSTM/GRU layer in LSTM/GRU models. Finally, sigmoid was used as an activation function at the output layer because sigmoid function produces values in the range of 0 to 1. Thus, it is more suitable and adaptable for the phishing detection model [32] since phishing detection is a binary classification problem in which the output of the classifier is either 0 (phishing) or 1 (legitimate).

4. Results and Discussion

This section presents and discusses the results obtained from numerous experiments to examine the impact of hyperparameter tuning on the performance accuracy of four DL models. Various issues that arose from these experimental results are also highlighted to manifest the overlooked problems that need to be resolved. Moreover, this section also discusses the perspectives that motivate researchers to explore new directions in phishing detection and DL.
Results obtained from Experiment 1 to 5 described in Section 3.4 are provided in Appendix A as listed in Table 6 below.
After conducting a series of experiments, the optimal set of parameters with the highest performance accuracies for various DL models was summarized in Table 7. In the experiment setup, 30 website features were used as input vectors, hence, there were 30 neurons in the input layer of all four DL architectures. Furthermore, since phishing detection is a binary classification problem, only one neuron in the output layer could classify the URL as either legitimate or phishing.

4.1. Results with DNN

For the DNN algorithm (Figure 6), the neural network with three hidden layers and neurons of (16 4 2) (16 neurons in the first hidden layer, 4 neurons in the second hidden layer, and 2 neurons in the third hidden layer) achieved higher accuracy than other DNN architectures. Other parameters such as learning rate, batch size, and epoch were set to 0.001, 32, and 500, respectively. The highest accuracy recorded for this set of parameters was 97.29%. These results were achieved after a series of experiments provided in Appendix A (Table A5, Table A6, Table A7 and Table A8).

4.2. Results with CNN

Similar to the DNN algorithm, CNN also achieved the best detection accuracy with a batch size of 32. However, the CNN structure was different from that of DNN since the CNN model consisted of convolutional layers with a different number of kernels and kernel size (Figure 7). From the experiment, 16 kernels of size three (3) were found to have the highest accuracy of 96.56%. Other parameters, including learning rate, dropout rate, and epoch, were set to 0.005, 0.5, and 50, respectively. Results obtained from this set of experiments are provided in Appendix A (Table A9, Table A10, Table A11, Table A12, Table A13, Table A14 and Table A15).

4.3. Results with LSTM

Likewise, different sets of parameters were tested for the LSTM model and provided in Appendix A (Table A16, Table A17, Table A18, Table A19 and Table A20). The obtained results indicated that the same dropout rate (0.5) and batch size (32) were acquired to produce the highest performance accuracy (97.20%). Nevertheless, only one LSTM layer was needed in the network architecture (Figure 8) because the LSTM algorithm took longer to train. Therefore, adding more layers into the neural network only caused a high computation problem, which compromised the efficiency of the phishing detection system.

4.4. Results with GRU

Last but not least, parameter settings for the GRU algorithm were set to (30 128 128 1) (Figure 9), learning rate = 0.001, dropout rate = 0.5, batch size = 32, and epoch = 200. The highest accuracy that the model could achieve with this set of parameters was 96.70%. Detailed experiments of how to achieve this optimal set of parameters are provided in Appendix A (Table A21, Table A22, Table A23, Table A24 and Table A25). Similar to LSTM, GRU also required a long duration of training time. As a result, a more complex network architecture with more layers or neurons only increased the computational cost and reduced the model efficiency.
The loss and accuracy versus the number of epochs for different DL algorithms during training and validation are shown in Figure 10; as the number of epochs increased, the performance accuracy increased while the loss function decreased.
The performance metrics of four DL models are displayed in Table 8 and Figure 11. It is observed from the experiments that the accuracy of the DNN model was slightly higher than that of the other three DL algorithms. In addition, the amount of time required to train and test the DNN model was relatively low. DNN also had the smallest parameter size and occupied the least memory storage. In contrast, CNN had the lowest measure for accuracy compared with the other three DL mechanisms, yet it required the shortest duration for model training and testing. The parameter size for CNN was more significant than that for DNN, but CNN consumed the most computational power in terms of GPU storage.
Meanwhile, to achieve an almost equivalent accuracy level as DNN, many iterations were involved in the training phase of LSTM and GRU models, which made their training time longer than the others. As the number of neurons in LSTM and GRU models was higher, their parameters were also more significant. However, the GPU capacity requirement for LSTM and GRU was less than that for CNN. To sum up, no DL algorithm provided the best measure across all performance metrics. Each DL technique has its pros and cons; therefore, selecting an appropriate DL approach is a challenging task that can affect the outcomes of a phishing detection model.
In Table 9, results obtained from the empirical analysis of this study are compared with those attained from other authors using the same dataset. In previous studies, the authors used only one type of DL algorithm, such as DNN, CNN, or MLP (multi-layer perceptron). However, four different DL architectures from various categories (supervised, unsupervised, and hybrid) were implemented in our research work. Moreover, dropout rate and batch size, which were not specified in some studies [6,18,28], were included in our empirical analysis. In addition, it is also observed from the table that although different authors used the same algorithm, their optimal set of parameters and accuracy results were not the same. This implies that researchers still had to perform manual parameter tuning to obtain the optimal parameter settings for their DL models. The authors in [64] suggested that this process could be optimized using swarm intelligence (Bat, Hybrid Bat, and Firefly Algorithm). Yet, their accuracy measures were either equivalent or lower than other authors. Meanwhile, the accuracies obtained from our study are almost as high as other authors. Moreover, we also measured additional metrics to obtain a more comprehensive analysis of different DL algorithms’ performance in detecting phishing websites.
Table 10 provides a comparison of various performance metrics between our research work and previous studies using the same dataset. Compared with [64], our study achieved higher accuracy and F1-Score for DNN and included various metrics not measured in [64]. Since MLP is a subset of DNN, MLP results from [6] could be compared with DNN from our empirical analysis. The obtained results showed that our DNN model outperformed the MLP algorithm in all four metrics: ACC, PR, RC, and F1-Score. In addition, although DNN and CNN accuracies in [18,28] demonstrated slightly better results than ours, some of the performance metrics were not calculated in these studies. In [18], for instance, AUC, time complexity, and memory constraints were not included in the evaluation process. Even though training time, testing time, and parameter size were measured in [28], other metrics (FPR, FNR, AUC, and memory storage) were not reported.
On the contrary, our study provided a complete set of performance metrics for four different DL algorithms. Conventional metrics (FPR, FNR, ACC, PR, RC, F1, and AUC) were used to evaluate the effectiveness of the DL mechanism in detecting phishing websites. In contrast, additional metrics (training time, testing time, parameter size, and memory usage) were utilized to assess the computational complexity of the phishing detection model.

4.5. Issues and Perspectives

Parameter Tuning. It can be observed from the experiments that parameter tuning is the common problem among all four DL algorithms. The parameters in the DL models consist of the number of layers in the neural networks, number of neurons (units) in each layer, number of kernels and kernel size (for CNN), learning rate, dropout rate, number of epochs, batch size, etc. [17]. There is no standard and specific guideline for setting these parameters so that the highest performance accuracy can be achieved. Manual fine-tuning and conducting a series of experiments through trial and error are standard practices among researchers who work with DL models. However, this process is tedious, time-consuming, and labor-intensive. One possible solution to overcome the problem of manual parameter tuning is to use optimization techniques to fine-tune the parameters and shorten the tuning process [64].
Accuracy Deficiency. Accuracy deficiency is another widespread issue among DL models, as accuracy is one of the most critical metrics that are used to evaluate the performance of the selected DL algorithm. Several factors affect the accuracy of a phishing detection model, including the quality of the dataset, the extracted features, the chosen classifier, the parameter settings, etc. Some of the current studies in the literature focused on solving the problem of accuracy deficiency by applying the existing DL algorithms [6,7,43]. In contrast, other researchers combined multiple DL techniques in an ensemble model to enhance the detection accuracy [11,31,37]. Ensemble DL (EDL) models are formed by stacking various DL algorithms in parallel and are divided into two categories: homogeneous and heterogeneous. A homogeneous EDL architecture is constructed by combining DL algorithms of the same type (CNN-CNN, LSTM-LSTM, GRU-GRU, etc.). In contrast, a heterogeneous EDL model incorporates DL mechanisms of different kinds (CNN-LSTM, CNN-GRU, LSTM-GRU, etc.). By doing so, the strengths of individual algorithms are merged while their weaknesses are resolved.
Computation Complexity. Computational complexity is another factor that needs to be considered in the design and implementation of DL architecture. Computational complexity can be divided into time complexity and memory constraints. Time complexity involves training and testing, while memory constraints refer to the parameter size and GPU storage capacity. DL generally requires a massive amount of data and substantial training time [65]. Although DL performs better than traditional machine learning on larger datasets, training with a large amount of data is also challenging and time-consuming. Since big datasets might consist of millions of instances, a longer time is needed to train the neural network to achieve high-performance accuracy.
Moreover, limited processing and storage facilities might also cause a delay in the training duration of a DL model [2]. Therefore, selecting an appropriate DL algorithm that can produce maximum accuracy with a minimum amount of time and computational consumption is essential. Plus, reducing the complexity of neural network architecture is also an alternative that can be applied to decrease training time. Last but not least, big data or cloud-based technologies can be integrated with DL to enhance the processing and storage capabilities, leading to a more robust and efficient model for phishing detection.

5. Conclusions and Future Works

In conclusion, many different DL algorithms exist, which numerous researchers in previous studies have implemented to detect phishing websites. However, choosing the right approach best suited for a specific application or dataset is a challenging task. To solve this problem, an empirical study was conducted in this paper, based on some of the most frequently-used DL techniques, such as DNN, CNN, LSTM, and GRU. Different neural network architectures were tested for each of these DL algorithms to find the optimal set of parameter settings that produce the highest performance accuracy. The empirical experiments were performed on the UCI dataset, consisting of 11055 phishing and benign URLs with 30 website features. Various performance metrics were measured to evaluate the effectiveness and the feasibility of the DL-based phishing detection model. The results obtained from the experiments indicated that among the four DL techniques, there was no single algorithm that produced the best measures in all performance metrics. Researchers and developers need to select the best suited to their particular applications or according to specific requirements. They can also combine different DL algorithms in a hybrid or ensemble model to join their advantages and cure their disadvantages.
As part of our future work, we plan to experiment with other DL algorithms that are relatively new and have not been fully explored in the phishing detection domain, such as Autoencoder (AE), Generative Adversarial Network (GAN), or Deep Reinforcement Learning (DRL). In addition to homogeneous EDL models, we will also implement heterogeneous EDL architectures by integrating multiple DL algorithms of different genres. Plus, we plan to use a more significant and unbalanced dataset in the experiment set up to reflect real-life scenarios, as we live in a big data era and phishing is an imbalanced classification problem, where the number of phishing URLs is much smaller than legitimate URLs.

Author Contributions

Conceptualization, N.Q.D. and A.S.; methodology, N.Q.D. and A.S.; software, N.Q.D. and A.S.; validation, N.Q.D. and A.S.; formal analysis, N.Q.D. and A.S.; investigation, N.Q.D. and A.S.; resources, N.Q.D. and A.S.; data curation, N.Q.D. and A.S.; writing—original draft preparation, N.Q.D.; writing—review and editing, N.Q.D., A.S., O.K., T.Y. and H.F.; visualization, N.Q.D.; supervision, A.S.; project administration, A.S.; funding acquisition, A.S. and O.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported/funded by the Ministry of Higher Education under the Fundamental Research Grant Scheme (FRGS/1/2018/ICT04/UTM/01/1). The authors sincerely thank Universiti Teknologi Malaysia (UTM) under Research University Grant Vot-20H04, Malaysia Research University Network (MRUN) Vot 4L876, for the completion of the research. Faculty of Informatics and Management, University of Hradec Kralove, SPEV project Grant Number: 2102/2021.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://www.kaggle.com/isatish/phishing-dataset-uci-ml-csv (accessed on 5 May 2021).

Acknowledgments

We are grateful for the support of Michal Dobrovolny and Sebastien Mambou in consultations regarding application aspects from Hradec Kralove University, Czech Republic.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Previous research works on DNN.
Table A1. Previous research works on DNN.
ReferenceDatasetDataset SizeNumber of NeuronsLearning RateBatch SizeEpoch
InputHiddenOutput
[18]UCI11,0553020/10/520.01-<200
[19]PhishTank73,575-20/40---100
Yandex
[22]UCI17,700------
DMO10,000
[42]PhishTank21191019/100/200/30010.0001-6000
Alexa1407
[21]PhishTank17,000------
DMOZ20,000
PageRank480
WHOIS480
[64]PhishTank
Yahoo
Own dataset
11,0553020/10/520.00132150
13539-2
58,645111-2
88,657111-2
[20]PhishTank
Alexa
30001020/100/200/300/400/50010.001 3100
Table A2. Previous research works on CNN.
Table A2. Previous research works on CNN.
ReferenceDatasetDataset SizeNumber of KernelsKernel SizePooling SizeStrideLearning rateDropout RateBatch SizeEpoch
[37]NA2000-5----16-
[7]PhishTank318,64225633--0.512820
Com Crawl73,575
Yandex83,857
Alexa82,888
[8]PhishTank245664
32
16
16
------
Millersmiles
Yahoo
Starting point
[10]PhishTank21,303-128310.0010.56410
Alexa/Amazon24,800
[11]PhishTank4,820,940 323 × 123 × 10.00010.5--
Openphish16
Alexa8
[25]DMOZ
Own dataset
3816323 × 3(2,2)---3261
32
64
[26]ILSVRC-2012-CLS12 ----0.01-325000
[30]PhishTank2,585,146642---0.2643
6430.5
[42]PhishTank211932/64/64/128/128/264/512-210.001--200
Alexa1407
[31]Alexa, DMOZ, etc., Sophos611,894124,5746454-0.001--500
[29]PhishTank13,6528/16/32/64/841/3/5/7/9------
Crawler10,000
[24]PhishTank10,604-2220.1
0.01
0.001
0.0005
0.54515
Common Crawl10,604
[32]PhishTank245,385-5/6/7--0.010.9204832
Alexa245,023
[27]PhishTank43,984-5----1050
5000 Best Websites45,000
[33]PhishTank1,021,758------6420/45/64
DMOZ989,021
[35]PhishTank97,4002565/6/74-0.0001-32200
Virus Total
Yandex97,400
[28]UCI11,0558/16/32/64102----220
5
[34]Own dataset340,0001282--0.010.510030
65,0001284
[38]PhishTank206,200200221----
MalwarePatrol
DMOZ, Alexa
Table A3. Previous research works on LSTM.
Table A3. Previous research works on LSTM.
ReferenceDatasetDataset SizeNumber of LayersNumber of UnitsLearning RateDropout RateBatch SizeEpoch
[41]PhishTank1.5 million11280.0001--25
Common Crawl2
[46]PhishTank153,78811000.0010.26430
Openphish7212
Alexa170,552
[42]PhishTank2119-40.001--700
Alexa1407
[31]Alexa, DMOZ, etc., Sophos611,894
124,574
1700.001--500
[44]Vaderetro20001----200
Alexa1,000,000
[43]PhishTank200051280.001-128-
Yahoo Directory2000
[33]PhishTank1,021,758----6420/140/
256/578
DMOZ989,021
[35]PhishTank97,4001320.0001-32200
Virus Total
Yandex97,400
[10]PhishTank21,30311280.0010.56410
Alexa/Amazon24,800
[11]PhishTank4,820,94021280.00010.5--
Openphish
Alexa
[47]UCI24561-0.001--200
20.0001600
30.0001800
40.01900
50.00011000
[45]PhishTank450,176110-0.2-10
[36]OpenPhish
Alexa
52,000------
Table A4. Previous research works on GRU.
Table A4. Previous research works on GRU.
ReferenceDatasetDataset SizeNumber of LayersNumber of UnitsLearning RateDropout RateBatch SizeEpoch
[41]PhishTank
Common Crawl
1.5 million1
2
1280.0001--25
[34]Own dataset340,000
65,000
1640.010.510030
[48]PhishTank
Common Crawl
759,3612600.001-25620
Table A5. Performance metrics of DNN at different learning rate. (Architecture = (30 16 1), batch size = 32, epoch = 50).
Table A5. Performance metrics of DNN at different learning rate. (Architecture = (30 16 1), batch size = 32, epoch = 50).
Experiment (Exp.)Learning RateFPR (%)FNR
(%)
Precision (%)Recall (%)F1-Score
(%)
AUC
(%)
Accuracy (%)Time
(s)
D1-10.00016.657.2292.7894.6293.6998.2193.0343
D1-20.00058.245.1694.8493.6394.2398.4793.4926
D1-30.0015.114.6395.3796.0695.7198.9795.1645
D1-40.0053.806.2093.8097.1995.4699.1294.8054
D1-50.016.954.3295.6894.2794.9798.5894.4853
D1-60.055.207.8892.1296.5094.2698.5793.1753
D1-70.16.357.7692.2494.9893.5998.0192.8554
Table A6. Performance metrics of different architectures of DNN. (Learning rate = 0.001, batch size = 32, epoch = 50).
Table A6. Performance metrics of different architectures of DNN. (Learning rate = 0.001, batch size = 32, epoch = 50).
Exp.Hidden LayerNeurons in Each Hidden LayerFPR (%)FNR
(%)
Precision (%)Recall (%)F1-Score
(%)
AUC
(%)
Accuracy (%)Time
(s)
D2-11(30 20 1)5.885.9694.0495.3594.6999.0094.0853
D2-2(30 16 1)5.114.6395.3796.0695.7198.9795.1645
D2-3(30 8 1)5.864.4095.6095.6795.6498.8594.9852
D2-42(30 20 16 1)6.475.0794.9394.7794.8598.5394.3053
D2-5(30 20 8 1)5.694.3895.6295.3095.4699.0695.0251
D2-6(30 20 4 1)10.172.4497.5691.2194.2899.0693.8549
D2-7(30 16 8 1)6.613.6396.3794.6695.5199.0395.0336
D2-8(30 16 4 1)4.594.4795.5396.1795.8598.9395.4832
D2-9(30 8 4 1)7.694.2295.7893.4994.6298.8994.1750
D2-103(30 20 16 8 1)4.674.3295.6896.2395.9598.9595.5253
D2-11(30 20 16 4 1)10.082.7297.2891.1294.1098.8393.7153
D2-12(30 20 16 2 1)3.816.2793.7397.1995.4398.7794.7553
D2-13(30 20 8 4 1)4.685.4394.5796.4795.5198.8294.8923
D2-14(30 20 8 2 1)4.615.7794.2296.6895.4498.6494.7153
D2-15(30 20 4 2 1)9.823.2196.7991.8994.2798.7493.7153
D2-16(30 16 8 4 1)4.015.6294.3896.9195.6399.1995.0753
D2-17(30 16 8 2 1)3.645.1794.8397.2796.0399.1195.4853
D2-18(30 16 4 2 1)3.894.3595.6597.1696.3998.5395.8447
D2-19(30 8 4 2 1)7.553.0097.0093.4795.2099.0494.8452
Table A7. Performance metrics of DNN at different batch size. Architecture = [30 16 4 2 1], learning rate = 0.001, epoch = 50.
Table A7. Performance metrics of DNN at different batch size. Architecture = [30 16 4 2 1], learning rate = 0.001, epoch = 50.
Experiment (Exp.)Batch
Size
FPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
D3-184.235.4094.6096.6395.6099.0095.12124
D3-21610.253.0396.9791.6494.2398.4293.6254
D3-3323.894.3595.6597.1696.3998.5395.8447
D3-4646.944.9695.0494.5194.7898.2494.173
D3-51284.525.8794.1396.5095.3098.3394.713
D3-62565.586.6793.3395.5694.4397.9393.803
D3-75126.206.6793.3395.2294.2798.5693.533
D3-810247.577.0692.9493.9393.4496.5292.723
Table A8. Performance metrics of DNN at different epochs. Architecture = [30 16 4 2 1], learning rate = 0.001, batch size = 32.
Table A8. Performance metrics of DNN at different epochs. Architecture = [30 16 4 2 1], learning rate = 0.001, batch size = 32.
Experiment (Exp.)Number of EpochsFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
D4-1503.894.3595.6597.1696.3998.5395.8447
D4-21005.333.0796.9395.8496.3898.9795.93103
D4-31503.295.4394.5797.4896.0098.9795.48153
D4-42003.553.1096.9097.1397.0198.9596.70203
D4-52502.634.9295.0897.9696.5099.0896.07253
D4-63006.722.5297.4894.7896.1198.2895.61303
D4-75003.012.4797.5397.5397.5399.4097.29503
D4-87003.733.6996.3197.0996.7098.4896.29703
Table A9. Performance metrics of CNN at different learning rates. (Number of kernels = 16, kernel size = 3, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A9. Performance metrics of CNN at different learning rates. (Number of kernels = 16, kernel size = 3, dropout rate = 0.5, batch size = 32, epoch = 50).
Experiment (Exp.)Learning RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C1-10.00015.236.6593.3596.1894.7598.3193.9452
C1-20.00055.185.7594.2596.3295.2798.9294.4853
C1-30.0014.575.9694.0496.6695.3399.0494.6253
C1-40.0053.503.3996.6197.0996.8599.5196.5654
C1-50.013.258.0491.9697.6694.7298.9493.8954
C1-60.0516.343.9496.0685.1390.2797.3289.7853
Table A10. Performance metrics of CNN at different dropout rates. (Number of kernels = 16, kernel size = 3, learning rate = 0.005, batch size = 32, epoch = 50).
Table A10. Performance metrics of CNN at different dropout rates. (Number of kernels = 16, kernel size = 3, learning rate = 0.005, batch size = 32, epoch = 50).
Experiment (Exp.)Dropout RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C5-10.15.312.9897.0295.4096.2199.3695.9354
C5-20.25.373.3296.6895.5796.1299.4595.7553
C5-30.33.535.3694.6497.2095.9099.2195.4381
C5-40.43.964.2395.7796.9396.3499.2795.8854
C5-50.53.503.3996.6197.0996.8599.5196.5654
Table A11. Performance metrics of CNN for different kernel sizes. (Number of kernels = 16, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A11. Performance metrics of CNN for different kernel sizes. (Number of kernels = 16, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Experiment (Exp.)Kernel
Size
FPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C2-118.2412.6087.4094.0790.6194.8589.1554
C2-226.009.8190.1995.5092.7797.8191.7753
C2-333.503.3996.6197.0996.8599.5196.5654
C2-4430945.1994.8197.0295.9099.1995.3454
C2-556.423.5096.5094.9695.7299.3095.2153
C2-666.264.4695.5494.6695.1099.1094.7153
Table A12. Performance metrics of CNN for the different number of kernels. (Kernel size = 3, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A12. Performance metrics of CNN for the different number of kernels. (Kernel size = 3, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Experiment (Exp.)Number of KernelsFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C3-183.416.6193.3997.5195.4198.8194.7154
C3-2163.503.3996.6197.0996.8599.5196.5654
C3-3324.875.2994.7196.3195.5099.2494.8954
C3-4644.444.1995.8196.5196.1699.4695.7053
C3-51283.015.7094.3097.7395.9999.2895.43153
C3-625610.931.9598.0590.3894.0699.1393.67199
C3-75126.564.5695.4494.5795.0098.9694.53253
Table A13. Performance metrics of different architectures of CNN. (Kernel size = 3, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A13. Performance metrics of different architectures of CNN. (Kernel size = 3, learning rate = 0.005, dropout rate = 0.5, batch size = 32, epoch = 50).
Exp.Conv. LayerNumber of KernelsFPR
(%)
FNR
(%)
Precision (%)Recall
(%)
F1-Score(%)AUC
(%)
Accuracy (%)Time
(s)
C6-11644.444.1995.8196.5196.1699.4695.7053
C6-2324.875.2994.7196.3195.5099.2494.8954
C6-3163.503.3996.6197.0996.8599.5196.5654
C6-483.416.6193.3997.5195.4198.8194.7154
C6-52(64 64)0.8310.9789.0399.4393.9499.0592.90108
C6-6(64 32)4.098.5091.5097.1394.2398.6193.26104
C6-7(64 16)2.6310.7189.2997.9793.4398.3492.63104
C6-8(64 85.137.4292.5896.2294.3798.6393.53103
C6-9(32 32)2.579.1390.8798.1194.3598.8993.5391
C6-10(32 16)6.148.0691.9495.2593.5798.3492.76104
C6-11(32 8)11.345.7694.2491.2792.7397.8491.7754
C6-12(16 16)6.228.3791.6394.8793.2297.9792.5853
C6-13(16 8)6.758.3991.6194.7693.1697.7192.3154
C6-14(8 8)8.4210.0289.9893.8191.8596.5890.6454
Table A14. Performance metrics of CNN for different batch sizes. (Number of kernels = 16, kernel size =3, learning rate = 0.005, dropout rate = 0.5, epoch = 50).
Table A14. Performance metrics of CNN for different batch sizes. (Number of kernels = 16, kernel size =3, learning rate = 0.005, dropout rate = 0.5, epoch = 50).
Experiment (Exp.)Batch
Size
FPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C7-184.346.0893.9296.6795.2798.8194.66166
C7-2163.645.6394.3797.2695.7999.1995.2185
C7-3323.503.3996.6197.0996.8599.5196.5654
C7-4644.946.0593.9596.0494.9998.8394.4454
C7-51283.266.1196.7497.5995.7099.1595.0714
C7-62563.314.9595.0597.5096.2699.2995.758
C7-75123.896.1193.8996.9795.4198.9694.843
C7-810243.376.4393.5797.5095.4998.9494.844
Table A15. Performance metrics of CNN for different number of epochs. (Number of kernels = 16, kernel size =3, learning rate = 0.005, dropout rate = 0.5, batch size = 32).
Table A15. Performance metrics of CNN for different number of epochs. (Number of kernels = 16, kernel size =3, learning rate = 0.005, dropout rate = 0.5, batch size = 32).
Experiment (Exp.)Number of EpochsFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(s)
C8-1503.503.3996.6197.0996.8599.5196.5654
C8-21004.814.6295.3896.4395.9099.3095.30103
C8-31502.935.1894.8297.7096.2499.3295.79153
C8-42005.173.8496.1695.8596.0099.3495.57204
C8-52503.634.9795.0397.1396.0799.2995.61254
C8-63003.415.6694.3497.4095.8599.2295.30302
C8-75004.574.1695.8496.3196.0899.2895.66503
C8-87003.916.5393.4797.2395.3199.0694.53704
Table A16. Performance metrics of LSTM at different learning rates. (Number of layers = 1, units = 128, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A16. Performance metrics of LSTM at different learning rates. (Number of layers = 1, units = 128, dropout rate = 0.5, batch size = 32, epoch = 50).
Experiment (Exp.)Learning RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
L1-10.000119.167.2792.7383.7788.0295.5686.976.80
L1-20.00055.496.9693.0495.7794.3898.1993.675.90
L1-30.00114.826.6493.3687.1290.1397.0189.426.72
Table A17. Performance metrics of LSTM at different dropout rates. (Number of layers = 1, units = 128, learning rate = 0.0005, batch size = 32, epoch = 50).
Table A17. Performance metrics of LSTM at different dropout rates. (Number of layers = 1, units = 128, learning rate = 0.0005, batch size = 32, epoch = 50).
Experiment (Exp.)Dropout RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
L2-10.18.288.1291.8893.8692.8697.2491.816.23
L2-20.28.285.0095.0093.4094.1998.3993.535.30
L2-30.37.846.3393.6793.5993.6398.1392.996.75
L2-40.411.536.9793.0390.6291.8197.5990.955.95
L2-50.55.496.9693.0495.7794.3898.1993.675.90
Table A18. Performance metrics of different architectures of LSTM. (Learning rate = 0.0005, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A18. Performance metrics of different architectures of LSTM. (Learning rate = 0.0005, dropout rate = 0.5, batch size = 32, epoch = 50).
Exp.No of LayerUnits per LayerFPR
(%)
FNR
(%)
Precision (%)Recall
(%)
F1-Score
(%)
AUC
(%)
Accuracy (%)Time
(min)
L3-112563.82104789.5397.3493.2798.2792.1319.78
L3-21285.496.9693.0495.7794.3898.1993.675.90
L3-36412.425.3494.6688.7491.6197.3291.183.40
L3-4329.0110.7389.2793.0291.1195.6890.002.57
L3-51612.5714.3285.6890.1087.8393.8586.432.57
L3-62(128 128)4.378.1891.8296.7594.2298.0693.4013.08
L3-7(128 64)6.887.5792.4395.0093.7097.7992.7210.97
L3-8(128 32)5.178.2391.7795.9593.8197.9893.0810.97
L3-9(128 16)9.615.9394.0791.8292.9398.2292.368.87
L3-103(128 128 128)10.876.9293.0891.0392.0496.7891.2722.38
L3-11(128 128 64)6.467.2792.7395.0693.8898.0693.0820.40
L3-12(128 128 32)9.375.1994.8193.2094.0097.7993.0318.92
L3-13(128 128 16)12.176.8093.2089.9691.5597.3390.7315.92
L3-14(128 64 64)11.717.4892.5290.4191.4597.2790.5914.27
L3-15(128 64 32)8.707.0592.9593.1093.0397.6292.2214.73
L3-16(128 64 16)3.289.2090.8097.6594.1098.0293.1713.53
L3-17(128 32 32)3.489.1990.8197.6194.0998.2593.0312.48
L3-18(128 32 16)6.296.8493.1695.1394.1398.1793.4010.13
L3-19(128 16 16)4.668.1791.8396.6394.1797.9593.2610.50
Table A19. Performance metrics of LSTM for different batch sizes. (Number of layers = 1, units = 128, learning rate = 0.0005, dropout rate = 0.5, epoch = 50).
Table A19. Performance metrics of LSTM for different batch sizes. (Number of layers = 1, units = 128, learning rate = 0.0005, dropout rate = 0.5, epoch = 50).
Experiment (Exp.)Batch
Size
FPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
L4-186.706.0393.9794.4494.2198.3593.6718.53
L4-2168.075.8694.1493.7693.9598.0893.1712.32
L4-3325.496.9693.0495.7794.3898.1993.675.90
L4-46411.336.2893.7290.4292.0497.1491.365.07
L4-51284.1015.0884.9297.3190.7096.3488.924.22
L4-62562.8119.8980.1198.3688.3095.0985.623.38
Table A20. Performance metrics of LSTM for the different number of epochs. (Number of layers = 1, units = 128, learning rate = 0.0005, dropout rate = 0.5, batch size = 32).
Table A20. Performance metrics of LSTM for the different number of epochs. (Number of layers = 1, units = 128, learning rate = 0.0005, dropout rate = 0.5, batch size = 32).
Experiment (Exp.)Number of EpochsFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
L5-1505.496.9693.0495.7794.3898.1993.675.90
L5-21005.476.7493.2695.7794.5098.7193.8013.40
L5-31503.868.5191.4997.1594.2398.0593.4020.42
L5-42003.773.9196.0996.9696.5399.2996.1641.12
L5-52503.284.9095.1097.4996.2899.2395.7954.23
L5-63005.917.4492.5695.4393.9698.2393.2265.07
L5-75002.974.1895.8297.7596.7798.5496.34120.27
L5-87001.803.5596.4598.6397.5399.1197.20169.17
Table A21. Performance metrics of GRU at different learning rates. (Number of layers = 1, units = 128, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A21. Performance metrics of GRU at different learning rates. (Number of layers = 1, units = 128, dropout rate = 0.5, batch size = 32, epoch = 50).
Experiment (Exp.)Learning RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
G1-10.000114.647.2192.7988.1490.4095.9989.375.45
G1-20.00056.367.5192.4995.0093.7398.3992.996.40
G1-30.0015.855.5094.5095.4994.9998.7694.355.40
G1-40.0054.909.0390.9796.1893.5097.7492.726.42
G1-50.019.067.8292.1892.8592.5196.9891.635.40
Table A22. Performance metrics of GRU at different dropout rates. (Number of layers = 1, units = 128, learning rate = 0.001, batch size = 32, epoch = 50).
Table A22. Performance metrics of GRU at different dropout rates. (Number of layers = 1, units = 128, learning rate = 0.001, batch size = 32, epoch = 50).
Experiment (Exp.)Dropout RateFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
G2-10.13.838.2691.7497.0094.3098.6693.625.40
G2-20.27.615.5594.4593.9194.1898.8193.536.42
G2-30.37.365.4594.5593.7694.1598.7793.676.42
G2-40.47.726.0193.9994.2294.1098.5093.265.40
G2-50.55.855.5094.5095.4994.9998.7694.355.40
Table A23. Performance metrics of different architectures of GRU. (Learning rate = 0.001, dropout rate = 0.5, batch size = 32, epoch = 50).
Table A23. Performance metrics of different architectures of GRU. (Learning rate = 0.001, dropout rate = 0.5, batch size = 32, epoch = 50).
Exp.No of LayerUnits per LayerFPR
(%)
FNR
(%)
Precision (%)Recall
(%)
F1-Score
(%)
AUC
(%)
Accuracy (%)Time
(min)
G3-112565.186.2793.7395.8994.8098.9494.2110.40
G3-21285.855.5094.5095.4994.9998.7694.355.40
G3-3646.057.6192.3995.0993.7298.0293.084.42
G3-4324.9111.0688.9496.6392.6397.8291.323.42
G3-5167.615.8694.1493.7693.9598.1193.352.98
G3-62(128 128)5.045.1794.8396.0095.4198.9994.8924.40
G3-7(128 64)6.015.7894.2295.2994.7598.8094.1222.40
G3-8(128 32)7.195.6294.3894.0794.2298.7093.6723.42
G3-9(128 16)5.326.3793.6395.9794.7898.7194.0822.42
G3-103(128 128 128)7.435.5694.4494.2294.3398.1993.6233.43
G3-11(128 128 64)10.233.6096.4091.7194.0098.6193.3532.87
G3-12(128 128 32)4.198.0591.9596.9394.3798.7393.5333.42
G3-13(128 128 16)7.765.7294.2894.1394.2098.6793.4034.43
G3-14(128 64 64)3.967.4692.5497.2394.8398.6893.9435.82
G3-15(128 64 32)4.077.0592.9596.9094.8898.9294.2133.32
G3-16(128 64 16)3.156.6793.3397.6595.4498.9594.8034.43
G3-17(128 32 32)6.126.3393.6795.3394.4998.6793.7612.43
G3-18(128 32 16)2.948.5191.4997.9094.5998.3993.7110.45
G3-19(128 16 16)9.844.3495.6691.6793.6298.7793.0811.45
Table A24. Performance metrics of GRU for different batch sizes. (Number of layers = 1, units = [128 128], learning rate = 0.001, dropout rate =0.5, epoch = 50).
Table A24. Performance metrics of GRU for different batch sizes. (Number of layers = 1, units = [128 128], learning rate = 0.001, dropout rate =0.5, epoch = 50).
Experiment (Exp.)Batch
Size
FPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
G4-188.164.1995.8193.2494.5198.8993.9835.42
G4-2162.2810.5189.4998.3593.7198.6792.7617.42
G4-3325.045.1794.8396.0095.4198.9994.8924.40
G4-4641.6312.2687.7498.8392.9698.9391.8611.43
G4-51285.037.1692.8496.0594.4298.7993.768.42
G4-62564.596.9993.0196.5094.7298.3694.036.42
Table A25. Performance metrics of GRU for a different number of epochs. (Number of units = [128 128], learning rate = 0.001, dropout rate = 0.5, batch size = 32).
Table A25. Performance metrics of GRU for a different number of epochs. (Number of units = [128 128], learning rate = 0.001, dropout rate = 0.5, batch size = 32).
Experiment (Exp.)Number of EpochsFPR (%)FNR
(%)
Precision (%)Recall
(%)
F1-Score (%)AUC
(%)
Accuracy (%)Time
(min)
G5-1505.045.1794.8396.0095.4198.9994.8924.40
G5-21003.544.6995.3197.3796.3398.8495.7916.42
G5-31503.584.4495.5697.2596.4098.6795.9335.43
G5-42002.483.9496.0698.0397.0498.7996.7055.07
G5-52507.975.7994.2194.3594.2898.4193.3158.15
G5-63005.804.7595.2595.6395.4498.7994.8083.43
G5-750016.008.5191.4987.3889.3994.5888.10140.43
G5-870050.7042.5557.4573.5164.5048.7955.09205.00

References

  1. Ahmad, R.; Alsmadi, I. Machine learning approaches to IoT security: A systematic literature review. Internet Things 2021, 14, 100365. [Google Scholar] [CrossRef]
  2. Amanullah, M.A.; Habeeb, R.A.A.; Nasaruddin, F.H.; Gani, A.; Ahmed, E.; Nainar, A.S.M.; Akim, N.M.; Imran, M. Deep learning and big data technologies for IoT security. Comput. Commun. 2020, 151, 495–517. [Google Scholar] [CrossRef]
  3. Liu, H.; Lang, B. Machine Learning and Deep Learning Methods for Intrusion Detection Systems: A Survey. Appl. Sci. 2019, 9, 4396. [Google Scholar] [CrossRef] [Green Version]
  4. Asharf, J.; Moustafa, N.; Khurshid, H.; Debie, E.; Haider, W.; Wahab, A. A Review of Intrusion Detection Systems Using Machine and Deep Learning in Internet of Things: Challenges, Solutions and Future Directions. Electronics 2020, 9, 1177. [Google Scholar] [CrossRef]
  5. Bello, I.; Chiroma, H.; Abdullahi, U.A.; Gital, A.Y.; Jauro, F.; Khan, A.; Okesola, J.O.; Abdulhamid, S.M. Detecting ransomware attacks using intelligent algorithms: Recent development and next direction from deep learning and big data perspectives. J. Ambient Intell. Humaniz. Comput. 2020, 12, 8699–8717. [Google Scholar] [CrossRef]
  6. Al-Ahmadi, S. PDMLP: Phishing Detection Using Multilayer Perceptron. Int. J. Netw. Secur. Its Appl. 2020, 12. SSRN:3624621. Available online: https://papers.ssrn.com/abstract=3624621 (accessed on 12 May 2021). [CrossRef]
  7. Aljofey, A.; Jiang, Q.; Qu, Q.; Huang, M.; Niyigena, J.-P. An Effective Phishing Detection Model Based on Character Level Convolutional Neural Network from URL. Electronics 2020, 9, 1514. [Google Scholar] [CrossRef]
  8. Al-Milli, N.; Hammo, B.H. A Convolutional Neural Network Model to Detect Illegitimate URLs. In Proceedings of the 2020 11th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 7–9 April 2020; pp. 220–225. [Google Scholar] [CrossRef]
  9. Feng, J.; Zou, L.; Nan, T. A Phishing Webpage Detection Method Based on Stacked Autoencoder and Correlation Coefficients. J. Comput. Inf. Technol. 2019, 27. [Google Scholar] [CrossRef]
  10. Feng, J.; Zou, L.; Ye, O.; Han, J. Web2Vec: Phishing Webpage Detection Method Based on Multidimensional Features Driven by Deep Learning. IEEE Access 2020, 8, 221214–221224. [Google Scholar] [CrossRef]
  11. Huang, Y.; Yang, Q.; Qin, J.; Wen, W. Phishing URL Detection via CNN and Attention-Based Hierarchical RNN. In Proceedings of the 2019 18th IEEE International Conference on Trust, Security and Privacy in Computing and Communications/13th IEEE International Conference on Big Data Science And Engineering (TrustCom/BigDataSE), Rotorua, New Zealand, 5–8 August 2019; pp. 112–119. [Google Scholar] [CrossRef]
  12. Chen, Z. Deep Learning for Cybersecurity: A Review. In Proceedings of the 2020 International Conference on Computing and Data Science (CDS), Stanford, CA, USA, 1–2 August 2020; pp. 7–18. [Google Scholar]
  13. Naway, A.; LI, Y. A Review on The Use of Deep Learning in Android Malware Detection. arXiv 2018, arXiv:181210360. Available online: http://arxiv.org/abs/1812.10360 (accessed on 3 April 2021).
  14. Sarker, I.H. Deep Cybersecurity: A Comprehensive Overview from Neural Network and Deep Learning Perspective. SN Comput. Sci. 2021, 2, 154. [Google Scholar] [CrossRef]
  15. Quang, D.N.; Selamat, A.; Krejcar, O. Recent Research on Phishing Detection Through Machine Learning Algorithm. In Advances and Trends in Artificial Intelligence. Artificial Intelligence Practices; Fujita, H., Selamat, A., Lin, J.C.-W., Ali, M., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 495–508. [Google Scholar]
  16. Wu, Y.; Wei, D.; Feng, J. Network Attacks Detection Methods Based on Deep Learning Techniques: A Survey. Secur. Commun. Netw. 2020, 2020, e8872923. [Google Scholar] [CrossRef]
  17. Mahdavifar, S.; Ghorbani, A.A. Application of deep learning to cybersecurity: A survey. Neurocomputing 2019, 347, 149–176. [Google Scholar] [CrossRef]
  18. Mahdavifar, S.; Ghorbani, A.A. DeNNeS: Deep embedded neural network expert system for detecting cyber attacks. Neural Comput. Appl. 2020, 32, 14753–14780. [Google Scholar] [CrossRef]
  19. Sahingoz, O.K.; Işılay Baykal, S.; Bulut, D. Phishing detection from urls by using neural networks. In Computer Science & Information Technology (CS & IT); AIRCC Publishing Corporation: Chennai, India, 2018; pp. 41–54. [Google Scholar] [CrossRef]
  20. Khan, M.F.; Al, E. Detection of Phishing Websites Using Deep Learning Techniques. Turk. J. Comput. Math. Educ. TURCOMAT 2021, 12, 3880–3892. [Google Scholar] [CrossRef]
  21. Sountharrajan, S.; Nivashini, M.; Shandilya, S.K.; Suganya, E.; Bazila Banu, A.; Karthiga, M. Dynamic Recognition of Phishing URLs Using Deep Learning Techniques. In Advances in Cyber Security Analytics and Decision Systems; Shandilya, S.K., Wagner, N., Nagar, A.K., Eds.; EAI/Springer Innovations in Communication and Computing; Springer International Publishing: Cham, Switzerland, 2020; pp. 27–56. ISBN 978-3-030-19353-9. [Google Scholar] [CrossRef]
  22. Selvaganapathy, S.; Nivaashini, M.; Natarajan, H. Deep belief network based detection and categorization of malicious URLs. Inf. Secur. J. Glob. Perspect. 2018, 27, 145–161. [Google Scholar] [CrossRef]
  23. Aldweesh, A.; Derhab, A.; Emam, A.Z. Deep learning approaches for anomaly-based intrusion detection systems: A survey, taxonomy, and open issues. Knowl.-Based Syst. 2020, 189, 105124. [Google Scholar] [CrossRef]
  24. Wei, W.; Ke, Q.; Nowak, J.; Korytkowski, M.; Scherer, R.; Woźniak, M. Accurate and fast URL phishing detector: A convolutional neural network approach. Comput. Netw. 2020, 178, 107275. [Google Scholar] [CrossRef]
  25. Liu, D.; Lee, J.-H.; Wang, W.; Wang, Y. Malicious Websites Detection via CNN based Screenshot Recognition. In Proceedings of the 2019 International Conference on Intelligent Computing and its Emerging Applications (ICEA), Tainan, Taiwan, 30 August–1 September 2019; pp. 115–119. [Google Scholar] [CrossRef]
  26. Phoka, T.; Suthaphan, P. Image Based Phishing Detection Using Transfer Learning. In Proceedings of the 2019 11th International Conference on Knowledge and Smart Technology (KST), Phuket, Thailand, 23–26 January 2019; pp. 232–237. [Google Scholar] [CrossRef]
  27. Xiao, X.; Zhang, D.; Hu, G.; Jiang, Y.; Xia, S. CNN–MHSA: A Convolutional Neural Network and multi-head self-attention combined approach for detecting phishing websites. Neural Netw. 2020, 125, 303–312. [Google Scholar] [CrossRef] [PubMed]
  28. Yerima, S.Y.; Alzaylaee, M.K. High Accuracy Phishing Detection Based on Convolutional Neural Networks. In Proceedings of the 2020 3rd International Conference on Computer Applications Information Security (ICCAIS), Riyadh, Saudi Arabia, 19–21 March 2020; pp. 1–6. [Google Scholar] [CrossRef]
  29. Wang, H.; Yu, L.; Tian, S.; Peng, Y.; Pei, X. Bidirectional LSTM Malicious webpages detection algorithm based on convolutional neural network and independent recurrent neural network. Appl. Intell. 2019, 49, 3016–3026. [Google Scholar] [CrossRef]
  30. Rasymas, T.; Dovydaitis, L. Detection of phishing URLs by using deep learning approach and multiple features combinations. Balt. J. Mod. Comput. 2020, 8, 471–483. [Google Scholar] [CrossRef]
  31. Srinivasan, S.; Vinayakumar, R.; Arunachalam, A.; Alazab, M.; Soman, K. DURLD: Malicious URL Detection Using Deep Learning-Based Character Level Representations. In Malware Analysis Using Artificial Intelligence and Deep Learning; Stamp, M., Alazab, M., Shalaginov, A., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 535–554. ISBN 978-3-030-62582-5. [Google Scholar] [CrossRef]
  32. Wang, W.; Zhang, F.; Luo, X.; Zhang, S. PDRCNN: Precise Phishing Detection with Recurrent Convolutional Neural Networks. Secur. Commun. Netw. 2019, 2019, e2595794. [Google Scholar] [CrossRef]
  33. Yang, P.; Zhao, G.; Zeng, P. Phishing Website Detection Based on Multidimensional Features Driven by Deep Learning. IEEE Access 2019, 7, 15196–15209. [Google Scholar] [CrossRef]
  34. Yang, W.; Zuo, W.; Cui, B. Detecting Malicious URLs via a Keyword-Based Convolutional Gated-Recurrent-Unit Neural Network. IEEE Access 2019, 7, 29891–29900. [Google Scholar] [CrossRef]
  35. M, Y.V.; Janet, B.; Reddy, S. Anti-phishing System using LSTM and CNN. In Proceedings of the 2020 IEEE International Conference for Innovation in Technology (INOCON), Bangluru, India, 6–8 November 2020; pp. 1–5. [Google Scholar] [CrossRef]
  36. jaysinha. Available online: https://jaysinha.me/files/phishx_preprint.pdf (accessed on 18 September 2021).
  37. Al-Ahmadi, S. A Deep Learning Technique for Web Phishing Detection Combined URL Features and Visual Similarity. Soc. Sci. Res. Netw. 2020. SSRN:3716033. Available online: https://papers.ssrn.com/abstract=3716033 (accessed on 10 March 2021). [CrossRef]
  38. Zhang, Q.; Bu, Y.; Chen, B.; Zhang, S.; Lu, X. Research on phishing webpage detection technology based on CNN-BiLSTM algorithm. J. Phys. Conf. Ser. 2021, 1738, 012131. [Google Scholar] [CrossRef]
  39. Chen, D.; Wawrzynski, P.; Lv, Z. Cyber security in smart cities: A review of deep learning-based applications and case studies. Sustain. Cities Soc. 2021, 66, 102655. [Google Scholar] [CrossRef]
  40. Elnagar, S.; Thomas, M. A Cognitive Framework for Detecting Phishing Websites. In Proceedings of the International Conference on Advances on Applied Cognitive Computing (ACC 2018), Las Vegas, NV, USA, 30 July–2 August 2018; pp. 60–61. [Google Scholar]
  41. Feng, T.; Yue, C. Visualizing and Interpreting RNN Models in URL-based Phishing Detection. In Proceedings of the 25th ACM Symposium on Access Control Models and Technologies, Barcelona, Spain, 10–12 June 2020; pp. 13–24. [Google Scholar] [CrossRef]
  42. Somesha, M.; Pais, A.R.; Rao, R.S.; Rathour, V.S. Efficient deep learning techniques for the detection of phishing websites. Sādhanā 2020, 45, 165. [Google Scholar] [CrossRef]
  43. Su, Y. Research on Website Phishing Detection Based on LSTM RNN. In Proceedings of the 2020 IEEE 4th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chongqing, China, 12–14 June 2020; Volume 1, pp. 284–288. [Google Scholar] [CrossRef]
  44. Torroledo, I.; Camacho, L.D.; Bahnsen, A.C. Hunting Malicious TLS Certificates with Deep Neural Networks. In Proceedings of the 11th ACM Workshop on Artificial Intelligence and Security; Association for Computing Machinery: New York, NY, USA, 2018; pp. 64–73. [Google Scholar] [CrossRef]
  45. Afzal, S.; Asim, M.; Javed, A.R.; Beg, M.O.; Baker, T. URLdeepDetect: A Deep Learning Approach for Detecting Malicious URLs Using Semantic Vector Models. J. Netw. Syst. Manag. 2021, 29, 21. [Google Scholar] [CrossRef]
  46. Rao, R.S.; Vaishnavi, T.; Pais, A.R. PhishDump: A multi-model ensemble based technique for the detection of phishing sites in mobile devices. Pervasive Mob. Comput. 2019, 60, 101084. [Google Scholar] [CrossRef]
  47. Wang, S.; Khan, S.; Xu, C.; Nazir, S.; Hafeez, A. Deep Learning-Based Efficient Model Development for Phishing Detection Using Random Forest and BLSTM Classifiers. Complexity 2020, 2020, e8694796. [Google Scholar] [CrossRef]
  48. Yuan, L.; Zeng, Z.; Lu, Y.; Ou, X.; Feng, T. A Character-Level BiGRU-Attention for Phishing Classification. In Information and Communications Security; Zhou, J., Luo, X., Shen, Q., Xu, Z., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 746–762. [Google Scholar] [CrossRef]
  49. Yi, P.; Guan, Y.; Zou, F.; Yao, Y.; Wang, W.; Zhu, T. Web Phishing Detection Using a Deep Learning Framework. Wirel. Commun. Mob. Comput. 2018, 2018, e4678746. [Google Scholar] [CrossRef]
  50. Robic-Butez, P.; Win, T.Y. Detection of Phishing websites using Generative Adversarial Network. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 3216–3221. [Google Scholar] [CrossRef]
  51. Sohn, I. Deep belief network based intrusion detection techniques: A survey. Expert Syst. Appl. 2021, 167, 114170. [Google Scholar] [CrossRef]
  52. Alotaibi, R.; Al-Turaiki, I.; Alakeel, F. Mitigating Email Phishing Attacks using Convolutional Neural Networks. In Proceedings of the 2020 3rd International Conference on Computer Applications Information Security (ICCAIS), Riyadh, Saudi Arabia, 19–21 March 2020; pp. 1–6. [Google Scholar] [CrossRef]
  53. Fang, Y.; Zhang, C.; Huang, C.; Liu, L.; Yang, Y. Phishing Email Detection Using Improved RCNN Model With Multilevel Vectors and Attention Mechanism. IEEE Access 2019, 7, 56329–56340. [Google Scholar] [CrossRef]
  54. Berman, D.S.; Buczak, A.L.; Chavis, J.S.; Corbett, C.L. A Survey of Deep Learning Methods for Cyber Security. Information 2019, 10, 122. [Google Scholar] [CrossRef] [Green Version]
  55. Chatterjee, M.; Namin, A.-S. Detecting Phishing Websites through Deep Reinforcement Learning. In Proceedings of the 2019 IEEE 43rd Annual Computer Software and Applications Conference (COMPSAC), Milwaukee, WI, USA, 15–19 July 2019; Volume 2, pp. 227–232. [Google Scholar] [CrossRef]
  56. Odeh, A.; Keshta, I.; Abdelfattah, E. Efficient Detection of Phishing Websites Using Multilayer Perceptron International Association of Online Engineering. 2020, pp. 22–31. Available online: https://www.learntechlib.org/p/217754/ (accessed on 10 March 2021).
  57. Saha, I.; Sarma, D.; Chakma, R.J.; Alam, M.N.; Sultana, A.; Hossain, S. Phishing Attacks Detection using Deep Learning Approach. In Proceedings of the 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT), Tirunelveli, India, 20–22 August 2020; pp. 1180–1185. [Google Scholar] [CrossRef]
  58. Ya, J.; Liu, T.; Zhang, P.; Shi, J.; Guo, L.; Gu, Z. NeuralAS: Deep Word-Based Spoofed URLs Detection AgaIInst Strong Similar Samples. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–7. [Google Scholar] [CrossRef]
  59. Adebowale, M.A.; Lwin, K.T.; Hossain, M.A. Deep Learning with Convolutional Neural Network and Long Short-Term Memory for Phishing Detection. In Proceedings of the 2019 13th International Conference on Software, Knowledge, Information Management and Applications (SKIMA), Island of Ulkulhas, Maldives, 26–28 August 2019; pp. 1–8. [Google Scholar] [CrossRef]
  60. Digwal, H.N.; Kavya, N.P. Detection of Phishing Website Based on Deep Learning. Int. J. Res. Eng. Sci. Manag. 2020, 3, 331–336. [Google Scholar]
  61. Pooja, A.S.S.V.L.; Sridhar, M. Analysis of Phishing Website Detection Using CNN and Bidirectional LSTM. In Proceedings of the 2020 4th International Conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 5–7 November 2020; pp. 1620–1629. [Google Scholar] [CrossRef]
  62. Kaggle. Available online: https://www.kaggle.com/isatish/phishing-dataset-uci-ml-csv (accessed on 12 April 2021).
  63. Github. Available online: https://github.com/quangdn83/WebsitePhishingDetection (accessed on 21 September 2021).
  64. Vrbančič, G.; Fister, I.; Podgorelec, V. Parameter Setting for Deep Neural Networks Using Swarm Intelligence on Phishing Websites Classification. Int. J. Artif. Intell. Tools 2019, 28, 1960008. [Google Scholar] [CrossRef]
  65. Chen, S.; Fan, L.; Chen, C.; Xue, M.; Liu, Y.; Xu, L. GUI-Squatting Attack: Automated Generation of Android Phishing Apps. IEEE Trans. Dependable Secure Comput. accepted. [CrossRef]
Figure 1. Confusion matrix.
Figure 1. Confusion matrix.
Applsci 11 09210 g001
Figure 2. Theoretical workflow.
Figure 2. Theoretical workflow.
Applsci 11 09210 g002
Figure 3. Correlation matrix of website features.
Figure 3. Correlation matrix of website features.
Applsci 11 09210 g003
Figure 4. Neural network architecture for (a) DNN and (b) CNN.
Figure 4. Neural network architecture for (a) DNN and (b) CNN.
Applsci 11 09210 g004
Figure 5. Neural network architecture for (a) LSTM and (b) GRU.
Figure 5. Neural network architecture for (a) LSTM and (b) GRU.
Applsci 11 09210 g005
Figure 6. Optimal DNN architecture.
Figure 6. Optimal DNN architecture.
Applsci 11 09210 g006
Figure 7. Optimal CNN architecture.
Figure 7. Optimal CNN architecture.
Applsci 11 09210 g007
Figure 8. Optimal LSTM architecture.
Figure 8. Optimal LSTM architecture.
Applsci 11 09210 g008
Figure 9. Optimal GRU architecture.
Figure 9. Optimal GRU architecture.
Applsci 11 09210 g009
Figure 10. Accuracy and loss of four DL algorithms.
Figure 10. Accuracy and loss of four DL algorithms.
Applsci 11 09210 g010
Figure 11. Performance metrics of four DL algorithms.
Figure 11. Performance metrics of four DL algorithms.
Applsci 11 09210 g011
Table 1. Frequency of parameter settings for DL architectures.
Table 1. Frequency of parameter settings for DL architectures.
CategoryAlgorithmFrequencyReference
NS 1RS 2PS 3FS 4
SingleDNN-1--[19]
MLP1-2-[6,56,57]
CNN--4-[7,8,25,27]
LSTM-11-[44,45]
BiLSTM1---[58]
BiGRU--1-[48]
DualCNN, CNN--1-[37]
CNN, LSTM1-4-[30,31,33,59,60]
CNN, BiLSTM3-1-[10,36,40,61]
CNN, GRU--1-[34]
MultipleCNN, RNN, MLP--1-[11]
DNN, DBM, SAE1---[21]
DNN, CNN, LSTM--1-[20]
LSTM, GRU, BiLSTM, BiGRU--1-[41]
DNN, CNN, LSTM, GRU---1Our study
Total72181
1 Not Specified. 2 Rarely Specified. 3 Partly Specified. 4 Fully Specified.
Table 2. Related works on parameter optimization for DL algorithms.
Table 2. Related works on parameter optimization for DL algorithms.
ReferenceAlgorithmParameters
Learning RateNetwork ArchitectureDropout RateBatch SizeEpoch
[9]SAE Applsci 11 09210 i001 Applsci 11 09210 i002xx Applsci 11 09210 i003
[18]DNN Applsci 11 09210 i004 Applsci 11 09210 i005xx Applsci 11 09210 i006
[49]DBN Applsci 11 09210 i007 Applsci 11 09210 i008xx Applsci 11 09210 i009
[50]GAN Applsci 11 09210 i010xx Applsci 11 09210 i011x
[24]CNN Applsci 11 09210 i012 Applsci 11 09210 i013xxx
[28]CNN, CNNx Applsci 11 09210 i014xxx
[22]DBN, DNN Applsci 11 09210 i015 Applsci 11 09210 i016xx Applsci 11 09210 i017
[32]CNN, BiLSTMx Applsci 11 09210 i018x Applsci 11 09210 i019 Applsci 11 09210 i020
[42]DNN, CNN, LSTM Applsci 11 09210 i021 Applsci 11 09210 i022xx Applsci 11 09210 i023
[29]CNN, RNN, BiLSTMx Applsci 11 09210 i024xxx
Our studyDNN, CNN, LSTM, GRU Applsci 11 09210 i025 Applsci 11 09210 i026 Applsci 11 09210 i027 Applsci 11 09210 i028 Applsci 11 09210 i029
Table 3. Performance metrics used in the previous studies and our research work.
Table 3. Performance metrics used in the previous studies and our research work.
ReferencePerformance Metrics
FPRFNRACCPRRCF1AUCTraining TimeTesting TimeOther
[7] Applsci 11 09210 i030 Applsci 11 09210 i031 Applsci 11 09210 i032 Applsci 11 09210 i033 Applsci 11 09210 i034 Applsci 11 09210 i035 Applsci 11 09210 i036
[10] Applsci 11 09210 i037 Applsci 11 09210 i038 Applsci 11 09210 i039 Applsci 11 09210 i040
[11] Applsci 11 09210 i041 Applsci 11 09210 i042 Applsci 11 09210 i043 Applsci 11 09210 i044
[24] Applsci 11 09210 i045 Applsci 11 09210 i046 Applsci 11 09210 i047 Applsci 11 09210 i048 Applsci 11 09210 i049 GPU memory requirement
Parameter set size, Loss
Number of URL per second
[28] Applsci 11 09210 i050 Applsci 11 09210 i051 Applsci 11 09210 i052 Applsci 11 09210 i053 Applsci 11 09210 i054 Applsci 11 09210 i055Parameter size
[32] Applsci 11 09210 i056 Applsci 11 09210 i057 Applsci 11 09210 i058 Applsci 11 09210 i059 Applsci 11 09210 i060 Applsci 11 09210 i061 Applsci 11 09210 i062
[33] Applsci 11 09210 i063 Applsci 11 09210 i064 Applsci 11 09210 i065 Applsci 11 09210 i066 Applsci 11 09210 i067Detection cost, Epoch/s
[34] Applsci 11 09210 i068 Applsci 11 09210 i069 Applsci 11 09210 i070 Applsci 11 09210 i071 Parameter size
[41] Applsci 11 09210 i072 Applsci 11 09210 i073 Applsci 11 09210 i074 Applsci 11 09210 i075 Parameter size
[48] Applsci 11 09210 i076 Applsci 11 09210 i077 Applsci 11 09210 i078 Applsci 11 09210 i079 Model storage spaceParameter size
Our study Applsci 11 09210 i080 Applsci 11 09210 i081 Applsci 11 09210 i082 Applsci 11 09210 i083 Applsci 11 09210 i084 Applsci 11 09210 i085 Applsci 11 09210 i086 Applsci 11 09210 i087 Applsci 11 09210 i088GPU memory stageParameter size
Table 4. List of phishing website features.
Table 4. List of phishing website features.
TypeNoFeatureNameDescriptionValue
Address bar-based1IP addressUsingIPHaving IP address in URL−1, 1
2URL lengthLongURLLong URL to hide the suspicious part−1, 0, 1
3Shortening serviceShortURLUsing URL shortening services “TinyURL”−1, 1
4@ SymbolSymbol@URL’s having @ symbol−1, 1
5“//” redirectingRedirecting//Having “//” within URL path for directing−1, 1
6Prefix suffixPrefixSuffixAdding prefix or suffix separated by (-) to the domain−1, 1
7Sub domainSubDomainsSub domain and multi sub domain−1, 0, 1
8SSL final stateHTTPSExistence of HTTPS and validity of the certificate−1, 0, 1
9Domain registrationDomainRegLenExpiry date of domains/Domain registration length−1, 1
10FaviconFaviconFavicon loaded from a domain−1, 1
11PortNonStdPortUsing non-standard port−1, 1
12HTTPS tokenHTTPSDomainURLThe existence of HTTPS token in the domain part of URL−1, 1
Abnormal-based13Request URLRequestURLRequest URL within a webpage/Abnormal request−1, 1
14URL of anchorAnchorURLURL within <a> tag/Abnormal anchor−1, 0, 1
15Links in tagsLinksInScriptTagsLinks in <Meta>, <Script> and <Link> tags−1, 0, 1
16SFHServerFormHandlerServer Form Handler−1, 0, 1
17EmailInfoEmailSubmitting information to E-mail−1, 1
18Abnormal URLAbnormalURLHost name is included in the URL/Whois−1, 1
HTML and JavaScript-based19RedirectingWebsiteForwardingNumber of times a website has been redirected0, 1
20On mouseoverStatusBarCustOn mouse over changes status bar/Status bar customization−1, 1
21Right clickDisableRightClickDisabling right click−1, 1
22Pop-up windowUsingPopupWindowUsing Pop-up window−1, 1
23Iframe redirectionIframeRedirectionUsing Iframe−1, 1
Domain-based24Age of domainAgeofDomainMinimum age of a legitimate domain is 6 months−1, 1
25DNS recordDNSRecordingExistence of DNS record for the domain−1, 1
26Website trafficWebsiteTrafficBeing among top 100,000 in Alexa rank−1, 0, 1
27Page rankPageRankHaving a page rank greater than 0.2−1, 1
28Google indexGoogleIndexWebsite indexed by Google−1, 1
29Link referenceLinksPoitingToPageNumber of links pointing to a page−1, 0, 1
30Statistical reportStatsReportTop 10 domain and top 10 Ips from PhishTank−1, 1
ResultclassPhishing or legitimate−1, 1
Table 5. List of parameters in various DL models.
Table 5. List of parameters in various DL models.
DL AlgorithmNumber of LayersNumber of UnitsNumber of KernelsKernel SizeLearning RateDropout RateBatch SizeNumber of Epochs
DNN Applsci 11 09210 i089 Applsci 11 09210 i090 Applsci 11 09210 i091 Applsci 11 09210 i092 Applsci 11 09210 i093
CNN Applsci 11 09210 i094 Applsci 11 09210 i095 Applsci 11 09210 i096 Applsci 11 09210 i097 Applsci 11 09210 i098 Applsci 11 09210 i099 Applsci 11 09210 i100
LSTM Applsci 11 09210 i101 Applsci 11 09210 i102 Applsci 11 09210 i103 Applsci 11 09210 i104 Applsci 11 09210 i105 Applsci 11 09210 i106
GRU Applsci 11 09210 i107 Applsci 11 09210 i108 Applsci 11 09210 i109 Applsci 11 09210 i110 Applsci 11 09210 i111 Applsci 11 09210 i112
Table 6. List of experiments for parameter optimization in Appendix A.
Table 6. List of experiments for parameter optimization in Appendix A.
ExperimentDescriptionParameterDL Algorithm
DNNCNNLSTMGRU
1Optimizing the learning rateLearning rateTable A5Table A9Table A16Table A21
2Optimizing the dropout rateDropout rate-Table A10Table A17Table A22
3Optimizing the neural network architectureNumber of layers/ Number of neurons per layerTable A6Table A13Table A18Table A23
Number of kernels-Table A12--
Kernel size-Table A11--
4Optimizing the batch sizeBatch sizeTable A7Table A14Table A19Table A24
5Optimizing the number of epochsEpochTable A8Table A15Table A20Table A25
Table 7. An optimal set of parameter settings for various DL algorithms.
Table 7. An optimal set of parameter settings for various DL algorithms.
DL
Algorithm
No of LayersNumber of NeuronsNumber of KernelsKernel SizeLearning RateDropout RateBatch SizeNo of Epochs
DNN5(30 16 4 2 1)--0.001-32500
CNN4(30 16 1)1630.0050.53250
LSTM3(30 128 1)--0.00050.532700
GRU4(30 128 128 1)--0.0010.532200
Table 8. Performance metrics of four DL algorithms.
Table 8. Performance metrics of four DL algorithms.
DL
Algorithm
Performance Metrics
FPR (%)FNR (%)ACC (%)PR (%)RC (%)F1(%)AUC (%)Training Time (min)Testing Time (s)Parameter SizeMemory
Storage (MB)
DNN3.012.4797.2997.5397.5397.5399.408.380.4701507172
CNN3.503.3996.5696.6197.0996.8599.510.90.2633745325
LSTM1.803.5597.2096.4598.6397.5399.11169.170.80466,689184
GRU2.483.9496.7096.0698.0397.0498.7955.071.447149,505184
Table 9. Comparison of parameter settings with other studies using the same dataset.
Table 9. Comparison of parameter settings with other studies using the same dataset.
ReferenceDL
Algorithm
Parameter SettingsACC (%)
Learning RateInput LayerHidden LayerOutput LayerNumber of KernelsKernel SizeDropout RateBatch SizeEpoch
[18]DNN0.0130(20 10 5)2----20097.50
[28]CNN-30(64 64)164/6412/6--22097.20
[6]MLP--(100 100)------96.65
[64]DNN.BA 10.018530(50 30)2---4415595.76
DNN.HBA 20.046230(42 30)2---10113595.00
DNN.FA 30.005330(50 30)2---3719296.65
Our studyDNN
CNN
LSTM
GRU
0.001
0.005
0.0005
0.001
30
30
30
30
(16 4 2)
(16)
(128)
(128 128)
1
1
1
1
-
16
-
-
-
3
-
-
-
0.5
0.5
0.5
32
32
32
32
500
50
700
200
97.29
96.56
97.20
96.70
1 Parameter settings for DNN using Bat Algorithm. 2 Parameter settings for DNN using Hybrid Bat Algorithm. 3 Parameter settings for DNN using Firefly Algorithm.
Table 10. Comparison of performance metrics with other studies using the same dataset.
Table 10. Comparison of performance metrics with other studies using the same dataset.
Ref.DL
Algorithm
Performance Metrics
ConventionalAdditional
FPR
(%)
FNR
(%)
ACC
(%)
PR
(%)
RC
(%)
F1
(%)
AUC
(%)
Training Time (min)Testing Time (s)Parameter sizeMemory
Storage (MB)
[18]DNN1.803.3097.5097.7096.7097.20-----
[28]CNN--97.2096.9098.1097.50-10.670.47227,985-
[6]MLP--96.6596.6596.6596.65-----
[64]DNN.BA 1--95.76--95.70-----
DNN.HBA 2--95.00--94.93-----
DNN.FA 3--96.65--96.61-----
Our studyDNN3.012.4797.2997.5397.5397.5399.408.380.4701,507172
CNN3.503.3996.5696.6197.0996.8599.510.90.2633,745325
LSTM1.803.5597.2096.4598.6397.5399.11169.170.80466,689184
GRU2.483.9496.7096.0698.0397.0498.7955.071.447149,505184
1 Parameter settings for DNN using Bat Algorithm. 2 Parameter settings for DNN using Hybrid Bat Algorithm. 3 Parameter settings for DNN using Firefly Algorithm.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Do, N.Q.; Selamat, A.; Krejcar, O.; Yokoi, T.; Fujita, H. Phishing Webpage Classification via Deep Learning-Based Algorithms: An Empirical Study. Appl. Sci. 2021, 11, 9210. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199210

AMA Style

Do NQ, Selamat A, Krejcar O, Yokoi T, Fujita H. Phishing Webpage Classification via Deep Learning-Based Algorithms: An Empirical Study. Applied Sciences. 2021; 11(19):9210. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199210

Chicago/Turabian Style

Do, Nguyet Quang, Ali Selamat, Ondrej Krejcar, Takeru Yokoi, and Hamido Fujita. 2021. "Phishing Webpage Classification via Deep Learning-Based Algorithms: An Empirical Study" Applied Sciences 11, no. 19: 9210. https://0-doi-org.brum.beds.ac.uk/10.3390/app11199210

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop