Abstract

Face recognition is an important technology with practical application prospect. One of the most popular classifiers for face recognition is support vector machine (SVM). However, selection of penalty parameter and kernel parameter determines the performance of SVM, which is the major challenge for SVM to solve classification problems. In this paper, with a view to obtaining the optimal SVM model for face recognition, a new hybrid intelligent algorithm is proposed for multiparameter optimization problem of SVM, which is a fusion of cultural algorithm (CA) and emperor penguin optimizer (EPO), namely, cultural emperor penguin optimizer (CEPO). The key aim of CEPO is to enhance the exploitation capability of EPO with the help of cultural algorithm basic framework. The performance of CEPO is evaluated by six well-known benchmark test functions compared with eight state-of-the-art algorithms. To verify the performance of CEPO-SVM, particle swarm optimization-based SVM (PSO-SVM), genetic algorithm-based SVM (GA-SVM), CA-SVM, and EPO-SVM, moth-flame optimization-based SVM (MFO-SVM), grey wolf optimizer-based SVM (GWO-SVM), cultural firework algorithm-based SVM (CFA-SVM), and emperor penguin and social engineering optimizer-based SVM (EPSEO-SVM) are used for the comparison experiments. The experimental results confirm that the parameters optimized by CEPO are more instructive to make the classification performance of SVM better in terms of accuracy, convergence rate, stability, robustness, and run time.

1. Introduction

As an important branch of pattern recognition, face recognition has the widespread application demands in the field, such as intelligent monitoring, virtual reality, medicine examination, and human-computer interaction. The key step in face recognition is classifier design, namely, how to use the extracted features to classify the new face images. Recently, many classifiers have been presented, such as decision tree [1], neural network [2], k-nearest (KNN) [3], and support vector machine (SVM) [4], among which SVM is the most popular one.

Although SVM has the advantages of good generalization ability, small training set size limitation, and high noise stability, it sometimes leads to overfitting problems. To address this problem, some researchers have applied genetic algorithm (GA) [5], particle swarm optimizer (PSO) [6], and simulated annealing (SA) [7] to parameter optimization of SVM. However, the implement of GA is complex although the search process has the characteristic of exclusivity, which makes the convergence speed slow. PSO and SA are successfully applied to solve some engineering optimization problems, but they have poor global search ability for parameters optimization. Therefore, it is worthy to design a new intelligent algorithm with superior performance for parameters optimization of SVM.

In the face of complex optimization problems which are difficult to deal with effectively by traditional methods, many new inspired algorithms have been proposed for solving problems by simulations of natural rules and social behaviours of biological population. Recently, some algorithms such as grey wolf optimizer (GWO) [8], moth-flame optimization (MFO) [9], tunicate swarm algorithm (TSA) [10], and dice game optimizer (DGO) [11] are widely used. In 2018, Dhiman and Kumar proposed a novel bio-inspired algorithm named emperor penguin optimizer [12], which is inspired by the budding behavior of emperor penguin. However, the experimental results showed the convergence rate of EPO will decrease continuously in the later stage, which will affect the efficiency of optimization. Afterwards, Baliarsingh et al. proposed a hybrid algorithm using EPO with social engineering optimizer (SEO), namely, emperor penguin and social engineering optimizer (EPSEO) [13] to improve the performance of EPO.

Cultural algorithm (CA) is a bio-inspired intelligent algorithm proposed by Reynolds in 1994 [14], which is by simulating the cultural evolution of human society. CA is a double-layer evolutionary mechanism, which uses all kinds of knowledge accumulated in belief space to update continuously, so as to guide the evolution of population space and accelerate the optimization efficiency of the algorithm. Therefore, it can provide a space that allows other evolutionary algorithms to be embedded to cooperate and promote each other. Recently, some researchers have proposed PSO, bacterial foraging algorithm (BFA), and firework algorithm (FA) as a population space to ingrate with CA [1517] and verified that the performance of the hybrid algorithms are significantly improved.

Based on the above observations, due to the fact that CA with single population does not make full use of the knowledge of belief space and the convergence speed of EPO will decrease continuously in the later stage, we propose a hybrid metaheuristic algorithm named cultural emperor penguin optimizer (CEPO) with the help of the embedded framework provided by CA. In CEPO, hybrid of two different evolution processes for locations is an effective mechanism to enhance the exploitation capability of EPO.

Specifically, CEPO are implemented on six well-known benchmark test functions compared with eight state-of-the-art optimization algorithms. Furthermore, CEPO is applied to multiparameter optimization problem of SVM for face recognition. In the process of feature, extraction feature representation vectors of face images are obtained by Gabor filter and Principal Component Analysis (PCA), which are as input to CEPO-SVM. PCA is used to make up Gabor wavelet transform disadvantages caused by increased dimensionality. Along with feature classification, penalty parameter and kernel parameter of SVM are optimized by CEPO to get the training model with high classification accuracy. To prove the performance of CEPO-SVM, eight state-of-the-art optimization algorithm-based SVM are used for the comparison experiments in terms of accuracy, convergence rate, stability, robustness, and run time.

The main contributions of this paper are as follows:(1)A new hybrid intelligent algorithm named CEPO inspired by EPO and the framework of CA is proposed(2)CEPO is applied to automatic learning of the network parameters of SVM to solve the classification problem for face recognition(3)We demonstrate that the performance of CEPO-SVM is superior on the basis of accuracy, convergence rate, stability, robustness, and run time, compared with eight state-of-the-art algorithm-based SVM(4)CEPO as a multidimensional search algorithm can not only obtain the optimal penalty parameter and kernel parameter of SVM but also can be employed in other classifiers and other real number optimization problems

The rest of the work is organized as follows. Section 2 describes feature extraction of face images including the Gabor face image representation and dimensionality reduction by Principal Component Analysis. Section 3 presents formulations of parameter optimization of SVM for classification as an optimization problem. The design of cultural emperor penguin optimizer for parameter optimization of SVM is depicted in Section 4. Section 5 describes the experimental setup. In section 6, the experiment and result analysis are presented. Section 7 provides the conclusion and future works.

2. The Feature Extraction for Face Images

At the preprocessing stage, each image is first converted into grayscale and then adjusted to the same size. Due to the superior robustness of the Gabor filter to changes in brightness and attitude of face images, the Gabor filter is used to capture more useful features from face images. Furthermore, to make up Gabor wavelet transform disadvantages caused by increased dimensionality, PCA is applied to dimensionality reduction. After the preprocessing stage, face images are divided into training sets and testing sets to obtain feature representation vectors, which are as the input to SVM. Figure 1 shows the block diagram of face recognition based on Gabor-PCA.

2.1. Gabor Face Image Representation

The two-dimensional Gabor filter can be defined as follows:where defines the pixel position in the spatial domain, defines the orientation of the Gabor filter, defines the scale of the Gabor filter. represents the radius of Gaussian function, which is used to limit the size of the Gabor filter, and is the wave vector of the filter at orientation and scale [18, 19]. In this paper, with five different scales, , and eight different orientations, , 40 different Gabor filters can be obtained. The real part and the magnitude of the 40 Gabor filters are shown in Figures 2 and 3, which exhibit desirable characteristics of spatial locality, spatial frequency, and orientation selectivity.

Let represent the grey level distribution of an image, and the Gabor wavelet representation of an image can be given bywhere defines the convolution result between image and the Gabor filters at different orientations and scales. In order to improve the computation speed, the convolution computation can be transformed by FFT [20]. To contain the different spatial localities, spatial frequencies, and orientation selectivities, all these representation results will be concatenated to derive an augmented feature vector . In [19], before the concatenation, each will be downsampled by a factor at different scales and orientations. Therefore, the augmented Gabor feature vector can be given bywhere denotes transpose operator, is the downsample factor, and is the concatenated column vector.

2.2. Principle Component Analysis for Feature Optimization

The augmented Gabor feature vector which is introduced in equation (3) has very high dimensionality: , where is the dimensionality of vector space. Obviously, the high dimensionality seriously affects the computing speed and recognition rate in the process of face recognition. In the method with PCA [21], we can find orthogonal basis for feature vector, sort dimensions in order of importance, and discard low significance dimensions. Let be the covariance matrix of the augmented feature vector :where is the expectation operator. The covariance matrix can be transformed into the following form:where is an orthogonal eigenvector matrix and is a diagonal eigenvalue matrix, and the diagonal elements are arranged in descending order .

An important property of PCA is that it is the optimal signal reconstruction in the sense of minimum mean-square error by using a subset of principal components to represent the original signal [19]. Following this property, PCA can be applied to dimensionality reduction:where , , and .The lower dimensional vector captures the most expressive features of the original data .

3. Formulations of Parameter Optimization of SVM for Classification

3.1. Model of Support Vector Machine

Support vector machine (SVM) [22] is a kind of binary classifier with strong learning ability and generalization ability. The key point of SVM is to find the optimal hyperplane for accurate classification of two classes and ensure that the classification gap is large enough. Figure 4 shows the optimal hyperplane of SVM, where indicates the hyperplane and margin represents the gap between class and class .

Suppose the training data can be given bywhere are the training samples, are the labels in -dimensional vector, is the number of training data, and each variable must subject to the criteria and . For linear data, the hyperplane that separates the given data can be determined:where is -dimensional vector and is a scalar. and determine all the boundaries. SVM could be classified according to the problem types, when the hyperplane with minimum margin width subject to

For the data that cannot be separated linearly, the relaxing variables are introduced to allow for optimal generalization and classification by controlling the size of them. Then, the optimal hyperplane can be obtained by the following optimization problem:where denotes the penalty parameter. To satisfy the Karush–Kuhn–Tucker (KKT) conditions, the Lagrange multipliers is introduced. The abovementioned optimization problem is converted into the dual quadratic optimization problem:

Thus, the ultimate classification function can be obtained by solving the dual optimization problem:where represents symbolic function.

Kernel functions can be used to map the data in the low-dimensional input space to the high-dimensional space by nonlinearity. Therefore, the linear indivisibility of the input can be transformed into the linear separable problem by this mapping. Kernel function can be defined as follows:

Then, the optimal function can be given by

Several types of kernel functions such as Linear Kernel, Polynomial Kernel, and Radial Basis Function (RBF) are widely used. However, RBF has advantages of realizing nonlinear mapping, less parameters, and less numerical difficulties. In this paper, RBF is selected for face recognition, which can be represented as follows:where denotes the kernel parameter.

3.2. Objective Function of SVM

With a view to obtaining training model with high classification accuracy, the penalty parameter which is used to adjust confidence scale and the kernel parameter which determines the width of kernel and the range of data points could be optimized by an effective intelligent algorithm. Furthermore, the mean square error of parameters of SVM can be selected as the objective function of the intelligent algorithm:where is the output value of the corresponding parameter and is the actual value of the corresponding parameter.

4. Design of Cultural Emperor Penguin Optimizer for Parameter Optimization of SVM

4.1. The Design of Cultural Emperor Penguin Optimizer

EPO is a novel optimization algorithm presented by Dhiman and Kumar in 2018 [12], which is inspired by the budding behavior of emperor penguin. In this paper, a hybrid algorithm named CEPO is proposed to solve the real numbers optimization problems, which is with the help of cultural algorithm basic framework in Figure 5. The key idea of CEPO is to obtain problem-solving knowledge from the budding behavior of emperor penguin and make use of that knowledge to guide the evolution of emperor penguin population in return.

Suppose CEPO is designed for general minimum optimization problem:where denotes the position of the emperor penguin in the -dimensional search space, . is the objective function, and denotes the objective value of the position . and represent the lower and upper boundary of the position of emperor penguins in the dimension.

Belief space of emperor penguin population in the generation defined is given by and , where is situational knowledge component. is normative knowledge which represents the value space information for each parameter in the dimension and in the generation. denotes . , where represents the interval of normative knowledge in the dimension. The lower boundary and the upper boundary are initialized according to the value range of variables given by the problem. represents the objective value of the lower boundary of the parameter, and represents the objective value of the upper boundary of the parameter.

The acceptance function is used to select the emperor penguins who can directly influence the current belief space. In CEPO, the acceptance function selects the cultural individual in proportion top 20% from the current population space to update the belief space.

Situational knowledge can be updated by update function:where is the optimal position of emperor penguin population space in the generation.

Assume that, for the cultural individual, a random variable lies in the range of is produced. The cultural individual affects the lower boundary of normative knowledge in the dimension when is satisfied. Normative knowledge can be updated by update function:

The cultural individual affects the upper boundary of normative knowledge in the dimension when is satisfied:

Situational knowledge and normative knowledge can be used to guide emperor penguin population evolution by the influence function. In CEPO, a selection operator is produced to select one of two ways to influence the of evolution emperor penguin population:where denotes the maximum number of iterations.

Assume that, for the emperor penguin, a random variable which lies in the range of is produced. The first way is to update the position of emperor penguin by changing the search size and direction of the variation with belief space, which will be implemented when satisfied . The position of emperor penguin in the dimension can be updated bywhere is a random number subjecting to the standard normal distribution. is the length of adjustable interval of the parameter in belief space in the generation. is set to be in the range of by [14].

The other way is by a series of steps in EPO which are the huddle boundary generation, temperature profile around the huddle computing, the distance calculation between emperor penguins, and the position update of emperor penguins, which will be carried when satisfied . The specific steps can be represented as follows:where represents the temperature profile around the huddle, is the time for finding best optimal solution, and is a random variable which lies in the range of .where denotes the distance between the emperor penguin and the optimal solution, represents the current optimal solution found in emperor penguin population space in the generation, represents the social forces of the emperor penguin that is responsible for convergence towards the optimal solution, and are used to avoid the collision between neighboring emperor penguins, and is a random variable which lies in the range of . can be computed as follows:where is the movement parameter which holds a gap between emperor penguins for collision avoidance and defines the absolute difference by comparing the difference between emperor penguins. in equation (28) is computed as follows:where represents the base of natural logarithm. and are two control parameters for better exploration and exploitation, which is in the range of and . Ultimately, the position of emperor penguin is updated as follows:

The algorithm procedure of CEPO is sketched in Algorithm 1.

Algorithm: CEPO
Input: the emperor penguin population
Output: the optimal position of emperor penguin
(1)Procedure CEPO
(2)Initialize the size of population, , the parameters of EPO and CA
(3)Initialize the situational knowledge and normative knowledge of the belief space
(4)Initialize the population space of Eps
(5)Compute the fitness of each EP
(6)Arrange the fitness value of each EP
(7)Initialize the belief space by acceptance proportion 20% and AdjustCulture
(8)Initialize the belief space
(9)While () do
(10)  Calculate and using equations. (9) and (10)
(11)  For1 to do
(12)   
(13)    For1 to do
(14)    If using equation (7)
(15)    Update the position of EPs using equation (8)
(16)    Else
(17)    Compute and using equations (12) and (13)
(18)    Compute using equation (14)
(19)    Compute the new position of EPs using equation (15)
(20)    Compute the fitness of current Eps
(21)    Update the better position of EP compared with the previous position
(22)    End if
(23)   End for
(24)  End for
(25)  Amend EPs which cross the boundary
(26)  Arrange the fitness value of each EP
(27)  Update the belief space by acceptance proportion top 20% and AdjustCulture
(28)  
(29)End while
(30)Return the optimal position
(31)End procedure
(32)Procedure AdjustCulture
(33) Update situational knowledge using equation (2)
(34)  For to do
(35)   
(36)   If
(37)   For to do
(38)   Update the lower boundary of normative knowledge using equation (3) and (4)
(39)   Else
(40)   For to do
(41)   Update the upper boundary of normative knowledge using equation (5) and (6)
(42)   End if
(43)  End for
(44)End for
(45)End procedure
4.2. CEPO for Parameter Optimization of SVM

During the iterative searching process, each emperor penguin adjusts the optimal position by comparing the value of the objective function, and the optimal position is considered to be the parameter pair ( and ) for repeated training of SVM and update of each emperor penguin. Figure 6 shows a schematic diagram of SVM based on CEPO.

5. The Experiments

5.1. Face Databases

In the following experiments, to test our proposed model, we select four face databases that are used commonly and internationally, i.e., YALE (http://cvc.yale.edu/projects/yalefaces/yalefaces.html), ORL (http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html), UMIST (http://images.ee.umist.ac.uk/danny/database.html), and FERET (http://www.frvt.org/). Specifically, the YALE database contains 165 images of 15 individuals, including changes in illumination, expression, and posture. The ORL database consists of 400 photographs of 40 human subjects with different expressions and variations at different scales and orientations. The UMIST database includes 564 images of 20 human subjects with different angles and poses. The FERET database consists of a total of 1400 images for 200 human subjects, corresponding to different poses, expressions, and illumination conditions. To fully verify the validity and rationality of the experiments, configuration of training and testing samples on four face databases is shown in Table 1.

5.2. State-of-the-Art Algorithms for Comparison

To validate the performance of CEPO, the eight state-of-the-art metaheuristics are used for comparison. These are Moth-Flame Optimization (MFO) [9], Grey Wolf Optimizer (GWO) [8], Particle Swarm Optimization (PSO) [6], Genetic Algorithm (GA) [5], Cultural Algorithm (CA) [14], Emperor Penguin Optimizer (EPO) [12], Cultural Firework Algorithm (CFA) [17],and Emperor Penguin and Social Engineering Optimizer (EPSEO) [13].

5.3. Experimental Setup

The experimental environment is MATLAB 2018a with LIBSVM on the computer with Intel Core i5 processor and 16 GB of RAM. Table 2 Shows the parameter settings of the proposed CEPO and eight competitor intelligent algorithms, i.e., MFO, GWO, PSO, GA, CA, EPO, CFA, and EPSEO. And the parameter values of these algorithms are set to be recommended in their original paper.

6. Simulation Results and Discussion

6.1. Performance of CEPO Comparison Experiment

Six benchmark test functions are applied to verify the superiority and applicability of CEPO. Table 3 shows the details of six benchmark test functions including single-peak benchmark functions which are Sphere (F1) and Schwefel 2.22 (F2) and multipeak benchmark functions which are Rastrigin (F3), Griewank (F4), Ackley (F5), and Schaffer (F6). For each benchmark test function, each algorithm carries out thirty independent experiments. The mean and the standard deviation of the thirty independent results are shown in Table 4. As we can see, the statistical results of CEPO for six benchmark test functions are obviously superior to the other eight comparison algorithms. Under the same optimization conditions, the two evaluation indexes of CEPO are both superior to several or even ten orders of magnitude of other algorithms, which proves that the CEPO has strong optimization performance. The optimal mean index shows that the CEPO maintains a high overall optimization level in thirty independent experiments and implies that CEPO has better single optimization accuracy. And the optimal standard deviation index proves that CEPO has strong optimization stability and robustness. Furthermore, under the same test constraints, CEPO attains the high optimization accuracy in the single-peak functions (F1-F2) and multipeak functions (F3-F6) optimization process, which indicates that the CEPO has strong local mining performance and good convergence accuracy, especially for the four multipeak functions (F3-F6) still maintain a high optimization objective value, which verifies its better global exploration performance and stronger local extreme value avoidance ability. If CEPO takes the target value optimization precision as the evaluation index, the preset precision threshold can be as high as 1.0E-12. Especially, for function F1、F3、F6, the threshold value can even be relaxed to 1.0E-25 under the premise of ensuring the success rate of optimization, and the precision can effectively meet the error tolerance range of engineering problems, which further proves that CEPO has better application potential in engineering optimization problems.

6.2. Performance of CEPO-SVM Comparison Experiment

To verify the superiority of the proposed CEPO-SVM classification model for face recognition, CEPO and eight competitor intelligent algorithms are used to iteratively optimize the penalty parameter and kernel parameter of SVM to assess the performance of CEPO. The average recognition rates of thirty independent runs are shown in Table 5. As can be seen clearly from Table 5, CEPO-SVM achieves the highest accuracy among all nine models. In case of YALE face database, an accuracy of 98.1% is obtained by CEPO-SVM model. For ORL, UMIST, and FERET face database, CEPO-SVM gets an accuracy of 97.8%, 98.4%, and 98.5%, respectively. Furthermore, the convergence rate of CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM are examined and depicted in Figures 7(a)7(d). From this figure, it can be observed that the recognition rate successively increases from iteration 1 to iteration 100 on all four face databases. For YALE database, the accuracy rate has not improved after 14th, 22nd, 31st, 19th, 18th, 21st, 20th, 23rd, and 24th iterations using CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM, respectively. Similarly, for ORL database, no improvement in accuracy is noticed after 11th, 20th, 22nd, 19th, 15th, 20th, 18th, 21st, and 24th iterations using CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM, respectively. In case of UMIST database, CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM converge after 15th, 24th, 25th, 19th, 20th, 26th, 18th, 24th, and 21st iterations, respectively. For FERET database, no rise in accuracy is seen after 16th, 22nd, 24th, 19th, 18th, 20th, 18th, 21st, and 25th iterations using CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM, respectively. Obviously, CEPO-SVM converges firstly among nine proposed models, which verifies that EPO with the help of the embedded framework provided by CA could remarkably enhance convergence rate. Therefore, the parameters of optimization by CEPO are superior to make the classification performance of SVM better.

6.3. Stability of CEPO-SVM Comparison Experiment

For all nine models, the stability of the models is measured by the mean and standard deviation of penalty parameter and kernel parameter . In this experiment, each model runs thirty experiments. Tables 69 show the comparison of mean and standard deviation of parameters among nine models on four face databases. As we can see from Tables 69, the mean of penalty parameter in CEPO-SVM is less than that of other models on four face databases, which could reduce the possibility of overlearning of SVM. And CEPO-SVM provides the lowest standard deviation of penalty parameter among all nine models on all four databases, which verifies the stability of CEPO-SVM is better. For all four databases, the mean and standard deviation of kernel parameter are close among the nine models. However, it can still be seen that CEPO-SVM has the minimum standard deviation of kernel parameter . Obviously, CEPO is more stable to optimize the parameters of SVM for face recognition.

6.4. Robustness of CEPO-SVM Comparison Experiment

To prove the robustness of CEPO-SVM, salt and pepper noise is added to each image of the original four databases. The noise databases generated are used for face recognition by CEPO-SVM, PSO-SVM, GA-SVM, EPO-SVM, CA-SVM, MFO-SVM, GWO-SVM, CFA-SVM, and EPSEO-SVM. For each noise databases, we selected 70% of noise samples as the training samples and the rest 30% of noise samples as the testing samples. Salt and pepper noise is a random black or white pixels in an image. Four different levels of noise image on YALE database which are 0%, 5%, 10%, and 15% are shown in Figure 8. From Figure 8, we can see that, as the noise level improves, the face image becomes more blurred. Figure 9 shows the detailed recognition rates among different models on four databases. As we can see from Figure 9, CEPO-SVM obtains the highest recognition rate at all four noise levels on among four databases in comparison with other eight models. Although the face image is seriously polluted by salt and pepper noise which is up to 15%, the recognition rate of CEPO-SVM can still attain 82.5% on YALE database, 83.5% on ORL database, 81.1% on UMIST database, and 82.1% on FERET database, which far exceed that of other eight models. Besides, with the increase of noise level, the recognition rate of CEPO-SVM on four databases decreases more slowly than that of other eight models. Obviously, the results show that CEPO-SVM can not only attain good noise-free face recognition accuracy but also achieve good accuracy under noise conditions, which verifies that robustness of CEPO-SVM is superior to the other eight models.

6.5. Run Time Analysis

The average training time of thirty independent runs by the proposed CEPO-SVM and eight competitor intelligent algorithm-based SVMs are shown in Table 10. As we can see from Table 10, in case of all four databases, the average training time by the proposed CEPO-SVM is more than that of six models with single algorithms, i.e., MFO-SVM, GWO-SVM, PSO-SVM, GA-SVM, EPO-SVM, and CA-SVM. This may be because hybrid of two algorithms increases the training time. Although CEPO-SVM takes more time than these six models with single algorithms, it can be considered better than these six models because of its high classification performance. However, for all four databases, the average training time by the proposed CEPO-SVM is less than that of two models with hybrid algorithms, i.e., CFA-SVM and EPSEO-SVM. Obviously, CEPO-SVM has better performance than these two models with hybrid algorithms.

6.6. Limitation Analysis

Like other hybrid algorithms, although CEPO provides superior performance in face recognition, due to large amount of computation, the proposed CEPO will cost more time for iterations compared with single algorithms. Besides, when CEPO is applied to solve super-dimensional optimization problems, the performance may be reduced.

7. Conclusions

In this paper, with respective advantages of EPO and CA, we propose a hybrid metaheuristic algorithm named CEPO with the help of the embedded framework provided by CA to promote each other and overcome each other’s self-defects. The proposed algorithm has been experimented on six benchmark test functions compared with other eight state-of-the-art algorithms. Furthermore, we apply this hybrid metaheuristic algorithm to automatic learning of the network parameters of SVM to solve the classification problem for face recognition. The proposed classifier model has been tested on four face databases that are used commonly and internationally by comparing with models with other eight state-of-the-art algorithms. The experimental results show that the proposed model has better performance in terms of accuracy, convergence rate, stability, robustness, and run time.

We provide future works in different directions. First, the proposed model has only been experimented on four face databases; however, other new face databases should be tested to extend the current work. Second, the proposed hybrid algorithm should be applied to other classifiers for face recognition. Lastly, the proposed hybrid algorithm should be applied to solve other optimization problems, such as digital filters design, cognitive radio spectrum allocation, and image segmentation.

List of Definitions of Partial Parameters

:The position of the emperor penguin in the -dimensional search space
:Situational knowledge component
:Normative knowledge in the dimension and in the generation
:The interval of normative knowledge in the dimension
:The lower boundary in the dimension and in the generation
:The objective value of the lower boundary of the parameter
:The higher boundary in the dimension and in the generation
:The objective value of the upper boundary of the parameter
:The distance between the emperor penguin and the optimal solution
:The movement parameter
:The scalar
:The penalty parameter
:The relaxing variables
:The training samples
:The Lagrange multipliers
:The scale of Gabor filter
:The orientation of Gabor filter
:The concatenated column vector
:The covariance matrix of the augmented feature vector
:The orthogonal eigenvector matrix
:The diagonal eigenvalue matrix

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This research was supported by the National Natural Science Foundation of China (Grant no. 61571149), Heilongjiang Province Key Laboratory Fund of High Accuracy Satellite Navigation and Marine Application Laboratory (Grant no. HKL-2020-Y01), and Initiation Fund for Postdoctoral Research in Heilongjiang Province (Grant no. LBH-Q19098).