Next Article in Journal
Analysis of Soot Deposition Mechanisms on Nickel-Based Anodes of SOFCs in Single-Cell and Stack Environment
Next Article in Special Issue
Improving the Imazapyr Degradation by Photocatalytic Ozonation: A Comparative Study with Different Oxidative Chemical Processes
Previous Article in Journal
Carbon Emission Reduction Potential in the Finnish Energy System Due to Power and Heat Sector Coupling with Different Renovation Scenarios of Housing Stock
Previous Article in Special Issue
Characterisation of Bioactive Ingredients in Extracts of Fresh and Dried Coniferous Trees for the Development of Sustainable Packaging Materials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of the Solubility of CO2 in Imidazolium Ionic Liquids Based on Selective Ensemble Modeling Method

1
College of Chemical Engineering, Zhejiang University of Technology, Hangzhou 310014, Zhejiang, China
2
Zhejiang Province Key Laboratory of Biomass Fuel, Hangzhou 310014, Zhejiang, China
*
Author to whom correspondence should be addressed.
Submission received: 13 September 2020 / Revised: 26 October 2020 / Accepted: 27 October 2020 / Published: 28 October 2020
(This article belongs to the Special Issue Green Sustainable Chemical Processes)

Abstract

:
Solubility data is one of the essential basic data for CO2 capture by ionic liquids. A selective ensemble modeling method, proposed to overcome the shortcomings of current methods, was developed and applied to the prediction of the solubility of CO2 in imidazolium ionic liquids. Firstly, multiple different sub–models were established based on the diversities of data, structural, and parameter design philosophy. Secondly, the fuzzy C–means algorithm was used to cluster the sub–models, and the collinearity detection method was adopted to eliminate the sub–models with high collinearity. Finally, the information entropy method integrated the sub–models into the selective ensemble model. The validation of the CO2 solubility predictions against experimental data showed that the proposed ensemble model had better performance than its previous alternative, because more effective information was extracted from different angles, and the diversity and accuracy among the sub–models were fully integrated. This work not only provided an effective modeling method for the prediction of the solubility of CO2 in ionic liquids, but also provided an effective method for the discrimination of ionic liquids for CO2 capture.

1. Introduction

With the increase of energy consumption in industrial production, reducing CO2 emissions and increasing CO2 absorption have become an essential means to alleviate environmental degradation [1]. Room–temperature ionic liquids, which are relatively new compounds, have gained much attention in recent years, and had the potential to be considered as an alternative to conventional volatile organic solvents in the reaction and separation processes. Information about the solubility and the rate of solubility is a crucial factor for consideration of ionic liquids in potential industrial processes [2,3]. A large number of ionic liquids can be synthesized due to their special ionic structure. Due to some difficulties associated with experimental measurements and the cost of ionic liquids, it is more advantageous to develop predictive methods for prediction of the phase behavior of such systems [4,5,6]. Therefore, the modeling prediction methods have become an important way to obtain the solubility data of CO2 in ionic liquids, which is divided into the mechanism modeling method and the data–driven modeling method.
In order to understand and predict the phase behavior of CO2 and ionic liquid mixtures, Perturbed Hard Sphere Chain Equation of State (PHSC) has been selected to simulate the CO2 absorption in a series of ionic liquids [7]. CO2 solubility in ionic liquids had been calculated based on two thermodynamic models, namely the UNIQUAC model and the quantum model, based on COSMO–RS theory of interacting molecular surface charge [8]. Venkatramanet et al. [9] used molecular descriptors based on quantum chemistry computations to predict the solubility of CO2 in different ionic liquids. Although the mechanism modeling method has the advantage of strong model interpretability, thermodynamic model is relatively complex and requires complicated mathematical operations [10]. Considering the complex parameters of mechanism model, a multi–model fusion method was proposed to predict the solubility of CO2 in ionic liquids [11].
Based on the radial basis function artificial neural network (RBFANN) and least squares support vector machine (LSSVM) combined with group contribution (GC) method, RBFANN–GC and LSSVM–GC were used to study the model of CO2 absorption in polyionic liquids [12]. Four different methods based on artificial intelligence were proposed to predict CO2 solubility in different ionic liquids [13]: the predictive capability of the Least Square Support Vector Machine (LSSVM), Adaptive Neuro–Fuzzy Inference System (ANFIS), Multi–Layer Perceptron Artificial Neural Network (MLP–ANN), and Radial Basis Function Artificial Neural Network (RBF–ANN) have been evaluated for estimating carbon dioxide solubility in 67 different ionic liquids [14]. The Classification And Regression Tree (CART) methodology in modeling CO2 solubility in different ionic liquids is also under investigation [15]. Artificial neural networks (ANNs) technique was proposed as a new approach to predict the solubility of CO2 in ethanol–[EMIM][Tf2N] ionic liquid mixtures [16]. Although the data–driven models have the advantage of high prediction accuracy, they are mainly based on the single models, which have shortcomings such as easy to fall into local optimality and cannot describe the global characteristics of the problem. As a result, the prediction performance of any of these models is limited.
A selective ensemble method based on data–driven model is proposed in this paper for the deficiency of the current mechanism model method and data–driven model method. Firstly, on the premise of ensuring the accuracy and diversity of the ensemble, fuzzy C–means and collinearity detection algorithms were used to eliminate redundant models. Then the sub–models with different predictive capabilities were integrated by using the information entropy method, so as to fully mine the performance of various models, realize the full coverage of information describing the problem and improve the prediction accuracy of the model. Finally, the method was applied to predict the solubility of CO2 in ionic liquids and to evaluate its prediction performance.

2. Methods

2.1. Sub–Model Training

Zhou et al. [17] proposed the theory of ‘many can be better than all’, which presupposed that the sub–models had a high degree of diversity and accuracy. The diversity of sub–models had an essential impact on improving the generalized performance of ensemble models [18,19]. The sub–models were established based on data, structural, and parameter diversities.

2.1.1. Data Diversity

Multiple datasets from the original dataset were generated based on data diversity to train different sub–models. The data sets should be different from each other to obtain different results from the trained sub–models. Bootstrap aggregation (Bagging), Adaptive Boosting (AdaBoost), and random subspace were commonly used to achieve data diversity. To generate several training sets with different attributes, the bootstrap algorithm was introduced to achieve the goal. When re–sampling was enough, about 36.8% of the given data sets did not appear in the constructed training set, which ensured the diversity of the data.

2.1.2. Structural Diversity

Different model structures were used to induce structural diversity, and three data–driven algorithms (Back Propagation Neural Network (BPNN), Extreme Learning Machine (ELM), and Radial Basis Function Neural Network (RBFNN)) were used to train the generated sub–training sets. These sub–models varied in size and architecture, and such collections were called heterogeneous ensemble [20]. To control the diversity in the heterogeneous integration, the ‘overproduce and choose’ strategy was performed. Firstly, a large number of models were trained; then, a selection or combination of these models was made to optimize performance, which purpose was to minimize the size of the ensemble model without significantly reducing the accuracy of the model [21,22].

2.1.3. Parameter Diversity

Parameter diversity uses different parameter sets to generate different sub–models. Even if the same training set is used, the output of the sub–model may vary with different parameter sets. The method of adjusting the internal parameters of the model was adopted to ensure the diversity of model parameters. For BPNN and ELM, the internal parameters of the adjusted model are the number of neurons, the activation function and the number of hidden layers. For RBFNN, the internal parameters of the adjusted model are the kernel function center and width.

2.2. Sub–Model Discrimination

To improve the prediction ability of the ensemble model or reduce the prediction cost, it is necessary to screen the established sub–models and avoid multicollinearity in the sub–model as much as possible. As one of the main techniques of unsupervised machine learning, fuzzy clustering analysis had been widely used in large–scale data analysis, data mining, pattern recognition, and other fields [23,24,25]. The fuzzy C–means clustering was used to screen the sub–model in this study.
N sub–models are defined, the parameters of each model are wi (I = 1,2,…,N), c sub–models are clustered centers and denoted by mj (j = 1,2,…,c), the sample set is S = {(x1,y1),(x2,y2),…,(xn,yn)}, where x is the input variable, y is the output variable, and n is the number of data in the sample set. For all the sub–models to be clustered, the difference between the models can be measured by the Euclidean distance between the sub–models. The calculation formula of the Euclidean distance is as follow:
D ( w i , m j ) = k d ( y ( w i , x k ) , y ( m j , x k ) )
where d ( r 1 , r 2 ) = r 1 r 2 2 is the distance between the sub–models, and y ( w i , x k ) and y ( m j , x k ) represent the output of x k on the parameters wi and mj, respectively. The results of the sub–models on the input data set were adopted to define the difference between the models. The larger the Euclidean distance was, the greater the difference between the two sub–models was.
To perform cluster analysis on the sub–models, the outputs of each sub–model on the sample point xk (k = 1,2,…,n) is composed into a vector, namely: zi = (y(wi,x1),y(wi,x2),…,y(wi,xn)) (I = 1,2,…,N), the outputs of N sub–models with dimension n can be obtained. To determine the optimal number of clusters output by the N sub–models, CH indicators and Davies—Bouldin (DB) could be used as evaluation indicators. In consideration of computational efficiency, the CH evaluation index was utilized to determine the optimal number of clusters. The CH index used the intra–class dispersion matrix to describe the tightness and the inter–class dispersion matrix to describe the separation. The specific calculation formula is as follow:
CH ( k ) = t r B ( k ) / ( k 1 ) t r W ( k ) / ( n k )
where n is the number of clusters, k is the current class, trB(k) is the trace of the inter–class dispersion matrix, and trW(k) is the trace of the intra–class dispersion matrix. The larger the CH is, the closer and the more disperse the class is. When the number of clusters is 1, the CH evaluation index cannot be used.
In the clustering process, when the difference between the two sub–models is very large, it means that the two sub–models are likely to be in different clusters, otherwise they may be in the same cluster. Due to the similarity of the sub–models in the same cluster, the output results obtained by these sub–models under the same input are similar [26]. When the fuzzy C–means algorithm is applied to over–generated sub–models, it is necessary to detect the collinearity of the sub–models in each cluster [27]. Belsleyet et al. [28] believed that the existence of collinearity between models will not only increase the workload of modeling, but also affect the actual performance of the model. The variance expansion coefficient (VIF) was used to judge the collinearity of each sub–model. The larger the value of VIF was, the more serious the collinearity was. 10 are taken as the judgment boundary: when VIF < 10, there is no multicollinearity; when 10 ≤ VIF ≤ 100, there is strong multicollinearity; when VIF ≥ 100, there is serious multicollinearity. The calculation formula of VIF is as follows:
VIF = 1 1 R 2 i
where Ri represents the multiple determination coefficient of the independent variable xi for the regression analysis of other independent variables.

2.3. Sub–Model Ensemble

The weight coefficient of the sub–model generally reflects the degree of influence of the sub–model on the ensemble model. The reasonable determination of the weight coefficients of the sub–models will directly affect the prediction accuracy of the model, so it is necessary to adopt appropriate methods to determine the weight coefficients of each sub–model. Information entropy is an effective measurement tool to describe information content (information structure, uncertainty, etc.). Using the information entropy method to determine the weight coefficient of each sub–model can effectively reduce the impact of the weak sub–model on the model performance [29]. In this paper, the information entropy method is used to obtain the weight coefficient of the optimal sub–model [11].

2.4. Implementation Step

A new selective ensemble modeling method was established, and its implementation process is shown in Figure 1, which mainly includes data collection and grouping, sub–model training, sub–model discrimination, and sub–model ensemble and model performance testing. The specific implementation steps were as follows:
(1) Data collection and grouping
The appropriate auxiliary variables were determined as the input variables of the model, and the dominant variable to be predicted were taken as the output variable of the model. Firstly, the collected original samples data set was normalized, and then the preprocessed original sample data set was randomly divided into a training set, a validation set, and a test set in an appropriate ratio. The training set was ensured to cover all the types of the experimental data and operating conditions.
(2) Sub–model training
Firstly, different training sets were generated based on the diversity of the data. Then, multiple BPNN sub–models, ELM sub–models, and RBFNN sub–models with different structural parameters were established by using the realization method of structure diversity and parameter diversity, so that these sub–models have the characteristics of high diversity and accuracy.
(3) Sub–model discrimination
The validation set was used to evaluate the predictive performance of all sub–models, and the Euclidean distance was used as the standard to evaluate the differences of sub–models. The fuzzy C–means algorithm was adopted to cluster all the sub–models, and the Calinski–Harabasz (CH) method was used to determine the optimal cluster number. After clustering, the collinearity detection method was used to eliminate some sub–models with high collinearity in the same cluster, and only some of the sub–models without collinearity were retained.
(4) Sub–model ensemble and model performance testing
The information entropy method was utilized to calculate the weight coefficients of the retained sub–models, so as to establish the selective ensemble model. Then the test set was used to evaluate the prediction performance of the model.

3. Results and Discussion

3.1. Data Collecting and Grouping

Six essential parameters, including temperature, pressure, critical temperature (Tc), critical pressure (Pc), molecular weight (MW), and eccentricity factor (w) were taken as input variables for the CO2 solubility predictive models [30,31]. Temperature and pressure will affect the solubility of CO2 in the ionic liquid. For the same ionic liquid, the solubility of CO2 in the ionic liquid increases when the temperature decreases or the pressure increases. Theoretically, Tc, Pc, M and w are the essential thermodynamic properties of ionic liquids. They can distinguish the species of ionic liquids and reflect the characteristics of ionic liquid structures [13,31]. In addition, the input variables of the model were only applicable to imidazolium ionic liquids. The solubility of CO2 in ionic liquids was selected as the output variable of the model.
Data of critical temperature (Tc), critical pressure (Pc), molecular weight (M) and eccentricity factor (w) of nine imidazolium ionic liquids were collected by referring to a large number of literatures, as shown in Table 1 [7,32,33,34,35,36,37,38,39,40,41]. The name and abbreviation of imidazolium ionic liquid are shown in Table 2. Meanwhile, a large number of data on the solubility of CO2 in the nine imidazolium ionic liquids were collected, as shown in Table 3. A total of 1468 sets of samples were collected. All the solubility of CO2 in ionic liquids in this paper was obtained in the equilibrium phase. The unit of the stoichiometry of reagents gas/ionic liquids is the molar ratio. For all sample data of each type of ionic liquid, 80% (1176 sets) were randomly selected as the training set for training the sub–model, 10% (146 sets) were randomly selected as the validation set for sub–model discrimination and sub–model ensemble, and the remaining 10% (146 sets) was used as the test set for the performance test of the ensemble model.

3.2. Selective Ensemble Model Developing

3.2.1. Sub–Model Training

In order to ensure the diversity of the data, the re–sampling technique (bootstrap) was used to generate 30 sub–training sets. In order to ensure the diversity of the sub–model structure and parameters, BPNN, ELM, and RBFNN were used to divide the generated 30 training sets randomly to these three algorithms, and 30 sub–models were obtained. All sub–models were implemented by MATLAB software (version 2016a, MathWorks, Natick, MA, USA). The structure and parameters of BPNN, ELM, and RBFNN sub–models were as follows:
(1) BPNN
Ten sub–models of BP neural network were established with a single hidden layer structure. The transfer function of the hidden layer was a tansig type excitation function, and the output layer was expanded by a purelin–type excitation function for range expansion. The training termination error was 4 × 10−4, and the learning rate was 0.05. The Levenberg–Marquardt algorithm was used in the training algorithm and the number of hidden layer nodes was 6–15.
(2) ELM
Ten sub–models were established by adjusting the number of hidden layers and activation functions of the extreme learning machine. There were 5 sub–models with sigmoid activation function (the number of hidden layer nodes was 113–117), and 5 sub–models with sin activation function (the number of hidden layer nodes was 115–119).
(3) RBFNN
Ten sub–models were established by adjusting the number of neurons and the activation function. Among them, the activation function used Gaussian kernel function, the center selection of the basis function adopted K–means clustering, the learning rate was 0.1, the training termination error was 1 × 10−4, and the number of neurons was 71–80.

3.2.2. Sub–Model Discrimination

The performance of the 30 sub–models established was evaluated by the validation set. Firstly, the fuzzy C–means algorithm was adopted to cluster all the sub–models, and then, the sub–models with high collinearity were eliminated based on the collinearity detection program. The performance indexes used for model evaluation included the mean absolute error (MAE), root mean square error (RMSE), and correlation coefficient (R2). The specific calculation formula for each index was as follows:
M A E = 1 N i = 1 N | x i x ^ i |
R M S E = i = 1 N ( x ^ i x i ) 2 N
R 2 = 1 i = 1 N ( x ^ i x i ) 2 i = 1 N ( x ^ i x ¯ ) 2
where N was the number of samples, xi was the predicted value of the sample i, x ^ i was the true value of the sample i, and x ¯ was the average of all samples.
The performance index data of each sub–model was obtained from the validation set. The performance index data of the BPNN, ELM, and RBFNN sub–models are shown in Table 4, Table 5 and Table 6 and Figure 2, respectively. It can be seen from Table 4, Table 5 and Table 6 that all BPNN, ELM, and RBFNN sub–models had good model performance.
The fuzzy C–means algorithm was used to cluster all the sub–models in Table 4, Table 5 and Table 6, and the CH index was used as the standard to evaluate the number of clusters. Formula (2) was applied to calculate the value of CH. The specific results are shown in Table 7. It can be seen from Table 7 that when the number of clusters was 3, the value of CH reaches the maximum; thus, the number of clusters was selected as 3. When three classes were selected as clustering target, the following clustering results could be obtained: the first class included seven sub–models (5 BPNN sub–models and 2 ELM sub–models), the second class included 12 sub–models (10 RBFNN sub–models and 2 BPNN sub–models), and the third class included 11 sub–models (3 BPNN sub–models and 8 ELM sub–models).
Since the sub–models might be collinearity after clustering, it is necessary to carry out collinearity detection on the sub–models in the cluster to eliminate the adverse effects of collinearity. Variance Inflation Factor (VIF) was applied to judge the collinearity. The criterion was that there was no collinearity when VIF < 10. According to the criterion, the following results were obtained: 3 BPNN sub–models in the first class, 3 RBFNN sub–models in the second class, 2 BPNN sub–models, and 1 ELM sub–model in the third class.

3.2.3. Sub–Model Ensemble

For the nine sub–models obtained by discrimination, the information entropy method was used to calculate the weight coefficient of each sub–model. The specific results were as follows:
Y = 0.1130 y 11 + 0.0898 y 12 + 0.0875 y 13 + 0.1755 y 21 + 0.1545 y 22 + 0.1633 y 23 + 0.0763 y 31 + 0.0713 y 32 + 0.0688 y 33
where Y is the output of the selective ensemble model based on information entropy, yij (i = 1,2,3, j = 1,2,3) is each sub–model, where i is the number of clusters, and j is the number of sub–models in the clustering number.

3.3. Model Performance Testing

In order to compare the predictive performance of the selective ensemble model based on information entropy (selective ensemble model), the optimal BPNN sub–model (optimal BPNN), the optimal ELM sub–model (optimal ELM), the optimal RBFNN sub–model (optimal RBFNN), and the fully integrated model based on information entropy (fully integrated model) were also established. The test set was used to conduct performance tests on all the above models. The prediction performance of each model is shown in Figure 3. As shown in Figure 3, all models can well realize the prediction of the solubility of CO2 in imidazolium ionic liquids. The histograms of the error distributions of the models are shown in Figure 4. Compared with the single optimal sub–model and the fully integrated model, from the perspective of error distribution, the selective ensemble model had smaller errors, which also verified the effectiveness of the proposed model. In other words, Figure 4 also proved the superiority of the selective ensemble model.
In order to quantitatively compare the prediction performance of the five models, Table 8 gave the specific results of MAE, RMSE, and R2 of the five models based on the testing set.
It can be seen from Table 8 that all models have good prediction performance due to the reasonable selection of relevant physical and chemical parameters and structural parameters as the input of the prediction model for the solubility of CO2 in ionic liquids.
Compared with the three optimal sub–models, the fully integrated model and the selective ensemble model made full use of the advantages of data diversity, parameter diversity, and structural diversity. The sub–models with different structures could excavate more global information contained in the data, and extract the useful information from the data by their different operation mechanisms. Both ensemble models were effective in reducing the error from predictions, thus improving the overall predictive performance. In addition, the information entropy method was used to reasonably select the combination weight coefficients of each sub–model. The models with different predictive abilities were assigned to different weight coefficients. In addition, the differences among the sub–models were fully considered, so that the prediction performance of the model was further improved.
Compared with the fully integrated model based on information entropy, the selective ensemble model based on information entropy used the fuzzy C–means algorithm and the collinearity detection method to screen the sub–models, which further ensured the diversity and accuracy of the models in different clusters, and removed the interference of some sub–models, thus ensuring the effectiveness of the selective ensemble model. Simultaneously, the selective ensemble model based on information entropy further fully mined the information inside the model, and extracted the useful information in the data from different angles to a great extent, and further improved the overall predictive performance.

4. Conclusions

In this paper, a selective ensemble modeling method for predicting the solubility of CO2 in imidazolium ionic liquids was proposed. The implementation process of the selective ensemble modeling method included sub–model training, sub–model discrimination, sub–model ensemble and model performance testing. Sub–model training made full use of the advantages of data diversity, structural diversity, and parameter diversity. Sub–model discrimination used a fuzzy C–means clustering algorithm and collinearity detection method to ensure model diversity and reduce model collinearity. Sub–model ensemble adopted the information entropy weighting method to effectively reduce the impact of weak sub–models on model performance. The result of the prediction performance on the solubility of CO2 in imidazolium ionic liquids showed that the solubility prediction model established by the selective ensemble modeling method had the best prediction performance compared with the other four models.
Although the prediction model established by the fusion modeling method had a good prediction effect for nine imidazolium ionic liquids in this study, it may not be applicable to predicting the solubility of CO2 in other ionic liquids. The research work not only provides a feasible method to obtain the solubility data of CO2 in ionic liquids, but also provides an effective means for further discrimination of ionic liquids, which has important practical significance.

Author Contributions

Conceptualization, L.X. and H.P.; Data curation, S.L.; Formal analysis, S.L.; Funding acquisition, H.P.; Investigation, L.X. and S.L.; Methodology, L.X.; Project administration, H.P.; Resources, H.P.; Software, S.L.; Supervision, H.P.; Validation, L.X. and H.P.; Visualization, S.L.; Writing—original draft, L.X.; Writing—review & editing, H.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (grant number 21676251) and Graduate Practice Base Project of Zhejiang University of Technology.

Acknowledgments

This work was funded by National Natural Science Foundation of China (grant number 21676251) and Graduate Practice Base Project of Zhejiang University of Technology. The authors would also like to acknowledge everyone who has provided helpful guidance and also like to thank the anonymous reviewers for their useful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xu, M.M.; Wang, S.J. Research progress of liquid–liquid phase variable solvent trapping CO2 technology. Chinese J. Chem. Eng. 2018, 69, 1809–1818. [Google Scholar]
  2. Taimoor, A.A.; Al–Shahrani, S.; Muhammad, A. Ionic liquid (1–butyl–3–metylimidazolium methane sulphonate) corrosion and energy analysis for high pressure CO2 absorption process. Processes 2018, 6, 45. [Google Scholar] [CrossRef] [Green Version]
  3. Leonzio, G.; Zondervan, E. Surface–response analysis for the optimization of a carbon dioxide absorption process using [hmim][Tf2N]. Processes 2020, 8, 1063. [Google Scholar] [CrossRef]
  4. Rogers, R.D. Chemistry: Ionic liquids–solvents of the future? Science. 2003, 302, 792–793. [Google Scholar] [CrossRef]
  5. Ding, J.; Xiong, Y.; Yu, D.H. Solubility of CO2 in ionic liquids–measuring and modeling methods. Chem. Ind. Eng. Prog. 2012, 31, 732–741. [Google Scholar]
  6. Bahmani, A.R.; Sabzi, F.; Bahmani, M. Prediction of solubility of sulfur dioxide in ionic liquids using artificial neural network. J. Mol. Liq. 2015, 211, 395–400. [Google Scholar] [CrossRef]
  7. Bazargani, Z.; Sabzi, F. Thermodynamic modeling of CO2 absorption in 1–butyl–3–methylimidazolium–based ionic liquids. J. Mol. Liq. 2016, 223, 235–242. [Google Scholar] [CrossRef]
  8. Kamgar, A.; Rahimpour, M.R. Prediction of CO2 solubility in ionic liquids with QM and UNIQUAC models. J. Mol. Liq. 2016, 222, 195–200. [Google Scholar] [CrossRef]
  9. Venkatraman, V.; Alsberg, B.K. Krakenx: Software for the generation of alignment–independent 3D descriptors. J. Mol. Mode. 2016, 22, 92–100. [Google Scholar] [CrossRef]
  10. Bavoh, C.B.; Lal, B.; Nashed, O.; Khan, M.S.; Keong, L.K.; Bustam, M.A. COSMO–RS: An ionic liquid prescreening tool for gas hydrate mitigation. Chinese J. Chem. Eng. 2016, 24, 1619–1624. [Google Scholar] [CrossRef]
  11. Xia, L.Y.; Wang, J.C.; Liu, S.S.; Li, Z.; Pan, H.T. Prediction of CO2 solubility in ionic liquids based on multi–model fusion method. Processes 2019, 7, 258. [Google Scholar] [CrossRef] [Green Version]
  12. Kardani, M.N.; Baghban, A.; Sasanipour, J.; Mohammadi, A.H.; Habibzadeh, S. Group contribution methods for estimating CO2 absorption capacities of imidazolium and ammonium–based polyionic liquids. J. Clean. Prod. 2018, 203, 601–618. [Google Scholar] [CrossRef]
  13. Tatar, A.; Naseri, S.; Bahadori, M.; Hezave, A.Z.; Kashiwao, T.; Bahadori, A.; Darvish, H. Prediction of carbon dioxide solubility in ionic liquids using MLP and radial basis function (RBF) neural networks. J. Taiwan. Inst. Chem. E. 2015, 60, 151–164. [Google Scholar] [CrossRef]
  14. Alireza, B.; Amir, H.M.; Mohammad, S.T. Rigorous modeling of CO2 equilibrium absorption in ionic liquids. Int. J. Greenh. Gas. Con. 2017, 58, 19–41. [Google Scholar]
  15. Ghiasi, M.M.; Mohammadi, A.H. Application of decision tree learning in modelling CO2 equilibrium absorption in ionic liquids. J. Mol. Liq. 2017, 242, 594–605. [Google Scholar] [CrossRef]
  16. Mirarab, M.; Sharifi, M.; Ghayyem, M.A.; Mirarab, F. Prediction of solubility of CO2 in ethanol–[EMIM][Tf2N] ionic liquid mixtures using artificial neural networks based on genetic algorithm. Fluid. Phase. Equilibr. 2014, 371, 6–14. [Google Scholar] [CrossRef]
  17. Zhou, Z.H.; Wu, J.; Wei, T. Ensembling neural networks: Many could be better than all. Artif. Intell. 2002, 137, 239–263. [Google Scholar] [CrossRef] [Green Version]
  18. MendesMoreira, J.; Soares, C.; Jorge, A.M.; De Sousa, J.F. Ensemble approaches for regression: A survey. Acm. Comput. Surv. 2012, 45, 11–40. [Google Scholar]
  19. Zhu, P.F.; Xia, L.Y.; Pan, H.T. Multi–model fusion modeling method based on improved kalman filter algorithm. Chinese J. Chem. Eng. 2015, 66, 1388–1394. [Google Scholar]
  20. Ren, Y.; Zhang, L.; Suganthan, P.N. Ensemble classification and regression–recent developments, applications and future directions. IEEE. Comput. Intell. M. 2016, 11, 41–53. [Google Scholar] [CrossRef]
  21. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wires. Data. Min. Knowl. 2018, 8, 1249–1266. [Google Scholar] [CrossRef]
  22. Martínez–Muoz, G.; HernándezLobato, D.; Suárez, A. An analysis of ensemble pruning techniques based on ordered aggregation. IEEE T. Pattern. Anal. 2009, 31, 245–259. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Mojarad, M.; Nejatian, S.; Parvin, H.; Mohammadpoor, M. A fuzzy clustering ensemble based on cluster clustering and iterative fusion of base clusters. Appl. Intell. 2019, 49, 2567–2581. [Google Scholar] [CrossRef]
  24. Bowen, F.; Yun, F.Q.; Wan, J.L. Fuzzy clustering ensemble model based on distance decision. Chin. J. Electron. 2018, 54, 823–831. [Google Scholar]
  25. Bagherinia, A.; MinaeiBidgoli, B.; Hossinzadeh, M.; Parvin, H. Elite fuzzy clustering ensemble based on clustering diversity and quality measures. Appl. Intell. 2019, 49, 1724–1747. [Google Scholar] [CrossRef]
  26. Li, K.; Huang, H.K. A selective neural network integration method based on clustering technology. Jcrd. 2005, 42, 594–598. [Google Scholar]
  27. Hashem, S. Optimal linear combinations of neural networks. Neural. Netw. 1997, 10, 599–614. [Google Scholar] [CrossRef]
  28. Fildes, R.B.R. Conditioning diagnostics: Collinearity and weak data in regression. J. Oper. Res. Soc. 1993, 44, 88–89. [Google Scholar]
  29. Rao, Y.Q.; Janet, E. Information entropy–based complexity measurement of manufacturing system and its application in scheduling. Chinese J. Chem. Eng. 2006, 42, 8–13. [Google Scholar]
  30. Zuan, P.; Huang, Y. Prediction of sliding slope displacement based on intelligent algorithm. Wireless. Pers. Commun. 2018, 102, 3141–3157. [Google Scholar] [CrossRef]
  31. Baghban, A.; Ahmadi, M.A.; Hashemi Shahraki, B. Prediction carbon dioxide solubility in presence of various ionic liquids using computational intelligence approaches. J. Supercrit. Fluids. 2015, 98, 50–64. [Google Scholar] [CrossRef]
  32. Yokozeki, A.; Shiflett, M.B.; Junk, C.P.; Grieco, L.M.; Foo, T. Physical and chemical absorptions of carbon dioxide in room–temperature ionic liquids. J. Phys. Chem. B. 2008, 112, 16654–16663. [Google Scholar] [CrossRef] [PubMed]
  33. Yim, J.H.; Lim, J.S. CO2 solubility measurement in 1–hexyl–3–methylimidazolium ([HMIM]) cation based ionic liquids. Fluid. Phase. Equilibr. 2013, 352, 67–74. [Google Scholar] [CrossRef]
  34. Tagiuri, A.; Sumon, K.Z.; Henni, A. Solubility of carbon dioxide in three [Tf2N] ionic liquids. Fluid. Phase. Equilibr. 2014, 380, 39–47. [Google Scholar] [CrossRef]
  35. Schilderman, A.M.; Raeissi, S.; Peters, C.J. Solubility of carbon dioxide in the ionic liquid 1–ethyl–3–methylimidazolium bis(trifluoromethylsulfonyl)imide. Fluid. Phase. Equilibr. 2007, 260, 19–22. [Google Scholar] [CrossRef]
  36. Kim, Y.S.; Jang, J.H.; Lim, B.D.; Kang, J.W.; Lee, C.S. Solubility of mixed gases containing carbon dioxide in ionic liquids: Measurements and predictions. Fluid. Phase. Equilibr. 2007, 256, 70–74. [Google Scholar] [CrossRef]
  37. Kim, Y.S.; Choi, W.Y.; Jang, J.H.; Yoo, K.P.; Lee, C.S. Solubility measurement and prediction of carbon dioxide in ionic liquids. Fluid. Phase. Equilibr. 2005, 228, 439–445. [Google Scholar] [CrossRef]
  38. Carvalho, P.J.; Álvarez, V.H.; Machado, J.J.B.; Pauly, J.; Daridon, J.L.; Marrucho, I.M.; Aznar, M.; Coutinho, J.A.P. High pressure phase behavior of carbon dioxide in 1–alkyl–3–methylimidazolium bis(trifluoromethylsulfonyl)imide ionic liquids. J. Supercrit. Fluids. 2009, 48, 99–107. [Google Scholar] [CrossRef]
  39. Blanchard, L.A.; Gu, Z.; Brennecke, J.F. High–pressure phase behaviour of ionic liquid/CO2 systems. J. Phys. Chem. B. 2001, 105, 2437–2444. [Google Scholar] [CrossRef] [Green Version]
  40. Bermejo, M.D.; Fieback, T.M.; Martín, Á. Solubility of gases in 1–alkyl–3methylimidazolium alkyl sulfate ionic liquids: Experimental determination and modeling. J. Chem. Thermodyn. 2013, 58, 237–244. [Google Scholar] [CrossRef]
  41. Afzal, W.; Liu, X.; Prausnitz, J.M. Solubilities of some gases in four immidazolium–based ionic liquids. J. Chem. Thermodyn. 2013, 63, 88–94. [Google Scholar] [CrossRef]
Figure 1. The implementation processes of the selective ensemble modeling method.
Figure 1. The implementation processes of the selective ensemble modeling method.
Processes 08 01369 g001
Figure 2. Performance indexes of the validation set: (a) Mean absolute error (MAE); (b) Root mean square error (RMSE); (c) R2.
Figure 2. Performance indexes of the validation set: (a) Mean absolute error (MAE); (b) Root mean square error (RMSE); (c) R2.
Processes 08 01369 g002aProcesses 08 01369 g002b
Figure 3. Prediction performances of five models: (a) Optimal BPNN; (b) Optimal ELM; (c) Optimal RBFNN; (d) Fully integrated model; and (e) Selective ensemble model.
Figure 3. Prediction performances of five models: (a) Optimal BPNN; (b) Optimal ELM; (c) Optimal RBFNN; (d) Fully integrated model; and (e) Selective ensemble model.
Processes 08 01369 g003
Figure 4. The histogram of the prediction error distribution of five models: (a) Optimal BPNN; (b) Optimal ELM; (c) Optimal RBFNN; (d) Fully integrated model; and (e) Selective ensemble model.
Figure 4. The histogram of the prediction error distribution of five models: (a) Optimal BPNN; (b) Optimal ELM; (c) Optimal RBFNN; (d) Fully integrated model; and (e) Selective ensemble model.
Processes 08 01369 g004
Table 1. Critical parameters of imidazolium ionic liquids.
Table 1. Critical parameters of imidazolium ionic liquids.
NO.Ionic LiquidsMW
(g/mol)
Tc
(K)
Pc
(MPa)
w
1[BMIM][BF4]226.03623.302.0400.8489
2[EMIM][TF2N]391.30788.053.3101.2250
3[EMIM][ETSO4]236.291061.104.0400.3368
4[HMIM][TF2N]447.921292.782.3890.3893
5[HMIM][TFO]316.341055.602.4950.4890
6[HMIM][BF4]254.08716.611.7940.6589
7[HMIM][MESO4]278.371110.842.9610.4899
8[BMMIM][TF2N]433.401255.802.0310.3193
9[HMIM][PF6]312.24759.161.5500.9385
Table 2. Name and abbreviation of imidazolium ionic liquid.
Table 2. Name and abbreviation of imidazolium ionic liquid.
NO.AbbreviationName
1[BMIM][BF4]1–butyl–3–methylimidazolium tetrafluoroborate
2[EMIM][TF2N]1–ethyl–3–methylimidazolium bis(trifluoromethylsulfonyl)imide
3[EMIM][ETSO4]1–ethyl–3–methylimidazolium ethylsulfate
4[HMIM][TF2N]1–hexyl–3–methylimidazoliumbis(trifluoromethylsulfonyl)imide
5[HMIM][TFO]1–hexyl–3–methylimidazolium trifluoromethanesulfonate
6[HMIM][BF4]1–hexyl–3–methylimidazolium tetrafluoroborate
7[HMIM][MESO4]1–hexyl–3–methylimidazolium methyl–sulfate
8[BMMIM][TF2N]1–butyl–2,3–dimethylimidazoliumbis(trifluoromethanesulfonyl)imide
9[HMIM][PF6]1–methyl–3–hexylimidazolium hexafluorophosphate
Table 3. Solubility data of CO2 in different imidazolium ionic liquids.
Table 3. Solubility data of CO2 in different imidazolium ionic liquids.
NO.Ionic LiquidsTemperature Range
(K)
Pressure Range
(MPa)
CO2 Solubility Range
(Mole Fraction)
NO. of SamplesRefs.
1[BMIM][BF4]278.47–368.220.01–67.620.003–0.610204[8,32,41]
2[EMIM][TF2N]450.49–292.750.00–43.250.000–0.782250[32,35,37,38,41]
3[EMIM][ETSO4]398.04–353.150.10–9.460.000–0.45782[39,40]
4[HMIM][TF2N]278.12–450.490.01–45.280.001–0.824394[8,33,38]
5[HMIM][TFO]303.15–373.151.25–100.120.267–0.81634[33]
6[HMIM][BF4]293.18–373.150.31–86.600.071–0.703160[8,33,37,39]
7[HMIM][MESO4]303.15–373.150.87–50.140.158–0.60248[33,39]
8[BMMIM][TF2N]298.15–343.150.01–1.900.002–0.21136[8,37]
9[HMIM][PF6]298.15–373.150.30–94.600.058–0.727160[8,34,36]
Table 4. Performance index data of back propagation neural network (BPNN) sub–models (Validation set).
Table 4. Performance index data of back propagation neural network (BPNN) sub–models (Validation set).
NO.MAERMSER2
10.00690.01180.9971
20.00710.01260.9967
30.00850.01380.9961
40.00940.01470.9955
50.00540.01050.9977
60.01060.01590.9947
70.01160.01700.9939
80.01310.01860.9928
90.00910.01430.9957
100.00770.01330.9963
Table 5. Performance index data of extreme learning machine (ELM) sub–models (Validation set).
Table 5. Performance index data of extreme learning machine (ELM) sub–models (Validation set).
NO.MAERMSER2
10.01270.01800.9932
20.01230.01720.9938
30.01080.01610.9946
40.00970.01500.9953
50.01370.01890.9926
60.01590.02160.9902
70.01080.01580.9948
80.01270.01820.9930
90.01100.01580.9948
100.00760.01320.9964
Table 6. Performance index data of radial basis function neural network (RBFNN) sub–models (Validation set).
Table 6. Performance index data of radial basis function neural network (RBFNN) sub–models (Validation set).
NO.MAERMSER2
10.00370.00890.9983
20.00420.00940.9982
30.00620.01110.9974
40.00340.00810.9986
50.00350.00760.9988
60.00390.00830.9986
70.00480.00950.9981
80.00470.00960.9981
90.00400.00890.9984
100.00460.01010.9979
Table 7. Calinski–Harabasz (CH) value.
Table 7. Calinski–Harabasz (CH) value.
NO. of ClusterCH Value
217.25
321.44
415.22
512.61
Table 8. Prediction performance data of five models (Testing set).
Table 8. Prediction performance data of five models (Testing set).
ModelMAERMSER2
Optimal BPNN0.00820.01370.9960
Optimal ELM0.00940.01500.9952
Optimal RBFNN0.00660.01180.9971
Fully integrated model0.00550.01030.9978
Selective ensemble model0.00490.00960.9981
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xia, L.; Liu, S.; Pan, H. Prediction of the Solubility of CO2 in Imidazolium Ionic Liquids Based on Selective Ensemble Modeling Method. Processes 2020, 8, 1369. https://0-doi-org.brum.beds.ac.uk/10.3390/pr8111369

AMA Style

Xia L, Liu S, Pan H. Prediction of the Solubility of CO2 in Imidazolium Ionic Liquids Based on Selective Ensemble Modeling Method. Processes. 2020; 8(11):1369. https://0-doi-org.brum.beds.ac.uk/10.3390/pr8111369

Chicago/Turabian Style

Xia, Luyue, Shanshan Liu, and Haitian Pan. 2020. "Prediction of the Solubility of CO2 in Imidazolium Ionic Liquids Based on Selective Ensemble Modeling Method" Processes 8, no. 11: 1369. https://0-doi-org.brum.beds.ac.uk/10.3390/pr8111369

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop