Next Article in Journal
Low-Complexity Joint Range and Doppler FMCW Radar Algorithm Based on Number of Targets
Next Article in Special Issue
Deep Neural Networks for the Classification of Pure and Impure Strawberry Purees
Previous Article in Journal
Estimating the Instantaneous Screw Axis and the Screw Axis Invariant Descriptor of Motion by Means of Inertial Sensors: An Experimental Study with a Mechanical Hinge Joint and Comparison to the Optoelectronic System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Category Tea Polyphenols Evaluation Model Based on Feature Fusion of Electronic Nose and Hyperspectral Imagery

1
School of Information and Computer, Anhui Agricultural University, Hefei 230036, China
2
New Rural Research Institute, Anhui Agricultural University, Hefei 230036, China
3
School of Electrical and Information Engineering, Anhui University of Technology, Ma’anshan 243032, China
4
State Key Laboratory of Tea Plant Biology and Utilization, Anhui Agricultural University, Hefei 230036, China
*
Authors to whom correspondence should be addressed.
Submission received: 14 November 2019 / Revised: 15 December 2019 / Accepted: 18 December 2019 / Published: 20 December 2019

Abstract

:
Tea polyphenols are important ingredients for evaluating tea quality. The rapid development of sensors provides an efficient method for nondestructive detection of tea polyphenols. Previous studies have shown that features obtained from single or multiple sensors yield better results in detecting interior tea quality. However, due to their lack of external features, it is difficult to meet the general evaluation model for the quality of the interior and exterior of tea. In addition, some features do not fully reflect the sensor signals of tea for several categories. Therefore, a feature fusion method based on time and frequency domains from electronic nose (E-nose) and hyperspectral imagery (HSI) is proposed to estimate the polyphenol content of tea for cross-category evaluation. The random forest and the gradient boosting decision tree (GBDT) are used to evaluate the feature importance to obtain the optimized features. Three models based on different features for cross-category tea (black tea, green tea, and yellow tea) were compared, including grid support vector regression (Grid-SVR), random forest (RF), and extreme gradient boosting (XGBoost). The results show that the accuracy of fusion features based on the time and frequency domain from the electronic nose and hyperspectral image system is higher than that of the features from single sensor. Whether based on all original features or optimized features, the performance of XGBoost is the best among the three regression algorithms (R2 = 0.998, RMSE = 0.434). Results indicate that the proposed method in this study can improve the estimation accuracy of tea polyphenol content for cross-category evaluation, which provides a technical basis for predicting other components of tea.

Graphical Abstract

1. Background

Tea polyphenols (TP), as the main biological active ingredient in tea, affects the aroma of tea and the volatility of flavor compounds [1]. Different tea varieties have different polyphenol contents, which is one of the key indicators for assessing tea quality [2] and which affects tea quality control. In addition, tea polyphenols have attracted the attention of scholars at home and abroad because of their various pharmacological effects [3], which makes the quantitative extraction and detection of tea polyphenol content especially important [4]. Therefore, the establishment of an efficient tea polyphenol detection model is of great significance for tea quality improvement and function expansion.
In recent years, many chemical analysis methods have been used to determine the total polyphenol content in tea, such as gas chromatography (GLC), capillary electrophoresis, and high-performance liquid chromatography (HPLC) [5,6,7], which have achieved good results. However, they still have some disadvantages, such as low detection efficiency, high destructiveness, and high detection cost, which cannot meet the requirements of real-time detection of tea quality. Therefore, it is important to find a rapid and nondestructive detection method. In the last ten years, with the development of spectroscopic instruments and data processing technologies, there have been many studies on the detection of phenolic components in tea. However, they mainly focus on the general models of green tea, black tea, and oolong tea to detect tea polyphenols [8,9,10,11,12,13,14]. At present, there is still a lack of a general model to evaluate the quality cross-category tea parameters. Therefore, the nondestructive detection of polyphenols from cross-category teas still faces great challenges.
With the rapid development of sensors, the wide application of electronic nose (E-nose), electronic tongue, and near-infrared technology [15,16,17,18,19,20,21,22] has made tea quality estimation easier. Especially, electronic nose technology has the convenience and objectivity of detecting food taste, which has been successfully applied to many aspects of tea research by simulating the human olfactory system, including in the tea fermentation process [23,24], tea classification [25,26,27,28], tea storage [29], and tea components [30]. However, the function of a single sensor always has certain limitations [31]. Therefore, it is necessary to study the combination of different technologies to capture more comprehensive information. Many studies report that combining multiple sensor signals can improve the results of tea quality estimates. For example, the combination of electronic nose and capillary electrophoresis [32], the combination of electronic nose and visible/near infrared spectroscopy [33], and the combination of electronic nose and electronic tongue technologies [34,35,36]. All of the above studies have achieved good results, regardless of whether they used a single sensor or a multisensor, which obtains signal features or functional group features that can reflect changes in the internal composition of the tea. However, the lack of spatial information limits the in-depth study of tea polyphenols. Actually, different varieties of tea might have different spatial characteristics, even if they come from the same category, not to mention that different categories of tea certainly have obviously different spatial characteristics. Therefore, it is necessary to study the spatial features of tea to make up for the lack of information. To date, hyperspectral imaging technology has been used to improve the evaluation of tea components, including polyphenols [37,38,39], amino acid [40], and catechins [41], due to the advantages of simultaneous acquisition of spatial image information and spectral information of the analyte. However, it remains unclear whether it is possible to improve the estimation model of the polyphenol content of cross-category tea based on the fusion features of different sensors.
In fact, how to effectively fuse multiple features based on E-nose and hyperspectral imagery (HSI) still faces many problems. On the one hand, there is the question of how to extract more meaningful features. On the other hand, there is the question of how to select more representative features. Previous studies have shown that multisource information fusion can more effectively detect the composition of tea [33]. Moreover, the time domain and frequency domain features are more effective in extracting internal quality from the sensor signal array [42]. However, there are still some obstacles to the acquisition of features of cross-category tea. The wide application of wavelet transforms [43,44] provides a new idea for the analysis of tea quality, since this method can extract the time domain and frequency domain features of the signal [45].
Therefore, this study focused on the feasibility of fusion features from multisource sensors, including E-nose and HSI, to estimate the polyphenol content of cross-category tea. To make full use of time domain features and frequency domain features, support vector regression (SVR) [46], random forest (RF) [47], and extreme gradient boosting (XGBoost) [48] are used to construct an estimation model to improve the accuracy of tea polyphenol content of cross-category.
The purpose of this study is to: (1) extract time domain features and frequency domain features from the electronic nose and hyperspectral systems, respectively; (2) fuse time domain features and frequency domain features based on E-nose and HSI to improve estimation models of polyphenol content for cross-category tea; and (3) compare and evaluate polyphenol content estimation models for cross-category tea based on three different regression methods.

2. Data and Methods

2.1. Sample Collection

A total of 110 samples of tea (three categories: yellow tea, black tea, and green tea) were collected from different provinces of China, which were kept in a closed jar and stored in a refrigerator at about 4 °C before test. In order to obtain a wide range of tea polyphenols, different geographical origins of tea samples were collected for the same variety of tea for experimentation.
For the same category of tea, different varieties (geographical origins) of tea samples were collected for experimentation. For example, three brands were collected for black tea, including Zhengshan Xiaozhong tea, Qimen Black tea (Anhui Huangshan and Anhui Qimen, China), and Jinjunmei tea. Therefore, 10 varieties were obtained, as shown in Table 1, 10–15 samples were selected for each variety, and 110 samples were used as research objects.
The tea polyphenol content was directly titrated by potassium permanganate. The tea polyphenol content was calculated according to the Equation (1):
X = ( A B ) × ω × 0.00582 / 0.318 m × V 1 / V 2
where X represents the content (%) of tea polyphenols; A and B represent the number of milliliters of potassium permanganate consumed by the sample and the number of milliliters of potassium permanganate consumed in the blank, respectively; and ω represents the concentration of potassium permanganate (%), m represents the mass (g) of the sample, and V1 and V2 represent the volume (mL) of the test solution and the test solution for measurement, respectively.

2.2. Sample Sampling

2.2.1. HSI Sampling

The hyperspectral image acquisition system used in this experiment included a near-infrared spectrograph (Imspector V17E, Spectral Imaging Ltd., Oulu, Finland), which covers the spectral range of 900–1700 nm, and a charge couple device (CCD)-based digital video camera (IPX-2M30, Imperx Inc., Boca Raton, FL, USA), two 150W halogen lamps (3900, Illumination Technologies Inc., New York, NY, USA), one data acquisition black box, reflective linear tube, electronically controlled displacement platform (MTS120, Beijing Optical Instrument Factory, Beijing, China), and the image acquisition and analysis software (Spectral Image Software, Isuzu Optics Corp., Zhubei, Taiwan). The four tungsten halogen lamps of the reflected light source are evenly distributed on the ring bracket in the dark box, and the light source is irradiated in a direction of 45° with respect to the vertical direction.
The parameters when collecting hyperspectral images are set as follows: exposure time of 20 ms, electronic control stage moving speed of 8 mm·s−1, image resolution of 636 × 815 pixels; spectral resolution of 5 nm, spectral range of 908–1700 nm, spectral sampling interval of 2 nm. Here, 20 g ± 0.5 g from each variety tea sample was evenly spread in a Φ 9 × 1 cm culture dish, and was placed on an electronically controlled stage in a black box to collect hyperspectral images of tea, which were black and white calibrated to remove noise interference and other light source interference.

2.2.2. E-Nose Sampling

The electronic nose (PEN3, Win Muster Air-sense Analytics Inc., Schwerin, Germany) used to collect the scent fingerprint of tea has many functions, including automatic adjustment, automatic calibration, and automatic enrichment, and consists of three units, including a gas sensor array, a signal preprocessing unit, and a pattern recognition unit. The gas sensor array is composed of 10 metal oxide sensors (MOS) (W1C, W5S, W3C, W6S, W5C, W1S, W1W, W2S, W2W, and W3S), which are defined as f0–f9 and are sensitive to different types of volatiles. The characteristics of each sensor are shown in Table 2.
Zero gas was pumped into the cleaning channel to reset the sensors before sampling, and 5 g of each tea sample was placed in a 100 mL glass beaker, sealed and placed for 30 min, and the gas in the headspace bottle was equilibrated and tested. The parameters of the electronic nose were set as follows: sampling interval of 1 s, sensor cleaning of 60 s, sensor return time of 10 s, sampling time of 75 s, and injection flow rate of 600 mL/min.

2.3. Feature Extraction

2.3.1. Feature Extraction from E-Nose System

Feature extraction is used to extract the useful information of a sensor signal by a certain means, so that the discrimination between different types of signals is highlighted and maximized, which is an important part of the model establishment process. The time domain feature mainly measures the change of the signal with time. By comparing the waveform shape of the electrical signal, small differences in odor can be obtained for different samples.
In order to achieve quantitative evaluation of multiple features, parameters including variance, integrals, steady state average, mean differential value, skewness, and kurtosis were selected as time domain features, as shown in Table 3. Variance describes the degree of data dispersion acquired by different sensors, the integral value reflects the total response of the sensor to the gas, the steady state average reflects the characteristic information of the sample, the average differential value reflects the average speed of the sensor’s response to the gas, and skewness and kurtosis reflect the distribution of signals [49,50,51,52].
Here, c i represents the response of the sensor to the second of the sample; c ¯ is the average of the response of the signal; N is the acquisition time of a sample, where N = 75 ; Δ t is the time interval between two adjacent acquisition points ( Δ t = 1 s); t 0 is the time corresponding to when the steady state is ready to be reached.
The frequency domain feature mainly depicts the frequency distribution of the signal. Signal-based frequency domain transformation generally analyzes the energy and entropy values of signals at various frequencies. Wavelet transform is one of the potential technologies for frequency domain information extraction. It can decompose signals into subsignals of different frequencies and effectively express different features. Therefore, the maximum energy and mean of the wavelet transform coefficients are used to represent the main characteristics and overall level of the sensor signal.
Here, X i is the wavelet coefficient of CWT processed by 2 i layers, i represents the different scale factors ( i = 1 , 2 , 3 10 ), the wavelet coefficient energy is the square of the wavelet coefficient at each scale, E i is the wavelet coefficient energy, and S is the sum of the wavelet coefficient energy. W M and W A represent the maximum and mean of the sum of the energy of the wavelet coefficients [53]:
E i = X i 2
S j = i = 1 n E i j   ( j = 1 , 2 , 10   for   E - nose )
W M = m a x | S 1 , S 2 , S j S n | ( j = 1 , 2 , 10   for   E - nose )
W A = 1 n j = 1 n S j ( j = 1 , 2 , 10   for   E - nose )

2.3.2. Feature Extraction from HSI

In image processing, the time domain is the scanning of signals at different times, the processing object is the image itself, and the frequency domain is a coordinate system used to describe the frequency characteristics of a signal. In the frequency domain, the information of the image appears as a combination of different frequency components. Wavelet transform is a time frequency analysis method based on Fourier transform, which can simultaneously represent features including time domain and frequency domain.
Therefore, the wavelet transform is used to obtain the energy ( W E ) and entropy ( W E N ) characteristics of sub-images of different frequencies from the hyperspectral image. In this study, the Daubechies wavelet is used to decompose the hyperspectral image in two layers, W E and W E N , for the image according to Equations (6) and (7) [54]:
W E = 1 M N m = 1 M n = 1 N | I m n Λ | 2
W E N = 1 M N m = 1 M n 1 N I m n Λ ( log 2 ( I m n Λ ) )
where M × N represents the image, ( x , y ) is a pixel, I m n Λ is the wavelet coefficient, and Λ = | LH , HL , HH | .

2.4. Methodology

2.4.1. Normalized Processing

To eliminate the differences between different features to achieve the comparability of multiple indicators, the data needs to be normalized; that is, a dimensionless processing method is required that can reduce the calculation amount and training time. The sensor characteristic data is mapped to 0–1 by the normalization method, and the calculation formula is as shown in the Formula (8).
V = V V min V max V min
where V and V represent the normalized value and original value, V max and V min represent the maximum and the minimum value of the original data.

2.4.2. Support Vector Regression

Support vector regression (SVR) is a linear regression through dimensional transformation. The value of SVR parameters (penalty parameters and kernel parameters) has a great influence on the evaluation performance of SVR. To improve the accuracy of the model, the grid algorithm is used to optimize the selection of SVR parameters [46]. Therefore, the SVR hyperparameter range is shown in Table 4.

2.4.3. Random Forest

Random forest (RF) is an integrated learning method based on bagging, which further improves the level of generalization for the model by selecting random variable sets and random samples from the calibration data set [47], to avoid overfitting caused by excessive eigenvalues and weakening of the influence of outliers on the model. In this study, RF is used to construct the prediction model, and the hyperparameter range of RF is shown in Table 5.

2.4.4. Extreme Gradient Boosting

Extreme gradient boosting (XGBoost) [48] is an integrated algorithm based on lifting trees., which uses the gradient descent architecture to enhance the integrated tree approach of weak learners (typically classification and regression tree). The hyperparameters of the model are optimized with grid search, as shown in Table 6, which is used to build the best model.

2.4.5. Feature Importance Assessment Method

To reduce the number of features and improve the accuracy of the model, feature selection based on the importance of the feature can eliminate irrelevant or redundant features. Therefore, feature selection based on the XGBoost method and correlation coefficient analysis methods are used to detect and eliminate useless features.
In the training process of the XGBoost model, the criterion for dividing each node of the tree is implemented to achieve the optimal features, which indicates the importance of this feature in dividing decision tree nodes. The importance of a feature is the sum of the occurrences of it in all trees. The larger the value, the more important this feature is. Correlation analysis can measure the closeness of the correlation between two characteristic factors [55]. Pearson’s correlation coefficient was used to analyze the correlation between tea polyphenols and the time domain and frequency domain features from HSI. The correlation coefficient value ranges from −1 to 1, and the larger the absolute value, the higher the correlation.
To ensure the representativeness of tea polyphenol content, the data set was divided into a calibration set and validation set according to different varieties of tea. The ratio was about 7:3, and the results are shown in Table 7. The content of tea polyphenols ranged from 10.65% to 29.65% in all samples. The data distribution trend of the calibration set was also consistent with the validation set, indicating that the data distribution of the two data sets was reasonable.
To further verify the universality of the model, the coefficient of determination (R2), adjusted determination coefficient (adjusted_R2) as shown in Formula (9), and the root mean squared error (RMSE) were used as indicators to interpret and quantify the model [56,57].
a d j u s t e d _ R 2 = 1 ( 1 R 2 ) ( n 1 ) n p 1
where n is the number of samples and p is the number of features.

3. Results and Analysis

3.1. Feature Extraction and Feature Selection from E-Nose System

Figure 1 shows the response of the sensor array to Jinjunmei tea. As can be seen from Figure 2, the sensor W1W has a large response to the odor of the tea, and the response value is large. The sensors W5S, W1S, W2W, and W2S are positively deviated by 1, indicating that the response value is enhanced. The sensors W3C, W5C, and W1C deviate by 1 in the negative direction, indicating that the signal changes less. Therefore, different teas have different odors and the sensor signals obtained are different. It can be seen from Figure 1 that the response of the sensor tends to stabilize after 55 s. To facilitate data processing and correct differentiation, multiple features were extracted for each response curve.
Figure 2 shows a characteristic histogram of the response signal of each sensor for a tea sample. It can be seen from the figure that different characteristics reflect different response information for the same sensor, revealing that the gas sensor has broad spectrum responsiveness. However, the same feature also differs for different sensors, which reflects the selectivity of the sensor. Therefore, the sample data pattern generated by the array can be used to express the difference in tea quality and to achieve a one-to-one correspondence between the response pattern and the sample. Additionally, the array can be used to estimate the composition of a tea sample. For a single feature, the electronic nose signal consists of 10 features corresponding to 10 gas sensors. For six time domain features, the electronic nose signal is represented by 10 × 6 features. Therefore, the initial time domain feature matrix from the electronic nose is 110 × 60 features.
The feature importance based on XGBoost is shown in Figure 3, where the vertical axis represents features and the horizontal axis represents the number of times the feature is used to divide the decision tree nodes. As can be seen from Figure 3, the F-score of each feature for the preferred sensor array is different and the difference in performance is large. Therefore, this study ranks the top 30% of sensors according to the importance of the feature, and 3 × 6 features are preferred by XGBoost. For variance, kurtosis, and skewness, f1, f3, and f9 are selected according to importance; f0, f1, and f3 are selected for integral value (INV); f1, f3, and f6 are selected for average differential value (ADV); and f3, f6, and f9 are selected for relative steady-state average value (RSAV). The numbers of decision tree nodes for the above features are all above 400.

3.2. Feature Extraction and Feature Selection from HSI

The 50 × 50 pixel area in the middle of the hyperspectral image was chosen as the region of interest (ROI). The spectral values of all the pixels of the ROI were extracted, and the average value was calculated as the spectral value of the sample. The spectral curves of the 110 tea samples are shown in Figure 4. There is significant noise at both ends of the spectrum. To improve the stability of the model, a total of 457 bands of 944–1688 nm were selected for further analysis. Successive projections algorithm (SPA) [40] was used to extract 1106 and 1375 nm as feature wavelengths for hyperspectral imaging. Time domain and frequency domain features were further extracted for hyperspectral images corresponding to feature wavelengths.
Twenty-four features were extracted from hyperspectral images corresponding to characteristic wavelengths of tea samples (as shown in Figure 5), which describe the features of tea hyperspectral images for different feature parameters (entropy and energy), wavelet decomposition of different layers (two layers), and different directions (three directions). There are different correlations between the features and tea polyphenols.
In this study, Pearson’s correlation analysis was performed between 24 features and tea polyphenol content, as shown in Figure 6. The top 25% features with the absolute value of the correlation coefficient were selected. Therefore, six features that were positively correlated with TP were selected for further modeling, namely B1_HL1_WE, B1_LH1_WE, B1_LH1_WEN, B2_LH1_WE, and B2_LH1_WEN; along with B2_HL2_WE, which was negatively correlated with TP. The absolute values of the correlation coefficient of the above features ranged from 0.4 to 0.59.

3.3. Different Methods for Estimation of Polyphenol Content in Cross-Category Tea

The estimation models of tea polyphenol content were constructed with different features of cross-category tea, including time domain features and frequency domain features acquired from E-nose and HSI; and fusion features, which were compared with the model based on the original features, including the SVR, RF, and XGBoost models, as shown in Table 8. For the calibration set, R2 ranged from 0.839 to 0.977 for Grid-SVR, R2 ranged from 0.758 to 0.876 for RF, and R2 ranged from 0.992 to 0.998 for XGBoost. For the validation set, R2 ranged from 0.703 to 0.816 for Grid-SVR, R2 ranged from 0.722 to 0.854 for RF, and R2 ranged from 0.744 to 0.90 for XGBoost, which indicates that the estimation results based on models of Grid-SVR, RF, and XGBoost perform well.

3.4. Results of Models Based on Different Features

In the model constructed using all the original features, for calibration, the R2 values of the model were 0.759–0.906 for E-nose, R2 ranged from 0.758 to 0.995 for HSI, and R2 was between 0.823 and 0.992 using the fusion features. For the validation set, R2 values of the models were 0.703–0.808, 0.641–0.744, and 0.791–0.871 for E-nose, HSI, and fusion features, respectively.
All of the models were built using the preferred features from E-nose and HSI, including the Grid-SVR, RF, and XGBoost; these models performed better than those based on all features. As shown in Figure 7, especially for validation, R2 for E-nose is between 0.754 and 0.81, R2 for HSI is between 0.694 and 0.747, and R2 for fusion features is between 0.816 and 0.90.

4. Discussion

4.1. Wavelet Transform and Features

Hyperspectral images with rich spectral information and image information contain numerous bands, which lead to excessive data dimensions. In addition, the spectral absorption features of tea polyphenols are mainly caused by the frequency doubling and frequency combination of basic chemical bonds, such as O–H and C–H in the molecules [58]. Therefore, in this study, the feature wavelengths of tea polyphenols were extracted according to the spectral curve, and the time domain features and frequency domain features were extracted according to the hyperspectral image corresponding to the feature wavelengths, which analyzes not only the spectral features, but also spatial features. From the spectral dimension, the hyperspectral images showed the performance of the material under different spectral signals. From spatial dimension analysis, an image is essentially a signal, which is a measure of the intensity variation between various locations of the image, including low frequency signals and high frequency signals. Moreover, the time domain and the frequency domain are the basic properties of the signal [59]. The time domain feature represents the time domain morphology and periodic variation of the signal, and the frequency domain analysis reflects the components of the signal. Therefore, the time domain features and frequency domain features extracted in this study are more representative.
Wavelet transform is the use of an orthogonal wavelet base to decompose the signal into components of different scales, which is equivalent to the partial decomposition of time series signals by using a set of high-pass and low-pass filters. Furthermore, the wavelet transform provides a “time frequency” window that varies with frequency, which obtains the detailed features of the signal by multiscale analysis of the signal stretching [45]. Therefore, regardless of the E-nose or HSI, the features extracted by wavelet transform are representative.

4.2. Different Features Affect Estimation Results

The number of features extracted will have a certain impact on the accuracy of the model [60]. Too few features cannot fully express the useful information of the data, which limits the accuracy of the model; too many features will increase the complexity and computation of the model. In order to obtain a good model estimation effect by applying the appropriate number of features, in this study, the feature selection methods were applied to the reduction of features from multisource sensors. Moreover, the cross-category tea polyphenol content prediction models constructed by feature fusion, such as Grid-SVR, RF, and XGBoost, showed that multisource feature fusion effectively improved the accuracy of the model.
Among them, the accuracy of the three models using the fusion features was higher than that of the individual features. For the Grid-SVR model, the R2 values of the model based on the fusion features were 5.5% and 13.1% higher than the accuracy of the E-Nose and HSI features alone. The R2 values of the RF model were increased by 3.1% and 12.6%, and the R2 values of the XGBoost model were increased by 0.3% and 1.2%, respectively. Therefore, feature fusion can achieve effective data compression, facilitate real-time processing, and provide the information needed for decision analysis.
To verify the universality of the models, adjusted R2 was used to re-evaluate the models, and the results are shown in Table 8. Regardless of whether using Grid-SVR, RF, or XGBoost models, the adjusted R2 values of models based on the fusion features were the highest, which indicates that this is important for feature fusion to improve the accuracy of the model. At the same time, the adjusted R-square value as another evaluation index can eliminate the impact of the number of samples on the accuracy of the model, which further illustrates that the three models established are all valid.

4.3. Different Regression Models Affect Estimation Results

Table 8 shows that the estimated model established by the machine learning method in this paper performs well. Among them, the accuracy of the XGBoost model was higher than the Grid-SVR model or the RF model, which increased by 0.72% and 14.7% with the E-Nose feature, by 13.9% and 22.4% with the HSI feature, and by 2.1% and 12.2% with the fusion feature, respectively. The accuracy of the XGBoost estimation model, whether based on a single feature or a fused feature, was high. In fact, all three methods have their own characteristics. XGBoost and RF are integrated learning methods based on boosting and bagging, respectively. Grid-SVR is an SVR model whose parameters are optimized by grids, and which performs well.
In addition, the methods based on fusion features obtained from E-nose and HSI can provide different methods for predicting the tea polyphenol content of cross-category tea. Wang et al. estimated the polyphenol content of five varieties of green tea [61]. Wang et al. estimated polyphenol contents four Chinese tea, including black tea, dark tea, oolong, and green tea [62]. Although they all achieved good results, they were all based on NIR (near-infrared reflectance) spectroscopy, lacking comprehensive feature fusion. However, the time domain features and frequency domain features from different sensors in this study can provide a new approach for estimating tea composition. The proposed method was used to obtain the expected predictions, despite dealing with different categories, different brand samples, different sensors, and different features. In the future, more varieties of tea should be collected to further improve the universality of the model and provide technical support for nondestructive testing of tea.

5. Conclusions

In this study, an estimation model for polyphenol content in cross-category tea based on feature fusion was proposed, including time and frequency domains, and cross-category tea varieties, such as black tea, green tea, and yellow tea. On the one hand, fusion-based models are superior to models based solely on electronic nose or hyperspectral features, as multisensor feature fusion provides more features, including time and frequency domain features, which reflect gas-sensitive information, spectral and spatial information. The Grid-SVR, RF, and XGBoost models were constructed to estimate the polyphenol content of cross-category teas, with the XGBoost model performing best out of the three models (R2 = 0.998 and adjusted R2 = 0.995 for calibration, R2 = 0.90 and adjusted R2 = 0.75 for validation), which provides a technical basis for quantitative estimation of tea polyphenol content of cross-category tea. In addition, the method proposed in this paper can be used to nondestructively detect the components of other varieties of tea, such as the amino acids and tea polyphenols of white tea, dark tea, and Pu’er tea in future research.

Author Contributions

B.Y. and M.W. designed the algorithms. M.W. and L.Q. performed the experiments and used the software. B.Y. wrote the paper. S.H., H.W., B.W., and J.N. revised the paper. All authors have read and agreed to the published version of the manuscript.

Funding

The Natural Science Foundation of Anhui Province (1808085MF195), the National Key R&D Program (2016YFD0300608), the Natural Science Research Project of Anhui Province (KJ2016A837), and the Open Fund of the Key Laboratory of Technology Integration and Application in Agricultural Internet of Things, the Ministry of Agriculture (2016KL02), the National Natural Science Foundation of China (331772057), and the Science and Technology Major Special Project of Anhui Province (1803071149) partially supported this work.

Conflicts of Interest

All the authors declare no conflict of interest.

References

  1. Alexandr, Y.; Boris, V.; Emilie, C.; Yakov, I. Determination of the chemical composition of tea by chromatographic methods: A review. J. Food Res. 2015, 4, 56–88. [Google Scholar]
  2. Juneja, L.; Chu, D.; Okubo, T.; Nagato, Y.; Yokogoshi, H. L-theanine—A unique amino acid of green tea and its relaxation effect in humans. Trends Food Sci. Technol. 1999, 10, 199–204. [Google Scholar] [CrossRef]
  3. Qian, Y.; Zhang, S.; Yao, S.; Xia, J.; Li, Y.; Dai, X.; Wang, W.; Jiang, X.; Liu, Y.; Li, M.; et al. Effects of vitro sucrose on quality components of tea plants (Camellia sinensis) based on transcriptomic and metabolic analysis. BMC Plant Biol. 2018, 18, 121. [Google Scholar] [CrossRef] [PubMed]
  4. Jia, W.; Liang, G.; Jiang, Z.; Wang, J. Advances in Electronic Nose Development for Application to Agricultural Products. Food Anal. Meth. 2019, 12, 2226–2240. [Google Scholar] [CrossRef]
  5. Nishitani, E.; Sagesaka, Y.M. Simultaneous determination of catechins, caffeine and other phenolic compounds in tea using new HPLC method. J. Food Compos. Anal. 2004, 17, 675–685. [Google Scholar] [CrossRef]
  6. Togari, N.; Kobayashi, A.; Aishima, T. Pattern recognition applied to gas chromatographic profiles of volatile components in three tea categories. Food Res. Int. 1995, 28, 495–502. [Google Scholar] [CrossRef]
  7. Kotani, A.; Takahashi, K.; Hakamata, H.; Kojima, S.; Kusu, F. Attomole catechins determination by capillary liquid chromatography with electrochemical detection. Anal. Sci. 2007, 23, 157–163. [Google Scholar] [CrossRef] [Green Version]
  8. Lee, M.; Hwang, Y.; Lee, J.; Choung, M. The characterization of caffeine and nine individual catechins in the leaves of green tea (Camellia sinensis L.) by near-infrared reflectance spectroscopy. Food Chem. 2014, 158, 351–357. [Google Scholar] [CrossRef]
  9. Ren, G.; Wang, S.; Ning, J.; Xu, R.; Wang, Y.; Xing, Z.; Wan, X.; Zhang, Z. Quantitative analysis and geographical traceability of black tea using Fourier transform near-infrared spectroscopy (FT-NIRS). Food Res. Int. 2013, 53, 822–826. [Google Scholar] [CrossRef]
  10. Chen, G.; Yuan, Q.; Saeeduddin, M.; Ou, S.; Zeng, X.; Ye, H. Recent advances in tea polysaccharides: Extraction, purification, physicochemical characterization and bioactivities. Carbohydr. Polym. 2016, 153, 663–678. [Google Scholar] [CrossRef]
  11. Sun, H.; Chen, Y.; Cheng, M.; Zhang, X.; Zheng, X.; Zhang, Z. The modulatory effect of polyphenols from green tea, oolong tea and black tea on human intestinal microbiota in vitro. J. Food Sci. Technol. 2018, 55, 399–407. [Google Scholar] [CrossRef] [PubMed]
  12. Hocker, N.; Wang, C.; Prochotsky, J.; Eppurath, A.; Rudd, L.; Perera, M. Quantification of antioxidant properties in popular leaf and bottled tea by high-performance liquid chromatography (HPLC), spectrophotometry, and voltammetry. Anal. Lett. 2017, 50, 1640–1656. [Google Scholar] [CrossRef]
  13. Dutta, D.; Das, P.; Bhunia, U.; Singh, U.; Singh, S.; Sharma, J.; Dadhwal, V. Retrieval of tea polyphenol at leaf level using spectral transformation and multi-variate statistical approach. Int. J. Appl. Earth Obs. Geoinf. 2015, 36, 22–29. [Google Scholar] [CrossRef]
  14. Pan, H.; Zhang, D.; Li, B.; Wu, Y.; Tu, Y. A rapid UPLC method for simultaneous analysis of caffeine and 13 index polyphenols in black tea. J. Chromatogr. Sci. 2017, 55, 495–496. [Google Scholar] [CrossRef]
  15. Sun, Y.; Wang, J.; Cheng, S.; Wang, Y. Detection of pest species with different ratios in tea plant based on electronic nose. Ann. Appl. Biol. 2019, 174, 209–218. [Google Scholar] [CrossRef]
  16. Yang, X.; Liu, Y.; Mu, L.; Wang, W.; Zhan, Q.; Luo, M.; Li, J. Discriminant research for identifying aromas of non-fermented Pu-erh tea from different storage years using an electronic nose. J. Food Process Preserv. 2018, 42, e13721. [Google Scholar] [CrossRef]
  17. Zhi, R.; Zhao, L.; Zhang, D. A framework for the multi-level fusion of electronic nose and electronic tongue for tea quality assessment. Sensors 2017, 17, 1007. [Google Scholar] [CrossRef] [Green Version]
  18. Jin, J.; Deng, S.; Ying, X.; Ye, X.; Lu, T.; Hui, G. Study of herbal tea beverage discrimination method using electronic nose. J. Food Meas. Charact. 2015, 9, 52–60. [Google Scholar] [CrossRef]
  19. Peng, W.; Wang, L.; Qian, Y.; Chen, T.; Dai, B.; Feng, B.; Wang, B. Discrimination of Unfermented Pu’er Tea Aroma of Different Years Based on Electronic Nose. Agric. Res. 2017, 6, 436–442. [Google Scholar] [CrossRef]
  20. Lelono, D.; Triyana, K.; Hartati, S.; Istiyanto, J. Classification of Indonesia black teas based on quality by using electronic nose and principal component analysis. In Proceedings of the 1st International Conference on Science and Technology, Advances of Science and Technology for Society, Yogyakarta, Indonesia, 21 July 2016; p. 020003. [Google Scholar]
  21. Sarkar, S.; Bhondekar, A.; Macaš, M.; Kumar, R.; Kaur, R.; Sharma, A.; Kumar, A. Towards biological plausibility of electronic noses: A spiking neural network based approach for tea odour classification. Neural Netw. 2015, 71, 142–149. [Google Scholar] [CrossRef]
  22. Tudu, B.; Jana, A.; Metla, A.; Ghosh, D.; Bhattacharyya, N.; Bandyopadhyay, R. Electronic nose for black tea quality evaluation by an incremental RBF network. Sens. Actuator B Chem. 2009, 138, 90–95. [Google Scholar] [CrossRef]
  23. Bhattacharya, N.; Tudu, B.; Jana, A.; Ghosh, D.; Bandhopadhyaya, R.; Bhuyan, M. Preemptive identification of optimum fermentation time for black tea using electronic nose. Sens. Actuator B Chem. 2018, 131, 110–116. [Google Scholar] [CrossRef]
  24. Ghosh, S.; Tudu, B.; Bhattacharyya, N.; Bandyopadhyay, R. A recurrent Elman network in conjunction with an electronic nose for fast prediction of optimum fermentation time of black tea. Neural Comput. Appl. 2017, 31, 1165–1171. [Google Scholar] [CrossRef]
  25. Chen, Q.; Zhao, J.; Chen, Z.; Lin, H.; Zhao, D. Discrimination of green tea quality using the electronic nose technique and the human panel test, comparison of linear and nonlinear classification tools. Sens. Actuator B Chem. 2011, 159, 294–300. [Google Scholar] [CrossRef]
  26. Kaur, R.; Kumar, R.; Gulati, A.; Ghanshyam, C.; Kapur, P.; Bhondekar, A. Enhancing electronic nose performance: A novel feature selection approach using dynamic social impact theory and moving window time slicing for classification of Kangra orthodox black tea (Camellia sinensis (L.) O. Kuntze). Sens. Actuator B Chem. 2012, 166, 309–319. [Google Scholar] [CrossRef]
  27. Yu, H.; Wang, J.; Zhang, H.; Yu, Y.; Yao, C. Identification of green tea grade using different feature of response signal from E-nose sensors. Sens. Actuator B Chem. 2008, 128, 455–461. [Google Scholar] [CrossRef]
  28. Dutta, R.; Hines, E.; Gardner, J.; Kashwan, K.; Bhuyan, M. Tea quality prediction using a tin oxide-based electronic nose: An artificial intelligence approach. Sens. Actuator B Chem. 2003, 94, 228–237. [Google Scholar] [CrossRef]
  29. Yu, H.; Wang, Y.; Wang, J. Identification of tea storage times by linear discrimination analysis and back-propagation neural network techniques based on the eigenvalues of principal components analysis of E-nose. Sens. Actuator B Chem. 2009, 9, 8073–8082. [Google Scholar] [CrossRef] [Green Version]
  30. Ghosh, D.; Gulati, A.; Joshi, R.; Bhattacharyya, N.; Bandyopadhyay, R. Estimation of Aroma Determining Compounds of Kangra Valley Tea by Electronic Nose System. In Proceedings of the Perception and Machine Intelligence-First Indo-Japan Conference, PerMIn 2012, Kolkata, India, 12–13 January 2012; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  31. Li, Z.; Thomas, C. Quantitative evaluation of mechanical damage to fresh fruits. Trends Food Sci. Technol. 2014, 35, 138–150. [Google Scholar] [CrossRef]
  32. Mirasoli, M.; Gotti, R.; Di Fusco, M.; Leoni, A.; Colliva, C.; Roda, A. Electronic nose and chiral-capillary electrophoresis in evaluation of the quality changes in commercial green tea leaves during a long-term storage. Talanta 2014, 129, 32–38. [Google Scholar] [CrossRef]
  33. Xu, S.; Sun, X.; Lu, H.; Zhang, Q. Detection of Type, Blended Ratio, and Mixed Ratio of Pu’er Tea by Using Electronic Nose and Visible/Near Infrared Spectrometer. Sensors 2019, 19, 2359. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Zou, G.; Xiao, Y.; Wang, M.; Zhang, H. Detection of bitterness and astringency of green tea with different taste by electronic nose and tongue. PLoS ONE 2018, 13, e0206517. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Banerjee, R.; Modak, A.; Mondal, S.; Tudu, B.; Bandyopadhyay, R.; Bhattacharyya, N. Fusion of electronic nose and tongue response using fuzzy based approach for black tea classification. Procedia Technol. 2013, 10, 615–622. [Google Scholar]
  36. Banerjee, R.; Chattopadhyay, P.; Tudu, B.; Bhattacharyya, N.; Bandyopadhyay, R. Artificial flavor perception of black tea using fusion of electronic nose and tongue response: A bayesian statistical approach. J. Food Eng. 2014, 142, 87–93. [Google Scholar] [CrossRef]
  37. Djokam, M.; Sandasi, M.; Chen, W.; Viljoen, A.; Vermaak, I. Hyperspectral imaging as a rapid quality control method for herbal tea blends. Appl. Sci. 2017, 7, 268. [Google Scholar] [CrossRef] [Green Version]
  38. Bian, M.; Skidmore, A.; Schlerf, M.; Liu, Y.; Wang, T. Estimating biochemical parameters of tea (Camellia sinensis (L.)) using hyperspectral techniques. ISPRS J. Photogramm. Remote Sens. 2012, 39, B8. [Google Scholar] [CrossRef] [Green Version]
  39. Tu, Y.; Bian, M.; Wan, Y.; Fei, T. Tea cultivar classification and biochemical parameter estimation from hyperspectral imagery obtained by UAV. Peer J. 2018, 6, e4858. [Google Scholar] [CrossRef]
  40. Yang, B.; Gao, Y.; Li, H.; Ye, S.; He, H.; Xie, S. Rapid prediction of yellow tea free amino acids with hyperspectral images. PLoS ONE 2019, 14, e0210084. [Google Scholar] [CrossRef] [Green Version]
  41. Ryu, C.; Suguri, M.; Park, S.; Mikio, M. Estimating catechin concentrations of new shoots in the green tea field using ground-based hyperspectral image. In Proceedings of the SPIE 8887, Remote Sensing for Agriculture, Ecosystems, and Hydrology XV, 88871Q, Dresden, Germany, 16 October 2013. [Google Scholar]
  42. Peng, C.; Yan, J.; Duan, S.; Wang, L.; Jia, P.; Zhang, S. Enhancing electronic nose performance based on a novel QPSO-KELM model. Sensors 2016, 16, 520. [Google Scholar] [CrossRef] [Green Version]
  43. Bruce, L.M.; Li, J.; Huang, Y. Automated detection of subpixel hyperspectral targets with adaptive multichannel discrete wavelet transform. IEEE Trans. Geosci. Remote Sens. 2002, 40, 977–980. [Google Scholar] [CrossRef]
  44. Cheng, T.; Rivard, B.; Sanchez-Azofeifa, A. Spectroscopic determination of leaf water content using continuous wavelet analysis. Remote Sens. Environ. 2011, 115, 659–670. [Google Scholar] [CrossRef]
  45. Coffey, M.A.; Etter, D.M. Image coding with the wavelet transform. In Proceedings of the ISCAS’95-International Symposium on Circuits and Systems, Seattle, WA, USA, 30 April–3 May 1995. [Google Scholar]
  46. Atzberger, C.; Guerif, M.; Baret, F.; Werner, W. Comparative analysis of three chemometric techniques for the spectroradiometric assessment of canopy chlorophyll content in winter wheat. Comput. Electron. Agric. 2010, 73, 165–173. [Google Scholar] [CrossRef]
  47. Leo, B. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar]
  48. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, New York, NY, USA, 13–17 August 2016; pp. 758–794. [Google Scholar]
  49. Yin, Y.; Chu, B.; Yu, H.; Xiao, Y. A selection method for feature vectors of electronic nose signal based on wilks Λ–statistic. J. Food Meas. Charact. 2014, 8, 29–35. [Google Scholar] [CrossRef]
  50. Yin, Y.; Yu, H.; Bing, C. A sensor array optimization method of electronic nose based on elimination transform of Wilks statistic for discrimination of three kinds of vinegars. J. Food Eng. 2014, 127, 43–48. [Google Scholar] [CrossRef]
  51. Yin, Y.; Zhao, Y. A feature selection strategy of E-nose data based on PCA coupled with Wilks Λ-statistic for discrimination of vinegar samples. J. Food Meas. Charact. 2019, 13, 2406–2416. [Google Scholar] [CrossRef]
  52. Zhang, S.; Xia, X.; Xie, C.; Cai, S.; Li, H.; Zeng, D. A method of feature extraction on recovery curves for fast recognition application with metal oxide gas sensor array. IEEE Sens. J. 2009, 9, 1705–1710. [Google Scholar] [CrossRef]
  53. Rosso, O.A.; Blanco, S.; Yordanova, J.; Kolev, V.; Figliola, A.; Schürmann, M.; Başar, E. Wavelet entropy: A new tool for analysis of short duration brain electrical signals. J. Neurosci. Methods 2001, 105, 65–75. [Google Scholar] [CrossRef]
  54. Gai, S. Efficient Color Texture Classification Using Color Monogenic Wavelet Transform. Neural Process. Lett. 2017, 46, 609–626. [Google Scholar] [CrossRef]
  55. Ahlgren, P.; Jarneving, B.; Rousseau, R. Requirements for a cocitation similarity measure, with special reference to Pearson’s correlation coefficient. J. Am. Soc. Inf. Sci. Technol. 2003, 54, 550–560. [Google Scholar] [CrossRef]
  56. Yang, B.; Wang, M.; Sha, Z.; Wang, B.; Chen, J.; Yao, X.; Cheng, T.; Cao, W.; Zhu, Y. Evaluation of Aboveground Nitrogen Content of Winter Wheat Using Digital Imagery of Unmanned Aerial Vehicles. Sensors 2019, 19, 4416. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Liao, J.G.; McGee, D. Adjusted coefficients of determination for logistic regression. Am. Stat. 2003, 57, 161–165. [Google Scholar] [CrossRef]
  58. Cen, H.; He, Y. Theory and application of near infrared reflectance spectroscopy in determination of food quality. Trends Food Sci. Technol. 2007, 18(2), 72–83. [Google Scholar] [CrossRef]
  59. Brandt, A. A signal processing framework for operational modal analysis in time and frequency domain. Mech. Syst. Signal Process. 2019, 115, 380–393. [Google Scholar] [CrossRef]
  60. Zhang, Y.; Ji, X.; Liu, B.; Huang, D.; Xie, F.; Zhang, Y. Combined feature extraction method for classification of EEG signals. Neural Comput. Appl. 2017, 28, 3153–3161. [Google Scholar] [CrossRef]
  61. Wang, X.; Huang, J.; Fan, W.; Lu, H. Identification of green tea varieties and fast quantification of total polyphenols by near-infrared spectroscopy and ultraviolet-visible spectroscopy with chemometric algorithms. Anal. Methods 2015, 7, 787–792. [Google Scholar] [CrossRef]
  62. Wang, J.; Wang, Y.; Cheng, J.; Wang, J.; Sun, X.; Sun, S.; Zhang, Z. Enhanced cross-category models for predicting the total polyphenols, caffeine and free amino acids contents in Chinese tea using NIR spectroscopy. LWT Food Sci. Technol. 2018, 96, 90–97. [Google Scholar] [CrossRef]
Figure 1. Sensor signal intensity of Jinjunmei tea.
Figure 1. Sensor signal intensity of Jinjunmei tea.
Sensors 20 00050 g001
Figure 2. Bar results of six kinds of features of each gas sensor for one tea sample.
Figure 2. Bar results of six kinds of features of each gas sensor for one tea sample.
Sensors 20 00050 g002
Figure 3. Feature importance: (a) variance value; (b) kurtosis coefficient; (c) integral value; (d) coefficient of skewness; (e) average differential value; (f) relative steady-state average value.
Figure 3. Feature importance: (a) variance value; (b) kurtosis coefficient; (c) integral value; (d) coefficient of skewness; (e) average differential value; (f) relative steady-state average value.
Sensors 20 00050 g003
Figure 4. Near-infrared hyperspectral original curve of three categories of tea.
Figure 4. Near-infrared hyperspectral original curve of three categories of tea.
Sensors 20 00050 g004
Figure 5. Original images and hyperspectral images based on characteristic wavelengths for three categories of tea. (A1): Zhengshan Xiaozhong ( Fujian); (A2): Qimen Black tea (Anhui Huangshan); (A3): Qimen Black tea ( Anhui Qimen); (A4): JinJunMei (Fujian); (B1): Huangshan Maofeng (Anhui); (B2): Liuan Guapian (Anhui); (C1): Junshan Yinzhen (Hunan); (C2): Huoshan Huangya tea (Anhui); (C3): Mengding Huangya tea (Sichuan) and (C4): Pingyang Huangtang tea (Zhejiang) represent the original images. A1_1106, A2_1106, A3_1106, A4_1106, B1_1106, B2_1106, C1_1106, C2_1106, C3_1106 and C4_1106 represent the 1106 nm hyperspectral images corresponding to the above original images; A1_1375, A2_ 1375, A3_1375, A4_ 1375, B1_1375, B2_1375, C1_1375, C2_1375, C3_1375 and C4_1375 represent the 1375 nm hyperspectral images corresponding to the above original images.)
Figure 5. Original images and hyperspectral images based on characteristic wavelengths for three categories of tea. (A1): Zhengshan Xiaozhong ( Fujian); (A2): Qimen Black tea (Anhui Huangshan); (A3): Qimen Black tea ( Anhui Qimen); (A4): JinJunMei (Fujian); (B1): Huangshan Maofeng (Anhui); (B2): Liuan Guapian (Anhui); (C1): Junshan Yinzhen (Hunan); (C2): Huoshan Huangya tea (Anhui); (C3): Mengding Huangya tea (Sichuan) and (C4): Pingyang Huangtang tea (Zhejiang) represent the original images. A1_1106, A2_1106, A3_1106, A4_1106, B1_1106, B2_1106, C1_1106, C2_1106, C3_1106 and C4_1106 represent the 1106 nm hyperspectral images corresponding to the above original images; A1_1375, A2_ 1375, A3_1375, A4_ 1375, B1_1375, B2_1375, C1_1375, C2_1375, C3_1375 and C4_1375 represent the 1375 nm hyperspectral images corresponding to the above original images.)
Sensors 20 00050 g005
Figure 6. Correlation between tea polyphenols and spatial features of hyperspectral images.
Figure 6. Correlation between tea polyphenols and spatial features of hyperspectral images.
Sensors 20 00050 g006
Figure 7. Estimation of polyphenol content of cross-category tea based on different features. Top: Grid-SVR model; middle: RF model; bottom: XGBoost model with original features (a,c,e) and preferred features (b,d,f).
Figure 7. Estimation of polyphenol content of cross-category tea based on different features. Top: Grid-SVR model; middle: RF model; bottom: XGBoost model with original features (a,c,e) and preferred features (b,d,f).
Sensors 20 00050 g007
Table 1. Geographical sources and descriptive statistics of tea polyphenol content (%).
Table 1. Geographical sources and descriptive statistics of tea polyphenol content (%).
Tea CategoryTea Variety (Geographical Origins)NumberRange (%)Mean ± SD (%)
Black teaZhengshan Xiaozhong (Fujian)1010.65–13.2111.832 ± 0.850
Qimen Black Tea (Anhui Huangshan)1013.66–16.5415.196 ± 0.810
Qimen Black Tea (Anhui Qimen)1016.51–22.8518.789 ± 1.567
JinJunMei (Fujian)1012.62–19.1616.99 ± 1.788
Green tea Huangshan Maofeng (Anhui)1525.32–29.4126.485 ± 1.195
Liuan Guapian (Anhui)1527.42–29.6528.535 ± 0.632
Yellow teaJunshan Yinzhen (Hunan)1014.55–19.616.34 ± 1.863
Huoshan Huangya Tea (Anhui)1011.88–16.6513.862 ± 1.367
Mengding Huangya Tea (Sichuan)1011.36–16.3514.334 ± 1.738
Pingyang Huangtang Tea (Zhejiang)1013.34–19.3616.735 ± 2.231
Table 2. Performance description of sensors for the PEN3 electronic nose (E-nose).
Table 2. Performance description of sensors for the PEN3 electronic nose (E-nose).
Array NumberSensor NameObject Substances of SensingComponentThreshold Value/(mL m−3)
f0W1CAromaticsC6H5CH310
f1W5SNitrogen oxidesNO21
f2W3CAmmonia and aromatic moleculesC6H610
f3W6SHydrogenH2100
f4W5CMethane, propane and aliphatic nonpolar moleculesC3H81
f5W1SBroad methaneCH4100
f6W1WSulfur-containing organicsH2S1
f7W2SBroad alcoholsCO100
f8W2WAromatics, sulfur-, and chlorine-containing organicsH2S1
f9W3SMethane and aliphaticsCH410
Table 3. Feature extraction from electronic nose signals.
Table 3. Feature extraction from electronic nose signals.
IndicesNameFormula
VARVariance value V A R = 1 N i = 1 N ( c i c ¯ ) 2
INVIntegral value I N V = c ( i ) Δ t
RSAVRelative steady-state average value R S A V = 1 N t 0 t o T c i
ADVAverage differential value A D V = 1 N 1 i = 1 N 1 c i + 1 c i Δ t
KURTKurtosis coefficient K U R T = 1 N i = 1 N ( c i c ¯ ) 4 ( 1 N i = 1 N ( ( c i c ¯ ) 2 ) 2 3 = m 4 m 2 2 3
SKEWCoefficient of skewness S K E W = N ( N 1 ) N 2 1 N i = 1 N ( c i c ¯ ) 3 ( 1 N i = 1 N ( ( c i c ¯ ) 2 ) 3 2
Table 4. The support vector regression (SVR) modeling hyperparameters.
Table 4. The support vector regression (SVR) modeling hyperparameters.
ParameterRangeOptimum Value
c0 to 2015
g0 to 105
s0 to 103
p0.001 to 10.01
Table 5. The random forest (RF) regression modeling hyperparameters.
Table 5. The random forest (RF) regression modeling hyperparameters.
ParameterRangeOptimum Value
n_estimators100 to 20001000
max_depth1 to 103
extra_ options. importance[0, 1]1
extra_ options. nPerm[0, 1]1
bootstrap[True, False]FALSE
Table 6. Extreme gradient boosting (XGBoost) regression modeling hyperparameters.
Table 6. Extreme gradient boosting (XGBoost) regression modeling hyperparameters.
ParameterRangeOptimum Value
learning_rate0.1 to 10.1
n_estimators100 to 1000400
max_depth1 to 105
gamma0.1 to 10.1
subsample0.1 to 10.9
min_child_weight3 to 105
Table 7. Descriptive statistics of tea polyphenol (TP) contents in the calibration and prediction sets.
Table 7. Descriptive statistics of tea polyphenol (TP) contents in the calibration and prediction sets.
Data SetNumberContent RangeMeanSD
Full11010.65–29.6518.785.82
Calibration set8010.65–29.6518.615.94
Validation set3011.31–29.1319.215.47
Table 8. Estimation models for polyphenol content of cross-category tea based on different variables with three techniques. Note: Grid-SVR = grid support vector regression; RF = random forest; XGBoost = extreme gradient boosting; RMSE = root mean square error.
Table 8. Estimation models for polyphenol content of cross-category tea based on different variables with three techniques. Note: Grid-SVR = grid support vector regression; RF = random forest; XGBoost = extreme gradient boosting; RMSE = root mean square error.
ModelFeaturesVariablesCalibrationValidation
NumberR2Adjusted_R2RMSER2Adjusted_ R2RMSE
Grid-SVRE-Nose200.9230.8191.6590.7540.4722.852
HSI60.8490.7052.3130.6940.4513.225
Fusion260.9770.9400.9060.8160.5612.856
RFE-Nose200.8480.6562.3180.7960.5512.637
HSI60.7650.5612.8810.7240.4962.982
Fusion260.8760.6952.0940.8540.6452.287
XGBoostE-Nose200.9950.9880.2740.8100.5792.422
HSI60.9870.9730.7050.7470.5323.099
Fusion260.9980.9950.4340.9000.7501.895

Share and Cite

MDPI and ACS Style

Yang, B.; Qi, L.; Wang, M.; Hussain, S.; Wang, H.; Wang, B.; Ning, J. Cross-Category Tea Polyphenols Evaluation Model Based on Feature Fusion of Electronic Nose and Hyperspectral Imagery. Sensors 2020, 20, 50. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010050

AMA Style

Yang B, Qi L, Wang M, Hussain S, Wang H, Wang B, Ning J. Cross-Category Tea Polyphenols Evaluation Model Based on Feature Fusion of Electronic Nose and Hyperspectral Imagery. Sensors. 2020; 20(1):50. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010050

Chicago/Turabian Style

Yang, Baohua, Lin Qi, Mengxuan Wang, Saddam Hussain, Huabin Wang, Bing Wang, and Jingming Ning. 2020. "Cross-Category Tea Polyphenols Evaluation Model Based on Feature Fusion of Electronic Nose and Hyperspectral Imagery" Sensors 20, no. 1: 50. https://0-doi-org.brum.beds.ac.uk/10.3390/s20010050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop