Next Article in Journal
Find Outliers of Image Edge Consistency by Weighted Local Linear Regression with Equality Constraints
Previous Article in Journal
Indirect Temperature Measurement in High Frequency Heating Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of the Gaussian Mixture Model to Classify Stages of Electrical Tree Growth in Epoxy Resin

by
Abdullahi Abubakar Mas’ud
1,2,*,
Arunachalam Sundaram
1,
Jorge Alfredo Ardila-Rey
3,
Roger Schurch
3,
Firdaus Muhammad-Sukki
4 and
Nurul Aini Bani
5
1
Department of Electrical and Electronic Engineering Technology, Jubail Industrial College, Al Jubail 35718, Saudi Arabia
2
Prince Saud bin Thunayan Research Centre, Royal Commission for Jubail, Al Jubail 35718, Saudi Arabia
3
Department of Electrical Engineering, Universidad Técnica Federico Santa María, Santiago de Chile 8940000, Chile
4
School of Engineering & the Built Environment, Edinburgh Napier University, Edinburgh EH10 5DT, Scotland, UK
5
Razak Faculty of Technology and Informatics, Universiti Teknologi Malaysia, Jalan Sultan Yahya Petra, Kuala Lumpur 54100, Malaysia
*
Author to whom correspondence should be addressed.
Submission received: 16 March 2021 / Revised: 31 March 2021 / Accepted: 2 April 2021 / Published: 6 April 2021
(This article belongs to the Section Sensor Materials)

Abstract

:
In high-voltage (HV) insulation, electrical trees are an important degradation phenomenon strongly linked to partial discharge (PD) activity. Their initiation and development have attracted the attention of the research community and better understanding and characterization of the phenomenon are needed. They are very damaging and develop through the insulation material forming a discharge conduction path. Therefore, it is important to adequately measure and characterize tree growth before it can lead to complete failure of the system. In this paper, the Gaussian mixture model (GMM) has been applied to cluster and classify the different growth stages of electrical trees in epoxy resin insulation. First, tree growth experiments were conducted, and PD data captured from the initial to breakdown stage of the tree growth in epoxy resin insulation. Second, the GMM was applied to categorize the different electrical tree stages into clusters. The results show that PD dynamics vary with different stress voltages and tree growth stages. The electrical tree patterns with shorter breakdown times had identical clusters throughout the degradation stages. The breakdown time can be a key factor in determining the degradation levels of PD patterns emanating from trees in epoxy resin. This is important in order to determine the severity of electrical treeing degradation, and, therefore, to perform efficient asset management. The novelty of the work presented in this paper is that for the first time the GMM has been applied for electrical tree growth classification and the optimal values for the hyperparameters, i.e., the number of clusters and the appropriate covariance structure, have been determined for the different electrical tree clusters.

1. Introduction

Electrical treeing is a key degradation phenomenon of high-voltage polymeric insulation [1]. When electrical trees are initiated, they grow until they bridge the entire insulation material, resulting in catastrophic failure of the power system plant. Electrical trees are strongly related to partial discharge (PD) activity, which is usually characterized using techniques such as phase-resolved PD (PRPD) patterns [2,3], and, to a lesser extent, pulse sequence analysis (PSA) [4], pulse waveform analysis [5], and nonlinear time series analysis [6,7].
When PD activity is severe, there is higher dissipation of energy and greater PD amplitudes, resulting in tree growth, and serious degradation [8,9]. As mentioned in the literature [10], PD activity might be undetectable when the tree structure is forming conductive channels within the insulation system, therefore, thorough knowledge of PD behavior related to treeing degradation is needed, especially for conditioning monitoring engineers in the industry. Understanding the tree phenomenon is crucial in determining the remaining lifetime of an electrical asset.
Several studies analyze tree growth from PD activity [10,11,12,13,14]. In particular, Lv et al. [10], Bao et al. [11], Zhou et al. [14], and Alapati [13] investigated PD development during the early stages of PD degradation in epoxy resin insulation, cross-linked polyethylene (XLPE) cable, and low-density polyethylene (LDPE). These studies found that the growth rate of electric trees is strongly influenced by the increasing level of voltages. However, for electrical tree propagation in an XLPE cable, the skewness (i.e., the extent to which the distribution deviates from the normal distribution) of the maximum amplitude-phase distributions decreased with the spread of the electrical stress, and the skewness can be considered as a parameter to fully identify different levels of electrical tree propagation. Furthermore, the tree pattern feature in cross-linked polyethylene (XLPE) cable is similar to the needle-plate electrode system.
In the case of the LDPE, there was a considerable decrease in PD repetition rate and PD magnitude of LDPE filled with alumina nanocomposites compared to unfilled LDPE. There was an increase in the PD inception voltage with 3% weight of the filler loadings and then decreases when the filler loadings reached 5%. On the other hand, other researchers investigated tree propagation mechanisms in XLPE cable insulation based on a double electrical tree structure [12]. It was found that five types of electrical tree structures (branch, forest, bine-branch, pine-branch, and mixed configurations) will propagate in XLPE cable insulation due to the effect of the irregular congregating state, differences in the crystalline structure, and the presence of residual stress in the semi-crystalline polymer.
Few investigations are found in the area of pattern identification of PD characteristics from the initial stage to breakdown of the insulation due to electrical treeing. Park et al. [15] evaluated and classified PD degradation of electric trees for cable insulation. The authors of this study utilized three classification techniques: Adaptive neuro-fuzzy combination (ANFIS), multi-layer perceptron (MLP), and principal component analysis (PCA). Compared to other defects, such as voids and metal surfaces, the findings specifically demonstrated different features of electric trees. The results clearly showed distinct characteristics of electrical trees as compared to that of other defects such as voids and metal surfaces. Among all the classification techniques, ANFIS showed higher identification potential and can be used for classifying electrical tree progression with about a 99% recognition rate. In another investigation, Salama et al. [16] applied a MLP neural network (NN) to discriminate between PD defects in voids and electrical trees. In particular, the algorithm could recognize discharge patterns from different degradation levels of the electrical trees. Although this algorithm was applied for the case of real power cable faults, it has not been applied to electrical tree faults degradation up to the breakdown stage. In another study, Park et al. [17] attempted to recognize three different electrical tree models using the adaptive network-based fuzzy inference system. The models considered samples with a needle-plane electrode, needle-void-plane electrode, and needle-metal strip-plane electrode. Statistical features extracted from the aforementioned tree models were applied as inputs to the ANFIS system. The results showed a good discrimination rate of these models up to 100%. However, this work was limited to electrical tree patterns only, without analyzing the progression and different stages of tree growth.
This paper studies the growth of electric trees in samples of epoxy resin under various voltage levels aiming to correlate PD activity with the stage of tree growth through the analysis of PRPD patterns using a Gaussian mixture-based model (GMM) clustering technique. GMM was chosen over other techniques because it is flexible and can perform hard clustering for complex data. Using this approach, it is expected to assess the remaining life of insulation subjected to electrical treeing degradation more accurately.
Section 2 details the experimental setup and the data capture procedure, describing the dataset, and their analysis. Section 3 explains the GMM model used in this work. Section 4 is the data processing technique adopted, while Section 5 describes this work’s results and conclusion.

2. Experiment and Data Capture Procedure

2.1. Experimental Setup

The samples were prepared using the conventional needle-to-plane configuration with a gap distance of ~2 mm between the needle tip and the bottom of the sample. The needle was a hypodermic needle Terumo with an approximate curvature radius of 3 µm, the insulating material was epoxy resin (Mepox–1685/L, a Bisphenol A diglycidyl ether (epoxy system) in Santiago, Chile and the cuboid dimensions were 10 × 10 mm base and 25 mm height. Electrical treeing experiments were carried out using the test circuit shown in Figure 1. The voltage source (Vac) was a transformer fed from the grid through a variac (Variable AC Transformer). The samples were fed through a limiting resistance (R) in order to reduce disturbances and protect the instruments in case of breakdown. PD measurements were carried out using the balanced circuit shown in standard IEC 60,270 [18]. The treeing sample (N2) and the dummy sample (N1, PD free) were placed into a transparent oil container to prevent unwanted surface discharges and allow visualization of tree growth using an optical camera. The signals from the treeing and dummy samples were subtracted in the subtracting circuit (SC), whose output was fed to a commercial PD system (Acquisition System) that continuously registered PD activity. The voltage was measured using a voltage divider (Vm), which was also used by the PD measurement system. The minimum value of PD magnitude for the measurement was set to 2 pC; however, to reduce the background noise recorded, a threshold between 10 to 15 pC was used for the analysis.
Before the electrical tree growth experiment itself, an incipient electrical tree needed to be created in each sample. To initiate an electrical tree, 12–16 kV 50 Hz voltage was applied to each sample until the camera optically observed a tree, and then the voltage was turned off and the sample was prepared for the tree growth experiment. By doing this, the initiation stage was separated from the propagation stage, which was the stage to be analyzed in this research. The electrical tree growth experiment was carried out in the selected samples by applying 12, 14, and 16 kVrms 50 Hz until breakdown, according to Table 1, where also the resulting time-to-breakdown (Time BD) is shown. This time is the duration of tree growth from the initial stage until final breakdown of the insulation. PD measurements were made using two simultaneous means: continuous recording with filming camera and taking pictures every 10 s. The utilization of this simultaneous registry system was required to correlate PD behavior (electrical response of the insulation) and tree propagation shape/length (physical damage).

2.2. Partial Discharge Recorded and Selected Data for Analysis

The measurements of the electrical tree growth experiment are shown in Figure 2, where the PD amplitude time series (left axis) and the tree length progression (right axis) are compressed in the same graph for each sample. Tree length was extracted from the tree images taken during the growth and was measured as the furthest tree extent from the needle tip in the direction of the plane electrode. In the graph, the length was represented in per unit values, i.e., the ratio between the length (L) and the length of the first tree branch that reached the plane electrode (Lmax). Note that dielectric breakdown did not occur immediately after the tree arrived at the counter electrode. In particular, in the cases of Samples A and B, a considerable amount of experimental time passed after the tree bridged the insulation.
Although PD was recorded during the entire experiment, tree growth analysis was carried out using ten selected windows or intervals of analysis to study the parameters’ evolution during tree growth. The selection of data and intervals followed the criteria previously described by Zheng et al. [19]. Each interval was selected to have at least 10,000 PD events and at least 10 s of continuous measuring time. In practice, this resulted in a total of 10,000–60,000 PD events (observations) per analysis interval for all the samples. The first analysis interval was chosen to start three minutes after the beginning of the test for a more stable PD activity, and the last interval was set to finish at least five minutes before the breakdown. The separation between intervals depended on the duration of each test. The intervals of analysis are shown as black bands in Figure 2a–e.
The results indicate that PD dynamics are different for every sample, depending on the stressing voltages and the stage of tree growth. For example, Samples A and B had irregular trends, and Sample B even had periods of no detected PD while the tree was growing. This phenomenon has been reported before and is due to the growth of ‘filamentary’ trees [19]. This is observed in Figure 3; though Samples B and C were both stressed at 14 kV, time series of PD amplitudes had different behavior, which was also observed when comparing Samples D and E, stressed at 16 kV. In particular, Sample D showed the highest PD amplitude values among all the samples, with a constantly increasing trend.
Images of the electrical trees of each sample at interval 6 are presented in Figure 4. It can be observed how Sample A, aged at 12 kV, presented the widest electrical tree. It is worth noting that the images are from the same interval (6th), but they do not correspond necessarily to similar stressing time; for Sample A, the 6th interval was at 130 min of aging, which is longer than any other total stressing time.

3. Gaussian Mixture Model Clustering Technique and Classification Model

Clustering techniques have been widely used in power system analysis and Rajabi et al. [20] have discussed a literature survey of various clustering techniques available and their application towards smart metering. Out of all the available models for unsupervised learning, the most popular is k-means clustering, which groups data according to a distance-based calculation with respect to a centroid [21]. The centroids are updated iteratively through a mean value and the clustered data will be in a circular shape. The drawback of the k-means clustering is that it fails to cluster data that are not in a circular shape, such as elliptical shape or irregular patterns. This drawback is overcome by GMM, which uses a probability density function (PDF) determining parameters by expectation-maximization (EM) technique. Compared to the k-means, the centroid formed by GMM takes into account the mean as well as the variance of the data, accommodating different sized clusters with varying correlations within them [22].
The clustering of unimodal distribution and multimodal distribution using GMM is explained in [22,23,24]. The comparison in [25] reveals that GMM takes more simulation time than k-means. Additionally, GMM can group complex patterns into similar components that match closely while k-means uses simple principles to produce only abstract information. The performance and comparison of the sampling methods used in GMM are reported in [26].
GMM can also be used for both hard and soft clustering of the dataset. In hard clustering, the GMM assigns each query data point to a particular cluster, which will maximize the posterior probability of the component given the data. In soft clustering, the GMM calculates the likelihood of the query data point belonging to a specific cluster and then assigns the query data point to a cluster, which would have maximum posterior probability, calculated using Bayes’ theorem. In this study, the versatile soft clustering GMM is utilized for unsupervised learning to model unknown data distribution by multivariate normal distributions. Unfortunately, k-means clustering has no means to measure the likelihood or uncertainty of cluster assignments. On the other hand, GMM uses probability distribution functions that can model any input dataset by assigning each point a probability to belong to a certain cluster. Hence, it is used for clustering in this work.
The various steps in the GMM are explained in [22]. The Gaussian model is formulated by Equations (1) and (2). Let X   =   x 1 , x 2 , x n   be a set of n   observations. The variable x i   is distributed among a mixture of M   components. The PDF of x i is written as shown in Equation (1), which is the weighted sum of Gaussian densities given by Equation (2) and the sum of weights i   =   1 M w i   =   1 , w i   represents the mixing probabilities.
p x i | λ   =   i   =   1 M w i g x i | μ i ,   Σ i
g x | μ i ,   Σ i   =   1 2 π D / 2 Σ i 1 / 2 e x p 1 2 x     μ i T Σ i 1 x     μ i
where M is the number of Gaussian densities, x-D is the dimensional continuously valued data vector, wi, I = 1…M is the weight of the mixture, g x | μ i ,   Σ i , i   =   1 , 2 M   is the component of Gaussian densities, μ is the mean vector of dimension D, Σ is the covariance matrix of dimension D×D, and λ i   =   w i , μ i ,   Σ i   is   the parameter of the GMM. In this study of clustering for insulation degradation using GMM the following covariance structure is adopted:
  • The covariance structure of the components will determine the shape and orientation of the ellipsoid drawn over the cluster. The covariance matrix is diagonal instead of being full to avoid the over-fitting problem, and major and minor axes of the ellipsoid are parallel and perpendicular to the abscissa and the ordinate. The covariance matrix is shared among the components; hence, the ellipse of each cluster has the same size and orientation.
  • The expectation-maximization (EM) algorithm fits the GMMs. The initial values of the parameters are set, and then the initial cluster assignments for data points are allowed to be selected randomly.
  • Regularization is applied in order to avoid the likelihood of data point becoming ill-conditioned and starts moving towards infinity.

3.1. The Expectation Maximization (EM) Algorithm

Expectation maximization (EM) is a mathematical algorithm used to find the correct parameters for a model. The estimated parameter of mean, variance, and weight are necessary to cluster the data, but this is possible only if the Gaussian family is known. The EM algorithm starts with random parameters, and then the optimal parameters are found by iteration. This algorithm has the capability to deal with latent variables. Assuming k clusters are to be assigned, then k distributions are required with mean and covariance values of µ1, µ2, …, µk and Σ1, Σ2, …, Σk, respectively. The EM algorithm generally has two main steps, i.e., the Expectation step (E-step) and the Maximization step (M-step) [27].

3.1.1. The Expectation Step (E-step)

In this step, using randomly initialized parameters, for every point xj, we obtain the likelihood of belonging to a certain cluster c1, c2, ..., ck. This is achieved using Equation (3).
r j c =   p r o b a b i l i t y   t h a t   x j b e l o n g s   t o   c s u m   o f   p r o b a b i l i t y   x j   b e l o n g s   t o   c 1 , c 2 c k   =   π c   N x j   ;   μ c   , Σ c Σ c π c N x j   ;   μ c   , Σ c
This value would be high if the point is allocated to the correct class, or vice versa.

3.1.2. The Maximization Step (M-step)

In this step, the parameter λ is updated as follows:
  • The weight is updated using Equation (4), which is the ratio of cluster points to the overall number of points.
    w c   =   N u m b e r   o f   p o i n t s   a s s i g n e d   t o   a   c l u s t e r T o t a l   n u m b e r o f   p o i n t s
  • Then, the covariance and the mean values are modified using Equations (5) and (6) in relation to the probability values for the data point and based on the values assigned to that particular distribution. Therefore, any data point having a high probability of being a member of the distribution should be contributing a higher portion.
    μ c = 1 N u m b e r   o f   p o i n t s   a s s i g n e d   t o   c l u s t e r Σ j r j c x j
    Σ c = 1 N u m b e r   o f   p o i n t s   a s s i g n e d   t o   c l u s t e r Σ j r j c ( x j μ c ) T ( x j   μ c )
Based on the updated parameters the E-step is repeated. These two steps are iterated until the optimal parameters are obtained using the log-likelihood function as described in [3].

4. Data Processing

Data pre-processing is an important step in machine learning models. To explain the data pre-processing steps, dataset A at the first interval (A1) is used. Sample A1 analysis is summarized in Table 2. The dataset has 10,139 observations with minimum values of Phi (phase angle) and Q (PD amplitude) are −1.790 × 10−5 and −7.220 × 10−11, respectively. The maximum values of Phi and Q are 9.997 × 10−1 and 8.880 × 10−11, respectively. In the data pre-processing step, the data is transformed using the normalized function available in MATLAB, which transforms the data with a mean of 0 and standard deviation of 1. The transformed variables Phi (Stdscale) and Q (Stdscale) are used by the clustering algorithm, which has a mean of 0 and standard deviation of 1 as shown in Table 2.

5. Results and Discussion

5.1. Gaussian Mixture Model Clustering

GMM clustering was applied to generate clusters for every training pattern from the initial stage to the breakdown of the insulation, for each dataset. The overall aim was to be able to understand and classify different PD patterns during the stages of tree growth. The important hyperparameters in GMM are the number of clusters k and appropriate covariance structure Σ. In GMM the covariance structure includes a covariance matrix, which can be diagonal or full, and the nature of the covariance matrix, which can be shared or unshared. When a diagonal covariance matrix is chosen, the minor and major axes of the confidence ellipsoids drawn over the clusters are parallel or perpendicular to the x and y axes. When a full covariance matrix is chosen, there is no restriction to the orientation of the minor and major axis of the confidence ellipsoids drawn over the clusters. A shared covariance matrix indicates that all confidence ellipsoids have the same size and orientation, whereas an unshared covariance matrix indicates different sizes and shapes of the confidence ellipsoids. Choosing the appropriate hyperparameter is a very important task, and the method adopted is presented in the next section.

5.1.1. Hyperparameter Tuning of GMM

The number of cluster components k and appropriate covariance structure Σ is unknown for each stage of electrical treeing in Samples A−E. The most commonly used technique to tune the hyperparameters is by comparing the Akaike information criterion (AIC) and Bayesian information criterion (BIC). AIC is the relative distance between the unknown true likelihood function of the data and the fitted likelihood function of the model. A lower AIC means the model is closer to reality. BIC is an estimate function of the posterior probability of a model being true under certain assumptions, so a lower BIC means the model is bound to be the genuine model. A detailed discussion on the importance of AIC and BIC related to model selection is available in [28]. The procedure to choose the optimal values for the hyperparameters namely the number of clusters k and appropriate covariance structure Σ is shown in Figure 5 and was coded using MATLAB. A regularization value of 0.01 is specified and the EM iteration is specified as 10,000 in order to avoid ill-conditioning of the covariance matrix during EM iteration. The number of clusters k takes a value from 1 to 12.
The procedure shown in Figure 5 is performed for all fifty datasets. The procedure is explained for the initial stage of electrical treeing in Sample A in the first interval and the same procedure is applicable for other datasets. For sample A1, the AIC and BIC values are summarized in Table 3 and Table 4, respectively. The bar plots grouped by the number of clusters are shown in Figure 6 and Figure 7, respectively. From the bar plot it is very clear that when the number of clusters is more than 1, the AIC and BIC values decrease, and after a specific number of clusters is reached, the variation of AIC values is significantly low. The point where the change in AIC value declines the most is the elbow point. The elbow point in Figure 8 and Figure 9 was determined by plotting a curb, and corresponds to the number of clusters k = 4 and covariance structure which is full and unshared. These are the best hyperparameters for dataset A1. The same procedure has to be repeated in all datasets and the best hyperparameter for each dataset in samples A to E are provided in Table 5, Table 6, Table 7, Table 8 and Table 9, respectively.
According to Table 5, Table 6, Table 7, Table 8 and Table 9, the minimum value of k is 4 and its maximum value is 8. The nature of the covariance matrix is mostly full with an unshared structure and rarely diagonal with an unshared structure. The values of hyperparameters selected for the GMM are shown in Table 10.
The shared covariance is false, which indicates a non-identical or unshared covariance matrix. The grid length is important in order to draw the confidence ellipsoids over the clusters. Grid length and the number of iterations for the EM algorithm are selected by trial and error.

5.2. GMM Results and Discussions

The six clusters with their centers and confidence ellipsoids as shown in Figure 10 indicate GMM models for the PRPD patterns of the initial stage (interval 1) of electrical treeing in Samples A, B, C, D and E. From Figure 10, it is clear that the ellipsoids are of different sizes and there are no restrictions to the orientation of their minor and major axis. Figure 11 shows the six clusters with their centers and confidence in the final stage (interval 10) of electrical treeing in Samples A, B, C, D and E ,which is close to the insulation breakdown. An interesting observation from Figure 11 is that none of the clusters are circular in shape and, therefore, K-means clustering could not be used here since it is not built to account for other shapes and a circular fit would be a bad fit to the data. In samples C and D, cluster overlapping can be clearly appreciated. The GMM algorithm produces the cluster centers based on the shape of the PRPD pattern and the PD activity and plots confidence ellipsoids with a 99% probability threshold as specified in the MATLAB program.
For each dataset sample, the normalized data is clustered into six groups, differentiated by color, using the GMM clustering. For each cluster in the two-dimensional (2D) plane, the midpoint of the cluster is also indicated in Figure 10 and Figure 11. In each case, the Phi and Q are normalized to return the vector-wise Z score of all the datasets (i.e., Samples A−E) with center 0 and standard deviation 1. The midpoints of the six clusters for each sample are shown in Appendix A.
For each dataset, the normalized data is clustered into six groups using the GMM clustering. From Figure 10 and Figure 11, it can be seen that the clusters are different, showing that the applied voltage and the stage of tree growth have a significant effect on the generated PD patterns. The initial and final stage clusters appear to be different for all the samples except for Sample A. For this sample, there is a similarity between the initial and the final stage clusters probably due to the continuous irregular trend in PD because of the formation of filamentary trees. The variance of the data in each cluster appears to be higher in the final tree stages when compared to the initial stages. This might be the PD mechanism over time and the increase in the number of PD events as the tree approaches the breakdown stage. It can be seen that the PRPD pattern clusters for all 5 samples at the initial stage appeared to be centered at different positions despite the fact that the applied voltages are close to each other. This shows that, even for the same applied voltage, the PD phenomenon in trees exhibits complex behavior and does not show any clear trend, making it difficult to evaluate in some instances. These conclusions can also be associated with the final stages of the treeing patterns in Figure 11.
As a further analysis of the GMM technique, Appendix A and Appendix B shows the cluster centers and their mean values/variances for all the samples (A−E), i.e., from the initial to the final stages of tree growth. In Appendix A, the cluster centers for the different sample datasets are shown. It can be seen that for samples A, B, and C, the positioning of the clusters from the beginning to the breakdown level varies significantly, although the first and the third cluster for all the samples A−E, fall within close range with insignificant variance in the X-axis in Appendix B. However, cluster 2 evidenced higher variance for samples A−C. This might be due to the stochastic nature of the PD mechanism of electric trees and the fact that the occurrence of the PD patterns within the first and last half of the AC power cycle are similar. Furthermore, it is interesting to note that the patterns for samples D and E have very similar pattern trends in the initial stage and the final breakdown stage, with insignificant variance among all the cluster shapes (see Appendix B, Table A6). This is a clear indication that with a lower breakdown time of insulation, treeing patterns are identical, and can easily be categorized. In the case of samples A−C, the results showed wider PD variability among the samples, showing their distinct nature. It was difficult to clearly differentiate between the initial and breakdown stage of the PD patterns. In addition, most of the pattern cluster centers show lower variance along the x-coordinates as compared to the y-coordinate, showing higher amplitude variation for the PD patterns as compared to the phase changes. The information in appendix B can serve as a statistical tool to predict and identify the applied voltages of the treeing patterns and their breakdown times for electrical trees in epoxy resin insulation. Assuming that there are new samples, they need to be classified as belonging to any of the confidence intervals of the samples in Appendix B and can be regarded as that particular sample.
In general, the results imply that the GMM can classify different degradation stages of the treeing patterns up to breakdown for samples having breakdown times higher than one hour, while it cannot effectively perform the same function for samples with shorter breakdown times, e.g., half an hour and lower. This might be because the PD mechanism has not been allowed to develop and it bridges the insulation in a shorter time, while more time before breakdown refers to a wider spread of treeing branches and more deterioration of the insulating material. To a certain extent, it can be said that the information in Appendix B almost correlates with the breakdown times but not necessarily the applied electrical stress across the sample defects. Although voltage can be regarded as one of the factors affecting the tree growth shape and structure in the cases analyzed here, the insulation breakdown time appears to be a key factor in determining the degradation levels of the PD patterns emanating from electrical trees.

5.3. Electrical Tree Pattern Recognition

As GMM is an easy and efficient technique for clustering, it has been shown, in the previous section, to aid the classification of electrical tree patterns. The fundamental flowchart for the electrical tree pattern recognition is shown in Figure 12. First, the data is captured and subsequent PRPD patterns are formed, followed by the data processing and the GMM clustering. The recognition of the tree growth clusters can be done by PD pattern judgment, i.e., comparing the clusters with the already established cluster confidence intervals in Appendix B. In this case, the level of the tree growth can be known, i.e., the initial, final, or other stages of degradation can be determined.

6. Conclusion

In this paper, the GMM has been utilized for clustering and classification applications in electrical trees emanating from epoxy resin insulation. Different PD samples were captured at different voltages from the initial to the final breakdown stage. The results show that PD dynamics vary with different stressing voltages and with the level of the tree growth. Depending on the sample and the applied electrical stress, there are different breakdown times. The GMM is chosen over other techniques in this work because it is robust and can perform hard clustering for complex data such as electrical tree patterns. The results clearly indicate that GMM can effectively classify patterns from the initial to the breakdown level for breakdown times above an hour, but not breakdown times of less than an hour, as the ones obtained with samples stressed at the highest stressing voltage of 16 kV. The PD patterns for shorter breakdown times possess identical clusters through the degradation stages. In this paper, the cluster centers and their confidence intervals have been developed to recognize the PD patterns in electrical trees at different stages ranging from the initial to the breakdown stage. However, the results presented in this paper can be further validated by experimenting with different samples captured at different voltages and breakdown times. Further research can also be conducted for different insulating materials such as polyethylene or cross-linked polyethylene to ascertain the efficiency of the proposed classification tool.

Author Contributions

Conceptualisation, A.A.M. and A.S.; Data curation, A.A.M. and A.S.; Formal analysis, A.A.M. and A.S.; Funding acquisition, J.A.A.-R.; Investigation, A.A.M. and A.S.; Methodology, A.A.M.; Project administration, A.A.M. and J.A.A.-R.; Resources, A.A.; Software, A.A.M. and A.S.; Validation, A.A.M. and J.A.A.-R.; Visualization, F.M.-S..; Writing-original draft, A.A.M.; Writing-review and editing, A.A.M., A.S., J.A.A.-R., R.S., F.M.-S., and N.A.B. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge funding support from Agencia Nacional de Investigacion y Desarrollo (ANID) under grants Fondecyt 11181177, 1200055 and Fondef 19I10165, UTFSM under grant PI_m_19_01 and Jubail Industrial College for providing the facilities to conduct this research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to privacy.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Appendix A

Table A1. Cluster centers for dataset A.
Table A1. Cluster centers for dataset A.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
A1−1.1861.173−1.1250.387−0.1354−0.56980.1385−0.80570.5116−1.0161.6920.7549
A2−1.056−0.7532−1.0090.7146−0.7347−0.0112−0.4768−0.41070.88481.1431.092−0.251
A3−1.058−0.8859−0.8825−0.13−0.7949−1.4030.88981.160.97810.035781.1391.042
A4−1.187−0.0547−1.1570.77610.3293−0.57110.7657−0.84120.88460.23571.3060.8736
A5−1.0061.686−0.95730.2095−0.59260.14350.859−1.8020.987−0.24181.26−0.3535
A6−1.2580.5283−1.2480.4826−0.05480.88860.3141−0.72690.5488−1.6031.2410.5868
A7−1.253−0.3607−0.44230.73350.07192−0.38530.5229−0.37160.59250.96931.06−0.5146
A8−1.4611.946−1.2140.1791−1.180.90870.4871−0.82430.5937−0.11021.9120.854
A9−1.2560.2578−1.0571.301−0.6930.92270.5859−0.8970.9851−1.1361.7910.1495
A10−1.3081.352−1.0350.8009−0.1624−0.06380.1342−0.68210.5402−1.2841.6750.3228
Table A2. Cluster centers for dataset B.
Table A2. Cluster centers for dataset B.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
B1−1.0580.9892−0.27750.08860.4043−0.7230.6774−1.3060.7123−0.87061.9660.7448
B2−1.052−0.897−0.8772−0.803−0.78550.17710.89651.1251.001−0.12121.1561.217
B3−1.124−0.871−0.8612−0.7610.82510.77491.017−0.04781.1181.3621.268−0.0238
B4−1.167−0.6658−0.9213−1.19−0.8323−0.0097−0.7085−1.3570.91221.0611.1340.4679
B5−1.385−1.039−1.0110.2924−1.014−0.23180.49210.77030.8353−0.23621.903−1.131
B6−1.0091.087−0.87590.7558−0.64020.22050.8734−0.76981.076−1.0971.2230.0073
B7−1.0640.7804−0.9420.94840.6927−0.59580.9069−2.6731.033−0.87641.0790.3853
B8−1.1970.8737−0.2211.2530.0151−1.0920.1591−0.66680.4250.31271.3480.9801
B9−1.1850.7059−1.1070.2628−0.1266−0.37750.7336−0.11640.8355−0.23721.706−0.1947
B10−1.080.7476−1.0520.01270.8639−0.67981.016−0.19251.1350.14451.34611.28
Table A3. Cluster centers for dataset C.
Table A3. Cluster centers for dataset C.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
C1−1.0590.8105−1.050.7709−0.91091.265−0.4285−0.50410.9417−0.78931.037−1.096
C2−1.1840.6434−1.163−0.7276−0.93621.2990.6173−0.52020.9107−1.1871.3960.4608
C3−1.5271.04−1.2281.0530.1526−0.50480.1862−0.22350.2575−1.0611.6530.638
C4−1.320.89080.1122−0.39290.1897−0.30520.2012−1.0570.2261−0.59831.6310.6416
C5−1.1670.5778−1.1291.2180.2461−0.60280.4939−1.5080.5652−0.48431.8450.6457
C6−1.0570.5739−0.9929−1.147−0.99110.7769−0.88451.2490.960.79480.965−1.215
C7−1.0390.721−1.035−1.0940.016621.0510.879−1.0460.92770.70061.009−1.649
C8−0.90870.7963−0.8923−1.1160.410.26990.9988−1.021.072−1.2781.0770.8119
C9−0.94120.7026−0.9178−1.116−0.86950.96571.0410.7011.055−1.1991.1050.9908
C10−0.95530.662−0.8707−0.10810.9538−0.93781.0610.97781.07−1.3841.0950.5495
Table A4. Cluster centers for dataset D.
Table A4. Cluster centers for dataset D.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
D1−1.3441.282−1.2460.4507−0.1909−0.39220.2312−1.0580.3376−0.72181.6240.7341
D2−1.3251.267−1.2690.3666−0.2559−0.26630.0334−0.69150.4417−0.98021.6190.7413
D3−1.5821.118−1.2951.007−0.2618−0.213−0.2295−0.21930.2859−0.81781.6230.6534
D4−1.2620.8567−0.448−2.087−0.2115−0.05350.2842−0.34030.3249−0.9621.6480.7744
D5−1.1550.786−1.130.33−0.0899−0.11650.4073−0.82351.4510.45041.690.6882
D6−1.1990.8595−1.1280.2565−0.0391−0.4619−0.0344−0.19490.43550.90971.640.4941
D7−1.3850.8288−1.1160.89−0.1785−0.11130.0824−0.73880.4462−1.0741.5710.5003
D8−1.2230.9173−0.3391−1.581−0.1936−0.0970.2715−1.2390.3142−0.52051.5850.5408
D9−1.0420.9082−0.99330.4894−0.98540.22180.0139−0.33290.4613−1.1361.6380.5217
D10−1.1310.8649−1.1130.3652−0.0428−0.40830.2569−0.1880.4195−1.0311.5870.5078
Table A5. Cluster centers for dataset E.
Table A5. Cluster centers for dataset E.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
E1−1.3644.287−1.250.1643−1.0480.41620.306−0.97290.568−0.18961.6011.363
E2−1.1870.6647−1.1580.3211−0.0214−2.7690.3147−0.66050.91890.99221.6420.7291
E3−1.3215.777−1.040.29630.1226−0.27360.282−2.8140.7842−0.26252.0891.537
E4−1.1020.4142−0.23982.350.4716−2.9430.5297−0.3440.6166−0.43421.940.4602
E5−1.2985.447−1.1540.0571−1.132.8240.4232−0.02490.6845−2.311.7290.0095
E6−1.1552.081−1.1030.07180.4351−1.3650.48310.01771.7840.066761.81.247
E7−1.1031.011−1.0530.3248−0.9927−0.07640.4207−1.0480.5579−0.00671.7730.6432
E8−1.0570.7747−1.0210.24940.5946−0.30080.7169−1.0440.8952−0.17771.9840.3507
E9−1.0360.9421−1.030.27270.57720.30240.5828−0.18030.5843−0.92981.8730.3772
E10−1.1620.8128−1.070.08720.5089−0.47310.5709−0.60041.8290.80531.853−1.366

Appendix B

Table A6. Mean values of the cluster centers and their variances.
Table A6. Mean values of the cluster centers and their variances.
SampleCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6
XYXYXYXYXYXY
AMean−1.20290.48886−1.0130.54543−0.3947−0.01410.422−0.620.75064−0.31.41680.34645
Std.Dev0.136591.015120.23030.406290.47130.76430.41530.7370.208250.9430.31720.564852
BMean−1.13210.1711−0.8150.08594−0.0598−0.25370.6064−0.5230.90833−0.061.41291.373295
Std.Dev0.108870.905720.30890.800680.73020.55430.53181.10060.218650.8110.32543.54433
CMean−1.11580.74183−0.917−0.266−0.17390.32770.4165−0.2950.79859−0.651.28130.07783
Std.Dev0.191420.146030.38030.953240.69440.8510.65920.95670.327440.7930.32190.984708
DMean−1.26480.96884−1.0080.04874−0.2449−0.18980.1317−0.5830.49178−0.591.62250.61561
Std.Dev0.152990.183770.33671.028630.27220.20470.1910.38150.343110.7010.03510.113455
EMean−1.17852.22115−1.0120.41947−0.0482−0.46580.463−0.7670.92226−0.241.82840.535085
Std.Dev0.113782.113480.28070.686290.72341.65520.14070.82440.483850.9170.15180.830019

References

  1. Champion, J.V.; Dodd, S.J.; Alison, J.M. The correlation between the partial discharge behaviour and the spatial and temporal development of electrical trees grown in an epoxy resin. J. Phys. D. Appl. Phys. 1996, 29, 2689. [Google Scholar] [CrossRef]
  2. Conti, M.; Cavallini, A.; Montanari, G.C.; Guastavino, F. Identification of electrical tree growth in insulation systems by Fuzzy logic techniques based on partial discharge acquisition. In Proceedings of the 2004 IEEE International Conference on Solid Dielectrics ICSD 2004, Toulouse, France, 5–9 July 2004. [Google Scholar]
  3. Iddrissu, I.; Lv, Z.; Rowland, S.M. The dynamic character of partial discharge in epoxy resin at different stages of treeing. In Proceedings of the 2016 IEEE International Conference on Dielectrics, ICD 2016, Montpellier, France, 3–7 July 2016. [Google Scholar]
  4. Hoof, M.; Patsch, R. Analyzing partial discharge pulse sequencesA new approach to investigate degradation phenomena. In Proceedings of the Conference Record of IEEE International Symposium on Electrical Insulation, Pittsburgh, PA, USA, 5–8 June 1994. [Google Scholar]
  5. Sarathi, R.; Oza, K.H.; Pavan Kumar, C.L.G.; Tanaka, T. Electrical treeing in XLPE cable insulation under harmonic AC voltages. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 3177–3185. [Google Scholar] [CrossRef]
  6. Chen, X.; Xu, Y.; Cao, X. Nonlinear time series analysis of partial discharges in electrical trees of XLPE cable insulation samples. IEEE Trans. Dielectr. Electr. Insul. 2014, 21, 1455–1461. [Google Scholar] [CrossRef]
  7. Donoso, P.; Schurch, R.; Ardila, J.; Orellana, L. Analysis of Partial Discharges in Electrical Tree Growth Under Very Low Frequency (VLF) Excitation Through Pulse Sequence and Nonlinear Time Series Analysis. IEEE Access 2020, 8, 163673–163684. [Google Scholar] [CrossRef]
  8. Vogelsang, R.; Fruth, B.; Farr, T.; Fröhlich, K. Detection of electrical tree propagation by partial discharge measurements. Euro. Trans. Electr. Power 2005, 15, 271–278. [Google Scholar] [CrossRef]
  9. Schurch, R.; Rowland, S.M.; Bradley, R.S. Partial discharge energy and electrical tree volume degraded in epoxy resin. In Proceedings of the Annual Report—Conference on Electrical Insulation and Dielectric Phenomena, CEIDP, Ann Arbor, MI, USA, 18–21 October 2015. [Google Scholar]
  10. Lv, Z.; Rowland, S.M.; Chen, S.; Zheng, H.; Iddrissu, I. Evolution of partial discharges during early tree propagation in epoxy resin. IEEE Trans. Dielectr. Electr. Insul. 2017, 24, 2995–3003. [Google Scholar] [CrossRef]
  11. Bao, M.; Yin, X.; He, J. Analysis of electrical tree propagation in XLPE power cable insulation. Phys. B Condens. Matter 2011, 406, 1556–1560. [Google Scholar] [CrossRef]
  12. Xiaoquan, Z.; George, C. Propagation mechanism of electrical tree in XLPE cable insulation by investigating a double electrical tree structure. IEEE Trans. Dielectr. Electr. Insul 2008, 15, 12–23. [Google Scholar]
  13. Alapati, S.; Thomas, M.J. Electrical treeing and the associated PD characteristics in LDPE nanocomposites. IEEE Trans. Dielectr. Electr. Insul. 2012, 19, 697–704. [Google Scholar] [CrossRef]
  14. Zhou, T.; Liu, L.; Liao, R.; Chen, G. Study on propagation characteristics and analysis of partial discharges for electrical treeing in XLPE power cables. In Proceedings of the 2010 IEEE International Conference on Solid Dielectrics, ICSD 2010, Potsdam, Germany, 4–9 July 2010. [Google Scholar]
  15. Park, S.H.; Jung, H.E.; Yun, J.H.; Kim, B.C.; Kang, S.H.; Lim, K.J. Classification of defects and evaluation of electrical tree degradation in cable insulation using pattern recognition method and weibull process of partial discharge. In Proceedings of the 2008 International Conference on Condition Monitoring and Diagnosis, CMD 2008, Beijing, China, 21–24 April 2008. [Google Scholar]
  16. Salama, M.M.A.; Mazroua, A.A.; Bartnikas, R. Neural Network System Using the Multi-Layer Perceptron Technique for the Recognition of PD Pulse Shapes due to Cavities and Electrical Trees. IEEE Trans. Power Deliv. 1995, 10, 92–96. [Google Scholar]
  17. Park, S.H.; Lim, K.J.; Kang, S.H. Recognition of electrical tree progress in solid insulator. In Proceedings of the the IEEE International Conference on Properties and Applications of Dielectric Materials, Bali, Indonesia, 26–30 June 2006. [Google Scholar]
  18. International Electrotechnical Commission Standard (IEC 60270). “High-Voltage Test Techniques—Partial Discharge Measurements; Jabatan Standard: Cyberjaya, Malaysia, 1986; ISBN 5990722001. [Google Scholar]
  19. Zheng, H.; Rowland, S.M.; Iddrissu, I.; Lv, Z. Electrical treeing and reverse tree growth in an epoxy resin. IEEE Trans. Dielectr. Electr. Insul. 2017, 24, 3966–3973. [Google Scholar] [CrossRef]
  20. Rajabi, A.; Eskandari, M.; Ghadi, M.J.; Li, L.; Zhang, J.; Siano, P. A comparative study of clustering techniques for electrical load pattern segmentation. Renew. Sustain. Energy Rev. 2020, 120, 109628. [Google Scholar] [CrossRef]
  21. Qiu, L.; Yuan, S.; Mei, H.; Fang, F. An improved gaussian mixture model for damage propagation monitoring of an aircraftwing spar under changing structural boundary conditions. Sensors 2016, 16, 291. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Cluster Using Gaussian Mixture Model—MATLAB & Simulink—MathWorks India. Available online: https://www.mathworks.com/help/stats/clustering-using-gaussian-mixture-models.html (accessed on 22 March 2021).
  23. Sugiyama, M. Maximum Likelihood Estimation for Gaussian Mixture Model. In Introduction to Statistical Machine Learning; Elsevier: Amsterdam, The Netherlands, 2016; pp. 157–168. [Google Scholar]
  24. O’Hagan, A.; Murphy, T.B.; Scrucca, L.; Gormley, I.C. Investigation of parameter uncertainty in clustering using a Gaussian mixture model via jackknife, bootstrap and weighted likelihood bootstrap. Comput. Stat. 2019, 34, 1779–1813. [Google Scholar] [CrossRef] [Green Version]
  25. Cho, J.; Jung, Y.; Kim, D.S.; Lee, S.; Jung, Y. Moving object detection based on optical flow estimation and a gaussian mixture model for advanced driver assistance systems. Sensors 2019, 19, 3217. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Patel, E.; Kushwaha, D.S. Clustering Cloud Workloads: K-Means vs. Gaussian Mixture Model. Procedia Comput. Sci. 2020, 171, 158–167. [Google Scholar] [CrossRef]
  27. Reynolds, D. Gaussian Mixture Models. Encycl. Biom. 2009, 741, 659–663. [Google Scholar]
  28. Vrieze, S.I. Model selection and psychological theory: A discussion of the differences between the Akaike information criterion (AIC) and the Bayesian information criterion (BIC). Psychol. Methods 2012, 17, 228. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Test circuit for electrical tree growth experiments.
Figure 1. Test circuit for electrical tree growth experiments.
Sensors 21 02562 g001
Figure 2. Time series of partial discharge (PD) amplitude and tree length for each sample: (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E different.
Figure 2. Time series of partial discharge (PD) amplitude and tree length for each sample: (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E different.
Sensors 21 02562 g002
Figure 3. Picture of Sample B at interval 2, showing a filamentary tree growth.
Figure 3. Picture of Sample B at interval 2, showing a filamentary tree growth.
Sensors 21 02562 g003
Figure 4. Images of electrical trees at interval 6 in each sample: (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E.
Figure 4. Images of electrical trees at interval 6 in each sample: (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E.
Sensors 21 02562 g004
Figure 5. Procedure to choose the hyperparameters.
Figure 5. Procedure to choose the hyperparameters.
Sensors 21 02562 g005
Figure 6. Bar Plot of Akaike information criterion values for each fit in dataset A1.
Figure 6. Bar Plot of Akaike information criterion values for each fit in dataset A1.
Sensors 21 02562 g006
Figure 7. Bar Plot of Bayesian information criterion values for each fit in dataset A1.
Figure 7. Bar Plot of Bayesian information criterion values for each fit in dataset A1.
Sensors 21 02562 g007
Figure 8. Plot of Akaike information criterion values for each fit in dataset A1.
Figure 8. Plot of Akaike information criterion values for each fit in dataset A1.
Sensors 21 02562 g008
Figure 9. Plot of Bayesian information criterion values for each fit in dataset A1.
Figure 9. Plot of Bayesian information criterion values for each fit in dataset A1.
Sensors 21 02562 g009
Figure 10. Clusters for the different datasets at the initial stage of degradation (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E.
Figure 10. Clusters for the different datasets at the initial stage of degradation (a) Sample A, (b) Sample B, (c) Sample C, (d) Sample D, and (e) Sample E.
Sensors 21 02562 g010
Figure 11. Clusters for the different samples at the final stage (interval 10): (a) Sample A (b) Sample B (c) Sample C (d) Sample D and (e) Sample E.
Figure 11. Clusters for the different samples at the final stage (interval 10): (a) Sample A (b) Sample B (c) Sample C (d) Sample D and (e) Sample E.
Sensors 21 02562 g011
Figure 12. Flowchart for GMM Electrical tree pattern recognition.
Figure 12. Flowchart for GMM Electrical tree pattern recognition.
Sensors 21 02562 g012
Table 1. Samples, voltages, and the resulting time to breakdown of each experiment.
Table 1. Samples, voltages, and the resulting time to breakdown of each experiment.
SampleVoltage (kV)Time BD (h)
A123.8
B141.5
C141.4
D160.6
E160.47
Table 2. Statistical evaluation of Sample A at the first time interval.
Table 2. Statistical evaluation of Sample A at the first time interval.
CountMeanStandard DeviationMinimumMaximum
Phi10139.04.460 × 10−13.053 × 10−11.790 × 10−59.997 × 10−1
Q10139.0−1.297 × 10−124.454 × 10−11−7.220 × 10−118.880 × 10−11
Phi (Stdscale)10139.08.964 × 10−171.000 × 10−1.460 × 101.813 × 10
Q (Stdscale)10139.01.971 × 10−181.00 × 10−1.591 × 102.022 × 10
Table 3. Akaike information criterion (AIC) values for dataset A at the first interval (A1).
Table 3. Akaike information criterion (AIC) values for dataset A at the first interval (A1).
k ValueCovariance Structure Σ
Diagonal-SharedFull-SharedDiagonal-UnsharedFull-Unshared
157,55355,53057,55355,530
246,94640,75841,38835,693
326,41026,14123,56921,568
423,53822,95721,61619,709
522,01722,01520,16118,875
621,77621,00219,39618,683
720,98122,02419,44118,252
820,30020,49519,13518,178
920,52720,26318,82117,898
1020,28620,25919,06917,885
1120,13920,05618,75818,064
1220,29720,29819,08917,907
Table 4. Bayesian information criterion (BIC) values for dataset A at the first interval (A1).
Table 4. Bayesian information criterion (BIC) values for dataset A at the first interval (A1).
k ValueCovariance Structure Σ
Diagonal-SharedFull-SharedDiagonal-UnsharedFull-Unshared
157,58255,56657,58255,566
246,99640,81641,45335,772
326,48226,22123,67021,691
423,63223,05821,75319,875
522,13222,13820,33419,084
621,91321,14619,60618,936
721,14022,19019,68618,548
820,48120,68319,41718,517
920,72920,47219,13918,281
1020,51020,49019,42318,312
1120,38520,30919,14818,534
1220,56520,57319,51618,420
Table 5. The choice of best hyperparameter for dataset A.
Table 5. The choice of best hyperparameter for dataset A.
DatasetAICBIC
k ValueΣ Choicek ValueΣ Choice
A14Full-unshared4Full-unshared
A26Full-unshared6Full-unshared
A35Full-unshared5Full-unshared
A46Full-unshared6Full-unshared
A58Full-unshared8Full-unshared
A68Full-unshared8Full-unshared
A77Full-unshared7Diagonal-unshared
A85Full-unshared5Full-unshared
A98Diagonal-unshared8Diagonal-unshared
A106Diagonal-unshared6Diagonal-unshared
Table 6. The choice of best hyperparameter for dataset B.
Table 6. The choice of best hyperparameter for dataset B.
DatasetAICBIC
k ValueΣ Choicek ValueΣ Choice
B18Diagonal-unshared8Diagonal-unshared
B26Full-unshared6Full-unshared
B38Full-unshared8Full-unshared
B46Full-unshared6Full-unshared
B56Diagonal-unshared6Diagonal-unshared
B66Diagonal-unshared6Diagonal-unshared
B77Diagonal-unshared7Diagonal-unshared
B88Diagonal-unshared8Diagonal-unshared
B97Full-unshared6Diagonal-unshared
B106Full-unshared6Full-unshared
Table 7. The choice of best hyperparameter for dataset C.
Table 7. The choice of best hyperparameter for dataset C.
DatasetAICBIC
k ValueΣ Choicek ValueΣ Choice
C14Full-unshared4Full-unshared
C26Full-unshared6Full-unshared
C36Full-unshared6Full-unshared
C46Full-unshared6Full-unshared
C55Full-unshared5Full-unshared
C65Full-unshared5Full-unshared
C76Full-unshared5Diagonal-unshared
C86Full-unshared6Full-unshared
C96Full-unshared5Diagonal-unshared
C106Full-unshared6Full-unshared
Table 8. The choice of best hyperparameter for dataset D.
Table 8. The choice of best hyperparameter for dataset D.
DatasetAICBIC
k ValueΣ Choicek ValueΣ Choice
D15Full-unshared5Full-unshared
D26Diagonal-unshared6Diagonal-unshared
D36Full-unshared6Full-unshared
D46Full-unshared6Full-unshared
D56Full-unshared6Full-unshared
D65Full-unshared5Full-unshared
D76Full-unshared6Full-unshared
D85Full-unshared5Full-unshared
D96Full-unshared6Full-unshared
D106Full-unshared6Full-unshared
Table 9. The choice of best hyperparameter for dataset E.
Table 9. The choice of best hyperparameter for dataset E.
DatasetAICBIC
k ValueΣ Choicek ValueΣ Choice
E16Full-unshared6Full-Unshared
E26Full-unshared6Full-Unshared
E36Full-unshared6Full-Unshared
E46Full-unshared6Full-Unshared
E58Full-unshared8Full-Unshared
E66Full-unshared6Full-Unshared
E77Full-unshared7Full-unshared
E85Full-unshared5Full-unshared
E96Full-unshared6Full-Unshared
E105Diagonal-unshared5Diagonal-unshared
Table 10. The parameter values for Gaussian mixture-based model (GMM).
Table 10. The parameter values for Gaussian mixture-based model (GMM).
ParameterValue
Number of GMM components/clusters6
Covariance matrixFull
Shared covarianceFalse
Grid length500
Number of iterations for the EM algorithm1000
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mas’ud, A.A.; Sundaram, A.; Ardila-Rey, J.A.; Schurch, R.; Muhammad-Sukki, F.; Bani, N.A. Application of the Gaussian Mixture Model to Classify Stages of Electrical Tree Growth in Epoxy Resin. Sensors 2021, 21, 2562. https://0-doi-org.brum.beds.ac.uk/10.3390/s21072562

AMA Style

Mas’ud AA, Sundaram A, Ardila-Rey JA, Schurch R, Muhammad-Sukki F, Bani NA. Application of the Gaussian Mixture Model to Classify Stages of Electrical Tree Growth in Epoxy Resin. Sensors. 2021; 21(7):2562. https://0-doi-org.brum.beds.ac.uk/10.3390/s21072562

Chicago/Turabian Style

Mas’ud, Abdullahi Abubakar, Arunachalam Sundaram, Jorge Alfredo Ardila-Rey, Roger Schurch, Firdaus Muhammad-Sukki, and Nurul Aini Bani. 2021. "Application of the Gaussian Mixture Model to Classify Stages of Electrical Tree Growth in Epoxy Resin" Sensors 21, no. 7: 2562. https://0-doi-org.brum.beds.ac.uk/10.3390/s21072562

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop