Next Article in Journal
Coinductive Natural Semantics for Compiler Verification in Coq
Next Article in Special Issue
Assessing and Comparing COVID-19 Intervention Strategies Using a Varying Partial Consensus Fuzzy Collaborative Intelligence Approach
Previous Article in Journal
Numerical Solution of Stieltjes Differential Equations
Previous Article in Special Issue
Using the Interval Type-2 Fuzzy Inference Systems to Compare the Impact of Speed and Space Perception on the Occurrence of Road Traffic Accidents
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ridge Fuzzy Regression Modelling for Solving Multicollinearity

1
Department of Statistics, North Carolina State University, Raleigh, NC 27695, USA
2
Department of Applied Mathematics, Hanyang University, Gyeonggi-do 15588, Korea
*
Author to whom correspondence should be addressed.
Submission received: 14 August 2020 / Revised: 5 September 2020 / Accepted: 10 September 2020 / Published: 12 September 2020
(This article belongs to the Special Issue Applications of Fuzzy Optimization and Fuzzy Decision Making)

Abstract

:
This paper proposes an α -level estimation algorithm for ridge fuzzy regression modeling, addressing the multicollinearity phenomenon in the fuzzy linear regression setting. By incorporating α -levels in the estimation procedure, we are able to construct a fuzzy ridge estimator which does not depend on the distance between fuzzy numbers. An optimized α -level estimation algorithm is selected which minimizes the root mean squares for fuzzy data. Simulation experiments and an empirical study comparing the proposed ridge fuzzy regression with fuzzy linear regression is presented. Results show that the proposed model can control the effect of multicollinearity from moderate to extreme levels of correlation between covariates, across a wide spectrum of spreads for the fuzzy response.

1. Introduction

Often times in practical applications, the available data may not always be precise. The researcher may be only accessible to minimum and maximum values of data. Sometimes the data may not even be given in numbers. For instance, consider linguistics data such as “young”, “tall”, or “high”, and medicine data such as “healthy” and “not healthy”. In such cases where the given data are imprecise and vague, classical representation of numbers may be insufficient. The fuzzy set theory introduced by Zadeh [1,2] can handle such uncertainty in data. In the view of fuzzy set theory, uncertain data are what is called fuzzy. Fuzzy data are prevalent in various fields such as linguistics, survey, medicine and so forth [3,4,5,6,7]. The development of fuzzy set theory has led to statistical methods for analyzing fuzzy data. When the measure of indeterminacy is needed, the neutrosophic set introduced by Smarandache [8] considered the measure of indeterminacy in addition to the fuzzy set. The neutrosophic statistics based on the the neutrosophic set can be applied for the analysis of the data when data are selected from the population with uncertain, fuzzy, and imprecise observations [9].
In 1982, Tanaka et al. [10] proposed the fuzzy linear regression model which generalizes the usual linear regression model to fuzzy data. Fuzzy regression models have been since then widely used to analyze fuzzy data [11,12,13,14,15,16].
In classical linear regression models, the multicollinearity phenomenon is frequently observed in which two or more explanatory variables are highly linearly related. Common examples of collinear covariates are: a person’s height and weight, a person’s level of education, gender, race, and starting salary. When multicollinearity occurs, the least squares estimator may not be obtainable or be subject to very high variance. Once the researcher identifies the collinear variables, there are several techniques the researcher can use to handle multicollinearity. Among these techniques, the two most widely used approaches are lasso regression and ridge regression. Lasso regression developed by Tibshirani [17] and ridge regression developed by Hoerl and Kennard [18] improve model performance by adding a penalty term to the classical linear regression model. Both methods aim to shrink the model parameters towards zero. This induces a sparse model which increases the model bias, but decreases the model variance even more, thus improving overall performance. Ridge regression decreases the parameters of low contributing variables towards zero, but not exactly to zero, and stabilizing the parameter variance of the least squares estimator in the presence of multicollinearity. Lasso regression sets the model parameters exactly to zero, removing low contributing variables as well as improving model fitting. However, sometimes the researcher may want to include all the available covariates in the model without having to reduce the dimension of the data. In such cases, ridge regression is preferred to lasso regression.
Similar to classical linear regression models, multicollinearity occurs frequently in fuzzy linear regression models as well, causing problems in the estimation procedure. Often times the number of covariates is not particularly large for fuzzy data. Consequently, dropping any explanatory variables may not be an option. As in the classical statistical setting, we prefer to use ridge regression to lasso regression to handle multicollinearity in such datasets. In this paper, we incorporate fuzzy set theory with ridge regression developed by Hoerl and Kennard [18] to handle multicollinearity observed in fuzzy data. Only some works have suggested ridge estimation methods for fuzzy linear regression, and are limited to obtaining fuzzy ridge estimators which are dependent on the distance between fuzzy numbers [19,20,21]. We instead propose an α -level estimation algorithm for ridge fuzzy regression modelling. The proposed algorithm is an extension of the ridge regression model introduced in Choi et al. [22]. By applying α -levels to the estimation procedure, we are able to construct a fuzzy ridge estimator which does not depend on the distance between fuzzy numbers. Simulation experiments show the proposed ridge fuzzy regression model can solve moderate to severe degrees of multicollinearity across a wide range of spreads for the fuzzy response. An empirical study using Tanaka’s house prices data [10] with multicollinearity, the most widely applied data set in the fuzzy linear regression literature, is conducted to demonstrate the practical implementations.
The rest of this paper is organized as follows. Section 2 introduces key definitions and results from fuzzy set theory. Section 3 describes the classical ridge regression, followed by a step-by-step procedure for the proposed α -level estimation algorithm of ridge fuzzy regression modeling. Section 4 and Section 5 illustrates the performance of the model with simulation studies and a numerical example, respectively. Section 6 concludes the paper.

2. Fuzzy Numbers

A fuzzy set is a set of ordered pairs A = { ( x , μ A ( x ) ) : x X } where μ A ( x ) : X [ 0 , 1 ] is a membership function which represents the degree of membership of x in a set A. Please note that when A is a crisp (classical) set, its membership function can take only the values one or zero depending on whether or not x does or does not belong to A. In this case, μ A ( x ) reduces to the indicator function I A ( x ) of a set A. For any α in [ 0 , 1 ] , the α -level set of a fuzzy set A is a crisp set A ( α ) = { x X : μ A ( x ) α } which contains all the elements in X with membership value in A greater than or equal to α . The α -level set of a fuzzy set A can also be represented by A ( α ) = [ l A ( α ) , r A ( α ) ] . Here l A ( α ) and r A ( α ) are the left and right end-points of the α -level set, respectively. Zadeh’s [23] resolution identity theorem states that a fuzzy set can represented by its membership function or by its α -level set. Let A be a fuzzy number with membership function μ A ( x ) and α -cut A ( α ) . Then we have μ A ( x ) = S u p { α · I A ( α ) ( x ) : α [ 0 , 1 ] } .
A fuzzy number is a normal and convex subset of the real line R with bounded support. The support of a fuzzy set A is defined by s u p p ( A ) = { x R | μ A ( x ) > 0 } . The following parametric class of fuzzy numbers, the so-called L R -fuzzy numbers denoted by A = ( a m , s l , s r ) L R , is often used as a special case:
μ A ( x ) = L a m x s l if x a m R x a m s r if x > a m for x R
where L , R : R [ 0 , 1 ] are fixed, left-continuous, and non-increasing functions with R ( 0 ) = L ( 0 ) = 1 , and R ( 1 ) = L ( 1 ) = 0 . L and R are called left and right shape functions of A, respectively. a m is the mean value of A, and s l , s r > 0 are each the left and right spreads of A. The spreads s l and s r represent the fuzziness of the fuzzy number and can be symmetric or asymmetric. If s l = s r = 0 , the L R -fuzzy number becomes a precise real number with no fuzziness. Thus, a precise real number can be considered to be a special case of a fuzzy number. For a precise observation a R , its corresponding membership function is μ a ( x ) = 1 .
In the fuzzy set theory, triangular and trapezoidal fuzzy numbers are special cases of L R -fuzzy numbers and are used extensively [24]. The membership function of a triangular fuzzy number A = ( a l , a m , a r ) T is given by
μ A ( x ) = x a l a m a l if x a m a r x a r a m if x > a m for x R
where a l , a m , and a r are the left end-point, mid-point, and right end-point, respectively.

3. Ridge Fuzzy Regression

In this section, we propose the α -level estimation algorithm for the proposed ridge fuzzy regression model. This algorithm modifies the method based on Choi et al. [22] to estimate the fuzzy parameters. The term α -level estimation indicates that our algorithm uses α -levels to describe fuzzy data. By using α -level, we are able to develop a ridge fuzzy estimator which is not restricted to the distance between fuzzy numbers. We first briefly examine the original formulation of ridge regression model for crisp data.

3.1. Ridge Regression

Given a data set { y i , x i 1 , x i 2 , , x i p } i = 1 N , a multiple linear regression model assumes that the relationship between a dependent variable y i , i = 1 , , N and a set of explanatory variables x i 1 , x i 2 , , x i p , i = 1 , , N is linear. The model takes the form
y i = β 0 + β 1 x i 1 + β p x i p + ϵ i = X i t β i = 1 , , N
or written alternatively in matrix notation as Y = X β + ϵ . A vector Y = ( y 1 , , y N ) t is a vector of observations on the dependent variable, X = ( X 1 t , , X N t ) t is a matrix of explanatory variables, β = ( β 0 , β 1 , , β p ) t is a vector of regression coefficients to be estimated, and ϵ = ( ϵ 1 , , ϵ N ) t is a vector of error terms. The standard estimator for β is the least squares estimator defined by
β ^ = ( X t X ) 1 X t Y .
In the presence of multicollinearity, i.e., in state of extreme correlations among the explanatory variables β ^ is poorly determined and susceptible to high variance. Thus, we may deliberately bias the regression coefficient estimates so as to control their variance. In this manner, the ridge regression estimator was introduced by Hoerl and Kennard [18] as a penalized least squares estimator. It is achieved by minimizing the residual sum of squares (RSS) subject to a constraint on the size of the estimated coefficient vector [25]:
R S S ( λ ) = ( Y X β ) t ( Y X β ) + λ β t β .
Here λ 0 is a shrinkage parameter which controls the size of the coefficients. The larger the value of λ , the greater the amount of shrinkage, and we have coefficients close to zero. The smaller the value of λ is close to 0, we obtain the least squares solutions. Please note that by convention the input matrix X is assumed to be standardized and Y is assumed to be centered before solving RSS( λ ). The ridge regression solution is
β ^ ridge = ( X t X + λ I ) 1 X t Y
where I is the p × p identity matrix. The shrinkage parameter λ is usually selected via K-fold cross validation. Cross validation is a simple and powerful tool often used to calculate the shrinkage parameter and the prediction error in ridge regression. The entire dataset is divided into K parts, and trains the model on all but the kth part. The model is validated on the k t h part, iterating for all k = 1 , , K . The choice of K is K = 5 or K = 10 in general.

3.2. Ridge Fuzzy Regression Algorithm

Let us consider a set of observations
{ y i , x i 1 , x i 2 , , x i p } i = 1 N = { ( y i l , y i m , y i r ) T , ( x 1 l , x 1 m , x 1 r ) T , , ( x p l , x p m , x p r ) T } i = 1 N
where the dependent variable y i , i = 1 , , N and the explanatory variables x i 1 , x i 2 , , x i p , i = 1 , , N are triangular fuzzy numbers. We assume a linear relationship between the dependent and explanatory variables:
y i = A 0 A 1 x i 1 A p x i p ϵ i , i = 1 , , N
where { A j } j = 0 p = { ( A j l , A j m , A j r ) T } j = 0 p are the fuzzy regression parameters and { ϵ i } i = 1 N = { ( ϵ i l , ϵ i m , ϵ i r ) T } i = 1 N are the fuzzy error terms. ⊕ and ⊙ represent addition and multiplication between two fuzzy numbers, respectively. Often the N equations are stacked together and written in matrix notation as
Y = X A ϵ .
For more details on arithmetic operations between fuzzy numbers, see [10,26]. Please note that the above fuzzy variables can be symmetric or asymmetric, and be extended to various forms such as normal, parabolic, or square root fuzzy data. Since crisp sets are a special case of fuzzy sets, fuzzy inputs and fuzzy outputs, or fuzzy inputs and crisp outputs combinations are also possible. For illustration purposes, in this section, we present our ridge fuzzy regression model using triangular membership functions.
We divide the given data into training and test sets. The model is computed from the training set { y i , x i 1 , x i 2 , , x i p } i = 1 n , and later its performance is evaluated on the test set { y i , x i 1 , x i 2 , , x i p } i = 1 m . Note again that N is the total number of observations, n is the number of observations for the training set, and m is the number of observations for the test set, such that n + m = N . We fit our ridge fuzzy regression model on the training set by the following estimation algorithm:
Step 1: Create α -level sets of the triangular fuzzy input and output as illustrated in Figure 1. For any α -level in [ 0 , 1 ] ,
y i ( α ) = [ l y i ( α ) , r y i ( α ) ] = [ ( α 1 ) s i l y + y i m , ( α 1 ) s i r y + y i m ]
x i j ( α ) = [ l x i j ( α ) , r x i j ( α ) ] = [ ( α 1 ) s i j l x + x i j m , ( α 1 ) s i j r x + x i j m ] i = 1 , , n , j = 1 , , p
where s i l y , s i r y , s i j l x , s i j r x 0 are the left and right spreads of the dependent and explanatory variables, respectively. The α -levels are denoted by the sequence ( α k ) k = 0 K for some K with α k [ 0 , 1 ] .
Step 2: Perform ridge regression of Y ( α k ) on X ( α k ) for each k = 0 , , K . Find the intermediate estimators l A ¯ ( α k ) and r A ¯ ( α k ) of l A ( α k ) and r A ( α k ) by minimizing the following respective ridge loss functions (see Figure 2).
( l Y ( α k ) l X ( α k ) l A ( α k ) ) t ( l Y ( α k ) l X ( α k ) l A ( α k ) ) + λ l A ( α k ) t l A ( α k ) ( r Y ( α k ) r X ( α k ) r A ( α k ) ) t ( r Y ( α k ) r X ( α k ) r A ( α k ) ) + λ r A ( α k ) t r A ( α k )
We assume the endpoints of the α -level set of Y has been centered and the endpoints of α -level set of X has been standardized as is by convention in classical ridge regression [25].
Step 3: Obtain the estimators l A ˜ ( α k ) and r A ˜ ( α k ) of l A ( α k ) and r A ( α k ) by modifying the intermediate estimators l A ¯ ( α k ) and r A ¯ ( α k ) so that the estimated coefficients form the membership function of a triangular fuzzy number. For this the following operations are performed (see Figure 3).
l A ˜ ( α K ) = r A ˜ ( α K ) = l A ¯ ( α K ) = r A ¯ ( α K ) l A ˜ ( α k ) = min { l A ¯ ( α k ) , l A ˜ ( α k + 1 ) } k = K 1 , K 2 , , 0 r A ˜ ( α k ) = max { r A ¯ ( α k ) , r A ˜ ( α k + 1 ) } k = K 1 , K 2 , , 0
Step 4: Estimate the triangular fuzzy coefficient A ^ = ( A l ^ , A m ^ , A r ^ ) T and its membership function μ A ^ by fitting a linear regression line through l A ˜ ( α k ) and r A ˜ ( α k ) for k = 0 , , K , respectively. A constraint is given so that μ A ^ satisfy the condition of μ A ^ ( l A ^ ( 1 ) ) = μ A ^ ( r A ^ ( 1 ) ) = 1 (see Figure 4).
Step 5: Symmetric fuzzy inputs or outputs do not always guarantee that the estimated membership function μ A ^ will also be symmetric. To reduce the difference between the true values with the fitted values we consider the following candidates:
μ A ^ ( x ) = x l A ^ ( 0 ) l A ^ ( 1 ) l A ^ ( 0 ) if x l A ^ ( 1 ) = r A ^ ( 1 ) r A ^ ( 0 ) x r A ^ ( 0 ) r A ^ ( 1 ) if x > l A ^ ( 1 ) = r A ^ ( 1 ) for x R
or
μ A ^ ( x ) = x l A ^ ( 0 ) l A ^ ( 1 ) l A ^ ( 0 ) if x l A ^ ( 1 ) l A ^ ( 0 ) + x l A ^ ( 1 ) l A ^ ( 0 ) if x > l A ^ ( 1 ) for x R
where r A ^ ( 0 ) is chosen as r A ^ ( 0 ) = 2 × l A ^ ( 1 ) l A ^ ( 0 ) , or
μ A ^ ( x ) = r A ^ ( 0 ) + x r A ^ ( 0 ) r A ^ ( 1 ) if x r A ^ ( 1 ) r A ^ ( 0 ) x r A ^ ( 0 ) r A ^ ( 1 ) if x > r A ^ ( 1 ) for x R
where l A ^ ( 0 ) is chosen as l A ^ ( 0 ) = 2 × r A ^ ( 1 ) r A ^ ( 0 ) .
We present two performance criteria based on Diamond’s fuzzy distance measure [27] to evaluate the proposed fuzzy estimators. Denote the dependent variable as y i = ( y i l , y i m , y i r ) T , i = 1 , , n , and its predicted value as y i ^ = ( y i l ^ , y i m ^ , y i r ^ ) T = ( X i l t A l ^ , X i m t A m ^ , X i r t A r ^ ) T , i = 1 , , n . Here n is the number of observations for the training set. We defined RMSE F (root mean square error for fuzzy numbers) and MAPE F (mean absolute percentage error for fuzzy numbers) as below.
RMSE F = 1 n i = 1 n { ( y i l y i l ^ ) 2 + ( y i m y i m ^ ) 2 + ( y i r y i r ^ ) 2 }
MAPE F = 100 % n i = 1 n ( | y i l y i l ^ | + | y i m y i m ^ | + | y i r y i r ^ | )
Compute the RMSE F for each of the membership functions, then select the one which minimizes the criterion.
Step 6: Repeat Steps 1–5 for selected α -level sequences ( α k ) k = 0 K with α k equally spaced between 0 and 1. Choose the optimal set of α -levels which minimizes RMSE F . Finally, compute the fuzzy ridge coefficient estimate A ^ based on that selected sequence.

4. Simulation Study

A simulation study was conducted to illustrate the performance of the proposed ridge fuzzy regression model in the presence of multicollinearity. Simulation results are compared with the fuzzy linear regression model with varying degrees of correlation. The fuzzy least squares estimator is obtained by setting the tuning parameter λ as zero in Step 2 of Section 3.2.
We generated N = 100 observations for each of the p = 4 crisp explanatory variables. The number of data dimensions is in line with commonly found fuzzy data. Following Gibbons [28], the explanatory variables x i j are generated by
x i j = ( 1 ρ 2 ) 1 / 2 z i j + ρ z i p , i = 1 , , n j = 1 , , p
where ρ is a given constant and z i j are generated from independent normal distributions with mean 50 and variance 1. Here x i j are assumed to be non-negative so as to reflect the non-negative characteristics of real world fuzzy data. The degree of linear association between explanatory variables is controlled via ρ , where in this case is the correlation between any two explanatory variables is ρ 2 . Three different sets of correlation are considered corresponding to ρ = 0.8 , 0.9 , and 0.99 . Each value of ρ stands for moderate, high, and very high correlation between the variables. Observations on the fuzzy dependent variable are generated by
y i l = A 0 l + A 1 l x i 1 + + A p l x i p + ϵ i 1 y i m = A 0 m + A 1 m x i 1 + + A p m x i p + ϵ i 2 y i r = A 0 r + A 1 r x i 1 + + A p r x i p + ϵ i 3 , i = 1 , , N
where ϵ i 1 , ϵ i 2 , ϵ i 3 are generated from independent normal distributions with mean 0 and variance σ 2 . Four values of σ are investigated in this study: 0.5, 1, 1.5, and 2. Large values of σ correspond to bigger variation in the spreads of the fuzzy dependent variable. S l y = ( s 1 l y , , s N l y ) t the vector of left spreads and S r y = ( s 1 r y , , s N r y ) t the vector of right spreads are determined by
s i l y = y i m y i l s i r y = y i r y i m i = 1 , , N .
Cases of asymmetric spreads, S l y S r y , and symmetric spreads, S l y = S r y are also compared. The supposed parameters of the model are: A l = ( 0 , 0.1 , 0.15 , 0.2 , 0.25 ) , A m = ( 0 , 0.4 , 0.45 , 0.5 , 0.55 ) , and A r = ( 0 , 0.7 , 0.75 , 0.8 , 0.85 ) . In order to analyze the effects of factors ρ and σ , we controlled for the effects of varying α -level sequences in Step 6 of Section 3.2. For both models we fixed the α -level sequence as (0, 0.25, 0.5, 0.75, 1).
200 replicates for each scenario are generated. The explanatory variables and the fuzzy coefficients remain fixed, while the error terms and hence the fuzzy dependent variable changes. We separated the simulated data into training and test sets. Once the ridge fuzzy regression model and the fuzzy linear regression model are fit to the training data, RMSE F and MAPE F are computed from the test set for t = 1 , , 200 replicates. Let RMSE F t and MAPE F t be the performance measures when the fuzzy model is applied to the replicate t. The following quantities are then computed for each fuzzy estimator:
Ave . RMSE F = 1 200 t = 1 200 RMSE F t
Ave . MAPE F = 1 200 t = 1 200 MAPE F t .
In addition, we fit the ridge regression model and the linear regression model on the mid-point of our training data { y i m , x i 1 m , x i 2 m , , x i p m } i = 1 n for comparison with fuzzy methods. The test Ave. RMSE and Ave. MAPE values of 200 replicates are recorded for both models. The output from numerical experiments is suggested below in Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6. Measures of performance are summarized for all combinations of factors ρ , σ and whether the fuzzy output is symmetric or not. The following remarks can be made on the basis of Table 1, Table 2, Table 3, Table 4, Table 5 and Table 6:
  • Ave. RMSE and Ave. MAPE do not depend on whether the spreads are symmetric or not as they are computed from the mid-points of the generated data. Ridge regression achieves smaller Ave. RMSE than linear regression in all cases. Ridge regression achieves smaller or nearly equal Ave. MAPE with linear regression in all cases. If the Ave. RMSE of ridge regression is smaller than linear regression a similar pattern is observed for Ave . RMSE F values. This relationship is observed for Ave. MAPE and Ave . MAPE F as well.
  • Ave . RMSE F increases as σ increases for both models. As σ and ρ increases, the Ave . RMSE F difference between ridge fuzzy regression and fuzzy linear regression increases as well. Ave . RMSE F values are larger when the spreads are symmetric. In all scenarios, ridge fuzzy regression Ave . RMSE F values almost always outperform those of fuzzy linear regression.
  • Ave . MAPE F exhibit near identical patterns with Ave . RMSE F . For both ridge fuzzy regression and fuzzy linear regression, Ave . MAPE F is larger for bigger σ values. The difference between the two models increases as σ and ρ increases. When the fuzzy dependent variable is symmetric the Ave . MAPE F values are larger than when it is asymmetric. Ave . MAPE F is in general lower for ridge fuzzy regression than fuzzy linear regression for all σ and ρ combinations and asymmetric, symmetric outputs.

5. Empirical Study

In this section, we demonstrate the performance of the proposed ridge fuzzy regression model on an illustrative example taken from Tanaka [10]. The performance of the ridge fuzzy regression estimator is compared with the fuzzy least squares estimator for crisp explanatory variables and a fuzzy dependent variable. The linear regression fuzzy model from Tanaka [10] is further compared to illustrate the performance of the ridge fuzzy regression model. For both the ridge fuzzy regression and the linear fuzzy model, the α -level sequences α k = r × k , k = 0 , , K for some r and K are chosen as candidates for Step 6 of the estimation algorithm in Section 3.2. The list of α -level sequences is presented in Table 7.

Example: House Prices Data

Tanaka et al. [10] presents a data set concerning the price mechanism of prefabricated houses. The relationship between five crisp inputs (rank of material, first floor space (m2), second first floor space (m2), number of rooms and number of Japanese-style rooms) and a fuzzy output (house price) is investigated. The complete data is shown in Table 8. The fitted values for the ridge fuzzy model and the linear fuzzy model is shown in Table 9. Results show the predicted values from the ridge fuzzy regression more accurately describes the original data than fuzzy linear regression. This is again clarified in Figure 5. In the triangular fuzzy plot of the observed and fitted values, a comparison of the two models is shown. The black triangles correspond to the observed values, the red triangles in Figure 5a to the ridge fuzzy fitted values, and the blue triangles in Figure 5b to the fuzzy linear fitted values. Both methods estimated the mid-points of the fuzzy dependent variable well. The spreads however, are shorter for the proposed ridge fuzzy regression than the other. The fitted equation for the ridge fuzzy regression is given by
Y ^ = ( 1839.23 , 1156.78 , 474.32 ) + ( 1874.56 , 1874.56 , 1874.56 ) x 1 + ( 73.73 , 75.29 , 76.85 ) x 2 + ( 59.04 , 65.57 , 72.10 ) x 3 + ( 149.93 , 149.93 , 149.93 ) x 4 + ( 543.50 , 587.74 , 631.98 ) x 5
and for the fuzzy linear regression, the fitted equation is
Y ^ = ( 2038.12 , 1129.61 , 221.09 ) + ( 2386.56 , 2386.56 , 2386.56 ) x 1 + ( 87.13 , 93.37 , 99.60 ) x 2 + ( 71.72 , 82.13 , 92.54 ) x 3 + ( 376.35 , 376.35 , 376.35 ) x 4 + ( 285.03 , 188.25 , 91.48 ) x 5 .
Please note that the fitted equation for the linear regression fuzzy model shown in Tanaka et al. [10] is
Y ^ = ( 10 , 220 , 11 , 040 , 11 , 860 ) + ( 1810 , 1810 , 1810 ) x 1 + ( 1770 , 2140 , 2510 ) x 2 + ( 870 , 870 , 870 ) x 3 + ( 540 , 540 , 540 ) x 4 + ( 180 , 180 , 180 ) x 5 .
An analysis of the α -level sequences used in Step 6 of the estimation algorithm is presented in Figure 6. The α -level sequence which minimizes RMSE F was chosen as the optimal α -level sequence for each of the models. The red dots in Figure 6a,b each indicate the chosen α -level sequence based on RMSE F . For the ridge fuzzy regression, α k = r × k , k = 0 , , K with r = 0.01 , K = 100 was chosen. In the case of fuzzy linear regression, r = 0.5 , K = 2 was selected.
In Table 10 the performance measures RMSE F and MAPE F ridge fuzzy regression are compared with the fuzzy linear regression model and the linear regression fuzzy model from Tanaka et al. [10]. Clearly both measures are greatly reduced for the ridge fuzzy regression compared to the other models, suggesting that the proposed ridge fuzzy regression model provides a better fit of the data in comparison to the two methods.

6. Conclusions

This paper proposes an α -level estimation algorithm for ridge fuzzy regression modeling, extending the ridge regression model introduced in Choi et al. [22]. As shown in simulation studies and an empirical study, the proposed ridge fuzzy regression model can handle fuzzy data sets with crisp inputs and triangular fuzzy outputs. The same procedure is available with fuzzy inputs and fuzzy outputs, or fuzzy inputs and crisp outputs. In previous works, estimation methods for ridge fuzzy regression depend on the distance between fuzzy numbers. By incorporating α -levels to ridge fuzzy regression, we are able to construct the ridge fuzzy estimator without having to define the distance between fuzzy numbers. Simulation results show the ridge fuzzy regression model reduces the effect of multicollinearity over a wide range of spreads for the fuzzy response, for various levels of correlation between inputs. In the illustrative example taken from Tanaka et al. [10], we have shown the practical implementations of our method. Comparison is made with fuzzy linear regression with respect to RMSE and MAPE for fuzzy numbers. Overall these results demonstrate the effectiveness of ridge regression in fuzzy data.
An importance point to note is that typically ridge regression is preferred over lasso regression when the objective of research is to handle multicollinearity while not wanting to remove low contributing variables. However, when the dimension of the data is large and dropping collinear variables is necessary, one may use lasso regression rather than ridge regression. To manage such cases, in future studies we plan to extend the proposed α -level estimation algorithm for ridge fuzzy models to lasso fuzzy regression models. Lasso fuzzy regression will be especially useful for modeling correlated genetic data sets.
In addition, the present study can be extended for neutrosophic statistics [9,29,30,31,32,33,34,35] as future research.

Author Contributions

Conceptualization, H.K. and H.-Y.J.; methodology, H.-Y.J.; software, H.K.; validation, H.K. and H.-Y.J.; formal analysis, H.K.; investigation, H.K.; resources, H.K.; data curation, H.K.; writing–original draft preparation, H.K.; writing–review and editing, H.K. and H.-Y.J.; visualization, H.K.; supervision, H.-Y.J.; project administration, H.-Y.J.; funding acquisition, H.-Y.J. Both authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No.2017R1C1B1005069, No.2019R1I1A1A01046810).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zadeh, L. Fuzzy sets. Inf. Control 1965, 8, 338–353. [Google Scholar] [CrossRef] [Green Version]
  2. Zadeh, L. Probability measures of fuzzy events. J. Math. Anal. Appl. 1968, 23, 421–427. [Google Scholar] [CrossRef] [Green Version]
  3. Barro, S.; Marin, R. Fuzzy Logic in Medicine; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  4. Bellamy, J.E. Medical diagnosis, diagnostic spaces, and fuzzy systems. J. Am. Vet. Med. Assoc. 1997, 210, 390–396. [Google Scholar] [CrossRef] [PubMed]
  5. Jung, H.Y.; Choi, H.; Park, T. Fuzzy heaping mechanism for heaped count data with imprecision. Soft Comput. 2018, 22, 4585–4594. [Google Scholar] [CrossRef]
  6. Jung, H.Y.; Leem, S.; Lee, S.; Park, T. A novel fuzzy set based multifactor dimensionality reduction method for detecting gene-gene interaction. Comput. Biol. Chem. 2016, 65, 193–202. [Google Scholar] [CrossRef]
  7. Jung, H.Y.; Leem, S.; Park, T. Fuzzy set-based generalized multifactor dimensionality reduction analysis of gene-gene interactions. BMC Med. Genom. 2018, 11. [Google Scholar] [CrossRef]
  8. Smarandache, F. A Unifying Field in Logics: Neutrosophic Logic. Neutrosophy, Neutrosophic Set, Neutrosophic Probability: Neutrsophic Logic. Neutrosophy, Neutrosophic Set, Neutrosophic Probability; Infinite Study: Austin, TX, USA, 2005. [Google Scholar]
  9. Aslam, M.; Arif, O.H.; Sherwani, R.A.K. New diagnosis test under the neutrosophic statistics: An application to diabetic patients. BioMed Res. Int. 2020, 2020, 2086185. [Google Scholar] [CrossRef] [Green Version]
  10. Tanaka, H.; Uejima, S.; Asai, K. Linear regression analysis with fuzzy model. IEEE Trans. Syst. Man Cybern. 1982, 12, 903–907. [Google Scholar] [CrossRef]
  11. Chang, P.T.; Lee, E. Fuzzy least absolute deviations regression and the conflicting trends in fuzzy parameters. Comput. Math. Appl. 1994, 28, 89–101. [Google Scholar] [CrossRef]
  12. Choi, S.H.; Jung, H.Y.; Lee, W.J.; Yoon, J.H. Fuzzy regression model with monotonic response function. Commun. Korean Math. Soc. 2018, 33, 973–983. [Google Scholar] [CrossRef]
  13. Icen, D.; Gunay, S. Design and implementation of the fuzzy expert system in Monte Carlo methods for fuzzy linear regression. Appl. Soft Comput. 2019, 77, 399–411. [Google Scholar] [CrossRef]
  14. Jung, H.Y.; Yoon, J.H.; Choi, S.H. Fuzzy linear regression using rank transform method. Fuzzy Sets Syst. 2015, 274, 97–108. [Google Scholar] [CrossRef]
  15. Lee, W.J.; Jung, H.Y.; Yoon, J.H.; Choi, S.H. The statistical inferences of fuzzy regression based on bootstrap techniques. Appl. Soft Comput. 2015, 19, 883–890. [Google Scholar] [CrossRef]
  16. Sohn, S.Y.; Kim, D.H.; Yoon, J.H. Technology credit scoring model with fuzzy logistic regression. Appl. Soft Comput. 2016, 43, 150–158. [Google Scholar] [CrossRef]
  17. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. (Methodol.) 1996, 58, 267–288. [Google Scholar] [CrossRef]
  18. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 1970, 12, 55–67. Available online: http://0-www-jstor-org.brum.beds.ac.uk/stable/1267351 (accessed on 14 May 2020). [CrossRef]
  19. Hong, D.H.; Hwang, C. Ridge Regression Procedures For Fuzzy Models Using Triangular Fuzzy Numbers. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2004, 12, 145–159. [Google Scholar] [CrossRef]
  20. Hong, D.H.; Hwang, C.; Ahn, C. Ridge estimation for regression models with crisp inputs and Gaussian fuzzy output. Fuzzy Sets Syst. 2004, 142, 307–319. [Google Scholar] [CrossRef]
  21. Donoso, S.; Marin, N.; Vila, M.A. Quadratic Programming Models for Fuzzy Regression. In Proceedings of the International Conference on Mathematical and Statistical Modeling in Honor of Enrique Castillo, Ciudad Real, Spain, 28–30 June 2006. [Google Scholar]
  22. Choi, S.H.; Jung, H.Y.; Kim, H. Ridge Fuzzy Regression Model. Int. J. Fuzzy Syst. 2019. [Google Scholar] [CrossRef]
  23. Zadeh, L. The concept of a linguistic variable and its application to approximate reasoning—I. Inf. Sci. 1975, 8, 199–249. [Google Scholar] [CrossRef]
  24. Dubois, D.; Prade, H. Fuzzy Sets and Systems: Theory and Applications; Academic Press: Cambridge, MA, USA, 1980. [Google Scholar]
  25. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  26. Tanaka, H.; Lee, H. Interval regression analysis by quadratic programming approach. IEEE Trans. Fuzzy Syst. 1998, 6, 473–481. [Google Scholar] [CrossRef]
  27. Diamond, P. Fuzzy least squares. Inf. Sci. 1988, 46, 141–157. [Google Scholar] [CrossRef]
  28. Gibbons, D.G. A Simulation Study of Some Ridge Estimators. J. Am. Stat. Assoc. 1981, 76, 131–139. [Google Scholar] [CrossRef]
  29. Aslam, M.; Albassam, M. Application of neutrosophic logic to evaluate correlation between prostate cancer mortality and dietary fat assumption. Symmetry 2019, 11, 330. [Google Scholar] [CrossRef] [Green Version]
  30. Aslam, M. Introducing Kolmogorov–Smirnov tests under uncertainty: An application to radioactive data. ACS Omega 2019, 5, 914–917. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Aslam, M.; Al Shareef, A.; Khan, K. Monitoring the temperature through moving average control under uncertainty environment. Sci. Rep. 2020, 10, 1–8. [Google Scholar] [CrossRef] [PubMed]
  32. Aslam, M. On detecting outliers in complex data using Dixon’s test under neutrosophic statistics. J. King Saud-Univ.-Sci. 2020, 32, 2005–2008. [Google Scholar] [CrossRef]
  33. Albassam, M.; Khan, N.; Aslam, M. The W/S test for data having neutrosophic numbers: An application to USA village population. Complexity 2020, 2020, 3690879. [Google Scholar] [CrossRef]
  34. Aslam, M.; Arif, O.H. Test of Association in the Presence of Complex Environment. Complexity 2020, 2020, 2935435. [Google Scholar] [CrossRef]
  35. Aslam, M.; Arif, O.H. Multivariate Analysis under Indeterminacy: An Application to Chemical Content Data. J. Anal. Methods Chem. 2020, 2020, 1406028. [Google Scholar] [CrossRef]
Figure 1. Fuzzy output data.
Figure 1. Fuzzy output data.
Mathematics 08 01572 g001
Figure 2. Intermediate estimators l A j ¯ ( α k ) and l A j ¯ ( α k ) for the α -level sequence (0, 0.25, 0.5, 0.75, 1).
Figure 2. Intermediate estimators l A j ¯ ( α k ) and l A j ¯ ( α k ) for the α -level sequence (0, 0.25, 0.5, 0.75, 1).
Mathematics 08 01572 g002
Figure 3. Modified estimators l A j ˜ ( α k ) and l A j ˜ ( α k ) for the α -level sequence (0, 0.25, 0.5, 0.75, 1).
Figure 3. Modified estimators l A j ˜ ( α k ) and l A j ˜ ( α k ) for the α -level sequence (0, 0.25, 0.5, 0.75, 1).
Mathematics 08 01572 g003
Figure 4. Estimated fuzzy coefficient A j ^ = ( A j l ^ , A j m ^ , A j r ^ ) T and its membership function.
Figure 4. Estimated fuzzy coefficient A j ^ = ( A j l ^ , A j m ^ , A j r ^ ) T and its membership function.
Mathematics 08 01572 g004
Figure 5. The triangular fuzzy plot of observed and fitted values. (a): Ridge fuzzy regression and (b): Fuzzy linear regression.
Figure 5. The triangular fuzzy plot of observed and fitted values. (a): Ridge fuzzy regression and (b): Fuzzy linear regression.
Mathematics 08 01572 g005
Figure 6. Analysis of the α-level sequences in Step 6 of the estimation algorithm. (a): Ridge fuzzy regression and (b): Fuzzy linear regression.
Figure 6. Analysis of the α-level sequences in Step 6 of the estimation algorithm. (a): Ridge fuzzy regression and (b): Fuzzy linear regression.
Mathematics 08 01572 g006
Table 1. The performance measures when ρ = 0.8 and the dependent variable is an asymmetric triangular fuzzy number.
Table 1. The performance measures when ρ = 0.8 and the dependent variable is an asymmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.1.1614.45910.95519.419
Fuzzy Reg.1.7068.97923.13040.243
Ave. R M S E Ridge Reg.0.4780.9571.4371.917
Linear Reg.0.4810.9621.4441.925
Ave. M A P E F Ridge Fuzzy Reg.1.84%7.64%19.26%34.43%
Fuzzy Reg.2.83%15.81%41.14%71.71%
Ave. M A P E Ridge Reg.0.31%0.61%0.91%1.21%
Linear Reg.0.31%0.61%0.92%1.23%
Table 2. The performance measures when ρ = 0.8 and the dependent variable is a symmetric triangular fuzzy number.
Table 2. The performance measures when ρ = 0.8 and the dependent variable is a symmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.3.6048.15019.94422.572
Fuzzy Reg.6.27319.22839.02061.870
Ave. R M S E Ridge Reg.0.4780.9571.4371.917
Linear Reg.0.4810.9621.4441.925
Ave. M A P E F Ridge Fuzzy Reg.5.89%13.64%25.52%38.88%
Fuzzy Reg.10.72%33.72%69.03%109.8%
Ave. M A P E Ridge Reg.0.31%0.61%0.91%1.21%
Linear Reg.0.31%0.61%0.92%1.23%
Table 3. The performance measures when ρ = 0.9 and the dependent variable is an asymmetric triangular fuzzy number.
Table 3. The performance measures when ρ = 0.9 and the dependent variable is an asymmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.1.2104.61211.62520.995
Fuzzy Reg.2.87917.32138.91863.449
Ave. R M S E Ridge Reg.0.4780.9591.4401.921
Linear Reg.0.4810.9621.4441.925
Ave. M A P E F Ridge Fuzzy Reg.2.02%8.29%21.47%39.05%
Fuzzy Reg.5.19%32.26%72.62%118.4%
Ave. M A P E Ridge Reg.0.32%0.64%0.95%1.27%
Linear Reg.0.32%0.64%0.96%1.28%
Table 4. The performance measures when ρ = 0.9 and the dependent variable is a symmetric triangular fuzzy number.
Table 4. The performance measures when ρ = 0.9 and the dependent variable is a symmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.3.3068.35616.03625.076
Fuzzy Reg.8.37830.69160.04191.664
Ave. R M S E Ridge Reg.0.4780.9591.4401.921
Linear Reg.0.4810.9621.4441.925
Ave. M A P E F Ridge Fuzzy Reg.5.60%14.71%28.85%45.47%
Fuzzy Reg.15.23%56.89%111.7%170.6%
Ave. M A P E Ridge Reg.0.32%0.64%0.95%1.27%
Linear Reg.0.32%0.64%0.96%1.28%
Table 5. The performance measures when ρ = 0.99 and the dependent variable is an asymmetric triangular fuzzy number.
Table 5. The performance measures when ρ = 0.99 and the dependent variable is an asymmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.0.9522.3786.82214.514
Fuzzy Reg.34.201101.27171.70243.33
Ave. R M S E Ridge Reg.0.4400.8741.3101.746
Linear Reg.0.4430.8851.3281.771
Ave. M A P E F Ridge Fuzzy Reg.1.80%4.70%14.53%31.64%
Fuzzy Reg.75.07%222.3%377.1%535.0%
Ave. M A P E Ridge Reg.0.37%0.74%1.11%1.49%
Linear Reg.0.38%0.76%1.14%1.52%
Table 6. The performance measures when ρ = 0.99 and the dependent variable is a symmetric triangular fuzzy number.
Table 6. The performance measures when ρ = 0.99 and the dependent variable is a symmetric triangular fuzzy number.
σ :0.51.01.52.0
Ave. R M S E F Ridge Fuzzy Reg.2.4716.26613.19523.077
Fuzzy Reg.55.455142.97233.84325.40
Ave. R M S E Ridge Reg.0.4400.8741.3101.746
Linear Reg.0.4430.8851.3281.771
Ave. M A P E F Ridge Fuzzy Reg.4.70%12.53%27.65%49.26%
Fuzzy Reg.121.5%313.5%513.2%714.9%
Ave. M A P E Ridge Reg.0.37%0.74%1.11%1.49%
Linear Reg.0.38%0.76%1.14%1.52%
Table 7. The list of α -level sequences for Step 6 of the estimation algorithm.
Table 7. The list of α -level sequences for Step 6 of the estimation algorithm.
α k = r × k , k = 0 , , K
No.rK
10.01100
20.0250
30.02540
40.0425
50.0520
60.110
70.156
80.25
90.254
100.33
110.52
1211
Table 8. Houses prices data.
Table 8. Houses prices data.
No. Y = ( y m , S l y = S r y ) x 1 x 2 x 3 x 4 x 5
1(6060, 550)138.0936.4351
2(7100, 50)162.1026.5061
3(8080, 400)163.7644.7171
4(8260, 150)174.5238.0981
5(8650, 750)175.3841.4072
6(8520, 450)252.9926.4942
7(9170, 700)262.9326.4952
8(10,310, 200)272.0433.1263
9(10,920, 600)276.1243.0672
10(12,030, 100)290.2642.6472
11(13,940, 350)385.7031.3363
12(14,200, 250)395.2727.6463
13(16,010, 300)3105.9827.6463
14(16,320, 500)379.2566.8163
15(16,990, 650)3120.5032.2563
Table 9. Fitted values of house prices data.
Table 9. Fitted values of house prices data.
No. Y = ( y m , S l y = S r y ) Ridge Fuzzy Reg.Fuzzy Reg.
1(6060, 550)(5812.32, 1024.08)(5735.21, 1621.95)
2(7100, 50)(6819.04, 996.70)(6785.09, 1668.23)
3(8080, 400)(7988.07, 1118.22)(8059.27, 1868.14)
4(8260, 150)(8214.21, 1091.78)(8143.87, 1866.29)
5(8650, 750)(9233.67, 1158.99)(8684.11, 2002.89)
6(8520, 450)(8894.65, 1026.66)(8884.70, 1708.12)
7(9170, 700)(9493.11, 1042.17)(9436.42, 1,770.07)
8(10,310, 200)(11,051.53, 1143.94)(10,266.90, 1992.65)
9(10,920, 600)(11,272.79, 1170.98)(11,276.09, 2024.78)
10(12,030, 100)(12,309.87, 1190.31)(12,561.81, 2108.54)
11(13,940, 350)(13,837.21, 1153.57)(13,781.85, 2059.16)
12(14,200, 250)(14,315.80, 1144.40)(14,372.32, 2080.39)
13(16,010, 300)(15,122.16, 1161.12)(15,372.29, 2147.15)
14(16,320, 500)(15,677.92, 1375.23)(16,093.51, 2388.31)
15(16,990, 650)(16,517.66, 1213.89)(17,106.59, 2285.64)
Table 10. RMSE F and MAPE F fuzzy performance measures for ridge fuzzy regression, fuzzy linear regression, and the linear regression fuzzy model from Tanaka et al. [10].
Table 10. RMSE F and MAPE F fuzzy performance measures for ridge fuzzy regression, fuzzy linear regression, and the linear regression fuzzy model from Tanaka et al. [10].
RMSEFMAPEF
Ridge Fuzzy Reg.1327.7818%
Fuzzy Reg.2321.2433%
Tanaka et al.349,851.855%

Share and Cite

MDPI and ACS Style

Kim, H.; Jung, H.-Y. Ridge Fuzzy Regression Modelling for Solving Multicollinearity. Mathematics 2020, 8, 1572. https://0-doi-org.brum.beds.ac.uk/10.3390/math8091572

AMA Style

Kim H, Jung H-Y. Ridge Fuzzy Regression Modelling for Solving Multicollinearity. Mathematics. 2020; 8(9):1572. https://0-doi-org.brum.beds.ac.uk/10.3390/math8091572

Chicago/Turabian Style

Kim, Hyoshin, and Hye-Young Jung. 2020. "Ridge Fuzzy Regression Modelling for Solving Multicollinearity" Mathematics 8, no. 9: 1572. https://0-doi-org.brum.beds.ac.uk/10.3390/math8091572

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop