Next Article in Journal
How New HRM Practices, Organizational Innovation, and Innovative Climate Affect the Innovation Performance in the IT Industry: A Moderated-Mediation Analysis
Next Article in Special Issue
A Hybrid Unequal Clustering Based on Density with Energy Conservation in Wireless Nodes
Previous Article in Journal
An Empirical Study on the Key Factors of Intelligent Upgrade of Small and Medium-sized Enterprises in China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Decision-Making Method based on Mixed Integer Linear Programming and Rough Set: A Case Study of Diesel Engine Quality and Assembly Clearance Data

1
School of Reliability and Systems Engineering, Beihang University, Beijing 100191, China
2
Henan Diesel Engine Industry Co., Luoyang 471039, China
*
Author to whom correspondence should be addressed.
These two authors contributed equally to this work.
Sustainability 2019, 11(3), 620; https://0-doi-org.brum.beds.ac.uk/10.3390/su11030620
Submission received: 19 December 2018 / Revised: 18 January 2019 / Accepted: 22 January 2019 / Published: 24 January 2019

Abstract

:
The purpose of this paper is to establish a decision-making system for assembly clearance parameters and machine quality level by analyzing the data of assembly clearance parameters of diesel engine. Accordingly, we present an extension of the rough set theory based on mixed-integer linear programming (MILP) for rough set-based classification (MILP-FRST). Traditional rough set theory has two shortcomings. First, it is sensitive to noise data, resulting in a low accuracy of decision systems based on rough sets. Second, in the classification problem based on rough sets, the attributes cannot be automatically determined. MILP-FRST has the advantages of MILP in resisting noisy data and has the ability to select attributes flexibly and automatically. In order to prove the validity and advantages of the proposed model, we used the machine quality data and assembly clearance data of 29 diesel engines of a certain type to validate the proposed model. Experiments show that the proposed decision-making method based on MILP-FRST model can accurately determine the quality level of the whole machine according to the assembly clearance parameters.

1. Introduction

Diesel engine is the power core of the ship. In the manufacturing process of the diesel engine, the assembly quality affects the performance indexes of the diesel engine, which is an important factor to measure the quality of the whole engine. Previous studies on the relationship between assembly clearance and machine quality have mainly focused on the mechanical principle, while the data mining method is still less used to mine the relationship between assembly clearance and machine quality. With the development of data mining technology and the accumulation of a large number of raw data, it is possible to apply data mining methods to solve this problem.
With the development of computers and the internet, the amount of data is increasing. Data always contain noise; therefore, it is necessary to handle noisy data to obtain accurate results. Some researchers have put forward methods to address this issue. Among these methods, the two classic methods are fuzzy set [1] and evidence theory [2]. However, these methods sometimes require additional information or prior knowledge, such as fuzzy membership functions, basic probability assignment functions and statistical probability distributions, which are not always easy to obtain.
Rough set theory provides a new way to address vagueness and uncertainty [3]. The core concept of rough set theory is to deduce imprecise data, or to find the correlation between different data by representing the given finite set as an upper approximation set or a lower approximation set.
Despite the advantages of rough set theory, some challenges need to be overcome in practical applications. These problems can be classified into two categories: (1) There are certain limitations of a rough set in practical applications. Many extension and correction theories to the classical rough set have been developed. For example, a model integrating distance and partition distance with a rough set on the basis of a rough set based on a similarity relation was proposed [4]. This model also provided a new understanding of the classification criteria of rough set equivalence classes. Φ-Rough set, another extension of the rough set theory based on the similarity relation, was proposed [5]. Φ-Rough set replaces the indiscernibility relation of the crisp rough set theory by the notion of the Φ-approximate equivalence relation. Use of dynamic probabilistic rough sets with incomplete data addressed the challenge of processing such dynamic and incomplete data [6]. Based on a three-way classification of attributes into the pair-wise disjoint sets of core, marginal, and non-useful attributes, the relationships between the corresponding classes of classification-based and class-specific attributes were examined [7]. (2) A rough set is sensitive to noisy data. The accuracy of a decision-making model based on a rough set is low when applied to the analysis of datasets containing noisy data [8]. To strengthen its ability to resist noisy data, the variable precision rough set (in short, VPRS) was proposed [9]. VPRS has been applied in several fields, such as data mining [10], decision systems [11], and expert systems [12], and provides satisfactory results. Similarly, knowledge reduction is an important research direction of VPRS. However, the related methods and theories are not mature. The most popular study based on VPRS is attribute reduction. In addition, variable precision threshold beta is usually determined by experts. Hence, some researchers have proposed the selection method of beta, which can reduce the difficulty of beta determination due to a lack of prior knowledge [13,14]. Zavareh M. and Maggioni V. proposes an approach to analyze water quality data that is based on rough set theory [15]. Bo C. studies multigranulation neutrosophic rough sets (MNRSs) and their applications in multi-attribute group decision-making [16]. Akram M., Ali G. and Alsheh N. O. introduce notions of soft rough m-polar fuzzy sets and m-polar fuzzy soft rough sets as novel hybrid models for soft computing, and investigate some of their fundamental properties [17]. Jia X. et al. propose an optimization representation of decision-theoretic rough set model. An optimization problem is proposed by considering the minimization of the decision cost [18]. Cao T. et al. discusses the use of parallel computation to obtain rough set approximations from large-scale information systems where missing data exist in both condition and decision attributes [19].
The linear programming method is a classical mathematical method, and its principle is to solve a series of linear constraint equations or inequalities on the premise of satisfying the method of solving linear functions of extreme objectives. Though its mathematical model is simple, it is widely used in various fields, such as location problems, route planning problems, manufacturing problems, marketing problems, and resource allocation problems. Application of the linear programming method can provide an effective and feasible decision-making basis for the aforementioned problems.
The linear programming problem (LP) is a problem of solving the maximum or minimum value of a linear function under a set of linear equality and inequality constraints [20]. The general linear programming model is composed of the following elements parameters and variables, objective functions, and constraint conditions.
There has been much research on optimization problems based on linear programming theory, such as the theory of linear programming and establishment of a mathematical model [21], as well as the combination of practical production, and enterprise and establishment of a linear programming model to solve the problem of allocating enterprise production resources [22]. Linear programming methods have also been used to optimize input–output models, and to establish a multi-objective linear programming model to maximize economic benefits and to minimize resource utilization [23]. One of the most important application fields of linear programming are location problems. A mixed integer linear programming model has been built to select the location of renewable energy facilities [24], and to study a multi-stage facility location problem [25]. To solve the vehicle routing problem of a distribution center, a two-stage solution was proposed.
One of the preprocessing methods of noisy data is regression. Hence, it is obvious that the linear programming model has a strong ability to resist noisy data, and designing a decision-making model based on rough set achieves the ideal accuracy. The integration of the rough set with the linear programming model will not only improve the inadequacies of the rough set, but will also make the decision-making model reach optimal accuracy, theoretically. There are a few studies on the integration of the rough set and linear programming model so far. Zhang et al. proposed a multi-objective linear programming method based on the rough set, to develop a classification for data mining. Based on their model, an improved model to predict hot spots of protein interactions was proposed [26]. However, among all of the above studies, the rough set was only used to reduce the attribute set. Because nonlinear models are considered to be the only way to describe the rough set, there are no studies on the application of linear programming methods to optimize decision-making models based on the rough set.
The biggest weakness of the decision model based on the rough set is its sensitivity to noisy data. VPRS only broadens the requirement of the upper and lower approximations in the definition, and the selection of precision often has strong subjectivity and lacks scientific evidence. VPRS can only be used as an auxiliary method to improve the resistance of the rough set model to noisy data, rather than the main method. Therefore, in this study, we extend the rough set theory via mixed-integer linear programming and we propose a model called the mixed-integer linear programming model for rough set-based classification with flexible attribute selection (in short, MILP-FRST). This model includes the advantages of MILP in resisting noisy data, and it has the ability to select attributes flexibly and automatically. MILP-FRST is able to divide the universe by attribute sets, calculate the lower approximation set under the condition of the presupposed variable precision and the minimum support number, and calculate the decision accuracy and screen out attributes. We set the maximum number of elements in the determination area as the objective function of the model. The processing of attribute selection, and partitioning of the attribute set for the universe are maximized by the objective function. During implementation, attributes that have a significant influence on the accuracy of the decision system will be selected, and the attribute set partition scheme is calculated to achieve the highest accuracy of the decision-making system. In addition, rough set models are often considered to be nonlinear. This paper first describes the related concepts and theories using linear models, which are an extension of rough set theory.
Next, we use the model to mine the correlation between the assembly clearance of diesel engine and the quality of the whole engine based on the dataset, which contains 28 attributes of the assembly clearance parameters and the whole machine quality of 29 diesel engines. Before applying the model, we carry out data pretreatment, and we screen out 15 principal components. These components cover the vast majority of information on the assembly clearance parameters of all diesel engines. Then, we input these data into the model. The experimental results verified the effectiveness and advantages of MILP-FRST.
The rest of the paper is organized as follows. In Section 2, we introduce the concept of the rough set and functional dependence. In Section 3, we build a mixed-integer linear programming model for rough set-based classification with flexible attribute selection. In Section 4, we use the clearance parameter data of 29 diesel engines and the quality data of the whole engine to verify the validity and accuracy of the model. Finally, in Section 5, conclusions are presented.

2. Rough Set Theory

2.1. Concepts and Definitions of Rough Sets

Consider a rough set based on an information system [27]: I S = ( I , A ) , where I is the universe; A is the attribute set. Both I and A are nonempty finite sets.
If the information system meets the conditions that A = C D , this information system can be called a decision-making system D S = ( I , C D ) , where C is the conditional attribute set, and D is the decisive attribute set.
Definition 1.
Indiscernibility relations [27]: In an information system I S = ( I , A ) , set B is a subset of the attribute set A, binary relation I N D ( B ) = { ( x , y ) I × I : a B , a ( x ) = a ( y ) } is the indiscernibility relation of IS, recorded as I N D ( B ) , where x and y are elements of the universe; a is an attribute of the attribute set; and a ( x ) is the value of the element x in attribute a.
Definition 2.
Equivalence class [27]: In an information system I S = ( I , A ) , set B is a subset of the attribute set A. The indiscernibility relation I N D ( B ) divides the universe I into several equivalence classes, where I / I N D ( B ) is the set of all equivalence classes, and [ x ] I N D ( B ) is the equivalence class containing element x.
Definition 3.
Upper and lower approximation [27]: In an information system I S = ( I , A ) , set B is a subset of the attribute set A, and set X is a subset of the universe I:
B _ X = { i I | [ i ] I N D ( B ) X }
B ¯ X = { i I | [ i ] I N D ( B ) X }
where B _ X is the lower approximation; B ¯ X is the upper approximation.
Definition 4.
The accuracy and membership grade [27] are:
a B = | B _ X | | B ¯ X |
ρ B = 1 a B
where a B is the accuracy of rough set X, and ρ B is the membership grade of rough set X.
Definition 5.
The membership function [27] is:
μ ( x , X ) = | [ x ] B X | | [ x ] B |
The membership function indicates the membership degree of element x to the rough set X.
Definition 6.
The accuracy of the decision-making system [27] is:
λ = | k = 1 K c B _ X k | | I |
Given that the strict definitions of the upper and lower approximations make the rough set sensitive to noisy data, the rough set cannot adapt well to all situations in practical applications. VPRS decreases the influence of missing data, incorrect data, and noisy data. In VPRS, an approximation variable precision β , which ranges from 0.5 to 1, represents the tolerance degree of the rough set to noisy data and incorrect data. β can be defined as follows:
Definition 7.
B _ X β = { i I | | [ i ] I N D ( B ) X | | [ i ] I N D ( B ) | β }
B ¯ X β = { i I | | [ i ] I N D ( B ) X | | [ i ] I N D ( B ) | > 1 β }
where set B is a subset of the attribute set A; set X is a subset of the universe I; β is the variable precision, which ranges from 0.5 to 1; B _ X β is the lower approximation; and B ¯ X β is the upper approximation.
Compared with the rough set, VPRS extends the range of the upper and lower approximation, thus restricting the sensitivity of the rough set model to noisy data.
The rough set uses the indiscernibility relation to classify equivalence classes, but it is unsuitable for numerical data, especially in the case of the application of big data and high accuracy. One way to overcome this difficulty is to use the similarity relation, so we can extend the rough set based on the similarity relation.
This extension essentially involves modifications of the two concepts of the rough set, the indiscernibility relation and equivalence class. The classical indiscernibility relation is more suitable for those descriptive attributes, while elements of an attribute set that satisfy the indiscernibility relation are divided into an equivalence class. However, when dealing with numeric data, the effect of this method will be considerably reduced. The rough set, based on the similarity relation extends the indiscernibility relation into a similarity relation, and the equivalence class classified by the indiscernibility relation is replaced by a similarity relation.

2.2. Rough Set and Functional Dependence

With the establishment of the rough set model with variable precision based on the similarity relation, the decision rules between approximately equivalence classes divided by the conditional attribute set and the approximate equivalence classes divided by decisive attribute set can be worked out. However, this is not enough to explain the correlation between the conditional attribute set and the decisive attribute set, and so, functional dependence is introduced. Although the rough set and functional dependence are two different fields, many concepts of functional dependence can be explained from the perspective of the rough set.
Definition 8.
Functional dependence, the complete dependence between universe I and attribute A, can be expressed as C d , where C A , d A .
Definition 9.
Partial dependence, the partial dependence between universe I and attribute A, can be expressed as C p d , where C A , d A .
Inference 1.
Complete dependence of attribute set, any attribute D , D A , C d works. Accordingly, there is functional dependence between C and D , and this relationship can be expressed as C D , where D is an attribute set.
Inference 2.
Partial dependence of attribute set, any attribute D , D A , C p d works. Accordingly, there is partial functional dependence between C and D , and this relationship can be expressed as C p D , where D is an attribute set.
We now explain complete dependence and partial dependence from the perspective of the rough set. For a decision-making system based on the rough set, λ = 1 means that the decisive attribute set completely depends on the conditional attribute set, that is, there is complete dependence between the decisive attribute set and conditional attribute set. 0 < λ < 1 means that there are some factors that affect the decisive attribute set in addition to the conditional attribute set, that is, the decisive attribute set partially depends on the conditional attribute set, so the following inferences are introduced.
Inference 3.
Complete dependence in the rough set, in a decision-making system D S = ( U , C D ) , only occurs when p λ = 1 , and the complete dependence C D comes into effect.
Inference 4.
Partial dependence in the rough set, in a decision-making system D S = ( U , C D ) , 0 < λ < 1 indicates that there is partial dependence C p D between C and D, and the degree of partial dependence p equals λ .
After calculating the accuracy λ of the model, according to inference 3 and inference 4, if λ = 1 , the decisive attribute set completely depends on the conditional attribute set. If 0 < λ < 1 , the decisive attribute set partially depends on the conditional attribute set. In other words, there is a certain correlation between the conditional attribute set and the decisive attribute set, and λ can be used as a parameter to measure the degree of correlation.
This section proposes the rough set model with variable precision, based on the similarity relation and some inferences related to functional dependence. This method will not only dig out the correlation between the conditional attribute set and the decisive attribute set, but also use accuracy λ to measure the degree of correlation.

3. A Mixed-Integer Linear Programming Model for Rough Set-Based Classification with Flexible Attribute Selection

There is no doubt that the decision-making model based on the rough set has the congenital defect of the rough set; thus, adding variable precision to extend the rough set into the rough set with variable precision is necessary while building the model. Nevertheless, adding variable precision only broadens the range of the upper and lower approximations. The choice of precision is often subjective and lacks scientific basis. Above all, variable precision can only be used as an auxiliary method to improve the ability of the rough set model to reduce noisy data’s bad influence on accuracy.
In this study, we build a mixed-integer linear programming model for a rough set-based classification with flexible attribute selection, which has a strong ability to overcome the noise sensitivity of the rough set model. Meanwhile, this study explains the rough set model, which is often considered to be nonlinear, by using a linear model for the first time. It is also an extension of the rough set.

3.1. Rough Set Model Based on Mixed Integer Linear Programming

Applying mixed integer linear programming to optimize the rough set model is essentially explaining the definition that is related to the rough set by linear programming. The rough set in a linear model enables the maximum accuracy of dividing the equivalence class, so that the decision-making system based on the rough set can correctly determine the correlation between the conditional attribute set and the decisive attribute set.
This model focuses on the rough set based on the similarity relation, and compares the similarity of each attribute in the attribute set. Next, the elements that satisfy the similarity threshold on each attribute are selected as the elements to be divided into an approximate equivalence class.
This model can also screen out attributes in the attribute set and take the attribute makes a considerable impact in dividing the universe into the final attribute set to reduce the dimension of the attributes.
We use the following notations:
  • I: Universe of elements.
  • k c : A set of approximate equivalence classes obtained by partitioning the conditional attribute set in the universe.
  • k d : A set of approximate equivalence classes obtained by partitioning the decisive attribute set in the universe.
  • C: Conditional attribute set.
  • D: Decisive attribute set.
  • N: Minimum support number of the conditional attribute set.
  • β : Variable precision.
  • α c : Similarity threshold of the conditional attribute set.
  • α d : Similarity threshold of the decisive attribute set.
  • M: A large number.
  • X c i : Value of each element in each conditional attribute.
  • X d i : Value of each element in each decisive attribute.
  • ω _ c i j : For any two elements i and j in universe I, if ω _ c i j = 1 ,   i and j are in the same approximate equivalence class divided by the conditional attribute set; otherwise, ω _ c i j = 0 .
  • s l c : s l c = 1 if attribute c will be selected as a new attribute set to divide universe; otherwise, s l c = 0 .
  • q i k : For any element i in universe I and any approximate equivalence class k in the set of approximate equivalence classes divided by the conditional attribute set, q i k = 1 if i belongs to k; otherwise, q i k = 0 .
  • s s i j c : Any two elements i and j in universe I and any attribute c in the conditional attribute set. s s i j c = 1   if   value   of   i   and   j on attribute c satisfies the corresponding similarity threshold α c ; otherwise, s s i j c = 0 .
  • Q k : Number of elements in the approximate equivalent class k, which is obtained from the partition of the conditional attribute set to the universe.
  • ω _ d i j : ω _ d i j = 1 if any two elements i and j belong to the same approximate equivalence class divided by the decisive attribute set; otherwise, ω _ d i j = 0 .
  • s l d : s l d = 1 if an attribute d in decisive attribute set will be selected as a new conditional attribute set to divide the universe; otherwise, s l d = 0 , and d will be eliminated.
  • q i k : q i k = 1 if any element i in universe belongs to the approximate equivalent class k ; otherwise, q i k = 0 .
  • s s i j d : s s i j d = 1 if value of any two points i and j on attribute d satisfies the corresponding similarity threshold α d ; otherwise , s s i j d = 0 .
  • Q k : Number of elements in the approximate equivalent class k , which is obtained from the partition of decisive attribute set to the universe.
  • e i k k : e i k k = 1 if point i not only belongs to the approximate equivalent class k of the conditional attribute set but also belongs to the approximate class k of decisive attribute set; otherwise, e i k k = 0 .
  • E k k : The number of elements is not only the approximate equivalence class k of the conditional attribute set, but also the approximate equivalence class k of the decisive attribute set.
  • f k :   f k = 1 if the number of elements in the approximate equivalence class k of the conditional attribute set satisfies the minimum support threshold, so that the approximate equivalence class k can be a lower approximation set; otherwise, f k = 0 .
  • L k k : L k k = 1 if the approximate equivalence class k in k c is the lower approximation set of the approximate equivalence class k in k d ; otherwise, L k k = 0 .
  • Y k : If the approximate equivalence class k in k c is the lower approximation set, Y k is the number of elements of lower approximation set k .
The objective function and constraints of the model are as follows:
Objective function: Maximize ( k = 1 K c Y k )
Subject to:
1)
M s s i j c α c | X c i X c j | ,   i I , j I , c C ;
2)
M ( 1 s s i j c ) | X c i X c j | α c ,   i I , j I , c C ;
3)
ω _ c i j s s i j c + ( 1 s l c ) , i I , j I , c C ;
4)
s s i j c 1 s l c , i I , j I , c C ;
5)
ω _ c i j 1 c C ( 1 s s i j c ) , i I , j I , c C ;
6)
M s s i j d α d | X d i X d j | , i I , j I , d D ;
7)
M ( 1 s s i j d ) | X d i X d j | α d , i I , j I , d D ;
8)
ω _ d i j s s i j d + ( 1 s l d ) , i I , j I , d D ;
9)
s s i j d 1 s l d , i I , j I , d D ;
10)
ω _ d i j 1 d D ( 1 s s i j d ) , i I , j I , d D ;
11)
q 11 = 1 ;
12)
k k c q i k = 1 , i I ;
13)
q i k + q j k 1 + ω c i j , i I , j I , k k c ;
14)
Q k = i I q i k , k k c ;
15)
q 11 = 1 ;
16)
k k d q i k = 1 , i I , k k d ;
17)
q i k + q j k 1 + ω d i j , i I , j I , k k d ;
18)
Q k = i I q i k , i I , k k d ;
19)
2 e i k k q i k + q i k , i I , k k c , k k d ;
20)
E k k = i I e i k k ,   k k c , k k d ;
21)
N f k N + ( Q k N ) ;
22)
c a r d ( I ) L k k c a r d ( I ) + ( E k k Q k β ) , k k c , k k d ;
23)
L k k f k , k k c , k k d ;
24)
Y k Q k , k k c ;
25)
Y k M k k d L k k , k k c .
In MILP-FRST, the objective function and constraints are critical parts. These parts introduce the concept of the rough set and the way to complete related theories.
The objective function in the model is the number of elements that belong to the conditional attribute set and the decisive attribute set. For MILP-FRST, it is obvious that the maximum accuracy is essentially the number of elements in the maximum region by integrating the objective function with the definition of precision in the rough set. The goal of constructing this objective function is to determine the method of division to find a more accurate correlation between the conditional attribute set and decisive attribute set.
The description of concepts related to the rough set and the complement of related theories are both completed in the process of setting constraints. These descriptions and complements consist of filtering out attributes from the conditional attribute set and the decisive attribute set, dividing the universe by the decisive attribute set, dividing the universe by the conditional attribute set, calculating the lower approximation set, calculating the number of elements, and limiting the coverage of the lower approximation set. Each constraint will be explained as follows.
The process of choosing the attributes and dividing the universe will be completed in the model.
s s i j c = 1 , if the distance between two elements of the attribute c is closer than the corresponding similarity threshold α c ; otherwise, s s i j c = 0 . The constraints are established as follows:
M s s i j c α c | X c i X c j | ,   i I , j I , c C
M ( 1 s s i j c ) | X c i X c j | α c ,   i I , j I , c C
where i and j are two elements of the same condition attribute c, and both i and j are natural numbers.
If attribute c is selected to divide the universe, s l c = 1 , we can establish constraint (11). Otherwise, s l c = 0 , as shown in constraint (12), and attribute c has no influence on dividing the universe, that is, the two elements have an indiscernibility relation on attribute c. Constraint (11) is defined as the necessary condition that indicates that when classifying two elements into an approximate equivalence class, it is not enough to make ω _ c i j = 1 . The condition of ω _ c i j = 1 means that all of the attributes of the attribute set meet the corresponding similarity threshold, so that constraint (13) is established. Elements i and j have an indiscernibility relation under the condition that all s s i j c in attribute set C are 1:
ω _ c i j s s i j c + ( 1 s l c ) , i I , j I , c C
s s i j c 1 s l c , i I , j I , c C
ω _ c i j 1 c C ( 1 s s i j c ) , i I , j I , c C
Constraints (9)–(13) initially divide the universe by the conditional attribute set, and select attributes from the conditional attribute set. Attribute sets divide the universe in accordance with the similarity between the elements of the attribute.
The processes of dividing the universe and filtering out attributes are almost the same for the conditional attribute set and decisive attribute set. Therefore, we establish constraints (14)–(18) to divide the universe by the decisive attribute set and filter out attributes from the decisive attribute set:
M s s i j d α d | X d i X d j | , i I , j I , d D
M ( 1 s s i j d ) | X d i X d j | α d , i I , j I , d D
ω _ d i j s s i j d + ( 1 s l d ) , i I , j I , d D
s s i j d 1 s l d , i I , j I , d D
ω _ d i j 1 d D ( 1 s s i j d ) , i I , j I , d D
We can obtain ω _ c and ω _ d through constraints (9)–(18), but there is much to be done to fulfil the process of dividing the universe. Each element in the universe should be allocated into k c or k d .
To complete model building, we need to specify the initial element and the initial equivalence class, and set the initial element belong to the initial equivalence class. As the initial element and the initial equivalence class are only numbers, there is no specific meaning, and so this set will not affect the results of the model calculation. According to the definition of q i k , i = 1 is the number of elements, k = 1 is the number of the approximation equivalence class, and q 11 = 1 means dividing this element into this approximation equivalence class. We can establish constraint (19):
q 11 = 1
Each element belongs to only one approximate equivalence class. However, not every predetermined approximate equivalence class has its own elements. When the number of approximate equivalence classes is unknown, the number of approximate equivalence classes in the set of approximate equivalence classes may be redundant. If the number of the provided approximate equivalence classes is less than the number of actual approximate equivalence classes, the model will not be solvable, so we establish constraint (20):
k k c q i k = 1 , i I
Only when it is confirmed that the two elements i and j can be classified into the same approximate equivalence class can elements i and j be classified into an approximate equivalence class. The value of q i k and q j k can be 1 at the same time only when ω _ c i j = 1. We establish constraint (21):
q i k + q j k 1 + ω c i j , i I , j I , k k c
Variable Q k counts the number of elements allotted into each approximate equivalence class divided by the conditional attribute set. We establish constraint (22):
Q k = i I q i k , k k c
Similarly, constraints (23)–(18) implement the process of allotting the element of the decisive attribute set:
q 11 = 1
k k d q i k = 1 , i I , k k d
q i k + q j k 1 + ω d i j , i I , j I , k k d
Q k = i I q i k , i I , k k d
The above constraints complete the process of selecting attributes and dividing the universe.
Constraints (27)–(31) implement the process of defining the lower approximation set and setting the minimum support threshold.
If one element belongs to the approximate equivalence class k and the approximate equivalence class k on the basis of the definition of the lower approximate set, this element will be selected, so we establish constraint (27):
2 e i k k q i k + q i k ,   i I , k k c , k k d
The number of elements obtained by constraint (19) should be counted, so we establish constraint (28):
E k k = i I e i k k ,   k k c , k k d
The minimum support threshold requires that the lower approximation set should meet the requirement of the minimum support number. Constraints (29) and (31) complete the limitation of the minimum support number for the lower approximate set. In constraints (29) and (31), f k shows whether the number of elements in the corresponding approximate equivalence class satisfies the minimum support number; if Q k < N , then f k must be 0. MILP-FRST introduces variable precision as an auxiliary method of improving the ability of resisting noisy data. Constraint (30) realizes the process of defining the lower approximate set:
N f k N + ( Q k N )
c a r d ( I ) L k k c a r d ( I ) + ( E k k Q k β ) , k k c , k k d
L k k f k , k k c , k k d
Finally, the number of elements in the lower approximate set is counted. If the approximate equivalence class obtained by conditional attribute set does not belong to any approximate equivalence class obtained by the decisive attribute set, this approximate equivalence class will be deemed to be an uncertain region, so the number of elements in the certain region is 0. Otherwise, this approximate equivalence class is a certain region, so the number of elements in the region equals the number of element points in this approximate equivalence class. Above all, we establish constraints (32) and (33):
Y k Q k , k k c
Y k M k k d L k k , k k c

3.2. Characteristics of the Model

Aiming at solving the problem that the rough set model has weak resistance to noisy information in a dataset, this study proposes the mixed-integer linear programming model for rough set-based classification with flexible attribute selection (MILP-FRST). This model integrates the mixed integer linear programming method with the rough set model to define the related concepts and to describe the related theories. It is not only an optimization of the original mining model, but also an extension of the rough set theory. This model has the following characteristics:
(1) The model can realize the process of filtering out attributes from the attribute set. In a practical application, the first step of analyzing a high-dimension dataset is the descending dimension. After the dimensionality reduction, the dataset can only contain partial information of the raw dataset; specifically, implementation of the dimensionality reduction process is at the expense of sacrificing the information contained in some raw datasets. MILP-FRST is able to eliminate the attributes that have little influence on the decisive accuracy, and to automatically complete the process of attribute selection. Therefore, only a simple preprocessing process based on data quality analysis needs to be performed, and the maximum extent of all of the information contained in the raw dataset is preserved.
(2) The model implements the partition of the attribute set to the universe, defines the lower approximation set and the lower approximation set, sets the variable precision, restricts the support of the lower approximation set, and calculates the determined region, and so on. All of the above are implemented in the linear model. The attribute set partitioning scheme that allows the decisive accuracy to reach the optimal value can be obtained.
(3) The model has strong extensibility. According to the specific object of this study, we can select the attribute set, and specific division of the universe and method to adapt to the dataset composed of various data types.

4. Application Study on Data from Diesel Engines

In this section, we report the results of computational experiments on an assembly clearance parameter dataset from a diesel engine to test the models and compare them. The MIP solver AMPL/CPLEX (version 12.6.0.1) was used to solve problem instances. All computational experiments were performed on a MacBook with a 2.90 GHz Intel Core i7 Processor and 8 GB memory.
This paper takes a certain type of marine diesel engine as the verification object. At present, this type of diesel engine has been put into the market for many years, and the production enterprises have accumulated a lot of valuable data. Table 1 lists the main technical parameters of this type of diesel engine.
Figure 1 the side view and main view of this diesel engine.

4.1. Data Set Introduction.

The object of study is 29 16-cylinder diesel engines of the same type. The data set includes assembly clearance parameter data and quality grade data of the diesel engine. The assembly clearance parameter data of the diesel engine is numerical data, and the quality grade data of the diesel engine is classified data.
The marine diesel engine has a complex structure and many components, so there are many assembly clearance parameters. Chybowski L. and Gawdzixuska K. put forward the latest technology of component importance analysis for complex technical systems [28,29,30]. Choosing important components in complex systems is a key step. This type of diesel engine mainly includes four parts assembly clearance parameters: 2K, 5K, 10K and 11K. Among them, 2K refers to the mating clearance parameters of the crankshaft and the seat hole of the main bearing, 5K refers to the mating clearance parameters of the camshaft and the seat hole, 10K refers to the meshing clearance parameters of the gears, and 11K refers to the mating clearance parameters of the gear hole and the bearing. Table 2 lists the components involved in four types of assembly clearance parameters and the number of parameters.
A total of 28 assembly clearance parameters of the diesel engine were selected, that is, the experimental data set is 28-dimensional. The quality grade data comes from the test run of the diesel engine by the manufacturer before the diesel engine is delivered, including tests on flammability, diesel viscosity, and reliability. Through various test runs, the manufacturer determines the quality grade of the diesel engine. The quality grades are divided into three grades, Qualified, First grade, and High grade. Table 3 shows the part of data of the 28 assembly clearance parameters and the corresponding quality grades of the diesel engine.

4.2. Data Pre-Treatment

After the correlation analysis of the dataset, it is obvious that there is a strong correlation between the assembly clearance parameters of the same part of the diesel engine, and this strong correlation will affect the effectiveness and efficiency of the model. Therefore, according to the correlation analysis of the assembly clearance parameters of the diesel engine, the principal component analysis method is used to reduce the dimension of the dataset. Taking all diesel engine assembly clearance parameters as the input, principal component analysis is carried out, and the cumulative variance contribution rate of each principal component is obtained.
As listed in Table 4, the cumulative variance contribution rate of the first 15 principal components is up to 89%; that is, these 15 principal components can cover most of the information of the assembly clearance parameters. A new dataset made up of these 15 principal components is presented in Table 5.
The new dataset simplifies the original dataset and retains most of the information contained in the original dataset. Consequently, we can avoid a series of problems that the high-dimensional datasets creates in data mining. Simultaneously, the simplification of the original dataset can improve the efficiency of the model.
Finally, we need to integrate the assembly clearance parameters and whole-quality grades after dimension reduction, and obtain the final dataset that is directly applied to the subsequent computation (see Table 6).

4.3. Demonstration of the Process of the Model

Considering the specific object and dataset, the whole quality grades of diesel engine are known in this case, so that the result of partitioning the decisive attribute set to the universe is known in this instance. Therefore, we can simplify the model. We first remove the selected attributes in the decisive attribute set, and relative variables and constraints in the dividing universe. Then, we transfer the variable q i k into a parameter matrix, which is known to be a parameter that describes the result of universe partitioning by the decisive attribute set.
The model will be implemented in the MIP solver AMPL/CPLEX (version 12.6.0.1). In the operation of the model, the following parameters need to be set in advance:
1)
Minimum support number of the lower approximation set N = 3,
2)
Variable precision β = 0.9 ,
3)
A large number M = 999,
4)
The initial condition of the conditional attribute set partitioning universe q 11 = 1 ,
5)
List of the threshold of the similarity of the conditional attribute set α c = [ 0.0495 , 0.0369 , , 0.099 ] , where c = 15 is the number of assembly clearance parameters in the instance,
6)
The initial number of the approximate equivalence class of the partition of the conditional attribute set to the universe is 10.
The model input consists of the principal component data of the assembly clearance parameters obtained by dimensionality reduction processing, quality grades of the diesel engine, and preset parameters described above. The output of the model includes the selection results of the principal components of the input, division results of the universe according to the conditional attribute set, calculation results of the lower approximation set, and calculation results of the number of elements in the determined region.
Fifteen principal component attributes are included in the conditional attribute set, which is composed of the assembly clearance parameters of a diesel engine. The model can filter the attributes from the attribute set to eliminate the attributes that have little impact on the accuracy of the decision system, and its filtering result are expressed by the variable s l c . The result is:
s l c = { 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 c = 1 c = 2 c = 3 c = 4 c = 5 c = 6 c = 7 c = 8 c = 9 c = 10 c = 11 c = 12 c = 13 c = 14 c = 15
If the s l value of attribute c is 1, this attribute will be selected; otherwise, this attribute will be eliminated. Therefore, the result shows that all 15 principal component attributes will be selected.
The conditional attribute set partitioning the universe is an important step in the calculation process of the model. Meanwhile, it is also the prerequisite for the subsequent calculation; k = 10 represents the 10 approximate equivalence classes. If a diesel engine belongs to an approximate equivalence class, the value of the element in the matrix is 1; otherwise, it is 0. The result is:
Q k = { 4 4 4 3 4 3 0 3 1 3 k = 1 k = 2 k = 3 k = 4 k = 5 k = 6 k = 7 k = 8 k = 9 k = 10
This result indicates the number of diesel engines in each approximate equivalence class obtained by the partitioning of conditional attribute set to the universe. Among the 10 approximate equivalence classes, one has not been allocated any element; this approximate equivalence class will be deleted. One has been allocated only one element, and its number is less than the minimum support number; therefore, it will also be deleted. Only eight approximate equivalence classes can be regarded as the lower approximation set.
E = [ 0 0 0 0 4 0 0 0 0 3 0 4 4 3 0 0 0 3 0 0 4 0 0 0 0 3 0 0 1 0 ]
The E matrix is the most important part of the model output. The E matrix represents the number of elements that not only belong to approximate equivalence class c, but also belong to one quality grade. E matrix is an important basis for solving the lower approximation set. In the E matrix for this case, the 10 lines indicates that the number of approximate equivalence classes determined by conditional attribute set partitioning of the universe is 10. Similarly, the number of approximate equivalence classes determined by decisive attribute set partitioning of the universe is 3.
Y k = { 4 4 4 3 4 3 0 3 0 3 k = 1 k = 2 k = 3 k = 4 k = 5 k = 6 k = 7 k = 8 k = 9 k = 10
Y k is the number of elements in each lower approximation set. It can be concluded that eight approximate equivalence classes meet the condition of being members of the lower approximate set by analyzing the minimum support number and variable precision. Hence, the number of elements in the determined area is:
k = 1 15 Y k = 28
The area of the model is:
λ = k = 1 15 Y k | I | = 0.97
On the basis of the inferences of the rough set and the function dependence, 0 < λ < 1 . Thus, there is partial dependence between the conditional attribute set and the decisive attribute set of the decision system:
{ assembly   clearance   parameter } 0.97 { quality   grade }

4.4. Performance Comparison of Models

To validate the effectiveness and advantages of the model, experiments are performed to compare the accuracy of the models. The model that our model is compared to is the Φ-rough set.
As listed in Table 7, obviously, the accuracy of model MILP-FRST is higher than that of the model Φ-Rough set. The accuracy is close to one, which shows that our proposed model can establish an accurate decision-making rule between the diesel engine assembly clearance parameters and whole machine quality grades, and excavate a higher correlation between them.
MILP-FRST is an extension of the rough set. An obvious characteristic of a linear model is its ability to find the optimum solution. This characteristic enables the model to find the best way to classify attributes, even if the dataset can also obtain ideal results merely through simple data preprocessing, and it considerably increases the ability to resist noisy data.

4.5. Extension of the Model

As mentioned above, the minimum support number of the lower approximation set N and variable precision β are set in advance. The classification quality reflects the degree of dependence of decisive attribute D on the conditional attribute C and the uncertainty of decision system. The classification quality is inversely proportional to the uncertainty. In the model, the classification quality largely depends on the value of β [31]. Practically, the user does not always know how to set the value of β to obtain the maximum accuracy model. The minimum support number of the lower approximation set N also influences the accuracy of the model. Similarly, the user always sets the value of N in accordance with their experience. To determine the best value of β and N, we propose an algorithm. We set up a loop traverses all the values between 0.5–1, adding 0.01 each time. The results of the different N values are shown in Figure 2.
As shown in Figure 3, it is obvious that with the increase in N, the average of λ decreases and the variance of λ increases.
Although the model’s accuracy increases when β decreases, for this model, the smaller β does not make the model better. β represents the tolerance degree of the rough set to noisy and incorrect information in the dataset, while a smaller β tolerates more noisy data, but this violates our purpose in building MILP-FRST. Therefore, how to balance β against λ requires further research. As for the minimum support number N, our results indicate that the smaller the N, the better the λ. This rule is useful in practical applications, but when using this rule, the characteristics of the object should be considered to determine the appropriate N.

5. Conclusions

The purpose of this paper is to establish an accurate decision-making method between the quality level of the diesel engine and the parameters of assembly clearance. Therefore, a novel mixed-integer linear programming model for the rough set-based classification with flexible attribute selection, called MILP-FRST, is presented. First, the correlation between the conditional attribute set and decisive attribute set according to the inference of the rough set and function dependence is calculated. Second, by integrating the data mining model with the mixed integer linear programming theory, the optimization method is studied with regard to the sensitivity of the rough set to noisy data. Integrating the MILP model with the rough set model, the related theories and concepts of the rough set are implemented in the model, and the extension of the rough set research is completed. Finally, a case study on test data from a diesel engine is carried out. Experiments show that the decision-making method proposed in this paper can realize the quantitative discussion of the relationship between assembly clearance and the quality level of the whole machine. Also, the effectiveness of and the advantages of the MILP-FRST are verified. Furthermore, the extension of the MILP-FRST indicates that the usage of the minimum support number N and related topics are worthy of more in-depth study.

Author Contributions

Conceptualization, X.Y. and W.C.; methodology, Y.X. and X.Y.; software, X.Y. and J.L.; validation, X.Y. and S.Z.; formal analysis, S.Z.; investigation, J.L. and Y.W.; resources, W.C. and Y.W.; data curation, Y.X.; writing—original draft preparation, X.Y.; writing—review and editing, S.Z.; visualization, J.L.; supervision, S.Z.; project administration, W.C.; funding acquisition, W.C.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No.71501007 & 71672006 & 71871003). The study is also sponsored by the Aviation Science Foundation of China (2017ZG51081), the Technical Research Foundation (JSZL2016601A004).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zadeh, L.A. Fuzzy sets, information and control. Inf. Control 1965, 8, 338–383. [Google Scholar] [CrossRef]
  2. Bundy, A.; Wallen, L. Dempster-Shafer Theory; Springer: Berlin/Heidelberg, Germany, 1984. [Google Scholar]
  3. Pawlak, Z. Rough set. Int. J. Comput. Inf. Sci. 1982, 11, 341–356. [Google Scholar] [CrossRef]
  4. Liang, J.; Li, R.; Qian, Y. Distance: A more comprehensible perspective for measures in rough set theory. Knowl. Based Syst. 2012, 27, 126–136. [Google Scholar] [CrossRef]
  5. Xiao, Y.; Kaku, I.; Chang, W. Φ-Rough Sets Theory and Its Usage on Mining Approximate Dependencies; Springer: Berlin/Heidelberg, Germany, 2008; Volume 5227, pp. 922–934. [Google Scholar]
  6. Luo, C.; Li, T.; Yao, Y. Dynamic probabilistic rough sets with incomplete data. Inf. Sci. 2017, 417, 39–54. [Google Scholar] [CrossRef]
  7. Yao, Y.; Zhang, X. Class-specific attribute reducts in rough set theory. Inf. Sci. 2017, 418, 601–618. [Google Scholar] [CrossRef]
  8. Zhang, Z.; Shi, Y.; Gao, G. A rough set-based multiple criteria linear programming approach for the medical diagnosis and prognosis. Expert Syst. Appl. 2009, 36, 8932–8937. [Google Scholar] [CrossRef]
  9. Ziarko, W. Variable precision rough set model. J. Comput. Syst. Sci. 1993, 46, 39–59. [Google Scholar] [CrossRef] [Green Version]
  10. Tao, Z.; Bao-Dong, X.U.; Wang, D.W.; Ran, L.I. Rough Rules Mining Approach Based on Variable Precision Rough Set Theory. Inf. Control 2004, 33, 17–18. [Google Scholar]
  11. Beynon, M.J. Introduction and Elucidation of the Quality of Sagacity in the Extended Variable Precision Rough Sets Model. Electron. Notes Theor. Comput. Sci. 2003, 82, 30–39. [Google Scholar] [CrossRef] [Green Version]
  12. Griffiths, B.; Beynon, M.J. Expositing stages of VPRS analysis in an expert system: Application with bank credit ratings. Expert Syst. Appl. 2005, 29, 879–888. [Google Scholar] [CrossRef]
  13. Su, C.T.; Hsu, J.H. Precision parameter in the variable precision rough sets model: An application. Omega 2006, 34, 149–157. [Google Scholar] [CrossRef]
  14. Wang, X.Y. New method of obtaining variable precision value based on variable precision rough set model. Comput. Eng. Appl. 2010, 46, 48–50. [Google Scholar]
  15. Zavareh, M.; Maggioni, V. Application of Rough Set Theory to Water Quality Analysis: A Case Study. Data 2018, 3, 50. [Google Scholar] [CrossRef]
  16. Bo, C.; Zhang, X.; Shao, S.; Smarandache, F. New Multigranulation Neutrosophic Rough Set with Applications. Symmetry 2018, 10, 578. [Google Scholar] [CrossRef]
  17. Akram, M.; Ali, G.; Alsheh, N.O. A New Multi-Attribute Decision-Making MethodBased on m-Polar Fuzzy Soft Rough Sets. Symmetry 2017, 9, 271. [Google Scholar] [CrossRef]
  18. Jia, X.; Tang, Z.; Liao, W.; Shang, L. On an optimization representation of decision-theoretic rough set model. Int. J. Approx. Reason. 2014, 55, 156–166. [Google Scholar] [CrossRef]
  19. Cao, T.; Yamada, T.; Unehara, M.; Suzuki, I.; Nguyen, D. Parallel Computation of Rough Set Approximations in Information Systems with Missing Decision Data. Computers 2018, 7, 44. [Google Scholar] [CrossRef]
  20. Zhang, J.Z. Linear Programming; Science Press: Beijing, China, 1990. [Google Scholar]
  21. Tian, Y.P. Strengthening the Project Cost Control by Using Linear Programming Theory. Railw. Eng. Cost Manag. 2013, 28, 38–40. [Google Scholar]
  22. Gu, M.C. Application of linear Programming Theory in Enterprise Production Planning. J. Gansu Radio TV Univ. 2010, 20, 40–42. [Google Scholar]
  23. Gong, Q.H.; Yang, L.; Huang, G.Q. Research on Industrial Structure Adjustment Model Based on Resources and Linear Programming. Sci. Technol. Manag. Res. 2011, 31, 26–28. [Google Scholar]
  24. Kelechi, O.; Tokos, H. An MILP Model for the Optimization of Hybrid Renewable Energy System. Comput. Aided Chem. Eng. 2016, 38, 2193–2198. [Google Scholar]
  25. Boujelben, M.K.; Gicquel, C.; Minoux, M. A MILP model and heuristic approach for facility location under multiple operational constraints. Comput. Ind. Eng. 2016, 98, 446–461. [Google Scholar] [CrossRef]
  26. Chen, R.Y.; Zhang, Z.; Wu, D.; Zhang, P.; Zhang, X.; Wang, Y.; Shi, Y. Prediction of protein interaction hot spots using rough set-based multiple criteria linear programming. J. Theor. Biol. 2011, 269, 174–180. [Google Scholar] [CrossRef] [PubMed]
  27. Pawlak, Z. Theoretical Aspect of Reasoning About Data; Kluwer Academic Publishers: Dordrecht, The Netherlands, 1991. [Google Scholar]
  28. Chybowski, L.; Gawdzińska, K. On the Present State-of-the-Art of a Component Importance Analysis for Complex Technical Systems. Adv. Intell. Syst. Comput. 2016, 445, 691–700. [Google Scholar]
  29. Chybowski, L.; Gawdzińska, K. On the Possibilities of Applying the AHP Method to a Multi-criteria Component Importance Analysis of Complex Technical Objects. Adv. Intell. Syst. Comput. 2016, 445, 701–710. [Google Scholar]
  30. Chybowski, L.; Gawdzińska, K. Selected issues regarding achievements in component importance analysis for complex technical systems. Sci. J. Marit. Univ. Szcz. 2017, 52, 137–144. [Google Scholar]
  31. Zhang, R.; Xiong, S.; Chen, Z. Construction method of concept lattice based on improved variable precision rough set. Neurocomputing 2016, 188, 326–338. [Google Scholar] [CrossRef]
Figure 1. Side view and main view of this diesel engine.
Figure 1. Side view and main view of this diesel engine.
Sustainability 11 00620 g001
Figure 2. (a) Results when N = 2; (b) Results when N = 1; (c) Results when N = 1; (d) Results when N = 1.
Figure 2. (a) Results when N = 2; (b) Results when N = 1; (c) Results when N = 1; (d) Results when N = 1.
Sustainability 11 00620 g002
Figure 3. Comparison of the average and variance of λ.
Figure 3. Comparison of the average and variance of λ.
Sustainability 11 00620 g003
Table 1. Main technical parameters.
Table 1. Main technical parameters.
Cylinder number16
Overload capacity110% (1 h, allowed for 12 h)
Cylinder Diameter/Stroke175/190 (mm)
Rated power1658~2032 kW
Exhaust backpressure<2.5 (kPa)
Lubricating oil consumption rate≤1.3 (g/kW·hr)
Calibrated speed1500~1800 (r/min)
Single Cylinder Exhaust Volume4.43 (L)
Fuel consumption ratio≤200 + 5% (g/kW·hr)
Compression ratio13.5:1
Explosion pressure16.0 (MPa)
Table 2. Assembling clearance parameters adopted.
Table 2. Assembling clearance parameters adopted.
Parameter TypesRelated ComponentsNumber of Parameters
2KSpindle hole crankshaft and bearing seat5
5KCamshaft and its seat7
10KGears10
11KGears and bearings6
Table 3. Assembly clearance data and quality grade of the diesel engine.
Table 3. Assembly clearance data and quality grade of the diesel engine.
Diesel Engine ID2K12K2……11K6Machine Quality Grade
10.1930.176……0.11Qualified
20.1830.183……0.16Qualified
30.1830.179……0.11First grade
40.1740.164……0.18High grade
50.1610.167……0.15Qualified
60.1760.175……0.15First grade
290.1760.159……0.13High grade
Table 4. Results of the principal component analysis.
Table 4. Results of the principal component analysis.
ComponentAttributePrincipal Component Contribution Rate
TotalRatio of Variance (%)Cumulative (%)TotalRatio of Variance (%)Cumulative (%)
110.93419.52419.52410.93419.52419.524
26.88912.30231.8266.88912.30231.826
34.9788.89040.7174.9788.89040.717
44.1127.34448.0604.1127.34448.060
53.7886.76554.8253.7886.76554.825
63.1445.61560.4393.1445.61560.439
72.6724.77165.2112.6724.77165.211
82.3284.15769.3682.3284.15769.368
92.1913.91273.2802.1913.91273.280
101.9013.39576.6751.9013.39576.675
111.7603.14279.8171.7603.14279.817
121.4952.66982.4861.4952.66982.486
131.4472.58385.0691.4472.58385.096
141.2002.14387.2131.2002.14387.213
151.0161.81489.0271.0161.81489.027
160.9231.64890.675
170.8161.45792.131
180.7181.28393.414
190.6641.18694.600
200.6061.08295.682
210.4670.83396.516
220.4380.78397.299
230.4320.77298.071
240.3970.70998.780
250.2520.45099.229
260.1840.32899.557
270.1640.29299.850
280.0840.150100.00
Table 5. A new dataset made up of these 15 principal components.
Table 5. A new dataset made up of these 15 principal components.
IDPC1PC2PC3……PC15
10.908−0.379−0.489……1.228
2−0.5601.453−0.921……−0.206
………………………………
130.339−0.882−0.015……0.892
141.5090.4540.669……−0.116
15−0.297−0.9250.043……1.370
………………………………
20−0.295−0.2850.915……0.243
………………………………
29−1.7820.146−1.386……−0.116
Table 6. The final dataset.
Table 6. The final dataset.
IDPC1PC2……PC15Level
10.908−0.379……1.228qualified
2−0.5601.453……0.206qualified
………………………………
130.339−0.882……0.892First grade
141.5090.454……0.116First grade
………………………………
20−0.295−0.285……0.243First grade
………………………………
29−1.7820.146……0.116High grade
Table 7. Comparison of the accuracy of the two models.
Table 7. Comparison of the accuracy of the two models.
Model λ
Φ-Rough set0.68
MILP-FRST0.97

Share and Cite

MDPI and ACS Style

Chang, W.; Yuan, X.; Wu, Y.; Zhou, S.; Lei, J.; Xiao, Y. Decision-Making Method based on Mixed Integer Linear Programming and Rough Set: A Case Study of Diesel Engine Quality and Assembly Clearance Data. Sustainability 2019, 11, 620. https://0-doi-org.brum.beds.ac.uk/10.3390/su11030620

AMA Style

Chang W, Yuan X, Wu Y, Zhou S, Lei J, Xiao Y. Decision-Making Method based on Mixed Integer Linear Programming and Rough Set: A Case Study of Diesel Engine Quality and Assembly Clearance Data. Sustainability. 2019; 11(3):620. https://0-doi-org.brum.beds.ac.uk/10.3390/su11030620

Chicago/Turabian Style

Chang, Wenbing, Xinglong Yuan, Yalong Wu, Shenghan Zhou, Jingsong Lei, and Yiyong Xiao. 2019. "Decision-Making Method based on Mixed Integer Linear Programming and Rough Set: A Case Study of Diesel Engine Quality and Assembly Clearance Data" Sustainability 11, no. 3: 620. https://0-doi-org.brum.beds.ac.uk/10.3390/su11030620

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop