Next Article in Journal
Effect of Composition on the Mechanical Properties and Wear Resistance of Low and Medium Carbon Steels with a Biomimetic Non-Smooth Surface Processed by Laser Remelting
Next Article in Special Issue
Experimental and Numerical Investigation of Reaction Behavior of Carbon Composite Briquette in Blast Furnace
Previous Article in Journal
Texture Hardening Observed in Mg–Zn–Nd Alloy Processed by Equal-Channel Angular Pressing (ECAP)
Previous Article in Special Issue
Prediction of Central Carbon Segregation in Continuous Casting Billet Using A Regularized Extreme Learning Machine Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Statistical Modeling to Predict the Electrical Energy Consumption of an Electric Arc Furnace Producing Stainless Steel

Royal Institute of Technology; Brinellvägen 23, 114 28 Stockholm, Sweden
*
Author to whom correspondence should be addressed.
Submission received: 29 November 2019 / Revised: 20 December 2019 / Accepted: 21 December 2019 / Published: 24 December 2019
(This article belongs to the Special Issue Mathematical Modeling and Simulation in Ironmaking and Steelmaking)

Abstract

:
The non-linearity of the Electric Arc Furnace (EAF) process and the correlative behavior between the process variables impose challenges that have to be considered if one aims to create a statistical model that is relevant and useful in practice. In this regard, both the statistical modeling framework and the statistical tools used in the modeling pipeline must be selected with the aim of handling these challenges. To achieve this, a non-linear statistical modeling framework known as Artificial Neural Networks (ANN) has been used to predict the Electrical Energy (EE) consumption of an EAF producing stainless steel. The statistical tools Feature Importance (FI), Distance Correlation (dCor) and Kolmogorov–Smirnov (KS) tests are applied to investigate the most influencing input variables as well as reasons behind model performance differences when predicting the EE consumption on future heats. The performance, measured as kWh per heat, of the best model was comparable to the performance of the best model reported in the literature while requiring substantially fewer input variables.

1. Introduction

In the light of the increased use of Electric Arc Furnaces (EAF) to produce steel, it has become of more interest to study ways of reducing the raw material and energy consumption of the process. Not only will successful attempts reduce the environmental impact, but also increase the financial result for the company producing the steel. One common tool to evaluate new production strategies is through the act of modeling. Modeling is a valuable tool because it enables process engineers to evaluate proposed changes without interfering with the process, which may be more or less harmful to the steel plant supply chain. These modeling approaches can be, for example, physicochemical models using established relationships such as the mass- and energy balance equations [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19]. Another approach is through Computational Fluid Dynamics (CFD) modeling [20]. While these models are mainly based on physicochemical equations, there is another type of model approach that is purely based on data. These models are known as statistical models and have frequently been used to predict the Electrical Energy (EE) consumption, per heat [21,22,23,24], or per ton of tapped steel, from the EAF [25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42].
Several gaps in previous attempts to model the EE consumption using statistical modeling have recently been highlighted [43]. One such gap is related to the black-box behavior of non-linear statistical models, such as Artificial Neural Networks (ANN), which imposes lack of trust in the model in the eyes of the process engineers. Other gaps are related to the statistical modeling framework itself. For example, lack of a detailed description of the data- and modeling pipeline, which includes variable selection, data collection, data cleaning, parameter-search, modelling, and evaluation of model robustness. Another gap is the lack of connection between metallurgical knowledge and statistical modeling, both of which must be carefully considered if one aims to create a statistical model that is of practical use in the steel plant. In some cases, this has led to a lack of relevant variables and in some other cases an overuse of the number of variables. The aim of this article is to address these gaps in the context of the aforementioned statistical modeling of an EAF producing stainless steel.
ANN will be used to predict the EE consumption of an EAF producing stainless steel. This is a suitable algorithm for this prediction problem since some important physicochemical relations governing the EE consumption are non-linear. The choice of variables to model the EE consumption will be motivated using both metallurgical reasoning and statistical reasoning. First, the energy balance equation will be used to highlight the physicochemical relations between the variables governing the EE consumption [43]. Second, the expected correlative relation between the variables with respect to the EE consumption will be highlighted. The common thread in both approaches is the use of process knowledge in the reasoning.
The data- and modeling pipeline will be explained in detail for reproducibility and for reference to future studies aiming to use statistical models to predict the EE consumption of the EAF. To validate the models and to address the black-box behavior of the ANN models, statistical analysis tools such as Feature Importance (FI), Kolmogorov–Smirnov (KS) test, and Distance correlation (dCor), will be used. These methods have not previously been used in the context of statistical modeling of the EE consumption in the EAF. The consequences and aspects of this approach will also be discussed.
The results of the modeling shows that it is not necessary to use a high amount of input variables to create the best performing model. It is enough to select a subset of a set of well-chosen variables using process knowledge. Using dCor, KS tests, and FI, as complementary tools can aid the process engineers in both finding the most important variables with respect to the model performance and the reasons why a model performs differently on previously unseen heats (test data).

2. Background

2.1. EAF Process

The EAF is the main melting process in the mini-mill, type of steel plant. It uses raw material such as steel scrap and alloys to create molten steel for further processing in downstream processes in the steel plant.
The EAF process begins with the charging phase during which raw materials are added to the furnace. The melting phase starts when the electrodes are powered on and bored down into the raw material. This phase lasts until enough of the raw material has melted to make room for the second basket of raw material. This is followed by yet another melting phase. During the melting phases, burners are activated to remove cold spots, which facilitates an even melting behavior of the charged scrap. After most of the raw material is melted, the refining phase starts during which the steel is adjusted to a pre-specified composition. Additional raw material such as carbon and silicon are added in combination with oxygen lancing to facilitate exothermic chemical reactions. This reduces the amount of EE needed to heat the steel. Lastly, the steel is tapped into a ladle for further processing in the steel plant. Any necessary preparations are then made for the furnace, such as fettling of the refractories, before the next heat starts. This generalized EAF process is illustrated in Figure 1.

2.2. Energy Balance Equation

The energy balance equation of the EAF process makes it possible to express EE consumption, E E l , as a sum of ingoing and outgoing energy factors (Equations (1)–(3)).
E t o t , i n = E E l + E C h e m + E B u
E t o t , o u t = E S t e e l + E S l a g + E G a s + E D u s t + E C o o l i n g + E R a d + E C o n v + E E l , l o s s
E t o t , i n E t o t , o u t = 0
Each of the energy factors are related to physical and chemical entities (see Table 1).
Calculations of ingoing and outgoing energy in the EAF process have previously been calculated [9,25,44,45,46]. This is compiled in Table 2.
The ingoing and outgoing energy of the EAF, as percentage of each total, provide guidance in choosing input variables to a statistical model predicting the EE consumption.

2.3. Non-Linearity

Some of the terms in the energy balance equation listed in Table 1 are related to non-linear physical phenomena in the process. Examples include, E G a s and E R a d . Breaking down the Tap-to-Tap time (TTT) into its smaller components Charging, Melting, Refining, Extended Refining, Tapping, and Preparation, the process evidently becomes considerably more non-linear. The reason is because the energy loss through convection, conduction, and radiation are very different from these sub-processes For example, the energy loss through radiation is a lot higher when the steel is molten, i.e., during refining, compared to the energy loss during the first charging phase. Adding the dynamic timelines of each sub-process due to varying amounts of scrap, refining times, and various delays between heats, the physicochemical relation to the EE consumption becomes even more complex.
Because of the non-linear and non-normal distributed outcome of the process, linear statistical models, such as Multivariate Linear Regression (MLR), and linear statistical metrics, such as the Pearson correlation, are sub-optimal as tools in the context of predicting the EE consumption.

2.4. Statistical Modeling

Statistical modeling differs from other types of mathematical modeling such as those based on physical and chemical relations. See the energy balance equation in Section 2.2 for an example of a physicochemical model. Physical models act directly upon the values of the input variables to predict the output variable. Statistical models, on the other hand, act on the value with respect to previous values of the input variable it has observed in the context of the values of the other input variables. The value of the output variable is always a probability or a value with the highest probability. This means that statistical models do not adhere to the physicochemical relations between the variables. The connection between the physicochemical relations and the model is, by and large, dependent on the data quality, variations in the data, and correlative behavior between the variables. Hence, statistical models can seem counter-intuitive in the eyes of a practitioner of physicochemical-based modeling.
Data quality: A bad data quality will negatively impact the performance of a statistical model since uncertainties are imposed. There are many different ways to get good data quality. The variables must be correctly defined with respect to what they represent. This may not always be the case because everything that can go wrong is not possible to account for. For example, if the furnace is considered closed when the roof is closed, then the charging time of the furnace could still be counted if the roof is not completely shut even though the furnace is in the melting stage. Connecting plant events to logged data is not always trivial and fault free. Other examples of where errors are prevalent are from online measurements and estimations of energies based on those measurements as well as manual logging of data.
Data variability: Near constant variables add little value to a statistical model because the variations in the data are what the statistical model learns and then uses to make predictions on test data. Statistical models require variations in the data to function properly. This is perhaps one of the most counter-intuitive traits of statistical modeling since a constant term in an arbitrary physicochemical model is important for its prediction. For example, a statistical model for an EAF where the total amount of ingoing lance oxygen is constant will not receive any benefit from including that variable. This is in contrast to a physical model where the amount of added oxygen leads to exothermic chemical reactions.
Correlation: Strongly correlated variables are in a statistical sense redundant variables. Weakly correlated variables with respect to the output variable may be redundant. The degree of redundancy is closely related to the correlation between the input variable and other input variables. It may be that the input variable in question does not add any further information to the statistical model with respect to lowering the prediction error. Because of this, some variables that are considered important for predicting EE consumption in a physical model may appear as not important for the statistical model. On one hand, correlation does not imply causation, on the other hand, statistical models cannot distinguish correlation from causation. However, correlation can point to areas were causation may exist. This stresses the importance of possessing domain-specific knowledge about the process to distinguish between causative and non-causative relations between the variables governing the statistical model. Three cases of correlation related to the EAF process with respect to the EE consumption are presented, which highlight the complex relations between the different entities in the EAF process governing the EE consumption. These cases are also illustrated in Figure 2.
First, the composition of the ingoing material affects the EE consumption exothermic reactions. Here, the composition together with oxygen lancing give rise to exothermic reactions which in turn reduce the EE consumption.
Second, the various sub-processes of the EAF affect the EE consumption in various rates. For example, during the melting process, EE is added in a more intense rate than, for example, during the refining process. Furthermore, the energy loss through radiation, convection, and conduction is more intense during the refining process when most of the steel is melted. Adding all sub-processes together gives us the total process time. The delays in the process are added to the sub-process where the delays occur. Thus, the time variables are both interrelated and have a complex relation to the EE consumption.
Third, the raw material types that are added to the EAF have a more complicated relation to EE consumption. Raw material types correlate positively to the EE consumption since they directly add to the total weight of the ingoing material to be melted. The raw material types are also correlated with one another since, given a pre-determined charge weight, the raw material types combined must make up to that specific weight. If one raw material type is not available, another is used as replacement. Raw material types can also contribute negatively to the EE consumption if the raw material type contains C, Si, Al, Fe, and Cr, all of which react exothermic with oxygen.
Variable selection comes down to how much additional useful information, with respect to predictive performance, the model gains after adding another input variable. The three points previously mentioned also tie closely to the specific EAF where the data comes from. Even though the interrelations between the input variables to the EE consumption are similar between EAF furnaces, it is not possible to know, ad hoc, the specific impact on the predictive performance by a setup of input variables.
In the supervised statistical modeling framework, which will be used in this paper, each row of input data has a corresponding output value during the training phase and test phase. The supervised statistical modeling framework can be abstracted into the following steps.
  • Select statistical model and values for the model-specific parameters.
  • Train the model using training data until the model accuracy converges, i.e., stops improving.
  • Test the model on previously unseen data.
  • Record the accuracy on test data and evaluate its practical applicability.
  • If the accuracy on the test data is satisfactory, deploy the model into production.
  • Re-train the model if the model accuracy has deteriorated. This deterioration is bound to happen over time in a production setting, partly due to changes in the process.
To find the optimal combination of model parameters, multiple models are trained for each combination of parameters. This is known as grid-search or parameter-search.
There are also subcategories of supervised statistical models. In the scope of this paper, these are divided into linear and non-linear statistical models. Linear statistical models can learn linear relationships between the input variables and the output variables. One such example is Multivariate Linear Regression (MLR). On the other hand, non-linear statistical models can learn both linear and non-linear relationships between the input variables and the output variable. One such example is ANN. Both ANN and MLR have been commonly used to predict the EE consumption in the EAF [43]. While an ANN model can learn non-linear relations in the data, the model is almost impossible to interpret. This is commonly referred to as a black-box model and is one of the main reasons why such models are not widely accepted in practice in the context of steel process modeling. However, by using statistical analysis tools, it is possible for the process engineers to get insights on which input variables are the most important for the model to make accurate predictions as well as reasons behind performance differences on test data.
Furthermore, it is always possible to create a more complex statistical model to any given problem. Nevertheless, one should always strive for model parsimony, which is achieved when picking the simplest of a group of models achieving similar performance. The model complexity is related to both the number of input variables as well as the chosen type of statistical model.

3. Method

The workflow explained in this section can be followed in Figure 3.
The software used for all parts of the experiments was Python provided by Anaconda distribution. The hardware and software specifics are shown in Appendix A.1.

3.1. Furnace Information

The data used in this study comes from an Alternate Current (AC) EAF producing stainless steel. The furnace nominal capacity is 80 tons of molten steel and the electrical system has a maximum power of 80 MW. The charged raw material is 100% scrap, and alloys. A preheater, which uses part of the furnace off-gas enthalpy as energy source and heat transfer medium, is employed to remove entrained moist and to burn off residual oil and grease in the scrap. Conventional oxy-fuel burners are used to facilitate an even temperature distribution during melting. Oxygen lancing is used to facilitate oxidation of elements such as C and Si. The number of heats produced per year is approximately 5000.

3.2. Data

3.2.1. Variable Selection

The variable selection process is partly based on the discussion regarding the energy balance equation, non-linearity, and the reasoning about the correlative relations of the EAF process with respect to the EE consumption, (see Section 2.2, Section 2.3 and Section 2.4). The variables can be divided into four categories:
  • Time: The logged times for the whole process as well as the various EAF sub-processes are important for predicting the EE demand. Each process sub-stage, Charging, Melting, Extended refining, Refining, and Tapping, contribute differently to the heat loss. For example, the heat loss is higher when molten steel is present compared to when the first bucket of scrap is charged. Both the TTT and the Process Time were included. The total time imposed by delays, defined as the sum of the deviation from nominal time of each EAF sub-process, was also included.
    Some obvious correlations can be expected with regards to these time variables. As Charging, Melting, Extended refining, Refining, and Tapping makes up the Process Time and the majority of TTT, these variables are expected to be relatively highly correlated.
  • Chemical: Oxidation and the oxyfuel-burners can account for as much as 50% and 11% of the total ingoing energy, respectively (see Table 2). The contribution by the burners was accounted for by one variable propane. The oxygen gas injected through the burners was not included due to a near stoichiometric relationship between oxygen gas and propane (5:1). The wt.% of C, Si, Al, O of the total charged metallic material and the oxygen lancing were also included to account for exothermic chemical reactions occurring during the process.
    The wt.% Fe, Cr, and Ni of the total charged metallic material were included to account for possible deviations in the process due to the steel grade being produced.
    The contents of the charged oxide bearing raw material are, of course, expected to impact the thermodynamics and kinetics of each reaction. However, there are more complicated factors in play between the metallic elements in the steel melt and the oxides in the slag. To account for these complex effects, the following oxides, in wt.% of total charged oxide bearing raw material, were included: C r 2 O 3 , M g O , C a O , F e O , S i O 2 , and A l 2 O 3 .
    It is known, by experience, that the specific heat per volume unit of oxygen is inversely proportional to total amount of lanced oxygen gas.
  • Charged material: The different material types are expected to contribute differently to the melting behavior and heat transfer of the scrap. Hence, all 8 of the available Material Types were included in the list of input variables. The Total Weight of the ingoing material was included because it is closely connected to the required energy to melt the steel. It was also divided into two separate variables, Metal Weight and Slag Weight, which represent the total weight of metallic material and total weight of oxide bearing raw material, respectively.
    The sum of all Material Types equals the Total Weight. Hence, the Material Types are expected to be correlated with the Total Weight. Furthermore, the sum of the Slag Weight and Metal Weight equals the Total Weight and are also expected to be correlated.
  • Energy: There are numerous energy variables that can be included in the list of input variables. However, most are in one way or another connected to the already mentioned variables. For example, heat loss is linearly related to time and amount of ingoing scrap is linearly related to the heat required to heat and melt the scrap.
    The preheating energy was included because the actual preheating is conducted before the start of the EAF process itself. In the steel plant of study, the scrap is preheated with partial off-gas from the EAF operation. Hydrocarbons such as grease and oil, and moisture entrenched in the scrap are burned off. Due to the varying amounts of moisture and hydrocarbons in the scrap, the preheating energy as ingoing variable is therefore motivated.
    EE is well defined in the transformer system and is subject to a negligible error. This means that the logged EE value can be trusted and is therefore also taken as the true value when training the statistical model.
The selected variables are shown in Table 3.
Correlations between the variables in the 4 categories are also expected. For example, EE demand should correlate with Total Weight. Another example is Material Types and chemical compositions as these material categories are partly based on chemical analyses.
The reason correlation is frequently referred to is because it is one of the main underlying factors that connects the input variables to the output variable by the trained statistical model. Furthermore, input variables that are correlated “soaks up” parts of the “correlation potential” with respect to the output variable. This is especially the case when the correlated input variables are correlated with the output variable. A simple example is the following correlative relationship between input variables A and B, and output variable C: A corr B corr C .

3.2.2. Variable Batches

The variables in each variable batch in Table 4 are explicitly shown in Table 5.
The motivation behind using a setup of variable batches is to investigate the true impact of each variable type. The impact of a variable on the model outcome can be diminished if that variable is partly represented by another variable, i.e., correlated, (see Section 2.4). One clear example is the percentage of C r and a Material Type containing high amounts of C r . Likewise, the total process time is partly represented by the melting time because the melting time adds to the total process time.
The decision to divide the variables into the specific variable batches shown in Table 4 was two-fold. First, there are multiple variables that represents the same effect of a physical phenomenon on the EE consumption, as motivated previously. Second, the number of model types to create given the number of combinations using each variable is very large ( 2 35 3.44 · 10 9 ) and it is not possible to model all variants within a reasonable time frame. Hence, bundling together variables that are of the same type, for example Material Types, will reduce the number of unique variable batch combinations to 64 ( 2 6 ).
The base variables were included in all variable batches because they are believed to be strongly related to the EE consumption. The specific variables in the base variable group are shown in Table 5.

3.2.3. Selection of Test Data

The test data is commonly selected from a random sample of the complete data set. However, in this regard, the training data and the test data become chronologically intertwined. This shortcoming has been discussed in previous research [43]. From a practical process perspective, a statistical model will predict on data from heats that are produced after the training of model. This means that these heats will always be from a future point in time. To consider this in a model evaluation, the test data should be selected in chronological order with respect to the training data. In the experiments, the test data will be from heats produced within 30 days after the last heat in the training data.

3.3. Data Treatment

3.3.1. Purpose

The purpose of data treatment is to clean, and possibly repair, erroneous data or remove data that is not part of regular production. Data treatment is a double-edged sword. On one hand, the aim of any given model is to predict well on as many heats as possible. On the other hand, including unrepresentative data makes it harder to optimize the model enough to make it practically useful.
There are two different categories for data treatment. One is domain-specific, where the values in the data are assessed to what is physically possible within the application of the model. For example, if the EAF has a maximum capacity of 80t, then any value above that value should be viewed as an erroneous value. The second strategy category is statistical outlier detection. For example, removing all values that are above 3 σ from the mean value of a variable that is normal distributed. However, it is important to consider that statistical outlier detection methods do not take the application domain into account. These methods must be used with caution and carefully assessed on a case-to-case basis.
The data used in the modeling was aggregated from two disparate data flows. The first data flow consists of raw material compositions, raw material types, total charge weight, metal and oxide bearing raw material weights, energy data, and added propane through burners and oxygen gas through lance. The second data flow consists of process time, times for different sub-processes during the EAF process, and delay times.
All data from the two data flows was gathered after the 31 March 2016 and before 18 October 2018.

3.3.2. Domain-Specific Methods

Data flow 1: 12.1% of the raw material entries lacked oxide compositions. These entries were repaired using previously recorded oxide compositions for the same material entries. This is not expected to dramatically affect the performance of the models. An average of 176 kg of added material per heat, amounting to an approximate 0.2% of the total weight, had not available (NA) values for Rawtype, oxide, and metal composition. These raw material types were categorized into a variable called Type N. The total number of heats removed due to a raw material input with zero or negative weight was 59. The resulting number of heats at the end of the flow was 12,805.
Data flow 2: Only heats with logged events “heat started” and “heat ended” were removed. This was to safeguard against any complications such an error would give rise to in the rest of the entries. No other data treatment was commenced in this stage. A total of 12,649 heats were gathered from this data flow.
Aggregated data: Aggregating the two data flows resulted in a total of 12,587 heats. The resulting number of heats are less than from both data flows due to mismatching overlap of heats. After the aggregation of the two data flows, domain-specific data treatment rules were applied and are shown in Table 6.

3.3.3. Statistical Methods

Statistical data cleaning methods were not applied to the data set in this study. There are three reasons for this:
  • A more robust outlier detection algorithm was used in an attempt to remove outliers [47]. However, the resulting data set became too small after applying the algorithm on just a few of the variables.
  • It makes sense to not apply statistical outlier detection methods on some variables. For example, the wt% content in the charged material can vary tremendously depending on what scrap types are available. Since multiple steel grades are produced using different charging strategies, the wt% of elements will also vary. Some stainless steel types need more wt% Cr and wt% Ni than others.
  • Most conventional outlier detection methods assume that the data is normal distributed, or normal distributed with various levels of skewness and kurtosis. This is not always the case in EAF production data, see Figure 4.

3.4. Modeling

3.4.1. Artificial Neural Networks (ANN)

ANN, as mentioned earlier, is one type of non-linear statistical model framework. The idea is to connect the input variables to the output variable using a network of nodes that are fully connected [48]. The first layer is the input layer and the last layer is the output layer. The intermediate layers are called hidden layers. See Figure 5 for an illustration of a simple ANN.
The number of hidden layers and the number of nodes in each hidden layer determine the complexity of the model. In each of the hidden layers, and in the output layer, each node multiplies a weight with each of the values from the nodes in the previous layer. The resulting values are summed together. This process can be mathematically expressed as:
s j = i = 1 P w i · x i
where P is the number of nodes in the previous layer and j is the jth node in the layer.
A function, known as activation function, is then applied on s j resulting in the value that the current node sends forward in the network. The hyperbolic tangent (tanh) and the logistic sigmoid functions are two commonly used activation functions.
During the training phase, the training data is sequentially fed into the network upon which the network weights are updated in the direction of minimizing the loss function. The loss function can be, for example, the mean squared error (MSE) or R-square ( R 2 ). As the output value is a function of all the weights in the network, it is also possible to express the loss function as a function of the values of those weights. The updating of the weights is also known as backpropagation, because the errors are propagated back through the network and the weights updated. Since the loss space is a function of all weights in the network, finding a good enough local minima requires a well-engineered algorithm. These types of algorithms are known as gradient-descent algorithms because their aim is to descent to the most optimal local minima in loss space [48].
The sequentially feeding of training data, and update of the weights, is done numerous iterations until the accuracy improvement vanishes. One also uses a validation set, a subsample of the training data, during the training phase to ensure that the model does not overfit on the training data. Overfitting means that the model has learned the training data so well that it fails to predict well on unseen data. This is one of the drawbacks with neural network models because of its ability to learn complex relations even though the relations are not of value to solving the prediction problem. The validation set is the data set that the model is benchmarked against during the training phase to account for this drawback. After the training phase is completed, the test data is predicted by the model and model performance metrics are calculated.

3.4.2. Model Performance Metrics

To compare the performance of the models, R 2 and the regular error metric will be used.
One should use the adjusted- R 2 formula if one aims to compare R 2 between models with different number of input variables. This is because each added predictor slightly increases the R 2 -value given that the number of data points is fixed [49]. The formula for the adjusted- R 2 can be written as follows
R ¯ 2 = 1 ( 1 R 2 ) n 1 n p 1
where R 2 is the standard R-square, n is the number of data points, and p is the number of input variables.
The regular error metric is advisable to use over the absolute error metric since an overestimated prediction is vastly different than an underestimated prediction in a practical steel plant context. The regular error metric for mean values is written as follows
E μ = 1 n i = 1 n ( y i y ^ i )
where y i is the true value, y ^ i as the predicted value, and i 1 , 2 , , n . The standard deviation, E σ , is derived from E μ the ordinary way.

3.4.3. Hyperparameter Optimization

Hyperparameter optimization aims to find the combination of parameters that creates the best model with respect to minimizing the model error [50]. A common method is to use a pre-specified grid of these parameters where the framework trains one, or more, models for each combination of the parameters. This specific method is also known as grid-search. For the experiments in this paper, parameters include model-specific parameters as well as application-/domain-specific parameters both of which together totals 36864 parameter combinations. See Table 7.
Parameters which were the same for all model combinations are explained below:
  • Validation fraction: 0.2 , which specifies the fraction of the training data used as validation set.
  • Gradient-descent algorithm: An adaptive learning rate optimization algorithm known as Adam. This algorithm has been shown to outperform other gradient-descent algorithms in a variety of models and datasets. The algorithm-specific parameter values were selected as recommended in the paper [51].
  • Early stopping set to True. Specified by the following items. 1. Number of iterations with no change = 20. Specifies how many iterations of increasing or equal performance are required before the training phase stops. 2. Tolerance = 10 7 . The tolerance of improvement to reset the number of iterations with no change.
Each combination of parameters represents one trained model type. Furthermore, each model type will be instanced 10 times to investigate the stability of the parameter selection and to reduce the impact of randomness. The metrics used to evaluate the model types based on the 10 instances are explained in Table 8.
To determine stability, the idea was to keep the R ¯ m i n 2 and R ¯ m a x 2 as close to the R ¯ μ 2 as possible. Hence, for filter out the best model type of each variable batch, the following algorithm was used.
  • Filter out any model that does not pass the following condition: R ¯ m a x 2 R ¯ m i n 2 0.05 . The motivation using this specific condition was to ensure that only stable models were selected while still enabling at least one model type from each variable batch to pass the condition.
  • Sort the models on decreasing R ¯ μ 2 .
  • Pick the first model in the list.
The reason adjusted R-square was used for determining stability is because the metric indicates goodness of fit for the model. Mean model error and standard deviation of model error were included to determine the best model out of models who have close to equal R-square.

3.4.4. Algorithmic Approach

In the algorithmic approach, process knowledge will not be the basis of choosing variable batches. By contrast, the variables in each batch will be chosen based on their one-to-one correlation with the EE consumption. In this manner, the approach will be mostly algorithmic except for the chosen setup of available variables, which are the same as in the domain-specific approach. The proposed algorithm is described below:
  • Calculate the pair-wise correlation values between the input variables and the output variable (EE consumption).
  • Use all input variables in the first model.
  • For each subsequent model, remove the input variable with lowest correlation value with respect to the output variable. Save the performance metric results from all models for further analysis.
In the experiments, dCor will be used as correlation metric due to its ability to detect non-linear correlative relations between variables. dCor is explained in Section 3.4.5.
The rest of the model-specific and domain-specific parameters will be the same as in the domain-specific approach. This results in a total of 35 · 2 = 70 domain-specific parameter combinations and a total of 2 · 3 · 48 = 288 model-specific parameter combinations. Using the algorithmic approach, a total of 70 · 288 = 20,160 model types were trained. To filter out the best model for each variable batch, 10 instances of each parameter combination were created and the same algorithm as in the domain-specific approach was used.

3.4.5. Model Analysis

To ensure the reliability of a model, its transparency must be highlighted. To achieve this, three different statistical methods will be used. Two of these methods, KS test and dCor, are model independent and point to reasons behind model performance deviation from the training and test data. The third method, FI, is model dependent and explains how important each input variable from the model’s viewpoint. By using FI in combination with dCor and KS tests, the goal is to home in on the variables that are the strongest reason behind model performance deviation between the training and test data.
dCor: Correlation metrics are used to investigate whether a variable is correlated with another variable. Correlated variables may or may not be causative. Nevertheless, two correlated variables may provide a hint to a possible causative relationship. Using domain-specific knowledge it is possible to assess the relative strength of the causative relation based on the correlation value, if there is evidence of such from physical deliberations.
The commonly used Pearson correlation metric only detects linear and monotonic relationships between two variables [52]. This severely limits the relevance of the correlation metric within the scope of this study since it is well known that parameters governing the EAF process are non-linear with respect to the EE consumption, (see Section 2.2 and Section 2.3).
Alternative correlation metrics, such as dCor, can detect both linear and non-linear relationships between variables [53]. The resulting mathematical expression for dCor is similar to the Pearson correlation coefficient:
d C o r ( V 1 , V 2 ) = d C o v ( V 1 , V 2 ) d V a r ( V 1 ) d V a r ( V 2 )
where d C o v ( V 1 , V 2 ) is the distance covariance and d V a r ( V 1 ) , d V a r ( V 2 ) is the distance variance of the random variables V 1 and V 2 , respectively. The square root of the latter is the distance standard deviations of V 1 and V 2 .
To calculate dCor, the task is first to calculate the n x n distance matrices of each random variable, V 1 and V 2 :
a j , k = | | V 1 , j V 1 , k | |
b j , k = | | V 2 , j V 2 , k | |
where j , k = 1 , 2 , , n and | | · | | is the Euclidean norm. Using the distance matrices, the doubly centered distances are calculated as:
A j , k = a j , k a ¯ j a ¯ k + a ¯
B j , k = b j , k b ¯ j b ¯ k + b ¯
where a ¯ j is the row mean of the distance matrix, a ¯ k is the column mean of the distance matrix, and a ¯ is the grand mean of the distance matrix for random variable V 1 . Analogous for random variable V 2 . The distance covariance is then calculated as the following arithmetic average:
d C o v 2 ( V 1 , V 2 ) = 1 n 2 j = 1 n k = 1 n A j , k B j , k
Analogous, the distance variance for V 1 and V 2 are:
d V a r 2 ( V 1 ) = d C o v 2 ( V 1 , V 1 ) = 1 n 2 j = 1 n k = 1 n A j , k 2
d V a r 2 ( V 2 ) = d C o v 2 ( V 2 , V 2 ) = 1 n 2 j = 1 n k = 1 n B j , k 2
The dCor metric assumes values between 0 and 1, where 0 implies that the variables are independent and 1 implies that the variables are equal.
It has also been theoretically proven that dCor asymptotically detect deviations from independence. This implies that dCor is, at least in theory, a solid tool to detect any relationship between two variables given that enough data is provided. Some alternative correlation metrics with the similar characteristics as dCor are Hoeffding’s D measure, HHG (Heller, Heller and Gorfine) measure and MI (Mutual Information) [52]. The use of any particular correlation metric depends in large on the number of observations and the type of relationship one intends to identify and the “nature” of the variables. This implicitly justifies that any correlation metric is not necessarily the “holy grail” of determining variable dependence [52]. It also means that a correlation metric may give low values if the pattern one intends to identify is challenging for the metric to detect. However, due to the limitations of the current study to the statistical modeling of the EE consumption in the EAF, dCor is chosen as the preferred metric due to its ability to detect both linear and non-linear relationships. Furthermore, for large number of observations, dCor has proven to be reliable for many different types of dependencies [52].
dCor will be calculated for each input variable to the output variable for both the training and the test data. A change in dCor between the training and test data indicates that the relation between the input variable and the output variable has changed. The relationship between the input variables and the output variable is what a statistical model learns using the training data. Thus, the change in dCor between the training and test data highlights the variable as a reason behind the model performance deviation.
In the analysis, input variable pairs with dCor values at or above 0.1 will be referred to as having relatively high correlation values. The reasoning behind this limit is two-fold. First, most of the dCor values between the input variables pairs were shown to be lower than 0.1 and most of the expected correlations between the input variable pairs was higher than 0.1. Second, to the authors knowledge, a clear guidance in the literature to what is considered a low or high dCor value does not exist.
KS tests: The KS-value is the maximum distance between two cumulative distribution functions (CDF). The CDF are either two samples of the same variable or one sample and one ideal CDF, such as the CDF for the normal distribution. Furthermore, the KS test is a non-parametric statistical test which means that it does not make any presumptions about the distributions governing the samples [54]. Variables from the EAF production are often from varying classes and superpositions of distributions, see Figure 4. The KS test is conducted by calculating the confidence, the p-value, of the KS-value under the null hypothesis, H 0 , that the two samples are from the same distribution. The KS-value takes values between 0 and 1 where 0 indicates the distributions are identical and 1 indicates that the distributions are completely different. The p-value is the probability that the two samples are in fact from the same distribution. Hence, a high KS-value with a low p-value strongly indicates that the two distributions underlying the two samples are different.
In the experiments, the two-sample KS test is of particular interest because it enables the study of the difference in two sampled distributions of the same variable. The two-sample KS test equations follows:
D n 1 , n 2 = s u p x | F n 1 ( x ) G n 2 ( x ) |
where F n 1 and G n 2 are the two distribution functions with n 1 and n 2 are the number of samples from the two distributions, respectively. x is the total sample space. s u p is the supremum. The two-sample KS test is illustrated in Figure 6.
The null hypothesis, which is that the samples come from the same distribution, is rejected if:
D n 1 , n 2 > c ( α ) n 1 + n 2 n 1 · n 2
where α is the significance level and c is the threshold value calculated using α and the cumulative KS distribution [55].
One drawback of KS test is that it is sensitive to any difference between the two distributions. A large, but local, difference between the distributions will be captured by the KS test, but may not necessarily be representative over the complete distribution space.
The two-sample KS test will be used to investigate the distribution difference between the training data and the test data for both the input variables and the output variable. Since the neural network weights are adapted to the training data, any significant difference in distribution between the training and the test data for any variable (more affect on the most important variables) will affect the performance of the model on previously unseen data (test data). The aim is to pinpoint variables that are a probable cause to the performance difference rather than to prove specific performance changes by each variable. The latter approach is difficult, if at all doable, partly due to the complex interrelations between the variables in the data.
KS-values above or equal to 0.2 with p-values at or below 0.05 will be considered to be variables of interest in the analysis. The main reason is to avoid capturing all variables while at the same time capturing variables that change significantly enough between the training and test data. The implications of this limit on the KS test will be discussed. Distributions plots for variables above the KS-value threshold can further help to identify if the difference between the training and test data is significant enough.
FI: It is important to investigate how much each input variable affects the output variable. A useful way to do this is to use interpretable machine learning algorithms. One such algorithm is called FI which ranks each input feature, e.g., variable, in order of importance with respect to all predictions. Hence, FI is known as a global interpretable machine learning model. Each variable is singly broken and the recorded error from the model on this new data set is compared with the error from the original, unbroken, data set. If the error is higher, then the input variable is of some predictive importance with respect to the output variable. If the error is statistically indifferent, then one can conclude that the input variable is of little importance to the model [56]. The algorithm is described as follows:
  • Train a model and record its error L ( X ) .
  • Permute one of the input variables, x ¯ j , in the input matrix X.
  • Apply the permuted input matrix, X j , to the trained model and record the error, L ( X j ) .
  • Repeat steps 2 and 3 for all input variables j 1 , 2 , , m .
  • Order all variables in the order of decreasing L ( X j ) .
  • Normalize all L ( X j ) on m a x ( L ( X j ) ) (optional).
where L is the model loss function, X is the complete and non-permuted input data matrix, and m is the total number of variables.
FI will be applied to both the training data and the test data. FI on training data tells us how much the model relies on each input variable for making predictions on new data. On the other hand, FI on test data reveals how important each input variable is to the actual model performance on unseen data [57]. Hence, both approaches can be viewed as complementary tools for model transparency. If the training and test data for a given variable come from the same distribution, then the FI value will be the same given that the number of data points is large enough. A larger deviation in FI indicates that the underlying relations between the input variables and the output variable have changed. FI will be calculated 20 times for each variable and data set to account for randomness in the permutations. The presented values will be the mean FI-values for each variable.

4. Results

4.1. Modeling

See Table 9 for the domain-specific cleaning results. Only 10% of the original data was removed after all cleaning steps.
In a practical application context, only the best model will be selected. However, some interesting observations were made when analyzing the results from both the domain approach and the algorithmic approach. Therefore, a total of 6 models were selected for further analysis. D1 was selected because it uses all input variables, D64 was selected because it uses only the base variables in the domain approach. D15, D17, and D31 were selected because they are the top 3 best performing models on the test data from the domain approach. A21 was selected because it is the best performing model from the algorithmic approach. Hence, we refer to D15, D17, D31, and A21, as the best models. The true EE consumption is plotted against the predicted EE consumption by models D15 and A21 in Figure 7. The performance of the selected models are presented in Table 10. The performance of the best model from each variable batch is presented in Figure A1 and Figure A2 for the domain approach and algorithmic approach, respectively.
A significant negative mean error on the test data compared to the training data can be observed for all selected models. This is also consistent for a majority of the best models from the variable batches, as can be seen in Figure A1 and Figure A2. The minimum and maximum errors are in the ranges of −3.7 MWh to −4.8 MWh and 5.4 MWh to 6.6 MWh on the test data, for the best models.
Comparing the performance from the best model to reported ANN models in the literature is shown in Table 11. The standard deviation of error is better than all reported models evaluated on test data, while the mean error is worse. The minimum and maximum error are comparable. However, the number of variables is significantly lower for the best model from the experiments done in this paper; 20 compared to 82 to 100 in the literature. Furthermore, the results from the models in the present paper are the average over 10 iterations. No such approach to combat model instability was done on the models in the literature. In addition, one has to keep in mind that the reported ANN models, evaluated on test data, in the literature lack R 2 for all models and % cleaned data for all but one model [43].
The best model from the algorithmic approach is almost as good as the best model from the domain approach with a difference in R ¯ μ 2 of only 0.025.
The model with the least amount of input variables, D64, has a R ¯ μ 2 that is only 0.001 higher than the R ¯ μ 2 for the model with the highest number of input variables, D1.
The performance on the test data is better than on the training data for D15, D31, D64, and D21. Significantly so for D31 and D64 where a difference in R ¯ μ 2 0.062 and 0.096 are observed. This is somewhat unexpected since a model usually performs worse on test data.

4.2. Model Analysis

Assuming a threshold p-value of 0.05, all variables except Delays, Type B, Type C, and Type F, are of particular interest for further analysis of the KS-values, see Table 12. MgO shows an unusually large KS-value of 0.85. Al, Metal Weight, Slag Weight, and Cr 2 O 3 have KS-values of 0.34, 0.41, 0.5, and 0.36, respectively. Then there is a plethora of variables that have KS-values between 0.20 and 0.30. From the Base variables these are, Total Weight, Propane, O 2 -lance, Preheater. From the time variables, Refining is the only variable. From the oxide composition, these variables are CaO, FeO, Al 2 O 3 . For the raw material types, Type N satisfies the criteria. EE consumption has a KS-value of 0.25. The distributions for the training and test data for EE consumption are shown in Figure 8.
The dCor between the input variables and the EE consumption are shown in Table 12. Changes in dCor above 0.1 between the training and test data are observed for Delays, TTT, Total Weight, Propane, Process Time, Extended refining, C, Si, O, Metal Weight, Type C, Type E, and Type N, raw materials.
Expected correlations between variables, as discussed in Section 2.4, can be observed in Table A4. dCor calculations between time variables and time variables have higher dCor values compared to dCor calculations between time variables against other variable types. The intra-correlation for metal composition, oxide compositions, and raw material types, are expected since the sum of the compositions are equal to one and the sum of the raw material types are bound by the total weight. The inter-correlation between metal compositions, oxide compositions, and the raw material types, indicates the complex relationship between these variables. Furthermore, the differences in dCor between the training and test data for all variables are explicitly shown in Table A3. While some larger changes in dCor can be observed for the time variables, the variables with the most number of larger dCor changes are the weight variables, metal composition, oxide composition, and raw material type variables.
FI for the 6 selected models are presented in Table 13. Delays, TTT, and Total Weight are the three most important variables for all models, except for model D15, with respect to the training data. Some significant changes in FI from the training and test data sets are also observed. These are marked with bold and underlined numbers in Table 13.
Training and test data distributions for some of the most important variables for the 6 selected models are shown in Figure 8. These are Delays, TTT, Process Time, Total Weight, and Metal Weight. EE is also plotted because of its KS-value of 0.25. Metal compositions, oxide compositions, and raw material types were not included because of their complex relationships shown by the relatively high inter-correlation and intra-correlation, (see Table A4 and Table A5). Analyzing their combined impact on the EE consumption is too extensive for this article.
The EE consumption have increased in the test data set. So have the TTT, Process Time, Total Weight, and Metal Weight. The delay is about equal on average between the training and test data sets.

4.3. Grid-Search Metadata

The parameters from the models passing the R ¯ m a x 2 R ¯ m i n 2 0.05 criterion (see Section 3.4.3) and the parameters from the best models are summarized in Table 14 and Table 15, respectively. Validating on chronologically ordered data during training is not always beneficial as the ordered is only present in 27% of the best models. Number of hidden layers, which is a proxy for model complexity, is equal to one for 87% of the best models. A similar reasoning can be made for the algorithmic approach, which metadata is presented in the same tables.

5. Discussion

5.1. Modeling

The best model from the domain approach has a R ¯ μ 2 of 0.706 while the best model from the algorithmic approach has a R ¯ μ 2 of 0.731. These values were calculated from the average of 10 model instances, and the difference is not large. This means that the algorithmic approach can be used as a tool to select variables that governs a model that will have a top-tier accuracy. However, it is important to consider the domain-specific reasoning behind the initial selection of the 35 available input variables. It is, therefore, not suggested that one should blindly select variables stored in an arbitrary EAF-database and then perform the algorithmic variable selection explained in this paper. While this could work, it is not what is intended using the algorithmic approach.
Using a large scale grid-search with over 57,000 model types, using both a domain-specific approach and an algorithmic approach, the best model type achieved an R μ 2 of 0.731, a mean error of −554 kWh/heat, a standard deviation of error of 1126 kWh, a minimum error of −3819 kWh/heat, and a maximum error of 5735 kWh/heat. The best model in the literature had a mean error of approximately zero, a standard deviation of error of 1300 kWh/heat, a minimum error of −3500 kWh/heat, and a maximum error of 6000 kWh/heat. The errors from both models are similar. However, the literature model did not report averages of 10 model iterations. Therefore, it is not possible to know if the performance is stable or purely based on luck. The models in this paper were selected after a filtering criterion, R ¯ m a x 2 R ¯ m i n 2 0.05 (see Section 3.4.3). After this filtering, the best models based on the highest R ¯ μ 2 on the test data were selected. In addition, the model from the literature did not report R μ 2 and was only evaluated on 20 test data points compared to 362 test data points for the best model produced in this paper. Assuming a production of 20 heats per day, this represents a model evaluation difference of approximately 17 production days. Furthermore, neither the data treatment specifics nor the percentage of cleaned data were specified for that model. Model reporting shortfalls, of which some are mentioned above, are frequently occurring in the available literature on the subject of statistical modeling to predict the EE consumption in the EAF. It is important to clearly describe all steps in the data- and modeling pipeline that impose changes to the end result. In the context of this paper, the end result is the best model. The best model presented in this paper is also less complex than the best reported model in the literature even though both models perform similarly. While the number of hidden layers are equal between both models, the model reported in the literature has 100 input variables compared to 20 input variables for the best model produced in this paper. In the interest of model parsimony, simplicity should be followed when creating any model. On the same note, the model using 6 of the base variables has an R μ 2 that is not significantly better (a difference of only 0.001) than the model using all 36 input variables, and the best model uses only 20 variables. Hence, adding a lot of input variables does not necessarily create a model with better performance. One has to search for the optimal combination of variables using either a purely domain-specific approach or an algorithmic approach, such as the one presented in this paper. Table 14 and Table 15 further support model simplicity since models with 2 layers are not preferred over models with 1 layers with respect to model stability (the R ¯ m a x 2 R ¯ m i n 2 0.05 criterion, see Section 3.4.3) or with respect to the best model from each variable batch.
3 of the 6 selected models perform better on the test data than on the training data. This is somewhat contradictory since models usually perform worse on test data. However, the selection of the best models from each variable batch was based on the performance on the test data. The performance on the training data was not considered since, in a practical environment, only the performance on test data is important.

5.2. Model Analysis

Investigating the dCor-matrix for the training data, Table A4, one can observe many relatively high 1-to-1 correlation among the metal composition, oxide composition, and raw material type variables. This is in line with what is expected from a process-metallurgical perspective. Because of this trait, these variables share information between one another to a large extent. For example, Type D is relatively highly correlated with all metal composition and oxide composition variables. Simplified, Type D has the potential to account for the metal and oxide composition which renders those variables less important to the models performance. This is the reason adding a lot of input variables does not necessarily create a model with better performance, and the reason variables that are important in a physical model may be next to worthless in a statistical model. Changes in dCor from the training data to the test data, see Table A3, indicate that the underlying relation between the variables have changed. This is not as frequently occurring for the time variables as for the metal composition, oxide composition, and raw material types where changes to both lower and higher dCor are present. Due to changes in dCor, changes in the model performance are also expected. However, analyzing a 36 × 36 correlation matrix in depth is both impractical and does not show which variables are the most important for the performance of the model.
Table 12 shows the results from the KS tests performed on each variable on training and test data sets. KS-values greater than or equal to 0.20 were considered of significant change given that the p-value is 0.05, or less. In conjunction with FI for the selected models, see Table 13, it is possible to identify reasons behind the performance differences from training to test. Delays, TTT, Process Time, Total Weight, and Metal Weight are some of the most important variables for the 6 selected models. Out of these variables, Total Weight and Metal Weight are variables with KS-values above 0.20. Process Time, TTT, and Delays, had KS-values of 0.18, 0.13, and 0.04, respectively. Furthermore, the KS-value for EE consumption is 0.25, which indicates that the EE consumption has also changed from the training to test data. However, KS-values do only provide a difference in the CDF and do not indicate whether the variable has increased or decreased. Observing the variable distributions for the training and test data for these variables in Figure 8, it is clear that all of these variables, except Delays, have increased from the training data to test data. Since the EE consumption increases with the amount of charged materials and longer process times, and because these input variables are important for the models, it is highly probable that these variables give rise to the decrease in Δ μ and the change in R ¯ μ 2 for the 6 selected models.
Please note that the limitation of demarcating “interesting” variables to only those at or above 0.2 KS-value becomes clear when observing the clear and consistent increase in Process Time and TTT in Figure 8. Nevertheless, using KS test in conjunction with FI and the additional distribution plots enabled the discovery of the implications on the model performance due to the changes in these variables.
Δ μ is defined as the average error for all predictions across all 10 model instances. Each error is in turn defined as the true value minus the predicted value. Since all 6 selected models get a significant negative Δ μ on the test data, the models overestimate the EE consumption due to the changes in the variables with the highest FI. In particular, the Total Weight, Metal Weight, TTT and Process Time. This behavior is not unexpected since any statistical model will predict the value of the highest probability with respect to the data it has used to adapt its parameters. In this case, the 10,966 training data points have adapted the ANN weights. Any change to the distribution of the variables with high FI will affect the models prediction on data sampled from this aggregated distribution.

6. Conclusions

One of the aims with the present work was to address the previously reported shortcomings in the statistical modeling approach to predict the EE consumption of the EAF [43]. In addition to addressing these shortcomings with a detailed data- and modeling pipeline, the choice of statistical modeling framework and statistical analysis tools was adapted to the non-linearity of the EAF process and the complex correlative relations between the selected process variables. The main conclusions of the modeling approach, and suggested future work, may be summarized as follows:

6.1. Modeling

  • In the interest of model parsimony, increasing the number of input variables does not automatically increase model performance. This is both observed when comparing the best model from the experiments to the best model reported in the literature and when comparing model D64, which uses 6 variables, and model D1, which uses 35 variables. It is sufficient to choose several variables from an initial selection demarcated by the use of process knowledge.
  • Similar performance was retrieved by the best models from the algorithmic approach and the domain approach, respectively. Given a setup of variables that are selected with process expertise, an algorithmic approach can find an optimal selection of variables. This is important from a practical applicability standpoint since less time need to be invested in searching for the optimal number of variables and the optimal subset of variables.
  • Neural networks with many layers, increasing the model complexity, are not necessary to produce state-of-the-art model performance. This was observed both in the grid-search performed in the experiments and in comparing the best model from the experiments with the best models reported in the literature.
  • The mean error is slightly higher, and the standard deviation of error is better for the best model compared to the best model reported in the literature. The minimum and maximum errors are approximately the same as the best model reported in the literature. However, these are −3.8 MWh to 5.7 MWh, respectively, which are quite high from a practical application context.

6.2. Model Analysis

  • Using KS test and FI as complementary tools, it was possible to identify an increase in Total Weight, Metal Weight, and Process Times as a highly probable cause behind the performance change from the training to test data for the 6 selected models.
  • An analysis of the Δ μ error from the selected models on the test data indicates that the models overestimate the EE consumption with regards to the change in the most important variables to the models. In particular, the Total Weight, Metal Weight, TTT, and Process Time variables.
  • High intra-correlation between the metal composition, oxide composition, and raw material type variables was found using dCor on the training data. This explains why some of these variables get a high FI in some of the models. The changes in dCor from the training to test data were also prominent for these variables and can explain part of the performance changes for the selected models attributing higher FI to some of these metal composition, oxide composition, and raw material type variables.

6.3. Future Work

Based on the present work carried out and the conclusions drawn, it is suggested to further investigate the following:
  • Use upsampling on the training data to get more data points as outliers. Could potentially solve the problem with large maximum and minimum errors.
  • Develop a more advanced variable selection algorithm based on dCor that takes into account one-to-many correlations for each input variable. The current algorithm only considers the correlation between the EE and each input variable.
  • Apply dCor to groups of variables and the EE consumption instead of variables in a 1-to-1 fashion. This feature of dCor should be explored further.
  • Use other non-linear correlation metrics to determine appropriate variable selection. These include Hoeffding’s, MI, and HHG.
  • Investigate the effects of different types of scrap on the EE consumption using new classifications based on density, and heat and melting behavior.
  • The effect on models by normalized variables that are not normally distributed is not known. See Figure 4 for a few examples.

Author Contributions

Conceptualization, L.S.C., P.B.S. and P.G.J.; Methodology, L.S.C. and P.B.S.; Validation, L.S.C., P.B.S. and P.G.J.; Formal analysis, L.S.C.; Investigation, L.S.C. and P.B.S.; Resources, P.B.S. and P.G.J.; Data curation, L.S.C.; Writing–original draft preparation, L.S.C.; Writing–review and editing, L.S.C., P.B.S. and P.G.J.; Visualization, L.S.C. and P.B.S.; Software, L.S.C.; Supervision, P.B.S. and P.G.J.; Project administration, P.G.J.; Funding acquisition, P.G.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by “Hugo Carlssons Stiftelse för vetenskaplig forskning” in the form of a scholarship granted to the corresponding author.

Acknowledgments

We want to thank the plant engineers Pär Ljungqvist, Christoffer Schmidt, and Jesper Janis at the Outokumpu Stainless Avesta mill for their support and data provisioning during this project.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

E t o t , i n Total ingoing energy
E E l Total Electrical Energy (EE) output from transformer
E C h e m Total energy from chemical reactions in steel and slag
E B u Total energy input from burner
E t o t , o u t Total outgoing energy
E S t e e l Total energy output into steel
E S l a g Total energy lost in slag
E G a s Total energy lost in gas
E D u s t Total energy lost in dust
E C o o l i n g Total energy lost in cooling water
E R a d Total energy lost through radiation
E C o n v Total energy lost through convection
E E l , l o s s Total energy lost in electrical system and arc transfer
T C S The temperature of the cooling panels
T E A F The temperature of the surface area subject to radiation losses
T s Temperature of ingoing material and gas at the start of the EAF process
T T a p Temperature of the steel at tapping
T O f f g a s Temperature of the off-gas leaving the EAF through the off-gas system
T H 2 O Temperature of the cooling water
T H The temperature of the surface area subject to convection losses
T A m b The temperature of the air surrounding the EAF
m S t e e l Mass of ingoing metallic material
m S l a g Mass of ingoing oxidic material
m ˙ D u s t Mass flow of dust in the off-gas system
c S t e e l The heat capacity of steel at constant pressure
c S l a g The heat capacity of slag at constant pressure
c D u s t The heat capacity of dust at constant pressure
c G a s The heat capacity of EAF ambient gas at constant pressure
c p ( r e a c t a n t s ) The heat capacity of reactants at constant pressure
c p ( p r o d u c t s ) The heat capacity of products at constant pressure
Δ H 298 o ( r e a c t a n t s ) Standard heat of formation for reactants at 298K
Δ H 298 o ( p r o d u c t s ) Standard heat of formation for products at 298K
Δ H M e l t , s t e e l Heat of fusion for steel
Δ H M e l t , s l a g Heat of fusion for slag
kConductivity of the cooling panels
hHeat transfer coefficient of the EAF ambient gas
ϵ Emissivity factor of the radiating surface area of the EAF
σ Stefan-Boltzmann constant
η E l Efficiency factor for the transformer system
η A r c Efficiency factor for the energy transferred from the arcs
η F u e l Efficiency factor for burning the fuel in the burners
h F u e l Heat generated per volume unit of fuel
V F u e l Volume of the fuel consumed by the burners
A H Surface area of the EAF subject to convection losses
A A C Surface area of the cooling panels
A E A F Surface area of the EAF subject to radiation losses
P A v g Average power of the transformer system
PFurnace pressure
MMolar mass of the furnace gas
RUniversal gas constant
V ˙ G a s Volume flow of gas in the off-gas system
t T T T Tap-to-tap time
t P O N Power-on time
R 2 Coefficient of determination
R ¯ 2 Coefficient of determination adjusted for number of data points and variables
nNumber of data points
pNumber of input variables
y i True value of the output variable for data point i
y ^ i Predicted value of the output variable for data point i
PNumber of nodes in the previous layer.
s j Summation of the input values for jth node in the current layer.
w i Weight of node i in the previous layer.
x i Value of node i in the previous layer.
Δ μ Mean error of the mean error of the 10 model instances on the test data
Δ σ Standard deviation of the mean error of the 10 model instances on the test data
Δ m i n Minimum of the mean error of the 10 model instances on the test data
Δ m a x Maximum of the mean error of the 10 model instances on the test data
R ¯ μ 2 Mean adjusted R-square of the 10 model instances
R ¯ σ 2 Standard deviation of adjusted R-square of the 10 model instances
R ¯ min 2 Minimum of adjusted R-square of the 10 model instances
R ¯ max 2 Maximum of adjusted R-square of the 10 model instances
V 1 Random variable
V 2 Random variable
d C o r ( V 1 , V 2 ) Distance correlation between V 1 and V 2
d C o v ( V 1 , V 2 ) Distance covariance between V 1 and V 2
d V a r ( V 1 ) Distance standard deviation for V 1
d V a r ( V 2 ) Distance standard deviation for V 2
d V a r 2 ( V 1 ) Distance variance for V 1
d V a r 2 ( V 2 ) Distance variance for V 2
a j , k Distance between values j and k for random variable V 1
b j , k Distance between values j and k for random variable V 2
V 1 , j Value j for random variable V 1
V 2 , j Value j for random variable V 2
A j , k Doubly centered distance for value j and k for random variable V 1
B j , k Doubly centered distance for value j and k for random variable V 2
a j , k
a j ¯ Row mean of the distance matrix for the random variable V 1
a k ¯ Column mean of the distance matrix for the random variable V 1
a ¯ Grand mean of the distance matrix for random variable V 1
b j , k
b j ¯ Row mean of the distance matrix for the random variable V 2
b k ¯ Column mean of the distance matrix for the random variable V 2
b ¯ Grand mean of the distance matrix for random variable V 2
n 1 Number of samples from the first distribution in the KS test
n 2 Number of samples from the second distribution in the KS test
D n 1 , n 2 KS-value from the KS test between F n 1 and G n 2
xThe total sample space in the KS test
F n 1 First distribution function in the two-sample KS test
G n 2 Second distribution function in the two-sample KS test
H 0 Null hypothesis
α Significance level
cThreshold value calculated from the cumulative KS distribution
LModel error function
XInput matrix
x ¯ j Input variable j
X j Input matrix with permuted variable j
E μ Mean error
E σ Standard deviation of error

Abbreviations

The following abbreviations are used in this manuscript:
CFDComputational Fluid Dynamics
EEElectrical Energy
EAFElectric Arc Furnace
ACAlternate Current
MLRMultivariate Linear Regression
ANNArtificial Neural Network
MSEMean Squared Error
FIFeature Importance
KSKolmogorov–Smirnov
CDFCumulative distribution function
dCorDistance correlation
MIMutual Information
HHGHeller, Heller and Gorfine
TTTTap-to-Tap Time
TRTraining
TETest
NANot Available

Appendix A

Appendix A.1. Hardware and Software

Table A1. Hardware specifications.
Table A1. Hardware specifications.
Computer modelDell Latitude E5570
CPUIntel Core i7 2376 MHz
RAM16,203 MB
Table A2. Software specifications for the experiments.
Table A2. Software specifications for the experiments.
PurposeSoftware/PackageVersion
Operating systemMicrosoft Windows 7 Professional6.1.7601 Service Pack 1 Build 7601
Programming languagePython 33.7.1
Python distributionAnaconda 34.6.7
Data handlingPandas0.23.4
Numpy1.17.4
Statistical modelingScikit-learn0.20.1
Feature importanceeli50.8.1
Distance correlationdcor0.3
KS testscipy0.15.0
VisualizationMatplotlib3.0.2

Appendix A.2. Variable Batch Performance

Figure A1. The modeling results from the domain approach. Top panels: Mean error plots for the best models of the 64 variable batches. The solid line with circles, Δ μ . The dashed line without circles, Δ σ . The dashed lines with circles, Δ m i n and Δ m a x , respectively. The solid line represents zero error and is used as reference for clarity purposes. The R ¯ μ 2 of each model. The model with the highest R ¯ μ 2 is model 15 with R ¯ μ 2 = 0.731 . Bottom panels: The difference in Δ μ and R ¯ μ 2 between the training and the test data.
Figure A1. The modeling results from the domain approach. Top panels: Mean error plots for the best models of the 64 variable batches. The solid line with circles, Δ μ . The dashed line without circles, Δ σ . The dashed lines with circles, Δ m i n and Δ m a x , respectively. The solid line represents zero error and is used as reference for clarity purposes. The R ¯ μ 2 of each model. The model with the highest R ¯ μ 2 is model 15 with R ¯ μ 2 = 0.731 . Bottom panels: The difference in Δ μ and R ¯ μ 2 between the training and the test data.
Metals 10 00036 g0a1
Figure A2. The modeling results from the algorithmic approach. Top panels: Mean error plots for the best models for each of the 36 models in the algorithmic approach. The solid line with circles, Δ μ . The dashed line without circles, Δ σ . The dashed lines with circles, Δ m i n and Δ m a x , respectively. The solid line represents zero error and is used as reference for clarity purposes. The R ¯ μ 2 of each model. The model with the highest R ¯ μ 2 is model 21 with R ¯ μ 2 = 0.716 . Bottom panels: The difference in Δ μ and R ¯ μ 2 between the training and the test data.
Figure A2. The modeling results from the algorithmic approach. Top panels: Mean error plots for the best models for each of the 36 models in the algorithmic approach. The solid line with circles, Δ μ . The dashed line without circles, Δ σ . The dashed lines with circles, Δ m i n and Δ m a x , respectively. The solid line represents zero error and is used as reference for clarity purposes. The R ¯ μ 2 of each model. The model with the highest R ¯ μ 2 is model 21 with R ¯ μ 2 = 0.716 . Bottom panels: The difference in Δ μ and R ¯ μ 2 between the training and the test data.
Metals 10 00036 g0a2

Appendix A.3. Distance Correlation (dCor) Matrices

Table A3. Difference in Variable-variable distance correlation (dCor) between the training and test data. dCor values at ± 0.1 , or above, is bold and underlined.
Table A3. Difference in Variable-variable distance correlation (dCor) between the training and test data. dCor values at ± 0.1 , or above, is bold and underlined.
DelaysTTTProcessChargingExt. Ref.MeltingRefiningTapping
Delays0.0−0.035−0.0540.018−0.112−0.041−0.063−0.048
TTT−0.0350.0−0.0460.036−0.121−0.022−0.076−0.028
Process−0.054−0.0460.00.074−0.212−0.009−0.059−0.009
Charging0.0180.0360.0740.0−0.058−0.088−0.078−0.059
Ext. ref.−0.112−0.121−0.212−0.0580.0−0.103−0.0150.007
Melting−0.041−0.022−0.009−0.088−0.1030.0−0.017−0.067
Ext. ref.−0.063−0.076−0.059−0.078−0.015−0.0170.0−0.002
Tapping−0.048−0.028−0.009−0.0590.007−0.067−0.0020.0
Propane−0.181−0.188−0.241−0.073−0.121−0.068−0.017−0.142
O 2 −0.044−0.056−0.044−0.081−0.06−0.031−0.037−0.024
TotWeight−0.0180.0070.081−0.003−0.054−0.007−0.0780.016
Metal−0.047−0.0050.087−0.031−0.055−0.016−0.0370.015
Slag−0.027−0.032−0.014−0.042−0.067−0.0160.013−0.112
%Fe−0.043−0.034−0.052−0.008−0.125−0.056−0.092−0.028
%O−0.052−0.055−0.065−0.045−0.06−0.06−0.039−0.073
%Al−0.055−0.039−0.031−0.053−0.053−0.042−0.115−0.021
%Cr−0.056−0.06−0.021−0.036−0.068−0.0560.022−0.049
%Si−0.048−0.055−0.077−0.058−0.058−0.052−0.113−0.093
%C−0.042−0.015−0.031−0.063−0.162−0.0730.079−0.018
%Ni−0.067−0.059−0.09−0.028−0.112−0.06−0.078−0.039
%FeO−0.066−0.046−0.035−0.08−0.035−0.062−0.067−0.039
%SiO 2 −0.044−0.041−0.031−0.041−0.068−0.057−0.116−0.058
%Al 2 O 3 −0.089−0.076−0.056−0.134−0.088−0.043−0.166−0.041
%Cr 2 O 3 −0.116−0.128−0.072−0.112−0.114−0.05−0.119−0.031
%Mgo−0.074−0.082−0.087−0.044−0.056−0.026−0.038−0.021
%CaO−0.039−0.029−0.026−0.048−0.065−0.045−0.084−0.054
Type A−0.061−0.062−0.054−0.051−0.059−0.073−0.229−0.041
Type B−0.033−0.03−0.055−0.053−0.088−0.038−0.016−0.057
Type C−0.006−0.016−0.049−0.012−0.035−0.078−0.192−0.054
Type D−0.036−0.026−0.038−0.033−0.063−0.077−0.137−0.03
Type E−0.049−0.054−0.056−0.047−0.058−0.062−0.015−0.088
Type F−0.034−0.022−0.023−0.043−0.02−0.042−0.071−0.096
Type G−0.06−0.07−0.161−0.038−0.112−0.081−0.202−0.076
Type N−0.133−0.164−0.127−0.048−0.126−0.087−0.091−0.031
PreHeater−0.055−0.055−0.029−0.069−0.066−0.052−0.151−0.066
EE−0.135−0.134−0.135−0.07−0.2260.02−0.0510.023
PropaneO2TotWeightMetalSlag%Fe%O%Al
Delays−0.181−0.044−0.018−0.047−0.027−0.043−0.052−0.055
TTT−0.188−0.0560.007−0.005−0.032−0.034−0.055−0.039
Process−0.241−0.0440.0810.087−0.014−0.052−0.065−0.031
Charging−0.073−0.081−0.003−0.031−0.042−0.008−0.045−0.053
Remelting−0.121−0.06−0.054−0.055−0.067−0.125−0.06−0.053
Melting−0.068−0.031−0.007−0.016−0.016−0.056−0.06−0.042
Ext. ref.−0.017−0.037−0.078−0.0370.013−0.092−0.039−0.115
Tapping−0.142−0.0240.0160.015−0.112−0.028−0.073−0.021
Propane0.00.1740.2010.197−0.057−0.077−0.1440.012
O 2 0.1740.00.1350.141−0.025−0.048−0.049−0.022
TotWeight0.2010.1350.00.001−0.219−0.08−0.2090.076
Metal0.1970.1410.0010.0−0.058−0.186−0.154−0.096
Slag−0.057−0.025−0.219−0.0580.00.098−0.0660.155
%Fe−0.077−0.048−0.08−0.1860.0980.0−0.0750.025
%O−0.144−0.049−0.209−0.154−0.066−0.0750.0−0.089
%Al0.012−0.0220.076−0.0960.1550.025−0.0890.0
%Cr−0.038−0.020.044−0.1250.0180.133−0.038−0.066
%Si−0.0250.045−0.219−0.041−0.094−0.183−0.326−0.059
%C0.056−0.126−0.012−0.034−0.046−0.007−0.068−0.004
%Ni−0.058−0.055−0.137−0.2540.118−0.058−0.0810.033
%FeO0.0620.0490.09−0.0660.141−0.055−0.139−0.05
%SiO 2 0.017−0.0030.143−0.1310.222−0.122−0.145−0.046
%Al 2 O 3 −0.168−0.059−0.011−0.125−0.054−0.125−0.296−0.133
%Cr 2 O 3 −0.127−0.0980.056−0.1940.193−0.182−0.088−0.108
%MgO−0.0350.065−0.018−0.0790.117−0.180−0.2360.030
%CaO0.0380.0140.123−0.1450.264−0.119−0.223−0.052
Type A−0.036−0.042−0.211−0.292−0.025−0.171−0.093−0.224
Type B−0.057−0.047−0.089−0.099−0.036−0.168−0.214−0.095
Type C−0.020.003−0.110−0.119−0.085−0.054−0.15−0.071
Type D−0.047−0.019−0.140−0.1440.010.0−0.094−0.109
Type E−0.123−0.02−0.160−0.1300.043−0.117−0.189−0.027
Type F−0.0250.018−0.02−0.036−0.034−0.096−0.114−0.038
Type G−0.0280.150−0.2050.0060.101−0.366−0.344−0.241
Type N0.1640.2440.1060.135−0.089−0.108−0.104−0.036
PreHeater−0.055−0.035−0.066−0.043−0.137−0.095−0.06−0.086
EE−0.1350.0090.2090.208−0.071−0.086−0.225−0.042
%Cr%Si%C%Ni%FeO%SiO2%Al2O3
Delays−0.056−0.048−0.042−0.067−0.066−0.044−0.089
TTT−0.06−0.055−0.015−0.059−0.046−0.041−0.076
Process−0.021−0.077−0.031−0.09−0.035−0.031−0.056
Charging−0.036−0.058−0.063−0.028−0.08−0.041−0.134
Ext. ref.−0.068−0.058−0.162−0.112−0.035−0.068−0.088
Melting−0.056−0.052−0.073−0.06−0.062−0.057−0.043
Refining0.022−0.1130.079−0.078−0.067−0.116−0.166
TT−0.049−0.093−0.018−0.039−0.039−0.058−0.041
Propane−0.038−0.0250.056−0.0580.0620.017−0.168
O 2 −0.020.045−0.126−0.0550.049−0.003−0.059
TotWeight0.044−0.219−0.012−0.1370.090.143−0.011
Metal−0.125−0.041−0.034−0.254−0.066−0.131−0.125
Slag0.018−0.094−0.0460.1180.1410.222−0.054
%Fe0.133−0.183−0.007−0.058−0.055−0.122−0.125
%O−0.038−0.326−0.068−0.081−0.139−0.145−0.296
%Al−0.066−0.059−0.0040.033−0.05−0.046−0.133
%Cr0.0−0.0660.0280.016−0.136−0.092−0.135
%Si−0.0660.0−0.146−0.184−0.101−0.225−0.337
%C0.028−0.1460.0−0.040.0240.035−0.069
%Ni0.016−0.184−0.040.0−0.047−0.065−0.130
%FeO−0.136−0.1010.024−0.0470.0−0.040.025
%SiO 2 −0.092−0.2250.035−0.065−0.040.0−0.082
%Al 2 O 3 −0.135−0.337−0.069−0.1300.025−0.0820.0
%Cr 2 O 3 −0.009−0.258−0.006−0.118−0.087−0.0910.116
%MgO−0.058−0.1960.025−0.1580.053−0.016−0.351
%CaO−0.077−0.2380.045−0.053−0.110−0.071−0.117
Type A−0.143−0.174−0.065−0.185−0.244−0.0490.129
Type B−0.084−0.213−0.053−0.106−0.089−0.165−0.177
Type C0.003−0.1550.085−0.0580.027−0.06−0.261
Type D−0.004−0.1310.183−0.017−0.05−0.039−0.194
Type E−0.015−0.217−0.029−0.087−0.165−0.197−0.238
Type F−0.021−0.165−0.041−0.065−0.018−0.056−0.036
Type G−0.212−0.231−0.154−0.291−0.112−0.339−0.122
Type N−0.0430.0680.021−0.126−0.028−0.056−0.094
PreHeater−0.092−0.04−0.112−0.0960.043−0.036−0.119
EE−0.008−0.1110.129−0.09−0.091−0.032−0.08
%Cr2O3%MgO%CaOType AType BType CType DType E
Delays−0.116−0.074−0.039−0.061−0.033−0.006−0.036−0.049
TTT−0.128−0.082−0.029−0.062−0.03−0.016−0.026−0.054
Process−0.072−0.087−0.026−0.054−0.055−0.049−0.038−0.056
Charging−0.112−0.044−0.048−0.051−0.053−0.012−0.033−0.047
Ext. ref.−0.114−0.056−0.065−0.059−0.088−0.035−0.063−0.058
Melting−0.05−0.026−0.045−0.073−0.038−0.078−0.077−0.062
Refining−0.119−0.038−0.084−0.229−0.016−0.192−0.137−0.015
Tapping−0.031−0.021−0.054−0.041−0.057−0.054−0.03−0.088
Propane−0.127−0.0350.038−0.036−0.057−0.02−0.047−0.123
O 2 −0.0980.0650.014−0.042−0.0470.003−0.019−0.02
TotWeight0.056−0.0180.123−0.211−0.089−0.110−0.140−0.160
Metal−0.194−0.079−0.145−0.292−0.099−0.119−0.144−0.130
Slag0.1930.1170.264−0.025−0.036−0.0850.010.043
%Fe−0.182−0.180−0.119−0.171−0.168−0.0540.0−0.117
%O−0.088−0.236−0.223−0.093−0.214−0.150−0.094−0.189
%Al−0.1080.03−0.052−0.224−0.095−0.071−0.109−0.027
%Cr−0.009−0.058−0.077−0.143−0.0840.003−0.004−0.015
%Si−0.258−0.196−0.238−0.174−0.213−0.155−0.131−0.217
%C−0.0060.0250.045−0.065−0.0530.0850.183−0.029
%Ni−0.118−0.158−0.053−0.185−0.106−0.058−0.017−0.087
%FeO−0.0870.053−0.110−0.244−0.0890.027−0.05−0.165
%SiO 2 −0.091−0.016−0.071−0.049−0.165−0.06−0.039−0.197
%Al 2 O 3 0.116−0.351−0.1170.129−0.177−0.261−0.194−0.238
%Cr 2 O 3 0.0−0.226−0.0520.03−0.141−0.136−0.017−0.173
%Mgo−0.2260.0−0.254−0.126−0.206−0.033−0.044−0.164
%CaO−0.052−0.2540.0−0.076−0.236−0.003−0.008−0.243
Type A0.03−0.126−0.0760.0−0.106−0.225−0.226−0.127
Type B−0.141−0.206−0.236−0.1060.00.011−0.025−0.210
Type C−0.136−0.033−0.003−0.2250.0110.0−0.122−0.065
Type D−0.017−0.044−0.008−0.226−0.025−0.1220.0−0.01
Type E−0.173−0.164−0.243−0.127−0.210−0.065−0.010.0
Type F0.017−0.034−0.058−0.065−0.0280.006−0.084−0.118
Type G−0.398−0.048−0.302−0.256−0.072−0.209−0.281−0.230
Type N−0.088−0.086−0.054−0.107−0.054−0.039−0.063−0.078
PreHeater−0.1790.027−0.05−0.112−0.062−0.091−0.124−0.015
EE−0.004−0.048−0.074−0.03−0.069−0.141−0.084−0.172
Type FType GType NPreHeaterEE
Delays−0.034−0.06−0.133−0.055−0.135
TTT−0.022−0.07−0.164−0.055−0.134
Process−0.023−0.161−0.127−0.029−0.135
Charging−0.043−0.038−0.048−0.069−0.07
Ext. ref.−0.02−0.112−0.126−0.066−0.226
Melting−0.042−0.081−0.087−0.0520.02
Refining−0.071−0.202−0.091−0.151−0.051
Tapping−0.096−0.076−0.031−0.0660.023
Propane−0.025−0.0280.164−0.055−0.135
O 2 0.0180.1500.244−0.0350.009
TotWeight−0.02−0.2050.106−0.0660.209
Metal−0.0360.0060.135−0.0430.208
Slag−0.0340.101−0.089−0.137−0.071
%Fe−0.096−0.366−0.108−0.095−0.086
%O−0.114−0.344−0.104−0.06−0.225
%Al−0.038−0.241−0.036−0.086−0.042
%Cr−0.021−0.212−0.043−0.092−0.008
%Si−0.165−0.2310.068−0.04−0.111
%C−0.041−0.1540.021−0.1120.129
%Ni−0.065−0.291−0.126−0.096−0.09
%FeO−0.018−0.112−0.0280.043−0.091
%SiO 2 −0.056−0.339−0.056−0.036−0.032
%Al 2 O 3 −0.036−0.122−0.094−0.119−0.08
%Cr 2 O 3 0.017−0.398−0.088−0.179−0.004
%MgO−0.034−0.048−0.0860.027−0.048
%CaO−0.058−0.302−0.054−0.05−0.074
Type A−0.065−0.256−0.107−0.112−0.03
Type B−0.028−0.072−0.054−0.062−0.069
Type C0.006−0.209−0.039−0.091−0.141
Type D−0.084−0.281−0.063−0.124−0.084
Type E−0.118−0.230−0.078−0.015−0.172
Type F0.0−0.195−0.02−0.028−0.005
Type G−0.1950.00.116−0.068−0.004
Type N−0.020.1160.0−0.126−0.205
PreHeater−0.028−0.068−0.1260.0−0.014
EE−0.005−0.004−0.205−0.0140.0
Table A4. Variable-variable dCor for the training data. dCor values at ± 0.1 , or above, is bold and underlined.
Table A4. Variable-variable dCor for the training data. dCor values at ± 0.1 , or above, is bold and underlined.
DelaysTTTProcessChargingExt. Ref.MeltingRefiningTappingPropane
Delays1.00.9440.4880.5570.1480.1310.0540.1990.09
TTT0.9441.00.5580.5570.1890.1460.1390.2180.102
Process0.4880.5581.00.3510.290.2030.3090.440.212
Charging0.5570.5570.3511.00.0130.0290.0240.0250.072
Ext. ref.0.1480.1890.290.0131.00.0220.0480.0720.051
Melting0.1310.1460.2030.0290.0221.00.1030.0180.066
Refining0.0540.1390.3090.0240.0480.1031.00.1150.144
Tapping0.1990.2180.4400.0250.0720.0180.1151.00.118
Propane0.090.1020.2120.0720.0510.0660.1440.1181.0
O 2 0.0420.0290.0440.0520.0330.0460.0980.0570.322
TotalWeight0.0570.0820.2060.070.0140.1100.0780.1310.301
Met0.0430.0820.1860.0530.0110.1060.1640.1120.298
Slag0.0350.0420.0680.0340.0180.0670.2080.0540.089
%Fe0.0430.060.0660.0720.0450.0590.1510.0520.066
%O0.0330.0410.0360.0370.0130.0290.1340.080.045
%Al0.0460.0510.0630.0550.0330.0550.1530.0620.124
%Cr0.050.0450.0710.0860.0390.0620.1980.0510.059
%Si0.0460.0340.040.0310.0310.0630.0670.0490.192
%C0.0320.0650.0740.0260.030.040.1930.0570.159
%Ni0.0340.050.0390.0520.0390.0590.1810.0440.052
%FeO0.0380.0550.0420.0570.040.0380.1240.0360.166
%SiO 2 0.0460.060.0660.0570.0310.0620.1870.0380.105
%Al 2 O 3 0.0210.0290.0510.0240.0180.0470.0490.0370.059
%Cr 2 O 3 0.0270.030.0550.030.0160.0610.1270.0820.056
%MgO0.0260.0330.0370.0590.0170.0540.1290.0710.086
%CaO0.0470.0620.0670.0540.0290.0690.1850.0410.127
Type A0.0170.0260.0290.0330.0140.0350.1070.050.073
Type B0.0310.030.0290.0150.0130.0250.0680.0380.04
Type C0.0540.0790.0620.0630.0730.0350.1810.0530.096
Type D0.0470.0870.0850.0490.060.0360.2350.0670.068
Type E0.0410.0430.0450.0350.0140.0470.1340.0470.059
Type F0.0170.0220.0250.0160.0140.020.0450.0240.031
Type G0.0190.0260.0320.0360.0280.0370.060.0710.092
Type N0.0250.0340.070.0420.0480.0430.1080.1180.278
PreHeater0.0210.0320.0510.0220.0250.0290.0550.0520.132
EE0.1350.2620.3990.0250.2370.1070.4480.1520.082
O 2 TotWeightMetalSlag%Fe%O%Al%Cr%Si
Delays0.0420.0570.0430.0350.0430.0330.0460.050.046
TTT0.0290.0820.0820.0420.060.0410.0510.0450.034
Process0.0440.2060.1860.0680.0660.0360.0630.0710.04
Charging0.0520.070.0530.0340.0720.0370.0550.0860.031
Ext. ref.0.0330.0140.0110.0180.0450.0130.0330.0390.031
Melting0.0460.1100.1060.0670.0590.0290.0550.0620.063
Refining0.0980.0780.1640.2080.1510.1340.1530.1980.067
Tapping0.0570.1310.1120.0540.0520.080.0620.0510.049
Propane0.3220.3010.2980.0890.0660.0450.1240.0590.192
O 2 1.00.2350.2620.0670.0560.0470.0710.0550.137
TotWeight0.2351.00.7930.3420.2730.1510.2740.2600.154
Metal0.2620.7931.00.1700.3090.1730.1320.2200.274
Slag0.0670.3420.1701.00.3410.3790.5000.3540.287
%Fe0.0560.2730.3090.3411.00.5330.3360.5710.426
%O0.0470.1510.1730.3790.5331.00.2060.2660.603
%Al0.0710.2740.1320.5000.3360.2061.00.4790.174
%Cr0.0550.2600.2200.3540.5710.2660.4791.00.233
%Si0.1370.1540.2740.2870.4260.6030.1740.2331.0
%C0.0420.1120.1510.1860.2740.1290.2000.2690.071
%Ni0.0510.2460.2490.3870.8220.3770.4500.4880.241
%FeO0.1340.2280.1840.3560.3560.2000.6110.3330.225
%SiO 2 0.0860.3580.2030.5700.4290.2550.6120.4050.205
%Al 2 O 3 0.0480.2220.0910.2430.2690.1770.1180.1260.154
%Cr 2 O 3 0.040.3180.1160.4470.3180.2480.2710.3080.096
%MgO0.1380.1840.1500.3280.1320.1580.2120.1370.151
%CaO0.0990.3210.2050.5770.4340.2390.5850.4110.263
Type A0.030.2450.1090.4260.3180.2310.2530.1840.100
Type B0.0380.0850.1010.1150.2960.1400.0990.1460.165
Type C0.1270.2550.2140.3890.4500.2590.4840.4240.176
Type D0.0790.2260.2300.5470.5480.3080.5080.5330.196
Type E0.070.1000.1790.3610.5090.6580.2540.3540.655
Type F0.0590.0340.0570.0740.0820.0810.040.0560.046
Type G0.2550.0860.2380.3890.1110.2350.0630.0610.388
Type N0.3340.2810.2860.0490.0510.0440.0760.0550.203
PreHeater0.0540.0630.0910.0590.0350.0480.0630.0290.07
EE0.1230.3250.3440.1080.0950.0320.0980.1310.105
%C%Ni%FeO%SiO2 %Al2O3%Cr2O3%Mgo%CaO
Delays0.0320.0340.0380.046 0.0210.0270.0260.047
TTT0.0650.050.0550.06 0.0290.030.0330.062
Process0.0740.0390.0420.066 0.0510.0550.0370.067
Charging0.0260.0520.0570.057 0.0240.030.0590.054
Ext. ref.0.030.0390.040.031 0.0180.0160.0170.029
Melting0.040.0590.0380.062 0.0470.0610.0540.069
Refining0.1930.1810.1240.187 0.0490.1270.1290.185
Tapping0.0570.0440.0360.038 0.0370.0820.0710.041
Propane0.1590.0520.1660.105 0.0590.0560.0860.127
O 2 0.0420.0510.1340.086 0.0480.040.1380.099
TotWeight0.1120.2460.2280.358 0.2220.3180.1840.321
Metal0.1510.2490.1840.203 0.0910.1160.1500.205
Slag0.1860.3870.3560.570 0.2430.4470.3280.577
%Fe0.2740.8220.3560.429 0.2690.3180.1320.434
%O0.1290.3770.2000.255 0.1770.2480.1580.239
%Al0.2000.4500.6110.612 0.1180.2710.2120.585
%Cr0.2690.4880.3330.405 0.1260.3080.1370.411
%Si0.0710.2410.2250.205 0.1540.0960.1510.263
%C1.00.2370.1250.195 0.0680.1590.1300.209
%Ni0.2371.00.3600.434 0.2360.3490.1520.426
%FeO0.1250.3601.00.719 0.2430.1740.2510.667
%SiO 2 0.1950.4340.7191.0 0.3110.5300.3340.891
%Al 2 O 3 0.0680.2360.2430.311 1.00.6310.1210.343
%Cr 2 O 3 0.1590.3490.1740.530 0.6311.00.1450.524
%MgO0.1300.1520.2510.334 0.1210.1451.00.177
%CaO0.2090.4260.6670.891 0.3430.5240.1771.0
Type A0.1410.3220.2270.495 0.4480.5080.2250.450
Type B0.0660.2630.2120.191 0.0840.0820.1190.179
Type C0.2880.4400.4920.496 0.0920.2630.1830.470
Type D0.3550.5270.5240.603 0.1380.3850.2120.576
Type E0.1460.3730.2200.264 0.2430.2170.1770.296
Type F0.0310.0620.0690.084 0.0560.1050.0470.076
Type G0.1060.0890.2250.175 0.0990.0670.2190.180
Type N0.1260.040.1100.073 0.0390.0360.1180.064
PreHeater0.0940.0350.1280.106 0.0350.0260.1030.089
EE0.2260.0970.0320.079 0.1300.0950.1930.057
Type AType BType CType DType EType FType GType NPreHeater
Delays0.0170.0310.0540.0470.0410.0170.0190.0250.021
TTT0.0260.030.0790.0870.0430.0220.0260.0340.032
Process0.0290.0290.0620.0850.0450.0250.0320.070.051
Charging0.0330.0150.0630.0490.0350.0160.0360.0420.022
Ext. ref.0.0140.0130.0730.060.0140.0140.0280.0480.025
Melting0.0350.0250.0350.0360.0470.020.0370.0430.029
Refining0.1070.0680.1810.2350.1340.0450.060.1080.055
Tapping0.050.0380.0530.0670.0470.0240.0710.1180.052
Propane0.0730.040.0960.0680.0590.0310.0920.2780.132
O 2 0.030.0380.1270.0790.070.0590.2550.3340.054
TotWeight0.2450.0850.2550.2260.1000.0340.0860.2810.063
Metal0.1090.1010.2140.2300.1790.0570.2380.2860.091
Slag0.4260.1150.3890.5470.3610.0740.3890.0490.059
%Fe0.3180.2960.4500.5480.5090.0820.1110.0510.035
%O0.2310.1400.2590.3080.6580.0810.2350.0440.048
%Al0.2530.0990.4840.5080.2540.040.0630.0760.063
%Cr0.1840.1460.4240.5330.3540.0560.0610.0550.029
%Si0.1000.1650.1760.1960.6550.0460.3880.2030.07
%C0.1410.0660.2880.3550.1460.0310.1060.1260.094
%Ni0.3220.2630.4400.5270.3730.0620.0890.040.035
%FeO0.2270.2120.4920.5240.2200.0690.2250.1100.128
%SiO 2 0.4950.1910.4960.6030.2640.0840.1750.0730.106
%Al 2 O 3 0.4480.0840.0920.1380.2430.0560.0990.0390.035
%Cr 2 O 3 0.5080.0820.2630.3850.2170.1050.0670.0360.026
%MgO0.2250.1190.1830.2120.1770.0470.2190.1180.103
%CaO0.4500.1790.4700.5760.2960.0760.1800.0640.089
Type A1.00.0750.3290.4140.1940.0540.050.0420.035
Type B0.0751.00.2520.1980.2050.0560.1580.0340.039
Type C0.3290.2521.00.8070.2940.1160.1120.0840.055
Type D0.4140.1980.8071.00.3720.0430.070.0430.041
Type E0.1940.2050.2940.3721.00.0590.2570.1070.075
Type F0.0540.0560.1160.0430.0591.00.0260.0260.03
Type G0.050.1580.1120.070.2570.0261.00.2380.032
Type N0.0420.0340.0840.0430.1070.0260.2381.00.062
PreHeater0.0350.0390.0550.0410.0750.030.0320.0621.0
EE0.0640.1110.1500.1870.1280.0410.1510.0980.06
EE
Delays0.135
TTT0.262
Process0.399
Charging0.025
Ext. ref.0.237
Melting0.107
Refining0.448
Tapping0.152
Propane0.082
O 2 0.123
TotWeight0.325
Metal0.344
Slag0.108
%Fe0.095
%O0.032
%Al0.098
%Cr0.131
%Si0.105
%C0.226
%Ni0.097
%FeO0.032
%SiO 2 0.079
%Al 2 O 3 0.130
%Cr 2 O 3 0.095
%MgO0.193
%CaO0.057
Type A0.064
Type B0.111
Type C0.150
Type D0.187
Type E0.128
Type F0.041
Type G0.151
Type N0.098
PreHeater0.06
EE1.0
Table A5. Variable-variable distance correlation for the test data.
Table A5. Variable-variable distance correlation for the test data.
DelaysTTTProcessChargingExt. Ref.MeltingRefiningTapping
Delays1.00.9790.5420.5390.260.1720.1170.247
TTT0.9791.00.6040.5210.310.1680.2150.246
Process0.5420.6041.00.2770.5020.2120.3680.449
Charging0.5390.5210.2771.00.0710.1170.1020.084
Ext. ref.0.260.310.5020.0711.00.1250.0630.065
Melting0.1720.1680.2120.1170.1251.00.120.085
Refining0.1170.2150.3680.1020.0630.121.00.117
Tapping0.2470.2460.4490.0840.0650.0850.1171.0
Propane0.2710.290.4530.1450.1720.1340.1610.26
O 2 0.0860.0850.0880.1330.0930.0770.1350.081
TotWeight0.0750.0750.1250.0730.0680.1170.1560.115
Metal0.090.0870.0990.0840.0660.1220.2010.097
Slag0.0620.0740.0820.0760.0850.0830.1950.166
%Fe0.0860.0940.1180.080.170.1150.2430.08
%O0.0850.0960.1010.0820.0730.0890.1730.153
%Al0.1010.090.0940.1080.0860.0970.2680.083
%Cr0.1060.1050.0920.1220.1070.1180.1760.1
%Si0.0940.0890.1170.0890.0890.1150.180.142
%C0.0740.080.1050.0890.1920.1130.1140.075
%Ni0.1010.1090.1290.080.1510.1190.2590.083
%FeO0.1040.1010.0770.1370.0750.10.1910.075
%SiO 2 0.090.1010.0970.0980.0990.1190.3030.096
%Al 2 O 3 0.110.1050.1070.1580.1060.090.2150.078
%Cr 2 O 3 0.1430.1580.1270.1420.130.1110.2460.113
%MgO0.10.1150.1240.1030.0730.080.1670.092
%CaO0.0860.0910.0930.1020.0940.1140.2690.095
Type A0.0780.0880.0830.0840.0730.1080.3360.091
Type B0.0640.060.0840.0680.1010.0630.0840.095
Type C0.060.0950.1110.0750.1080.1130.3730.107
Type D0.0830.1130.1230.0820.1230.1130.3720.097
Type E0.090.0970.1010.0820.0720.1090.1490.135
Type F0.0510.0440.0480.0590.0340.0620.1160.12
Type G0.0790.0960.1930.0740.140.1180.2620.147
Type N0.1580.1980.1970.090.1740.130.1990.149
PreHeater0.0760.0870.080.0910.0910.0810.2060.118
EE0.270.3960.5340.0950.4630.0870.4990.129
PropaneO 2 TotWeightMetalSlag%Fe%O%Al
Delays0.2710.0860.0750.090.0620.0860.0850.101
TTT0.290.0850.0750.0870.0740.0940.0960.09
Process0.4530.0880.1250.0990.0820.1180.1010.094
Charging0.1450.1330.0730.0840.0760.080.0820.108
Ext. ref.0.1720.0930.0680.0660.0850.170.0730.086
Melting0.1340.0770.1170.1220.0830.1150.0890.097
Refining0.1610.1350.1560.2010.1950.2430.1730.268
Tapping0.260.0810.1150.0970.1660.080.1530.083
Propane1.00.1480.10.1010.1460.1430.1890.112
O 2 0.1481.00.10.1210.0920.1040.0960.093
TotWeight0.10.11.00.7920.5610.3530.360.198
Metal0.1010.1210.7921.00.2280.4950.3270.228
Slag0.1460.0920.5610.2281.00.2430.4450.345
%Fe0.1430.1040.3530.4950.2431.00.6080.311
%O0.1890.0960.360.3270.4450.6081.00.295
%Al0.1120.0930.1980.2280.3450.3110.2951.0
%Cr0.0970.0750.2160.3450.3360.4380.3040.545
%Si0.2170.0920.3730.3150.3810.6090.9290.233
%C0.1030.1680.1240.1850.2320.2810.1970.204
%Ni0.110.1060.3830.5030.2690.880.4580.417
%FeO0.1040.0850.1380.250.2150.4110.3390.661
%SiO 2 0.0880.0890.2150.3340.3480.5510.40.658
%Al 2 O 3 0.2270.1070.2330.2160.2970.3940.4730.251
%Cr 2 O 3 0.1830.1380.2620.310.2540.50.3360.379
%MgO0.1210.0730.2020.2290.2110.3120.3940.182
%CaO0.0890.0850.1980.350.3130.5530.4620.637
Type A0.1090.0720.4560.4010.4510.4890.3240.477
Type B0.0970.0850.1740.20.1510.4640.3540.194
Type C0.1160.1240.3650.3330.4740.5040.4090.555
Type D0.1150.0980.3660.3740.5370.5480.4020.617
Type E0.1820.090.260.3090.3180.6260.8470.281
Type F0.0560.0410.0540.0930.1080.1780.1950.078
Type G0.120.1050.2910.2320.2880.4770.5790.304
Type N0.1140.090.1750.1510.1380.1590.1480.112
PreHeater0.1870.0890.1290.1340.1960.130.1080.149
EE0.2170.1140.1160.1360.1790.1810.2570.14
%Cr%Si%C%Ni%Feo%SiO2%Al2O3
Delays0.1060.0940.0740.1010.1040.090.11
TTT0.1050.0890.080.1090.1010.1010.105
Process0.0920.1170.1050.1290.0770.0970.107
Charging0.1220.0890.0890.080.1370.0980.158
Ext. ref.0.1070.0890.1920.1510.0750.0990.106
Melting0.1180.1150.1130.1190.10.1190.09
Refining0.1760.180.1140.2590.1910.3030.215
Tapping0.10.1420.0750.0830.0750.0960.078
Propane0.0970.2170.1030.110.1040.0880.227
O 2 0.0750.0920.1680.1060.0850.0890.107
TotWeight0.2160.3730.1240.3830.1380.2150.233
Metal0.3450.3150.1850.5030.250.3340.216
Slag0.3360.3810.2320.2690.2150.3480.297
%Fe0.4380.6090.2810.880.4110.5510.394
%O0.3040.9290.1970.4580.3390.40.473
%Al0.5450.2330.2040.4170.6610.6580.251
%Cr1.00.2990.2410.4720.4690.4970.261
%Si0.2991.00.2170.4250.3260.430.491
%C0.2410.2171.00.2770.1010.160.137
%Ni0.4720.4250.2771.00.4070.4990.366
%FeO0.4690.3260.1010.4071.00.7590.218
%SiO 2 0.4970.430.160.4990.7591.00.393
%Al 2 O 3 0.2610.4910.1370.3660.2180.3931.0
%Cr 2 O 3 0.3170.3540.1650.4670.2610.6210.515
%MgO0.1950.3470.1050.310.1980.350.472
%CaO0.4880.5010.1640.4790.7770.9620.46
Type A0.3270.2740.2060.5070.4710.5440.319
Type B0.230.3780.1190.3690.3010.3560.261
Type C0.4210.3310.2030.4980.4650.5560.353
Type D0.5370.3270.1720.5440.5740.6420.332
Type E0.3690.8720.1750.460.3850.4610.481
Type F0.0770.2110.0720.1270.0870.140.092
Type G0.2730.6190.260.380.3370.5140.221
Type N0.0980.1350.1050.1660.1380.1290.133
PreHeater0.1210.110.2060.1310.0850.1420.154
EE0.1390.2160.0970.1870.1230.1110.21
%Cr 2 O 3 %Mgo%CaOType AType BType CType DType E
Delays0.1430.10.0860.0780.0640.060.0830.09
TTT0.1580.1150.0910.0880.060.0950.1130.097
Process0.1270.1240.0930.0830.0840.1110.1230.101
Charging0.1420.1030.1020.0840.0680.0750.0820.082
Ext. ref.0.130.0730.0940.0730.1010.1080.1230.072
Melting0.1110.080.1140.1080.0630.1130.1130.109
Refining0.2460.1670.2690.3360.0840.3730.3720.149
Tapping0.1130.0920.0950.0910.0950.1070.0970.135
Propane0.1830.1210.0890.1090.0970.1160.1150.182
O 2 0.1380.0730.0850.0720.0850.1240.0980.09
TotWeight0.2620.2020.1980.4560.1740.3650.3660.26
Metal0.310.2290.350.4010.20.3330.3740.309
Slag0.2540.2110.3130.4510.1510.4740.5370.318
%Fe0.50.3120.5530.4890.4640.5040.5480.626
%O0.3360.3940.4620.3240.3540.4090.4020.847
%Al0.3790.1820.6370.4770.1940.5550.6170.281
%Cr0.3170.1950.4880.3270.230.4210.5370.369
%Si0.3540.3470.5010.2740.3780.3310.3270.872
%C0.1650.1050.1640.2060.1190.2030.1720.175
%Ni0.4670.310.4790.5070.3690.4980.5440.46
%FeO0.2610.1980.7770.4710.3010.4650.5740.385
%SiO 2 0.6210.350.9620.5440.3560.5560.6420.461
%Al 2 O 3 0.5150.4720.460.3190.2610.3530.3320.481
%Cr 2 O 3 1.00.3710.5760.4780.2230.3990.4020.39
%MgO0.3711.00.4310.3510.3250.2160.2560.341
%CaO0.5760.4311.00.5260.4150.4730.5840.539
Type A0.4780.3510.5261.00.1810.5540.640.321
Type B0.2230.3250.4150.1811.00.2410.2230.415
Type C0.3990.2160.4730.5540.2411.00.9290.359
Type D0.4020.2560.5840.640.2230.9291.00.382
Type E0.390.3410.5390.3210.4150.3590.3821.0
Type F0.0880.0810.1340.1190.0840.110.1270.177
Type G0.4650.2670.4820.3060.230.3210.3510.487
Type N0.1240.2040.1180.1490.0880.1230.1060.185
PreHeater0.2050.0760.1390.1470.1010.1460.1650.09
EE0.0990.2410.1310.0940.180.2910.2710.3
Type FType GType NPreHeaterEE
Delays0.0510.0790.1580.0760.27
TTT0.0440.0960.1980.0870.396
Process0.0480.1930.1970.080.534
Charging0.0590.0740.090.0910.095
Ext. ref.0.0340.140.1740.0910.463
Melting0.0620.1180.130.0810.087
Refining0.1160.2620.1990.2060.499
Tapping0.120.1470.1490.1180.129
Propane0.0560.120.1140.1870.217
O 2 0.0410.1050.090.0890.114
TotWeight0.0540.2910.1750.1290.116
Metal0.0930.2320.1510.1340.136
Slag0.1080.2880.1380.1960.179
%Fe0.1780.4770.1590.130.181
%O0.1950.5790.1480.1080.257
%Al0.0780.3040.1120.1490.14
%Cr0.0770.2730.0980.1210.139
%Si0.2110.6190.1350.110.216
%C0.0720.260.1050.2060.097
%Ni0.1270.380.1660.1310.187
%FeO0.0870.3370.1380.0850.123
%SiO 2 0.140.5140.1290.1420.111
%Al 2 O 3 0.0920.2210.1330.1540.21
%Cr 2 O 3 0.0880.4650.1240.2050.099
%MgO0.0810.2670.2040.0760.241
%CaO0.1340.4820.1180.1390.131
Type A0.1190.3060.1490.1470.094
Type B0.0840.230.0880.1010.18
Type C0.110.3210.1230.1460.291
Type D0.1270.3510.1060.1650.271
Type E0.1770.4870.1850.090.3
Type F1.00.2210.0460.0580.046
Type G0.2211.00.1220.10.155
Type N0.0460.1221.00.1880.303
PreHeater0.0580.10.1881.00.074
EE0.0460.1550.3030.0741.0

References

  1. MacRosty, R.; Swartz, C. Dynamics Optimization of Electric Arc Furnace Operation. Inst. Chem. Eng. 2007, 53, 640–653. [Google Scholar] [CrossRef]
  2. Ledesma-Carrión, D. Energy Optimization of Steel in Electric Arc Furnace. Glob. J. Technol. Optim. 2016, 7, 1–10. [Google Scholar]
  3. Gerardi, D.; Marlin, T.; Swartz, C. Optimization of Primary Steelmaking Purchasing and Operation under Raw Material Uncertainty. Ind. Eng. Chem. Res. 2013, 52, 12383–12398. [Google Scholar] [CrossRef]
  4. Morales, R.; Rodríguez-Hernández, H.; Conejo, A. A Mathematical Simulator for the EAF Steelmaking Process Using Direct Reduced Iron. ISIJ Int. 2005, 41, 426–435. [Google Scholar] [CrossRef]
  5. Nyssen, P.; Colin, R.; Junqué, J.-L.; Knoops, S. Application of a dynamic metallurgical model to the electric arc furnace. La Revue de Métallurgie 2004, 10, 317–326. [Google Scholar] [CrossRef]
  6. Çamdali, Ü. Determination of the Optimum Production Parameters by Using Linear Programming in the AC Electric Arc Furnace. Can. J. Met. Mater. Sci. 2013, 44, 103–110. [Google Scholar] [CrossRef]
  7. MacRosty, R.; Swartz, C. Dynamic Modeling of an Industrial Electric Arc Furnace. Ind. Eng. Chem. Res. 2005, 44, 8067–8083. [Google Scholar] [CrossRef]
  8. Mapelli, C.; Baragiola, S. Evaluation of energy and exergy performances in EAF during melting and refining period. Ironmak. Steelmak. 2006, 33, 379–388. [Google Scholar] [CrossRef]
  9. Kirschen, M.; Badr, K.; Pfeifer, H. Influence of Direct Reduced Iron on the Energy Balance of the Electric Arc Furnace in Steel Industry. Energy 2011, 36, 6146–6155. [Google Scholar] [CrossRef]
  10. Logar, V.; Dovžan, D.; Škrjanc, I. Modeling and Validation of an Electric Arc Furnace: Part 1, Heat and Mass Transfer. ISIJ Int. 2012, 52, 402–412. [Google Scholar] [CrossRef] [Green Version]
  11. Logar, V.; Dovžan, D.; Škrjanc, I. Modeling and Validation of an Electric Arc Furnace: Part 2, Thermo-chemistry. ISIJ Int. 2012, 52, 413–423. [Google Scholar] [CrossRef] [Green Version]
  12. Çamdalı, Ü.; Tunç, M. Modelling of Electric Energy Consumption in the AC Electric Arc Furnace. Int. J. Energy Res. 2002, 26, 935–947. [Google Scholar] [CrossRef]
  13. Opitz, F.; Treffinger, P. Physics-Based Modeling of Electric Operation, Heat Transfer, and Scrap Melting in an AC Electric Arc Furnace. Met. Mater. Trans. B 2016, 47, 1489–1503. [Google Scholar] [CrossRef]
  14. Morales, R.; Conejo, A.; Rodríguez, H. Process Dynamics of Electric Arc Furnace during Direct Reduced Iron Melting. Met. Mater. Trans. B 2002, 33, 187–199. [Google Scholar] [CrossRef]
  15. Prakash, S.; Mukherjee, K.; Singh, S.; Mehrotra, S.P. Simulation of energy dynamics of electric furnace steelmaking using DRI. Ironmak. Steelmak. 2007, 34, 61–70. [Google Scholar] [CrossRef]
  16. Kho, T.S.; Swinbourne, D.R.; Blanpain, B.; Arnout, S.; Langberg, D. Understanding stainless steelmaking through computational thermodynamics Part 1: Electric Arc Furnace Melting. Miner. Process. Extr. Met. 2010, 119, 1–8. [Google Scholar] [CrossRef]
  17. Kirschen, M.; Risonarta, V.; Pfeifer, H. Energy efficiency and the influence of gas burners to the energy related carbon dioxide emissions of electric arc furnaces in steel industry. Energy 2009, 34, 1065–1072. [Google Scholar] [CrossRef]
  18. Trejo, E.; Martell, F.; Micheloud, O.; Teng, L.; Llamasa, A.; Montesinos-Castellanosa, A. A novel estimation of electrical and cooling losses in electric arc furnaces. Energy 2012, 42, 446–456. [Google Scholar] [CrossRef]
  19. Fathi, A.; Saboohi, Y.; Škrjanc, I.; Logar, V. Comprehensive Electric Arc Furnace Model for Simulation Purposes and Model-Based Control. Steel Res. Int. 2017, 88, 1600083. [Google Scholar] [CrossRef]
  20. Odenthal, H.-J.; Kemminger, A.; Krause, F.; Sankowski, L.; Uebber, N.; Vogl, N. Review of Modeling and Simulation of the Electric Arc Furnace (EAF). Steel Res. Int. 2018, 89, 1–36. [Google Scholar] [CrossRef]
  21. Baumert, J.-C.; Engel, R.; Weiler, C. Dynamic modelling of the electric arc furnace process using artificial neural networks. Revue de Métallurgie 2002, 99, 839–849. [Google Scholar] [CrossRef]
  22. Baumert, J.-C.; Vigil, J.R.; Nyssen, P.; Schaefers, J.; Schutz, G.; Gillé, S. Improved Control of Electric arc Furnace Operations by Process Modelling; European Commission: Luxembourg, 2005; ISBN 92-894-9789-0. [Google Scholar]
  23. Mathy, C.; Terho, K.; Chouvet, M.; Coq, X.L.; Baumert, J.; Engel, R.; Hoffmann, J. Production of Steel at Lower Operating Costs in EAF; European Commission: Luxembourg, 2003; ISBN 92-894-6377-5. [Google Scholar]
  24. Chen, C.; Liu, Y.; Kumar, M.; Qin, J. Energy Consumption Modelling Using Deep Learning Technique—A Case Study of EAF. In Proceedings of the 51st CIRP Conference on Manufacturing Systems, Stockholm, Sweden, 16–18 May 2018. [Google Scholar]
  25. Sandberg, E. Energy and Scrap Optimisation of Electric Arc Furnaces by Statistical Analysis of Process Data. Ph.D. Thesis, Luleå University of Technology, Luleå, Sweden, 2005. [Google Scholar]
  26. Köhle, S.; Lichterbeck, R.; Paura, G. Verbesserung der energetischen Betriebsführung von Drehstrom-Lichtbogenöfen; European Commission: Brussels, Belgium, 1996; ISBN 92-827-6467-2. [Google Scholar]
  27. Köhle, S. Effects on the Electric Energy Consumption of Arc Furnace Steelmaking. In Proceedings of the 4th European Electric Steel Congress, Madrid, Spain, 3–6 November 1992. [Google Scholar]
  28. Köhle, S. Variables influencing electric energy and electrode consumption in electric arc furnaces. Met. Plant Technol. Int. 1992, 6, 48–53. [Google Scholar]
  29. Bowman, B. Performance comparison update—AC vs DC furnaces. Iron Steel Eng. 1995, 72, 26–29. [Google Scholar]
  30. Kleimt, B.; Köhle, S. Power consumption of electric arc furnaces with post-combustion. Met. Plant Technol. Int. 1997, 3, 56–57. [Google Scholar]
  31. Köhle, S. Improvements in EAF operating practices over the last decade. In Proceedings of the Electric Furnace Conference, Pittsburgh, PA, USA, 14–16 November 1999. [Google Scholar]
  32. Köhle, S. Recent improvements in modelling energy consumption of electric arc furnaces. In Proceedings of the 7th European Electric Steelmaking Conference, Venice, Italy, 26–29 May 2002. [Google Scholar]
  33. Köhle, S.; Hoffmann, J.; Baumert, J.; Picco, M.; Nyssen, P.; Filippini, E. Improving the Productivity of Electric Arc Furnaces; European Commission: Luxembourg, 2003; ISBN 92-894-6136-5. [Google Scholar]
  34. Kleimt, B.; Köhle, S.; Kühn, R.; Zisser, S. Application of models for electrical energy consumption to improve EAF operation and dynamic control. In Proceedings of the 8th European Electric Steelmaking Congress, Birmingham, UK, 9–11 May 2005; pp. 183–197. [Google Scholar]
  35. Kirschen, M.; Zettl, K.-M.; Echterhof, T.; Pfeifer, H. Models for EAF energy efficiency. Steel Times Int. 2017, 44, 1–6. [Google Scholar]
  36. Conejo, A.; Cárdenas, J. Energy Consumption in the EAF with 100% DRI. In Proceedings of the Iron & Steel Technology Conference, Cleveland, OH, USA, 1–4 May 2006; Volume 1. [Google Scholar]
  37. Czapla, M.; Karbowniczek, M.; Michaliszyn, A. The Optimisation of Electric Energy Consumption in the Electric Arc Furnace. Arch. Met. Mater. 2008, 53, 559–565. [Google Scholar]
  38. Sandberg, E.; Lennox, B.; Undvall, P. Multivariate Prediction of End Conditions for Electric Arc Furnaces. In Proceedings of the 2nd International Conference on Process Development in Iron and Steelmaking, Luleå, Sweden, 6–9 June 2004. [Google Scholar]
  39. Sandberg, E.; Lennox, B.; Marjanovic, O.; Smith, K. Multivariate process monitoring of EAFs. Ironmak. Steelmak. 2005, 32, 221–226. [Google Scholar] [CrossRef]
  40. Sandberg, E.; Lennox, B.; Undvall, P. Scrap management by statistical evaluation of EAF process data. Control Eng. Pract. 2007, 15, 1063–1075. [Google Scholar] [CrossRef]
  41. Gajic, D.; Savic-Gajic, I.; Savic, I.; Georgieva, O.; Gennaro, S.D. Modelling of electrical energy consumption in an electric arc furnace using artificial neural networks. Energy 2016, 108, 132–139. [Google Scholar] [CrossRef]
  42. Haupt, M.; Vadenbo, C.; Zeltner, C.; Hellweg, S. Influence of Input-Scrap Quality on the Environmental Impact of Secondary Steel Production. J. Ind. Ecol. 2016, 21, 391–401. [Google Scholar] [CrossRef]
  43. Carlsson, L.S.; Samuelsson, P.B.; Jönsson, P.G. Predicting the Electrical Energy Consumption of Electric Arc Furnaces Using Statistical Modeling. Metals 2019, 9, 959. [Google Scholar] [CrossRef] [Green Version]
  44. Pfeifer, H.; Kirschen, M. Thermodynamic analysis of EAF electrical energy demand. In Proceedings of the 7th European Electric Steelmaking Conference, Venice, Italy, 26–29 May 2002. [Google Scholar]
  45. Steinparzer, T.; Haider, M.; Zauner, F.; Enickl, G.; Naussed, M.M.; Horn, A.C. Electric Arc Furnace Off-Gas Heat Recovery and Experience with a Testing Plant. Steel Res. Int. 2014, 85, 519–526. [Google Scholar] [CrossRef]
  46. Keplinger, T.; Haider, M.; Steinparzer, T.; Trunner, P.; Patrejko, A.; Haselgrübler, M. Modeling, Simulation, and Validation with Measurements of a Heat Recovery Hot Gas Cooling Line for Electric Arc Furnaces. Steel Res. Int. 2018, 89, 1800009. [Google Scholar] [CrossRef] [Green Version]
  47. Carling, K. Resistant outlier rules and the non-Gaussian case. Comp. Stat. Data Anal. 2000, 33, 249–258. [Google Scholar] [CrossRef] [Green Version]
  48. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; The MIT Press: Cambridge, MA, USA, 2016; ISBN 9780262035613. [Google Scholar]
  49. Ajossou, A.; Palm, R. Impact of Data Structure on the Estimators R-square and Adjusted R-square in Linear Regression. Int. J. Math. Comput. 2013, 20, 84–93. [Google Scholar]
  50. Claesen, M.; De Moor, B. Hyperparameter Search in Machine Learning. arXiv 2015, arXiv:1502.02127. [Google Scholar]
  51. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  52. De Siqueira Santos, S.; Yasumasa Takahashi, D.; Nakata, A.; Fujita, A. A comparative study of statistical methods used to identify dependencies between gene expression signals. Brief. Bioinform. 2014, 1, 906–918. [Google Scholar] [CrossRef] [Green Version]
  53. Székely, G.J.; Rizzo, M.L. Brownian distance covariance. Ann. Appl. Stat. 2009, 3, 1236–1265. [Google Scholar] [CrossRef] [Green Version]
  54. Dodge, Y. Kolmogorov–Smirnov Test. In The Concise Encyclopedia of Statistics; Springer: New York, NY, USA, 2009; pp. 283–287. ISBN 978-0-387-32833-1. [Google Scholar]
  55. Pratt, J.W.; Gibbons, J.D. Kolmogorov-Smirnov Two-Sample Tests. In Concepts of Nonparametric Theory; Springer: New York, NY, USA, 1981; pp. 318–344. ISBN 978-1-4612-5931-2. [Google Scholar]
  56. Fisher, A.; Rudin, C.; Dominici, F. All Models are Wrong but Many are Useful: Variable Importance for Black-Box, Properietary, or Misspecified Prediction Models, using Model Class Reliance. arXiv 2018, arXiv:1801.01489v3. [Google Scholar]
  57. Molnar, C. Interpretable Machine Learning—A Guide for Making Black Box Models Explainable. 2019. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 19 October 2019).
Figure 1. An idealized timeline over the EAF process for one heat. The Extended Refining is unique to the steel plant of study.
Figure 1. An idealized timeline over the EAF process for one heat. The Extended Refining is unique to the steel plant of study.
Metals 10 00036 g001
Figure 2. Graph-model describing the expected correlations between the input variables and the output variable in a statistical model predicting the EE consumption. The arrows illustrate the correlative relation between the variables. The circled signs illustrate the positive or negative contributions imposed on EE.
Figure 2. Graph-model describing the expected correlations between the input variables and the output variable in a statistical model predicting the EE consumption. The arrows illustrate the correlative relation between the variables. The circled signs illustrate the positive or negative contributions imposed on EE.
Metals 10 00036 g002
Figure 3. Flowchart explaining the data and statistical modeling pipeline. * The flowchart is split to emphasize the two different approaches used to select the variable batches for the grid-search.
Figure 3. Flowchart explaining the data and statistical modeling pipeline. * The flowchart is split to emphasize the two different approaches used to select the variable batches for the grid-search.
Metals 10 00036 g003
Figure 4. Examples of distribution of variables from real EAF production. All variables are normalized. Top left: Propane. Top right: Refining time. Bottom left: Total weight. Bottom right: Material Type B.
Figure 4. Examples of distribution of variables from real EAF production. All variables are normalized. Top left: Propane. Top right: Refining time. Bottom left: Total weight. Bottom right: Material Type B.
Metals 10 00036 g004
Figure 5. An Artificial Neural Network (ANN) with two input nodes, one hidden layer with three nodes, and one output node. Each line drawn between the nodes illustrates the forward flow of calculations in the network.
Figure 5. An Artificial Neural Network (ANN) with two input nodes, one hidden layer with three nodes, and one output node. Each line drawn between the nodes illustrates the forward flow of calculations in the network.
Metals 10 00036 g005
Figure 6. An illustration of the two-sample Kolmogorov–Smirnov (KS)-test for the random variables X1 and X2, where X 1 N o r m ( 200,25 ) and X 2 N o r m ( 200,35 ) Left: The cumulative distribution functions (CDF) of X1 and X2 with the D-value as the difference between the upper and lower dashed lines. 100 samples were drawn from each distribution. Right: The probability density functions of X1 and X2.
Figure 6. An illustration of the two-sample Kolmogorov–Smirnov (KS)-test for the random variables X1 and X2, where X 1 N o r m ( 200,25 ) and X 2 N o r m ( 200,35 ) Left: The cumulative distribution functions (CDF) of X1 and X2 with the D-value as the difference between the upper and lower dashed lines. 100 samples were drawn from each distribution. Right: The probability density functions of X1 and X2.
Metals 10 00036 g006
Figure 7. The predicted and true EE consumption. The dashed lines represent a perfect model where all predicted EE consumption values are equal the true EE consumption values. All values are normalized. Left: Model D15. Right: Model A21.
Figure 7. The predicted and true EE consumption. The dashed lines represent a perfect model where all predicted EE consumption values are equal the true EE consumption values. All values are normalized. Left: Model D15. Right: Model A21.
Metals 10 00036 g007
Figure 8. Distributions for the training data (black) and test data (gray) for EE and the 5 of the variables with higher FI for the 6 selected models. The vertical dashed lines show the mean of each distribution. Both the training and test data are normalized with the mean error and standard deviation of error of the training data.
Figure 8. Distributions for the training data (black) and test data (gray) for EE and the 5 of the variables with higher FI for the 6 selected models. The vertical dashed lines show the mean of each distribution. Both the training and test data are normalized with the mean error and standard deviation of error of the training data.
Metals 10 00036 g008
Table 1. Ingoing and outgoing energy terms governing the energy balance equation of the Electric Arc Furnace (EAF) process.
Table 1. Ingoing and outgoing energy terms governing the energy balance equation of the Electric Arc Furnace (EAF) process.
Energy FactorDescriptionEquationProportionality
In E E l Total Electrical Energy (EE) output from transformer
E E l = η A r c · η E l · P A v g · t P O N
E E l t P O N
E C h e m Total energy from chemical reactions in steel and slag
E C h e m = Δ H 298 o ( p r o d u c t s ) Δ H 298 o ( r e a c t a n t s )
+ T s T T a p c p ( p r o d u c t s ) c p ( r e a c t a n t s ) d T
E C h e m T s ; T T a p ; c o m p o s i t i o n
E B u Total energy input from burner
E B u = η F u e l h F u e l V F u e l
E B u V F u e l ; η F u e l
Out E S t e e l Total energy output to steel
E S t e e l = H M e l t , s t e e l + m S t e e l · c S t e e l T s T T a p d T
E S t e e l T s ; T T a p ; c o m p o s i t i o n
E S l a g Total energy lost in slag
E S l a g = H M e l t , s l a g + m S l a g · c S l a g T s T T a p d T
E S l a g T s ; T T a p ; c o m p o s i t i o n
Out E G a s Total energy lost in gas
E G a s = P M V ˙ G a s R · t T T T · c G a s T s T O f f g a s d T T
E G a s l n ( T s ) ; l n ( T T a p ) ; t T T T ; V ˙ G a s
E D u s t Total energy lost in dust
E D u s t = m ˙ D u s t · t T T T · c D u s t T s T O f f g a s d T
E D u s t T s ; T T a p ; t T T T ; m ˙ D u s t
E C o o l i n g Total energy lost in cooling water
E C o o l i n g = k A C S ( T C S T H 2 O ) · t T T T
E C o o l i n g T C S ; t T T T
E R a d Total energy lost through radiation
E R a d = ϵ σ A E A F T E A F 4 · t T T T
E R a d T E A F 4 ; t T T T
E C o n v Total energy lost through convection
E C o n v = h A H ( T H T A m b ) · t T T T
E C o n v T H ; T A m b ; t T T T
E E l , l o s s Energy lost in electrical system and arc transfer
E E l , l o s s = ( 1 η A r c ) ( 1 η E l ) · E E l
E E l , l o s s t P O N
Table 2. Ingoing and outgoing energy in the EAF process as percentages of energy sources and energy sinks, respectively. Values are computed from a synthesis of reported values [9,25,44,45,46].
Table 2. Ingoing and outgoing energy in the EAF process as percentages of energy sources and energy sinks, respectively. Values are computed from a synthesis of reported values [9,25,44,45,46].
Energy FactorPercentage of in and out Energy Balance
InElectric40–66%
Oxidation20–50%
Burner/fuel2–11%
OutLiquid steel45–60%
Slag and dust4–10%
Off-gas11–35%
Cooling8–29%
Radiation and electrical losses2–6%
Table 3. The variables used in the models.
Table 3. The variables used in the models.
VariablesUnitDefinition
Delays m i n The sum of all delays, which is defined as all deviations from the nominal time of each sub-process.
Tap-to-Tap time (TTT) m i n The between the end of the tapping from the previous heat to the end of tapping of the current heat
Charging m i n Total time needed to charge all baskets
Melting m i n Total melting time for all scrap baskets
Refining m i n Total refining time
Extended refining m i n Total Extended refining time.
Tapping m i n Total tapping time
Total Weight k g Total weight of all materials added during the EAF process
Propane m 3 Total amount of propane gas added by burners
O 2 -lance m 3 Total amount of oxygen added by lance
Preheater energy k W h Estimated thermal energy added to the scrap baskets by the preheater
EE consumption k W h Total EE consumption for the heat. This is the output variable in the models.
Process Time m i n Defined as the sum of Charging, Melting, Refining, Extended refining, and Tapping.
C w t % The weight percent with respect to the total charged metallic material during the heat. Hence, added dolomite, lime, and carbon by lance are included
Si w t %
Cr w t %
Fe w t %
Ni w t %
O w t %
Al w t %
C r 2 O 3 w t % The weight percent with respect to the total charged oxide bearing raw material during the heat.
M g O w t %
C a O w t %
F e O w t %
S i O 2 w t %
A l 2 O 3 w t %
Metal Weight k g Total weight of metallic material
Slag Weight k g Total weight of oxide bearing raw material
Type A k g All raw material types as defined by the plant engineers.
Type B k g
Type C k g
Type D k g
Type E k g
Type F k g
Type G k g
Type N k g
Table 4. The variable batches governing the domain-specific variable selection approach. The variables contained within each variable group are shown in Table 5.
Table 4. The variable batches governing the domain-specific variable selection approach. The variables contained within each variable group are shown in Table 5.
Variable Batch12345678910111213141516
Basexxxxxxxxxxxxxxxx
Process Timexxxxxxxxxxxxxxxx
Sub-processesxxxxxxxxxxxxxxxx
Metallic elementsxxxxxxxx
Oxide compoundsxxxx xxxx
Met-Slag weightxx xx xx xx
Material Typesx x x x x x x x
Variable Batch17181920212223242526272829303132
Basexxxxxxxxxxxxxxxx
Process Time
Sub-processesxxxxxxxxxxxxxxxx
Metallic elementsxxxxxxxx
Oxide compoundsxxxx xxxx
Met-Slag weightxx xx xx xx
Material Typesx x x x x x x x
Variable Batch33343536373839404142434445464748
Basexxxxxxxxxxxxxxxx
Process Timexxxxxxxxxxxxxxxx
Sub-processes
Metallic elementsxxxxxxxx
Oxide compoundsxxxx xxxx
Met-Slag weightxx xx xx xx
Material Typesx x x x x x x x
Variable Batch49505152535455565758596061626364
Basexxxxxxxxxxxxxxxx
Process Time
Sub-processes
Metallic elementsxxxxxxxx
Oxide compoundsxxxx xxxx
Met-Slag weightxx xx xx xx
Material Typesx x x x x x x x
Table 5. Input variables for each variable group. The metallic elements and oxide compounds are the composition of the total charged metallic material and oxide bearing raw material, respectively.
Table 5. Input variables for each variable group. The metallic elements and oxide compounds are the composition of the total charged metallic material and oxide bearing raw material, respectively.
Variable GroupVariablesNo. VariablesVariable GroupVariablesNo. Variables
BaseDelays6Sub-processesCharging5
TTT Melting
Total Weight Refining
Propane Extended refining
O 2 -lance Tapping
Preheater energy Metallic elementsC7
Met-SlagMetal Weight2 Si
weightSlag Weight Cr
MaterialType A8 Fe
typesType B Ni
Type C O
Type D Al
Type E Oxide C r 2 O 3 6
Type F compounds M g O
Type G C a O
Type N F e O
Process TimeProcess Time1 S i O 2
A l 2 O 3
Table 6. The domain-specific cleaning rules that were applied to the data. The rules were applied to the training data and test data separately to simulate an applied data filtering system. Rules applied to the training data must be applied to the test data to safeguard against data points outside the scope of the statistical model.
Table 6. The domain-specific cleaning rules that were applied to the data. The rules were applied to the training data and test data separately to simulate an applied data filtering system. Rules applied to the training data must be applied to the test data to safeguard against data points outside the scope of the statistical model.
FilterMotivation
Removal of all Trial heatsTrial heats are not part of regular production since the aim is primarily to investigate the properties for new scrap types
Heats with EE above 60 MWhIdentified as abnormal EE by the process engineers
Heats with Total charged weight above 110,000 kgPhysically impossible weight due to furnace size limitations
45 min < TTT < 180 min45 min are considered unusually short and above 180 min are likely due to a longer delay in the process or a scheduled stop. Usually, the TTT is aimed at 60–70 min
Delays < 180 minHeats with delays over 3 h are because of longer stops due to, for example, broken equipment.
Table 7. The parameters used in the grid-search of which 2 · 3 · 48 = 288 are model-specific and 64 · 2 = 128 are domain-specific. A total of 288 · 128 = 36,864 parameter combinations.
Table 7. The parameters used in the grid-search of which 2 · 3 · 48 = 288 are model-specific and 64 · 2 = 128 are domain-specific. A total of 288 · 128 = 36,864 parameter combinations.
Model-Specific
ParameterDescriptionValues#
Activation functionActivation functions have different gradient intensity in the updating step of each iteration and different upper and lower bounds. This influences the training phase. [ t a n h , l o g i s t i c ] 2
Learning rateA larger learning rate means that the gradient-descent algorithm takes a larger step in error space and may therefore miss more optimal local minimum points. Smaller steps, i.e., smaller learning rates, are preferable to find more optimal local minimum points. [ 0.001 , 0.01 , 0.1 ] 3
Hidden node topologyIncreasing number of layers and nodes means increased model complexity. Here, the goal is to investigate whether or not more hidden layers leads to a better model. One should always strive for model simplicity when two models are statistically significant equally good. ( z ) or ( z , z ) where z 1 , 2 , , 24 48
Domain-Specific
ParameterDescriptionValues#
Input variablesIt is not possible to know, a priori, which combination of input variables is the most relevant for a statistical model to predict the output variable with high precision and accuracy. See Section 2.4 for reasoning.See Table 4.64
Validation setThe effect of a randomized validation set and ordered validation set with respect to the rest of the training data will be investigated. o r d e r e d , r a n d o m i z e d 2
Table 8. The metrics used to evaluate the performance of each model type.
Table 8. The metrics used to evaluate the performance of each model type.
SymbolDefinition
R ¯ μ 2 Mean adjusted R-square of the 10 model instances on the test data
R ¯ σ 2 Standard deviation of adjusted R-square of the 10 model instances on the test data
R ¯ m i n 2 Minimum adjusted R-square of the 10 model instances on the test data
R ¯ m a x 2 Maximum adjusted R-square of the 10 model instances on the test data
Δ μ Mean error of the mean error of the 10 model instances on the test data
Δ σ Standard deviation of the mean error of the 10 model instances on the test data.
Δ m i n Minimum error of the mean error of the 10 model instances on the test data.
Δ m a x Maximum error of the mean error of the 10 model instances on the test data.
Table 9. The remaining data from the domain-specific cleaning rules that were applied to the data. The training and test data were selected as described in Section 3.2.3.
Table 9. The remaining data from the domain-specific cleaning rules that were applied to the data. The training and test data were selected as described in Section 3.2.3.
FilterTraining DataTest DataTotal Data
Before cleaning12,18340412,587
Removal of all Trial heats11,53038611,916
Heats with EE above 60 MWh11,53038611,916
Heats with Total charged weight above 110,000 kg11,53038611,916
45 min < TTT < 180 min10,99036211,352
Delays < 180 min10,96636211,328
After cleaning (No.)10,96636211,328
(% data loss)10.0%10.4%10.0%
Table 10. Performance of the selected models on the training and test data. Models called D are from the Domain approach and models called A are from the algorithmic approach. TR and TE are short for training and test, respectively.
Table 10. Performance of the selected models on the training and test data. Models called D are from the Domain approach and models called A are from the algorithmic approach. TR and TE are short for training and test, respectively.
ModelD1D15D17D31D64A21
No. Variables35203419616
DatasetTRTETRTETRTETRTETRTETRTE
R ¯ μ 2 0.7900.6980.7170.7310.7660.7060.6600.7220.6010.6970.7250.706
Δ μ 1−331−2−554−23−2662−710−8−686−5−566
Δ σ 114511671392112612351176156211051691121613781185
Δ m i n −5679−4249−6606−3819−5275−3662−7526−4089−8149−4805−6519−4333
Δ m a x 531054425564573556576072646260208520608760386610
R ¯ μ , t r a i n 2 R ¯ μ , t e s t 2 0.092−0.0140.060−0.062−0.0960.019
Δ μ , t r a i n Δ μ , t e s t 332552243712676561
Table 11. The best model compared with ANN models created by Baumert et al. [21,22,23]. Only models that were tested on test data and presented the number of test heats are reported. Dashed entries indicate missing information.
Table 11. The best model compared with ANN models created by Baumert et al. [21,22,23]. Only models that were tested on test data and presented the number of test heats are reported. Dashed entries indicate missing information.
ModelBaumert C1Baumert D1Baumert E1Baumert E2D15
No. Hidden layers11010111
Total No. Hidden nodes750505811
No. Variables10084829520
Train/test split375/20668/150707/62921/101110,966/362
% cleaned data-2.6%--10%
R 2 ----0.731
Mean error (kWh/heat)~0−100~0−300−554
Standard deviation of error (kWh/heat)13003300180019001126
Minimum error (kWh/heat)−3500--−8000−3819
Maximum error (kWh/heat)6000--50005735
Table 12. Two-sample KS tests for the training data against the test data for all input variables and the EE consumption. Distance correlation (dCor) for training and test data sets are calculated as a one-to-one correlation between the input variables and EE consumption. The difference in dCor is defined as d C o r ( t r a i n i n g ) d C o r ( t e s t ) .
Table 12. Two-sample KS tests for the training data against the test data for all input variables and the EE consumption. Distance correlation (dCor) for training and test data sets are calculated as a one-to-one correlation between the input variables and EE consumption. The difference in dCor is defined as d C o r ( t r a i n i n g ) d C o r ( t e s t ) .
VariableKS-Valuep-ValuedCor EE TraindCor EE TestdCor EE Diff
Delays0.040.490.1350.270−0.135
TTT0.130.160.2620.396−0.134
Total Weight0.240.00.3250.1160.209
Propane0.220.00.0820.217−0.135
O 2 -lance0.240.00.1230.1140.009
Preheater energy0.230.00.0600.074−0.014
Process Time0.180.00.3990.534−0.135
Charging0.080.010.0250.095−0.07
Melting0.080.020.1070.0870.02
Refining0.240.00.4480.499−0.051
Extended refining0.080.030.2370.463−0.226
Tapping0.180.00.1520.1290.023
C0.070.040.2260.0960.129
Si0.160.00.1050.216−0.111
Cr0.190.00.1310.139−0.008
Fe0.130.00.0950.181−0.086
Ni0.160.00.0970.187−0.09
O0.110.00040.0320.257−0.225
Al0.340.00.0980.140−0.042
C r 2 O 3 0.360.00.0950.099−0.004
M g O 0.850.00.1930.241−0.048
C a O 0.240.00.0570.131−0.074
F e O 0.200.00.0320.123−0.091
S i O 2 0.130.00.0790.111−0.032
A l 2 O 3 0.260.00.1300.210−0.08
Metal Weight0.410.00.3440.1350.208
Slag Weight0.50.00.1080.179−0.071
Type A0.180.00.0640.094−0.03
Type B0.070.060.1110.180−0.069
Type C0.070.080.1500.291−0.141
Type D0.140.00.1870.271−0.084
Type E0.180.00.1280.300−0.172
Type F0.020.980.0410.046−0.005
Type G0.190.00.1510.155−0.004
Type N0.230.00.0980.303−0.205
EE consumption0.250.0---
Table 13. Feature importance (FI) for the selected models. Models called D are from the Domain approach and models called A are from the algorithmic approach. TR and TE are short for training and test, respectively. Significant changes in FI are indicated in underlined bold font.
Table 13. Feature importance (FI) for the selected models. Models called D are from the Domain approach and models called A are from the algorithmic approach. TR and TE are short for training and test, respectively. Significant changes in FI are indicated in underlined bold font.
D1D15D17D31D64A21
TRTETRTETRTETRTETRTETRTE
Delays0.850.751.371.300.650.621.211.131.901.710.710.58
TTT1.151.341.351.580.941.141.711.962.582.910.820.88
TotWeight1.150.700.160.021.160.630.150.040.130.030.430.40
Propane0.070.010.050.000.030.000.020.00.020.0--
O 2 -lance0.030.00.020.00.020.00.00.00.020.0--
Preheater0.030.020.030.010.020.000.020.010.020.01--
Process Time0.080.050.230.49------0.200.30
Charging0.010.00.010.010.010.00.010.0----
Melting0.00.010.00.00.00.00.00.0----
Refining0.090.060.050.040.090.020.050.05--0.030.03
Ext. ref.0.030.060.030.070.030.150.030.12--0.030.06
Tapping0.030.00.050.040.000.000.00.0--0.070.06
C0.010.0--0.010.00----0.010.0
Si0.090.14--0.080.09------
Cr0.020.0--0.010.0----0.010.01
Fe0.130.27--0.02−0.02------
Ni0.070.05--0.010.04------
O0.080.11--0.070.07------
Al0.040.03--0.030.01------
C r 2 O 3 0.03−0.01--0.010.00------
M g O 0.150.0--0.130.01----0.070.0
C a O 0.440.82--0.320.61------
F e O 0.050.14--0.030.10------
S i O 2 0.240.46--0.190.38------
A l 2 O 3 0.060.05--0.070.03----0.01−0.01
Metal0.440.29--0.490.25----0.150.16
Slag0.060.01--0.040.00------
Type A0.030.00.020.000.010.010.010.01----
Type B0.020.00.040.020.010.010.030.02----
Type C0.040.020.070.010.060.00.010.01--0.050.04
Type D0.020.040.050.080.020.010.030.07--0.010.02
Type E0.050.140.020.00.020.040.00.0--0.010.04
Type F0.00.00.010.00.00.00.00.0----
Type G0.100.130.050.040.090.080.040.02--0.080.12
Type N0.020.00.030.010.020.00.00.0----
Table 14. Parameter metadata for the models passing the R ¯ m a x 2 R ¯ m i n 2 0.05 criterion (see Section 3.4.3). The upper row is for the domain-specific approach and the bottom row is for the algorithmic approach.
Table 14. Parameter metadata for the models passing the R ¯ m a x 2 R ¯ m i n 2 0.05 criterion (see Section 3.4.3). The upper row is for the domain-specific approach and the bottom row is for the algorithmic approach.
Learning rate0.10.010.001Activation functionLTOrderedYesNoHidden layers12
2%19%79%68%32%39%61%60%40%
Learning rate0.10.010.001Activation functionLTOrderedYesNoHidden layers12
9%32%59%65%35%48%52%55%45%
Table 15. Parameter metadata for the best model from each variable batch. The upper row is for the domain-specific approach and the bottom row is for the algorithmic approach.
Table 15. Parameter metadata for the best model from each variable batch. The upper row is for the domain-specific approach and the bottom row is for the algorithmic approach.
Learning rate0.10.010.001Activation functionLTOrderedYesNoHidden layers12
0%6%94%69%31%19%81%88%12%
Learning rate0.10.010.001Activation functionLTOrderedYesNoHidden layers12
5%40%54%63%37%23%77%69%31%

Share and Cite

MDPI and ACS Style

Carlsson, L.S.; Samuelsson, P.B.; Jönsson, P.G. Using Statistical Modeling to Predict the Electrical Energy Consumption of an Electric Arc Furnace Producing Stainless Steel. Metals 2020, 10, 36. https://0-doi-org.brum.beds.ac.uk/10.3390/met10010036

AMA Style

Carlsson LS, Samuelsson PB, Jönsson PG. Using Statistical Modeling to Predict the Electrical Energy Consumption of an Electric Arc Furnace Producing Stainless Steel. Metals. 2020; 10(1):36. https://0-doi-org.brum.beds.ac.uk/10.3390/met10010036

Chicago/Turabian Style

Carlsson, Leo S., Peter B. Samuelsson, and Pär G. Jönsson. 2020. "Using Statistical Modeling to Predict the Electrical Energy Consumption of an Electric Arc Furnace Producing Stainless Steel" Metals 10, no. 1: 36. https://0-doi-org.brum.beds.ac.uk/10.3390/met10010036

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop