Home Comparison of data-driven prediction methods for comprehensive coke ratio of blast furnace
Article Open Access

Comparison of data-driven prediction methods for comprehensive coke ratio of blast furnace

  • Xiuyun Zhai EMAIL logo and Mingtong Chen
Published/Copyright: January 30, 2023

Abstract

The emission of blast furnace (BF) exhaust gas has been criticized by society. It is momentous to quickly predict the comprehensive coke ratio (CCR) of BF, because CCR is one of the important indicators for evaluating gas emissions, energy consumption, and production stability, and also affects composite economic benefits. In this article, 13 data-driven prediction techniques, including six conventional and seven ensemble methods, are applied to predict CCR. The result of ten-fold cross-validation indicates that multiple linear regression (MLR) and support vector regression (SVR) based on radial basis function are superior to the other methods. The mean absolute error, the root mean square error, and the coefficient of determination (R 2) of the MLR model are 1.079 kg·t−1, 1.668, and 0.973, respectively. The three indicators of the SVR model are 1.158 kg·t−1, 1.878, and 0.975, respectively. Furthermore, AdaBoost based on linear regression has also strong prediction ability and generalization performance. The three methods have important significances both in theory and in practice for predicting CCR. Moreover, the models constructed here can provide valuable hints into realizing data-driven control of the BF process.

1 Introduction

Despite high energy consumption and a large environmental load during the ironmaking process, a blast furnace (BF) is still a crucial component of the whole system of steel production [1,2,3,4]. Fundamentally, BF is an input and output system that generates molten iron through a series of extraordinarily complex physical and chemical processes with the cooperation of main and auxiliary materials [5,6]. A schematic diagram of a representative BF is shown in Figure 1. The main materials refer to iron resources, such as iron ore, sinter, and pellets. Auxiliary materials concern energy sources or other necessary materials for transforming main materials, e.g., coke, coal, limestone, and oxygen-enriched water and refractory materials. The production stability of BF is directly influenced by the operation characteristics of main and auxiliary materials [4,7]. The principle of BF ironmaking becomes entangled because of the interaction of various factors [8,9].

Figure 1 
               Schematic diagram of BF.
Figure 1

Schematic diagram of BF.

Energy consumption, resource shortages, and environmental pollution caused by the development of the iron and steel industry are becoming more and more serious. A lower comprehensive coke ratio (CCR) can decrease coke, heavy oil, and other fuels consumed in producing course of BF and is an effective means to solve the above problems [10]. With the benefit of mathematics and computer technologies to simulate complex BF ironmaking processes, the methods of predicting and controlling various variables by optimizing the BF parameters have recently gained increasing attention [3,11,12,13,14,15]. They contain two approaches: the traditional math approach and machine learning technology. It is a notoriously difficult task to construct a mathematical model on the basis of the mechanism because BF is a complex industrial reactor, including the interacting effects of multiphases accompanied simultaneously by multiphase coupling and multiphysics field coexistence, and so on [14]. In contrast, the machine learning technique based on data-driven is a fast and efficient means that has been widely and successfully employed in many industrial processes since many enterprises have accumulated a large amount of historical data [13,16,17,18,19]. Zhang et al. employed a variety of techniques to predict the current time and multistep-ahead hot metal temperature (HMT) of BF, such as random forest (RF), boosting regression tree, neural network–based methods, and Gaussian process regression [13]. Zhai et al. constructed a support vector regression (SVR) model based on the radial basis function (RBF) to predict the BF fuel ratio with merely six parameters [17]. Zhang et al. have demonstrated that the ensemble pattern trees method is superior to several conventional methods for predicting the HMT of BF [14].

Although predictive techniques are emerging in an endless stream, research has not made an appearance on comparing data-driven methods in detail for predicting the CCR of BF. The present work provides a study on predicting CCR with the 13 prediction technologies. They include six conventional methods, namely, multiple linear regression (MLR), decision tree regression (CART), Lasso, elastic net (EN), k-nearest neighbor (KNN), and SVR, as well as seven ensemble methods, namely, AdaBoost, RF, AdaBoost based on KNN (KNN-AdaBoost), gradient boosting regression (GBR), AdaBoost based on linear regression (LR-AdaBoost), extremely randomized trees (ERTs), and AdaBoost based on RF (RF-AdaBoost). The results of ten-fold cross-validation and external validation demonstrated that MLR, SVR, and LR-AdaBoost were superior to other methods. The evaluation indicators of the models indicated that they were highly valuable in theory and practice to realize data-driven control of the BF process.

The remaining structure and organization of this article are as follows: The 13 predictive methods are succinctly illustrated in Section 2. In Section 3, the evaluation indicators of the models are explicated. The comparative analysis and application research of the predictive techniques are presented in Section 4. The conclusions of the work are discussed in Section 5.

2 Predictive techniques

In the work, a number of predictive techniques are employed for predicting the CCR of BF, namely, MLR, CART, Lasso, EN, KNN, SVR, AdaBoost, RF, KNN-AdaBoost, GBR, LR-AdaBoost, ERT, and RF-AdaBoost. These methods are briefly depicted as follows.

2.1 MLR

MLR attempts to establish the relationship between a response variable and two or more explanatory variables by fitting a linear equation to the observed data [20]. The value of the response variable y has a relationship with every value of the independent variable x. The population regression line for p explanatory variables x 1 , x 2 , , x p is defined to be μ y = β 0 + β 1 x 1 + β 2 x 2 + + β p x p . This line describes how the mean response μ y changes with the explanatory variables. The observed values for y vary about their means μ y and are assumed to have the same standard deviation σ. Formally, the MLR model for n observations can be represented as follows:

(1) y i = β 0 + β 1 x i 1 + β 2 x i 2 + + β p x i p + ε i for i = 1 , 2 , , n ,

where ε is the notation for the model deviations.

2.2 CART

CART [21,22] is a supervised learning technique that is suitable for both classification and regression problems. It provides a flexible tree-like structure, where internal nodes, branches, and leaf nodes represent the features of a dataset, the decision rules, and the outcomes, respectively. Two kinds of nodes are in a decision tree, i.e., a decision node, including multiple branches for making decisions, and a leaf node for outputting a decision without further branches.

The principle of CART is described as follows. Without losing generality, X and Y are given as the input vector and the response vector, respectively. The training set is defined as D = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x N , y N ) } , where x i = ( x i ( 1 ) , x i ( 2 ) , , x i ( n ) ) is known as feature vector, n represents the number of features, i = 1 , 2 , , N , N is sample size.

The feature space is divided by the heuristic method. In each division, each value of the features in the current set will be investigated one by one, and the optimal one will be selected as the segmentation point according to the least-squares error criterion. The j-th feature of the training set is denoted as x ( j ) , the value of which is s. x ( j ) and s are regarded as segmentation variable and segmentation point, respectively. Suppose two regions: R 1 ( j , s ) = { x | x ( j ) s } and R 2 ( j , s ) = { x | x ( j ) > s } . Solve the following equation to obtain optimal j and s:

(2) min j , s min c 1 x i R 1 ( j , s ) ( y i c 1 ) 2 + min c 2 x i R 2 ( j , s ) ( y i c 2 ) 2 ,

where c 1 and c 2 denote the fixed output values of the two regions. The above formula can also be written as follows:

(3) min j , s x i R 1 ( j , s ) ( y i c ˆ 1 ) 2 + x i R 2 ( j , s ) ( y i c ˆ 2 ) 2 ,

where c ˆ 1 = 1 N 1 x i R 1 ( j , s ) y i and c ˆ 1 = 1 N 2 x i R 2 ( j , s ) y i .

After finding the optimal segmentation point (j, s), the input space is divided into two regions. Repeat the above division process for each region until the termination condition is reached. Based on the above process, a least-squares decision tree is constructed.

2.3 Lasso

Lasso regression [23,24], sometimes called L1 regularization of LR, is a modified form of LR. A regularization parameter as punishment is used to control model complexity. It is a biased estimation for dealing with complex collinearity data and obtains a more accurate model by constructing a penalty function. Suppose a regression function:

(4) h θ ( x ) = θ 0 x 0 + θ 1 x 1 + + θ n x n .

Its loss function can be expressed as follows:

(5) J L ( θ ) = 1 2 m i = 1 m ( h θ ( x ( i ) y ( i ) ) 2 ,

where X R n × m is the input vector, Y R m is the response vector, n and m are the feature number and the sample size, respectively. The emergence of Lasso is to tackle two problems: over-fitting of LR and the occurrence of X transpose multiplied by irreversible X in course of solving θ by regular methods. To this end, a regularizer is introduced into the loss function, which is as follows:

(6) J L ( θ ) = 1 2 m i = 1 m ( h θ ( x ( i ) y ( i ) ) 2 + λ j = 1 n θ j ,

where λ is a regularization parameter. If λ is set too big, it will cause an under-fitting result; otherwise, it will cause over-fitting.

In addition, the loss function can also be matrixed into the following form:

(7) J L ( θ ) = arg min θ R n | | y X θ | | 2 2 + λ | | θ | | 2 2 .

2.4 EN

EN, the mixture of Ridge and Lasso regressions, is a linear regression method with L1 and L2 norms as the priori regularizers [25]. A model with only a few non-zero and sparse parameters is learned by this combination, which is just like Lasso. But it still maintains some regular properties like Ridge regression. The matrix of the loss function of EN is described as follows:

(8) J E ( θ ) = arg min θ R n | | y X θ | | 2 + λ 2 | | θ | | 2 2 + λ 1 | | θ | | 1 .

2.5 KNN

KNN is one of the most basic and simplest machine learning algorithms [26]. Its idea is very simple: Each of the n-dimensional input vectors corresponds to a point in the feature space, and the output is a category label or a prediction value. When it is used for regression prediction, the average of those target values ( y i ) of k nearest samples is taken as the prediction value y ˆ of the new sample [27]. The formula is as follows:

(9) y ˆ = 1 K i = 1 K y i .

2.6 SVR

Based on kernel functions and ε insensitive function, SVR first proposed by Cortes and Vapnik [28] has been applied in many fields successfully [29,30], such as character recognition [31], drug design [32], and combinatorial chemistry [33]. The input and output of the i-th sample ( i = 1 , 2 , , l ) are denoted as x i R n and y i , respectively. The solution of the nonlinear SVR could be obtained via the following equation:

(10) min α , α W ( α , α ) = min α , α 1 2 i = 1 l j = 1 l ( α i α i ) ( α j α j ) K ( x i , x j ) + i = 1 l [ ( ε y i ) α i + ( ε + y i ) α i ] ,

s.t. 0 α i C , i = 1 , 2 , , l

0 α i C , i = 1 , 2 , , l

i = 1 l ( α i α i ) = 0 ,

where α i and α i represent Lagrange coefficients; C is the penalty factor; and K ( ) denotes the kernel function. The regression function f ( x ) is

(11) f ( x ) = i = 1 N sv sv ( α i α i ) K ( x i , x ) ,

where N SV denotes the number of SVs.

2.7 AdaBoost

In view of multiple variations of AdaBoost, its R 2 regression is illustrated as an example here [34]. Denoting T = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x m , y m ) } as the sample set. The iteration number of the weak learner is K.

The initial weight of the sample set is expressed as follows:

D ( 1 ) = ( ω 11 , ω 12 , , ω 1 m ) ; ω 1 i = 1 m ; i = 1 , 2 , , m .

Supporting k = 1 , 2 , , K :

  1. Gain a weak learner G k ( x ) by training the sample set with weights D k .

  2. Calculate the maximum error of the training set.

    (12) E k = max | y i G k ( x i ) | , i = 1 , 2 , , m .

  3. Calculate the relative error for each sample. Exponent error is expressed as follows:

    (13) e k i = 1 exp | y i G k ( x i ) | E k .

  4. Calculate the regression error rate as follows:

    (14) e k = i = 1 m ω k i e k i .

  5. Calculate the coefficient of the weak learner as follows:

    (15) α k = e k 1 e k .

  6. Update the weight distribution of the sample set as follows:

(16) ω k + 1 , i = ω k i Z k α k 1 e k i .

where the normalization factor Z k is expressed as follows:

(17) Z k = i = 1 m ω k i α k 1 e k i .

Finally, the strong learner is constructed as follows:

(18) f ( x ) = k = 1 K ( ln 1 α k ) g ( x ) ,

where g ( x ) , k = 1 , , K , is the median value of all α k G k ( x ) .

2.8 RF

RF [35] is an ensemble of B trees { T 1 ( X ) , , T B ( X ) } , where X = { x 1 , , x p } is a p-dimensional input vector [36]. The ensemble produces B outputs { Y ˆ 1 = T 1 ( X ) , , Y ˆ B = T B ( X ) } , where Y ˆ b , b = 1 , , B , is a prediction value of the output variable generated by the b-th tree. The outputs of all trees are aggregated to generate one final prediction, Y ˆ . In regression, it is the average prediction of all the individual trees.

The training set, D = { ( X 1 , Y 1 ) , , ( X n , Y n ) } , where X i , i = 1 , , n , is an input vector and Y i is an output vector. The training process is as follows:

  1. A bootstrap sample with replacement is randomly drawn from the training data including n samples.

  2. For each bootstrap sample, a tree will be grown by the following process: at each node, the optimal segmentation is chosen among the subsets, including random m try (rather than all) variables. Here, m try is the only tunable parameter in RF. The tree is grown to its maximum size (i.e., until no further splits are possible) and not pruned back.

  3. Repeat the above steps until B (a sufficiently large number) trees are grown.

  4. Average the predictions of the individual trees to obtain the final prediction as follows:

(19) Y ˆ = 1 B b = 1 B Y b .

2.9 KNN-AdaBoost, LR-AdaBoost, and RF-AdaBoost

Ensemble learning algorithms can significantly promote the generalization ability of learning systems by training a certain number of weak learners and combining them into a strong learner. KNN-AdaBoost, LR-AdaBoost, and RF-AdaBoost are all ensemble algorithms that improve AdaBoosts by training KNN, LR, or RF as weak learners, respectively [37,38,39]. The training process is the same as AdaBoost, and only one difference can be found in the first step. KNN, RF, or LR algorithms are selected as the weak learners G k ( x ) , and the sample set is trained with the weights D k in iterations.

2.10 GBR

GBR fits the negative gradient by iterative process [40,41]. Considering a regression question, n samples, ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) given are fitted into a function F ( x ) with minimal error.

Set the fitted values y ˆ = F ( x ) , square error function L ( y , y ˆ ) = ( y y ˆ ) 2 / 2 , and the total error for all samples J = i = 1 n L ( y i , y ˆ i ) .

The target value y i of the sample is known, and the final result of error J is a numerical scalar. Error J is a function of the n-variable ( y ˆ 1 , y ˆ 2 , , y ˆ n ) for a total of n samples.

Calculate the gradient of the function J as follows:

(20) J = J y ˆ 1 J y ˆ i J y ˆ n ,

J y ˆ i = i = 1 n L ( y i , y ˆ i ) y ˆ i = L ( y i , y ˆ i ) y ˆ i = ( ( y i y ˆ i ) 2 / 2 ) y ˆ i = ( y i y ˆ i ) ,

(21) y i y ˆ i = J y ˆ i .

It can be seen from the above derivation that the residual error is equal to the negative gradient y y ˆ = J .

The weak models are established through iteration, gradually enhanced (boosting) and combined into a strong model.

2.11 ERT

ERT, proposed by Geurts et al. [42] in 2006, is very similar to the RF algorithm, which is composed of many decision trees [43,44]. The two major differences between the two algorithms are described as follows:

  1. The bagging model is applied in the RF algorithm. While all samples are trained to generate each decision tree in ERT, i.e., each decision tree is constructed with the same training samples.

  2. A random subset of RF needs to find the best bifurcation attribute. While the bifurcation value of ERT is obtained completely at random, so as to realize the bifurcation of a decision tree. The ERT model is trained with random features and threshold values. Therefore, the training process of ERT is much faster than RF.

3 Evaluation criteria

The four evaluation criteria, including root mean square error (RMSE), correlation coefficient (R), mean absolute error (MAE), and R 2 (coefficient of determination), were employed to evaluate the results of TFCV, model prediction, and external validation. They are defined as follows:

(22) RMSE = 1 N i = 1 N ( y i y ˆ i ) 2 ,

(23) R = i = 1 N ( y i m 0 ) ( y ˆ i m ˆ 0 ) i = 1 N ( y i m 0 ) 2 i = 1 N ( y ˆ i m ˆ 0 ) 2 ,

(24) MAE = 1 N i = 1 N | y i y ˆ i | ,

(25) R 2 = 1 i = 1 N ( y i y ˆ i ) 2 i = 1 N ( y i m 0 ) 2 ,

where N, y i , y ˆ i , m 0 , and m ˆ 0 denote the number of samples, the measured value, the predicted value, the mean value of y i , and the mean value of y ˆ i , respectively.

4 Results and discussions

In this section, the genetic algorithm (GA) and recursive feature elimination (RFE) are adopted to select variables to form the optimal feature sets. Thirteen technologies, including MLR, CART, Lasso, EN, KNN, SVR, AdaBoost, RF, KNN-AdaBoost, GBR, LR-AdaBoost, ERT, and RF-AdaBoost, were used for predicting the CCR of the industrial BF. The comparative analysis of the above techniques is presented as follows.

4.1 Dataset

The dataset consists of two parts: the training set for modeling and the test set for validation. After excluding the outliers caused by the abnormal state of BF (such as blowing down, record fault, or overhaul) and data with missing values, the training set with 326 samples was established by collecting the historical data throughout a year from an industrial BF (an internal volume of 2,000 m3) with a sampling interval of 1 day. The test set includes 87 samples collected from the same BF in the first 3 months of next year. The disciplinarian of BF burden distribution remained the same during the period of data collection [17].

The dependent variable of the dataset is CCR (kg·t−1). The data for every CCR is the average of a day. According to the experience of BF experts, 36 process variables (shown in Table 1) were chosen as the independent variables of the dataset. The quantity and quality of data are two key factors influencing industrial optimization and constructing a reasonable model. A common thought is that the amount of data needed to construct a plausible data-driven model is at least three times of feature number. The quantity of data in the work meets the above requirement. The quality of the data depends on the spatial coverage of the target variable and the uncertainties associated with the data [45]. Normally, data with a normal distribution is in favor of establishing a reasonable model. Data uncertainty, such as experimental error and input error, could influence the quality of the data. To further analyze the quality of the data, the distribution curves of CCR and the independent variables were plotted in Figures S1 and S2 of Supporting Information, respectively. It can be seen from the two figures that all variables are basically consistent with normal distribution, and the stability of the data is higher because the data show relative concentration. Most of the data measurement instruments have random and systematic errors, so the data collected from the BF system has some degree of uncertainty. On the whole, the quality of the data set used is suitable for data mining and industrial optimization, although the BF data have some inevitable errors [17,19].

Table 1

The list of the independent variables

No. Meanings Features No. Meanings Features
1 Grade of iron (%) GI 19 Iron losses (%) LI
2 Pig iron [Ti] (%) PITi 20 Unit consumption of nut coke (kg·t−1) Ic
3 Pig iron [Si] (%) PISi 21 Small sinter (kg·t−1) SS
4 Blast tmperature (°C) T B 22 Comprehensive ironmaking strength (t/m3·d) SC
5 Top gas pressure (MPa) P Top 23 Feed batch FB
6 Blast volume (m3·min−1) Q B 24 Gas utilization rate (%) R Gas
7 Burden ratio (t·t−1) P 25 Unit consumption of iron ore (kg·t−1) U IO
8 Utilization coefficient (t·m−3·day−1) η v 26 Permeability index IP
9 Slag ratio (kg·t−1) R Sl 27 Gray iron ratio (%) R GI
10 Basicity of slag (%) R Bas 28 Top temperature (°C) T Top
11 Oxygen-enriched rate (%) R Ox 29 Sinter of each batch (t·batch−1) S B
12 Coke ash (%) CA 30 Small sinter of each batch (t·batch−1) SSB
13 Coke sulfur (%) CS 31 Pellet 1 of each batch (t·batch−1) P 1B
14 Coke M40 (%) M 40 32 Pellet 2 of each batch (t·batch−1) P 2B
15 Coke M10 (%) M 10 33 Huili mine of each batch (t·batch−1) M B
16 Coke CSR (%) CSR 34 Batch weight of coke (t·batch−1) C BW
17 Coke <25 mm (%) C 25 35 Blast speed (m·s−1) V B
18 Clinker rate (%) R CL 36 Coal injection ratio (kg·t−1) Y Coal

4.2 Selecting variables

Before establishing the models, variable selection can not only reduce the dimension of feature space to further decrease the risk of over fitting but also better remove some variables unrelated to the target variable and noise interference. Meanwhile, it can also greatly reduce training time and further improve the prediction accuracy and generalization performance of the model [17,46]. Apparently, the model’s performance will be harshly affected, and the model is too complex if the variables irrelevant to the output are not deleted before training a model.

In the work, the primary feature screening was conducted using the Pearson correlation coefficients between the features. First, the max relevance min redundancy (mRMR) method [47] was employed to sort the features. The sorting result is represented in Figure S3. Second, the Pearson correlation coefficients between the features were calculated. Its matrix is shown in Figure S4. If the correlation score between the features is higher than 0.9, the feature with the lower mRMR score will be omitted. After PISi, R CL, and SSB were omitted, the 33 features were retained by the primary screening.

GA based on the learners and RFE based on the base_estimators were used to further optimize the feature sets, respectively. GA, first put forward by Holland [48], is a global optimization algorithm of random search that simulates the processes of inheritance and evolution in natural conditions. It has highly collateral, random, and adaptive features, is an effective method to remove from local optima present on the response surface, and can solve a wide variety of optimization problems with the requirement of no knowledge about the response surface or gradient present on it [49].

RFE is a method that works by selecting data features recursively based on the smallest feature value [50]. In the RFE concept, RFE based on the base_estimators works by eliminating irrelevant features in each iteration, namely the lowest-weight feature. This method is divided into three stages as described in ref. [51]. The two feature selection processes are related to the learners or the base_estimators and are automatically performed during the training of the learners or the base_estimators.

Figure 2 illustrates the implementation processes of variable selection of GA based on the seven learners, respectively. The learners employed are CART, KNN, SVR, KNN-AdaBoost, GBR, LR-AdaBoost, and RF-AdaBoost. Genetic algebra and the best feature set for each GA-learner are listed in Table 2. Figure 3 shows the processes of feature selection of RFE based on the six base_estimators, namely, MLR, Lasso, EN, AdaBoost, RF, and ETR. Table 3 illustrates the best feature set for each RFE – base_estimator. In the variable screening process, the optimal feature sets will be found when R is the maximum or RMSE is the minimum. It can be seen from Tables 2 and 3 that the best feature set is different with different learners or base_estimators, and the feature number for different optimal feature set varies greatly. The least feature number is 4, and the most is 18.

Figure 2 
                  The procedures of variable selection in GA-learners: (a) GA-CART, (b) GA-KNN, (c) GA-SVR, (d) GA-KNN-AdaBoost, (e) GA-GBR, (f) GA-LR-AdaBoost, and (g) GA-RF-AdaBoost.
Figure 2

The procedures of variable selection in GA-learners: (a) GA-CART, (b) GA-KNN, (c) GA-SVR, (d) GA-KNN-AdaBoost, (e) GA-GBR, (f) GA-LR-AdaBoost, and (g) GA-RF-AdaBoost.

Table 2

Genetic algebra and the optimal feature sets in GA-learners

Learner Genetic algebra Selected features
CART 22 GI, T B, P Top, P, CA, C 25, SC, FB, U IO, S B, P 2B, M B, Y Coal
KNN 18 P, η v , R Sl, CSR, LI, Ic, SC, FB, R Gas, U IO, IP, Y Coal
SVR 20 P, M 10, U IO, Y Coal
KNN-AdaBoost 24 GI, P, η v , R Sl, R Bas, R Ox, CA, CS, M 40, CSR, LI, Ic, FB, R Gas, U IO, S B, P 1B, Y Coal
GBR 19 PITi, P Top, P, R Ox, CS, M 40, M 10, CSR, C 25, Ic, SS, SC, FB, R Gas, U IO, IP, T Top, M B, C BW, Y Coal
LR-AdaBoost 15 GI, PITi, P, R Sl, U IO, R GI, T Top, S B, P 1B, P 2B, Y Coal
RF-AdaBoost 17 P Top, P, η v , R Sl, M 40, SC, U IO, P 2B, Y Coal
Figure 3 
                  The procedures of feature selection in RFE based on the base_estimators: (a) MLR, (b) Lasso, (c) EN, (d) AdaBoost, (e) RF, and (f) ETR.
Figure 3

The procedures of feature selection in RFE based on the base_estimators: (a) MLR, (b) Lasso, (c) EN, (d) AdaBoost, (e) RF, and (f) ETR.

Table 3

The optimal feature sets in RFE based on the base_estimators

Base_estimator Selected features
MLR PITi, P Top, R Ox, M 10, C 25, Ic, SS, SC, FB, R Gas, U IO, S B, P 1B, P 2B, M B, C BW, Y Coal
Lasso Q B, CSR, Ic, FB, U IO, IP, R GI, T Top, V B, Y Coal
EN Q B, CSR, Ic, FB, R Gas, U IO, IP, R GI, T Top, C BW, V B, Y Coal
AdaBoost PITi, P, η v , R Sl, LI, Ic, SC, U IO, C BW, Y Coal
RF PITi, P, η v , LI, SC, U IO, Y Coal
ETR P, η v , LI, SC, U IO, C BW, Y Coal

The top three most frequency features selected by the various GA-learners are P, U IO, and Y Coal in the feature subsets. P refers to the amount of ore smelted by unit coke, whose higher value may result in lower CCR. U IO is the amount of iron ore consumption for generating 1 t of pig iron, and the larger U IO content may lead to the higher CCR. In terms of the mechanism of BF operation, Y Coal has large contributions to CCR and their relationship is related to the quality and dosage of pulverized coal and coke.

As shown in Figure 4, the scatter plots between the top important features and CCR are generated for further analysis based on the training data. Figure 4(a) indicates that the larger P tends to result in a smaller CCR value and has a negative correlation with CCR. Meanwhile, Figure 4(b) shows the positive correlation between U IO and CCR, and Figure 4(c) reveals the positive one. Therefore, the relationships between the three features and CCR are consistent with the mechanism of BF operation.

Figure 4 
                  Relationships between the top three most important features: (a) P, (b) U
                     IO, and (c) Y
                     Coal and CCR.
Figure 4

Relationships between the top three most important features: (a) P, (b) U IO, and (c) Y Coal and CCR.

4.3 CCR prediction

The parameters of the models were optimized with the grid search technique and TFCV. The results were compared by three evaluation criteria, namely RMSECV, MAE_CV, and R 2_CV, where “CV” represents “cross validation.” The three methods with the top performances were discovered by TFCV and used to establish the models. The generalization capabilities and robustness of the models were evaluated by using RMSE, MAE, R, and R 2.

4.3.1 Model validation

The TFCV results of six conventional and seven ensemble methods constructed with the optimal features obtained from feature selection are shown in Table 4. Compared with the other methods, KNN-AdaBoost has an inferior prediction performance, for it achieved the largest RMSECV and MAE_CV and the lowest R 2_CV in TFCV. MLR, SVR based on RBF, and LR-AdaBoost exhibit the best predictive properties among all the technologies. LR-AdaBoost is an ensemble learning algorithm that improves AdaBoost with linear regression as weak learner. According to the excellent performance of the MLR and LR-AdaBoost models, a conclusion can be drawn that a certain linear relationship exists between the target variable and the independent variables. Due to sufficiently high R 2_CV and very low RMSECV and MAE_CV, the three models have excellent generalization performance. Therefore, the further optimization of other methods would not be considered.

Table 4

The TFCV results of the thirteen methods

Methods RMSECV MAE_CV R 2_CV
MLR 1.668 1.079 0.973
CART 5.630 3.860 0.757
Lasso 3.018 2.365 0.928
EN 4.776 3.635 0.819
KNN 4.233 2.890 0.882
SVR 1.878 1.158 0.975
AdaBoost 5.073 3.808 0.803
RF 3.711 2.741 0.885
KNN-AdaBoost 6.404 4.882 0.674
GBR 3.459 2.578 0.904
LR-AdaBoost 2.276 1.638 0.953
ERT 4.053 2.976 0.867
RF-AdaBoost 4.226 2.931 0.858

Figure 5 reveals the box plots of RMSECV and R 2_CV for six conventional methods. In Figure 5(a), the technologies are ranked from low to high according to the center of RMSECV: MLR, SVR, Lasso, KNN, EN, and CART. Especially, the center values of RMSECV distribution of the MLR and SVR models are far less than other methods. As shown in Figure 5(b), the zones of R 2_CV for MLR, SVR, and Lasso are much narrower than those of the other methods, and their center values are much higher than those of the others. According to the above analysis, MLR, SVR, and Lasso are more reliable and more precise for predicting CCR than the other techniques.

Figure 5 
                     Box plots of (a) RMSECV and (b) R
                        2_CV for six conventional methods.
Figure 5

Box plots of (a) RMSECV and (b) R 2_CV for six conventional methods.

Figure 6 represents the box plots of RMSECV and R 2_CV of TFCV for seven ensemble methods. In Figure 6(a), the methods are ordered according to the central values of RMSECV in ascending order: LR-AdaBoost, GBR, RF, ERT, RF-AdaBoost, AdaBoost, and KNN-AdaBoost. The RMSECV range of LR-AdaBoost is the narrowest among the seven methods. In Figure 6(b), the centers of LR-AdaBoost, RF, GBR, ERT, and RF-AdaBoost are much higher than those of the others. The methods with a narrow distribution of R 2_CV include LR-AdaBoost, RF, and ERT. On the whole, LR-AdaBoost is the most accurate and stable technique for predicting CCR among all the ensemble techniques.

Figure 6 
                     Box plots of (a) RMSECV and (b) R
                        2_CV for seven ensemble methods.
Figure 6

Box plots of (a) RMSECV and (b) R 2_CV for seven ensemble methods.

In addition, it can be seen from the above two figures that MLR has the superior prediction performance, while SVR and LR-AdaBoost trail in second and third, respectively.

4.3.2 Model training and testing

Based on the above comparison and analysis of the thirteen methods in TFCV, MLR, SVR, and LR-AdaBoost have the top three performances for CCR prediction. This section will illustrate how they would be employed to establish the models and how well the models performed in external validation.

By setting a regression equation, the MLR model can estimate the dependent variable with multiple independent variables and then predict the dependent variable. The independent variables were standardized because of their different units. The larger the value of the standardized regression coefficient, the greater the influence of the independent variable on the dependent variable. The MLR model constructed with the optimal feature set is shown as follows:

(26) CCR = 4.16 PI Ti + 9.05 P Top 1.64 R Ox + 0.82 M 10 0.19 C 25 + 0.24 Ic 0.05 SS + 287.4 SC 1.88 FB 0.41 R Gas + 0.25 U IO 9.05 S B 8.76 P 1B 9.5 P 2B 9.61 M B + 9.25 C BW + 0.21 Y Coal + 348.52 .

Parameter tuning of SVR and LR-AdaBoost is a crucial step to develop models with high generalization performance. The three hyperparameters, C, ε, and σ need to be optimized using the grid search method based on leave-one-out cross-validation before SVR modeling [17]. In the work, the range of ε was from 0.01 to 0.1 with a step of 0.01; C changed from 1 to 100 with an interval of 2; σ varied from 0.5 to 1.4 with a step of 0.1. As can be seen from Figure 7, the higher the C and σ, the larger the RMSE. The RMSE value ranges from 1.867 to 5.470. Obviously, the reasonable selection of C, ε, and σ has a great influence on the prediction performance of the SVR model. After conducting the grid search shown in Figure 7, the optimal C, ε, and σ were determined to be 23, 0.01, and 0.7, respectively, when RMSE (equal to 1.867) was lowest.

Figure 7 
                     The variety of RMSE with σ and C in optimizing SVR’s hyperparameters.
Figure 7

The variety of RMSE with σ and C in optimizing SVR’s hyperparameters.

The SVR model constructed with the optimal C, ε, and σ is shown as follows:

(27) y = i = 1 n β i exp ( 0.7 | | x x i | | 2 ) + 0 .64286,

where x is the unknown vector, x i is the support vector of the SVR model, n is the corresponding sample number, and β i is the Lagrange multiplier of the support vector.

The two hyperparameters of the LR-AdaBoost model, namely n_estimators and learning_rate need to be tuned by using the grid search method. Here, n_estimators is the number of weak learners. Generally, too small n_estimators could cause under-fitting, otherwise could cause over-fitting. Furthermore, learning_rate is the weight reduction factor of the weak learner. In this work, n_estimators varied from 1 to 10 with a step of 1 and from 20 to 100 with a step of 10. Learning_rate changed from 0.1 to 1 with an interval of 0.1. The result indicated that n_estimators and learning_rate were equal to 2 and 0.1, respectively, when RMSE was at a minimum. The optimization process of the parameters is shown in Figure 8, from which it can be found that n_estimators have a more influential impact on RMSE than learning_rate. When n_estimators are less than 20, a smaller RMSE can be obtained.

Figure 8 
                     The variety of RMSE with learning_rate and n_estimators in optimizing LR-AdaBoost’s hyperparameters.
Figure 8

The variety of RMSE with learning_rate and n_estimators in optimizing LR-AdaBoost’s hyperparameters.

The MLR, SVR, and LR-AdaBoost models were employed to predict the samples in the test set. The evaluation criteria of the three models, RMSE, MAE, R, and R 2 are listed in Table 5. The subscripts “TR” and “TE” represent “training” and “test,” respectively. It can be seen from the table that MLR has the lowest RMSETR and MAETR, and the highest R TR and R 2 TR. Moreover, it also performs best in the external test. In model training, SVR is superior to LR-AdaBoost. The reverse is true for the performance indicators generated by the external test of the two technologies. This shows that LR-AdaBoost has stronger extrapolation performance than SVR. Figure 9 represents the scatter plots of the three models. It can be seen from the figure that they could primely fit the experimental data since the predictive data are closely distributed along the diagonal direction. In particular, the training result of SVR is slightly better than LR-AdaBoost, but LR-AdaBoost in external validation is more reliable than SVR. It can be seen from the above analysis that the three methods all possess high prediction accuracy and practical values.

Table 5

The prediction results of the MLR, SVR, and LR-AdaBoost models

Indicator MLR SVR LR-AdaBoost
RMSETR 1.337 1.812 1.855
MAETR 1.037 1.084 1.123
R TR 0.993 0.988 0.986
R 2 TR 0.987 0.976 0.973
RMSETE 1.593 2.464 1.863
MAETE 1.328 1.887 1.508
R TE 0.989 0.981 0.987
Figure 9 
                     The scatter plots of the (a) MLR, (b) SVR, and (c) LR-AdaBoost models.
Figure 9

The scatter plots of the (a) MLR, (b) SVR, and (c) LR-AdaBoost models.

5 Conclusions

In the work, a comparative study of data-driven prediction methods for the CCR of BF was carried out. Six conventional methods, including MLR, CART, Lasso, EN, KNN, and SVR, and seven ensemble methods, namely, AdaBoost, RF, KNN-AdaBoost, GBR, LR-AdaBoost, ERT, and RF-AdaBoost, were investigated from the application point of view. The TFCV results found that KNN-AdaBoost had the lowest competitiveness for predicting CCR. Meanwhile, MLR had the optimal prediction performance among all the methods, while SVR and LR-AdaBoost trailed in second and third place. Furthermore, the SVR model possessed better training performance than the LR-AdaBoost model, but the LR-AdaBoost method appeared more reliable for external validation. On the whole, the three methods owned high practical values and generalization performance because of their models’ extremely high R and very low RMSE and MAE. The study integrated 13 techniques to solve the industrial optimization problem with strong noise and multivariable couple. The method outlined here can provide valuable hints into industry optimization with the assistance of machine learning and has important instructions and practical value for assisting managers to control BF parameters and detect BF status.

Acknowledgments

The authors gratefully acknowledge the financial supports from the Sichuan Science and Technology Program (No.2022YFG0318) and Sichuan Technology & Engineering Research Program for Vanadium Titanium Materials of China (No. 2020-2FTGC-YB-01).

  1. Funding information: The work was supported by the Sichuan Science and Technology Program (No. 2022YFG0318) and Sichuan Technology & Engineering Research Program for Vanadium Titanium Materials of China (No. 2020-2FTGC-YB-01).

  2. Author contributions: X.Z. contributed to the writing of the original draft, review, and editing and assisted with the methodology, formal analysis, validation, resources, and funding acquisition. M.C. contributed to the writing of the original draft and assisted with the formal analysis, visualization, project administration, and investigation.

  3. Conflict of interest: The authors state that there is no conflict of interest.

  4. Data availability statement: All data, models, or code that support the findings of this study are available from the corresponding author upon reasonable request.

References

[1] Kuang, S., Z. Li, and A. Yu. Review on modeling and simulation of blast furnace. Steel Research International, Vol. 89, No. 1, 2017, id. 1700071.10.1002/srin.201700071Search in Google Scholar

[2] Li, J. P., C. C. Hua, Y. N. Yang, and X. P. Guan. Fuzzy classifier design for development tendency of hot metal silicon content in blast furnace. IEEE Transactions on Industrial Informatics, Vol. 14, No. 3, 2018, pp. 1115–1123.10.1109/TII.2017.2770177Search in Google Scholar

[3] Roche, M., M. Helle, J. van der Stel, G. Louwerse, L. Shao, and H. Saxen. On-line estimation of liquid levels in the blast furnace hearth. Steel Research International, Vol. 90, No. 3, 2019, id. 1800420.10.1002/srin.201800420Search in Google Scholar

[4] Nielson, S., T. Okosun, B. Damstedt, M. Jampani, and C. Q. Zhou. Tuyere-level syngas injection in the blast furnace: a computational fluid dynamics investigation. Processes, Vol. 9, No. 8, 2021, id. 1447.10.3390/pr9081447Search in Google Scholar

[5] Li, Z. N., M. S. Chu, Z. G. Liu, G. J. Ruan, and B. F. Li. Furnace heat prediction and control model and its application to large blast furnace. High Temperature Materials and Processes, Vol. 38, 2019, pp. 884–891.10.1515/htmp-2019-0049Search in Google Scholar

[6] Dong, X. F., P. Zulli, and M. Biasutti. Prediction of blast furnace hearth condition: Part II – A transient state simulation of hearth condition during blast furnace shutdown. Ironmaking & Steelmaking, Vol. 47, No. 5, 2020, pp. 561–566.10.1080/03019233.2018.1561386Search in Google Scholar

[7] Bernasowski, M., A. Klimczyk, and R. Stachura. Support algorithm for blast furnace operation with optimal fuel consumption. Journal of Mining and Metallurgy, Section B: Metallurgy, Vol. 55, No. 1, 2019, pp. 31–38.10.2298/JMMB180206010BSearch in Google Scholar

[8] Guha, M. Revealing cohesive zone shape and location inside blast furnace. Ironmaking & Steelmaking, Vol. 45, No. 9, 2018, pp. 787–792.10.1080/03019233.2017.1338385Search in Google Scholar

[9] La, G. H., J. S. Choi, and D. J. Min. Investigation on the reaction behaviour of partially reduced iron under blast furnace conditions. Metals-Basel, Vol. 11, No. 5, 2021, id. 839.10.3390/met11050839Search in Google Scholar

[10] Li, S., J. C. Chang, M. S. Chu, J. Li, and A. M. Yang. A blast furnace coke ratio prediction model based on fuzzy cluster and grid search optimized support vector regression. Applied Intelligence, Vol. 52, 2022, pp. 13533–13542.10.1007/s10489-022-03234-8Search in Google Scholar

[11] Roche, M., M. Helle, J. van der Stel, G. Louwerse, L. Shao, and H. Saxen. Off-line model of blast furnace liquid levels. ISIJ International, Vol. 58, No. 12, 2018, pp. 2236–2245.10.2355/isijinternational.ISIJINT-2018-417Search in Google Scholar

[12] Shiau, J. S. and C. K. Ho. A visualization technique to predict abnormal channeling phenomena in the blast furnace operation. Mining, Metallurgy & Exploration, Vol. 36, No. 2, 2019, pp. 423–430.10.1007/s42461-018-0003-0Search in Google Scholar

[13] Zhang, X., M. Kano, and S. Matsuzaki. A comparative study of deep and shallow predictive techniques for hot metal temperature prediction in blast furnace ironmaking. Computers & Chemical Engineering, Vol. 130, 2019, id. 106575.10.1016/j.compchemeng.2019.106575Search in Google Scholar

[14] Zhang, X., M. Kano, and S. Matsuzaki. Ensemble pattern trees for predicting hot metal temperature in blast furnace. Computers & Chemical Engineering, Vol. 121, 2019, pp. 442–449.10.1016/j.compchemeng.2018.10.022Search in Google Scholar

[15] Li, J. P., C. C. Hua, Y. N. Yang, and X. P. Guan. Data-driven Bayesian-based Takagi-Sugeno fuzzy modeling for dynamic prediction of hot metal silicon content in blast furnace. IEEE Transactions on Systems, Man, and Cybernetics: Systems, Vol. 52, No. 2, 2022, pp. 1087–1099.10.1109/TSMC.2020.3013972Search in Google Scholar

[16] Sun, W. Q., Z. H. Wang, and Q. Wang. Hybrid event-, mechanism- and data-driven prediction of blast furnace gas generation. Energy, Vol. 199, 2020, id. 117497.10.1016/j.energy.2020.117497Search in Google Scholar

[17] Zhai, X. Y., M. T. Chen, and W. C. Lu. Fuel ratio optimization of blast furnace based on data mining. ISIJ International, Vol. 60, No. 11, 2020, pp. 2471–2476.10.2355/isijinternational.ISIJINT-2020-238Search in Google Scholar

[18] Li, J. L., R. J. Zhu, P. Zhou, Y. P. Song, and C. Q. Zhou. Prediction of the cohesive zone in a blast furnace by integrating CFD and SVM modelling. Ironmaking & Steelmaking, Vol. 48, No. 3, 2021, pp. 284–291.10.1080/03019233.2020.1771893Search in Google Scholar

[19] Hu, Y., H. Zhou, S. Yao, M. Kou, Z. Zhang, L. P. Wang, et al. Comprehensive evaluation of the blast furnace status based on data mining and mechanism analysis. International Journal of Chemical Reactor Engineering, Vol. 20, No. 2, 2022, pp. 225–235.10.1515/ijcre-2021-0160Search in Google Scholar

[20] Li, W. B. and Z. Y. Chen. Breathing rate estimation based on multiple linear regression. Computer Methods in Biomechanics and Biomedical Engineering, Vol. 25, No. 7, 2022, pp. 772–782.10.1080/10255842.2021.1977801Search in Google Scholar PubMed

[21] Al-Najjar, H. A. H. and B. Pradhan. Spatial landslide susceptibility assessment using machine learning techniques assisted by additional data created with generative adversarial networks. Geoscience Frontiers, Vol. 12, No. 2, 2021, pp. 625–637.10.1016/j.gsf.2020.09.002Search in Google Scholar

[22] Huang, F. M., Z. Ye, S. H. Jiang, J. S. Huang, Z. L. Chang, and J. W. Chen. Uncertainty study of landslide susceptibility prediction considering the different attribute interval numbers of environmental factors and different data-based models. Catena, Vol. 202, No. 2, 2021, id. 117406.10.1016/j.catena.2021.105250Search in Google Scholar

[23] Shahzad, S. J. H., E. Bouri, T. Ahmad, and M. A. Naeem. Extreme tail network analysis of cryptocurrencies and trading strategies. Finance Research Letters, Vol. 44, 2022, id. 102106.10.1016/j.frl.2021.102106Search in Google Scholar

[24] Liu, D., S. Baldi, W. W. Yu, J. D. Cao, and W. Huang. On training traffic predictors via broad learning structures: A benchmark study. IEEE Transactions on Systems, Man, and Cybernetics: Systems, Vol. 52, No. 2, 2022, pp. 749–758.10.1109/TSMC.2020.3006124Search in Google Scholar

[25] Scannell Bryan, M., J. Sun, J. Jagai, D. E. Horton, A. Montgomery, R. Sargis, et al. Coronavirus disease 2019 (COVID-19) mortality and neighborhood characteristics in Chicago. Annals of Epidemiology, Vol. 56, 2021, pp. 47–54.10.1016/j.annepidem.2020.10.011Search in Google Scholar PubMed PubMed Central

[26] Sharif, M., M. A. Khan, M. Rashid, M. Yasmin, F. Afza and U. J. Tanik. Deep CNN and geometric features-based gastrointestinal tract diseases detection and classification from wireless capsule endoscopy images. Journal of Experimental & Theoretical Artificial Intelligence, Vol. 33, No. 4, 2021, pp. 577–599.10.1080/0952813X.2019.1572657Search in Google Scholar

[27] Ebrahimi-Khusfi, Z., R. Taghizadeh-Mehrjardi, and M. Mirakbari. Evaluation of machine learning models for predicting the temporal variations of dust storm index in arid regions of Iran. Atmospheric Pollution Research, Vol. 12, No. 1, 2021, pp. 134–147.10.1016/j.apr.2020.08.029Search in Google Scholar

[28] Cortes, C. and V. Vapnik. Support-vector networks. Machine Learning, Vol. 20, No. 3, 1995, pp. 273–297.10.1007/BF00994018Search in Google Scholar

[29] Yang, X., L. Li, Q. L. Tao, W. C. Lu, and M. J. Li. Rapid discovery of narrow bandgap oxide double perovskites using machine learning. Computational Materials Science, Vol. 196, 2021, id. 110528.10.1016/j.commatsci.2021.110528Search in Google Scholar

[30] Zhao, Q. Z., Y. Liu, W. Q. Yao, and Y. B. Yao. Hourly rainfall forecast model using supervised learning algorithm. IEEE Transactions on Geoscience and Remote Sensing, Vol. 60, 2022, pp. 1–9.10.1109/TGRS.2021.3054582Search in Google Scholar

[31] Gu, E. X. Convolutional neural network based Kannada-MNIST classification. 2021 IEEE International Conference on Consumer Electronics and Computer Engineering, IEEE, Guangzhou, China, 15–17 January 2021, pp. 180–185.10.1109/ICCECE51280.2021.9342474Search in Google Scholar

[32] Rodriguez-Perez, R. and J. Bajorath. Evolution of support vector machine and regression modeling in chemoinformatics and drug discovery. Journal of Computer-Aided Molecular Design, Vol. 36, No. 5, 2022, pp. 355–362.10.1007/s10822-022-00442-9Search in Google Scholar PubMed PubMed Central

[33] Zhen, Z., T. Potta, N. A. Lanzillo, K. Rege, and C. M. Breneman. Development of a web-enabled SVR-based machine learning platform and its application on modeling transgene expression activity of aminoglycoside-derived polycations. Combinatorial Chemistry & High Throughput Screening, Vol. 20, No. 1, 2017, pp. 41–55.10.2174/1386207319666161228124214Search in Google Scholar PubMed

[34] Chen, W. and Y. Li. GIS-based evaluation of landslide susceptibility using hybrid computational intelligence models. Catena, Vol. 195, 2020, id. 104777.10.1016/j.catena.2020.104777Search in Google Scholar

[35] Du, P. J., A. Samat, B. Waske, S. C. Liu, and Z. H. Li. Random forest and rotation forest for fully polarized SAR image classification using polarimetric and spatial features. ISPRS Journal of Photogrammetry and Remote Sensing, Vol. 105, 2015, pp. 38–53.10.1016/j.isprsjprs.2015.03.002Search in Google Scholar

[36] Speiser, J. L., M. E. Miller, J. Tooze, and E. Ip. A comparison of random forest variable selection methods for classification prediction modeling. Expert Systems with Applications, Vol. 134, 2019, pp. 93–101.10.1016/j.eswa.2019.05.028Search in Google Scholar PubMed PubMed Central

[37] Li, H. H., B. Zhang, W. W. Hu, Y. Liu, C. W. Dong, and Q. S. Chen. Monitoring black tea fermentation using a colorimetric sensor array-based artificial olfaction system. Journal of Food Processing and Preservation, Vol. 42, No. 1, 2018, id. e13348.10.1111/jfpp.13348Search in Google Scholar

[38] Wei, M. M., H. X. Lu, and H. H. Yang. Research on boold species identification algorithm based on RF_AdaBoost model. Chemical Journal of Chinese Universities-Chinese, Vol. 41, No. 1, 2020, pp. 94–101.Search in Google Scholar

[39] Gupta, K. K., K. Kalita, R. K. Ghadai, M. Ramachandran, and X. Z. Gao. Machine learning-based predictive modelling of biodiesel production – A comparative perspective. Energies, Vol. 14, No. 4, 2021, id. 1122.10.3390/en14041122Search in Google Scholar

[40] Touzani, S., J. Granderson, and S. Fernandes. Gradient boosting machine for modeling the energy consumption of commercial buildings. Energy and Buildings, Vol. 158, 2018, pp. 1533–1543.10.1016/j.enbuild.2017.11.039Search in Google Scholar

[41] Natekin, A. and A. Knoll. Gradient boosting machines, a tutorial. Frontiers in Neurorobotics, Vol. 7, 2013, id. 21.10.3389/fnbot.2013.00021Search in Google Scholar PubMed PubMed Central

[42] Geurts, P., D. Ernst, and L. Wehenkel. Extremely randomized trees. Machine Learning, Vol. 63, No. 1, 2006, pp. 3–42.10.1007/s10994-006-6226-1Search in Google Scholar

[43] Saeed, U., S. U. Jan, Y. D. Lee, and I. Koo. Fault diagnosis based on extremely randomized trees in wireless sensor networks. Reliability Engineering & System Safety, Vol. 205, 2021, id. 107284.10.1016/j.ress.2020.107284Search in Google Scholar

[44] Wei, J., Z. Li, M. Cribb, W. Huang, W. Xue, L. Sun, et al. Improved 1 km resolution PM2.5 estimates across China using enhanced space-time extremely randomized trees. Atmospheric Chemistry and Physics, Vol. 20, No. 6, 2020, pp. 3273–3289.10.5194/acp-20-3273-2020Search in Google Scholar

[45] Tao, Q., P. Xu, M. Li, and W. Lu. Machine learning for perovskite materials design and discovery. NPJ Computational Materials, Vol. 7, No. 1, 2021, pp. 171–88.10.1038/s41524-021-00495-8Search in Google Scholar

[46] Shi, L., D. P. Chang, X. B. Ji, and W. C. Lu. Using data mining to search for perovskite materials with higher specific surface area. Journal of Chemical Information and Modeling, Vol. 58, No. 12, 2018, pp. 2420–2427.10.1021/acs.jcim.8b00436Search in Google Scholar PubMed

[47] Yu, D. R., S. An, and Q. H. Hu. Fuzzy mutual information based min-redundancy and max-relevance heterogeneous feature selection. International Journal of Computational Intelligence Systems, Vol. 4, No. 4, 2011, pp. 619–633.10.1080/18756891.2011.9727817Search in Google Scholar

[48] Holland, J. H. Genetic algorithms. Scientific American, Vol. 267, No. 1, 1992, pp. 66–72.10.1038/scientificamerican0792-66Search in Google Scholar

[49] Holland, J. H. Building blocks, cohort genetic algorithms, and hyperplane-defined functions. Evolutionary Computation, Vol. 8, No. 4, 2000, pp. 373–391.10.1162/106365600568220Search in Google Scholar PubMed

[50] Jeon, H. and S. Oh. Hybrid-recursive feature elimination for efficient feature selection. Applied Sciences-Basel, Vol. 10, No. 9, 2020, id. 3211.10.3390/app10093211Search in Google Scholar

[51] Bustamam, A., A. Bachtiar, and D. Sarwinda. Selecting features subsets based on support vector machine recursive features elimination and one dimensional-naïve Bayes classifier using support vector machines for classification of prostate and breast Cancer, 4th International Conference on Computer Science and Computational Intelligence 2019 (ICCSCI). pp. 450–458.10.1016/j.procs.2019.08.238Search in Google Scholar

Received: 2022-08-27
Revised: 2022-12-01
Accepted: 2022-12-02
Published Online: 2023-01-30

© 2023 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Research Articles
  2. First-principles investigation of phase stability and elastic properties of Laves phase TaCr2 by ruthenium alloying
  3. Improvement and prediction on high temperature melting characteristics of coal ash
  4. First-principles calculations to investigate the thermal response of the ZrC(1−x)Nx ceramics at extreme conditions
  5. Study on the cladding path during the solidification process of multi-layer cladding of large steel ingots
  6. Thermodynamic analysis of vanadium distribution behavior in blast furnaces and basic oxygen furnaces
  7. Comparison of data-driven prediction methods for comprehensive coke ratio of blast furnace
  8. Effect of different isothermal times on the microstructure and mechanical properties of high-strength rebar
  9. Analysis of the evolution law of oxide inclusions in U75V heavy rail steel during the LF–RH refining process
  10. Simultaneous extraction of uranium and niobium from a low-grade natural betafite ore
  11. Transfer and transformation mechanism of chromium in stainless steel slag in pedosphere
  12. Effect of tool traverse speed on joint line remnant and mechanical properties of friction stir welded 2195-T8 Al–Li alloy joints
  13. Technology and analysis of 08Cr9W3Co3VNbCuBN steel large diameter thick wall pipe welding process
  14. Influence of shielding gas on machining and wear aspects of AISI 310–AISI 2205 dissimilar stainless steel joints
  15. Effect of post-weld heat treatment on 6156 aluminum alloy joint formed by electron beam welding
  16. Ash melting behavior and mechanism of high-calcium bituminous coal in the process of blast furnace pulverized coal injection
  17. Effect of high temperature tempering on the phase composition and structure of steelmaking slag
  18. Numerical simulation of shrinkage porosity defect in billet continuous casting
  19. Influence of submerged entry nozzle on funnel mold surface velocity
  20. Effect of cold-rolling deformation and rare earth yttrium on microstructure and texture of oriented silicon steel
  21. Investigation of microstructure, machinability, and mechanical properties of new-generation hybrid lead-free brass alloys
  22. Soft sensor method of multimode BOF steelmaking endpoint carbon content and temperature based on vMF-WSAE dynamic deep learning
  23. Mechanical properties and nugget evolution in resistance spot welding of Zn–Al–Mg galvanized DC51D steel
  24. Research on the behaviour and mechanism of void welding based on multiple scales
  25. Preparation of CaO–SiO2–Al2O3 inorganic fibers from melting-separated red mud
  26. Study on diffusion kinetics of chromium and nickel electrochemical co-deposition in a NaCl–KCl–NaF–Cr2O3–NiO molten salt
  27. Enhancing the efficiency of polytetrafluoroethylene-modified silica hydrosols coated solar panels by using artificial neural network and response surface methodology
  28. High-temperature corrosion behaviours of nickel–iron-based alloys with different molybdenum and tungsten contents in a coal ash/flue gas environment
  29. Characteristics and purification of Himalayan salt by high temperature melting
  30. Temperature uniformity optimization with power-frequency coordinated variation in multi-source microwave based on sequential quadratic programming
  31. A novel method for CO2 injection direct smelting vanadium steel: Dephosphorization and vanadium retention
  32. A study of the void surface healing mechanism in 316LN steel
  33. Effect of chemical composition and heat treatment on intergranular corrosion and strength of AlMgSiCu alloys
  34. Soft sensor method for endpoint carbon content and temperature of BOF based on multi-cluster dynamic adaptive selection ensemble learning
  35. Evaluating thermal properties and activation energy of phthalonitrile using sulfur-containing curing agents
  36. Investigation of the liquidus temperature calculation method for medium manganese steel
  37. High-temperature corrosion model of Incoloy 800H alloy connected with Ni-201 in MgCl2–KCl heat transfer fluid
  38. Investigation of the microstructure and mechanical properties of Mg–Al–Zn alloy joints formed by different laser welding processes
  39. Effect of refining slag compositions on its melting property and desulphurization
  40. Effect of P and Ti on the agglomeration behavior of Al2O3 inclusions in Fe–P–Ti alloys
  41. Cation-doping effects on the conductivities of the mayenite Ca12Al14O33
  42. Modification of Al2O3 inclusions in SWRH82B steel by La/Y rare-earth element treatment
  43. Possibility of metallic cobalt formation in the oxide scale during high-temperature oxidation of Co-27Cr-6Mo alloy in air
  44. Multi-source microwave heating temperature uniformity study based on adaptive dynamic programming
  45. Round-robin measurement of surface tension of high-temperature liquid platinum free of oxygen adsorption by oscillating droplet method using levitation techniques
  46. High-temperature production of AlN in Mg alloys with ammonia gas
  47. Review Article
  48. Advances in ultrasonic welding of lightweight alloys: A review
  49. Topical Issue on High-temperature Phase Change Materials for Energy Storage
  50. Compositional and thermophysical study of Al–Si- and Zn–Al–Mg-based eutectic alloys for latent heat storage
  51. Corrosion behavior of a Co−Cr−Mo−Si alloy in pure Al and Al−Si melt
  52. Al–Si–Fe alloy-based phase change material for high-temperature thermal energy storage
  53. Density and surface tension measurements of molten Al–Si based alloys
  54. Graphite crucible interaction with Fe–Si–B phase change material in pilot-scale experiments
  55. Topical Issue on Nuclear Energy Application Materials
  56. Dry synthesis of brannerite (UTi2O6) by mechanochemical treatment
  57. Special Issue on Polymer and Composite Materials (PCM) and Graphene and Novel Nanomaterials - Part I
  58. Heat management of LED-based Cu2O deposits on the optimal structure of heat sink
  59. Special Issue on Recent Developments in 3D Printed Carbon Materials - Part I
  60. Porous metal foam flow field and heat evaluation in PEMFC: A review
  61. Special Issue on Advancements in Solar Energy Technologies and Systems
  62. Research on electric energy measurement system based on intelligent sensor data in artificial intelligence environment
  63. Study of photovoltaic integrated prefabricated components for assembled buildings based on sensing technology supported by solar energy
  64. Topical Issue on Focus of Hot Deformation of Metaland High Entropy Alloys - Part I
  65. Performance optimization and investigation of metal-cored filler wires for high-strength steel during gas metal arc welding
  66. Three-dimensional transient heat transfer analysis of micro-plasma arc welding process using volumetric heat source models
Downloaded on 10.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/htmp-2022-0261/html
Scroll to top button