Home An efficient recurrent neural network with ensemble classifier-based weighted model for disease prediction
Article Open Access

An efficient recurrent neural network with ensemble classifier-based weighted model for disease prediction

  • Tamilselvi Kesavan EMAIL logo and Ramesh Kumar Krishnamoorthy
Published/Copyright: August 25, 2022
Become an author with De Gruyter Brill

Abstract

Day-to-day lives are affected globally by the epidemic coronavirus 2019. With an increasing number of positive cases, India has now become a highly affected country. Chronic diseases affect individuals with no time identification and impose a huge disease burden on society. In this article, an Efficient Recurrent Neural Network with Ensemble Classifier (ERNN-EC) is built using VGG-16 and Alexnet with weighted model to predict disease and its level. The dataset is partitioned randomly into small subsets by utilizing mean-based splitting method. Various models of classifier create a homogeneous ensemble by utilizing an accuracy-based weighted aging classifier ensemble, which is a weighted model’s modification. Two state of art methods such as Graph Sequence Recurrent Neural Network and Hybrid Rough-Block-Based Neural Network are used for comparison with respect to some parameters such as accuracy, precision, recall, f1-score, and relative absolute error (RAE). As a result, it is found that the proposed ERNN-EC method accomplishes accuracy of 95.2%, precision of 91%, recall of 85%, F1-score of 83.4%, and RAE of 41.6%.

1 Introduction

Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) is a novel coronavirus that causes coronavirus disease (COVID-19) in the respiratory tract [1]. This epidemic causes serious disasters across the world at an instance and affects nearly 216 nations, with 44 million cases and 1.1 million deaths, according to the statistics of WHO on November 02, 2020 [2]. Because of insufficient amenities for intensive care units, severe illness has reached a point, where health systems, including those in developed countries, have failed. Medical sciences adopted machine learning method for this disease prediction [3]. To be specific, RT-PCR is the gold standard in COVID-19 prediction, but it consumes a lot of time, laboratory, and complex manual process, which are in shortage. Moreover, RT-PCR sensitivity varies highly and the results are not clear and consistent, and in China its initial findings are with low relative sensitivity [4]. Its succeeding results produce increased variable positive rate based on the specimen gathered, and the positive rate also decreases with time after the initial symptom. The process of optimization and improvement of these results are proceeded by the researchers. Machine learning tasks are enhanced by an approach called ensemble learning [5]. Individual classifier set is combined together in an ensemble classifier using a process similar to majority voting, which is the combination of component prediction. This ensemble classifier performs better compared to the traditional classifiers [6]. A single base learner or algorithm is present in the members of homogeneous ensemble learning [7]. In the meantime, different structures are associated with members. However, various base learners are present in the members of heterogeneous ensemble learning [8]. Many machine learning approaches are motivated towards the enhancement of COVID disease risk prediction and performance of classification. Recurrent Neural Network with ensemble classifiers are mainly used for various disease predictions such as heart disease, Alzheimer’s disease, etc. VGG means Visual Geometry Group with multiple layers in convolution neural network architecture. The VGG-16 or VGG-19 consists of 16 and 19 convolution layers. Alexnet is the ImageNet Classification with Deep Convolutional Neural Networks, which is used in the machine learning field for classification. Based on this, this article proposes a type of homogeneous ensemble learning approach. The contribution of this work is as follows:

  • To construct an Efficient Recurrent Neural Network with Ensemble Classifier (ERNN-EC) by utilizing VGG-16 and Alexnet with a weighted model for predicting the disease and its level.

This article is presented as follows: Section 1 defines the background of COVID disease prediction and the role of neural network-based classification in disease prediction. In Section 2, the literature for various disease predictions using neural network classification is given. Section 3 explains the proposed neural network for disease prediction. In Section 4, the experimental analysis is given with graphs by comparing them with two standard methods. Finally, Section 5 concludes the article by presenting the conclusion and future work.

2 Related works

In, ref. [9], an artificial neural network-dependent classifier utilizing multilayer perceptron with backpropagation algorithm was developed for peak event prediction (peak demand days) in patients with respiratory diseases such as asthma and chronic obstructive pulmonary disease. Peak event class achieves a precision of 77.1% and a recall of 78.0%, whereas nonpeak events achieves a precision of 83.9% and a recall of 83.2%. The entire system accuracy is 81.0%. Graph Sequence Recurrent Neural Network (GS-RNN) is proposed in ref. [10] in which FoG patterns are characterized through the emerging graph recurrent cells, which carry dynamic structure’s graph sequences as inputs. Intended for unavailable prior edge annotation cases, a data-driven dependent adjacency estimation approach is also developed and this first approach of vision-based FoG detection utilizing DNN that is intended for dynamic structure’s sequences of graph. An appropriate classifier is selected in a block-based neural network (BBNN) in ref. [11] because of its capability of evolving internal structures and dynamic environment adaptability. Application characteristics are incorporated systematically in this architecture for structuring Hybrid rough-Block-Based Neural Network (R-BBNN). Hybrid particle swarm optimization having wavelet mutation is the global training algorithm developed for optimizing the parameter of R-BBNN. Improved CNN-based trained network named VggNet16-wbn is presented in ref. [12], which is the derivation of VggNet16 (pre-trained CNN algorithm) (Table 1).

Table 1

Comparison of GS-RNN and the hybrid R-BBNN

S. No GS-RNN Hybrid R-BBNN
1 GS-RNN optimization is based on weight regularization, activation function optimization, and batch normalization R-BBNN optimization is a weight-based genetic algorithm
2 GS-RNN estimates the sample value at time frame utilizing the values of multiple frames R-BBNN model includes some restrictions such as 2D array and integer weights for estimating sample values
3 The gradient problem vanishing is mitigated by this algorithm The engineering problems such as pattern classification and mobile robot control

Transfer learning, which is a deep learning approach, as well as automated ocular detection method, is discussed in this article. A CNN-based trained network architecture was present depending on network assessment from six CNN pre-trained networks for detecting conjunctiva. Dual-input neural network presented in ref. [13] was used in the feature extraction and deep learning method integration. Multiple domains are used for the extraction of ECG and PCG features and ratio of information gain is used for significant feature selection. In this article, signals of ECG and decayed PCG are integrated as five-channel signal. Automated detection of kidney disease using sensing method is presented in ref. [14]. The concentration of salivary urea is observed the disease detection. The saliva sample’s urea levels are monitored by this approach. Moreover, the raw signals are analyzed from the sensor and the 1D DL-CNN algorithm is implemented by incorporating support vector machine (SVM) classifier. Some of the previous methods of leveraged artificial neural networks architectures and several well-performed methods that are used in the statistical approaches, such as LRM, for developing prediction models to estimate the liver failure likelihood are discussed. These statistics approaches permit the researchers to develop the prediction models for predicting results based on the independent variable set. Sometimes, incorrect independent variables are incorporated by the researchers. Likewise, statistical models that appeared susceptible to bullishness and had high prediction power resulted in sampling bias. All these limitations can be easily overcome by ERNN-EC as described in Section 3.

3 System model

The neural network-based disease prediction architecture is constructed in a homogeneous manner as shown in Figure 1. Initially, the datasets are preprocessed and trained using Recurrent Neural Network with Perception Algorithm (RNNPA). Second, the ensemble classifiers, such as VGG-16 and Alexnet, are used for better classification with the reduction of error in the output side. Finally, the learning process is done using weighted model for finalizing the level of disease.

Figure 1 
               System architecture for neural network-based disease prediction.
Figure 1

System architecture for neural network-based disease prediction.

4 Dataset description

The clinical data and demographics of the COVID-19-positive patient’s clinical data and its demography are included in the dataset. This dataset contains COVID-19 records of 287 patients with binary class labels known as “survived” and “deceased,” which are 243 and 44, respectively. The BodyTemp 1 field designates the first temperature of body that is reserved when the patient is admitted to the hospital. The reading of the last temperature of the body reserved prior to the discharge of patient is indicated in BodyTemp 2. Likewise, breath shortness is represented by shortness of breath, chronic disease diabetes mellitus is represented by chr-dm, hypertension is represented by chr-Htn, cardiovascular diseases are represented by chr-cardiac, and dyslipidemia and cough are represented by chr-dlp.

5 Preprocessing and training using RNNPA

In RNNPA, the neuron’s preliminary mission within the input layer is to partition the input signal xi among the neurons in the hidden layer, which is represented in Figure 2. Each j th neuron in the hidden layer enhances its x i input signals by weighting them with the respective connection strengths w j i from the input layer, and its output y j is determined as a sum function f , as provided in the following equation:

(1) y ( j ) = f i = 0 n w ( j i ) x ( i ) .

The working of RNNPA is summarized as follows:

Figure 2 
               Architecture of recurrent neural network with perception.
Figure 2

Architecture of recurrent neural network with perception.

Step 1: For processing, input layer is provided with input data, from which predicted output is obtained; X = x 1 , x 2 , x 3 , , x n 1 , x n is the sequence provided. As t is the timestep, the input is x t R k . At the same time [15], the hidden state, h t R d , is given as shown in the following equation:

(2) h t = f ( h t 1 , x t ) = f ( U h t 1 + W x t + b ) .

Step 2: The value of error is calculated by the difference between predicted output and actual output.

Step 3: Back-propagation with the Stochastic-based Nelder–Mead method is used in this network to reduce complexity while using the standard back-propagation method. Here, in every iteration, the training instance subset or batch is used. As a result, the error computed is biased as per optimum, but it was performed quickly. Every iteration in the whole training set needs numerous iterations in the small batches, and adjusting weights are provided in equations (3)–(8):

(3) i t = β ( U t h t 1 + W i x t + b i ) ,

(4) f t = β ( U f h t 1 + W f x t + b f ) ,

(5) 0 t = β ( U o h t 1 + W o x t + b o ) ,

(6) c t = tanh ( U c h t 1 + W c x t + b c ) ,

(7) c t = f t c t 1 + i t c t ,

(8) h t = O t tanh ( c t ) .

Step 4: The weights of the output layer node and last hidden layers node weight are weighed for adjusting weights by the network that works backward. Input x with non-linear function sigmoid ( σ ) or tanh described in equation (9) [16] is triggered in hidden neurons, and linear function f ( x ) = x is triggered in output neuron.

(9) σ ( x ) = 1 1 + e x .

Step 5: As back propagation is completed, the forwarding process begins again.

The reset gate with reduced error permits the hidden layer to drop any information without future prediction, and the amount of information that are controlled by the updated gate from the prior hidden layer will remained in the present hidden layer. Particularly, the input vector x t is appended to train this model, and an additional dimension denotes the time as log of the number of days between t , time of event, and, T is date of index. Logarithmic transformation is applied for minimizing the duration’s skewed distribution

Input: Training data

Output: Error reduced data

Training data (T) = Weighted Difference (Vector.T1,Vector.T2)

( y j , x j ) j = 0 n j < n , where x j and y j = 0 , 1 standardize the weights ω r .

For each and every feature, weak classifier h i is trained.

A classifier h i with error ε i is considered to match the weight ω r

ε i = ω r , j h i ( y j ) x j n j = 1 .

Set iter = 0

While iter = 0

For each view

H j ( x ) c l a s s c j

i f H j ( x ) = w i c j i ( x )

H i ( x ) = H j ( x )

End if

Base = Base + Weighted

Trial = Aggregate(Target,Base)

if M F ( Trial ) < M F ( Base ) then

Trial = Base

else

Target = Base

end if

6 Ensemble classifier

Two RNN models, VGG-16 and Alexnet, used in this experiment, use 13 convolutional layers and 3 fully connected layers. Input for the neural network is an image with 224 × 224 × 3 size. The filters are 3 × 3 matrices and the stride of which is fixed to 1. The padding size is always 1, while max-pooling is carried out in a pixel window of size 2 × 2 , with stride of 2. Rectified Linear Unit (ReLU) layer, which always follows the convolution layer in VGG-16, boosts the nonlinearity of VGG-16 . The convolution layers present within both the pooling layers possess same channel number, kernel size, and stride. Actually, collecting two 3 × 3 convolution layers and three 3 × 3 convolution kernels is equal to a single 5 × 5 and 7 × 7 convolution layer, respectively. Stacking 2 or 3 small convolution kernels works much quicker than a single huge convolution kernel. Moreover, parameter numbers have been minimized. ReLU layers that are inserted between undersized convolution layers are really useful. To eliminate the trade-off problem, transfer learning as well as multiple pre-trained Alexnet is applied. Suitable approaches are required in this Alex-Net model for faster training, and overfitting is prevented because of its complicated structure, several training parameters, and huge training data. As Alex-Net model is built and nonlinearity of ReLU in the data structures directly utilized for making initialized approach is highly consistent with theory and network training began directly from initial point for enhancing the speed of training, local response normalization was carried out in the process known as “near-suppression” operation, which efficiently enhances the performance generalization of this approach by normalizing the local input regions. The proposed framework for ensemble learning incorporates multiple fusion levels of miscellaneous classifiers that are trained on various sets of features. Likewise, the extraction of feature set done by particular approach is processed further to obtain various subsets of features using various methods of feature selection. In the third layer, multiple classifiers’ training, base classifiers are trained by using m learning algorithms on every feature set Fi, thus Ei is the preliminary ensemble formed on every n feature sets that are present in primary fusion layer. Finally, primary ensembles n designed on feature set n are bonded further for creating the final ensemble, because the final classification is finished as the final ensemble output are presented in the final fusion layer. The cross-validated loss for base learner j , which is defined first [17], is given as shown in the following equation:

(10) R c v = v = 1 v i val ( v ) n l ( y i , p i j ) ,

where, in the v th fold, the indices set of observations is indicated by val( v ), and the i th observation with prediction is defined as p i j , using the base learner j that is trained on the entire data except the fold v .

The task of binary classification is considered simple as it can be easily generalized for multi-class classification and regression. The single algorithms of Super Learner’s simple version are first studied using negative (Bernoulli) log-likelihood as loss function and are represented in the following equations:

(11) l ( y , p ) = [ y log ( p ) + ( 1 y ) log ( 1 p ) ] ,

(12) R c v = v = 1 v i val ( v ) n y i log j = 1 m a j p j i + ( 1 y i ) log ( i a j p j i ) ,

where, i th unit predicted probability from j th base learner is denoted by p j i , which is trained on the entire data except v th fold. For K -class classification having softmax output, such as neural networks, the score level in ensemble is given as shown in the following equation:

(13) P z i ( a ) = exp j = 1 m a j . s j [ j , z ] k = 1 k exp j = 1 m a j . s j [ j , z ] .

where the i th unit ensemble prediction and class z with weight vector “a” is P z i ( a ) . m X K matrix is s i and model j and class k score is stand by s i [ j , k ] .

7 Learning process using weighted model

The variable x ( i ) ’s weighted mean is given as shown in the following equation:

(14) X ( i ) = 1 N i = 1 N w i x i j , w i > 0 .

The weighted mean occupies more contribution of data elements having increased weight, and it occupies less contribution of data elements with low weights. There should not be any negative weights, and there are zero weights. The algorithm of data partitioning is developed by introducing the randomization, i.e., at every partitioning, a random variable dataset is used by this algorithm.

In the latter, individual outputs and different weights are combined to take any of the forms described below [18]:

  • For every classifier, particular weight is assigned, and the combined output for class c j i is given as shown in the following equation:

(15) H j ( x ) = 1 T i = 1 T w i c j i ( x ) ,

where Hj is the classifier with the assigned weight wi.

  • For every classifier h i , class-specific weight w ( i ) j is assigned, and the combined output for class cji is given as shown in the following equation:

(16) H j ( x ) = 1 T i = 1 T w i ( j ) c j i ( x ) .

  • For every sample in a class of every classifier, weight is applied, and the combined output of cji is given as shown in the following equation:

(17) H j ( x ) = i = 1 T n i = 1 T w i ( j ) H j ( x ) ,

where w i ( j ) is the weight assigned to instance x k of c j for H j .

8 Detection of disease and its level

Later, in the training of classifiers, an equivalent process is carried out for the testing set: the feature selected is acquired from the training set. The mean values are obtained by imputing the missing values from the training set and processing the z -score depending on the consistent mean and standard deviation features from the training set.

At last, the trained classifier’s output on the testing set with the real labels is compared for the performance evaluation. Every result of validation and associated attribute subsets are registered in a list that is sorted as per metrics. Ties are solved in the following order: Level 1, very low; Level 2, low; Level 3, Moderate; and Level 4, high. In the conclusion of process, top subset is the best subset. During the testing stage, the algorithm is applied on both the training union and the dataset of testing, acquiring the attribute subset’s rank. In the final case, the top of the rank subset is called the optimal attribute subset (OAS). It is evident that the final rank will not be obtainable in practice, as the dataset for testing denotes samples which are not found in samples obtained in the future. The attribute rank combinations creat the OAS that is utilized for defining a performance metric, which is utilized only for the purpose of evaluation.

9 Performance analysis

Performance evaluation was carried out using the parameters such as accuracy, Precision, recall, F 1-score, and relative absolute error (RAE). These parameters are contrasted with two existing methods, such as GS-RNN and Hybrid R-BBNN, together with the proposed ERNN-EC (Table 2).

Table 2

Parameters for operation

Operational parameters Values
Size of the batch 15
Rate drop for learning 1 × 1 0 2
Period of learn rate drop 30
Hidden layers 8
Epochs 160
Batch normalization epsilon 1 0 4
Initial learn rate 1 × 1 0 2

The investigation parameters, including accuracy, precision, recall, F1-score and RAE, are discussed as follows:

  • Accuracy: The network model capability is generally predicted by accuracy. The classifier capacity is computed by true positive (TP) and true negative (TN), which are used for the calculation of lack and being there of disease. The false prediction amount obtained by the model is recognized by false negative (FN) and false positive (FP). The accuracy formula is provided in the following equation:

(18) Accuracy = TP + TN TP + TN + FP + FN .

Table 3 represents the accuracy of comparison among the existing GS-RNN and R-BBNN methods with the ERNN-EC method.

Table 3

Analysis of accuracy

Number of samples GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
200 88 90 92
400 90 91 94
600 92 93 95
800 93 95 97
1,000 94 96 98

Figure 3 illustrates the comparison of accuracy between existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method, where X axis represents the number of datasets utilized for examination and accuracy acquired in percentage is represented in Y axis. When compared, existing GS-RNN and R-BBNN methods achieve 91.4 and 93%, whereas the proposed ERNN-EC method achieves 95.2% of accuracy which is 3.8% better than GS-RNN and 2.2% better than R-BBNN.

Figure 3 
               Comparison of accuracy.
Figure 3

Comparison of accuracy.

Precision: The model of disease classification’s entire achievement is indicated by Precision. The classification function’s likelihood is forecasted for getting a positive rate when the disease is present. The true positive (TP) amount is recognized and it is calculated as given in the following equation:

(19) Precision ( P ) = TP TP + FP .

Table 4 shows the comparison of precision between the existing GS-RNN and R-BBNN with the proposed ERNN-EC method.

Table 4

Analysis of precision

Number of samples GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
200 82 84 87
400 84 86 88
600 85 89 92
800 87 91 93
1,000 90 92 95

Figure 4 illustrates the comparison of precision between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method, where X axis represents the number of datasets utilized for evaluation and precision acquired in percentage is represented in Y axis. When compared, the existing GS-RNN and R-BBNN methods achieve 85.6 and 88.4%, whereas the proposed ERNN-EC method achieves 91% of precision which is 5.4% better than GS-RNN and 2.6% better than R-BBNN.

Figure 4 
               Comparison of precision.
Figure 4

Comparison of precision.

Recall designates the classifier likelihood that attains negative result when there is no disease and also called true negative (TN) rate, which is computed by the following equation:

(20) Recall ( R ) = TP TP + FN .

Table 5 represents the comparison results of recall among the existing GS-RNN and R-BBNN with the proposed ERNN-EC method.

Table 5

Analysis of recall

Number of samples GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
200 74 77 80
400 76 79 82
600 79 82 85
800 80 85 88
1,000 83 87 90

Figure 5 illustrates the comparison of recall between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method, where X axis represents the number of datasets used for analysis and the recall acquired in percentage is represented in Y axis. When compared, the existing GS-RNN and R-BBNN methods achieve 78.4 and 82%, whereas the proposed ERNN-EC method achieves 85% of recall which is 6.6% better than GS-RNN and 3% better than R-BBNN.

Figure 5 
               Comparison of recall.
Figure 5

Comparison of recall.

F 1 -score is a feat for establishing prediction performance. It is built by evaluating the precision and recall’s harmonic portion. If a score value is computed as 1, then it is restrained as highly brilliant, and if results are bad, then the score value is 0. The true negative rate is not considered by F 1-score in its account. The F 1-score is computed as shown in the following equation:

(21) F 1 -score = 2 P R P + R .

Table 6 shows the comparison of F 1-score between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method.

Table 6

Analysis of F 1-score

Number of samples GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
200 71 75 79
400 73 79 81
600 75 81 84
800 78 82 85
1,000 81 85 88

Figure 6 represents the F 1-score’s comparison between the existing GS-RNN and R-BBNN methods and proposed ERNN-EC method, where X axis represents the number of datasets utilized for evaluation and Y axis indicates the F 1-score acquired in percentage. When compared, the existing GS-RNN and R-BBNN methods achieve 75.6 and 80.4%, whereas the proposed ERNN-EC method achieves 83.4% of F 1-score, which is 7.8% better than GS-RNN and 3% better than R-BBNN.

  • RAE

The comparison of a mean error (residual) to errors obtained by an unimportant or naive model ratio is known as RAE. The formula of RAE is given as shown in the following equation:

(22) RAE = i = 1 n ( p i A i ) 2 i = 1 n A i .

Table 7 shows the comparison of RAE between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method.

Figure 6 
               Comparison of 
                     
                        
                        
                           F
                        
                        F
                     
                  1-score.
Figure 6

Comparison of F 1-score.

Table 7

Analysis of RAE

Number of samples GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
200 46 40 33
400 48 42 37
600 51 45 42
800 53 49 46
1,000 55 52 50

Figure 7 illustrates the comparison of RAE between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method, where X axis indicates the number of datasets utilized for evaluation and RAE acquired in percentage is represented as Y axis. When compared, the existing GS-RNN and R-BBNN methods achieve 50.6 and 45.6%, whereas the proposed ERNN-EC method achieves 41.6% of RAE, which is 9% better than GS-RNN and 4% better than R-BBNN.

Figure 7 
               Comparison of RAE.
Figure 7

Comparison of RAE.

Table 8 shows the overall comparison between the existing GS-RNN and R-BBNN methods and the proposed ERNN-EC method.

Table 8

Overall comparative analysis between existing and proposed methods

Parameters GS-RNN [10] R-BBNN [11] ERNN-EC [proposed]
Accuracy (%) 91.4 93 95.2
Precision (%) 85.6 88.4 91
Recall (%) 78.4 82 85
F 1-score (%) 75.6 80.4 83.4
RAE (%) 50.6 45.6 41.6

10 Conclusion

In the neural network field, significant role is performed by deep learning and data mining. The core objective is to enhance the accuracy of predictive model. The COVID-19 disease is efficiently predicted in this research. This approach uses ERNN-EC to randomly partition a dataset into smaller subsets based on the classification entered into the model with every partition. An accuracy-based weighted aging classifier is used in the creation of the homogeneous ensemble. This approach is contrasted with the two existing methods, like GS-RNN and Hybrid R-BBNN, in terms of parameters such as accuracy, precision, recall, F 1-score, and RAE. As a result, it is found that the proposed ERNN-EC method accomplishes an accuracy of 95.2%, precision of 91%, recall of 85%, F 1-score of 83.4%, and RAE of 41.6%. The future work concentrates on constructing optimization-based autoregressive models for improving the accuracy of neural networks.

Acknowledgments

All the contents and images used in the articles are cited and acknowledged in the reference section.

  1. Conflict of interest: The authors state no conflict of interest.

References

[1] Li J, Xi L, Wang X, Fang F, Lv X, Zhang D, et al. Radiology indispensable for tracking COVID-19. Diagnostic Intervent Imaging 2020;102(2):69–75. 10.1016/j.diii.2020.11.008Search in Google Scholar

[2] Salehi S, Abedi A, Balakrishnan S, Gholamrezanezhad A. Coronavirus disease 2019 (COVID-19): a systematic review of imaging findings in 919 patients. Am J Roentgenol. 2020;215(1):87–93. 10.2214/AJR.20.23034Search in Google Scholar

[3] Wang S, Zha Y, Li W, Wu Q, Li X, Niu M, et al. A fully automatic deep learning system for COVID-19 diagnostic and prognostic analysis. Europ Respirat J. 2020;56(2):2000775. 10.1183/13993003.00775-2020Search in Google Scholar

[4] Huang C, Wang Y, Li X, Ren L, Zhao J, Hu Y, et al. Clinical features of patients infected with 2019 Novel Coronavirus in Wuhan China. Te Lancet. 2020;395:497–506. 10.1016/S0140-6736(20)30183-5Search in Google Scholar

[5] Guan WJ, Hu Y, Ni ZY. Clinical characteristics of Coronavirus disease 2019 in China. N Engl J Med. 2020;382(18):1708–20. 10.1056/NEJMoa2002032Search in Google Scholar PubMed PubMed Central

[6] Shang F, Zhou K, Liu H, Cheng J, Tsang IW, Zhang L, et al. VR-SGD: a simple stochastic variance reduction method for machine learning. IEEE Trans Knowledge Data Eng. Jan. 2020;32(1):188–202. 10.1109/TKDE.2018.2878765. Search in Google Scholar

[7] Ebenuwa SH, Sharif MS, Alazab M, Al-Nemrat A. Variance ranking attributes selection techniques for binary classification problem in imbalance data. IEEE Access. 2019;7:24649–66. 10.1109/ACCESS.2019.2899578. Search in Google Scholar

[8] Rahim R, Murugan S, Priya S, Magesh S, Manikandan R. Taylor based grey wolf optimization algorithm (TGWOA) for energy aware secure routing protocol, Int J Comput Netw Appl (IJCNA). 2020;7(4):93–102. 10.22247/ijcna/2020/196041Search in Google Scholar

[9] Khatri KL, Tamil LS. Early detection of peak demand days of chronic respiratory diseases emergency department visits using artificial neural networks. IEEE J Biomed Health Informatics. 2017;22(1):285–90. 10.1109/JBHI.2017.2698418Search in Google Scholar PubMed

[10] Hu K, Wang Z, Wang W, Martens KAE, Wang L, Tan T. Graph sequence recurrent neural network for vision-based freezing of gait detection. IEEE Trans Image Process. 2019;29:1890–901. 10.1109/TIP.2019.2946469Search in Google Scholar PubMed

[11] San PP, Ling SH, Nguyen H. Evolvable rough-block-based neural network and its biomedical application to hypoglycemia detection system. IEEE Trans Cybernetics. 2013;44(8):1338–49. 10.1109/TCYB.2013.2283296Search in Google Scholar PubMed

[12] Zamani NSM, WMDW Zaki, Huddin AB, Hussain A, Mutalib HA Ali A. Automated pterygium detection using deep neural network. IEEE Access 2020;8:191659–72. 10.1109/ACCESS.2020.3030787Search in Google Scholar

[13] Li H, Wang X, Liu C, Wang Y, Li P, Tang H. Dual-input neural network integrating feature extraction and deep learning for coronary artery disease detection using electrocardiogram and phonocardiogram. IEEE Access 2019;7:146457–69. 10.1109/ACCESS.2019.2943197Search in Google Scholar

[14] Bhaskar N, Manikandan S. A deep-learning-based system for automated sensing of chronic kidney disease. IEEE Sensors Letters 2019;3(10):1–4. 10.1109/LSENS.2019.2942145Search in Google Scholar

[15] Rajanayagam J, Frank R, Shepherd RW, Lewindon PJ. Artificial neural network is highly predictive of outcome in pediatric acute liver failure. Pediatric Transplant. 2013;17(6):535–42. 10.1111/petr.12100Search in Google Scholar PubMed

[16] Valero D, Bung DB. Artificial Neural Networks and pattern recognition for air-water flow velocity estimation using a single-tip optical fibre probe. J Hydro Environ Res. 2018;19:150–9. 10.1016/j.jher.2017.08.004Search in Google Scholar

[17] Lai CC, Su KL. Development of an intelligent mobile robot localization system using Kinect RGB-D mapping and neural network. Comput. Electr. Eng. 2018;67:620–8. 10.1016/j.compeleceng.2016.04.018Search in Google Scholar

[18] Erkaymaz O, Ozer M, Perc M. Performance of small-world feedforward neural networks for the diagnosis of diabetes. Appl Math Comput. 2017;311:22–8. 10.1016/j.amc.2017.05.010Search in Google Scholar

Received: 2021-10-30
Revised: 2021-12-29
Accepted: 2022-05-07
Published Online: 2022-08-25

© 2022 Tamilselvi Kesavan and Ramesh Kumar Krishnamoorthy, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Research Articles
  2. Construction of 3D model of knee joint motion based on MRI image registration
  3. Evaluation of several initialization methods on arithmetic optimization algorithm performance
  4. Application of visual elements in product paper packaging design: An example of the “squirrel” pattern
  5. Deep learning approach to text analysis for human emotion detection from big data
  6. Cognitive prediction of obstacle's movement for reinforcement learning pedestrian interacting model
  7. The application of neural network algorithm and embedded system in computer distance teach system
  8. Machine translation of English speech: Comparison of multiple algorithms
  9. Automatic control of computer application data processing system based on artificial intelligence
  10. A secure framework for IoT-based smart climate agriculture system: Toward blockchain and edge computing
  11. Application of mining algorithm in personalized Internet marketing strategy in massive data environment
  12. On the correction of errors in English grammar by deep learning
  13. Research on intelligent interactive music information based on visualization technology
  14. Extractive summarization of Malayalam documents using latent Dirichlet allocation: An experience
  15. Conception and realization of an IoT-enabled deep CNN decision support system for automated arrhythmia classification
  16. Masking and noise reduction processing of music signals in reverberant music
  17. Cat swarm optimization algorithm based on the information interaction of subgroup and the top-N learning strategy
  18. State feedback based on grey wolf optimizer controller for two-wheeled self-balancing robot
  19. Research on an English translation method based on an improved transformer model
  20. Short-term prediction of parking availability in an open parking lot
  21. PUC: parallel mining of high-utility itemsets with load balancing on spark
  22. Image retrieval based on weighted nearest neighbor tag prediction
  23. A comparative study of different neural networks in predicting gross domestic product
  24. A study of an intelligent algorithm combining semantic environments for the translation of complex English sentences
  25. IoT-enabled edge computing model for smart irrigation system
  26. A study on automatic correction of English grammar errors based on deep learning
  27. A novel fingerprint recognition method based on a Siamese neural network
  28. A hidden Markov optimization model for processing and recognition of English speech feature signals
  29. Crime reporting and police controlling: Mobile and web-based approach for information-sharing in Iraq
  30. Convex optimization for additive noise reduction in quantitative complex object wave retrieval using compressive off-axis digital holographic imaging
  31. CRNet: Context feature and refined network for multi-person pose estimation
  32. Improving the efficiency of intrusion detection in information systems
  33. Research on reform and breakthrough of news, film, and television media based on artificial intelligence
  34. An optimized solution to the course scheduling problem in universities under an improved genetic algorithm
  35. An adaptive RNN algorithm to detect shilling attacks for online products in hybrid recommender system
  36. Computing the inverse of cardinal direction relations between regions
  37. Human-centered artificial intelligence-based ice hockey sports classification system with web 4.0
  38. Construction of an IoT customer operation analysis system based on big data analysis and human-centered artificial intelligence for web 4.0
  39. An improved Jaya optimization algorithm with ring topology and population size reduction
  40. Review Articles
  41. A review on voice pathology: Taxonomy, diagnosis, medical procedures and detection techniques, open challenges, limitations, and recommendations for future directions
  42. An extensive review of state-of-the-art transfer learning techniques used in medical imaging: Open issues and challenges
  43. Special Issue: Explainable Artificial Intelligence and Intelligent Systems in Analysis For Complex Problems and Systems
  44. Tree-based machine learning algorithms in the Internet of Things environment for multivariate flood status prediction
  45. Evaluating OADM network simulation and an overview based metropolitan application
  46. Radiography image analysis using cat swarm optimized deep belief networks
  47. Comparative analysis of blockchain technology to support digital transformation in ports and shipping
  48. IoT network security using autoencoder deep neural network and channel access algorithm
  49. Large-scale timetabling problems with adaptive tabu search
  50. Eurasian oystercatcher optimiser: New meta-heuristic algorithm
  51. Trip generation modeling for a selected sector in Baghdad city using the artificial neural network
  52. Trainable watershed-based model for cornea endothelial cell segmentation
  53. Hessenberg factorization and firework algorithms for optimized data hiding in digital images
  54. The application of an artificial neural network for 2D coordinate transformation
  55. A novel method to find the best path in SDN using firefly algorithm
  56. Systematic review for lung cancer detection and lung nodule classification: Taxonomy, challenges, and recommendation future works
  57. Special Issue on International Conference on Computing Communication & Informatics
  58. Edge detail enhancement algorithm for high-dynamic range images
  59. Suitability evaluation method of urban and rural spatial planning based on artificial intelligence
  60. Writing assistant scoring system for English second language learners based on machine learning
  61. Dynamic evaluation of college English writing ability based on AI technology
  62. Image denoising algorithm of social network based on multifeature fusion
  63. Automatic recognition method of installation errors of metallurgical machinery parts based on neural network
  64. An FCM clustering algorithm based on the identification of accounting statement whitewashing behavior in universities
  65. Emotional information transmission of color in image oil painting
  66. College music teaching and ideological and political education integration mode based on deep learning
  67. Behavior feature extraction method of college students’ social network in sports field based on clustering algorithm
  68. Evaluation model of multimedia-aided teaching effect of physical education course based on random forest algorithm
  69. Venture financing risk assessment and risk control algorithm for small and medium-sized enterprises in the era of big data
  70. Interactive 3D reconstruction method of fuzzy static images in social media
  71. The impact of public health emergency governance based on artificial intelligence
  72. Optimal loading method of multi type railway flatcars based on improved genetic algorithm
  73. Special Issue: Evolution of Smart Cities and Societies using Emerging Technologies
  74. Data mining applications in university information management system development
  75. Implementation of network information security monitoring system based on adaptive deep detection
  76. Face recognition algorithm based on stack denoising and self-encoding LBP
  77. Research on data mining method of network security situation awareness based on cloud computing
  78. Topology optimization of computer communication network based on improved genetic algorithm
  79. Implementation of the Spark technique in a matrix distributed computing algorithm
  80. Construction of a financial default risk prediction model based on the LightGBM algorithm
  81. Application of embedded Linux in the design of Internet of Things gateway
  82. Research on computer static software defect detection system based on big data technology
  83. Study on data mining method of network security situation perception based on cloud computing
  84. Modeling and PID control of quadrotor UAV based on machine learning
  85. Simulation design of automobile automatic clutch based on mechatronics
  86. Research on the application of search algorithm in computer communication network
  87. Special Issue: Artificial Intelligence based Techniques and Applications for Intelligent IoT Systems
  88. Personalized recommendation system based on social tags in the era of Internet of Things
  89. Supervision method of indoor construction engineering quality acceptance based on cloud computing
  90. Intelligent terminal security technology of power grid sensing layer based upon information entropy data mining
  91. Deep learning technology of Internet of Things Blockchain in distribution network faults
  92. Optimization of shared bike paths considering faulty vehicle recovery during dispatch
  93. The application of graphic language in animation visual guidance system under intelligent environment
  94. Iot-based power detection equipment management and control system
  95. Estimation and application of matrix eigenvalues based on deep neural network
  96. Brand image innovation design based on the era of 5G internet of things
  97. Special Issue: Cognitive Cyber-Physical System with Artificial Intelligence for Healthcare 4.0.
  98. Auxiliary diagnosis study of integrated electronic medical record text and CT images
  99. A hybrid particle swarm optimization with multi-objective clustering for dermatologic diseases diagnosis
  100. An efficient recurrent neural network with ensemble classifier-based weighted model for disease prediction
  101. Design of metaheuristic rough set-based feature selection and rule-based medical data classification model on MapReduce framework
Downloaded on 12.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jisys-2022-0068/html
Scroll to top button