Startseite DeepCOVIDNet-CXR: deep learning strategies for identifying COVID-19 on enhanced chest X-rays
Artikel Open Access

DeepCOVIDNet-CXR: deep learning strategies for identifying COVID-19 on enhanced chest X-rays

  • Gokhan Altan ORCID logo EMAIL logo und Süleyman Serhan Narli ORCID logo
Veröffentlicht/Copyright: 8. Oktober 2024
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

Objectives

COVID-19 is one of the recent major epidemics, which accelerates its mortality and prevalence worldwide. Most literature on chest X-ray-based COVID-19 analysis has focused on multi-case classification (COVID-19, pneumonia, and normal) by the advantages of Deep Learning. However, the limited number of chest X-rays with COVID-19 is a prominent deficiency for clinical relevance. This study aims at evaluating COVID-19 identification performances using adaptive histogram equalization (AHE) to feed the ConvNet architectures with reliable lung anatomy of airways.

Methods

We experimented with balanced small- and large-scale COVID-19 databases using left lung, right lung, and complete chest X-rays with various AHE parameters. On multiple strategies, we applied transfer learning on four ConvNet architectures (MobileNet, DarkNet19, VGG16, and AlexNet).

Results

Whereas DarkNet19 reached the highest multi-case identification performance with an accuracy rate of 98.26 % on the small-scale dataset, VGG16 achieved the best generalization performance with an accuracy rate of 95.04 % on the large-scale dataset.

Conclusions

Our study is one of the pioneering approaches that analyses 3615 COVID-19 cases and specifies the most responsible AHE parameters for ConvNet architectures in the multi-case classification.

Introduction

Coronavirus Disease 2019 (COVID-19) is a pandemic with high mortality and contagiousness rates in the global population. It is in the same respiratory disease family as severe acute respiratory syndrome (SARS) and Middle East respiratory syndrome (MERS). Although SARS and MERS have affected specific locations in the world with mortality rates of 10 and 37 %, the prevalence of COVID-19 has already outperformed them in an exponential acceleration [1]. COVID-19 infects the lungs, reveals ruinous effects on the respiratory tract, and causes death related to gasping for each age group. Developments in medical image processing have enhanced computerized techniques for diagnosing and monitoring diseases.

Using novel techniques in diagnosis models procures quick and standardized abnormality detection on medical image processing [2].

Deep Learning (DL) has the ability to analyze the entire medical image through the advantages of feature learning. In contrast, conventional medical image processing depends on the local and morphological features to characterize the pathological region of interest. Convolutional neural networks (ConvNet), the most popular DL algorithms, extract low- and high-level features with sequential convolutions [3]. The depth of feature learning defines the complexity of the ConvNet architecture for feeding a deeper supervised classifier with responsible feature maps in DL. ConvNet comprises two main stages: feature learning and supervised learning. Whereas responsible feature maps are sequentially transferred among convolution blocks, fully connected (FC) layers support supervised learning with many units, regularization methods, and optimization algorithms. Besides the widespread use of ConvNet in various machine learning tasks, the primary justification of ConvNet is transfer learning that enables reusing the same knowledge to reduce the training time for related tasks. AlexNet [3], VGGNet [4], GoogLeNet [5], DarkNet19 [6], Inception modules, EfficientNet [7], and ResNet [8], 9] are among the most popular ConvNet architectures.

The easy accessibility of chest X-rays made them a widely used diagnostic tool for detecting COVID-19 [8], [10], [11], [12]. Even though most researchers shared the experimental results of chest X-ray analysis to catch COVID-19, the number of cases is small to produce a robust evaluation and modelling of DL architectures with high generalization capability.

Yoon et al. analyzed the correlation of chest X-ray and CT images in identifying COVID-19 with nine COVID-19-positive cases. They highlighted the efficiency of chest X-rays at least as well as CT [10]. Moreover, the popularity of feature learning on the ConvNet architectures brings stimulating approaches, including COVID-Net [12] and ResNet50 [8]. Nevertheless, the COVID-19-positive cases were insufficient in 31 and 25 for COVID-Net and ResNet50, respectively. Similarly, He et al. used a ConvNet to identify COVID-19-positive cases on 100 chest X-rays [9]. Das et al. fine-tuned InceptionNetV3 architecture to separate COVID-19 from healthy subjects. They experimented with various combinations of six COVID-19 chest X-ray datasets. In addition, they classified multi-cases (COVID-19, healthy, tuberculosis, and pneumonia cases) [13]. Asnaoui and Chawki also compared the deep architectures on chest X-rays and CT. They reported the performance of VGG16, VGG19, ResNet50, MobileNet, Inception-ResNetV2, and DenseNet201 on 231 COVID-19-positive cases [14]. Mukherjee et al. proposed a ConvNet architecture for various batch sizes, pooling, and convolutional filter sizes on 439 COVID-19-positive cases. They also outperformed popular ConvNet architectures, including, MobileNet, InceptionNetV3, and ResNet50 [15]. Minaee et al. utilized transfer learning and different adaptive optimization techniques and batch sizes on ResNet18, ResNet50, SqueezeNet, and DenseNet121 architectures. They experimented with two datasets on 184 COVID-19-positive cases and 5 K non-COVID cases [16]. Singh et al. proposed a multi-stage COVID-19 detection model using VGG16, ResNet50, DenseNet121, and DenseNet169 on 573 COVID-19-positive cases (473 train, 50 validation, and 50 test). They handled the lung segmentation module and generative adversarial networks to generate more practical chest X-rays in the preprocessing stage. They emphasized the effect of preprocessing for a robust evaluation, even for ConvNets [17]. Oh et al. also adapted a segmentation module and histogram equalization for the identification of multi-cases (healthy, pneumonia, and COVID-19) using ResNet-18 architecture on 180 COVID-19-positive cases (126 train, 18 validation, and 36 test) [18]. Ozturk et al. applied transfer learning using DarkNet19 architecture on 125 COVID-19-positive cases (5-fold cross-validation). They experimented with binary-case (COVID-19 and non-COVID-19) and multi-case classification (healthy, pneumonia, and COVID-19) and reported the achievements of DarkNet19 in learning realistic pathological patterns in chest X-rays [6]. Chowdhury et al. combined various COVID-19 chest X-ray databases for detecting COVID-19 using ConvNet architectures. They re-trained popular ConvNet architectures, including, MobileNetV2, SqueezeNet, ResNet18, ResNet101, and DenseNet201, on 423 COVID-19-positive cases (304 training, 34 validation, and 85 test). They reached the outperforming achievements against ChexNet using ResNet18 for binary-cases (COVID-19 and non-COVID-19) classification [19]. Rajpurkar et al. developed ChexNet, an initiator of reliable ConvNet analysis, using over 112 K chest X-rays with 14 pulmonary diseases [20]. Even though Conventional ChexNet did not include COVID-19-positive cases, the DenseNet architecture with 121 layers has become a baseline for much existing research on chest X-ray analysis. Haghanifar et al. re-trained the weights of ChexNet architecture, which used the DenseNet121 framework as a baseline feature learning, on 428 images (80 % train and 20 % test) for a variety of fully-connected layer parameters in the supervised stage [21]. Taspinar et al. used the feature learning stage of the VGG19 architecture to feed the conventional machine learning algorithms, including support vector machines and artificial neural networks on 800 COVID-19-positive cases. They reported the impact of feature learning on multi-case classification (healthy, pneumonia, and COVID-19) [22].

Among a considerable number of researchers on COVID-19 identification on chest X-rays in Table 1, most focused on proposing multi-stage ConvNet models (lung segmentation and pathology identification), compulsive generative DL algorithms for improving the quality, and global preprocessing stages for noise removals on small number of COVID-19-positive cases. The literature has limitations in being used as a guiding diagnostic with clinical relevance. The proposals must analyze large number of COVID-19 datasets for a high generalization capability. Moreover, the architectures need to apply adaptive preprocessing techniques with a detailed parameter presentation. Herein, the proposal aimed to identify multi-cases (COVID-19, healthy, and pneumonia) using adaptive techniques as preprocessing and designate the impact of localization parameters for chest X-rays with shallow ConvNet architectures. The main significance of the paper is analyzing COVID-19 on separate lung lobes for a specific impact definition on the performance of feature learning with popular ConvNet architectures, designating ConvNet-based adaptive preprocessing parameters for generating reliable lung anatomy of airways. In this paper, we generated new presentations of raw chest X-rays using various grid sizes and clip limits on Adaptive histogram equalization (AHE) [23], which enables global enhancing the chest X-ray without loss of contrast. We fed the popular shallow ConvNet architectures, including AlexNet, MobileNet, DarkNet19, and VGG16, using a reliable airway quality of the left lung, right lung, and both lungs from chest X-rays. In this manner, the generalization capacity of different chest X-ray sections for various AHE parameters could be designated, particularly for ConvNet architectures for future chest X-ray analysis. Feeding the ConvNet architectures with more significant and responsible pathology of lung airways to clinicians could implement more acceptable clinical relevance using the advantages of DL.

Table 1:

A complete comparison of the state-of-the-art in terms of overall accuracy and case-based classification performances (%) for COVID-19, normal, and pneumonia.

Related works Pre-processing Architecture Accuracy Case Subjects Precision Recall F1 score
Singh et al. [17] GAN VGG19 + pruned naive 98.67 COVID-19 473 100 100 100
Normal 8,851 98.00 98.00 98.00
Pneumonia 6,041 98.00 98.00 98.00
Wang et al. [12] Cropping, data augmentation (flip) Tailored COVID-Net 93.34 COVID-19 358 98.91 91.00 94.79
Normal 8,066 90.47 95.00 92.68
Pneumonia 5,538 91.26 94.00 92.61
Oh et al. [18] HE, segmentation (DenseNet103), ROI ResNet18 + adam optimizer 88.90 COVID-19 180 76.90 100 86.94
Normal 191 95.70 90.00 92.76
Pneumonia 54 90.30 93.00 91.63
Das et al. [13] Truncated InceptionNet 99.96 COVID-19 162 95.00 99.00 97.00
Normal 1,583
Pneumonia 4,280
Ozturk et al. [6] DarkNet19 + adam optimizer 87.02 COVID-19 127 80.70 97.87 88.46
Normal 500 89.64 86.64 88.11
Pneumonia 500 85.71 85.37 85.54
El asnaoui and chawki [14] Intensity normalization, AHE Inception ResNetV2 92.18 COVID-19 231 93.85 82.80 87.98
Normal 1,583 94.40 97.75 96.05
Pneumonia 2,780 88.88 95.77 92.20
Ucar and korkmaz [36] Data augmentation (shearing, noise, brightness) SqueezeNet + bayesian optimization 98.26 COVID-19 76 100 99.35 99.67
Normal 1,583 98.04 97.40 97.72
Pneumonia 4,290 96.73 98.01 97.37
Chowdhury et al. [19] DenseNet201 + SGDM 97.94 COVID-19 423 99.29 99.06 99.17
Normal 1,485 97.97 97.85 97.91
Pneumonia 1,579 97.51 97.71 97.61
Haghanifar et al. [21] HE, AHE, segmentation (U-net) DenseNet121 87.21 COVID-19 780 94.20 90.28 92.20
Normal 5,000 82.52 95.17 88.39
Pneumonia 4,600 92.42 77.98 84.58
Mangal et al. [35] DenseNet121 90.52 COVID-19 155 96.77 74.36 84.10
Normal 1,583 98.86 99.49 99.17
Pneumonia 4,273 86.80 100 92.93
Taspinar et al. [22] VGG19 + stacking model 96.86 COVID-19 800 99.46 98.80 99.13
Normal 1,341 96.06 96.27 96.16
Pneumonia 1,345 96.21 96.36 96.29
Apostolopoulos and mpesiana [33] Distortion fixing MobileNetV2 96.78 COVID-19 224 98.66 83.71 90.57
Normal 700 94.26 94.26 96.07
Pneumonia 504 93.65 96.13 94.87
Luz et al. [7] Data augmentation (rotation, zoom, and flipping) EfficientNet 93.94 COVID-19 183 96.77 99.99 98.36
Normal 8,066 94.00 93.07 93.53
Pneumonia 5,538 93.00 93.00 93.00
Jain et al. [34] Data augmentation (rotation, zoom, and shearing) Xception 97.97 COVID-19 576 98.75 91.86 95.18
Normal 1,583 97.79 92.86 95.26
Pneumonia 4,273 96.51 99.22 97.85
  1. GAN, generative adversarial networks; HE, histogram equalization; Adam, adaptive moment estimation; Bo, bayesian optimization; SGDM, stochastic gradient descent with momentum.

This paper presents a multi-case (COVID-19, pneumonia, and normal) identification evaluation for various AHE parameters with popular ConvNet architectures. Furthermore, we compared the effect of reliable AHE parameters for possible clinical relevance instead of raw chest X-rays. The key contributions of the proposed system are as follows:

  1. Whereas most researchers analyzed a small number of COVID-19-positive cases, a disadvantage of ConvNet for unbalanced data, our proposal utilized one of the latest datasets with 3615 COVID-19-positive cases.

  2. Rather than using raw chest X-rays, the proposal fed the ConvNet architectures using enhanced images with a more significant and responsible pathology visualization for a clinical relevance

  3. The paper performed a complete comparison to state the impact of various AHE parameters with each ConvNet for creating a baseline for future chest X-ray strategies

  4. Multi-case identification performances were evaluated separately for chest X-rays using fine-tuning on the best performance to reduce the training time of ConvNets.

  5. Transfer learning on the weights of ConvNets trained small-scale datasets for large-scale COVID-19 dataset

The remaining paper gives detailed information about the chest X-ray database, AHE, ConvNet architectures, feature learning, and fine-tuning in Section 2. The structure of the proposed model, iteration parameters, AHE variations, experimental setup, and achievements are shared in Section 3. The efficiency of ConvNets in identifying COVID-19, superiority, and limited aspects of the proposed models are handled in the last section.

Materials and methods

Chest X-ray database

Chest X-rays, one of the most common non-invasive diagnostics for respiratory and cardiac diseases, enable visualization of internal organs, tissues, and pathology using low-dose radiation depending on ray permeability.

We used two large-scale chest X-ray datasets, including the NIH (National Institutes of Health Clinical Center) Chest X-Ray database (www.kaggle.com/nih-chest-xrays/data), Pneumonia Chest X-Ray Images (www.kaggle.com/paultimothymooney/chest-Xray-pneumonia), and COVID-19 Radiography database (www.kaggle.com/tawsifurrahman/covid19-radiography-database) in the experiments.

The NIH chest X-ray dataset comprises over 112 K chest X-rays from 30,805 patients with no findings and 14 pulmonary diseases, including pneumonia and more.

To check the performance of ConvNets on identifying diseases with similar symptoms on chest X-rays, we excluded the remaining pathological conditions except for bacterial pneumonia and healthy. Due to the similarity of pneumonia (bacterial) symptoms and COVID-19 (viral pneumonia), the researchers frequently study multi-case identification. Chest X-Ray Images (Pneumonia) dataset is comprised of over 5.8 K chest X-rays with pneumonia and no-finding. The chest X-rays with COVID-19 were gained from the COVIDx dataset, the largest open-access dataset regarding the number of COVID-19-positive cases. It has 3,616 chest X-rays with COVID-19 positive and continues to be updated due to ongoing cases worldwide. It was incorporated as part of the Italian Society of Medical, Interventional Radiology (SIRM) COVID-19 database, Novel Corona Virus 2019 Dataset, and resources of many publications.

The chest X-rays have a wide variation in specifications, including resolution and depth. Therefore, each image needs to be resized into a standardized form. However, the chest X-ray with pneumonia and no finding has a rectangular form (e.g.: 1900 × 1400), whereas COVID-19 cases have a square form (299 × 299). In this manner, resizing the rectangular chest X-ray into a square form causes loss in the anatomical shape of the lungs by stretching. Therefore, we applied zero paddings to the chest X-rays to avoid distortion after resizing with a square form along with keeping the anterior shape.

We planned three strategies to evaluate the impact of AHE:

Strategy 1: Training/testing ConvNet architectures on a small-scale fold (428 COVID-19, 500 pneumonia, and 500 healthy) using AHE to compare the performances of architectures.

Strategy 2: Testing the best models in Strategy 1 using a discrete large-scale testing fold (1500 COVID-19, 1500 pneumonia, and 1500 healthy) to evaluate the generalization performance of ConvNet architectures with AHE.

Strategy 3: Re-Training/testing the ConvNet architectures on a large-scale dataset (3615 COVID-19, 3500 pneumonia, 3500 normal) using transfer learning on the own weights in Strategy 1 to evaluate the effect of small-scale datasets for continual transfer learning with AHE.

Adaptive histogram equalization on chest X-rays

AHE is a prosperous image contrast enhancement technique that provides more precise presentations for various image types. Capabilities of contrast enhancement and noise reduction for medical images are the most frequent preference reasons for using AHE [23]. It enables reproducible enhancing presentation with many contrasts in chest X-rays using interactive contrast density windows and cumulative distribution of pixels [24]. It has a common use in medical images to improve the visibility of the patterns and support the diagnosis due to ensuring detailed presentations for the region surrounding the black pixels [25]. The complexity of AHE is O(2 (m + i)) for an n-by-n medical image with an intensity level of i and m-by-m contextual region size [23], 25].

AHE has two parameters the grid size, which divides the chest X-rays into contextual tiles, and the clip limit, which is a thresholding value for contrast limiting to prevent over-saturation in homogeneous subsections [25]. Herein, we experimented with a brute force algorithm using AHE parameters at the ranges of [0.1–1] by increasing the size of 0.1 for clip limit and 1/2, 1/4, 1/7, 1/8 [23], 1/16 [25], 1/28, 1/56 of input chest X-rays for grid sizes on each pre-trained architecture [32]. We shared the classification performances for only AHE parameters with the highest achievements for each pre-trained architecture in ascending order according to the highest accuracy.

Chest X-rays have a large amount of information belonging to many viscera, including the heart, diaphragm, lungs, hilar arteries, clavicle, and rib cage. As shown in Figure 1, the histogram plot for raw chest X-ray and AHE application present different densities. In comparison, local image enhancement techniques improve the presentation by equally analyzing pixel intensity over the possible grayscale; the variety of density in viscera necessitates regional adaptive pixel intensity proportional [25]. Hence, using a whole chest X-ray with AHE implements a density-based enhancement referring to global enhancement techniques, which are incompetent for medical images by maintaining in restricted size of sub-regions [24]. Raw chest X-ray, depicted in Figure 1a, is composed of similar grayscale density due to the similar X-ray permeability of the tissues and noise arising out of device specifications (see Figure 1b). On the other hand, chest X-ray with AHE enables the pixel density to reach a more uniform distribution with local enhancement methods, as is seen in Figure 1d. This density-based enhancement provides a sharper acquisition of pathological regions and airways, minor alterations (see Figure 1c), as well as the detection of distinctive features for machine learning.

Figure 1: 
A random chest X-ray with COVID-19 (a), histogram plot for image (a) (b), AHE with clip limit of 0.4 and grid size of 112px (1/2) on image (a) (c), histogram plot for image (c) (d).
Figure 1:

A random chest X-ray with COVID-19 (a), histogram plot for image (a) (b), AHE with clip limit of 0.4 and grid size of 112px (1/2) on image (a) (c), histogram plot for image (c) (d).

Transfer learning through ConvNets

ConvNet is one of the most potent and trending DL algorithms for classifying various images, even for time series. The most representative specifications of ConvNet are feature learning and transfer learning approaches. Although ConvNet fine-tunes the pre-trained shared weights, it still has many classification parameters to perform a global optimization depending on the depth of hidden layers, feature learning stages, and the nodes at each dense layer [2]. This case brings about wasting a long time in training for a robust ConvNet architecture. Even though using pre-defined classification parameters reduces the training time depending on the statistical and probabilistic similarity of the features, it is still the prior focus of the researchers to shorten training and prediction time to handle big data problems [2].

ConvNet architectures are comprised of sequential layers/blocks (including convolutional layer (Conv), ReLU (rectified linear unit), pooling, dense, dropout, and batch normalization layers) in different orders. Conv layer is the principal procedure of ConvNet to extract a set of significant and dominant pixel sets concerning the filter. ConvNet applies a set of convolutional kernels at each Conv. The convolution function generates a novel presentation of the subsection on the chest X-ray in a sequential manner using filter kernels at each separate layer. The convolution function is as follows:

(1) f c k = d x 1 , y 1 j d ( x 1 , y 1 ) × i c k

where j d (x 1,y 1) and i c k are a subsection of the chest X-ray and index of the kth filter kernel at the cth layer, respectively.

The pooling layer is a down-sampling procedure applied after Conv to maintain the relevancy intact of feature maps. It stacks the corresponding values within the receptive regions considering the pooling operation (max, min, and average). The pooling layer (Pool max ) outputs a down-sampled view of feature maps using the pooled feature map of the kernel-based feature maps and the type of pooling operation (see Eq. (2)).

(2) Pool max = max ( f c k )

The dropout regularization layer randomly eliminates the nodes at the previous layer within a given rate to avoid ConvNet architecture from overfitting in training. The batch normalization layer performs re-normalization considering the output of feature maps. Eq. (3) and (4) stand for average and variance for cth mini-batch of feature activation map:

(3) µ x c = 1 m i = 1 m x i c

(4) σ x c 2 = 1 m i = 1 m ( x i c µ x c ) 2

(5) BN γ , β ( x c ˆ ) = γ c x c μ x c σ x c 2 + β c

Eq. (5) indicates batch normalization (BN) for x c , that is cth mini-batch of feature activation map. γ and β are learnable parameters during training. The dense layer stands for fully connected layers, the series of ConvNet for supervised learning with softmax output function. Mathematical definition of softmax(z i ) is indicated in Eq. (6) where z i stands for ith the output of a fully connected layer.

(6) s o f t m a x ( z i ) = e z i j = 1 n c l a s s e z j

Transfer learning is a novel procedure that reuses the knowledge obtained from a ConvNet architecture to solve a related task by re-training the pre-trained shared weights on new tasks. Transfer learning is usually implemented in various strategies. It is used to transfer the feature learning strategy into novel ConvNet architectures [26]. The ConvNet model is utilized as a feature extractor. The extracted features are fed into new neural network models or related machine learning algorithms.

It is used to train small-scale data with the capabilities of large-scale data in the cases of fewer instances for training. Lastly, it is used to fine-tune the ConvNet weights for optimal generalization and circumvent computational costs [26]. ConvNet architecture is adapted or refined using adjustments, pruning, and parameter tuning during the re-training of the architecture. In this way, achieving high performance depends on training on a comparatively large-scale dataset.

The statistical test characteristics (Eq. (7)(10)), including accuracy, precision, recall, and f1 score, were calculated to evaluate the performance of pre-trained architectures [2].

(7) A c c u r a c y = T N + T P A l l s u b j e c t s

(8) P r e c i s i o n = T P T P + F P

(9) R e c a l l = T P T P + F N

(10) F 1 S c o r e = 2 * P r e c i s i o n * R e c a l l P r e c i s i o n + R e c a l l

where TP is the number of correctly predicted classes on chest X-ray, and TN presents the number of correctly predicted other classes. Whereas FP is the number of falsely predicted classes, FN offers the number of missing other classes in the testing set.

We re-trained four popular ConvNet architectures, including AlexNet, MobileNet, DarkNet19, and VGG16 feeding chest X-rays for various AHE parameters.

  1. MobileNet is a lightweight ConvNet architecture that applies depth-wise separability for each channel of input images for embedded systems. It was trained on the large-scale ImageNet dataset. It has 23 layers (22 Conv layers and a fully connected layer) [27].

  2. DarkNet19 is a fast and thriving comparatively lightweight ConvNet architecture that adapted the backbone of YOLOv2. It is similar to VGG architectures. However, it uses average pooling instead of max pooling. It has 20 layers (19 Conv layers and a fully connected layer) [6].

  3. AlexNet is a pioneer ConvNet architecture in terms of GPU implementation on the large-scale ImageNet dataset. It has a simple architecture with eight layers (five Conv layers and three fully-connected layers) [3].

  4. VGG16 is a deep ConvNet architecture that does not presume on residual designing. It was trained on the large-scale ImageNet dataset. It has a lightweight architecture with sequential Conv blocks and shallow depth layers (13 Conv layers and three fully connected layers) [4].

Each of them is an efficient ConvNets with lightweight architectures. The essential characteristics of the experimented ConvNets are to be equipped with low- and high-level feature learning capabilities, to have high generalization in fully-connected layers with optimization and regularization methods, to handle the various presentations in Conv blocks, and to have the ability to reduce training time by transfer learning on pre-trained weights without many iterations.

One of the novelties of the work is adapting ConvNets on small-scale datasets into transfer learning of large-scale datasets for stating the advantages of analyzing big data using faster strategies in DL. To evaluate the main benefits of transfer learning on ConvNet architectures, we experimented with the same variations on the dataset, the same strategies in transfer learning, and the same parameters in fine-tuning on ConvNets.

Results

Computer-aided identification of rare pathology and incipient abnormalities is of great importance for COVID-19. Furthermore, it is possible to reveal ConvNet models using the density of pathologies, obstructions, and fluid accumulation resulting from bacterial- and viral pulmonary diseases. Therefore, minimizing the adherence to clinical specialists for the robust and practical identification of pandemic diseases is the prime necessity to increase the prevalence. The structure of the proposed strategies is indicated in Figure 2. It depicts the sequence of the proposed three strategies for transfer learning. Strategy 1 is initially attaining the pre-trained weights for ConvNet architectures by training with the small-scale dataset to pave for other strategies with related data training. Strategy 2 is the test of ConvNet architectures on Strategy 1 using the large-scale dataset. Strategy 3 performs transfer learning on pre-trained weights from Strategy 1 on the large-scale dataset. Consequently, the proposed model visualizes the most responsible regions for a better understanding with clinical relevance for multi-cases on chest X-rays.

Figure 2: 
The structure of the proposed model. Strategy 1 (transfer learning on small-scale dataset). Strategy 2 (testing strategy 1 using the large-scale dataset). Strategy 3 (transfer learning on the large-scale dataset).
Figure 2:

The structure of the proposed model. Strategy 1 (transfer learning on small-scale dataset). Strategy 2 (testing strategy 1 using the large-scale dataset). Strategy 3 (transfer learning on the large-scale dataset).

We evaluated the impact of AHE with popular ConvNet architectures to identify chest X-rays with COVID-19, pneumonia, and healthy. We applied AHE for various parameters as a preprocessing stage to designate the most responsible vision on chest X-rays depending on each ConvNet architecture. Although ConvNet architectures have high performances for identifying COVID-19 using raw chest X-rays, we focused on ensuring intact lung anatomy by padding and enhancing airway visibility to support the clinical relevance and validity of the proposals.

We fined-tuned the pre-trained ImageNet weights on both ConvNet architectures using Keras implementation. We trained each ConvNet architecture (Strategy 1) using the cross-entropy loss function, a learning rate of 0.00001, a patience period of 5 epochs, and a batch size of 16 images for 100 epochs. We selected the adaptive moment estimation (Adam) optimizer to adaptively decrease the learning rate when learning stagnates in a period of epochs during training [31]. The cross-entropy loss function, which approaches minimizing the probabilistic differences between predicted and actual labels, is adapted to the fine-tuning procedures. Finally, we utilized weight decay, validation set (20 % of the training set), and L1 regularization in training to avoid overfitting.

The statistical test characteristics were calculated using a weighted average to evaluate the class-based efficiency of the experimented AHE parameters with each Strategy on ConvNet architectures [2]. The results were averaged to produce the performance of ConvNet architectures. Furthermore, we visualized the most responsible sections on chest X-rays for each ConvNet using gradient-based localization techniques on feature maps.

The highest classification performances for AHE on complete chest X-ray analysis were achieved using DarkNet19 architecture with rates of 98.26 , 98.33, 98.26, and 98.25 % for accuracy, precision, recall, and f1 score, respectively. The most responsible AHE parameters are 1/28 and 0.6 for AHE grid size and clip limit, respectively. Table 2 presents classification performances and AHE parameters for ConvNets. Applying AHE in various parameters has outperformed models without AHE for both ConvNet architectures. The highest performance improvement in terms of overall accuracy is reported using 1/8 of grid size and 0.1 of clip limit on AHE using MobileNet architecture for Strategy 1 by 84.62–95.46 %.

Table 2:

The COVID-19 identification performances (%) on strategy 1 for the best three AHE parameters and without AHE. The best classification performances in terms of overall accuracy were highlighted in bold for each ConvNet architecture.

Architecture AHE disk AHE clip limit Accuracy Precision Recall f1 score
AlexNet 1/56 0.1 96.50 96.55 96.48 96.50
1/4 0.8 96.51 96.56 96.50 96.49
1/8 0.5 96.85 96.86 96.79 96.84
Without AHE 93.36 93.33 93.37 93.34
DarkNet19 1/56 0.4 97.90 97.95 97.92 97.88
1/28 0.1 98.25 98.26 98.21 98.24
1/28 0.6 98.26 98.33 98.26 98.25
Without AHE 96.15 96.17 96.07 99.15
MobileNet 1/7 0.7 95.46 95.47 95.41 95.46
1/8 0.6 95.46 95.44 95.43 95.41
1/8 0.5 95.46 95.43 95.44 95.43
Without AHE 84.62 85.25 84.67 84.31
VGG16 1/28 0.2 97.20 97.33 97.11 97.17
1/8 0.1 97.20 97.22 97.17 97.19
1/56 0.1 97.55 97.56 97.58 97.54
Without AHE 94.41 94.45 94.46 94.35

We tested the trained ConvNet architectures in Strategy 1 using a discrete large-scale testing fold to evaluate the generalization performances. The highest large-scale testing identification performances for AHE were achieved also using DarkNet19 architecture with rates of 71.33 , 76.81, 72.54, and 74.55 % for accuracy, precision, recall, and f1 score, respectively. The most responsible AHE parameters are 1/16 and 0.7 for AHE grid size and clip limit on complete chest X-rays, respectively. Table 3 presents the testing performances with the large-scale dataset for the best ConvNet architectures. Whereas DarkNet19 has the highest generalization capability with an accuracy rate of 71.33 %, the remaining ConvNet architectures predicted unsuccessfully for training with small-scale datasets. Especially, MobileNet is the weakest architecture for Strategy 2. This case indicates the disadvantages and the lack of clinical validity for related research.

Table 3:

The COVID-19 identification performances (%) on strategy 2 for the best ConvNet architectures. The tested large-scale chest X-rays do not include in the training of strategy 1. The highest testing performances were highlighted in bold for each statistical test metric.

Architecture AHE disk AHE clip limit Accuracy Precision Recall F1 score
AlexNet 1/28 0.1 47.83 91.87 47.82 58.07
1/2 0.7 50.84 90.60 50.78 60.96
1/8 0.5 56.50 91.26 56.55 66.22
DarkNet19 1/56 0.1 66.07 72.94 66.11 68.56
1/56 1 67.08 79.56 67.23 69.94
1/16 0.7 71.33 76.81 72.54 74.55
MobileNet 1/56 0.6 41.39 85.04 41.24 49.75
1/28 0.2 35.63 85.98 35.67 42.69
1/28 0.5 36.62 86.97 36.55 44.05
VGG16 1/8 1 48.36 91.35 48.32 58.21
Without AHE 44.88 89.49 44.81 54.67
1/56 0.1 48.21 87.71 48.27 59.08

Transfer learning was applied to the weights of the best five ConvNets on Strategy 1 using a large-scale database. The highest classification performances for Strategy 3 were achieved also using VGG16 architecture with rates of 95.04 , 95.11, 94.99, and 95.05 % for accuracy, precision, recall, and f1 score, respectively. The most responsible AHE parameters for VGG16 are 1/8 and 0.1 for AHE grid size and clip limit, respectively. Table 4 presents classification performances and the best AHE parameters for four ConvNets. Applying transfer learning on the weights of the small-scale COVID-19 dataset reached high classification performances for DarkNet19, VGG16, and AlexNet architectures. MobileNet reached a low generalization capability considering the classification performance on the small-scale dataset. Whereas the patience parameter for early stopping was set to 5 epochs, the training of Strategy 3 was halted at average epoch sizes of 18, 12, and 16 for AlexNet, DarkNet, and VGG16, respectively. Besides achieving the lowest classification performances for transfer learning on Strategy 3, MobileNet passed over the patience parameter by finalizing the training at 100 epochs.

Table 4:

The COVID-19 identification performances (%) on strategy 3 by transfer learning with the large-scale dataset on the weights of the best three ConvNets in strategy 1. The highest five testing performances were highlighted in bold for both ConvNet architectures.

Architecture AHE disk AHE clip limit Accuracy Precision Recall F1 score
AlexNet 1/8 0.5 91.27 91.67 91.23 91.45
1/56 0.1 91.74 92.03 91.33 91.68
1/4 0.8 92.43 92.60 92.24 92.42
DarkNet19 1/2 1 88.32 90.32 86.58 88.41
1/28 0.6 90.49 91.33 89.04 90.17
1/56 0.4 93.22 93.69 92.81 93.25
MobileNet 1/8 0.6 74.10 74.02 74.08 74.05
1/8 0.5 74.41 74.58 74.22 74.40
1/7 0.7 75.89 75.65 75.73 75.69
VGG16 1/28 0.2 93.63 93.76 93.54 93.65
1/4 0.2 94.98 94.98 94.92 94.95
1/8 0.1 95.04 95.11 94.99 95.05

It would be easy to understand the impact and responsibility of ConvNet architectures on different Strategies using the same AHE parameters. Therefore, we presented Table 5 that includes a cross check for the achievements with the same AHE parameters of the best pre-trained architectures on each strategy. Although the DarkNet19 architecture does not give the best results for each strategy on the same AHE parameters, it is clearly seen that it does not cause very large fluctuations on performance rates. The VGG16 architecture, on the other hand, showed inconstant rates up to almost 2x of overall accuracy. This situation shows that the VGG16 is quite ineffective compared to DarkNet19 in COVID-19 identification without fine-tuning on new data.

Table 5:

The cross check on the COVID-19 identification performances (%) using the same AHE parameters for the best ConvNet architectures on each strategy.



Architectures
AHE disk AHE clip limit Strategy Accuracy Precision Recall f1 score
DarkNet19 1/28 0.6 #1 98.26 98.33 98.26 98.25
#2 67.09 79.60 67.90 69.94
#3 90.49 91.34 90.48 90.17
DarkNet19 1/16 0.7 #1 97.20 97.28 97.21 97.21
#2 71.33 76.81 72.54 74.55
#3 84.31 88.26 84.31 84.68
VGG16 1/8 0.1 #1 96.50 96.59 96.51 96.45
#2 48.22 87.71 48.21 59.08
#3 95.04 95.11 94.99 95.05

DarkNet19 reached the highest overall accuracy rates for Strategies 1 and 2. The best identification rates are achieved with rates of 100 and 99.10 % for pneumonia and normal chest X-ray using DarkNet19 on Strategy 1, respectively. Although Strategy 2 has low performance in terms of overall accuracy, DarkNet19 is also the best architecture and reaches a pneumonia identification accuracy of 96.00 %. VGG16 has the highest capacity for transfer learning on Strategy 3. Figure 3 depicts the confusion matrices for the best ConvNet architectures. The confusion matrix for DarkNet19 with the small-scale dataset on Strategy 2 has been plotted for 1,428 chest X-rays with homogeneously distributed multi-classes on the left of Figure 3. The confusion matrix of the highly responsible VGG16 architecture for Strategy 3 on the large-scale dataset has been plotted for 3,128 chest X-rays on the right of Figure 3. The receiver operating characteristic (ROC) and learning curves (loss-epoch) have been implemented in Figure 4 for a better understanding of the training performance of the models. We included the ROC and learning curves of ConvNet architectures with the best achievements for Strategies 1 and 3. VGG16 in Strategy 3 has a better and faster-converging capability than DarkNet19 in Strategy 1 in the training of the COVID-19 identification model on enhanced chest X-rays. Areas under ROC curves, the aggregate measure of model performance, are 94.98 and 95.87 % for Strategy 3 and Strategy 1, respectively. The learning curve measures the correlation between training losses and the epochs for ConvNet architectures. Learning curves indicates that training DarkNet19 in Strategy 1 and VGG16 in Strategy 3 (without early stopping) have similar convergence by epoch. Strategy 1 can reach a lower loss by epoch on training. In contrast, Strategy 1 is prone to stop training with sufficiently minimum loss at early epochs for a small patience value but reaches a slightly higher loss than Strategy 3.

Figure 3: 
The confusion matrices for the best ConvNet architectures, DarkNet19 on strategies 1 and 2 (Left); VGG16 on strategy 3 (Right).
Figure 3:

The confusion matrices for the best ConvNet architectures, DarkNet19 on strategies 1 and 2 (Left); VGG16 on strategy 3 (Right).

Figure 4: 
The receiver operating characteristic (ROC) and learning (loss-epoch) curves for the ConvNet architectures with the highest achievements in strategies 1 and 3.
Figure 4:

The receiver operating characteristic (ROC) and learning (loss-epoch) curves for the ConvNet architectures with the highest achievements in strategies 1 and 3.

Discussion

As the amount of COVID-19 cases continues to increase with rapid acceleration, developing robust diagnostics with clinical validity is crucial to edge over the progression of the pandemic. In spite of the fact that the RT-PCR test is the most preferred diagnostic tool, its precision is also limited [28]. Although it gives definite results in identifying COVID-19-positive cases, definitive non-COVID-19 cases need to be evaluated at the various clinical stages. Hence, it is substantial to develop alternative diagnostic and visualization techniques to procure the analytical validity and testing for agreement. The majority of COVID-19 identification research on chest X-ray analysis has focused on obtaining a high generalization capability using small-scale datasets due to the scantiness in quantity. The chest X-ray analysis commonly focused on feeding ConvNet architectures with small-scale datasets due to the lack of chest X-rays with COVID-19. Table presents a complete comparison with related works on identifying multi-cases (COVID-19, pneumonia, and normal) in terms of databases, ConvNet architectures, preprocessing, number of subjects in the analysis, and classification performances.

Many researchers analyzed chest X-rays for binary case classification (COVID-19 vs. non-COVID-19 and COVID-19 vs. normal). Mukherjee et al. proposed a tailored ConvNet architecture for distinguishing two cases (COVID-19 – non-COVID-19). Their proposal outperformed popular architectures with rates of 99.69 , 100, 99.38, and 99.69 % for accuracy, recall, precision, and f1 score, respectively [15]. Minaee et al. re-trained ResNet18, ResNet50, SqueezeNet, and DenseNet121 architectures with adaptive optimization techniques by transfer learning on ImageNet weights. They reached identification performance rates of 92.28 , 92.10, 99.93, and 95.85 % for accuracy, recall, precision, and f1 score using a small-scale dataset for identifying COVID-19 and non-COVID-19 [16]. He et al. used a ConvNet to separate COVID-19 and non-COVID-19 cases. In training, they experimented with various threshold values with a stochastic gradient descent algorithm. They achieved high classification performance rates of 87.98 , 90.00, and 87.84 % for accuracy, recall, and specificity, respectively [9]. Haghanifar et al. also tried to train DenseNet121, which is the feature learning backbone of ChexNet architecture, for binary classification (COVID-19, normal). They distinguished COVID-19 cases from normal chest X-rays with performance rates of 98.68 , 99.34, 99.17, and 99.25 % for accuracy, precision, recall, and f1 score, respectively. However, they reported the inefficacy of models for the localization of reasonable lung region feature maps [21]. Ozturk et al. also proposed a binary case classification (COVID-19 vs. normal) model using DarkNet19 architecture. Although applying none of preprocessing except resizing, they iterated on various numbers of filters for the feature learning stage of ConvNet architecture. They separated COVID-19 cases from healthy chest X-rays with identification performance rates of 98.08 , 98.03, 95.13, 95.30, and 96.51 % for accuracy, precision, recall, specificity, and f1 score, respectively [6]. Ismael and Sengur applied local texture descriptors to classify chest X-rays (180 COVID-19 and 200 normal) using transfer learning on popular ConvNet architectures, including, ResNet18, ResNet50, ResNet101, VGG16, and VGG19. They also adapted multiple conventional machine learning algorithms to evaluate the impact of feature learning stages. They reached the highest identification achievements using ResNet50 with linear support vector machines function with rates of 94.74 , 91.00, 98.89, and 94.79 % for accuracy, recall, sensitivity, and f1 score, respectively [29]. Nayak et al. also performed a binary classification on chest X-rays (203 COVID-19 and 203 normal) to compare the impact of ConvNet architectures, including VGG16, InceptionV3, ResNet34, MobileNetV2, AlexNet, GoogleNet, ResNet50, and SqueezeNet. They augmented the dataset using scaling, flipping, rotation, and adding Gaussian noise. They produced the best achievements using ResNet34 with rates of 98.33 , 96.77, 96.67, and 98.36 % for accuracy, precision, specificity, and f1 score, respectively [5]. Taspinar et al. compared various machine learning algorithms by feeding ConvNet-based feature maps and highlighted the efficiency of feature learning on VGG19 with the stacking classifier. They reached the best achievements using the stacking classifier with rates of 96.86 , 96.88, and 96.89 % for accuracy, recall, and f1 score, respectively [22].

The researchers analyzed the chest X-rays for multi-case classification presented in Table 1 with a detailed comparison in terms of preprocessing, ConvNet architectures, additional optimization algorithms, and case-specific classification performances. Although the demonstrated achievements are relatively prosperous, the main deficiency of the related papers is that the analyses included a tiny amount of COVID-19 cases in common. Furthermore, using unbalanced datasets fails COVID-19 identification and learning COVID-19 characteristics on chest X-rays. Thanks to the recent updates on COVID-19 databases, our proposal on Strategy 3 is based on one of the foremost and most comprehensive chest X-ray analyses regarding the amount of COVID-19 positive cases. Hence, transfer learning on the weights of ConvNet architecture trained with the related small-scale dataset reached a high distinguishing capability for COVID-19 cases, just as providing precipitated optimization on detailed ConvNet architectures.

Even if the most relevant papers in preprocessing with AHE have been conducted on small-scale datasets within COVID-19 cases, they highlighted the effect of using enhanced lung pathology to support the clinical validity of results on the various ConvNet architectures [14], 17], 18], 21]. Singh et al. utilized generative adversarial networks on AHE representations to generate more visual lung patterns as a preprocessing. They produced the feature learning stage of ConvNet architectures, including VGG19, VGG16, DenseNet121, DenseNet169, and ResNet50. They reported the efficiency of feature maps in VGG19 with pruned Naive classifier with rates of 98.67 and 98 % for overall accuracy and kappa score, respectively [17]. However, using a deep generative model as segmentation and chest X-ray enhancement technique unveils additional workload in COVID-19 identification. Moreover, deep generative models have the possibility of constituting non-existing lung pathology. Hence, the clinical relevance of their proposal is still limited.

On the other hand, Oh et al. used a two-step ConvNet architecture to segment the lungs and classify multi-cases on chest X-rays. They excluded the non-lung sections from the analysis using DenseNet103 architecture and afterwards fed the ROIs of the lung into the ResNet18 architecture. They classified multi-cases with performance rates of 88.90 , 83.40, 85.90, and 84.40 % for accuracy, precision, recall, and f1 score, respectively [18]. Similarly, Haghanifar et al. utilized U-Net architecture for lung segmentation and afterwards applied AHE for default parameters to the segmented lung areas. They reached an identification accuracy rate of 87.21 % by feeding the segmented lungs into DenseNet121 architecture [21]. However, using deep ConvNet models as lung segmentation is a disadvantage for wasting training time. Moreover, the achievements are in the moderate performance ranges against our achievements for a majority of strategies. El Asnaoui and Chawki used AHE to generate more responsible chest X-rays in an approach similar to ours by image enhancement. They performed transfer learning on the VGG16, VGG19, DenseNet201, Inception ResNetV2, InceptionV3, Resnet50, and MobileNetV2 architectures. They reached multi-case identification performance using Inception ResNetV2 and L2 regularization with the rates of 92.18 , 92.19, 92.32, and 92.08 % for accuracy, recall, precision, and f1 score, respectively [14]. However, the paper doesn’t explain the most responsible AHE parameters and experimented ranges for COVID-19 identification. In particular, default AHE parameters for grid size and clip limit factor have limitations for representing true clinical evaluations considering the experiments on various ConvNet architectures. Whereas the state-of-the-art has disadvantages of clinical relevance and clinical validity due to a limited range of parameters in image enhancement and ConvNet architectures with a high number of layers, our proposal experimented with a wide range of AHE parameters with shallow ConvNet architectures. Whereas AHE with 1/56 and 1/16 of grid size presented the most responsible feature maps for DarkNet19, 1/8 of grid size for AHE has the most characteristic features for VGG16 architecture.

The highest classification performances for the large-scale COVID-19 dataset within Strategy 3 were achieved using VGG16 architecture with rates of 95.04 , 95.11, 94.99, and 95.05 % for accuracy, precision, recall, and f1 score, respectively. The most responsible AHE parameters are 1/8 and 0.1 for AHE grid size and clip limit, respectively (see Table 6).

Table 6:

The highest DeepCOVIDNet-CXR model achievements for overall accuracy and case-based classification performances (%) for COVID-19, normal, and pneumonia.

Pre-processing Architecture Accuracy Case Subjects Precision Recall F1 score
DeepCOVIDNet-CXR AHE (1/56, 0.1) DarkNet19 + adam optimizer 98.26 COVID-19 428 100 94.11 96.97
Normal 500 95.28 100 97.58
Pneumonia 500 100 100 100
AHE (1/8, 0.1) VGG16 + adam optimizer 95.04 COVID-19 3,615 95.55 92.36 93.93
Normal 3,500 91.25 95.25 93.21
Pneumonia 3,500 98.58 97.47 98.02

Gradient-weighted class activation mapping (Grad-CAM) is a visualisation technique on CNN architectures to identify the most responsible feature activations with a heat map [30]. We customized Grad-CAM by setting a threshold rate for learned highly responsible features. Whereas Grad-CAM generates a heat map to visualize the level of responsibility for activation maps, we excluded the learned activation maps less than the threshold ratio and a single colour presentation to provide a clearer view of the highest responsibility for COVID-19. Figure 5 depicts the customized approach on Grad-CAM for random chest X-rays with COVID-19. The yellow sections are the highly responsible sections with a threshold rate of 0.75 for each Strategy. Most dispersion on chest X-rays mainly occurs on the lungs, even using no lung segmentation in the proposal. Results indicate the highly responsible feature activation map for COVID-19 for Strategy 1, Strategy 2, and Strategy 3. Strategy 1 highlighted the most responsible feature activation maps for a local region and non-lung (heart) region, as is seen in Figure 5a. Strategy 2 has similar achievements with a local pathological region and non-lung (armpit) section in Figure 5b. The highly responsible feature maps for identifying COVID-19 on Strategy 3 are mostly located in the lung regions for extended pathologies on sample chest X-rays with clinical relevance in Figure 5c.

Figure 5: 
The highly responsible sections with Grad-CAM for strategy 1 (a), strategy 2 (b), and strategy 3 (c).
Figure 5:

The highly responsible sections with Grad-CAM for strategy 1 (a), strategy 2 (b), and strategy 3 (c).

The necessity of clinical validation on predicted pathology constitutes the study’s weakness. Hence, it is more appropriate to use chest X-rays as a guiding evaluation or supportive visualization technique rather than an early diagnosis tool for COVID-19 due to the variety of medical devices with different specifications and patient-specific cases. Computerized techniques need a more detailed COVID-19 database with annotations and localization of pathological regions caused by COVID-19 for the clinical applicability of the proposals to overcome this limitation. An experienced clinician evaluated the Grad-CAM of the proposal for responsible lung regions with COVID-19. According to the clinician’s evaluation, although the proposal within Strategy 3 is relatively successful at identifying severe COVID-19 pathology, it still has limitations in locating mild-level abnormalities. Nevertheless, highly responsible feature maps can be located in the non-lung and cardiac regions. This assessment clearly shows that the ConvNet architectures need to be supported with larger numbers of COVID-19 cases before gaining clinical use. Generally, the most responsible case activation maps are on lung regions for the best proposals on each Strategy.

Shallow ConvNet architectures have the ability to perform accurate predictions with the advantages of chest X-ray enhancements in preprocessing and straight parameter optimization techniques in training. The lightweight concepts of shallow ConvNets gain an edge in the pre-definition of robust models using transfer learning. That is why we used the weights of Strategy 1 for Strategy 3 on a large-scale relevant dataset. It enables re-training the novel COVID-19 cases using shared weights. In accordance with the highest classification performances for four ConvNet architectures, DarkNet19 achieved the highest multi-case identification performances for Strategies 1 and 2. VGG16 is the highest-performance model in Strategy 3. Whereas training time is a curse for deep ConvNets, Strategy 3 reduced training time with transfer learning on the model trained using a relevant small-scale dataset.

Conclusions

The main superiorities of this study are feeding the ConvNet architectures with enhanced representations of chest X-rays and COVID-19 pathology using various AHE parameters on the largest balanced COVID-19 dataset. The AHE enhancement on chest X-rays took advantage of accurate multi-case classification performances with local preprocessing, which exhibits reproducible presentation for lung patterns. The achievements prove that.

  1. AHE with ConvNets is a pioneering approach to designate multi-case models for distinguishing similar pathological metaplasia such as pneumonia and COVID-19.

  2. It is a pioneering transfer learning application that asserts a base and open-upgrade architecture for future COVID-19 cases for better generalization.

  3. The most responsible AHE parameters for future strategies on chest X-rays for identification of COVID-19 was designated

  4. The impact of the left lung, right lung, and complete chest X-rays for multi-case classification individually was revealed

  5. Transfer learning on the weights of ConvNets, which was trained using balanced small-scale COVID-19 datasets, was evaluated using extended balanced COVID-19 dataset for better generalization.

Therefore, DL-supported computer-aided techniques can significantly identify and visualize pathologies for diagnosing pneumonia and COVID-19 cases.


Corresponding author: Gokhan Altan, Computer Engineering Department, Iskenderun Technical University, Hatay, Türkiye, E-mail:

Acknowledgments

The authors express their thanks to Dr. Sinan INCE for providing his support in the clinical validation of COVID-19 pathologies.

  1. Research ethics: The database is an open access Chest X-ray database. Not applicable.

  2. Informed consent: The database is an open access Chest X-ray database. Informed consent was obtained from all individuals included in the database.

  3. Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.

  4. Use of Large Language Models, AI and Machine Learning Tools: None declared.

  5. Research funding: None declared.

  6. Conflict of interest: Authors state no conflict of interest.

  7. Data availability: Not applicable.

References

1. Hossein Zahraei, SA. Covid-19 (SARS-CoV-2) vs Sars-CoV; summary of all things that healthcare providers should know. Am J Biomed Sci Res 2020;8:376–81. https://doi.org/10.34297/ajbsr.2020.08.001305.Suche in Google Scholar

2. Altan, G. A deep learning architecture for identification of breast cancer on mammography by learning various representations of cancerous mass. In: Kose, U, Alzubi, J, editors. Deep Learning for Cancer Diagnosis [Internet]. Singapore: Springer Singapore; 2021:169–87 pp.10.1007/978-981-15-6321-8_10Suche in Google Scholar

3. Krizhevsky, A, Sutskever, I, Hinton, GE. ImageNet classification with deep convolutional neural networks. Commun ACM 2017;60:84–90. https://doi.org/10.1145/3065386.Suche in Google Scholar

4. Simonyan, K, Zisserman, A. Very deep convolutional networks for large-scale image recognition. In: 3rd Int Conf Learn Represent. San Diego, CA, USA: ICLR); 2014.Suche in Google Scholar

5. Nayak, SR, Nayak, DR, Sinha, U, Arora, V, Pachori, RB. Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: a comprehensive study. Biomed Signal Process Control 2021;64. https://doi.org/10.1016/j.bspc.2020.102365.Suche in Google Scholar PubMed PubMed Central

6. Ozturk, T, Talo, M, Yildirim, EA, Baloglu, UB, Yildirim, O, Rajendra Acharya, U. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput Biol Med 2020;121. https://doi.org/10.1016/j.compbiomed.2020.103792.Suche in Google Scholar PubMed PubMed Central

7. Luz, E, Silva, P, Silva, R, Silva, L, Guimarães, J, Miozzo, G, et al.. Towards an effective and efficient deep learning model for COVID-19 patterns detection in X-ray images. Res Biomed Eng 2022;38:149–62. https://doi.org/10.1007/s42600-021-00151-6.Suche in Google Scholar

8. Sethy, PK, Behera, SK, Ratha, PK, Biswas, P. Detection of coronavirus disease (COVID-19) based on deep features and support vector machine. Int J Math Eng Manag Sci 2020;5:643–51. https://doi.org/10.33889/ijmems.2020.5.4.052.Suche in Google Scholar

9. He, K, Zhang, X, Ren, S, Sun, J. Deep residual learning for image recognition. In: Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit (CVPR). Nevada, USA: Las Vegas; 2016.10.1109/CVPR.2016.90Suche in Google Scholar

10. Yoon, SH, Lee, KH, Kim, JY, Lee, YK, Ko, H, Kim, KH, et al.. Chest radiographic and CT findings of the 2019 novel coronavirus disease (COVID-19): analysis of nine patients treated in korea. Korean J Radiol 2020;21:494–500. https://doi.org/10.3348/kjr.2020.0132.Suche in Google Scholar PubMed PubMed Central

11. Zhang, J, Xie, Y, Pang, G., Liao, Z., Verjans, J., Li, W, Li, Y, Shen, C, Xia, Y.. Viral pneumonia screening on chest X-rays using confidence-aware anomaly detection. IEEE Trans Med Imaging 2020;40(3):879–90. https://doi.org/10.1109/TMI.2020.3040950.Suche in Google Scholar PubMed PubMed Central

12. Wang, L, Lin, ZQ, Wong, A. COVID-Net: a tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci Rep 2020;10:19549. https://doi.org/10.1038/s41598-020-76550-z.Suche in Google Scholar PubMed PubMed Central

13. Das, D, Santosh, KC, Pal, U. Truncated inception net: COVID-19 outbreak screening using chest X-rays. Phys Eng Sci Med 2020;43:915–25. https://doi.org/10.1007/s13246-020-00888-x.Suche in Google Scholar PubMed PubMed Central

14. El Asnaoui, K, Chawki, Y. Using X-ray images and deep learning for automated detection of coronavirus disease. J Biomol Struct Dyn 2021;39:3615–26. https://doi.org/10.1080/07391102.2020.1767212.Suche in Google Scholar PubMed PubMed Central

15. Mukherjee, H, Ghosh, S, Dhar, A, Obaidullah, SM, Santosh, KC, Roy, K. Shallow convolutional neural network for COVID-19 outbreak screening using chest X-rays. Cogn Comput 2021. https://doi.org/10.1007/s12559-020-09775-9.Suche in Google Scholar PubMed PubMed Central

16. Minaee, S, Kafieh, R, Sonka, M, Yazdani, S, Jamalipour Soufi, G. Deep-COVID: predicting COVID-19 from chest X-ray images using deep transfer learning. Med Image Anal 2020;65. https://doi.org/10.1016/j.media.2020.101794.Suche in Google Scholar PubMed PubMed Central

17. Singh, RK, Pandey, R, Babu, RN. COVIDScreen: explainable deep learning framework for differential diagnosis of COVID-19 using chest X-rays. Neural Comput Appl 2021;33:8871–92. https://doi.org/10.1007/s00521-020-05636-6.Suche in Google Scholar PubMed PubMed Central

18. Oh, Y, Park, S, Ye, JC. Deep learning COVID-19 features on CXR using limited training data sets. IEEE Trans Med Imag 2020;39:2688–700. https://doi.org/10.1109/tmi.2020.2993291.Suche in Google Scholar

19. Chowdhury, MEH, Rahman, T, Khandakar, A, Mazhar, R, Kadir, MA, Bin, MZ, et al.. Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 2020;8:132665–76. https://doi.org/10.1109/access.2020.3010287.Suche in Google Scholar

20. Rajpurkar, P, Irvin, J, Zhu, K, Yang, B, Mehta, H, Duan, T, et al.. CheXNet: radiologist-level pneumonia detection on chest X-rays with deep learning. ArXiv 2017. [Epub ahead of print]. https://doi.org/10.48550/arXiv.1711.05225.Suche in Google Scholar

21. Haghanifar, A, Majdabadi, MM, Choi, Y, Deivalakshmi, S, Ko, S. COVID-CXNet: detecting covid-19 in frontal chest x-ray images using deep learning. Multimed Tools Appl 2022;81:30615–45. https://doi.org/10.1007/s11042-022-12156-z.Suche in Google Scholar PubMed PubMed Central

22. Taspinar, YS, Cinar, I, Koklu, M. Classification by a stacking model using CNN features for COVID-19 infection diagnosis. J X Ray Sci Technol 2022;30:73–88. https://doi.org/10.3233/xst-211031.Suche in Google Scholar

23. Li, Y, Wang, W, Yu, D. Application of adaptive histogram equalization to x-ray chest images. In: Da-Heng, W, Consortini, A, Breckinridge, JB, editors. In second international conference on optoelectronic science and engineering. Beijing, China: SPIE Digital Library (2321); 1994.10.1117/12.182056Suche in Google Scholar

24. Sherrier, RH, Johnson, GA. Regionally adaptive histogram equalization of the chest. IEEE Trans Med Imag 1987;6:1–7. https://doi.org/10.1109/tmi.1987.4307791.Suche in Google Scholar

25. Pizer, SM, Johnston, RE, Ericksen, JP, Yankaskas, BC, Muller, KE. Contrast-limited adaptive histogram equalization: speed and effectiveness. IEEE Comput Soc 1990. https://doi.org/10.1109/VBC.1990.109340.Suche in Google Scholar

26. Weiss, K, Khoshgoftaar, TM, Wang, D. A survey of transfer learning. J Big Data 2016;3:9. https://doi.org/10.1186/s40537-016-0043-6.Suche in Google Scholar

27. Howard, AG, Zhu, M, Chen, B, Kalenichenko, D, Wang, W, Weyand, T, et al.. MobileNets: efficient convolutional neural networks for mobile vision applications. ArXiv 2017. [Epub ahead of print]. https://doi.org/10.48550/arXiv.1704.04861.Suche in Google Scholar

28. Pankhania, M. Artificial intelligence and radiology: combating the COVID-19 conundrum. Indian J Radiol Imag 2021;31(1):4–10. https://doi.org/10.4103/ijri.ijri_618_20. 33814755.Suche in Google Scholar PubMed

29. Ismael, AM, Şengür, A. Deep learning approaches for COVID-19 detection based on chest X-ray images. Expert Syst Appl 2021;164. https://doi.org/10.1016/j.eswa.2020.114054.Suche in Google Scholar PubMed PubMed Central

30. Selvaraju, RR, Cogswell, M, Das, A, Vedantam, R, Parikh, D, Batra, D. Grad-CAM: visual explanations from deep networks via gradient-based localization. Int J Comput Vis 2020;128:336–59. https://doi.org/10.1007/s11263-019-01228-7.Suche in Google Scholar

31. Kingma, DP, Ba, J. Adam: a method for stochastic optimization. ArXiv 2014. [Epub ahead of print]. https://doi.org/10.48550/arXiv.1412.6980.Suche in Google Scholar

32. Narlı, SS. COVID-19 detection using deep learning on chest X-rays enhanced by adaptive methods [master’s thesis]. Hatay (Türkiye): Iskenderun Technical University; 2021.Suche in Google Scholar

33. Apostolopoulos, ID, Mpesiana, AT. Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys Eng Sci Med 2020;43(2):635–40. https://doi.org/10.1007/s13246-020-00865-4 Suche in Google Scholar PubMed PubMed Central

34. Jain, R, Gupta, M, Taneja, S, Hemanth, DJ. Deep learning based detection and analysis of COVID-19 on chest x-ray images. Appl Intell 2021;51(3):1690–1700. https://doi.org/10.1007/s10489-020-01902-1 Suche in Google Scholar PubMed PubMed Central

35. Mangal, A, Kalia, S, Rajagopal, H, Rangarajan, K, Namboodiri, V, Banerjee, S, Arora, C. CovidAID: COVID-19 detection using chest x-ray. ArXiv 2020. [Epub ahead of print]. https://doi.org/10.48550/arXiv.2004.09803 Suche in Google Scholar

36. Ucar, F, Korkmaz, D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from x-ray images. Med Hypotheses 2020;140:109761. https://doi.org/10.1016/j.mehy.2020.109761 Suche in Google Scholar PubMed PubMed Central

Received: 2021-09-03
Accepted: 2024-09-10
Published Online: 2024-10-08
Published in Print: 2025-02-25

© 2024 the author(s), published by De Gruyter, Berlin/Boston

This work is licensed under the Creative Commons Attribution 4.0 International License.

Heruntergeladen am 26.9.2025 von https://www.degruyterbrill.com/document/doi/10.1515/bmt-2021-0272/html
Button zum nach oben scrollen