Home A multiscale and dual-loss network for pulmonary nodule classification
Article Open Access

A multiscale and dual-loss network for pulmonary nodule classification

  • Ping Zhang , Cheng Li and Jian Mao EMAIL logo
Published/Copyright: October 4, 2025
Become an author with De Gruyter Brill

Abstract

Detecting malignancy in pulmonary nodules holds significant clinical importance, yet existing image classification methods often struggle with inadequate feature integration and ineffective loss functions. This study proposes two innovative strategies to address these limitations: first, we introduce a multiscale feature weighted fusion technique that enhances the integration of features across different scales, allowing the model to prioritize critical pixel locations essential for accurate diagnosis. Second, we combine contrastive loss with binary cross-entropy within our training framework to improve learning from both similarities and differences among paired samples, which fosters better discrimination between similar nodules while maintaining sensitivity to variations across classes. Besides, our proposed methodologies demonstrate promising performance improvements in detecting pulmonary nodule malignancy, leading to enhanced performance and reliability compared to conventional approaches.

1 Introduction

Lung cancer, with its high incidence and mortality rates, necessitates early diagnosis to improve survival outcomes. The classification of pulmonary nodules (as shown in Figure 1) as benign or malignant is essential for identifying potential malignancies and guiding treatment plans [1,2]. Traditionally, physicians evaluate these nodules using computed tomography (CT) images; however, this process can be time consuming and relies heavily on individual experience, which may affect both efficiency and accuracy. To address these challenges, computer-aided diagnosis (CAD) systems [3,4] have emerged as valuable tools that utilize advanced image processing and machine learning algorithms to automatically analyze CT images. By learning from extensive datasets to detect subtle imaging features, CAD systems leverage artificial intelligence to aid physicians in making faster and more accurate decisions, thereby significantly advancing the assessment of pulmonary nodules and enhancing modern medical imaging. Previously, many studies have indicated that machine learning methodologies play a crucial role in advancing the diagnostic accuracy of pulmonary nodules. For instance, Liu et al. developed a systematic approach that quantifies radiological traits and emphasizes semantic imaging features for predicting malignancy, resulting in enhanced predictive accuracy [5]. In another study, Ferreira et al.’s random forest algorithm demonstrated superior performance compared to other classifiers; notably, a decision tree using only two features achieved comparable sensitivity and specificity [6]. Meanwhile, Tu et al. effectively combined localized thin-section CT with radiomics feature extraction and machine learning to accurately classify early-detected pulmonary nodules, thereby reducing false positives while improving malignancy differentiation [7]. Yang et al. highlighted essential nodular features through a review focused on ground-glass opacity nodules, which were vital for developing effective prediction models [8]. Uthoff et al. contributed by standardizing perinodular features, while Yamada et al. improved classification accuracy with positron emission tomographic/CT images [9,10]. Furthermore, Chen et al.’s method had outperformed existing approaches in malignancy prediction; similarly, Liu et al.’s integration of CT findings with CEA levels yielded superior predictions [11,12]. Recently, many investigations have highlighted the effectiveness of deep learning techniques, particularly convolutional neural networks (CNNs), in detecting and classifying pulmonary nodules from CT scans. Wang et al. introduced a hybrid CNN model that enhanced malignancy risk assessment accuracy by integrating global and local features, surpassing traditional texture and shape-based methods [13]. Dou et al. developed a three-dimensional CNN that significantly reduced false positives by incorporating multilevel information, thereby improving detection efficiency [14]. Jin et al.’s deep 3D residual CNN further advanced lung cancer diagnosis by effectively targeting false positives [15]. Additionally, Feng et al. created a weaklysupervised CNN for automated nodule segmentation using image-level labels, successfully localizing nodules in CT images [16]. The integration of transfer learning has also played a crucial role; Zhao et al.’s approach outperformed other CNN strategies in distinguishing malignant from benign nodules [17]. Recent innovations such as the multiscale gradual integration CNN had shown significant improvements on datasets like LUNA16 [18]. Moreover, the growing body of literature indicated that hybrid models, attention mechanisms, and ensemble learning strategies – such as those developed by Yuans et al. and Xu et al. – were essential for achieving high sensitivity and accuracy in lung cancer risk assessment [19,20]. These advancements underscore the continuous evolution of deep learning methodologies within medical imaging, highlighting their potential to transform clinical practices and improve patient outcomes in lung cancer detection and management.

Figure 1 
               The samples of different pulmonary nodules.
Figure 1

The samples of different pulmonary nodules.

Previous methods in image classification for detecting pulmonary nodule malignancy have made significant strides, but they still face limitations that impact their effectiveness. A major issue is the inadequate integration of features from various scales, which leads to a lack of focus on the discriminative traits necessary for accurate diagnosis. Traditional approaches often rely on single-scale feature extraction and can overlook critical contextual information, affecting performance when distinguishing between benign and malignant nodules. In addition, conventional loss functions like binary cross-entropy (BCE) do not fully leverage the relationships among paired samples, impairing generalization across classes. To address these challenges, we propose two innovative strategies: first, we implement multiscale feature weighted fusion (MFWF) to enhance information integration across different scales through spatial resizing and the computation of a weight matrix that emphasizes crucial pixel positions. This pixel-wise weight matrix guides the weighted fusion of multiscale features, prioritizing the most salient information. By effectively combining hierarchical details with semantic information, this approach enhances the model’s ability to capture complex patterns and improves its discriminability. Second, the contrastive loss excels at pulling similar pairs closer together in the embedding space and pushing dissimilar pairs farther apart, effectively emphasizing relative relationships. Simultaneously, the BCE loss ensures the model remains sensitive to variations across differing classes by providing a consistent supervisory signal focused on classification accuracy for each sample individually. Therefore, to leverage these strengths, we combine contrastive loss with BCE loss in a dual-loss training framework. This approach not only enhances discrimination among similar samples by fine-tuning the embedding space for closely related examples, but also maintains robust sensitivity to interclass variance, preventing the model from collapsing distinguishable categories into indistinguishable clusters. Overall, these integrated methodologies improve classification performance and bolster the model’s ability to generalize across diverse patient datasets encountered in clinical settings. The main contributions of this article can be summarized as follows:

  1. We introduce the MFWF that enhances the integration of features across different scales, allowing the model to focus on critical pixel positions and improving its sensitivity to discriminative characteristics essential for accurate diagnosis.

  2. By employing dual loss, we enable the model to better learn from similarities and differences among paired samples, which enhances discrimination among similar nodules while maintaining sensitivity to class variations.

  3. Our proposed methods exhibit promising performance in detecting pulmonary nodule malignancy, highlighting improvements in accuracy and reliability when compared to existing approaches.

The rest of this article is organized as follows: In Section 2, it outlines our proposed MFWF approach and the dual-loss training strategy, detailing the technical implementation and theoretical foundations underpinning these methods. Subsequently, in Section 3, we present the experimental setup, including the datasets utilized for evaluation and the metrics employed to assess performance. Finally, we conclude with a summary of key findings and future directions for research in this domain in Section 4.

2 Methodology

The main structure of this article is illustrated in Figure 2, featuring a primary architecture based on the ResNet network. To effectively extract multiscale features of pulmonary nodules, we propose the MFWF module. Furthermore, to enhance discrimination among paired samples while preserving sensitivity to class variations, we have designed a dual-loss strategy that complements this approach.

Figure 2 
               The proposed model is structured based on the ResNet architecture. A multiscale feature fusion module is introduced to effectively extract multiscale features from pulmonary nodules. Furthermore, a dual-loss strategy is designed to enhance discrimination among paired samples while preserving sensitivity to class variations, complementing this approach.
Figure 2

The proposed model is structured based on the ResNet architecture. A multiscale feature fusion module is introduced to effectively extract multiscale features from pulmonary nodules. Furthermore, a dual-loss strategy is designed to enhance discrimination among paired samples while preserving sensitivity to class variations, complementing this approach.

2.1 Revisit of residual block

In this section, we first revisit the residual block, which serves as the fundamental building block of our architectures. The key innovation that the residual block introduces is the concept of skip connections, which facilitate gradient flow through the network and mitigate issues related to vanishing gradients. The mathematical formulation of a basic residual block can be expressed as follows:

(1) y = ( x , { θ } ) + x ,

where x represents the input to the residual block, ( x , { θ } ) denotes a series of operations parameterized by weights { θ } , and y is the output of the residual block. The addition operation in equation (1) allows for the identity mapping from input x , enabling a direct path for gradients during backpropagation. To elaborate further, each residual block typically consists of two or more convolutional layers along with batch normalization and ReLU activation. The common configuration within a residual block could be given as follows:

(2) y = ( ( x , θ 1 ) , θ 2 ) + x .

This structure provides not only depth but also enhances representational capacity while ensuring that valuable features are retained throughout multiple layers.

2.2 Multiscale feature weighted fusion

The process begins with inputting an image I R H × W × C , where H represents the height, W denotes the width, and C is the number of channels. By applying the ResNet network, this image undergoes multiple convolutional layers and activation functions to extract features at various scales and abstraction levels. At this stage, we obtain corresponding feature maps for i th layer, the output feature map from layer l i can be expressed as follows:

(3) F l i = f l i ( I ) R H l i × W l i × D l i .

To reduce dimensionality while enhancing the expressive power of features, we apply a 1 × 1 convolutional layer to each extracted feature map. This not only allows us to reduce the channel dimensions while preserving spatial information but also enhances the model’s ability to capture nonlinear representations. This operation can be described by the following equation:

(4) C ( F l i ) = W ( i ) * F l i + b ( i ) ,

where W ( i ) represents the weights and b ( i ) is its corresponding bias term. Next, for all processed multiscale feature maps, we perform spatial resizing to unify their dimensions to a specified size H f , W f . This step is essential for effective combination and fusion of different scales without losing important information or details. The adjustment can be formulated as follows:

(5) R i = R ( C ( F l i ) , H f , W f ) .

After resizing, we obtain a set of new feature maps expressed as follows:

(6) R features = { R 1 , R 2 , , R n } .

Subsequently, at each pixel position ( h , w ) , we calculate the average value across different scales to construct a weight that provides insight into which scale contributes more significantly. This operation can be defined as follows:

(7) M ( h , w ) = 1 n i = 1 n R i ( h , w ) .

Following this calculation, we build a weight matrix reflecting the importance of each scale through normalization:

(8) W ( h , w ) = M ( h , w ) j = 1 n M j ( h , w ) .

Once these steps are completed, we acquire a weight matrix for every pixel that emphasizes important features. Next, these weighted feature maps are summation combined to derive a final fused representation defined by the following equation:

(9) F final = t = 1 n W t R t .

Finally, to achieve classification objectives, we employ a linear layer yielding class predictions according to:

(10) P ( y I ) = g ( F final , W p , b p ) ,

where g ( ) refers to softmax, and parameters W p , b p , represent weights and biases learned by the classifier.

2.3 Dual-loss strategy

In the binary classification task of pulmonary nodule malignancy, we first prepare a dataset consisting of pairs of pulmonary nodule images along with their corresponding labels. For each input image pair ( I 1 , I 2 ) , where I 1 and I 2 represent two pulmonary nodule images, if both belong to the same category (either both benign or both malignant), then the label is set to y = 0 ; if they belong to different categories (one benign and the other malignant), then the label is set to y = 1 . Next, we use the proposed ResNet as a feature extractor by loading pre-trained weights and removing the last fully connected layer, retaining only the part used for generating feature embeddings. During this process, for each input image I s , we obtain the feature representation through forward propagation:

(11) Z s = f ( I s ) ,

where Z s is the output feature vector, and f denotes the feedforward process through ResNet. Subsequently, to optimize model performance, we first define Euclidean distance to measure the difference in feature embeddings between the two pulmonary nodule images:

(12) D ( Z 1 , Z 2 ) = Z 1 Z 2 .

Then we compute the overall contrastive loss L contrastive ( y , D ( Z 1 , Z 2 ) ) :

(13) L c ( y , D ( Z 1 , Z 2 ) ) = 0.5 D ( Z 1 , Z 2 ) 2 if y = 0 0.5 ( max ( 0 , m D ( Z 1 , Z 2 ) ) ) 2 if y = 1 .

In this context, L c ( y , D ( Z 1 , Z 2 ) ) is the contrastive loss; when samples are from the same class (i.e., y = 0 ), we minimize their distance, whereas when samples are from different classes (i.e., y = 1 ), we desire that their distance exceeds some threshold m . Simultaneously, we use BCE loss to assess how well the network predicts given inputs. Corresponding to each category, the network produces a logits value that is transformed into probability via a sigmoid function:

(14) p ( y I s ) = σ ( z s ) , z s = g ( I s ) ,

where p ( y I s ) indicates the probability of malignancy given provided pulmonary nodule image I s and z s is its logits value after mapping; while function g ( I s ) maps inputs into output space via the neural network structure. Thus, the BCE loss can be defined as follows:

(15) L bce ( y , p ( y I ) ) = [ y log ( p ( y I ) ) + ( 1 y ) log ( 1 p ( y I ) ) ] .

Finally, these two losses are combined into an overall loss function:

(16) L dual = L contrastive + L bce ,

where L dual represents total loss including contrastive loss and binary cross-entropy loss. During training, paired data will be passed into model preferentially, followed by performing forward propagation in every epoch leading towards obtaining feature embedding accordingly. Full consideration regarding total loss would include comprehensive optimization schemes toward respective parameter updates throughout iterations.

3 Experiment

In this section, we outline a series of experiments aimed at validating the effectiveness of our proposed methodologies for detecting pulmonary nodule malignancy. We begin by describing the dataset utilized, along with the implementation details and evaluation metrics. Following this, we conduct a systematic set of experiments to assess and confirm the efficacy of our proposed model.

3.1 Datasets

The LIDC-IDRI (Lung Image Database Consortium and Image Database Resource Initiative, https://www.cancerimagingarchive.net/collection/lidc-idri/) is a public database dedicated to the detection and classification of pulmonary nodules. It consists of CT scan images from 1,018 patients, with slice thicknesses varying from 0.6 to 5 mm. To ensure accuracy, each nodule’s diagnostic information – including location, diameter, malignancy classified on a scale of 1 to 5, calcification level, and spiculation – has been independently annotated by four radiologists. This research aims to develop a network that classifies nodules as benign or malignant based on a voting strategy; specifically, if at least half of the experts rate a nodule above 3, it will be classified as malignant; otherwise, it will be deemed benign. To simplify model complexity, only the center cross-section image of each nodule is used as input.

3.2 Implementation details

The classification network was developed in an experimental environment using Keras, leveraging the power of an Nvidia GTX 3090 GPU. In addition, we implemented a learning rate adaptation strategy known as ReduceLROnPlateau, which adjusts the learning rate based on validation performance metrics. To further enhance model generalization, we applied various data augmentation techniques such as random rotations, vertical and horizontal flips, and brightness adjustments. The experimental results are validated by cross-validation to ensure generalizability, where performance metrics are averaged over all folds.

3.3 Evaluation metrics

Accuracy measures the overall model performance in correctly classifying instances. It is calculated as the ratio of true positives (TP), which are correctly identified positive cases, and true negatives (TN), which are correctly identified negative cases, to the total number of instances assessed (including false positives (FP) and false negatives (FN)):

(17) Accuracy = TP + TN TP + TN + FP + FN .

Sensitivity indicates the model’s ability to identify actual positive instances. It is defined as the proportion of true positives relative to all actual positives (TP plus FN):

(18) Sensitivity = TP TP + FN .

Specificity assesses the model’s effectiveness in identifying negative instances. It expresses the proportion of TN out of all actual negatives (TN plus FP):

(19) Specificity = TN TN + FP .

The area under receiver operating characteristic (ROC) curve (AUC) summarizes the model’s capability to discriminate between classes; higher values indicate better performance. AUC is evaluated by plotting the true positive rate against the false positive rate across various thresholds.

3.4 Evaluation of multiscale feature weighted fusion

This experiment evaluates the impact of MFWF on model performance. A comparison is made between the classification different evaluation metrics of the model employing the proposed MFWF and that without (w/o) MFWF as shown in Table 1. The results indicate an effective enhancement in the recognition of critical features across varying scales, suggesting that the integration of information from multiple scales allows for improved attention to essential characteristics relevant to accurate classification. This improvement highlights the effectiveness of our approach in addressing the limitations associated with single-scale methods, ultimately leading to better diagnostic capabilities in classifying pulmonary nodules.

Table 1

Evaluation of multiscale feature weighted fusion

Accuracy (%) Sensitivity (%) Specificity (%) AUC
w/o MFWF 90.12 90.00 90.23 0.902
w/ MFWF 91.34 92.44 90.26 0.924

3.5 Different fusion modes of MFWF

In the MFWF process, we examine various fusion techniques such as averaging, concatenation, and linear fusion alongside simple summation for integrating features from different layers. As shown in Figure 3, our experimental results demonstrate that summation consistently outperforms the other methods due to its ability to emphasize significant features captured in deeper layers while avoiding dimensionality issues present in concatenation. Unlike averaging, which dilutes key information by giving equal weight to all inputs, summation effectively enhances classification performance by retaining critical details. While linear blending shows promise in specific contexts through its adaptability, it does not achieve the robust performance exhibited by the summation method.

Figure 3 
                  Different fusion modes of MFWF. a, b, c, d represents the averaging, concatenation, linear, and summation, respectively.
Figure 3

Different fusion modes of MFWF. a, b, c, d represents the averaging, concatenation, linear, and summation, respectively.

3.6 Impact of dual-loss training strategy

The experiments conducted to evaluate the benefits of combining L dual with L contrastive and L bce demonstrate enhanced generalization performance, as shown in Table 2. The dual-loss strategy ultimately achieves superior results. This improvement can be attributed to its ability to facilitate learning from both similarities and differences among paired samples. By integrating both types of information, the model gains a more nuanced understanding of the relationships between benign and malignant nodules, surpassing the performance of using a single loss function ( L bce or L contrastive ) alone.

Table 2

Impact of dual-loss training strategy

Accuracy (%) Sensitivity (%) Specificity (%) AUC
L bce 90.72 91.53 90.01 0.905
L contrastive 89.16 90.02 88.37 0.868
L bce + L contrastive 91.34 92.44 90.26 0.924

3.7 The integration of various input sizes

In the experiment, image sizes of 64 × 64 , 96 × 96 , 128 × 128 , 256 × 256 , and 512 × 512 are utilized. As illustrated in Figure 4, the results show that the model performs best at an input size of 128 × 128 . This optimal performance stems from the ability to balance sufficient detail capture with manageable computational complexity. At this size, the model effectively retains important features necessary for classification while avoiding information loss seen in smaller sizes and excessive noise found in larger sizes. Ultimately, the choice of 128 × 128 enables efficient focus on significant patterns, leading to improved feature extraction and enhanced classification accuracy in deep learning models.

Figure 4 
                  The integration of various input sizes.
Figure 4

The integration of various input sizes.

3.8 Comparison with state-of-the-art models

To validate the proposed methods further, a comparative analysis is implemented against several state-of-the-art image classification models currently employed for malignancy detection in pulmonary nodules. Notably, since the sizes of each method in the original research is not the same, for a fair comparison, we reemploy those methods on the same data to validate the effectiveness of our proposed method. All the hyperparameters are obeyed the original settings. Here, we compare those methods as follow: Nibali et al. [21], Shen et al. [22], Zhang et al. [23], Wang et al. [24], Zhang et al. [25], Huo et al. [26], Ding et al. [27] (only using the single stream and modality), and Manzari et al. [28]. As shown in Table 3 and Figure 5, the results indicate that the proposed approach consistently outperforms existing methodologies across all evaluation metrics. Notably, we conduct paired t-tests to compare our proposed method with other existing approaches, and the p -values obtained from these comparisons are all below 0.05, thereby confirming the statistical significance of the reported improvements. Furthermore, this improved performance can be attributed to the effective integration of multiscale features employed during training, which enhances the model’s ability to discern subtle differences between benign and malignant nodules. In addition, the use of contrastive loss facilitates better learning from similar samples, contributing to an overall increase in diagnostic accuracy and reliability. These findings underscore the advancements achieved by our proposed methods in enhancing the classification capabilities for pulmonary nodule malignancy.

Table 3

Compared with state-of-the-art methods

Accuracy (%) Sensitivity (%) Specificity (%) AUC
Nibali et al. [21] 87.85 84.88 90.74 0.875
Shen et al. [22] 86.16 83.42 88.84 0.862
Zhang et al. [23] 85.92 85.12 86.72 0.859
Wang et al. [24] 85.92 91.46 80.52 0.852
Zhang et al. [25] 90.61 91.22 90.02 0.908
Huo et al. [26] 85.08 91.71 78.62 0.841
Ding et al. [27] 84.96 84.39 85.51 0.847
Manzari et al. [28] 84.72 87.56 81.95 0.853
Ours 91.34 92.44 90.26 0.924
Figure 5 
                  The ROC comparisons with state-of-the-art methods.
Figure 5

The ROC comparisons with state-of-the-art methods.

3.9 Different depths of ResNet

The influence of different depths of ResNet on performance is investigated through experiments examining various architectures. As shown in Figure 6, results indicate that as the depth of ResNet increases, performance initially improves due to enhanced capabilities in capturing complex features and learning intricate hierarchical representations. This improvement arises because deeper networks can learn more abstract representations, allowing them to model complex patterns in the data effectively. However, after reaching a certain depth, this performance stabilizes, suggesting diminishing returns regarding further depth. The saturation occurs because very deep networks may struggle with issues like overfitting and increased difficulty during training. Overall, while increasing depth enhances classification performance up to a point by leveraging these advanced capabilities, careful consideration is necessary regarding training dynamics and computational resources in deeper configurations.

Figure 6 
                  Comparisons with different depths.
Figure 6

Comparisons with different depths.

4 Conclusion

In this article, we present innovative strategies to enhance the effectiveness of image classification for classifying pulmonary nodule malignancy. Our approach addresses limitations in previous methods, notably the inadequate integration of multiscale features and the suboptimal use of conventional loss functions. By implementing multiscale feature weighted fusion, we improve information integration across various scales, allowing our model to focus on key pixel positions essential for accurate diagnosis. Furthermore, the combination of contrastive loss with binary cross entropy facilitates superior learning from similarities and differences among paired samples, enhancing discrimination between similar nodules while maintaining sensitivity to class variations. The results demonstrate that our methodologies yield improvements in classification performance and reliability compared to existing approaches. Looking ahead, future research could explore further optimization of feature integration techniques and investigate additional hybrid loss functions that leverage more complex relationships within datasets, aiming to refine diagnostic accuracy even further and expand applications to other medical imaging scenarios.

  1. Funding information: This research received no external funding.

  2. Author contributions: Ping Zhang and Li Cheng developed the conceptual framework. Li Cheng handled data collection and analysis, while Jian Mao oversaw the project and ensured result integrity. All authors contributed to drafting and revising the manuscript.

  3. Conflict of interest: The authors declare that there is no conflict of interest regarding the publication of this article.

  4. Ethical approval: This research was based on publicly available data and did not involve any direct interaction with human subjects or use of animal subjects. Therefore, ethical approval was not applicable.

  5. Data availability statement: The data used to support the findings of are available on: https://www.cancerimagingarchive.net/collection/lidc-idri/.

References

[1] A. McWilliams, M. C. Tammemagi, J. R. Mayo, H. Roberts, G. Liu, K. Soghrati, et al., “Probability of cancer in pulmonary nodules detected on first screening CT,” New Engl. J. Med., vol. 369, no. 10, pp. 910–919, 2013. 10.1056/NEJMoa1214726Search in Google Scholar PubMed PubMed Central

[2] A. E. Prosper, M. N. Kammer, F. Maldonado, D. R. Aberle, and W. Hsu, “Expanding role of advanced image analysis in ct-detected indeterminate pulmonary nodules and early lung cancer characterization,” Radiology, vol. 309, no. 1, p. e222904, 2023. 10.1148/radiol.222904Search in Google Scholar PubMed PubMed Central

[3] M. Javaid, M. Javid, M. Z. U. Rehman, and S. I. A. Shah, “A novel approach to CAD system for the detection of lung nodules in CT images,” Comput. Methods Programs Biomed., vol. 135, pp. 125–139, 2016. 10.1016/j.cmpb.2016.07.031Search in Google Scholar PubMed

[4] R. Yuan, P. M. Vos, and P. L. Cooperberg, “Computer-aided detection in screening CT for pulmonary nodules,” Am. J. Roentgenol., vol. 186, no. 5, pp. 1280–1287, 2006. 10.2214/AJR.04.1969Search in Google Scholar PubMed

[5] Y. Liu, Y. Balagurunathan, T. Atwater, S. Antic, Q. Li, R. C. Walker, et al., “Radiological image traits predictive of cancer status in pulmonary nodules,” Clin. Cancer Res., vol. 23, no. 6, pp. 1442–1449, 2016. http://dx.doi.org/10.1158/1078-0432.ccr-15-3102. Search in Google Scholar

[6] J. R. Ferreira, M. C. Oliveira, and P. M. de Azevedo-Marques, “Characterization of pulmonary nodules based on features of margin sharpness and texture,” J. Digit. Imag., vol. 31, no. 4, 451–463, 2017. http://dx.doi.org/10.1007/s10278-017-0029-8. Search in Google Scholar PubMed PubMed Central

[7] S.-J. Tu, C.-W. Wang, K.-T. Pan, Y.-C. Wu, and C.-T. Wu, “Localized thin-section ct with radiomics feature extraction and machine learning to classify early-detected pulmonary nodules from lung cancer screening,” Phys. Med. Biol., vol. 63, 065005, 2018, http://dx.doi.org/10.1088/1361-6560/aaafab. Search in Google Scholar PubMed

[8] J. Yang, H. Wang, C. Geng, Y. Dai, and J. Ji, “Advances in intelligent diagnosis methods for pulmonary ground-glass opacity nodules,” BioMed. Eng. OnLine, vol. 17, no. 1, 20, 2018, http://dx.doi.org/10.1186/s12938-018-0435-2. Search in Google Scholar PubMed PubMed Central

[9] J. Uthoff, M. J. Stephens, J. D. Newell, E. A. Hoffman, J. Larson, N. Koehn, et al., “Machine learning approach for distinguishing malignant and benign lung nodules utilizing standardized perinodular parenchymal features from CT,” Med. Phys., vol. 46, no. 7, pp. 3207–3216, 2019, http://dx.doi.org/10.1002/mp.13592. Search in Google Scholar PubMed PubMed Central

[10] A. Yamada, A. Teramoto, M. Hoshi, H. Toyama, K. Imaizumi, K. Saito, and H. Fujita, “Hybrid scheme for automated classification of pulmonary nodules using pet/CT images and patient information,” Appl. Sci., vol. 10, no. 2, 4225, 2020, http://dx.doi.org/10.3390/app10124225. Search in Google Scholar

[11] K. Chen, Y. Nie, S. Park, K. Zhang, Y. Zhang, Y. Liu, et al., “Development and validation of machine learning-based model for the prediction of malignancy in multiple pulmonary nodules: Analysis from multicentric cohorts,” Clin. Cancer Res., vol. 27, no. 8, pp. 2255–2265, 2021, http://dx.doi.org/10.1158/1078-0432.ccr-20-4007. Search in Google Scholar

[12] M. Liu, Z. Zhou, F. Liu, M. Wang, Y. Wang, M. Gao, et al., “Ct and CEA-based machine learning model for predicting malignant pulmonary nodules,” Cancer Sci., vol. 113, no. 12, pp. 4363–4373, 2022. http://dx.doi.org/10.1111/cas.15561. Search in Google Scholar PubMed PubMed Central

[13] H. Wang, T. Zhao, L. C. Li, H. Pan, W. Liu, H. Gao, et al., “A hybrid CNN feature model for pulmonary nodule malignancy risk differentiation,” J. X-Ray Sci. Tech., vol. 26, no. 2, pp. 171–187, 2017. http://dx.doi.org/10.3233/xst-17302. Search in Google Scholar PubMed

[14] Q. Dou, H. Chen, L. Yu, J. Qin, and P.-A. Heng, “Multilevel contextual 3-D CNNs for false positive reduction in pulmonary nodule detection,” IEEE Trans. Biomed. Eng., vol. 64, no. 7, 2017, pp. 1558–1567, http://dx.doi.org/10.1109/tbme.2016.2613502. Search in Google Scholar PubMed

[15] H. Jin, Z. Li, R. Tong, and L. Lin, “A deep 3D residual CNN for false-positive reduction in pulmonary nodule detection,” Med. Phys., vol. 45, no. 5, pp. 2097–2107, 2018. http://dx.doi.org/10.1002/mp.12846. Search in Google Scholar PubMed

[16] X. Feng, J. Yang, A. F. Laine, and E. D. Angelini, “Discriminative localization in cnns for weakly-supervised segmentation of pulmonary nodules,” Multimodal Brain Image Anal., vol. 10435, pp. 568–576, 2017. http://dx.doi.org/10.1007/978-3-319-66179-7_65. Search in Google Scholar PubMed PubMed Central

[17] X. Zhao, S. Qi, B. Zhang, H. Ma, W. Qian, Y. Yao, and J. Sun, “Deep CNN models for pulmonary nodule classification: Model modification, model integration, and transfer learning,” J. X-Ray Sci. Tech., vol. 27, no. 4, pp. 615–629, 2019, http://dx.doi.org/10.3233/xst-180490. Search in Google Scholar

[18] B.-C. Kim, J. S. Yoon, J.-S. Choi, and H.-I. Suk, “Multi-scale gradual integration cnn for false positive reduction in pulmonary nodule detection,” Neural Netw., vol. 115, no., pp. 1–10, 2019, http://dx.doi.org/10.1016/j.neunet.2019.03.003. Search in Google Scholar PubMed

[19] Y. Xu, S. Wang, X. Sun, Y. Yang, J. Fan, W. Jin, et al., “Identification of benign and malignant lung nodules in CT images based on ensemble learning method,” Interdiscipl. Sci. Comput. Life Sci., vol. 14, no. 1, pp. 130–140, 2022, http://dx.doi.org/10.1007/s12539-021-00472-1. Search in Google Scholar PubMed

[20] H. Yuan, Z. Fan, Y. Wu, and J. Cheng, “An efficient multi-path 3D convolutional neural network for false-positive reduction of pulmonary nodule detection,” Int. J. Comput. Assisted Radiol. Surgery, vol. 16, pp. 2269–2277, 2021, http://dx.doi.org/10.1007/s11548-021-02478-y. Search in Google Scholar PubMed

[21] A. Nibali, Z. He, and D. Wollersheim, “Pulmonary nodule classification with deep residual networks,” Int. J. Comput. Assist. Radiol. Surgery, vol. 12, pp. 1799–1808, 2017. 10.1007/s11548-017-1605-6Search in Google Scholar PubMed

[22] W. Shen, M. Zhou, F. Yang, C. Yang, and J. Tian, “Multi-scale convolutional neural networks for lung nodule classification,” in Information Processing in Medical Imaging: 24th International Conference, IPMI 2015, Sabhal Mor Ostaig, Isle of Skye, UK, June 28–July 3, 2015, Proceedings 24. Springer, 2015, pp. 588–599. 10.1007/978-3-319-19992-4_46Search in Google Scholar PubMed

[23] G. Zhang, D. Zhu, X. Liu, M. Chen, L. Itti, Y. Luo, and J. Lu, “Multi-scale pulmonary nodule classification with deep feature fusion via residual network,”, J. Ambient Intel. Humanized Comput., vol. 14, pp. 1–12, 2023. Search in Google Scholar

[24] Y. Wang, H. Zhang, K. J. Chae, Y. Choi, G. Y. Jin, and S.-B. Ko, “Novel convolutional neural network architecture for improved pulmonary nodule classification on computed tomography,” Multidimen. Syst. Signal Proces., vol. 31, pp. 1163–1183, 2020. 10.1007/s11045-020-00703-6Search in Google Scholar

[25] Y. Zhang, J. Zhang, L. Zhao, X. Wei, and Q. Zhang, “Classification of benign and malignant pulmonary nodules based on deep learning,” in 2018 5th International Conference on Information Science and Control Engineering (ICISCE). IEEE, 2018, pp. 156–160. 10.1109/ICISCE.2018.00042Search in Google Scholar

[26] X. Huo, G. Sun, S. Tian, Y. Wang, L. Yu, J. Long, et al., “Hifuse: Hierarchical multiscale feature fusion network for medical image classification,” Biomed. Signal Proces. Control, vol. 87, p. 105534, 2024. 10.1016/j.bspc.2023.105534Search in Google Scholar

[27] Z. Ding, H. Li, Y. Guo, D. Zhou, Y. Liu, and S. Xie, “M4fnet: Multimodal medical image fusion network via multi-receptive-field and multiscale feature integration,” Comput. Biol. Med., vol. 159, p. 106923, 2023. 10.1016/j.compbiomed.2023.106923Search in Google Scholar PubMed

[28] O. N. Manzari, H. Ahmadabadi, H. Kashiani, S. B. Shokouhi, and A. Ayatollahi, “Medvit: a robust vision transformer for generalized medical image classification,” Comput. Biol. Med., vol. 157, p. 106791, 2023. 10.1016/j.compbiomed.2023.106791Search in Google Scholar PubMed

Received: 2024-09-27
Revised: 2025-03-02
Accepted: 2025-04-28
Published Online: 2025-10-04

© 2025 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Research Articles
  2. Intelligent data collection algorithm research for WSNs
  3. A novel behavioral health care dataset creation from multiple drug review datasets and drugs prescription using EDA
  4. Speech emotion recognition using long-term average spectrum
  5. PLASMA-Privacy-Preserved Lightweight and Secure Multi-level Authentication scheme for IoMT-based smart healthcare
  6. Basketball action recognition by fusing video recognition techniques with an SSD target detection algorithm
  7. Evaluating impact of different factors on electric vehicle charging demand
  8. An in-depth exploration of supervised and semi-supervised learning on face recognition
  9. The reform of the teaching mode of aesthetic education for university students based on digital media technology
  10. QCI-WSC: Estimation and prediction of QoS confidence interval for web service composition based on Bootstrap
  11. Line segment using displacement prior
  12. 3D reconstruction study of motion blur non-coded targets based on the iterative relaxation method
  13. Overcoming the cold-start challenge in recommender systems: A novel two-stage framework
  14. Optimization of multi-objective recognition based on video tracking technology
  15. An ADMM-based heuristic algorithm for optimization problems over nonconvex second-order cone
  16. A multiscale and dual-loss network for pulmonary nodule classification
  17. Special Issue on AI based Techniques in Wireless Sensor Networks
  18. Blended teaching design of UMU interactive learning platform for cultivating students’ cultural literacy
  19. Special Issue on Informatics 2024
  20. Analysis of different IDS-based machine learning models for secure data transmission in IoT networks
  21. Using artificial intelligence tools for level of service classifications within the smart city concept
  22. Applying metaheuristic methods for staffing in railway depots
  23. Interacting with vector databases by means of domain-specific language
  24. Data analysis for efficient dynamic IoT task scheduling in a simulated edge cloud environment
  25. Analysis of the resilience of open source smart home platforms to DDoS attacks
  26. Comparison of various in-order iterator implementations in C++
Downloaded on 17.10.2025 from https://www.degruyterbrill.com/document/doi/10.1515/comp-2025-0030/html
Scroll to top button