Abstract
Objectives
This study explores the prognostic relevance of hemogram parameters and derived ratios in stratifying hospitalization levels, including intensive care unit (ICU) admission, using machine learning models.
Methods
Clinical and laboratory data were retrospectively obtained from a tertiary-level teaching hospital encompassing a wide range of care settings. Five classification models – Multinomial Logistic Regression, Random Forest, Gradient Boosting Machine (GBM), Support Vector Machine, and Multi-Layer Perceptron Neural Network (MLP-ANN) – were trained and evaluated. External validation was conducted using an independent dataset from affiliated institutions. SHAP (Shapley Additive Explanations) analysis was applied to enhance model interpretability.
Results
Among the evaluated models, GBM and MLP-ANN demonstrated consistent performance across treatment categories, particularly in distinguishing outpatient from higher-acuity hospital admissions. SHAP analysis identified eosinophil and basophil counts as key predictors, with eosinophil levels inversely associated with hospitalization intensity. External validation supported the models’ applicability across patient subsets, although multicenter generalizability remains to be established.
Conclusions
This study demonstrates that machine learning models trained on routine hemogram data can effectively differentiate hospitalization levels, offering a scalable approach to early clinical risk assessment. Eosinophil count, in particular, emerged as a key variable with consistent predictive power across models, suggesting its potential relevance in acute care stratification. These findings support the integration of hematological parameters into real-time triage systems and pave the way for future research to validate their role in diverse clinical contexts.
Introduction
The complete blood count (CBC), also known as the hemogram, is one of the most commonly utilized laboratory investigations in healthcare settings and comprises a panel of sub parameters reflecting various hematological indices [1]. Each individual parameter is employed in the diagnosis and/or monitoring of specific clinical conditions.
Growing research highlights hemogram-derived indices – especially the neutrophil-to-lymphocyte (NLR), platelet-to-lymphocyte (PLR), and lymphocyte-to-monocyte (LMR) ratios – as novel immuno-inflammatory biomarkers. These metrics aid diagnosis and prognosis in diverse conditions, including non-alcoholic fatty liver disease, oral cancer, osteoporosis, venous thromboembolism, and traumatic brain injury [2], [3], [4].
During the COVID-19 pandemic, hemogram-derived parameters have been extensively investigated in relation to disease morbidity and mortality, and have been associated with critical clinical outcomes such as the need for intensive care among infected patients [5], 6].
Moreover, these parameters are attractive for clinical use due to their affordability, rapid turnaround time, and widespread availability in routine laboratory workflows. Their practical advantages make them suitable candidates for incorporation into cost-effective and scalable diagnostic strategies [7].
However, studies evaluating the predictive power of these biomarkers for outcomes such as hospitalization or intensive care unit (ICU) admission remain limited, and most have focused on specific disease populations [8]. Large-scale retrospective analyses conducted on the general patient population are currently lacking.
In particular, analyses utilizing large-scale databases have not yet reached a clear consensus regarding the prognostic utility of hemogram-derived ratios in predicting hospitalization and ICU requirements. Existing studies have predominantly focused on specific disease cohorts, such as patients with sepsis or malignancies [8], 9].
This study seeks to evaluate the association between hemogram parameters and the requirement for hospitalization and intensive care through a single-center retrospective analysis, with the aim of addressing a gap in the current literature and providing a foundational framework for the future development of clinical decision support systems.
Materials and methods
Study population/subjects
The study was run at SBU Izmir Tepecik Training and Research Hospital, a tertiary center, and its affiliated hospital. Ethics approval was obtained (No. 2025/02-23; 10 Mar 2025). We included patients who had their first CBC between 1 Jan and 31 Dec 2024 at the main hospital, while identically filtered data from the affiliated hospital – providing outpatient care and palliative care, obstetric, and gynecologic-oncology units – formed the external test set.
Inclusion criteria
Patients who underwent a CBC test.
Patients with clearly documented hospitalization status (admitted or not).
Patients with at least one complete and valid set of CBC parameters.
Patients admitted to the intensive care unit, inpatient wards, or managed on an outpatient basis.
Exclusion criteria
Patients with incomplete or erroneous laboratory data.
Patients with known chronic hematological diseases.
Patients receiving chemotherapy or immunosuppressive therapy.
Patients with acute trauma or those who underwent major surgery within the past 3 months.
Pregnant or postpartum patients.
Patients with chronic kidney disease (≥stage 4) or those undergoing hemodialysis.
Patients with congenital or acquired immunodeficiency syndromes.
Patients with duplicate CBC test entries during the study period.
Whole blood was drawn into K2-EDTA tubes by ward nurses or phlebotomy nurses and analyzed within 30 min on a Sysmex XN-1000 (Kobe, Japan). All CBC reagents, calibrators, and internal quality control (QC) materials were certified, manufacturer-supplied products.
Study design
The study commenced following the approval of the institutional ethics committee. All patient identifiers were anonymized to ensure confidentiality. A dataset comprising information on treatment type, age, sex, and hemogram results – including white blood cell (WBC), neutrophil (NEU), lymphocyte (LYM), monocyte (MONO), eosinophil (EOS), basophil (BASO), platelet count (PLT), hemoglobin (HGB), red cell distribution width (RDW), mean corpuscular volume (MCV), mean platelet volume (MPV), platelet distribution width (PDW), hematocrit (HCT), as well as calculated ratios such as NLR, PLR and LMR was compiled from 565,877 patients and exported to Microsoft Excel 2021 (USA) for analysis.
HGB and HCT more accurately represent oxygen-transport capacity than red-blood-cell count (RBC), which also lacks prognostic value for acute admission. Plateletcrit (PCT) is a calculated, not measured, parameter (PLT × MPV/10,000). Mean corpuscular hemoglobin (MCH) and its concentration (MCHC) are mainly anemia markers and do not indicate acute severity or systemic inflammation. Therefore, RBC, PCT, MCH, and MCHC were excluded.
After applying the exclusion criteria, the final dataset consisted of 430,561 patients. This dataset was subsequently imported into Python (version 3.11, USA) for machine learning-based analysis. Non-numeric categorical variables were recoded into numerical values as follows: male=0, female=1; care setting: outpatient 0, surgical ward 1, internal ward 2, ICU 3. Outpatients exceeded inpatients ten-to-one, so 10,000 cases per group were randomly sampled. After identical filtering, the affiliated hospital supplied 6,132 records for external testing.
After cleaning, data were stratified by the four-class outcome and randomly split 75:25 into training and internal-test sets, preserving class proportions. Figure 1 – an STARD flow diagram – shows screening, exclusions, final sample sizes, and allocation to training, internal, and external test sets.

The standards for reporting diagnostic accuracy (STARD) diagram. Cohort selection workflow. A step-wise exclusion of 565,877 candidate samples produces a balanced core dataset of 40,000 cases. The data are then split into training, internal test and external test cohorts (3:1:≈0.2). Sub-category counts (ICU, internal medicine, surgical, outpatients) are annotated in each block.
Data preprocessing and training of machine learning algorithms
The cleaned dataset was imported into the Python programming environment (v3.11) using PyCharm IDE. Model development and evaluation were performed with widely used libraries such as TensorFlow and scikit-learn, enabling efficient data preprocessing, training, and optimization [10].
We tested five classifiers – multinomial logistic regression (MN-LR), random forest (RF), gradient boosting (GBM), support-vector machines (SVM), and a multilayer perceptron neural network (MLP-ANN). All are proven for structured clinical data; ensemble (RF, GBM) and neural-network (MLP-ANN) models capture the nonlinear, high-dimensional patterns typical of biomedical datasets.
All procedures were conducted within the Python 3.11 programming environment. The libraries utilized are categorized as follows:
Data Processing and Analysis: pandas (v1.5), numpy (v1.23).
Machine Learning Model Development: scikit-learn (v1.2), xgboost (v1.6), lightgbm (v3.3), tensorflow (v2.10), keras (v2.10).
Model Evaluation and Visualization: matplotlib (v3.6), seaborn (v0.12), scipy. stats (v1.9), sklearn. metrics (v1.2), mlxtend (v0.21), shap (v0.47).
Model hyperparameters were optimized via grid-search cross-validation, and performance was subsequently assessed on an independent, stratified test set. Residual class imbalance was mitigated by applying inverse-frequency class weights during training, thereby limiting majority-class bias and ensuring balanced discrimination across outcome categories.
Following model training, performance evaluation was conducted using the designated test dataset.
Performance evaluation
The modeling process underwent comprehensive evaluation, including hyperparameter tuning and model selection through internal cross-validation. Model performance was assessed using multiple evaluation metrics. The following criteria were used for classification:
Classification Performance Metrics
Area Under the Receiver Operating Characteristic Curve (AUC-ROC)
Area Under the Precision-Recall Curve (AUC-PR)
Confusion Matrix Analysis
Sensitivity, Specificity, Positive Predictive Value (PPV), and Negative Predictive Value (NPV), F1 Score
Model Interpretability Metrics
Feature Importance Analysis
Two-Dimensional Decision Boundary Visualization
SHAP (Shapley Additive Explanations) Graphs: SHAP values were employed to enhance model transparency and enable instance-level interpretation of feature contributions
Validation Results of the Predictive Models (with external test set): were analyzed to ensure comprehensive assessment. This structured and multifaceted evaluation approach provides a robust framework for predicting treatment modality outcomes based on laboratory-derived data.
Statistical analysis
Statistical analyses were performed using Python (v3.11). Continuous variables were summarized as means and standard deviations, while categorical variables were expressed as counts and percentages. Differences between the development and external test sets were assessed using independent samples t-test for age and Pearson Chi-square test for sex distribution.
To compare hemogram parameters across treatment groups (outpatient, surgical ward, internal medicine ward, ICU), the Kruskal–Wallis test was applied for continuous variables due to non-normal distribution. For categorical comparisons, the Pearson Chi-square test was used. A two-tailed p-value of <0.05 was considered statistically significant.
Results
Dataset description and data pre-processing
The study included a total of 40,000 data points derived from hospital records. The dataset consisted of hemogram parameters in conjunction with demographic characteristics of the patients. In addition, a preprocessed external test comprising 6,132 records was appended to the main dataset. Baseline characteristics of the study population, including main dataset (train and internal test set) and external test set, are summarized in Table 1, while the descriptive statistics of the hemogram and related variables are presented in Table 2. When assessing the statistical significance of differences in variable means across treatment types, the following comparisons were found to be non-significant: for lymphocyte count in treatment types 0 vs. 2 and 0 vs. 3 within the training set; and for lymphocyte (0 vs. 3), platelet (0 vs. 2), and mean platelet volume (0 vs. 1 and 0 vs. 2) within the test set. All other pairwise comparisons between treatment types demonstrated statistically significant differences in mean values (p<0.05). The main dataset (mean age 48.7 years) contained slightly older women than men (50.3 vs. 47.3 year) and more females overall (52.6 %). Male density peaked at 66–80 years and fell >80 years. External-test patients were significantly older – as they came from palliative and gynecologic-oncology wards – and included a higher proportion of women (p<0.05). A detailed breakdown of age distribution by gender is illustrated in Figure SM1 (Supplementary Material 1).
The baseline characteristics of the study population.
Characteristics | Main set (n: 40.000) Value±SD |
Train set (n: 30.000) Value±SD |
Internal test set (n: 10.000) Value±SD |
External test set (n: 6.132) Value±SD |
p-Value |
---|---|---|---|---|---|
Age, years | 48.7 ± 24.0 | 48.6 ± 24.0 | 48.8 ± 24.1 | 54.2 ± 18.3 | |
Male | 47.2 ± 23.4 | 47.0 ± 23.5 | 47.9 ± 23.3 | 57.2 ± 17.6 | p<0.05 |
Female | 50.2 ± 24.5 | 50.4 ± 24.4 | 49.9 ± 24.8 | 52.2 ± 18.6 | p<0.05 |
Sex | |||||
Male | 18.980 (47.45 %) | 14.268 (39.6 %) | 4.712 (47.1 %) | 2,469 (40.2 %) | p<0.05 |
Female | 21.020 (52.55 %) | 15.732 (60.4 %) | 5.288 (52.9 %) | 3,663 (59.8 %) | p<0.05 |
-
SD, standard deviation. p-values for age comparisons were calculated using the independent samples t-test. Sex distribution differences were analyzed using the chi-square test. Age differences between the main and external test sets were statistically significant (p<0.05), as were differences in sex distribution (p<0.05).
Descriptive statistics of the study population.
Parameters/treatment type | Outpatients-0- | Surgery services-1- | Internal services-2- | Intensive care-3- | |
---|---|---|---|---|---|
Age, years | Train | 42.5 ± 22.5 | 46.0 ± 20.1 | 50.7 ± 25.0 | 55.2 ± 26.0 |
Int. Test | 42.4 ± 22.1 | 46.1 ± 20.3 | 51.0 ± 25.2 | 55.9 ± 26.0 | |
Ext. Test | 43.3 ± 9.2 | 50.4 ± 17.1 | 58.2 ± 20.3 | 64.8 ± 17.3 | |
WBC (103/µL) | Train | 7.87 ± 4.90 | 10.0 ± 4.24 | 9.61 ± 11.3 | 10.7 ± 6.62 |
Int. Test | 7.84 ± 6.04 | 9.94 ± 4.14 | 9.34 ± 6.25 | 10.4 ± 5.51 | |
Ext. Test | 7.60 ± 1.75 | 9.54 ± 3.66 | 8.48 ± 2.64 | 10.5 ± 5.30 | |
NEU (103/µL) | Train | 4.43 ± 2.25 | 7.71 ± 6.39 | 6.16 ± 7.42 | 8.14 ± 8.29 |
Int. Test | 4.38 ± 1.99 | 7.56 ± 5.96 | 6.48 ± 7.54 | 7.77 ± 7.03 | |
Ext. Test | 4.61 ± 1.63 | 6.85 ± 3.59 | 5.53 ± 2.40 | 7.93 ± 5.27 | |
LYM (103/µL) | Train | 2.26 ± 4.03 | 1.87 ± 2.07 | 2.99 ± 9.30 a | 2.43 ± 4.30 a |
Int. Test | 2.58 ± 5.54 | 1.95 ± 2.48 | 2.64 ± 4.43 | 2.69 ± 5.70 a | |
Ext. Test | 2.20 ± 0.18 | 1.68 ± 0.40 | 1.93 ± 0.57 | 1.61 ± 0.53 | |
MONO (103/µL) | Train | 0.61 ± 0.29 | 0.78 ± 0.70 | 0.89 ± 1.39 | 0.97 ± 1.79 |
Int. Test | 0.61 ± 0.27 | 0.78 ± 0.88 | 0.87 ± 1.07 | 0.95 ± 1.19 | |
Ext. Test | 0.58 ± 0.16 | 0.77 ± 0.29 | 0.76 ± 0.31 | 0.78 ± 0.31 | |
EOS (103/µL) | Train | 0.20 ± 0.21 | 0.02 ± 0.09 | 0.03 ± 0.16 | 0.06 ± 1.18 |
Int. Test | 0.20 ± 0.23 | 0.02 ± 0.08 | 0.03 ± 0.12 | 0.06 ± 0.16 | |
Ext. Test | 0.17 ± 0.12 | 0.21 ± 0.26 | 0.21 ± 0.16 | 0.17 ± 0.18 | |
BASO (103/µL) | Train | 0.03 ± 0.05 | 0.12 ± 0.18 | 0.19 ± 0.27 | 0.15 ± 016 |
Int. Test | 0.03 ± 0.06 | 0.11 ± 0.17 | 0.19 ± 0.24 | 0.15 ± 0.17 | |
Ext. Test | 0.04 ± 0.02 | 0.04 ± 0.02 | 0.04 ± 0.02 | 0.04 ± 0.03 | |
PLT (103/µL) | Train | 280 ± 94 | 261 ± 113 | 290 ± 179 | 265 ± 133 |
Int. Test | 279 ± 90 | 259 ± 109 | 289 ± 179 a | 266 ± 130 | |
Ext. Test | 275 ± 9.9 | 259 ± 33 | 277 ± 44 | 256 ± 48 | |
HGB, g/dL | Train | 12.9 ± 1.74 | 10.9 ± 1.73 | 10.9 ± 2.20 | 11.0 ± 2.37 |
Int. Test | 13.0 ± 1.70 | 10.9 ± 1.72 | 11.0 ± 2.22 | 11.1 ± 2.41 | |
Ext. Test | 13.4 ± 1.47 | 11.0 ± 1.85 | 11.2 ± 2.07 | 10.5 ± 2.15 | |
RDW, % | Train | 14.5 ± 2.11 | 15.5 ± 3.02 | 16.2 ± 3.57 | 15.9 ± 2.95 |
Int. Test | 14.5 ± 2.13 | 15.5 ± 3.03 | 16.2 ± 3.77 | 15.8 ± 3.01 | |
Ext. Test | 13.4 ± 1.36 | 14.3 ± 2.26 | 15.1 ± 2.93 | 15.7 ± 3.00 | |
MCV, fL | Train | 83.4 ± 7.35 | 84.7 ± 7.21 | 84.2 ± 7.65 | 87.5 ± 8.21 |
Int. Test | 83.2 ± 7.06 | 84.6 ± 7.28 | 84.1 ± 7.62 | 87.4 ± 8.06 | |
Ext. Test | 86.0 ± 4.95 | 85.7 ± 6.15 | 84.1 ± 6.87 | 87.4 ± 7.28 | |
MPV, fL | Train | 8.69 ± 1.01 | 8.75 ± 1.09 | 8.75 ± 1.12 | 9.35 ± 1.39 |
Int. Test | 8.72 ± 1.00 | 8.73 ± 1.11 a | 8.72 ± 1.10 a | 9.33 ± 1.37 | |
Ext. Test | 10.4 ± 0.79 | 10.3 ± 0.90 | 10.4 ± 0.99 | 10.6 ± 1.13 | |
PDW, % | Train | 16.8 ± 0.56 | 17.0 ± 0.70 | 16.9 ± 0.83 | 15.8 ± 2.69 |
Int. Test | 16.8 ± 0.56 | 17.0 ± 0.77 | 16.9 ± 0.87 | 15.7 ± 2.75 | |
Ext. Test | 12.1 ± 1.74 | 11.6 ± 2.00 | 12.1 ± 2.22 | 12.2 ± 2.56 | |
HCT, % | Train | 38.8 ± 4.71 | 32.6 ± 4.90 | 32.4 ± 6.33 | 32.8 ± 6.75 |
Int. Test | 38.6 ± 4.53 | 32.7 ± 4.86 | 32.5 ± 6.42 | 32.8 ± 6.85 | |
Ext. Test | 40.7 ± 3.81 | 33.4 ± 5.13 | 33.7 ± 5.80 | 31.9 ± 5.85 | |
NLR | Train | 2.13 ± 1.67 | 5.50 ± 6.00 | 4.25 ± 7.87 | 6.82 ± 10.4 |
Int. Test | 2.11 ± 1.67 | 5.39 ± 6.02 | 4.67 ± 12.0 | 6.35 ± 9.63 | |
Ext. Test | 2.11 ± 0.77 | 4.33 ± 2.59 | 3.14 ± 1.76 | 5.66 ± 4.73 | |
PLR | Train | 129.7 ± 75.6 | 178.8 ± 129.4 | 173.0 ± 178.3 | 191.6 ± 187 |
Int. Test | 129.6 ± 68.1 | 178.0 ± 139.4 | 176.0 ± 164.9 | 190.1 ± 191 | |
Ext. Test | 127.9 ± 9.52 | 161.4 ± 33.8 | 153.9 ± 40.8 | 172.2 ± 48.7 | |
LMR | Train | 4.35 ± 2.86 | 2.78 ± 2.47 | 3.76 ± 7.28 | 3.05 ± 5.17 |
Int. Test | 4.38 ± 3.42 | 2.85 ± 2.38 | 3.71 ± 7.64 | 3.39 ± 7.45 | |
Ext. Test | 4.09 ± 1.13 | 2.50 ± 1.12 | 2.98 ± 1.99 | 2.31 ± 1.11 |
-
WBC, white blood cell; NEU, neutrophil count; LYM, lymphocyte count; MONO, monocyte count; EOS, eosinophil count; BASO, basophil count; PLT, platelet count; HGB, hemoglobin concentration; RDW, red cell distribution width; MCV, mean corpuscular volume; MPV, mean platelet volume; PDW, platelet distribution width; HCT, hematocrit; NLR, neutrophil-lymphocyte ratio; PLR, platelet-lymphocyte ratio; LMR, lymphocyte-monocyte ratio; Int. Test, Internal Test Set; Ext. Test, External Test Set. aSignificance of the difference between the means: p>0.05.
For the variables incorporated into the predictive models, Pearson correlation analysis was applied to assess linear relationships, whereas Spearman rank correlation was used to evaluate monotonic associations. The correlation matrices and corresponding visualizations for these analyses are presented in Figure SM2 (Supplementary Material 1).
The performance of MN-LR, RF, GBM, MLP-ANN, and SVM models was comparatively evaluated based on their predictive capabilities, classification metrics, and interpretability. Classification metrics such as F1 score, sensitivity, specificity, and AUC-ROC were used to assess the models’ ability to discriminate between classes.
Each algorithm offers distinct advantages and limitations depending on the nature of the dataset and the classification problem. MN-LR is valued for its simplicity and computational efficiency; RF and GBM provide robust predictive performance and are less prone to overfitting; MLP-ANN is particularly effective for large and complex datasets; and SVM excels in smaller, well-separated datasets. Model selection was guided by the structural characteristics of the data and the specific classification objectives. All models were systematically analyzed to reflect the diverse methodological strengths of each algorithm.
Comparison of classification performance metrics
In this section, MN-LR, RF, GBM, MLP-ANN, and SVM were compared on sensitivity, specificity, PPV, NPV, and F1-score. Metrics that satisfied normality and equal-variance criteria (e.g., F1-score) were analyzed with one-way ANOVA, while non-normal metrics (e.g., sensitivity) used the Kruskal–Wallis test to detect performance differences among models.
All models outperformed MN-LR in nearly every treatment group. Performance was strongest for outpatients: RF and SVM best distinguished them from inpatients. GBM and SVM excelled for ICU cases, whereas accuracy for internal and surgical ward patients remained moderate across models.
Although MN-LR showed the lowest mean F1 (0.60 ± 0.14), its score did not differ significantly from RF (0.64 ± 0.14), GBM (0.68 ± 0.13), MLP-ANN (0.67 ± 0.13), or SVM (0.68 ± 0.12); all t-test p-values exceeded 0.47. Likewise, GBM, MLP-ANN, and SVM displayed no pairwise differences (p>0.92). AUC-ROC and AUC-PR curves nevertheless highlighted GBM and MLP-ANN as the most accurate models across treatment groups. Detailed metrics appear in Table 3 and Figure 2, with performance curves for the other models presented in Supplementary Material 1.
Data of classification performance metrics.
Model | Treatment type | Sensitivity (95 % CI) |
Specificity (95 % CI) |
Positive predictive value (95 % CI) |
Negative predictive value (95 % CI) |
F1 score (95 % CI) |
AUC-ROC | AUC-PR | Model Macro AUC-ROC/PR | |
---|---|---|---|---|---|---|---|---|---|---|
NM-LR | Internal test set | 0 | 0.91 (0.90–0.92) | 0.90 (0.89–0.91) | 0.76 (0.75–0.77) | 0.97 (0.96–0.97) | 0.83 (0.82–0.84) | 0.96 | 0.86 | 0.80/0.70 |
1 | 0.57 (0.55–0.58) | 0.84 (0.83–0.85) | 0.55 (0.53–0.56) | 0.85 (0.84–0.86) | 0.56 (0.55–0.57) | 0.80 | 0.61 | |||
2 | 0.46 (0.44–0.47) | 0.81 (0.80–0.82) | 0.45 (0.44–0.46) | 0.82 (0.81–0.83) | 0.45 (0.44–0.46) | 0.74 | 0.44 | |||
3 | 0.48 (0.46–0.49) | 0.90 (0.90–0.92) | 0.64 (0.62–0.65) | 0.84 (0.82–0.86) | 0.55 (0.53–0.56) | 0.82 | 0.66 | |||
RF | Internal test set | 0 | 0.91 (0.90–0.92) | 0.96 (0.95–0.96) | 0.87 (0.86–0.88) | 0.97 (0.96–0.97) | 0.89 (0.88–0.90) | 0.98 | 0.93 | 0.78/0.62 |
1 | 0.56 (0.55–0.57) | 0.88 (0.87–0.88) | 0.61 (0.60–0.62) | 0.86 (0.85–0.86) | 0.58 (0.57–0.59) | 0.83 | 0.66 | |||
2 | 0.55 (0.54–0.56) | 0.84 (0.82–0.85) | 0.53 (0.52–0.54) | 0.85 (0.84–0.85) | 0.54 (0.53–0.55) | 0.79 | 0.53 | |||
3 | 0.55 (0.54–0.56) | 0.86 (0.85–0.86) | 0.56 (0.55–0.57) | 0.85 (0.84–0.85) | 0.56 (0.55–0.57) | 0.82 | 0.65 | |||
GBM | Internal test set | 0 | 0.91 (0.90–0.92) | 0.95 (0.94–0.96) | 0.86 (0.85–0.87) | 0.98 (0.96–0.97) | 0.89 (0.88–0.89) | 0.98 | 0.93 | 0.83/0.70 |
1 | 0.60 (0.58–0.60) | 0.89 (0.88–0.90) | 0.64 (0.63–0.65) | 0.87 (0.86–0.87) | 0.62 (0.61–0.63) | 0.86 | 0.72 | |||
2 | 0.58 (0.57–0.59) | 0.83 (0.82–0.84) | 0.53 (0.51–0.54) | 0.86 (0.86–0.87) | 0.55 (0.54–0.56) | 0.82 | 0.61 | |||
3 | 0.61 (0.59–0.61) | 0.90 (0.90–0.91) | 0.67 (0.66–0.68) | 0.87 (0.87–0.88) | 0.64 (0.62–0.65) | 0.89 | 0.77 | |||
MLP-ANN | Internal test set | 0 | 0.91 (0.90–0.91) | 0.96 (0.95–0.97) | 0.87 (0.86–0.88) | 0.97 (0.96–0.98) | 0.89 (0.88–0.89) | 0.98 | 0.93 | 0.83/0.70 |
1 | 0.61 (0.60–0.62) | 0.88 (0.87–0.89) | 0.63 (0.62–0.64) | 0.87 (0.86–0.88) | 0.62 (0.61–0.63) | 0.86 | 0.73 | |||
2 | 0.57 (0.56–0.58) | 0.83 (0.83–0.84) | 0.53 (0.52–0.54) | 0.85 (0.84–0.86) | 0.54 (0.54–0.59) | 0.82 | 0.60 | |||
3 | 0.60 (0.59–0.61) | 0.89 (0.88–0.90) | 0.65 (0.64–0.66) | 0.87 (0.86–0.88) | 0.63 (0.62–0.65) | 0.89 | 0.76 | |||
SVM | Internal test set | 0 | 0.90 (0.89–0.90) | 0.96 (0.95–0.96) | 0.87 (0.87–0.88) | 0.97 (0.96–0.97) | 0.89 (0.88–0.89) | 0.97 | 0.91 | 0.82/0.69 |
1 | 0.61 (0.59–0.62) | 0.89 (0.88–0.90) | 0.65 (0.64–0.66) | 0.87 (0.87–0.88) | 0.63 (0.62–0.64) | 0.84 | 0.70 | |||
2 | 0.64 (0.63–0.65) | 0.81 (0.80–0.82) | 0.52 (0.52–0.54) | 0.87 (0.87–0.88) | 0.58 (0.57–0.59) | 0.77 | 0.52 | |||
3 | 0.57 (0.56–0.58) | 0.91 (0.91–0.92) | 0.69 (0.68–0.70) | 0.88 (0.85–0.88) | 0.62 (0.62–0.63) | 0.87 | 0.70 | |||
GBM | External test set | 0 | 0.91 (0.90–0.93) | 0.97 (0.96–0.98) | 0.91 (0.90–0.92) | 0.97 (0.96–0.98) | 0.92 (0.91–0.93) | 0.95 | 0.86 | |
1 | 0.68 (0.67–0.72) | 0.90 (0.89–0.91) | 0.69 (0.68–0.70) | 0.90 (0.89–0.91) | 0.70 (0.68–0.72) | 0.80 | 0.56 | |||
2 | 0.69 (0.66–0.71) | 0.89 (0.88–0.90) | 0.66 (0.65–0.67) | 0.89 (0.88–0.91) | 0.69 (0.67–0.71) | 0.79 | 0.55 | |||
3 | 0.76 (0.74–0.78) | 0.92 (0.91–0.93) | 0.74 (0.73–0.75) | 0.92 (0.91–0.93) | 0.77 (0.75–0.78) | 0.84 | 0.65 | |||
MLP-ANN | External test set | 0 | 0.90 (0.88–0.91) | 0.97 (0.96–0.97) | 0.90 (0.88–0.90) | 0.97 (0.96–0.97) | 0.90 (0.89–0.91) | 0.93 | 0.83 | |
1 | 0.68 (0.66–0.71) | 0.90 (0.88–0.91) | 0.69 (0.66–0.71) | 0.90 (0.88–0.91) | 0.69 (0.67–0.71) | 0.79 | 0.55 | |||
2 | 0.67 (0.64–0.69) | 0.89 (0.88–0.90) | 0.67 (0.64–0.69) | 0.89 (0.88–0.91) | 0.67 (0.65–0.69) | 0.78 | 0.53 | |||
3 | 0.74 (0.71–0.75) | 0.91 (0.90–0.92) | 0.74 (0.71–0.75) | 0.91 (0.90–0.92) | 0.74 (0.72–0.75) | 0.82 | 0.61 |
-
AUC-ROC, area under the receiver operating characteristic curve; AUC-PR, area under the precision-recall curve; NM-LR, multinomial logistic regression; RF, random forest; GBM, gradient boosting machine; MLP-ANN, multilayer perceptron – artificial neural network; SVM, support vector machine; CI, confidence interval. 0, outpatients; 1, surgical services; 2, internal services; 3, intensive care unit. Bold numbers indicate the highest score according to treatment type.

AUC-ROC and AUC-PR plots of the methods. Discriminative performance via ROC and PR curves. Top two rows: class-specific ROC and PR curves for GBM and MLP-ANN (colors denote treatment classes). Bottom row: macro-averaged ROC (left) and PR (right) curves comparing all models; the grey dashed line indicates random performance. 0, outpatients; 1, surgical services; 2, internal services; 3, intensive care units; NM-LR, multinomial logistic regression; RF, random forest; SVM, support vector machine; AUC-ROC, area under the receiver operating characteristic curve; AUC-PR, area under the precision-recall curve.
Based on the overall AUC-ROC and AUC-PR values, the models were ranked from highest to lowest performance as follows: GBM, MLP-ANN, SVM, MN-LR, and RF. The GBM model, which achieved the highest AUC values, yielded 26,748 true negatives (TN), 6,748 true positives (TP), 3,252 false negatives (FN), and 3,252 false positives (FP). In comparison, the MLP-ANN model correctly classified 26,588 TN and 6,588 TP, while misclassifying 3,412 FN and 3,412 FP. These results indicate that although both models performed well, GBM had slightly better overall classification accuracy due to fewer false predictions.
Decision-boundary plots highlighted contrasting architectures: GBM formed sharp, piecewise-linear splits characteristic of tree models ideal for tabular data, whereas MLP-ANN produced smoother, continuous transitions that capture complex nonlinear patterns – explaining its aptitude for high-dimensional feature spaces.
The confusion matrices and decision boundary plots for both GBM and MLP-ANN models are presented in Figure 3. Corresponding visualizations for the other models are provided in Supplementary Material 2.

Confusion matrix and decision boundary plots. Confusion matrices and decision boundaries for GBM and MLP-ANN. Left column (confusion matrix): True-vs-predicted counts across four treatment classes. Deep diagonal blues indicate correct classifications; off-diagonals reveal residual confusion. GBM achieves its best recall on class 0, whereas the ANN favours class 1. Right column (decision boundary): test-set points projected onto the first two PCA components; background shading depicts the multi-class decision surface. The GBM exhibits block-like partitions, while the MLP-ANN displays smoother, radial regions – consistent with tree-based vs. neural-network decision mechanisms. GBM, gradient boosting machine; MLP-ANN, multilayer perceptron – artificial neural network.
Validation results of the models (with external test set)
In accordance with the International Federation of Clinical Chemistry and Laboratory Medicine (IFCC) recommendations, external validation was conducted to ensure the generalizability and robustness of our models [11]. External validation centered on the top performers, GBM and MLP-ANN. Both retained strong results across all care categories (Table 3). GBM yielded F1 scores of 0.92, 0.70, 0.69, 0.77 for types 0–3 (macro-F1 = 0.77) and AUC-ROC/AUC-PR ranges of 0.79–0.95/0.55–0.86, confirming reliable discrimination, particularly for outpatients and ICU admissions.
Similarly, MLP-ANN yielded F1 scores of 0.90, 0.69, 0.67, and 0.74 (macro-F1 = 0.75), with AUC-ROC 0.78–0.93 and AUC-PR 0.53–0.83. GBM remained marginally superior, especially in ICU classification.
Both models generalized well to the external cohort, yet prospective multi-center studies are required before routine clinical use. Confusion matrices for GBM and MLP-ANN in Supplementary Material 2 confirm balanced performance across outcome categories.
Comparison of interpretability performance metrics
In both of the top-performing models – GBM and MLP-ANN – EOS count emerged as the most influential predictor, as illustrated in the feature importance rankings. For the GBM model, EOS was followed by BASO, age, and HCT, whereas in the MLP-ANN model, HCT, HGB, and WBC count were the next most important variables. Notably, the widely studied hemogram-derived ratios NLR, PLR and LMR – were ranked only as moderate to low contributors in both models.
Although eosinophil count is not typically regarded as a key clinical parameter for determining hospitalization needs, its high relative importance in both models suggests a statistically significant – though potentially non-causal – association with treatment classification outcomes. This highlights the ability of machine learning models to uncover patterns that may not be immediately apparent in traditional clinical assessments.
Figure 4 presents the feature importance plots and SHAP graphs, which collectively demonstrate both the magnitude and direction of each variable’s influence on individual predictions. The contrast between the GBM and ANN models in their handling of feature interactions reflects their differing algorithmic structures – GBM’s tree-based logic vs. ANN’s capacity for capturing nonlinear relationships.

Feature importance and SHAP plots global and local feature importance for GBM vs. MLP-ANN. The upper bar plot contrasts global importance scores (aggregate SHAP values; GBM blue, ANN orange). Bottom violin/“bee-swarm” summaries depict local SHAP distributions and sign (blue=low, pink=high) for each model. Features are ordered by model-specific importance. SHAP, shapley additive explanations; GBM, gradient boosting machine; MLP-ANN, multilayer perceptron – artificial neural network; WBC, white blood cell; NEU, neutrophil count; LYM, lymphocyte count; MONO, monocyte count; EOS, eosinophil count; BASO, basophil count; PLT, platelet count; HGB, hemoglobin concentration; RDW, red cell distribution width; MCV, mean corpuscular volume; MPV, mean platelet volume; PDW, platelet distribution width; HCT, hematocrit; NLR, neutrophil-lymphocyte ratio; PLR, platelet-lymphocyte ratio; LMR, lymphocyte-monocyte ratio.
Furthermore, the Matthews Correlation Coefficient (MCC) values were calculated as 0.57 for GBM and 0.55 for MLP-ANN, indicating a moderate level of classification performance for both models.
Discussion
Using complete blood counts from a randomly sampled derivation cohort of 40,000 patients, this single-centre retrospective study trained several machine-learning models. GBM and MLP-ANN performed best, achieving an internal F1 of 0.68 (Matthews=0.56) and class-specific AUC-ROC values of 0.79–0.95 across outpatient, surgical, medical and ICU categories. External validation in 6,132 additional cases confirmed comparable performance (GBM F1=0.92, 0.70, 0.69, 0.77). SHAP highlighted eosinophil and basophil counts, revealing a graded eosinophil decline from ambulatory care to ICU – supporting eosinopenia as a severity proxy.
Admission decisions still rely heavily on physician judgment, especially when disease-specific criteria are lacking. Given variable ICU-admission guidelines and high patient volumes, objective triage is essential. Machine-learning decision-support can flag patients needing ward or ICU care in real time, reducing diagnostic variability and streamlining workflow.
NLR, PLR, and LMR have been widely recognized for their prognostic utility in assessing disease severity, ICU need, and mortality [12], [13], [14], [15], [16], [17]. Although they are low-cost, rapidly obtained markers of systemic inflammation, their prognostic value has chiefly been shown in narrowly focused ICU or COVID-19 studies. To bridge this gap, we assessed these ratios across the full care continuum – from ambulatory clinics to ICUs – and used explainable machine learning to isolate each marker’s independent influence on admission triage.
A growing body of evidence highlights the prognostic relevance of NLR, PLR and LMR across various clinical contexts. Elevated NLR and LMR have been associated with non-alcoholic fatty liver disease (NAFLD) [4], while higher NLR and PLR levels correlate with poor outcomes in traumatic brain injury and hemorrhagic events [3]. In COVID-19, increased NLR and MLR, alongside reduced PLR, have been linked to disease severity, especially in early stages [18]. Among elderly patients with acute myocardial infarction, elevated NLR, PLR, and lymphocyte counts were associated with mortality [8]. Similar associations have been observed in oral cancer and osteoporosis, where high NLR and PLR relate to tumor progression, and low LMR to impaired anti-tumor immunity [19], 20].
Our study revealed an unexpectedly high feature importance for EOS levels – a novel finding in the existing literature. Although eosinophils play key roles in immune and allergic responses [21], they are rarely viewed as admission or ICU indicators. Stratified 5-fold cross-validation confirmed the signal (mean accuracy 89.9 ± 1.9 %), indicating minimal overfitting.
Eosinophil levels varied significantly by treatment type – highest in outpatients, lower in ward patients, and modestly increased in ICU cases. Despite this rebound, ICU levels remained well below outpatient values. The high variability in ICU eosinophil counts (SD=1.18) may reflect diverse clinical states. This trend supports previous findings linking eosinopenia to critical illness [22]. Unlike clinical reasoning, machine learning models such as GBM and MLP-ANN derive decisions from statistical patterns in the data rather than predefined clinical priorities [22]. In this framework, eosinophil (EOS) emerged as the most influential variable in both models – not necessarily due to its clinical primacy, but because its distribution provided strong discriminatory power within the dataset. It is plausible that EOS functioned as a proxy for unmeasured clinical variables, gaining prominence due to the absence of stronger predictors. As such, statistical importance does not equate to clinical relevance and should be interpreted with caution, ideally in conjunction with expert clinical judgment.
Eosinophils are immune-regulating white blood cells, whose production is controlled by cytokines such as GM-CSF, IL-3, and IL-5 [23], 24]. Their levels typically decline during acute inflammation and normalize during recovery. Acute stress or events like organ infarction may further suppress eosinophils via apoptosis mediated through the adrenocortical pathway [25]. Eosinopenia has been linked to increased mortality in several reports [26], [27], [28] or example, Korkmaz et al. found significantly lower eosinophil levels among non-survivors of cardiopulmonary arrest, identifying eosinopenia as a strong mortality predictor surpassing other laboratory markers in multivariate analysis [26].
In a large ICU study on COPD patients, Salturk et al. found that eosinophil levels >2 % were linked to lower morbidity and mortality, while lower levels were associated with worse clinical parameters and higher mortality, especially in ventilated patients [27]. Similarly, eosinopenia has been proposed as a mortality marker in critically ill pediatric patients [28].
Among non-cardiac vascular surgery patients, 90-day mortality was significantly higher in the eosinopenic group (OR: 1.97; 95 % confidence interval (CI): 1.42–2.73; p<0.001) [29]. Shivani et al. also observed an association between low eosinophil counts and increased mortality in surgical ICU patients [30]. Additionally, tumor-associated eosinophils have been linked to favorable prognosis, suggesting potential relevance in oncologic settings [31].
Eosinopenia is a common feature in critically ill patients and correlates with greater disease severity and mortality. This is largely attributed to stress-induced corticosteroid release and pro-inflammatory cytokines suppressing eosinophil production during acute illness [32]. In contrast, higher eosinophil levels in outpatients may reflect the absence of systemic inflammation or hypothalamic–pituitary–adrenal axis suppression [33].
Interestingly, a modest rebound in eosinophil counts was observed in some ICU patients, potentially indicating immunologic recovery or corticosteroid tapering. Such rebounds have been associated with improved outcomes in conditions like COVID-19 [34], 35].
Basophil counts peaked on internal-medicine wards, were intermediate in surgical wards and ICUs, and were lowest in outpatients – consistent with subacute or chronic inflammation (e.g., tissue repair, late-phase hypersensitivity). Their prominence in the GBM model therefore likely flags inpatient-versus-outpatient status rather than critical illness per se, underscoring the need to read feature rankings in clinical context.
In summary, eosinophil counts fell progressively from outpatients to ICUs – evidence of inflammation-driven immunosuppression – before a slight ICU rebound that may signal early recovery or treatment effects. By contrast, higher basophil counts on medical and surgical wards point to sustained or complex inflammation, mirroring earlier reports and meriting further study.
Besides eosinophils and basophils, other predictors – chiefly HCT and HGB – held strong SHAP weights, especially in the MLP-ANN. ICU patients’ lower HCT likely reflects hemodilution or inflammation-driven anemia, both typical in critical illness.
PDW, although seldom highlighted clinically, ranked highly in the GBM model, implying that platelet-size variability captures subclinical inflammatory or thrombotic activity relevant to triage. MCV and RDW contributed modest, consistent signals, likely reflecting nutritional or marrow status.
While age and sex demonstrated statistically significant differences across groups, their relative importance in SHAP was lower than expected. This finding reinforces that even significant demographic predictors may be eclipsed by dynamic laboratory markers when integrated into complex multivariate models.
The Matthews Correlation Coefficient (MCC) offers a balanced evaluation of model performance by combining accuracy, sensitivity, specificity, and F1-score – especially useful for imbalanced datasets. In our study, the GBM and ANN models achieved MCC values of 0.56 and 0.54, respectively, with GBM showing a slight edge. Despite class imbalances, both models demonstrated strong generalization, and MCC proved more informative than single metrics like accuracy or F1-score [36]. Therefore, the use of MCC in our study provides a more comprehensive reflection of the models’ classification performance.
Strengths and limitations
A key strength of this study lies in its large, heterogeneous patient population, drawn from a tertiary-level training and research hospital situated in a densely populated urban center with high patient turnover. This robust sample enhances the statistical power and relevance of the findings. Furthermore, the comparative evaluation of multiple machine learning algorithms provides a multifaceted understanding of predictive performance across diverse modeling approaches.
Nevertheless, the retrospective and single-center design may limit the generalizability of the results. To enhance external validity and promote clinical translation, future studies should adopt prospective, multicenter designs. Comparative assessments that integrate machine learning predictions with expert clinical judgment across varied clinical scenarios may also improve real-world applicability. Additionally, combining hemogram-derived ratios with other biochemical and clinical variables could yield more comprehensive and accurate prognostic models.
Conclusions
Accessible hemogram-derived ratios, particularly the eosinophil count, offer economical prognostic markers for broad clinical risk stratification; their incorporation into routine algorithms may streamline early admission decisions and resource allocation, yet prospective multi-center studies are needed to confirm their triage utility.
Acknowledgments
The authors would like to express their sincere gratitude to Prof. Dr. Ayfer ÇOLAK for her invaluable academic guidance throughout the study. We also thank Prof. Dr. Savaş YAKAN, Chief Physician of the hospital, for his administrative support and facilitation of the research process.
-
Research ethics: Ethical approval for the study was obtained from the Ethics Committee of SBU Izmir Tepecik Training and Research Hospital prior to study initiation (Resolution No. 2025/02-23, dated March 10, 2025). The study was conducted in accordance with the Declaration of Helsinki and complied with the ethical standards of the country in which it was carried out.
-
Informed consent: Not applicable.
-
Author contributions: All authors contributed substantially to the conception, design, analysis, and writing of this manuscript. All authors reviewed and approved the final version of the manuscript.
-
Use of Large Language Models, AI and Machine Learning Tools: Only English editing (Grammarly).
-
Conflict of interest: The authors declare that there is no conflict of interest regarding the publication of this manuscript. All authors have disclosed any relationships or activities that could be perceived as potential conflicts of interest – whether professional, financial, or involving direct or indirect benefits – according to the relevant editorial policies.
-
Research funding: This research received no specific grant from any funding agency in the public, commercial, or not-for-profit sectors.
-
Data availability: Not applicable.
References
1. Bernstein, LH, David, G, Rucinski, J, Coifman, RR. Converting hematology based data into an inferential interpretation. In: Hematology - science and practice [Internet]. Rijeka, Crotia: InTech; 2012. Available from: http://www.intechopen.com/books/hematology-science-and-practice/converting-hematology-based-data-to-useful-informationSearch in Google Scholar
2. Santo, SS, Marco, D, Giuseppe, C, Maria, F, Giulia, T, Davide, C. The NLR, PLR, LMR ratios in venous thromboembolism (VTE): a study on patients affected by VTE with active cancer compared to N0on-VTE. Angiol Vasc Surg 2024;9:1–5.10.24966/AVS-7397/100120Search in Google Scholar
3. Wang, T, Yang, Z, Zhou, B, Chen, Y. Relationship between NLR and PLR ratios and the occurrence and prognosis of progressive hemorrhagic injury in patients with traumatic brain injury. J Invest Surg 2025;38:1–8. https://doi.org/10.1080/08941939.2025.2470453.Search in Google Scholar PubMed
4. Yang, Y, He, X, Tan, S, Qu, X, Huang, W, Cai, J, et al.. The association between immunoinflammatory biomarkers NLR, PLR, LMR and nonalcoholic fatty liver disease: a systematic review and meta-analysis. Clin Exp Med 2025;25:39. https://doi.org/10.1007/s10238-024-01539-1.Search in Google Scholar PubMed PubMed Central
5. Tok, D, Bakır, F. Hemogram parameters and hemogram-derived ratios in covid-19 severity. Clin Med Health Res J 2023;3:626–31.10.18535/cmhrj.v3i5.262Search in Google Scholar
6. Colaneri, M, Genovese, C, Fassio, F, Canuti, M, Giacomelli, A, Ridolfo, AL, et al.. Prognostic significance of NLR and PLR in COVID-19: a multi-cohort validation study. Infect Dis Ther 2024;13:1147–57. https://doi.org/10.1007/s40121-024-00967-6.Search in Google Scholar PubMed PubMed Central
7. Chernysh, NY, Kulakevich, MV, YuI, Z. Modern hemogram parameters in the diagnosis of infectious pathology. Russ J Pers Med 2023;3:66–76.10.18705/2782-3806-2023-3-4-66-76Search in Google Scholar
8. Chen, Y, Chen, S, Han, Y, Xu, Q, Zhao, X. Neutrophil-to-lymphocyte ratio and platelet-to-lymphocyte ratio are important indicators for predicting in-hospital death in elderly AMI patients. J Inflamm Res 2023;16:2051–61. https://doi.org/10.2147/jir.s411086.Search in Google Scholar
9. Peng, H, Su, Y, Luo, J, Ding, N. Association between admission hemoglobin level and prognosis in sepsis patients based on a critical care database. Sci Rep 2024;14:5212. https://doi.org/10.1038/s41598-024-55954-1.Search in Google Scholar PubMed PubMed Central
10. JetBrains Team. Learn IDE features [Internet]; 2025. Available from: https://www.jetbrains.com/help/pycharm/feature-trainer.html [cited 23 Feb 2025].Search in Google Scholar
11. Master, SR, Badrick, TC, Bietenbeck, A, Haymond, S. Machine learning in laboratory medicine: recommendations of the IFCC working group. Clin Chem 2023;69:690–8. https://doi.org/10.1093/clinchem/hvad055.Search in Google Scholar PubMed PubMed Central
12. Asaduzzaman, MD, Bhuia, MR, Alam, ZN, Bari, MZJ, Ferdousi, T. Role of hemogram-derived ratios in predicting intensive care unit admission in COVID-19 patients: a multicenter study. IJID Reg 2022;3:234–41. https://doi.org/10.1016/j.ijregi.2022.04.011.Search in Google Scholar PubMed PubMed Central
13. Velazquez, S, Madurga, R, Castellano, JM, Rodriguez-Pascual, J, de Aguiar Diaz Obregon, SR, Jimeno, S, et al.. Hemogram-derived ratios as prognostic markers of ICU admission in COVID-19. BMC Emerg Med 2021;21:89. https://doi.org/10.1186/s12873-021-00480-w.Search in Google Scholar PubMed PubMed Central
14. Ince, FM, Alkan Bilik, O, Ince, H. Evaluating mortality predictors in COVID-19 intensive care unit patients: insights into age, procalcitonin, neutrophil-to-lymphocyte ratio, platelet-to-lymphocyte ratio, and ferritin lactate index. Diagnostics 2024;14:684. https://doi.org/10.3390/diagnostics14070684.Search in Google Scholar PubMed PubMed Central
15. Asan, A, Üstündağ, Y, Koca, N, Şimşek, A, Sayan, HE, Parıldar, H, et al.. Do initial hematologic indices predict the severity of COVID-19 patients? Turk J Med Sci 2021;51:39–44. https://doi.org/10.3906/sag-2007-97.Search in Google Scholar PubMed PubMed Central
16. Duran, M, Uludağ, Ö. Can platelet count and mean platelet volume and red cell distribution width be used as a prognostic factor for mortality in intensive care unit? Cureus 2020;12:e11630. https://doi.org/10.7759/cureus.11630.Search in Google Scholar PubMed PubMed Central
17. Saurabh, A, Dey, B, Raphael, V, Barman, B, Dev, P, Tiewsoh, I, et al.. Evaluation of hematological parameters in predicting intensive care unit admission in COVID-19 patients. SN Compr Clin Med 2022;4:39. https://doi.org/10.1007/s42399-021-01115-8.Search in Google Scholar PubMed PubMed Central
18. Purnamasidhi, CAW, Riandika, IGVC, Sukmawati, NMDD, Gayatri, AAAY, Utama, IMS, Somia, IKA, et al.. The relationship between neutrophil to lymphocyte ratio (NLR), platelet to lymphocyte ratio (PLR), and monocyte to lymphocyte ratio (MLR) with the clinical severity level of COVID-19 patients at Udayana University Hospital. Res Soc Dev 2024;13:e8413946864. https://doi.org/10.33448/rsd-v13i9.46864.Search in Google Scholar
19. Zhang, XJ, Zhao, H, Zhang, D, Ye, DD, Qin, J, Zhou, YR, et al.. Blood cell count-derived inflammation indices as predictors of the osteoporotic risk of postmenopausal women. Eur Rev Med Pharmacol Sci 2024;28:2207–16. https://doi.org/10.26355/eurrev-202403-35725.Search in Google Scholar
20. Govindasamy Sukumar, SP, Pethagounder Thangavelu, R, Rajaram Mohan, K, Reddy Gowda Venkatesha, R, Appusamy, K. The role of neutrophils-lymphocytes ratio (NLR), lymphocytes-monocytes ratio (LMR), platelets-lymphocytes ratio (PLR) in oral cancer and acute dental infections. Int Arab J Dent 2024;15:15–27. https://doi.org/10.70174/iajd.v15i2.1029.Search in Google Scholar
21. Ghaffari, S, Rezaei, N. Eosinophils in the tumor microenvironment: implications for cancer immunotherapy. J Transl Med 2023;21:551. https://doi.org/10.1186/s12967-023-04418-7.Search in Google Scholar PubMed PubMed Central
22. Yip, CB, Ho, KM. Eosinopenia as a predictor of unexpected re-admission and mortality after intensive care unit discharge. Anaesth Intensive Care 2013;41:231–41. https://doi.org/10.1177/0310057x1304100130.Search in Google Scholar
23. Bass, DA. Behavior of eosinophil leukocytes in acute inflammation. I. Lack of dependence on adrenal function. J Clin Invest 1975;55:1229–36. https://doi.org/10.1172/jci108041.Search in Google Scholar PubMed PubMed Central
24. Bass, DA, Gonwa, TA, Szejda, P, Cousart, MS, DeChatelet, LR, McCall, CE. Eosinopenia of acute infection. J Clin Invest 1980;65:1265–71. https://doi.org/10.1172/jci109789.Search in Google Scholar
25. Cayrol, C, Girard, J. Interleukin-33 (IL-33): a nuclear cytokine from the IL-1 family. Immunol Rev 2018;281:154–68. https://doi.org/10.1111/imr.12619.Search in Google Scholar PubMed
26. Korkmaz, İ, Tekin, YK, Tekin, G, Demirtaş, E, Yurtbay, S, Nur, N. Eosinophil cell count predicts mortality in the intensive care unit after return of spontaneous circulation. Rambam Maimonides Med J 2022;13:e0001. https://doi.org/10.5041/rmmj.10458.Search in Google Scholar PubMed PubMed Central
27. Salturk, C, Karakurt, Z, Adiguzel, N, Kargin, F, Sari, R, Celik, ME, et al.. Does eosinophilic COPD exacerbation have a better patient outcome than non-eosinophilic in the intensive care unit? Int J Chronic Obstr Pulm Dis 2015;10:1837–46. https://doi.org/10.2147/COPD.S88058.Search in Google Scholar PubMed PubMed Central
28. Kim, YH, Park, HB, Kim, MJ, Kim, HS, Lee, HS, Han, YK, et al.. Prognostic usefulness of eosinopenia in the pediatric intensive care unit. J Kor Med Sci 2013;28:114–9. https://doi.org/10.3346/jkms.2013.28.1.114.Search in Google Scholar PubMed PubMed Central
29. von Meijenfeldt, GCI, Chary, S, van der Laan, MJ, Zeebregts, CJAM, Christopher, KB. Eosinopenia and post-hospital outcomes in critically ill non-cardiac vascular surgery patients. Nutr Metabol Cardiovasc Dis 2019;29:847–55. https://doi.org/10.1016/j.numecd.2019.05.061.Search in Google Scholar PubMed
30. Shivani, AA, Prabhu, VV, Kulkarni, SH. Can eosinopenia be used as an indicator to predict mortality in surgical intensive care unit. Sch J. Appl Med Sci 2015;3:1320–2.Search in Google Scholar
31. Proffer, SL, Guo, R, Demer, AM, Peters, MS. Eosinophils in metastatic melanoma. Hum Pathol 2023;141:110–7. https://doi.org/10.1016/j.humpath.2023.08.001.Search in Google Scholar PubMed
32. Tanni, F, Akker, E, Zaman, MM, Figueroa, N, Tharian, B, Hupart, KH. Eosinopenia and COVID-19. J Osteopath Med 2020;120:504–8. https://doi.org/10.7556/jaoa.2020.091.Search in Google Scholar PubMed
33. Huang, R, Xie, L, He, J, Dong, H, Liu, T. Association between the peripheral blood eosinophil counts and COVID-19. Medicine (Baltim) 2021;100:e26047. https://doi.org/10.1097/md.0000000000026047.Search in Google Scholar
34. Mateos González, M, Sierra Gonzalo, E, Casado Lopez, I, Arnalich Fernández, F, Beato Pérez, JL, Monge Monge, D, et al.. The prognostic value of eosinophil recovery in COVID-19: a multicentre, retrospective cohort study on patients hospitalised in Spanish hospitals. J Clin Med 2021;10:305. https://doi.org/10.3390/jcm10020305.Search in Google Scholar PubMed PubMed Central
35. Yan, PD, Markovic, D, Hixson, RY, Shover, CM, Buhr, RG, Salehi-Rad, R, et al.. Eosinophil recovery in hospitalized COVID-19 patients is associated with lower rates of ICU admission and in-hospital mortality: an observational cohort analysis. Respir Med Res 2023;84:101031. https://doi.org/10.1016/j.resmer.2023.101031.Search in Google Scholar PubMed PubMed Central
36. Chicco, D, Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genomics 2020;21:6. https://doi.org/10.1186/s12864-019-6413-7.Search in Google Scholar PubMed PubMed Central
Supplementary Material
This article contains supplementary material (https://doi.org/10.1515/tjb-2025-0200).
© 2025 the author(s), published by De Gruyter, Berlin/Boston
This work is licensed under the Creative Commons Attribution 4.0 International License.