Previous Article in Journal
Enhancing Fingerprint Liveness Detection Accuracy Using Deep Learning: A Comprehensive Study and Novel Approach
 
 
Order Article Reprints
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Biased Deep Learning Methods in Detection of COVID-19 Using CT Images: A Challenge Mounted by Subject-Wise-Split ISFCT Dataset

1
Medical Image and Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan JM76+5M3, Iran
2
Law, Economics, and Data Science Group, Department of Humanities, Social and Political Science, ETH Zurich, 8092 Zurich, Switzerland
3
Institute for Intelligent Systems Research and Innovation, Deakin University, Melbourne, VIC 3125, Australia
4
Department of Radiology, School of Medicine, Isfahan University of Medical Sciences, Isfahan JM76+5M3, Iran
5
St. George’s Hospital, London SW17 0RE, UK
6
School of Engineering, University of British Columbia, Kelowna, BC V1V 1V7, Canada
7
Cyclica Inc., Toronto, ON M5J 1A7, Canada
8
Department of Engineering, Durham University, Durham DH1 3LE, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2023, 9(8), 159; https://doi.org/10.3390/jimaging9080159
Received: 19 June 2023 / Revised: 27 July 2023 / Accepted: 28 July 2023 / Published: 8 August 2023
(This article belongs to the Section Medical Imaging)

Abstract

:
Accurate detection of respiratory system damage including COVID-19 is considered one of the crucial applications of deep learning (DL) models using CT images. However, the main shortcoming of the published works has been unreliable reported accuracy and the lack of repeatability with new datasets, mainly due to slice-wise splits of the data, creating dependency between training and test sets due to shared data across the sets. We introduce a new dataset of CT images (ISFCT Dataset) with labels indicating the subject-wise split to train and test our DL algorithms in an unbiased manner. We also use this dataset to validate the real performance of the published works in a subject-wise data split. Another key feature provides more specific labels (eight characteristic lung features) rather than being limited to COVID-19 and healthy labels. We show that the reported high accuracy of the existing models on current slice-wise splits is not repeatable for subject-wise splits, and distribution differences between data splits are demonstrated using t-distribution stochastic neighbor embedding. We indicate that, by examining subject-wise data splitting, less complicated models show competitive results compared to the exiting complicated models, demonstrating that complex models do not necessarily generate accurate and repeatable results.

1. Introduction

The spread of corona virus disease in 2019 (COVID-19) led to a worldwide pandemic. This condition is caused by severe acute respiratory syndrome corona virus 2 (SARS-CoV-2). The gold standard diagnostic investigation for COVID-19 is the reverse-transcription-polymerase chain reaction (RT-PCR) test. Furthermore, chest computed tomography(CT) has also been accessible to clinicians for diagnosis [1]. The current literature has shown that artificial intelligence (AI) is a rapidly growing technological advancement, actively aiding the field of medical imaging in the fight against COVID-19. The traditional qualitative assessment of chest CT slices is very time-consuming and prone to error. CT data typically contain hundreds of slices that need to be read and assessed individually to reach an accurate diagnosis. Hence, AI-assisted deep learning (DL) techniques are preferred, and most of the recent publications have demonstrated promising outcomes on published datasets [2,3,4,5]. The main concern about the available models and the datasets is related to slice-wise splitting of the data, rather than focusing on subject-wise splits. In subject-wise data splits, all CT slices coming from one subject are allocated only to a test or training set, whereas in slice-wise split, this point is not considered [6]. Current studies of the application of DL in health-related data have shown that slice-wise splits overestimate the accuracy of the models and make them unreliable in clinical diagnostic applications [6,7]. In particular, multiple observations from each subject should be split only in the training set or the test set. Otherwise, complex DL algorithms are powerful enough to detect a confounding relationship between the subject and the diagnostic status. Accordingly, unrealistically high prediction accuracy may be reported, which should be avoided by focusing on subject-wise splitting of the dataset.
Considering the available CT datasets for the diagnosis of COVID-19, many well-known datasets are currently in the literature. Our analyses revealed that most publications have used slice-wise splitting, resulting in overestimated high reported accuracy (Table 1). Although in Table 1, we have presented some papers with subject-wise splits that have reported high performance among their outcomes, we indicate that the designed models in these papers typically use extremely selected and well-processed datasets that are not representative of real-world circumstances. We investigate this issue in more detail in the following sections.
To compare the proposed method with the published works and to show the disadvantage of reporting the results from slice-wise splits, we introduce a new publicly available CT slice dataset (ISFCT Dataset), accompanied by a subject-wise split, demographic data, and ground truth classification (https://zenodo.org/record/7997151, accessed on 2 June 2023). The ground truth is not only limited to COVID-19 and healthy classes, and it benefits from detailed information comprising eight characteristic lung features in the right or left lung for all CT slices from all subjects. Moreover, we applied a diverse range of previously published and new DL models to some of publicity existing datasets (elaborated in Table 1) and the new ISFCT dataset and evaluated the performance of the models using slice-wise and subject-wise strategies, generalization ability, and repeatability. We also evaluated the effect of decreasing or increasing the complexity of the models (for example, number of the layers) on the performance of the models. The utilized and newly designed DL models are freely available for further comparison by researchers (https://zenodo.org/record/7997151, accessed on 2 June 2023). It is expected that this dataset and models will improve reproducibility and make the associated research more discoverable with less need to duplicate efforts.

2. Materials and Methods

A comprehensive range of different algorithms in the literature was applied to the ISFCT dataset, and their performance was compared. Furthermore, two shallow models were developed to show how a simpler architecture could address our and other available datasets.

2.1. Dataset

The ISFCT Dataset was collected by Sepahan Radiology, Isfahan, Iran, and the study was approved by the Isfahan University of Medical Sciences Institutional Review Board (IRB) (IR.MUI.RESEARCH.REC.1399.003) and adheres to the tenets of the Declaration of Helsinki. Written informed consent was obtained from all participants.
A range of characteristic lung features is already known for COVID-19, including ground-glass opacity (F1),peripheral (F2), central (F3), peribronchovascular (F4), consolidation (F5), reversehalo (F6), crazypaving (F7), and atelectasis (F8) (eight subclasses) [25]. These radiological patterns are significant for determining the severity of COVID-19 [13]; however, in previous AI-based studies, such detailed features were all merged in the presence or absence of COVID-19. To address this issue in the ISFCT Dataset, eight characteristic lung features and their existence in the right (R) or left (L) lung were marked for each CT slice. Figure 1 and Table 2 demonstrate a summary and samples of the features. In binary classification, in which only health and COVID-19 are considered, the COVID-19 label is assigned according to presence of F1-R or F1-L, as is prevalent in other published works. According to previous studies, men appear to fare worse, and our findings showed a similar data distribution [26]. A previous study examining a mixture of patients suffering from COVID-19 and SARS observed that there are positive correlations of increasing age, male gender, and severity of disease with mortality, and our dataset agrees [27]. In accordance with our demographic data, the average age of patients infected by COVID-19 is 51 years old with a standard deviation of 14.61, and 57 percent of all patients are men.
The ISFCT Dataset contains data from 178 subjects diagnosed with COVID-19 (43,399 CT slices) and 156 healthy controls (39,767 CT slices). All CT slices have an equal size of 768 × 768 pixels, and each subject has (on average) 260 CT slices. The distribution of scans labeled with each of the eight characteristic lung features is presented in Figure 2. This dataset alsoincludes demographic details, including gender and age.

2.2. Data Preprocessing

Each individual is presented by multiple CT slices, and each CT slice may have one of the lung characteristic features in the left (L) or right (R) lung; the presence of each is considered presence of that feature in the CT slice. Many images contain more than one feature because of the characteristic lung changes during acute and chronic periods of disease [25]. The patterns observed on CT images of COVID-19 are dynamic and can be categorized into four stages. The earlystage spans from 0 to 4 days following the manifestation of the initial symptoms. In this stage, one can observe F1 unilaterally or bilaterally in the lower lobes. The second stage has a duration of 5–8 days and is known as the progressive stage, during which one can detect F1, F7 patterns, and F5. These features are distributed bilaterally in multiple lobes. The third stage, known as the peak, is from the ninth to the 13th day. During this period, dense F5 becomes more predominant. The fourth stage of the disease, absorption, becomes apparent after 14 days, around the time the infection becomes more controlled. At this stage, the F7 pattern and F5 are steadily absorbed. However, F1 remains. These radiological patterns of disease are significant for classification of severity of COVID-19 during assessment of the patients. As a result, some images are placed in different categories. In the supporting materials, a Microsoft Excel file is prepared, and for each patient, number 1 in each column (features) indicates the presence of any feature. The input CT slices are generally scaled to retain compatibility with the network structures. For most of the networks, the input size is fixed at 224 × 224 × 3. However, for some networks, such as EfficientNetB0 and Efficient CovidNet [33], the input dimensions are adjusted to the determined size by the published papers. To standardize the CT slices, the brightness of each image is transferred to the 0–1 interval. The different data split methods utilized in this paper are illustrated in Figure 3.

2.3. Algorithms Based on Transfer Learning

Most DL algorithms have numerous parameters, training of which requires a huge number of labeled data. Transfer learning is introduced to take advantage of such algorithms in applications with small numbers of data. For this purpose, a pre-trained DL algorithm with learned parameters is used, and only a small proportion of the parameters remained trainable with the new dataset (which is not essentially similar to the original data) [10,27]. We used famous pre-trained models (trained on the ImageNet dataset [34], including ResNet50V2, EfficientNetB0, EfficientNetB3,VGG16, EfficientNetB0, and EfficientNetB3 [35]). Additionally, a limited number of published algorithms trained with COVID-19 data (NASNet-based model [9], COVID-NET, and MobileNet) were used as a baseline for our transfer learning approaches.
For transfer learning, the kernels of the CNN layers were initialized with pre-trained ImageNet weights. For the EfficientNetB0, EfficientNetB3, ResNet50V2, VGG16, and VGG19 models, the weights of CNN levels of the models were frozen first, and all fully connected layers were removed and substituted with a new fully connected classifier. The architecture of this classifier was selected empirically. More details are provided in Table 3. For transfer learning on COVID-19-based networks, such as Efficient CovidNet, NASNet, and COVID-Net, the networks were pre-trained on ImageNet and then on provided COVID-19 datasets in their related papers.

2.4. Shallow Algorithms

To show the performance of shallow networks in addressing our and other datasets, we also developed M1- and M2-COVID networks. The M1-COVID network consists of three convolution operations. We adopted max-pooling after each convolution, and a batch-normalization layer was utilized. Rectified linear unit (ReLU) was considered as an activation function, and two dense layers were finally added. The M2-COVID network has similar structure to M1-COVID, with two additional convolutional layers, each followed by batchnormalization and max-pooling. Additionally, more dropouts between layers were added. In terms of hyperparameters, we employed Adam optimization and a batch size of 32 for both networks. The loss function was set to categorical cross-entropy.

2.5. Implementation Details

To train the models, the learning rate was initialized with a value of 0.0001. This value was determined by a grid search and then tuned using a cosine annealing learning rate scheduler [36,37] and a decay factor. In our dataset, the COVID-19 and non-COVID-19 case distributions were almost similar, but a weighting scheme was used to eliminate any bias in predictions, giving more weight to samples from smaller data groups. Softmax was used as an activation function in the last layer of all models. Additionally, the binary cross-entropy as a loss function and Adam as an optimizer showed better performance on examined models and were used to train the models.
For transfer learning models (based on networks pre-trained on ImageNet), the weights of all the layers of the base model were frozen, and the fully connected layers were removed and substituted by newly trained classifiers.
For networks pre-trained on COVID-19, two strategies were selected:
  • For evaluating the performance of these models on our new dataset, the model was trained from scratch;
  • For transfer learning, fully connected layers were substituted and retrained.

2.6. Performance Evaluation Metrics

Model accuracy and loss curves are presented for evaluation of the networks. For M1- and M2-COVID, confusion matrixes are also presented. The following metrics are al so provided to show the performance of the models:
A c c u r a c y = T P + T N T P + F T + T N + F N
S p e c i f i c i t y = T N T N + F P
P r e c i s i o n = T P T P + F P
S e n s i t i v i t y / R e c a l l = T P T P + F N
N P V = T N T N + F N
where true positive (TP) refers to the correctly classified COVID-19 samples; false-positive (FP) is the number of non-COVID-19 samples classified as COVID; true-negative (TN) indicates the non-COVID slices correctly classified as non-COVID; and false-negative (FN) refers to the number of COVID-19 samples classified as non-COVID.

3. Results

3.1. Comparison of Algorithms in Slice-Wise and Subject-Wise Splits

In this section, various models are assessed in two distinct ways using our new ISFCT dataset:
  • In slice-wise setting, the data are randomly divided, and CT slices from the same subject maybe present in both the training and test data;
  • In subject-wise setting, the data are split such that the training and test splits contain slides from different subjects.
The evaluated algorithms include transfer-learning models and shallow algorithms. Table 4 shows the performance in more detail. Based on the results, COVID-Net, M1-COVID, and M2-COVID achieved the best results on the ISFCT dataset with a slice-wise split.
The COVID network based on NASNet also showed good performance. These results indicate that our simple models, M1-COVID and M2-CVOID, have good performance in comparison with investigated transfer learning-based models and multiple COVID-based networks.
It is clear from Table 4 that the performance of the models decreased drastically when the data were split subject-wise. Achieving high accuracy with a subject-wise split remains a challenging task.
To provide a better view regarding the high performance of our two shallow models, in Figure 4, ROC, learning rate, and precision-recall curves with different thresholds for M1-COVID and M2-COVID models in the slice-wise approach are shown.
To provide a visual description of different splitting methods, each CT slice is represented by two features (one point) using t-distributed stochastic neighbor embedding (t-SNE). CT slices pertaining to each subject are depicted in different colors, and 10 individually clusters are found for 10 randomly selected subjects in Figure 5, indicating common characteristics of slices from each subject. Figure 6 provides visual comparison of slice-wise and subject-wise splitting in CT slices. There is no doubt that discrimination of COVID (blue) from healthy (red) is a challenging task. However, as shown with slice-wise splitting (left panel—Figure 6), test images (circle mark) are selected from subjects (clusters), which are learned by the algorithm during the training stage (x mark). In particular, in slice-wise splitting, the leakage of information yields unreliably high accuracy, which is not achieved with a subject-wise split (right panel—Figure 6).

3.2. Average Accuracy of DL Models in Classification of Sub-Class Features

In this section, the performances of DL models in the classification of sub-class features are presented. We conducted a comprehensive examination of the provided eight features in the ISFCT dataset. Our main aim was to evaluate how effectively models could detect individual features, which are considered significant in COVID-19 detection but have received less attention in previous studies. In our evaluation, we assessed model performance for each feature (both right and left) using a multi-classifier approach. This approach determined the presence or absence of the examined feature in each sample. We present the results in Table 5. However, it is important to note that there were an insufficient number of samples available for features 6 and 7 to properly evaluate model performance. As a result, these two features are not included in the table. In general, our investigation revealed that the examined models did not exhibit remarkable performance in accurately identifying the different features present in the provided samples. Therefore, it is crucial for future research to focus on developing models that effectively detect these features. Additionally, further investigation should explore the impact of the presence of these features on detecting the given disease.

4. Discussion

COVID-19 detection papers frequently employ explainable methods, such as Grad-Cam [38], occlusion sensitivity maps [39], and GSInquire [40], to demonstrate whether the detection decision of their proposed model is based on relevant information. These techniques are used to ensure that models are not drawing conclusions based on irrelevant data, which might result in situations in which correct decisions are made for the wrong reasons. However, these tools only show the decision of a proposed model in a specific dataset, and they cannot be interpreted as tools to show generalizability of the models. If a model can find the cause of a disease such as COVID-19 by analyzing CT slices, it should perform well on similar data. We examined a number of publicly available COVID-19 datasets. In some datasets, such as [9], the subject to which each slide belongs to was not determined. We tested M1-COVID on the dataset provided in [9]. The results after only 15 epochs were very promising (accuracy 0.99, precision 0.99, and recall 0.8), offering the conclusion that the data are divided slice-wise and randomly. We also examined the COVIDxV9B dataset provided by [2]. The authors proposed COVID-NET to be applied on this dataset, which is divided into training and test tests in a subject-wise manner. We applied multiple models, such as VGG16, Resnet, DenseNet201, EfficientNetB0, and M2-COVID, on this dataset. All investigated models have shown more than 86% accuracy, demonstrating that the dataset consists of an easy set of images based on these findings. However, when we applied the provided model, COVID-NET, to our dataset, the model performed poorly. We conclude that the COVIDxV9B collection (which has been declared to be a subject wise split) does not reflect all CT imaging data collected in health centers, and almost all of the models have very good performance on it. Although these models provide many achievements and new understandings, they are not applicable in real-world scenarios. We came to the conclusion that models perform effectively for two reasons: they are trained on a biased dataset with the data partitioned slice-wise, or they are trained on datasets that are in an ideal shape and selected with bias. According to these findings, creating models capable of operating on heterogeneous datasets that are similar in content but vary in other features is crucial. Our publicly accessible data on eight sub-classes demonstrate enhanced performance with relatively small networks but not very high performance when subject-wise splitting is considered. We are providing free access to the ISFCT dataset. Moving forward, our work will focus on utilizing and analyzing the dataset using a subject-wise split. This approach will enable us to gather more reliable and accurate findings over time.
It is important to mention that, although the dataset is extensive, it does have its drawbacks. Some features, such as reverse halo and crazy paving, occur rarely, making it difficult to train networks with precision. Moreover, the dataset lacks proper scoring for each feature, creating ambiguity in determining the significance of each feature in classifying a patient as normal or infected. However, there is potential for creating a more comprehensive dataset in the future, enabling a more in-depth analysis of how each feature impacts disease diagnosis and assigning appropriate scores for these features.

5. Conclusions

In this paper, the ISFCT Dataset is introduced to challenge the previously published biased works, which were designed without considering subject-wise splitting. Furthermore, each CT scan in this new dataset has more than only two labels (COVID-19 and healthy). In particular, F1 toF8 are introduced as features in the lung, and their presence in the right or left lungs is labeled.
Our primary focus in this work is to classify each subject-wise split CT scan to a healthy or COVID-19 set (based on the presence of F1-R or F1-L). We examine multiple models on our dataset using two methods of data splitting, slice-wise and subject-wise. With a slice-wise approach, data are randomly split into training and test sets. As a result, CT slices from the same individual are possibly present in both the training and test sets, which is a kind of information leakage and can lead to bias. The examined models include transfer learning using the most successful models and some state-of-the-art COVID-19 detection networks, as well as two shallow models. We show that shallow models perform acceptably well in a slice-wise split, with results comparable to state-of-the-art COVID-19 detection models. However, when the data are divided subject wise (such that the training and test sets comprise CT slices from different subjects), the performance of all developed models decreases drastically. We therefore conclude that the high performance of many published works is only due to slice-wise splits and leakage of the information, rather than the power of the designed networks, which perform similar to our proposed shallow models (M2). Generally, the bias problem in DL models due to incorrect data distribution is difficult to resolve with transfer learning or augmentation [41]. The comparably high performance of the proposed shallow network (M2) demonstrates that a deeper model does not always imply a more unbiased and accurate model. Furthermore, the model’s generalization does not necessarily improve by training models on out-of-field datasets. This finding shows that evaluated models that have shown promising results in previous papers were only customized to perform well with specific datasets and not in most cases, in which the data are not split in a subject-wise manner.

Author Contributions

S.P. and N.S. designed different deep learning models; G.J.S., S.S. (Shamim Shafieyoon), F.H., A.P.Z., S.S. (Samira Soleimany), A.Y., H.N. and P.T. prepared the dataset and the manual annotations; A.S.M., S.A.M.T. and H.R. provided guidance on deep learning models and the design of the study; A.H. and R.K. supervised the project and contributed to the design and modeling. All authors contributed to writing the original draft. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Vice Chancellery for Research and Technology, Isfahan University of Medical Sciences, under Grant No. 198337.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Isfahan University of Medical Sciences, approved by Vice-Chancellor in Research Affairs-Medical University of Isfahan, grant no. IR.MUI.RESEARCH.REC.1399.003, 27 March 2020.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Our dataset containing COVID-19 and normal cases, as well as the demographic details, is available at https://zenodo.org/record/7997151 (accessed on 2 June 2023), and the utilized and newly designed DL models are freely available for further comparison by researchers (https://zenodo.org/record/7997151, accessed on 2 June 2023).

Acknowledgments

The authors would like to thank Hamid Khayyam for his comments about the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shi, F.; Wang, J.; Shi, J.; Wu, Z.; Wang, Q.; Tang, Z.; He, K.; Shi, Y.; Shen, D. Review of artificial intelligence techniques in imaging data acquisition, segmentation, and diagnosis for COVID-19. IEEE Rev. Biomed. Eng. 2020, 14, 4–15. [Google Scholar] [CrossRef]
  2. Pan, F.; Ye, T.; Sun, P.; Gui, S.; Liang, B.; Li, L.; Zheng, D.; Wang, J.; Hesketh, R.L.; Yang, L. Time course of lung changes at chest CT during recovery from coronavirus disease 2019 (COVID-19). Radiology 2020, 295, 715–721. [Google Scholar] [CrossRef] [PubMed]
  3. Loey, M.; Manogaran, G.; Khalifa, N.E.M. A deep transfer learning model with classical data augmentation and CGAN to detect COVID-19 from chest CT radiography digital images. Neural Comput. Appl. 2020, 1–13. [Google Scholar] [CrossRef] [PubMed]
  4. Devnath, L.; Luo, S.; Summons, P.; Wang, D. Performance comparison of deep learning models for black lung detection on chest X-ray radiographs. In Proceedings of the 3rd International Conference on Software Engineering and Information Management, Sydney, Australia, 12–15 January 2020. [Google Scholar]
  5. Rajaraman, S.; Siegelman, J.; Alderson, P.O.; Folio, L.S.; Folio, L.R.; Antani, S.K. Iteratively pruned deep learning ensembles for COVID-19 detection in chest X-rays. IEEE Access 2020, 8, 115041–115050. [Google Scholar] [CrossRef]
  6. Saeb, S.; Lonini, L.; Jayaraman, A.; Mohr, D.C.; Kording, K.P. The need to approximate the use-case in clinical machine learning. Gigascience 2017, 6, gix019. [Google Scholar] [CrossRef]
  7. Chaibub Neto, E.; Pratap, A.; Perumal, T.M.; Tummalacherla, M.; Snyder, P.; Bot, B.M.; Trister, A.D.; Friend, S.H.; Mangravite, L.; Omberg, L. Detecting the impact of subject characteristics on machine learning-based diagnostic applications. NPJ Digit. Med. 2019, 2, 99. [Google Scholar] [CrossRef]
  8. Yang, X.; He, X.; Zhao, J.; Zhang, Y.; Zhang, S.; Xie, P. COVID-CT-dataset: A CT scan dataset about COVID-19. arXiv 2020, arXiv:2003.13865. [Google Scholar]
  9. Ghaderzadeh, M.; Asadi, F.; Jafari, R.; Bashash, D.; Abolghasemi, H.; Aria, M. Deep convolutional neural network–based computer-aided detection system for COVID-19 using multiple lung scans: Design and implementation study. J. Med. Internet Res. 2021, 23, e27468. [Google Scholar] [CrossRef] [PubMed]
  10. Zhao, W.; Jiang, W.; Qiu, X. Deep learning for COVID-19 detection based on CT images. Sci. Rep. 2021, 11, 14353. [Google Scholar] [CrossRef]
  11. Gorbalenya, A.; Baker, S.; Baric, R.; De Groot, R.; Drosten, C.; Gulyaeva, A.; Haagmans, B.; Lauber, C.; Leontovich, A.; Neuman, B. The species Severe acute respiratory syndrome-related coronavirus: Classifying 2019-nCoV and naming it SARS-CoV-2. Nat. Microbiol. 2020, 5, 536–544. [Google Scholar]
  12. Kolesnikov, A.; Beyer, L.; Zhai, X.; Puigcerver, J.; Yung, J.; Gelly, S.; Houlsby, N. Big Transfer (BiT): General Visual Representation Learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  13. Zhang, K.; Liu, X.; Shen, J.; Li, Z.; Sang, Y.; Wu, X.; Zha, Y.; Liang, W.; Wang, C.; Wang, K. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 2020, 181, 1423–1433.e11. [Google Scholar] [CrossRef] [PubMed]
  14. Kassania, S.H.; Kassanib, P.H.; Wesolowskic, M.J.; Schneidera, K.A.; Detersa, R. Automatic detection of coronavirus disease (COVID-19) in X-ray and CT images: A machine learning based approach. Biocybern. Biomed. Eng. 2021, 41, 867–879. [Google Scholar] [CrossRef] [PubMed]
  15. Cohen, J.P.; Morrison, P.; Dao, L.; Roth, K.; Duong, T.Q.; Ghassemi, M. COVID-19 image data collection: Prospective predictions are the future. arXiv 2020, arXiv:2006.11988. [Google Scholar] [CrossRef]
  16. Kaggle, R. Pneumonia Detection Challenge. 2020. Available online: https://www.kaggle.com/c/rsna-pneumonia-detection-challenge (accessed on 18 June 2022).
  17. Jaiswal, A.K.; Tiwari, P.; Rathi, V.K.; Qian, J.; Pandey, H.M.; Albuquerque, V.H.C. Covidpen: A novel COVID-19 detection model using chest x-rays and ct scans. Medrxiv 2020. [Google Scholar] [CrossRef]
  18. Kogilavani, S.; Prabhu, J.; Sandhiya, R.; Kumar, M.S.; Subramaniam, U.; Karthick, A.; Muhibbullah, M.; Imam, S.B.S. COVID-19 detection based on lung ct scan using deep learning techniques. Comput. Math. Methods Med. 2022, 2022, 7672196. [Google Scholar] [CrossRef]
  19. Chouat, I.; Echtioui, A.; Khemakhem, R.; Zouch, W.; Ghorbel, M.; Hamida, A.B. COVID-19 detection in CT and CXR images using deep learning models. Biogerontology 2022, 23, 65–84. [Google Scholar] [CrossRef]
  20. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Al Emadi, N. Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  21. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S. Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef]
  22. Zouch, W.; Sagga, D.; Echtioui, A.; Khemakhem, R.; Ghorbel, M.; Mhiri, C.; Hamida, A.B. Detection of COVID-19 from CT and Chest X-ray Images Using Deep Learning Models. Ann. Biomed. Eng. 2022, 50, 825–835. [Google Scholar] [CrossRef]
  23. Available online: https://github.com/UCSD-AI4H/COVID-CT/tree/master/Images-processed (accessed on 2 June 2023).
  24. Ortiz, A.; Trivedi, A.; Desbiens, J.; Blazes, M.; Robinson, C.; Gupta, S.; Dodhia, R.; Bhatraju, P.K.; Liles, W.C.; Lee, A. Effective deep learning approaches for predicting COVID-19 outcomes from chest computed tomography volumes. Sci. Rep. 2022, 12, 1716. [Google Scholar] [CrossRef]
  25. Misztal, K.; Pocha, A.; Durak-Kozica, M.; Wątor, M.; Kubica-Misztal, A.; Hartel, M. The importance of standardisation–COVID-19 CT & Radiograph Image Data Stock for deep learning purpose. Comput. Biol. Med. 2020, 127, 104092. [Google Scholar] [PubMed]
  26. Pijls, B.G.; Jolani, S.; Atherley, A.; Derckx, R.T.; Dijkstra, J.I.; Franssen, G.H.; Hendriks, S.; Richters, A.; Venemans-Jellema, A.; Zalpuri, S. Demographic risk factors for COVID-19 infection, severity, ICU admission and death: A meta-analysis of 59 studies. BMJ Open 2021, 11, e044640. [Google Scholar] [CrossRef]
  27. Jin, J.-M.; Bai, P.; He, W.; Wu, F.; Liu, X.-F.; Han, D.-M.; Liu, S.; Yang, J.-K. Gender differences in patients with COVID-19: Focus on severity and mortality. Front. Public Health 2020, 8, 152. [Google Scholar] [CrossRef]
  28. Gunderman, R.B. Essential Radiology: Clinical Presentation, Pathophysiology, Imaging, 2nd ed.; Thieme: Stuttgart, Germany, 2006. [Google Scholar]
  29. Koo, H.J.; Lim, S.; Choe, J.; Choi, S.-H.; Sung, H.; Do, K.-H. Radiographic and CT features of viral pneumonia. Radiographics 2018, 38, 719–739. [Google Scholar] [CrossRef]
  30. Hacking, C.; Weerakkody, Y.; Bell, D. Left lower lobe consolidation. Radiopedia 2021, 66157. [Google Scholar] [CrossRef]
  31. Rossi, S.E.; Erasmus, J.J.; Volpacchio, M.; Franquet, T.; Castiglioni, T.; McAdams, H.P. “Crazy-paving” pattern at thin-section CT of the lungs: Radiologic-pathologic overview. Radiographics 2003, 23, 1509–1519. [Google Scholar] [CrossRef]
  32. Chiarenza, A.; Ultimo, L.E.; Falsaperla, D.; Travali, M.; Foti, P.V.; Torrisi, S.E.; Schisano, M.; Mauro, L.A.; Sambataro, G.; Basile, A. Chest imaging using signs, symbols, and naturalistic images: A practical guide for radiologists and non-radiologists. Insights Imaging 2019, 10, 114. [Google Scholar] [CrossRef]
  33. Silva, P.; Luz, E.; Silva, G.; Moreira, G.; Silva, R.; Lucio, D.; Menotti, D. COVID-19 detection in CT images with deep learning: A voting-based scheme and cross-datasets analysis. Inform. Med. Unlocked 2020, 20, 100427. [Google Scholar] [CrossRef] [PubMed]
  34. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  35. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  37. Loshchilov, I.; Hutter, F. Stochastic gradient descent with warm restarts. In Proceedings of the 5th International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016; pp. 1–16. [Google Scholar]
  38. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  39. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In Computer Vision–ECCV 2014: 13th European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 818–833. [Google Scholar]
  40. Qiu Lin, Z.; Javad Shafiee, M.; Bochkarev, S.; St Jules, M.; Wang, X.Y.; Wong, A. Do Explanations Reflect Decisions? A Machine-centric Strategy to Quantify the Performance of Explainability Algorithms. arXiv 2019, arXiv:1910.07387. [Google Scholar]
  41. Cheplygina, V. Cats or CAT scans: Transfer learning from natural or medical image source data sets? Curr. Opin. Biomed. Eng. 2019, 9, 21–27. [Google Scholar] [CrossRef]
Figure 1. Samples of eight characteristic lung features and one healthy case.
Figure 1. Samples of eight characteristic lung features and one healthy case.
Jimaging 09 00159 g001
Figure 2. Distribution of CT slices in each subclasses (eight characteristic lung features).
Figure 2. Distribution of CT slices in each subclasses (eight characteristic lung features).
Jimaging 09 00159 g002
Figure 3. The proposed process for data selection and evaluation.
Figure 3. The proposed process for data selection and evaluation.
Jimaging 09 00159 g003
Figure 4. (a,d) ROC, (b,e) training, and validation learning curves, and (c,f) precision/recall curves for M1-COVID and M2-COVID using various thresholds.
Figure 4. (a,d) ROC, (b,e) training, and validation learning curves, and (c,f) precision/recall curves for M1-COVID and M2-COVID using various thresholds.
Jimaging 09 00159 g004
Figure 5. Visual presentation of common characteristics between the CT slices of each subject. Each CT slice is represented by two features (one point) using t-distributed stochastic neighbor embedding (t-SNE). CT slices pertaining to each subject are depicted in different colors, and 10 individual clusters are found for 10 randomly selected subjects.
Figure 5. Visual presentation of common characteristics between the CT slices of each subject. Each CT slice is represented by two features (one point) using t-distributed stochastic neighbor embedding (t-SNE). CT slices pertaining to each subject are depicted in different colors, and 10 individual clusters are found for 10 randomly selected subjects.
Jimaging 09 00159 g005
Figure 6. Visual comparison of the slice-wise and subject-wise splitting in CT slices. The goal is discrimination of COVID (blue) from healthy (red) data. In slice-wise splitting ((left) panel), test images (circle mark) are selected from the subjects (clusters), which are learned by the algorithm during the training stage (x mark). In particular, in slice-wise splitting, the leakage of information yields unreliably high accuracy, which is not achieved with a subject-wise split ((right) panel).
Figure 6. Visual comparison of the slice-wise and subject-wise splitting in CT slices. The goal is discrimination of COVID (blue) from healthy (red) data. In slice-wise splitting ((left) panel), test images (circle mark) are selected from the subjects (clusters), which are learned by the algorithm during the training stage (x mark). In particular, in slice-wise splitting, the leakage of information yields unreliably high accuracy, which is not achieved with a subject-wise split ((right) panel).
Jimaging 09 00159 g006
Table 1. Summary of the published works on usage of DL in diagnosis of COVID-19 with CT slices (slice-wise = IW, subject-wise = SW, C = COVID-19 and NC = non-COVID-19, N = normal, ND = not determined).
Table 1. Summary of the published works on usage of DL in diagnosis of COVID-19 with CT slices (slice-wise = IW, subject-wise = SW, C = COVID-19 and NC = non-COVID-19, N = normal, ND = not determined).
Published DL WorkDatasetSlice-Wise or Subject-WiseNumber of CT SlicesNumber of SubjectsDL ModelsReported Accuracy
Yang et al.
[8]
COVID-CT dataset
[8]
IWC: 349
NC: 463
C: 216
NC: VD
Transfer-Learning: DenseNet-169, ResNet-50, and contrastive self-supervised transfer learningF1-Score = 0.90
AUC = 0.98
Accuracy = 0.89
Ghaderzadeh et al.
[9]
Ghaderzadeh et al. [9]IWC: 7644
NC: 2509
C: 190
NC: 59
Transfer-Learning: NASNetSensitivity = 0.999
Specificity = 0.986
Accuracy = 0.996
Zhao et. al
[10]
COVIDx CT 2A [11]IWC, SARS-CoV-2 and CP: 194,922
N: ND
C, SARS-CoV-2, and CP: 3475
N: ND
BigTransfer (BIT)
[12]
Accuracy = 0.992
Sensitivity = 0.987
Specificity = 0.995
NPV= 0.996
PPV=0.985
Zhang et al. [13]CC-CCll;
Zhang et al. [13]
IWC: 156,070
viral pneumonia: 159,700
N: 95,459
C: 839
viral pneumonia: 874
N: 758
A 3D classification networksAccuracy = 0.924
Sensitivity = 0.943
Specificity = 0.911
Kassania et al. [14]Cohen et al. [15],
and KaggleRSNAPneumonia Detection dataset [16]
IWC: 20
N: 20
C: ND
N: ND
Examination of transfer learning with a range of methods, including:
DenseNet121, Xception, InceptionV3, DenseNet201, InceptionResNetV2, VGG16, VGG19, NASNETLarge, NASNetMobile, ResNet50v2, ResNet101V2, and ResNet152V2
Accuracy = 0.99
Precision = 0.99
Recall = 0.99
F1-Score= 0.99
Jaiswal et al. [17]COVID CT slices [8]IWC: 746
N: ND
C: ND
Normal: ND
COVIDPEN: Pruned EfficientNet-B0Accuracy = 0.85
AUC = 0.84
F1-Score = 0.86
Kogilavani et al. [18]KaggleIWC: 1958
N: 1915
C: ND
N: ND
VGG16, DenseNet, MobileNet, Xception, EfficientNet, and NASNetAccuracy = 0.97
F1-Score = 0.98
Precision = 0.99
Recall = 1
Chouat et al. [19]GitHub repository [8], and Kaggle [20,21]IWC: 408
N: 325
C: ND
N: ND
VGGNet-19, ResNet50, Xception, and InceptionV3Accuracy = 0.905
F1-Score = 0.905
Precision = 0.915
Recall= 0.903
Zouch et al. [22]Database of CT slices provided in GitHub [23]IWC: 349
NC: 408
C: ND
NC: ND
VGG19 and
ResNet50
Accuracy = 0.98
F1-Score = 100
Precision = 0.993
Recall = 100
Ortiz et al. [24]CC-CCll;
Zhang et al. [13]
SWC: 156,070
viral pneumonia: 159,700
N: 95,459
C: 839
viral pneumonia: 874
N: 758
Inception ResNetV2Accuracy = 0.95
AUC = 0.96
Sensitivity = 0.94
Wang et al. [2]COVIDx [2]SWC: 358 CXR
N: not determined
C: 266
NC: 5538
no pneumonia: 8066
COVID-NetAccuracy = 0.933
Sensitivity = 0.91
PPV = 0.989
Table 2. The explanation of eight characteristic lung features.
Table 2. The explanation of eight characteristic lung features.
FeatureExplanationSymbol
Ground-Glass OpacityThe hazy gray indicates increased density inside the lungs [25].F1
PeripheralThe feature is situated on the edge or periphery of the lung.F2
CentralThe feature is located in the middle of the lung.F3
PeribronchovascularThickening of the interstitial or bronchial wall [28].F4
ConsolidationThe alveolar air spaces are filled with fluid, cells, tissue, or other material [29].F5
Reverse HaloCentral ground-glass opacity surrounded by denser consolidation of a crescent shape or a complete ring of at least 2 mm in thickness [30].F6
Crazy PavingScattered diffuse ground-glass attenuation with superimposed interlobular septal thickening and intralobular lines [31].F7
AtelectasisComplete or partial collapse of the entire lung or area(lobe) of the lung [32].F8
Table 3. Architecture of different transfer learning models (FC: demonstrating how fully connected layers have been developed).
Table 3. Architecture of different transfer learning models (FC: demonstrating how fully connected layers have been developed).
Classification ModelsInput Slice SizeFCInitial Weights
ResNet50v2(224 × 24 × 3)AveragePooling2D
Flatten layer
Dense 256
Drop out 0.5
Dense 2
ImageNet
VGG16(224 × 224 × 3)MaxPooling2D
Flatten
Dense256
Dense 2
ImageNet
EfficientNetB0(224 × 224 × 3)MaxPooling2D
Flatten
Dense256
Dense 2
ImageNet
EfficientNetB3(300 × 300 × 3)MaxPooling2D
Flatten
Dense256
Dense 2
ImageNet
MobileNet(224 × 224 × 3)AveragePooling2D
Flatten layer
Dense 256
Dropout 0.5
Dense 2
ImageNet
COVIDnetwork based on NASNET(224 × 224 × 3)No changeImageNet and Ghaderzadeh et al. [9] COVID-19 dataset
COVID-Net(224 × 224 × 3)Dense 2COVIDx [2]
Table 4. Performance of the examined models based on different measures.
Table 4. Performance of the examined models based on different measures.
ModelsSlice-WiseSubject-Wise
ACCPrecisionRecallSpecificityNPVACCPrecisionRecallSpecificityNPV
Transfer Learning ModelsResNet50V20.840.640.460.760.600.590.350.110.840.55
EfficientNetB00.520.501000.440010.56
EfficientNetB30.500010.500.4800.40.60
VGG160.840.730.380.870.600.620.690.370.790.55
COVID network based on NASNET0.940.930.950.920.950.660.660.670.670.66
MobileNet0.520.520.560.470.500.490.470.400.570.50
COVID-NET0.9910.9910.990.600.730.540.70.50
Shallow
Models
M1-COVID0.990.990.990.9910.580.580.400.560.58
M2-COVID0.9910.9810.980.610.670.400.780.53
Table 5. The average accuracy of different models in sub-classes.
Table 5. The average accuracy of different models in sub-classes.
ModelsF1F2F3F4F5F8
Transfer Learning ModelsResNet50v20.750.570.500.470.570.51
EfficientNetB00.730.500.530.480.510.37
EfficientNetB30.740.400.690.400.480.40
VGG160.740.450.650.460.640.64
COVID network based on NASNET0.800.530.760.560.720.66
MobileNet0.670.530.600.530.510.68
COVIDNET0.750.400.790.610.650.60
Shallow ModelsM1-COVID0.730.500.770.710.740.40
M2-COVID0.500.500.620.580.590.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Parsarad, S.; Saeedizadeh, N.; Soufi, G.J.; Shafieyoon, S.; Hekmatnia, F.; Zarei, A.P.; Soleimany, S.; Yousefi, A.; Nazari, H.; Torabi, P.; et al. Biased Deep Learning Methods in Detection of COVID-19 Using CT Images: A Challenge Mounted by Subject-Wise-Split ISFCT Dataset. J. Imaging 2023, 9, 159. https://doi.org/10.3390/jimaging9080159

AMA Style

Parsarad S, Saeedizadeh N, Soufi GJ, Shafieyoon S, Hekmatnia F, Zarei AP, Soleimany S, Yousefi A, Nazari H, Torabi P, et al. Biased Deep Learning Methods in Detection of COVID-19 Using CT Images: A Challenge Mounted by Subject-Wise-Split ISFCT Dataset. Journal of Imaging. 2023; 9(8):159. https://doi.org/10.3390/jimaging9080159

Chicago/Turabian Style

Parsarad, Shiva, Narges Saeedizadeh, Ghazaleh Jamalipour Soufi, Shamim Shafieyoon, Farzaneh Hekmatnia, Andrew Parviz Zarei, Samira Soleimany, Amir Yousefi, Hengameh Nazari, Pegah Torabi, and et al. 2023. "Biased Deep Learning Methods in Detection of COVID-19 Using CT Images: A Challenge Mounted by Subject-Wise-Split ISFCT Dataset" Journal of Imaging 9, no. 8: 159. https://doi.org/10.3390/jimaging9080159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop