Abstract
BACKGROUND:
The infection caused by the SARS-CoV-2 (COVID-19) pandemic is a threat to human lives. An early and accurate diagnosis is necessary for treatment.
OBJECTIVE:
The study presents an efficient classification methodology for precise identification of infection caused by COVID-19 using CT and X-ray images.
METHODS:
The depthwise separable convolution-based model of MobileNet V2 was exploited for feature extraction. The features of infection were supplied to the SVM classifier for training which produced accurate classification results.
RESULT:
The accuracies for CT and X-ray images are 99.42% and 98.54% respectively. The MCC score was used to avoid any mislead caused by accuracy and F1 score as it is more mathematically balanced metric. The MCC scores obtained for CT and X-ray were 0.9852 and 0.9657, respectively. The Youden’s index showed a significant improvement of more than 2% for both imaging techniques.
CONCLUSION:
The proposed transfer learning-based approach obtained the best results for all evaluation metrics and produced reliable results for the accurate identification of COVID-19 symptoms. This study can help in reducing the time in diagnosis of the infection.
Introduction
The World Health Organisation (WHO) declared the infection caused by severe acute respiratory syndrome coronavirus (SARS-CoV-2) a global pandemic as the infection spread rapidly throughout the globe. SARS-CoV-2 was renamed COVID-19 by the WHO on Feb 11, 2020 [1]. The presence or absence of protective immunity due to infection or vaccination will affect future transmission and illness severity [2]. The issues caused by COVID-19 are highly complex, which can lead to the severe infections of the lungs and can spread to other organs. The symptoms of COVID-19 take 11–12 days to develop after infection [3, 4]. The timely identification of COVID-19 is much necessary else; it can cause severe issues to humans. The radiographic imaging tools for detection of COVID-19 are Computed Tomography (CT) and X-ray [5]. The computer-based analysis is quite demanding for the analysis of this infection as there is a lot of burdens nowadays on the medical system. In such a situation, computer analysis can save the time and burden of the doctors. The researchers are working across the globe for finding the best feasible solution to accurately detect the COVID-19 infection using computer-based analysis. The infections caused by viral pneumonia infections and COVID-19 are almost comparable [6]. Thus the detection of COVID-19 becomes a tedious task as features of COVID-19 overlaps with the other inflammatory diseases of the lung, including different types of pneumonia. The incorrect detection of COVID-19 may lead to the danger of life. Machine Learning-based methods are detecting the infection quite accurately as proposed in the available literature, but these approaches need handcrafted features for training [7, 8]. The features are extracted using some predefined methods and are then supplied to the classifier for learning [9].
Deep learning-based approaches [10, 11, 12] are gaining considerable attention for computer-based detection of COVID-19 as these approaches are capable of extracting the features on their own and produce more reliable result [13, 14]. Performing the specific task in one domain by using the knowledge acquired from other domain is termed transfer learning [15]. In transfer learning-based approaches, the knowledge is transferred from a source to target for improving the task-specific efficiency of the target [17].
The transfer learning-based methods are categorized into three categories, namely 1) Inductive transfer learning, 2) Transductive transfer learning, and 3) Unsupervised transfer learning. The task of source and target are same in inductive transfer learning while in transductive learning, the task of source and target are different. The unsupervised transfer learning is same as inductive learning, but the target tasks are unsupervised. The machine learning and deep learning approaches which utilized transfer learning showed remarkable improvement in the classification task. These methods reveal the features of the infection, which are not clearly visible in the original images as a better understanding of pathophysiology related to COVID-19 is important to improve the treatment process.
The convolutional neural network (CNN) architectures have shown good efficiency in classifying the infected images [18]. CNN-based binary classifier was used by Asnaoui and Chaw [19] to classify the CT images for detecting the infection. RCNN-based classification technique was proposed by Zreik et al. [20] to classify the coronary artery plaque. Li et al. developed the Cov-Net model to extract the features of COVID-19 and showed a remarkable improvement in the classification of infected images. Melanoma dermoscopy images were classified by Hagerty et al. [21] with remarkable accuracy using the deep learning-based method. VGG-16-based model for identification of lung infection was proposed by Li et al. [22]. Identification of thoracic diseases was made with the incorporation of DenseNet model [23]. An Inception Net-based model was used by Wang et al. [24] to identify abnormalities of COVID-19. An accuracy of 85.20% was achieved by classifying 1065 CT images. Resnet-based transfer learning technique was proposed by Ayrton, who achieved validation accuracy of about 96%. U-Net ++-based model was used by Chen et al. [25] for covid-1p classification, and the accuracy achieved was 98.85%. The pretrained U-net architecture for segmentation of 3D Lung images was used by Zheng et al. [26] and then they applied the DL method for determination of COVID-19 infection. The accuracy obtained by them was 95.9%. A CT analysis method for detection of COVID-19 infection using artificial intelligence was proposed by Gozes et al. [27]. Recently, DCNN-based structure was presented by Linda et al. [28] for detection of COVID-19 from chest X-ray images. The proposed model classified X-ray and CT images with significant increment in the results as compared with other methods. The results were compared using various evaluation metrics. To avoid the ambiguity caused by some metrics, additional metrics are used to confirm the effectiveness of results. Accuracy and F score sometimes misleads the classification judgment [29], so Matthews correlation coefficient (MCC) is used for better understanding [30]. The proposed classifier learns the features with the help of transfer learning and produces remarkable results for both CT as well as X-ray images. The main advantage of the proposed technique is the reduction in training time and requirement of fewer data along with remarkable improvement in image evaluation statistics. Finding a faster method to identify COVID-19 infected images is the driving force behind our endeavour. In order to save medical professionals’ time, it can aid in the automatic detection of COVID-19 images.
Methodology
The CT and X-Ray images in the presented work are classified using a transfer learning-based method because both image modalities are utilised to identify COVID-19 symptoms. In this work, the SVM classifier [29] is utilised for classification. The features of the images were extracted using MobileNet V2 architecture. The SVM classifier efficiently classified the infected images after receiving the derived features. The feature extraction process also utilised a number of different networks for comparison. As a comparison tool, the SVM classifier also received the characteristics that were retrieved by several network designs. Transfer learning was used to expedite the classification of photos that were important from a medical standpoint. The pipeline of the suggested work is depicted in Fig. 2.
Graphical abstract.
Pipeline of the proposed work.
Transfer learning eliminates the requirement to train the classifier from scratch since it makes use of the knowledge amassed by a network on a large training dataset with labels that are readily available [30]. In the field of machine and deep learning, transfer learning is a type of design process. Deep learning models typically extract the features of edges in the earliest layers for applications linked to computer vision. The specific features linked to the particular purpose for which the network is designed are extracted by the intermediary layers. Typically, the last layers in a transfer learning approach are taught. This method has the benefit of saving time while training large networks and removing the requirement for big training datasets. Thus, the expertise acquired by one network to do a particular task is utilised to complete a different task. We trained an SVM classifier for the classification of CT and X-ray images infected by COVID-19 using the knowledge gathered by the MobileNet model and other state of the art models for comparison. The entire transfer learning process is depicted in Fig. 3.
Feature transfer using transfer learning.
Depthwise separable convolution block.
MobileNet V2 architecture for classification.
The next-generation architecture for all-purpose computer vision applications is called MobileNet V2 [31] and was created by Google Brain. Low power and low latency models from the MobileNet family are employed to complete a variety of computer vision use cases. The depthwise separable convolution [33] is the fundamental building element for MobileNet V1 [32]. The linear bottlenecks between the layers and the shortcut connections between the bottleneck blocks are included in MobileNet version 2. The in-between inputs and outputs are encoded by the bottlenecks. The middle layers are responsible for converting lower-level components like pixels to higher-level images like images. Figure 5 depicts the MobileNet V2 bottleneck block that is still in place. Expansion Layer, Depth Wise Convolutional Layer, and Projection Layer are the names of the layers. Depth-wise convolution is incorporated, and each input channel receives a single filter. A pointwise 1
Depth wise convolution with one filter per input channel (input depth) can be written as:
Where
Equation (2) is the sum of depthwise convolution and 1
The SVM classifier receives the MobileNetV2 features that were finely extracted. The prime strength of the SVM is its non-probabilistic nature. It separates data across a decision boundary determined by only a small subset of the data. The subset of data which supports the decision boundary are known as support vectors. The other data does not have any effect in determining the position of the boundary. The SVM problem can be understood as a basic optimization problem in which a decision boundary is evaluated which gives maximum distance between the classes. It first finds the hyperplane that has the maximum margin and this hyperplane is used to predict the class of new data object. The equation of hyperplane is given as:
The polynomial kernel function was used to train the SVM. The data was split in to the ratio of 80:20 for training and testing. The MobileNet V2 extracted the features of COVID-19 from CT as well as X-Ray images. Both types of image features extracted by MobileNet V2 were fed to the SVM classifier for training. The SVM effectively classified the COVID-19 infected images for both modalities, i.e. CT and X-ray. Other five networks were also used to supply the extracted features to the SVM classifier for identification of images. These networks are Shufflenet, Alexnet, Darknet, Resnet and Inception Net V3. The presented work clearly finds the better method among the compared methods for better identification of COVID-19 infected images. SVM was used as a common platform for all networks. Thus better network can be judged based on the results provided by the SVM. According to the pathological studies of the lung on the autoptic tissues for COVID-19 and viral pneumonia, the widespread thrombosis with microangiopathy and vascular angiogenesis are the main discriminators for texture variations which are efficiently captured by the proposed method [22, 27].
X-ray dataset
The dataset used in this paper is obtained from the COVID-19 radiography database of Kaggle [32] which consists of 219 chest X-ray images of COVID-19 positive patients, 1341 images of normal chest X-rays and 1345 images of viral pneumonia infection. In addition to this, 301 images were taken from the GitHub repository of Cohen [33]. To increase the data for training, we augmented the data using data augmentation techniques as the deep learning networks need a large amount of dataset to be trained. The data was rotated at the angles of 30
CT dataset
The CT dataset is acquired from CT scans for COVID-19 classification database of Kaggle [34]. The database consists of images collected from Union Hospital (HUST-UH) and Liyuan hospital (HUST-LH). The database consists of 5705 non-informative CT (NiCT) images, 4001 positive CT (pCT) images and 9979 negative CT (nCT). Figure 6 shows the sample images of COVID-19 infection.
Sample images of COVID (
The dataset was resized according to the input requirement of each network. For Mobilenet V2, Shufflenet and Resnet the images were resized to 224
Evaluation metric comparison of CT images
Evaluation metric comparison of CT images
Evaluation metric comparison of X-Ray images
Evaluation metrics comparison of CT images.
Evaluation metrics comparison of X-Ray images.
This section presents the analysis of the comparative results obtained by the proposed and other existing classifiers models. The models used for comparison are Alexnet, Resnet101, Inception V3, Darknet and ShuffleNet. The results obtained are shown in three categories 1) The comparison of accuracy, F scores and MCC, 2) The parameters whose value should be minimum for achieving good performance and, 3) other parameters for firm confirmation of results obtained. The results are categorised into the above-mentioned categories to make them more suitable for analysis. The accuracy, F scores and MCC are evaluated together because the literature review suggests that accuracy and F scores can sometimes mislead the evaluation of classification results. So MCC is used to confirm the reliability of the results as mathematically it takes all four parameters of confusion matrix in a more balanced manner. The second category is of those parameters which should attain minimum value to confirm the effectiveness of classification. Six more evaluation metrics are used for the firm affirmation of results. Tables 1 and 2 show the accuracy, F scores and MCC values obtained by the classifiers for CT and X-Ray images, respectively. Figures 7 and 8 show the results obtained by other parameters which confirm the robustness of results obtained by the proposed method for both modalities.
The prime focus of this research is to detect the COVID-19 infection through computer-aided analysis. The accurate analysis and detection of COVID-19 infection are quite necessary in the current scenario as the pandemic has placed the whole world in an unprecedented situation by its life-threatening effects. The classifiers segregate the data into COVID (
Quantitative analysis
Accuracy is a measure of accurate identification of the class of a pixel. The evaluation of correctly identified pixels is made using accuracy. Accuracy is computed using the following equation:
The accuracy for both modalities shows an improvement of more than 1.5%. The distribution of data is not taken into account by the accuracy so the F-score [35] is a better parameter for judging the classification efficiency. It is given by the following equation:
and
Where
MCC has shown a remarkable improvement of respectively 2.6% and 3.8% for CT and X-ray images. This signifies the better classification ability of the proposed model as the MCC has achieved the highest score along with the accuracy and F-scores.
For more analysis and a better understanding of results, we used six additional evaluation metrics for examining the obtained results. Precision, Recall, Specificity, FM index, Geometric mean and Youden’s index values are used to compare the results of classification by various methods. Precision [37] which is also known as Positive prediction value (PPV) is the ratio of positive samples which were correctly classified to the total number of samples predicted as positive. It is given by the following equation:
Recall or sensitivity is given as the ratio of positive correctly classified samples to the total number of positive samples. It is also known as true positive rate (TPR) and hit rate. The following equation represents recall:
The recall is the ability of a model to find all relevant cases within a dataset while the ability of a classification model to identify only the relevant data points is termed precision. A good margin improves the values of precision and recall in both image modalities. Specificity, which is also termed inverse recall or true negative rate (TNR) is expressed as the fraction of correctly classified negative samples to the total count of negative samples. Specificity is expressed mathematically as:
Fowlkes-Mallows Index (FM Index) [38] gives a more accurate representation of unrelated data. The higher value of FM Index gives more significant similarity between classified and ground truth data. It is given by the following equation:
The prime goal of the classification process is to improve sensitivity without sacrificing the specificity. This is quite a tough task for imbalanced datasets. Thus Geometric Mean (GM) [39] represents the aggregation of both metrics using the following equation:
The values obtained by the Specificity, FM and GM has shown considerable improvement shown in the figure. The higher values of sensitivity and specificity indicate that the true values of positive and negative classified samples are greater in the results obtained by the classifier. Thus the classifier is said to have predicted the COVID (
Where
The proposed method classified the COVID-19 infected images with good accuracy which is supported by several other evaluation metrics [41, 42, 43, 44]. This method is time efficient and accurate as compared with the several methods. Moreover, the computation cost is less as compared to other methods because of depth wise separable convolution. Lesser number of convolution multiplications tends to save the spatial resolution of the patterns in the images.
The COVID-19 pandemic has caused a large number of deaths across the globe. The early and accurate diagnosis is necessary for the treatment of this infection. The current need in the diagnosis of COVID-19 is to save the precious time of doctors and to reduce the burden on the present medical system. In conclusion, a transfer learning-based approach is presented in this work for accurate identification of COVID-19 in CT and X-ray images. This work proposes a computer-aided analysis of accurate detection of COVID-19 infected images. The features finely extracted by MobileNet V2 model paved the way for this research. The obtained results were evaluated using the well-known evaluation metrics, and to the best of our knowledge, no study till now has not compared their method using such a large number of metrics. The advantage of this comparison is that we can be quite sure about the reliability of our results. MCC was used because sometimes accuracy and F score misleads the prediction of classification as mathematically it is a perfect balance between all four parameters of the confusion matrix. Youden’s Index, which is quite a trustworthy statistical measure for the medical image analysis, shows outstanding improvement, which proves the medical effectiveness of the results obtained by the proposed model. The CT scan and X-ray facilities are common in most medical organisations. Thus the proposed method can help for initial identification of the The infections caused by viral pneumonia infections and COVID-19 are almost comparable disease. We believe that with the increase in the dataset, the effectiveness of the classification process can be further improved.
Funding
The authors did not receive funding to carry out this research.
Availability of data and materials
Not applicable.
Code availability
Not applicable.
Ethics statement
Due to the nature of the study, formal consent and ethical approval were not required.
Footnotes
Conflict of interest
The authors declare that they have no conflict of interest.
