Introduction

Since the beginning of twenty-first century, several viruses belonging to the family of coronaviruses that cause serious impact on the health of human being have been emerging. Severe acute respiratory syndrome (SARS) and Middle East respiratory syndrome (MERS) belonging to the family of coronaviruses have been reported in 2003 and 2015, respectively. In December 2019, the Wuhan city in South china have been reported with a series of unidentified pneumonia cases. Most of these cases are associated with the Wholesale wet market which is famous for selling large variety of fresh meat and sea food. On 7 January 2020, Centers for Disease Control declared the virus as novel coronavirus pneumonia after making an examination of the samples gathered from the throat swab of the infected patients [1]. As the genome structure of novel coronavirus is identical to the structure of SARS-CoV, it was renamed as Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) by the International Committee on Taxonomy of Viruses [2, 3], and the disease was renamed on 11 February 2020 as COVID-19 by the World Health organization (WHO) [4]. On 11 March 2020, when the number of cases has raised thirteen time all over the world apart from China then the WHO declared the COVID-19 disease as pandemic. The clinical symptoms of COVID-19 infected patient are identical to the signs of illness caused by Influenza, bacterial pneumonia and respiratory syncytial virus (RSV). The symptoms like fever, dry cough, sore throat, shortness of breath, diarrhea and lung infiltrates are the most common signs of illness observed in coronavirus infected patients. In the beginning, the spread of novel COVID-19 is related with recent travel history to china. Later on, the increase in number of cases occurred because of person-to-person contact through respiratory droplets [5]. Therefore, the necessary measure in preventing the transmission of disease is the immediate diagnosis of COVID-19 infected patients. Presently, RT-PCR is the standard screening technique utilized for the detection of COVID-19 infection [6, 7]. Due to less sensitivity of RT-PCR in the identification of COVID-19 at the initial stage and unavailability of RT-PCR kits, computed radiography images are used as an alternative approach for the diagnosis of COVID-19 infection. Several clinical analyses reveal that most of the COVID-19 infected patients experience the symptom of lung infection [8]. Even though computed tomography (CT) images are more accurate in the diagnosis of diseases associated with lungs, chest X-ray images are widely used. This is mainly due to the availability of more number of X-ray devices and also because of its low cost and less radiation when compared to CT scans.

From the past few years, the usage of computer vision and deep learning algorithms for the automatic identification of different diseases and lesions and for the analysis of biomedical images has produced sufficient results [9]. These techniques have been effectively used in the determination of types of tumor [10, 11], classification of electrocardiogram (ECG) [12], detection of skin cancer [13], etc. In recent years, due to the substantial increase in the number of pneumonia cases it becomes one of the hot topic areas in the research. As deep learning techniques can be utilized with low-cost imaging techniques, the usage of deep learning methods on chest X-ray images is gaining popularity. Togaçar et al. [14] have used convolutional neural network (CNN) as feature extractor for the detection of pneumonia in chest X-ray images. Initially, they performed reduction in number of deep features in each deep model, such as AlexNet, VGG-16 and VGG-19. Later, they performed classification by using different approaches, such as decision tree, K-nearest neighbors, linear discriminant analysis, linear regression, and support vector machine (SVM). The acquired outcomes reveal that the deep features generate coherent features for the identification of pneumonia in X-ray images. A unique deep learning architecture for the categorization of child pneumonia images was proposed by Liang and Zeng [15]. The proposed model utilizes residual structure and dilated convolution to avoid the problem of depth model, such as degradation and overfitting. The development of prediction models for the detection of post-stroke pneumonia using deep learning techniques has been suggested by Ge et al. [16]. The temporal series of information available in electronic health records have been utilized by implementing classical methods and methods depending on the concepts of multiple layer perceptron (MLP) and recurrent neural networks (RNN). From the results, it can be deduced that the suggested model achieves enhanced performance when analyzed with the performance of traditional machine learning approaches.

The main intention of this paper is to improve the novel COVID-19 detection by using Deep learning architectures, such as VGG16 and MobileNet-V2 with ADAM and RMSprop optimizers. Further, we have applied techniques, such as image preprocessing, augmentation of data and transfer learning techniques to avoid the overfitting problem. Finally, to assess the performance of the models, the proposed models have been evaluated on chest X-ray image dataset.

Related Work

From the last few months, COVID-19 is rapidly spreading across different regions of the world. To inhibit the spread of the novel virus, it is required to identify the COVID-19 positive cases as soon as possible. As the signs of coronavirus infection are related to pneumonia, it can be revealed through the diagnosis of medical imaging datasets. Chest X-ray images are extensively used by the medical practitioners in the identification of COVID-19 disease because of the availability of sufficient number of X-ray devices. Several studies have been conducted on the diagnosis of pneumonia and other viral diseases using chest X-ray images [17,18,19,20]. As deep learning techniques are efficiently utilized in the screening, segmentation and classification of medical images [21], these techniques are used as alternative to the standard RT-PCR for the identification of COVID-19 infected images.

To assist clinical experts in effective identification of COVID-19 disease, an efficient detection method known as Cascaded-SEMEnet was developed by Lv et al. [22]. The proposed model consists of SEME-ResNet50 and SEME-DenseNet169. Vast input sizes and SE-structure are adopted by both cascaded network of the proposed model. Further, for the enhancement of data it utilizes MOEx and Histogram equalization techniques. Initially, SEME-ResNet50 was used for the screening of chest X-ray images and for identification of chest X-ray images as normal, viral pneumonia a bacteria class. Afterwards, SEME-DenseNet169 was utilized for further fine categorization of viral pneumonia to determine whether it has been generated by COVID-19. While training the SEME-DenseNet169, U-Net is used for preprocessing the data to remove the impact of non-pathological characteristics on the network. Moreover, from the outcomes it can be noticed that the suggested model produces an accuracy of 85.6% in deciding the category of pneumonia infection and accuracy of 97.1% in the fine categorization of COVID-19. Further, authors have utilized Grad-CAM for anticipating the judgment which provides assistance to the physicians in understanding the chest X-ray image for the efficient diagnosis of COVID-19 disease. Rahimzadeh et al. [23] have trained various deep convolutional paradigms for the efficient categorization of chest X-ray images as normal, pneumonia and COVID-19 infected classes using two publicly available datasets. These datasets consist of 180 X-ray images of COVID-19 infected patients. They also used some training approaches that assist the network models to perform better even if it consists of unbalanced datasets. Further, a neural network framework that combines both Xception and ResNet50 models has been proposed. The multiple features obtained from the two network models helps in achieving better accuracy results in the categorization of chest X-ray images. To reveal the original accuracy attainable in real situations, the proposed model was tested on 11,302 images. The results specify that the proposed model attains an average accuracy of 99.50% for identifying the COVID-19 infected images and an overall accuracy of 91.4% for the identification of all classes. To reduce the limitation of inadequacy of trained practitioners in remote villages, a novel approach for the automatic identification of COVID-19 infection from the chest X-ray images was proposed by Ozturk et al. [24]. The proposed model was mainly designed for the precise categorization of X-ray images as two class (COVID-19 and normal images) and multiple class (COVID-19, Pneumonia and Normal) images. The proposed model was implemented using 17 convolutional layers and by using various filtering techniques in every layer of the model. The experimental outcomes reveal that the suggested approach achieves 98.08% and 87.02% for the categorization of X-ray images into two class and multi-class images. The YOLO (you only look once) real-time detection system was developed by using Darknet approach as classifier. Moreover, the proposed model can be implemented to provide help to the radiologist in ascertaining the initial screening. It also provides prompt screening of the patients by implementing the proposed model via cloud. To overcome the challenges of unbiased learning, various random oversampling and weighted class loss function techniques have been suggested by Punn et al. [25]. These approaches provide fine-tuned transfer learning in different pre-trained deep learning models to execute classification of posteroanterior X-ray image into two classes, i.e., normal and COVID-19 infected images and multi-class images, i.e., normal, pneumonia and COVID-19 infected images. The effectiveness of the model is assessed by using evaluation metrics, such as accuracy, loss, recall, precision and area under curve (AUC). From the experimental results, it can be noticed that the effectiveness of the models mainly relies on the scenario. Among the available models, NASNetLarge achieves better score. Moreover, the authors also compared the performance of NASNetLarge with other newly proposed applications to show its efficiency in classification of X-ray radiography images. The following Table 1 represents the other research works carried on the diagnosis of coronavirus infection from chest X-ray radiographs using deep learning approaches.

Table 1 Diagnosis of novel coronavirus infection from X-ray radiographs by utilizing approaches of deep learning

Methodology

From the past few years, deep learning frameworks have been extensively utilized for the diagnosis of pneumonia [32, 33]. In this paper, we have chosen VGG16 and MobileNetV2 for the automatic diagnosis of the COVID infected images from the other pneumonia X-ray images using ADAM and RMSprop optimizer.

VGG16 Architecture

The architecture of VGG (Visual Geometry Group) was proposed by Simonyan and Zisserman in 2014 to win the competition of ILVSR (ImageNet) [34]. The model VGG16 belongs to CNN framework. As it consists of 16 layers, this model was named as VGG 16. The VGG16 model comprises of layers, such as Convolutional, Max Pooling, Activation and Fully connected layers. This framework has been trained on ImageNet dataset that consists of 14 million images and 1000 classes and attains top-5 accuracy of 92.7%.

It outperforms AlexNet network by substituting the large filters of size 11 × 5 used in the first and second convolution layers with small filters of size 3 × 3. Therefore, VGG16 is the modified version of AlexNet. This network is large even by modern standards. It has around 138 million parameters. Most of the parameters are used in the fully connected layers. It has a total memory of 96 MB per image for only forward propagation. The block diagram of VGG-16 architecture is mentioned in Fig. 1.

Fig. 1
figure 1

Block diagram of VGG16 architecture

MobileNetV2 Architecture

MobileNetV2 is issued as a component of TensorFlow-Slim Image Classification Library. MobileNetV2 is a convolutional neural network builds upon the ideas from MobileNetV1 [35]. The architecture of MobileNetV2 consists of 54 layers that takes as input an image of size 224 × 224. The main feature of this architecture is that it utilizes depth wise separable convolution as the effective building blocks of the model. Due to this feature, MobileNet-V2 makes use of two 1D convolutions with two kernels rather than operating on 2D convolution with one kernel. Moreover, it also adds two new features to the framework, such as usage of linear bottlenecks between the layers and shortcut connections between the bottlenecks.

As depicted in Fig. 2, the MobileNet-V2 consists of two types of blocks. One block k with stride1is the residual block and the other block with stride 2 is used for downsizing. Both the blocks consist of three layers. The first layer consists of 1 × 1 convolution with ReLU. Depth wise convolution of 3 × 3 is used in the second layer. The third layer again makes use of 1 × 1 convolution that is linear. The third layer does not make use of ReLU. If the third layer uses ReLU again, then the deep networks can be used only as linear classifier on the nonzero part of the output.

Fig. 2
figure 2

Convolutional residual blocks of MobileNet-V2

Each layer in the MobileNet-V2 is represented as series of one or more identical layers that are repeated for n number of times. The number of output channels is same in all the layers that have the same series. Stride s is used only in the first layer of each series, while all other layers make use of stride 1. The MobileNet-V2 makes use of 3 × 3 kernels in all the spatial convolutions. MobileNetV1 requires 1600 k memory space but MobileNetV2 requires only 400 k memory space (1/4 of MobileNetV1). So, the model has taken the MobileNet-V2 for data training. The overall architecture of the MobileNet-V2 is depicted in Fig. 3.

Fig. 3
figure 3

Block diagram of MobileNet-V2 architecture

The directed acyclic hypergraph ‘G’ generated by TensorFlow comprises of edges and nodes which are used to describe the operations and tensors in between computations. The total number of tensors that needs to be stored in memory can be reduced by the computations. In general, it explores all computational orders Σ(G) and selects the one that minimizes.

$$M(G) = \min_{\pi \in \sum (G)} \max_{i \in 1, \ldots m)} \left[ {\sum\limits_{A \in R(i,\pi ,G)} {\left| A \right|} } \right] + size(\pi_{i} )$$
(1)

In Eq. 1, the list of intermediate tensors linked to any of πi... πn nodes is denoted by R(i, π, G), size of the tensor A is denoted by |A| and the total amount of memory required for internal storage during each operation i is denoted by size (πi).

The amount of memory required is denoted by the maximum total size of the integrated inputs and outputs over each operation as shown in Eq. 2

$$M(G) = \max_{op \in G} \left[ {\sum\limits_{{A \in op_{out} }} {\left| {\text{A}} \right|} + \sum\limits_{{B \in op_{out} }} {\left| {\text{B}} \right| + \left| {op} \right|} } \right]$$
(2)

Bottleneck Residual Block

Equation 3 shows a bottleneck block operator \({\mathcal{F}}(x)\) expresses as a composition of three operator \({\mathcal{F}}(x) = [A \circ {\mathcal{N}} \circ B]x\), where A if linear transform A: \({\mathcal{R}}^{s \times s \times k} \to {\mathcal{R}}^{s \times s \times n} ,{\mathcal{N}}\) is nonlinear transform \({\mathcal{N}}:{\mathcal{R}}^{s \times s \times k} \to {\mathcal{R}}^{{s^{\prime } \times s\prime \times n}}\), B is also linear transform(o/p) B: \({\mathcal{R}}^{{s^{\prime } \times s^{\prime } \times n}} \to {\mathcal{R}}^{{s^{\prime } \times s^{\prime } \times k^{\prime } }}\). For this network \({\mathcal{N}} = ({\text{Relu}} \circ {\text{dwise}} \circ {\text{Relu}})\). The memory space needed to calculate F(X) can be as small as \(\left| {s^{2} k} \right| + \left| {s^{\prime 2} k^{\prime } } \right| + {\mathcal{O}}(\max (s^{2} ,s^{\prime 2} )).\)

The basic phenomenon of the algorithm mainly depends on the factor that the inner tensor denoted by I can be expressed as sequence of t tensors each of size n/t. The function used in bottle residual block can be represented as shown in Eq. 3

$${\mathcal{F}}(x) = \sum\limits_{t = 1}^{t} {(A_{i} \circ N \circ B_{i} )(x)}$$
(3)

Optimization Algorithm

Both VGG16 and MobileNet-V2 makes use of ADAM and RMSprop as optimizers for the identification of the COVID infected X-ray images from other pneumonia images. The algorithms for ADAM [36] and RMSprop [37] are given below.

figure a
figure b

Experimental Set up

The experiment has carried out on: Acer Predator PH315-51 Helios 300; CPU: Intel® Core™ i7 + 8th Gen -8750H @ 2.20 GHz 2.21 GHz; RAM: 8.00 GB; GPU: NVIDIA® GEFORCE® GTX 1060; Optane Memory: 16 GB; OS: Windows 10 Home, Anaconda Distribution package jupyter notebook and Python programming language along with tensorflow, keras, Numpy, matplotlib and OpenCV. Fig. 4 represents the in-detail steps in the proposed methodology.

Fig. 4
figure 4

Block diagram of the proposed approach

Description of Dataset

In this study, we considered the image dataset from the Kaggle COVID-19 X-ray dataset [38], organized into dataset folder and contains subfolders for each image category (COVID-19/Normal). This dataset comprises of 624 X-ray images of which 390 are COVID-19 X-Ray images and 234 are Normal X-ray images. Figure 5 depicts the examples of X-ray images of patients with COVID-19 infection and normal X-ray images. Figure 5a displays COVID-19 infected X-ray images with opacities in the lobes of the lungs, while Fig. 5b displays normal X-ray images with no abnormal opacification zones.

Fig. 5
figure 5

a. COVID infected X-ray radiographs with abnormal opacities in the lobes of the lungs. b. Normal X-ray images with no abnormal opacification zones

Data Preprocessing

To enhance the training process, the input images are pre-processed utilizing various preprocessing approaches. The main reason for preprocessing the images is to enhance the visual capacity of the input image by increasing contrast, by deleting low or high frequencies in the image and by reducing the noise present in the original image. In this paper, we applied image resizing and image normalization preprocessing techniques for enhancing the quality of image.

Image Resizing

The dimensions of the images in this dataset are of different ranges. Therefore, to achieve fixed dimensions of all images, the original images are resized to the size of 224 × 224 pixels to fit to the input image size in VGG16 and MobileNetV2 pretrained models.

Image Normalization

In this paper, we applied intensity normalization as preprocessing technique which is widely used in image processing applications [39]. The intensity values of all the images from 0 to 255 are normalized to the basic normal distribution by utilizing min–max normalization as shown in Eq. 15. As this technique removes bias and achieves uniform distribution, it assists in accelerating the convergence of the model.

$$x_{norm} = \frac{{x - x_{\min } }}{{x_{\max } - x_{\min } }}$$
(15)

where x is the pixel intensity. xmin and xmax are minimum and maximum intensity values of the input image in Eq. 15.

Approach of Data Augmentation

Data augmentation is a technique used by the practitioners to significantly enhance the diversity of available data available for training models, without obtaining any new data. To train the large neural networks, techniques of data augmentation, such as cropping, padding and horizontal flipping, are used. However, most of the approaches use only primitive type of augmentation for training the large neural networks. As the main focus of the neural network architecture is on analyzing the depth, less attention has been given in developing strong type of data augmentation techniques. In our research, we have used the following parameter values as shown in Table 2 in data augmentation technique to avoid the data overfitting.

Table 2 Parameters used in Data Augmentation approach

Data Splitting

In this paper, further image dataset is divided into 80:20 ratio where 80% of the image dataset is used for training the model and 20% of the image dataset is used for validating the model.

Transfer Learning

The process of replacing random initialization of weights with the existing convolutional neural network models that were trained on huge image datasets is known as transfer learning. Transfer learning can be performed in two ways. The first way is concerned with removing the fixed features and the whole classifier is trained by considering data from the intermediate layer. The second way is concerned with fine-tuning of the image dataset. In this, new classes are used to replace the last fully connected layer and then the entire convolutional neural network model is retrained with the new alterations. Further, the weights in the deeper layer are updated using these alterations [2, 13]. In this paper, we performed transfer learning by using the net weights of the VGG16 and MobileNetV2 models to replace the last fully connected layer. Further, we added the global average pooling and required final two class labels.

Result and Discussions

We validated our proposed models on 390 COVID and 234 normal Chest X-ray images using ADAM and RMSprop optimizers. It is also significant that transfer learning techniques have been used in the training process improve the learning ability of models. We performed transfer learning by using the net weights of the VGG16 and MobileNetV2 models to substitute the weights of the last fully connected layer. Further, we added the global average pooling and required final two class labels.

Performance Metrics Used for the Interpretation of Image Data

The efficiency of any machine learning model can be determined using performance metrics, such as F1_score, Recall, Precision, Specificity, Confusion Matrix, Mean and Standard Deviation. In this study, the efficiency of the proposed approach is also evaluated using these methods based on confusion matrix. The equations for Recall, Fall_out, Specificity, F1_score and Accuracy are specified in Eqs. 16 to 20

$${\text{Re}} call = \frac{TP}{P} = \frac{TP}{{TP + FN}}$$
(16)
$$Fall\_out = \frac{FP}{N} = \frac{FP}{{FP + TN}}$$
(17)
$$Specificity = \frac{TN}{N} = \frac{TN}{{FP + TN}} = 1 - Fall\_out$$
(18)
$$F1\_score = \frac{2 \times TP}{{2 \times TP + FP + FN}}$$
(19)
$$Accuracy = \frac{TP + TN}{{TP + FP + TN + FN}}$$
(20)

where True Positive, False Positive, True Negative and False Negative of dataset are denoted by TP, FP, TN and FN, respectively.

Classification Report

Table 3 highlights the detailed comparative analysis and best scores of the VGG16 and MobileNetV2 models using ADAM and RMSprop optimizers for the identification of COVID-19 infected images by means of discussed evaluation metrics. From Table 3, it can be observed that MobileNet-V2 using ADAM and RMSprop optimizer attains better performance in terms of precision, recall, and F1_score when compared with VGG 16 precision, recall, and F1_score for identification of COVID-19 infected chest X-ray images. Furthermore, MobileNet-V2 using ADAM and RMSprop achieves an overall accuracy of 92.13% and 92.91%, respectively, when compared with accuracy of VGG16 using ADAM and RMSprop optimizer. It can be also observed that MobileNetV2 using RMSprop and ADAM achieves better accuracy in identification of COVID and normal chest X-ray images when compared with the accuracy of VGG16 using ADAM and RMSprop optimizer.

Table 3 Classification report of VGG 16 and MobileNet-V2 with optimizer using ADAM and RMSprop

Confusion Matrix with Heat Map

Figures 6,7,8,9 represent the confusion matrix of VGG16 and MobileNet-V2 with ADAM and RMSprop optimizers that illustrates the detailed description of what happens with images after image classification.

Fig. 6
figure 6

Heat map representing the confusion matrix of VGG16 model using ADAM optimizer

Fig. 7
figure 7

Heat map depicting the confusion matrix of VGG16 model utilizing RMSprop optimizer

Fig. 8
figure 8

Heat map representing the confusion matrix of MobileNet-V2 model utilizing ADAM optimizer

Fig. 9
figure 9

Heat map representing the confusion matrix of MobileNet-V2 using RMSprop optimizer

From Fig. 6, it can be observed that the VGG16 model using ADAM optimizer was capable of recognizing 74 images as COVID infected images, while 6 images belonging to COVID class are wrongly identified as normal images. In normal class, 41 images were correctly identified as normal images, but 6 images were wrongly identified as COVID images. In Fig. 7, VGG 16 using RMSprop optimizer in COVID class was able to recognize 78 images correctly as COVID infected images, but 2 images were wrongly identified as normal images. While 37 images were identified as normal images and 10 were wrongly determined as COVID infected in Normal class. From Fig. 8, it can be observed that MobileNet-V2 with ADAM optimizer can diagnose 73 X-ray images correctly as COVID infected images in COVID class, but 7 images were wrongly diagnosed as normal images. While MobileNet-V2 with ADAM optimizer in normal class was able to recognize 44 images as normal images and 3 images as COVID-19 infected images. It can be observed from Fig. 9 that MobileNet-V2 with RMSprop optimizer was able to analyze 72 X-ray images correctly as COVID class images, while 8 images were wrongly identified as normal images. In normal class, MobileNet-V2 was able to diagnose 46 images as normal images and only one image was wrongly identified as COVID infected image. The accuracy, sensitivity and specificity of the VGG16 and MobileNet-V2 using ADAM and RMSprop optimizer are represented in Table 4.

Table 4 Analysis of accuracy, sensitivity and specificity of VGG 16 and MobileNet-v2 using ADAM and RMSprop optimizer

It can be also observed from Table 4 that MobileNet-V2 with ADAM optimizer and RMSprop optimizers achieves accuracy of 92.13% and 92.91%, sensitivity of 91.25% and 90% and specificity of 93.62% and 97.87%, respectively, when compared with VGG16 using ADAM (accuracy is 90.55%, sensitivity is 92.50%, specificity is 87.23%) and VGG16 using RMSprop optimizer (accuracy is 90.05%, sensitivity is 97.50%, specificity is 78.72%). Moreover, the results achieved in this study are analyzed with the outcomes obtained using these models in other studies to identify COVID-19 from X-ray images as shown in Table 5.

Table 5 Comparative study of the efficiency of the proposed methodology with the performance of same approaches in other studies

It can be noticed from Table 5 that our proposed approach achieves better accuracy and precision when compared with the same models in other recent studies. In the proposed work, the visual capacity of the image dataset has been enhanced by applying image resizing and normalization as data preprocessing techniques. Then, data augmentation technique is applied to avoid the overfitting of the model. Further, transfer learning results in saving training time of the module. Because of these techniques, the proposed work achieves better accuracy when compared with other previous studies.

Performance Testing of Four Different Method

This section is used to represent and discuss the classification results of chest X-ray images. The classification results are represented using train curve and validation curve. From the training dataset, the train curve is calculated which represents how appropriately the prototype is learning, while the validation curve which is determined from the hold-out validation dataset is used for representing how properly the model is generalizing. Figures 10,11,12,13 represent the training and validation curve of VGG16 with ADAM optimizer, VGG16 with RMSprop optimizer, MobileNetV2 with ADAM optimizer, MobilenetV2 with RMSprop optimizer.

Fig. 10
figure 10

Training and Validation curve of VGG16 with ADAM optimizer

Fig. 11
figure 11

Training and Validation curve of VGG16 with RMSprop optimizer

Fig. 12
figure 12

Training and Validation curve of MobileNet-V2 with ADAM optimizer

Fig. 13
figure 13

Training and Validation curve of MobileNet-V2 with RMSprop optimizer

Training and Validation Curve of VGG16 with ADAM Optimizer

Fig. 10 represents the accuracy and loss curve of VGG16 using ADAM optimizer. From the figure, it can be noticed that the train data accuracy curve is gradually raising from epoch 1 to epoch 22 and then it converges to a value around 90.98%. Similarly, the validation accuracy curve is also gradually raising from epoch 1 to epoch 25 and reaches a value of 90.55% for the epoch 25. It is also noticed that in the train data loss curve is gradually decreasing from epoch 1 to last epoch and reaches and the loss is equal to 0.26 for the epoch 25. The same is noticed for the validation data loss curve and the loss value for the last 25 epoch is 0.23.

Training and Validation Curve of VGG16 with RMSprop Optimizer

From Fig. 11, it can be observed that the train and test data accuracy curve of VGG 16 using RMSprop increases from epoch 1 to last epoch and reaches a value of 90.3% and 90.5% for the last epoch 25 of train and test data. In the same way, for the loss curve that represents train and validation data, the loss value gradually decreases from epoch 1 to last epoch and reaches a value of 0.25 and 0.24 for the last epoch of train and test data, respectively.

Training and Validation Curve of MobileNet-V2 with ADAM Optimizer

As shown in Fig. 12, the training and validation accuracy curve of the MobileNet-V2 increases from epoch 1 to 23 and then converge to a value around 95.5% and 92.1%, respectively, for the train and validation data. For the loss curve, the loss value gradually decreases from epoch 1 to epoch 25 and reaches a value of 0.11 for the train data and 0.18 for the test data.

Training and Validation curve of MobileNet-V2 with RMSprop optimizer

In Fig. 13, the accuracy curve increases gradually from epoch 1 to epoch 19 and then for the last epoch it reaches a value of 94.39% and 92.91% for the train and test data of MobileNet-V2 using RMSprop. In the case of loss curve, the value decreases from epoch 1 to epoch 25 and for the lost epoch the loss value is 0.16 for the train data and 0.20 for the test data.

Conclusion

The COVID-19 infection induced by SARS-CoV2 is one of the recent viruses that loomed in late December 2019 in the Wuhan city of South China. This virus causes pneumonia that causes infection to the lung air sacs. RT-PCR is the basic approach used in the identification of COVID-19 infection. The alternative approach that is used in the efficient detection of infection is computed radiography images as RT-PCR is less sensitive for the determination of novel coronavirus at the initial stage. Therefore, in this paper, we proposed VGG16 and Mobilenet-V2, which makes use of ADAM and RMSPROP optimizers for the automatic identification of the COVID-19 X-ray images from other X-ray images consisting of pneumonia. Then, the efficiency of the proposed methodology has been enhanced by the application of data preprocessing, augmentation of data and transfer learning approaches which are utilized to enhance the visual quality of the image as well as to overcome the overfitting problem. Precision, recall, f1score, and accuracy are applied to estimate the efficiency of the models. The performance of the proposed models has been validated on a publicly available COVID-19 dataset consisting of 390 COVID-19 and 234 normal X-ray images. From the experimental outcomes, it can be deduced that proposed MobileNet-V2 prototype using ADAM and RMSPROP optimizers achieved accuracy of 92.13% and 92.91%, sensitivity of 91.25% and 90% and specificity of 93.62% and 97.87%, respectively, when compared with the VGG16 using ADAM and RMSPROP optimizers. Though this model had performed incredibly well for image classification, there is no concept of image segmentation and localization. In future, we will look forward to apply Image segmentation using deep learning approaches.