COVID-19 Identification using a Convolution Neural Network Design from Chest X-Ray Images

DOI : 10.17577/IJERTV10IS070209

Download Full-Text PDF Cite this Publication

Text Only Version

COVID-19 Identification using a Convolution Neural Network Design from Chest X-Ray Images

Nanduri. Giri Raghava Vineetp, A. Arokiaraj Jovitp, Yeluguri. Yashwanth Reddy3, Chintala. Revanth4

Department of Information Technology, SRM Institute of Science and Technology, Kattankulathur, Chennai,


Abstract COVID-19, a new severe acute respiratory syndrome (SARS), coronavirus 2 has indeed been announced an illness epidemic by WHO and COVID-19 intensified pandemic has come as a surprise to healthcare delivery systems. Every individual must be tested using basic procedures such as RT- PCR with minimal testing kits, which can be a massive task, particularly because these tests have a long turnaround time and low tolerance. Pneumonia, which can be tested and found, using a chest X-ray, can be established after the COVID-19 infection. In light of the deep recession and scarcity of diagnostic kits in certain regions, chest X-rays are the safest choice for high-risk people in lockdown, since they are more easily obtainable in today's healthcare institutions than waiting long for findings from standard approaches. Researchers of this paper advocate using chest X-ray to prioritize the identification of individuals for improved RT-PCR monitoring, as well as to classify patients with an elevated danger of COVID that had a false negative outcome and should be screened again. Furthermore, researchers of this paper recommend using current computational intelligence technology to analyze COVID-19 patients invariably using chest X-ray snapshots, particularly in conditions where radiologists are not present. A Convolution Neural Network (CNN) algorithm is presented to triage individuals for appropriate research. Our VGG16 model detects COVID-19 infection with 98.73% precision using the publicly accessible COVID chest X-ray data set [1].

Keywords: Deep Learning, Convolution Neural Network (CNN), COVID 19 Detection, Chest X-Ray.


    COVID-19 (a newly identified infectious virus) infection rates have risen unexpectedly, putting a burden on healthcare sector organizations. Many nations' healthcare services are already overloaded. Now that the second wave of the pandemic has begun, staying healthy by wearing masks, sanitizing, and effectively screening contaminated patients are a key step in the battle against COVID-19. It is crucial to incorporate which individuals with high severe viral illness may have a COVID-19 infection so as to make enough use of limited resources. Researchers of this paper recommend that individuals with SARI symptoms undergo a chest X-ray so as to cure COVID-19 infection. Using our model, an X-ray can be divided into either of two groups: NORMAL or COVID.

    X-Ray image has some benefits over traditional diagnostic tests: [11]

    1. X-ray imaging is way additional common and less price than regular diagnostic tests.

    2. There is no need to transport digital X-ray images from one location to another. From the purpose of acquisition to the purpose of examination, the diagnostic method is very fast.

    3. Portable X-ray devices, unlike CT scans, allow research to take place within an isolation chamber, without a need for additional Protective Gear (PPE), In this situation, a finite and precious commodity. It further decreases the likelihood of people contracting an infection while in the health care centers.

      The development of a deep neural network machine – learning model for reasonably precise COVID-19 infection recognition from patients' chest X-ray photographs is the study's biggest contribution. In today's world, the overwhelming bulk of radiographs are viewed by non-radiologists. Furthermore, due to the virus's immediacy, many radiologists may be inexperienced with all of the infection's complications and may lack the expertise needed to make an extremely effective treatment. As a consequence, those at the frontline of this investigation should refer to this machine-driven approach.

      Researchers of this paper want to highlight that the developed model should be used as a screening tool to determine whether an individual with SARI is suitable for a COVID-19 infection examination, not as a substitute for conventional COVID-19 infection screening procedures.

      To aid in the advancement of research on this issue, researchers of this paper have made a development code and professional model openly accessible at Netwotk-Model-VGG16-for-Detection-of-COVID-19-

      from-Chest-X-Ray-Images. However, researchers of this paper noticed that each prototype and this analysis merely portrays our stated position of this fast- changing, based on the limited details currently accessible. If researchers of this paper gain more modern experience and better performance, researchers of this paper will continue to update the concept and study.


      1. Biosensors applications in fighting COVID-19 pandemic[5]

        Biosensors can detect non-polar molecules, which are impossible to detect with other instruments. These sensors have a high level of precision and a quick response time. This technology assisted in identifying the signs of viral infection during COVID-19. For COVID-19 patients, it tracks their breathing rhythm, pulse rate, warmth, and some real-time activity. This technology quickly contacts and advises the healthcare service provider when the patient's symptoms change. It enables simple surveillance of affected individuals without the risk of infection. In the COVID-19 disease outbreak, they have identified 7 big biosensor implementations. Using these biosensors more correctly and productively, COVID-19 virus tests can be carried out more efficiently. This era has revolutionized the healthcare enterprise with the aid of using permitting it to behavior its meant capabilities in actual time. Biosensors could be able to provide more productive and timely care of patients in the event of an illness or pandemic in the future.

      2. Recognition of COVID-19 Inflammation Using Deep Learning in Routine Bloodstream Tests [3]

        The COVID-19 disease outbreak, caused by the SARS- CoV-2 coronavirus, has now expanded to over 200 nations, with over 150 million cases reported (and a significantly larger number of diseased) and almost 3.16 million deaths since its epidemic (till 30, April 2021). Despite its established disadvantages, such as lengthy processing times (3-4 hours), possible reagent shortages, falsified rates as large as 15-20%, and a requirement for approved repositories, and complex facilities, the new gold standardized test for verification of infection is the multiplication (Real-time) adversely affect series of chemical reaction analysis of viral RNA (rRT-PCR).

        Consequently, alternative tests are required, which are faster, cheaper, and more available. They developed two classification techniques of the master learning

        system focused on hematochemical blood glucose levels (i.e. platelet levels, AST, ALT, GGT, ALP and, LDH) extracted from 279 individuals who were examined for the rRT-PCR tests conducted after admission to San Raffaele Hospital (Milan, Italy) with COVID-19 signs in these patients, 177 were optimistic and 102 were negative. Their exactness ranges around 82 and 86% and their resistance is between 92 and 95% corresponding to the benchmark, and they have developed two machine learning approaches for SARS- CoV-2 patients that are both positive and negative.

        They have developed a template Decision Tree for clinicians evaluating COVID-19 supect blood samples as basic decision support (even off-line). This research demonstrated the feasibility and scientifically sound use of blood testing and machine learning for the identification of COVID-19 positive patients as an option to rRT-PCR.

      3. Used multi-goal discovery regarding specifically dependent deep learning models to classify COVID-19 patients from chest computed tomography [12]

    The main purpose of their model is the classification of COVID-19 inflamed chest CT patients in photographs. The use of multi-target difference evolution (MODE) and convolution neural networks (CNNs) for the group of individuals is developed to provide a new deeper understanding of the variant, mostly focused on their being affected by COVID-19. A multifaceted function of health is intended to categorize the inflammatory COVID-19 cases in ways that reflect on vulnerability and uniqueness. Via the MODE algorithm, the hyper- parameters of CNN are tailored. Apprenticeship is given by methods of thought on COVID-19 patients' chest CT pics. Analogies with militant styles like convolution neural network (CNN) systems, adaptive neuro-fuzzy inference systems (ANFIS), and plastic neural networks (ANN) are also made with the Well established type measures.


    1. Description

      Researchers of this paper introduced a convolution neural network model for detecting COVID-19 inflammation in patients' chest X-ray photographs. Moreover, since the epidemic is unique, many pathologists will be unfamiliar with all of the nuances of the outbreak and may lack the expertise needed to make a massively effective assessment. As a result, the ones at the leading edge of this study will use this automatic technique as a reference. The system has a set of algorithms for pre-processing, feature extraction, and

      convolution neural network-based model (VGG16) which can identify whether the image is COVID positive or Normal.

    2. Architecture Diagram

      Figure 3.1 COVID detection using the VGG16 model.

      Figure 3.2 VGG16 Architecture

    3. Proposed Sub Modules

      The numerous modules of the project might be divided into segments.

      1. Datasets: We utilized chest longitudinal X-ray specimens in our study. The only X-ray pictures were of the rear and front (PA). COVID-19 individuals with X- ray pictures were divided into two categories, as were stable people with X-ray exposures. There are uniform sections in the database with 196 object classes, with a total of 392 data sets. The COVID-19 data set is composed of 196 COVID-19 X-rays across various sources[1],[2]. On 31 March 2020, all resources had been accessed. They are constructed from X-ray photographs obtained from different publications, libraries, and other media. Researchers of this paper used a list of stable individuals' chest X-ray photographs from the Kaggle assignment "Chest X-ray Images (Pneumonia)" [8] for this database. Researchers of this paper chose 196 tissue samples at probability sampling from the "natural" X-ray libraries., which correspond to stable individuals, at sporadic intervals.

        This connection was chosen even though it has been extensively cited in reports advocating COVID-19 monitoring systems in X-rays [6], [7].

        These files are all in JPG/JPEG or PNG (Portable Network Graphics) files in the database. The artifacts are also displayed. The training data set has pixel ranges of 249 to 255 pixels, and 4280 to 3520 pixels, a size of

        249 to 255 pixels, resolution of 3520 to 4280 pixels. However, all the files were pre-processed using the image compression technique. The hardware that is used for X-rays is indeed different and often unidentified; the image source [1],[4],[8] gives further information. Illustrations of photographs from the databases are shown in Figure 3.3. Figure 3.3(a) depicts a chest X-ray of a person with coronavirus infection in PA perspective, while Figure 3.3(b) depicts a chest X- ray of a stable individual from the Database.

        1. (b)

        Figure 3.3 Samples from the data set. (a) chest X-ray of an individual with coronavirus inflammation in PA view; (b) chest X-ray of a stable individual from database.

        In ways to construct somewhat more sustainable, unifying framework, image enhancement was used to provide more data sets. Augmenting the classification model is a common technique in the research [9]. The pictures that would be transformed were chosen at random. The affine transformations [10] are the maximum extensively used modifications for data augmentation. Rotation, a shift in distance, height, and magnification were the affine transformations used in this analysis.

      2. Data Pre-processing: the data set contains images with little noise, noise removal was not a necessary pre-processing step. To render model coaching computationally possible, the images within the data set have been resized. The input or target factor standardization technique has the effect of speeding up the training process.

      3. Feature Engineering: One-hot encoding – This method assigns 0 or 1 to several flag columns based on the values in a column. The relationship between the grouped and encoded columns is represented by these binary values.

        This technique converts your categorical information to a numerical format, which makes it less difficult for algorithms to know and permits you to cluster your categorical information while not losing any details.

      4. Feature Extraction steps: The transfer learning concept is used to remove features from X-ray images. To initiate, researchers of this paper chose a CNN design that performed well on the Training data. Second, researchers of this paper pick various CNN architecture combinations that have recently been developed on Image classification. Finally, any invisible layers are omitted in these frameworks, leaving only convolutional and pooling layers. These parameters are accountable for descriptor separation, whereas the absolutely interconnected layers are in charge of characterizing the characteristics and, as a result, the picture. To convert a CNN into a function representation, these surfaces must be removed. In this stage, the updated CNN's new output is a collection of specifications obtained from an input image. We build a semi dataset composed of evaluation metrics derived from each photograph in the fields of data for each CNN setup. We must first compress each picture in order to generate a sub data set towards the CNN's necessary size of the input that we selected. CNN is then fed every different style in the corresponding subset and its function list is collected and stored.

        • VGG16 (Architecture used)

        The cov1 layer receives a 224 x 224 RGB picture with a fixed size as a source. The image is enhanced by a series of fully connected layers, each with a completely narrow visual field: 33 (the shortest length that encompasses the concepts of left/right, up/down, and center). It also employs eleven solution process in one of the environments, which may be a concept of a dimensional combination of the input sources (observed via way of means of non-linearity). For 33 fully connected layers, the convolution pace is approximately one pixel, and the spatial spacing of the convolution layer enter is approximately one pixel, in order to preserve the spatial judgement during convolution. To do temporal pooling, five max-pooling layers are added after some of the completely linked surfaces (now not all the convolution layers are found thru the manner of the method of max-pooling). Path 2 over a 22-pixel frame completes the max-pooling.

        Regarding a stack of convolutional neural networks (with varying intensities in specific architectures), three Fully-Connected (FC) layers are added: The predominant have 4096 networks each, while the 0.33 participates in the thousand-manner ILSVRC class and hence has 1000 streams (one for every class). The last sheet is the soft-max surface. The absolutely wired

        surfaces of all environments are programmed in the same manner.

        The quasi of rectification (ReLU) is present in all neural networks. With the exception of one, none of the channels involve Local Response Regularization (LRN), which does not enhance efficiency on the ILSVRC datasets but improves storage overhead and processing time.

      5. Classification steps: There are three phases of identification: i) model planning, ii) model testing, and

    1. phases i) and ii) replication.

      The options extracted from the extractors conjure every sub-data set. 80% of those sub-datasets are used for coaching, whereas the remaining are used for checking. Additionally, we used data augmentation in dataset's training collection. Figure 3.4 displays the number of texture features for train and analysis.

      Figure 3.4 Data Split According to Class

      1. Training of model: Training the machine is like feeding the data to the suitable model to touch up the test data. During this process, researchers of this paper validate the channel with 80 percent of the sub-data set. Training sets are accustomed to tune and match the models. Testing sets are not touched, as a version ought to now no longer be judged primarily based totally on unseen data.

      2. Testing of model: At this level, researchers of this paper use the rescued classification model to perform a review on the resulting 20% of the sub-data set. The procedure applies a class to each specimen in the sub- dataset. In this step, the metrics are also determined.

      3. Practices 1 and 2 are repeated: the sub-datasets are arbitrarily separated into additional preparation and research sets. Subsets will stand out from the rest according to the genotype used.


    1. TABLE I


      Name of Model

      Accuracy Percentage




    2. Visualization

    Figure 4.1 Model Accuracy plot

    Figure 4.2 Model Loss plot

    Figure 4.3 Confusion matrix

    Figure 4.4 Report on Classification


    In order to choose the right treatment and avoid disease transmission, early diagnosis of recent corona virus- infected patients is crucial. Our results show that CNNs are a concise way of explaining x-ray scanning as routine or pessimistic for the COVID-19 by adding a transfer learning concept to classify these parameters using conventional machine learning techniques. Human evaluation of the proposed method has now been discontinued. As a result, a clinical diagnosis should not be replaced as a more detailed survey would be possible across a wider database. Our research helps in those cases to develop precise, automated, rapid, and utilizing chest X-ray images, a minimal procedure for diagnosing COVID-19 has been developed. We plan to leverage our collection further so that additional X-ray scans of COVID-19 victims are added as quickly and efficiently as possible and the efficiency of the recommended x-ray treatment on further lung problems is validated. We also want to make an unbalanced data set, available for the proposed approach to the analysis.


  1. Cohen, J.P., Morrison, P., Dao, L.: Covid-19 image data collection. arXiv 2003.11597 (2020).

  2. COVID-19 X rays. [Online]. Available:

  3. Detection of COVID-19 Infection from Routine Blood Exams with Machine Learning: A Feasibility Study

    Authors:Davide Brinati, Andrea Campagner, Davide Ferrari, Massimo Locatelli, Giuseppe Banfi & Federico 01 July 2020.

  4. Wang, X., Peng, Y., Lu, L., Lu, Z., Bagheri, M., Summers, R.M.: Chestx-ray8:Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 20972106 (2017)

  5. Shashi Bahl, Mohd Javaid, Ashok Kumar Bagha, Ravi Pratap Singh, Abid Haleem, Raju Vaishya, Rajiv Suman, Biosensors applications in fighting COVID-19 pandemic, Apollo medicine journal, July 29, 2020.

  6. I. D. Apostolopoulos and T. Bessiana, Covid-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks, arXiv preprint arXiv: 2003.11617, 2020.

  7. A. Narin, C. Kaya, and Z. Pamuk, Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks, arXiv preprint arXiv: 2003.10849, 2020.

  8. D. S. Kermany, M. Goldbaum, W. J. Cai, C. C. S. Valentim, H. Y. Liang, S. L. Baxter, A. McKeown, G. Yang, X. K. Wu, F. B. Yan, J. Dong, M. K. Prasadha, J. Pei, M. Y. L. Ting, J. Zhu, C. Li, S. Hewett,

    J. Dong, I. Ziyar, A. Shi, R. Z. Zhang, L. H Zheng, R. Hou, W. Shi, X. Fu, Y. O. Duan, V. A. N. Huu, C. Wen, E. D. Zhang, C. L. Zhang, O.

    L. Li, X. B. Wang, M. A. Singer, X. D. Sun, J. Xu, A. Tafreshi, M. A. Lewis, H. M. Xia, and K. Zhang, Identifying medical diagnoses and treatable diseases by image-based deep learning, Cell, vol. 172, no. 5,pp. 11221131.E9, Feb. 1122.

  9. Y. L. Tian, X. Li, K. F. Wang, and F. Y. Wang, Training and testing object detectors with virtual images, IEEE/CAA J. Autom. Sinica, vol. 5, no. 2, pp. 539546, Mar. 2018.

  10. A. Mikoajczyk and M. Grochowski, Data augmentation for improving deep learning in image classification problem, in Proc. Int. Interdisciplinary PhD Workshop, Swinoujcie, Poland, 2018, pp.117 122.

  11. Arpan Mangal1, Surya Kalia1, Harish Rajagopal2, Krithika Rangarajan3;1,Vinay Namboodiri2;4, Subhashis Banerjee1, and Chetan Arora1, CovidAID: COVID-19 Detection Using Chest X-Ray.

  12. Dilbag Singh,1 Vijay Kumar,2 Vaishali,3 and Manjit Kaur, Classification of COVID-19 patients from chest CT images using multi-objective differential evolutionbased convolutional neural networks.

Leave a Reply