Enhanced convolutional neural network enabled optimized diagnostic model for COVID-19 detection

Aaron Meiyyappan Arul Raj, Sugumar Rajendran, Georgewilliam Sundaram Annie Grace Vimal


Computed tomography (CT) films are used to construct cross-sectional pictures of a particular region of the body by using many x-ray readings that were obtained at various angles. There is a general agreement in the medical community at this time that chest CT is the most accurate approach for identifying COVID-19 disease. It was demonstrated that chest CT had a higher sensitivity than reverse transcription polymerase chain reaction (RT-PCR) for the detection of COVID-19 illness. This article presents gray-level co-occurrence matrix (GLCM) texture feature extraction and convolutional neural network (CNN)-enabled optimized diagnostic model for COVID-19 detection. In this diagnostic model, CT scan images of patients are given as input. Firstly, GLCM algorithm is used to extract texture features from the CT scan images. This feature extraction helps in achieving higher classification accuracy. Classification is performed using CNN. It achieves higher accuracy than the k-nearest neighbors (KNN) algorithm and multi-layer preceptor (MLP). The accuracy of GLCM based CNN is 99%, F1 score is 99% and the recall rate is also 98%. CNN has achieved better results than MLP and KNN algorithms for COVID-19 detection.


Accuracy; Convolutional neural network deep learning; COVID-19 detection; Diagnostic model; Gray-level co-occurrence matrix feature extraction; Precision; Recall

Full Text:


DOI: https://doi.org/10.11591/eei.v13i3.6393


  • There are currently no refbacks.

Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

Bulletin of EEI Stats

Bulletin of Electrical Engineering and Informatics (BEEI)
ISSN: 2089-3191, e-ISSN: 2302-9285
This journal is published by the Institute of Advanced Engineering and Science (IAES) in collaboration with Intelektual Pustaka Media Utama (IPMU).