Applied Soft Computing Journal: Gonçalo Marques, Deevyankar Agarwal, Isabel de La Torre Díez
Applied Soft Computing Journal: Gonçalo Marques, Deevyankar Agarwal, Isabel de La Torre Díez
article info a b s t r a c t
Article history: COVID-19 infection was reported in December 2019 at Wuhan, China. This virus critically affects
Received 17 August 2020 several countries such as the USA, Brazil, India and Italy. Numerous research units are working at their
Received in revised form 22 August 2020 higher level of effort to develop novel methods to prevent and control this pandemic scenario. The
Accepted 26 August 2020
main objective of this paper is to propose a medical decision support system using the implementation
Available online 29 August 2020
of a convolutional neural network (CNN). This CNN has been developed using EfficientNet architecture.
Keywords: To the best of the authors’ knowledge, there is no similar study that proposes an automated method
Automated decision support system for COVID-19 diagnosis using EfficientNet. Therefore, the main contribution is to present the results
Convolutional Neural Network (CNN) of a CNN developed using EfficientNet and 10-fold stratified cross-validation. This paper presents two
COVID-19 main experiments. First, the binary classification results using images from COVID-19 patients and
Deep learning normal patients are shown. Second, the multi-class results using images from COVID-19, pneumonia
Machine learning and normal patients are discussed. The results show average accuracy values for binary and multi-
class of 99.62% and 96.70%, respectively. On the one hand, the proposed CNN model using EfficientNet
presents an average recall value of 99.63% and 96.69% concerning binary and multi-class, respectively.
On the other hand, 99.64% is the average precision value reported by binary classification, and 97.54%
is presented in multi-class. Finally, the average F1-score for multi-class is 97.11%, and 99.62% is
presented for binary classification. In conclusion, the proposed architecture can provide an automated
medical diagnostics system to support healthcare specialists for enhanced decision making during this
pandemic scenario.
© 2020 Elsevier B.V. All rights reserved.
https://doi.org/10.1016/j.asoc.2020.106691
1568-4946/© 2020 Elsevier B.V. All rights reserved.
2 G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691
layers [35,36]. CNN’s represents a huge breakthrough in auto- this automated system have been evaluated by radiologists. The
matic image classification systems as there is no need for pre- dataset used includes 1125 images. In total, 125 samples are
processing the images, that was needed in traditional machine used for COVID-19 class, 500 for pneumonia class and for 500
learning algorithms [37–39]. normal class. The authors have used 5-fold cross-validation to
The main objective of this research is to propose a med- validate the performance of the proposed method. The average
ical decision support system using CNNs. To the best of the accuracy of 98.08% and 87.02% is reported for binary and multi-
authors’ knowledge, there is no similar study that proposes an class classification. The limitation on the number of samples used
automated method to the detected COVID in CT X-ray images for COVID-19 class is reported by the authors.
using EfficientNet [40,41]. Therefore, the main contribution of A deep learning model to improve the accuracy of binary
this paper is to present an automated medical diagnosis system classification of COVID-19 is proposed in [44]. The proposed CNN
implemented using EfficientNet. Numerous studies and applica- was implemented based on the VGG-19 classifier. The dataset
tions have been reported in the literature. The authors believe used includes a total of 364 X-ray scans. The model performance
that in the critical pandemic scenario is crucial to share all the has been validated using random sampling. The ratio used for
methods and materials to allow the readers to reproduce the train, validation and testing was 80:20:20. The number of samples
results. The authors share the python scripts developed in the for normal class and COVID-19 was 233 and 115 during training,
Google Colab platform. In this way, the software is accessible to 56 and 32 during validation, and 75 and 34 during testing. The
all the readers who can execute the scripts for future research results show an accuracy of 96.3%. The limited number of samples
activities. This architecture can be used for transfer learning, and for COVID-19 cases is stated by the authors.
it is more efficient than most of its predecessors such as VGG (e.g. Apostolopoulos et al. [45] propose a transfer learning ap-
VGG16 or VGG19), GoogLeNet (e.g. InceptionV3), and Residual proach using VGG-19 and MobileNet v2 for automated detec-
Network (e.g. ResNet50) [42]. The EfficientNet model consists of tion of patients with pneumonia and COVID-19. Two different
8 models from B0 to B7, with each subsequent model number datasets have been included in this study. One dataset of a total
referring to variants with more parameters and higher accuracy. of 1427 samples that include 504, 700 and 224 images of normal,
EfficientNet architecture uses transfer learning to save time and pneumonia and COVID-19, respectively, have been used. On the
computational power. Consequently, it provides higher accuracy other hand, a different dataset of 224 samples for COVID-19,
values than the competitor known models. This is due to the use 714 samples of pneumonia patients and 504 of normal patients
of a clever scaling at depth, width, and resolution. The authors is also included. The 10-fold cross-validation has been used to
have used the B4 model, as it contains 19 m parameters, that evaluate the proposed models. The VGG-19 and MobileNet v2 re-
is feasible for our experimental setup, as B5, B6 and B7 include ported 98.75% and 97.40% of accuracy for binary classification and
30M, 43M and 66M params, respectively [41]. Furthermore, the 93.48% and 92.85% for multi-class concerning the first dataset.
authors have used separated datasets to validate the proposed Furthermore, the MobileNet v2 has been applied in the second
CNN models using images that are not included during the testing dataset presenting an accuracy of 96.78% for binary classification
and training phase. The proposed model has been evaluated using and 94.72% for multiclass. The authors state that a more in-
stratified cross-validation 10-fold stratified. This paper presents depth analysis using more patient data concerning COVID-19
two main experiments including different datasets for testing and individuals is required.
validation to ensure the non-occurrence of overfitting. First, the The authors of [46] proposed a fast screen system for COVID-
binary classification results using images from COVID-19 patients 19 detection based on deep learning neural networks. The pre-
and normal patients are shown. Second, the multi-class results sented method is based on the nCOVnet and uses chest X-rays
using images from COVID-19, pneumonia and normal patients images. The dataset included in this study has a total of 337
are discussed. The source code is provided in this document as samples. In total, 192 of the samples are from COVID-19 positive
supplementary files. patients and 142 images of normal patients. The model’s per-
The remainder of this paper is structured as follows. Section 2 formance has been evaluated using random sampling using 70%
introduces the related work. The materials and methods used in for training and 30% for testing. The proposed system for binary
this research are described in Section 3. Section 4 presents the re- classification provides an accuracy of 88.10%. The authors of this
sults of the proposed CNN model. The discussion and comparison study state the limitations on the number of samples included in
of the proposed method with the related work available in state the used dataset.
of the art are presented in Section 5. Finally, the conclusions are A novel artificial neural network system for COVID-19 de-
presented in Section 6. tection is proposed in [47]. The proposed method is based on
Convolutional CapsNet and uses chest X-ray images. The system
2. Related work includes binary and multi-class classification features. The dataset
used includes a total of 3150 samples, 1050 images each class
Numerous researchers are working at their best effort using (normal, pneumonia and COVID-19). The 10-fold cross-validation
AI technologies to develop novel systems to support COVID-19 is used to evaluate the performance of the proposed system. The
diagnosis. These studies aim to create new automated systems results show an accuracy of 97.24% and 84.22% for binary and
for COVID-19 diagnosis. These methods should be used to sup- multi-class classification. The limitations reported by the authors
port medical staff in the current pandemic scenario. Further- focus on the hardware resources needed to process a massive
more, machine learning technologies can be used to decrease the number of images and the processing time.
stress factors that affect medical professionals during this pan- Nour et al. [48] propose a novel medical diagnosis model of
demic scenario concerning the increase of workflow in healthcare COVID-19 to support clinical applications. The system is based
facilities. on deep features and Bayesian optimization. The CNN model is
Ozturk et al. [43] propose an automated detection system applied for automated extraction of features that are often pro-
for COVID-19 cases using deep neural networks and chest X-ray cessed by different machine learning methods such as kNN, SVM
images. The proposed method is based on the DarkNet model and Decision Tree. The used dataset contains 2033 of samples,
for real-time detection and implements 17 convolutional layers. 135 for COVID-19, 939 for normal class and 941 for pneumonia.
This study aims to support the decision making of radiologists The authors have used data augmentation to increase the number
to validate their screening process. The heatmaps produced by of samples concerning COVID-19 class. The performance of the
G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691 3
Table 1
Related work on COVID-19 detection systems.
Reference Model Data used Number of images Classification Evaluation method
[43] DarkNet model Chest X-ray 1125 — total; 125 — COVID-19; 500 Binary and Multi-class 5-fold cross-validation
— Pneumonia; 500 — Normal
[44] VGG-19 Chest X-ray 545 — total; 181 — COVID-19; 364 - Normal Binary Random sampling 80:20:20
for train, validation and
testing.
[45] VGG-19 and Chest X-ray Dataset 1: 1427 — total; 504 — Normal; 700 Binary and Multi-class 10-fold cross-validation
MobileNet v2 — Pneumonia; 224 — COVID-19 Dataset 2:
1442 — total; 504 — Normal; 714
— Pneumonia; 224 — COVID-19
[46] nCOVnet Chest X-ray 337 — total; 192 — COVID-19; 142- Normal Binary Random sampling 70% for
training and 30% for testing
[47] CapsNet Chest X-ray 3150 — total; 1050 — Normal; 1050 — Binary and Multi-class 10-fold cross-validation
Pneumonia; 1050 — COVID-19
[48] Proposed CNN Chest X-ray 2033 — total; 135 — COVID-19, ; 939 Binary and Random sampling
— Normal; Multi-class 70% for training and 30%
941 — Pneumonia. for testing
[49] ResNet 18 Chest X-ray 746 — total 349 — COVID-19 397 — Normal. Binary Random sampling. 70% for
training and 30% for testing.
[50] Proposed Semi- Chest X-ray Dataset 1: 2482 — total; 1230 — Normal; Binary Random sampling 70% for
supervised 1252 — COVID-19 Dataset 2: 20 — COVID-19 training and 30% for testing
model
proposed system has been evaluated using 70% and 30% of the the equal number of samples that cover the analyzed classes to
dataset for training and testing, respectively. The proposed CNN properly validate the performance of the model. Consequently,
presents an accuracy of 97.14%. the authors have used the same number of images to train and
In [49], the authors have used augmentation for increasing the test. Table 2 presents the reference and number of images used by
size of training dataset by using stationary wavelets and com- the authors. In total, 404 samples have been used corresponding
pared different transfer learning CNN architectures. The dataset to Normal, Pneumonia and COVID-19. These samples have been
used includes 349 samples for COVID-19 and 397 samples for nor- used in the stratified 10-fold cross-validation.
mal class. The authors also applied data augmentation techniques Furthermore, the authors have tested the model using a sep-
to increase the number of samples for both classes. In this study, arate dataset for validation. The number of samples used to
70% of the samples are used for training, and 30% have been validate the proposed model was 96 for normal class and 100 for
considered for validation. The proposed method provides 99.4% pneumonia and COVID-19 class. These datasets were not used in
accuracy during testing for a binary classifier using the ResNet the training phase of the model, and this experiment has been
18 model. conducted to test the non-occurrence of overfitting.
Konar et al. [50] propose a semi-supervised shallow neural On the one hand, pneumonia and normal images have been
network model for automated diagnostic of COVID-19. This study retrieved from the dataset available in [51]. This dataset is public
included two datasets. One of them consists of a total of 2482 and contains validated chest X-ray images of the pneumonia
samples, from these 1252 samples are from COVID-19 positive patients and normal patients. It is freely available on the Kaggle
patients, and 1230 are from not infected patients. The second website. This dataset only contains the folders named as Pneu-
dataset includes 20 samples of COVID-19 positive patients. The monia and Normal, and there is no other information available.
proposed model has tested using random sampling with a ratio The authors have downloaded these folders directly from the
of 70% for training and 30% for testing. Moreover, the model has Kaggle website to be included in the proposed work. On the other
also been evaluated using 5 and 10-fold validation. The proposed hand, the COVID-19 Image DataSet has been used to retrieve the
method presents an accuracy of 93.1%. COVID-19 positive samples and is available in [52]. It is a public
In summary, several methods have been proposed in the lit- dataset of validated chest X-ray images of COVID-19 positive
erature for the automated diagnostic of COVID-19. These stud- patients. It is available on GitHub database repository. The images
ies use different number of images and datasets from multiple are collected from public sources as well as through indirect
sources. Moreover, different approaches have been used to eval- collection from hospitals and physicians. This project is approved
uate the performance of the models such as cross-validation and by the University of Montreal’s Ethics Committee (CERSES-20-
random sampling. Most of the studies state the limitation asso- 058-D). The Research community is adding images continuously
ciated with the number of samples to conduct the experiments. in this dataset. The main purpose of developing this dataset is
Table 1 summarizes the related work on COVID-19 detection to improve prognostic predictions to triage and manage patient
systems. care. The authors used only two attributes, the findings column
to identify COVID-19 images and the name of the image.
3. Methods and materials
3.2. Proposed CNN
This section presents the methods and materials used in this
study. Section 3.1 details the datasets of X-ray images used to test The authors have used the EfficientNetB4 model for the trans-
and train the proposed method. The proposed CNN is presented fer learning process and added a global_average_pooling2d layer
in Section 3.2. Finally, the validation method and experimental to minimize overfitting by reducing the total number of param-
setup are presented in Section 3.3. eters. In addition to this, a sequence of 3 inner dense layers
with RELU activation functions and dropout layers have been
3.1. X-ray Image DataSet added. In total, a 30% dropout rate has chosen randomly to avoid
overfitting. Finally, one output dense layer contains two output
The samples used to train and test the proposed method units in case of binary classification, and 3 output units for multi-
have been collected from public datasets. It is critical to ensure class classification, with softmax activation function that has been
4 G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691
Table 2
Dataset information.
Class Reference Number of images for training/ testing Number of images for validation
NORMAL [51] 404 96
PNEUMONIA 404 100
COVID-19 [52] 404 100
Table 3 segmentation, and detection. The authors have used the Compose
Layer types and parameters used in the proposed model. method of the Albumentaion library. This library reduces over-
Layer (type) Output shape Param # fitting, improve the performance of classifiers and the decrease
EfficientNetB4 (Model) 7 ×7× 1792 17,673,816 execution time [53]. After the implementation of this library for
global_average_pooling2d 1792 0
augmentation purposes in each fold, the model accuracy of the
dense (Dense) 128 229,504
dropout (Dropout) 128 0 model has increased, and the processing time decreased.
dense_1 (Dense) 64 8256 Finally, the ImageDataAugmentor is a custom image data gen-
dropout_1 (Dropout) 64 0 erator for Keras supporting the use of modern augmentation
dense_2 (Dense) 32 2080 modules (e.g. imgaug and albumentations) [54]. This library is
dropout_2 (Dropout) 32 0
used to configure the Image data generator according to the al-
dense_3 (Dense) 2/3 99
Total Parameters: 17,913,755 bumentations settings to decrease execution time. Data generator
Trainable Parameters: 17,788,555 is created by using the constructor of ImageDataAugmentor class
Non-trainable Parameters: 125,200 with two arguments. One is rescale whose value is set as 1/255 to
transform every pixel value from range [0, 255] to [0, 1]. Another
is the augment value that is configured as the output of the
added to create the proposed automated detection system. The compose function of the Albumentation library. Data generator
details of the layers and their order in the proposed model, output has used further to process the image datasets. Fig. 1 presents
shape of each layer, the number of parameters (weights) in each the block diagram of the proposed work.
layer, and the total number of parameters (weights) are presented
in Table 3. The total number of parameters is 17,913,755. 3.3. Validation and experimental setup
All the software and libraries used in the proposed work are
open source. To reproduce the results, the readers should use The model has been validated in two different phases. On the
Google Colab Notebook using the GPU run time type. This soft- one hand, the 10-fold cross-validation method has been using
ware can be used without costs since it is provided by Google for the same dataset for training and testing. On the other hand,
research activities using a Tesla K80 GPU of 12 GB. The Efficient- a separate dataset which contains samples that have not been
Net Models are pre-trained, scaled CNN models that can be used used during the training phase has been applied to validate
for transfer learning in image classification problems. The model the performance of the model. The confusion matrix has been
is developed by Google AI in May 2019 and is available from extracted. Consequently, the precision, recall and F1-score have
Github repositories. The ImageDataAugmentor is a custom image been computed concerning the separated classes. Finally, the av-
data generator for Keras which supports augmentation modules. erage values for each fold have been calculated. The experimental
It is also developed by Google AI and is available from Github setup used to conduct this study is detailed in Algorithm 1.
repositories. Finally, the Albumentations library is also developed
by Google AI and can be installed from Github Repositories. In 4. Results
summary, the software used can be used without license concerns
as it is free and open source. The experiments were carried out on Google Colab notebook
The authors have used three different main libraries in the using GPU run time type. The training of the proposed CNN model
proposed method. These libraries include the EfficientNet mod- was realized using stratified 10-fold cross-validation method. In
ule, the Albumentation module and the ImageDataAugmentor total, 11 epochs were used in each fold. Moreover, 69 steps for
module. multi-class and 46 steps for binary classification are used in each
On the one hand, the EfficientNet Models are based on simple epoch. The mini-batch size used was 16.
and highly effective compound scaling methods. This method The training of the model was completed in a total of 7590
enables to scale up a baseline ConvNet to any target resource iterations for multi-class and 5060 iterations for binary class.
constraints while maintaining model efficiency, used for transfer The time elapsed for the training of the model was 111.83 min
learning datasets. In general, EfficientNet models achieve both for multi-class CNN and 79.16 min for binary class. The initial
higher accuracy and better efficiency over existing CNNs such as learning rate was 0.0001. The authors employed a ReduceLROn-
AlexNet, ImageNet, GoogleNet, and MobileNetV2 [41]. Efficient- Plateau method since it reduces the learning rate when it stops
Net could serve as a new foundation for future computer vision improving. This callback monitors the improvement, and if no
tasks. There is no similar study that uses EfficientNet for transfer improvement is verified for a ‘patience’ number of epochs, the
learning concerning COVID-19 classification to the best of authors learning rate is reduced. The authors have defined patience=3
knowledge until this date. EfficientNet includes models from B0 and the min_lr=0.000001 in the proposed method. The ADAM
to B7, and each one has different parameters from 5.3M to 66M. optimization method was used as a solver. The training and
The authors used EfficentNetB4 that contains 19M parameters, as validation graphs with the loss, confusion matrix, and area under
it is suitable according to our resources and purpose. the curve of receiver operating characteristics for each fold of
On the other hand, the Albumentation library is widely used in the proposed CNN can be verified from the supplementary files.
industry, deep learning research, machine learning competitions, After training of all the 10-fold CNN models, the best model is
and open source projects. This module efficiently implements a identified and used for the validation testing, by using different
variety of image transform operations that are optimized for per- datasets. The performance reported in the validation experiment
formance. This library provides an image augmentation interface is promising. The scripts and detailed information concerning the
for different computer vision tasks, including object classification, experiments can be consulted in the supplementary files.
G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691 5
Table 4 Table 6
Results of binary classification for COVID-19 class. Results of binary classification between classes.
Fold Precision Recall F1 score Fold Accuracy Precision Recall F-1 Score
1 100% 100% 100% 1 100% 100% 100% 100%
2 97.61% 100% 98.79% 2 98.76% 98.88% 98.75% 98.76%
3 100% 97.56% 98.76% 3 98.76% 98.78% 98.78% 98.76%
4 100% 100% 100% 4 100% 100% 100% 100%
5 100% 100% 100% 5 100% 100% 100% 100%
6 100% 100% 100% 6 100% 100% 100% 100%
7 100% 100% 100% 7 100% 100% 100% 100%
8 97.56% 100% 98.76% 8 98.76% 98.78% 98.78% 98.76%
9 100% 100% 100% 9 100% 100% 100% 100%
10 100% 100% 100% 10 100% 100% 100% 100%
Average 99.51% 99.75% 99.63% Average 99.62% 99.64% 99.63% 99.62%
Table 5
Results of binary classification for normal class.
Fold Precision Recall F1 score
1 100% 100% 100%
2 97.61% 100% 98.79%
3 97.56% 100% 98.76%
4 100% 100% 100%
5 100% 100% 100%
6 100% 100% 100%
7 100% 100% 100%
8 100% 97.56% 98.76%
9 100% 100% 100%
10 100% 100% 100%
Average 99.51% 99.75% 99.63%
Table 9
Results of multi-class classification for pneumonia class.
Fold Precision Recall F1 Score
1 95.00% 92.68% 93.82%
2 95.12% 95.12% 95.12%
3 97.5% 95.12% 96.29%
4 95.12% 95.12% 95.12%
5 97.43% 95.00% 96.20%
6 100% 100% 100%
7 100% 92.50% 96.10%
8 100% 92.50% 96.10%
9 97.29% 90.00% 93.50%
10 97.29% 87.80% 92.30%
Average 97.47% 93.58% 95.45%
Table 10
Results of multi-class classification between all classes.
Fold Accuracy Precision Recall F-1 Score
Fig. 3. ROC curve of the validation testing for the binary classifier. 1 95.90% 95.89% 95.89% 95.88%
2 96.72% 96.70% 96.70% 96.70%
Table 7 3 97.54% 98.33% 97.54% 97.93%
Results of multi-class classification for COVID-19 class. 4 96.72% 96.70% 96.70% 96.70%
5 97.52% 98.31% 97.50% 97.90%
Fold Precision Recall F1 Score
6 99.17% 100% 99.18% 99.58%
1 100% 100% 100% 7 97.52% 98.44% 97.50% 97.90%
2 100% 100% 100% 8 97.52% 98.39% 97.50% 97.88%
3 100% 100% 100% 9 95.04% 97.45% 95.04% 96.18%
4 100% 100% 100% 10 93.38% 95.78% 93.43% 94.52%
5 100% 100% 100% Average 96.70% 97.59% 96.69% 97.11%
6 100% 100% 100%
7 100% 100% 100%
8 97.56% 100% 98.76%
9 97.56% 100% 98.76%
10 97.56% 100% 98.76%
Average 99.26% 100% 99.62%
Table 8
Results of multi-class classification for normal class.
Fold Precision Recall F1 Score
1 92.68% 95.00% 93.82%
2 95.00% 95.00% 95.00%
3 97.50% 97.50% 97.50%
4 95.00% 95.00% 95.00%
5 97.50% 97.50% 97.50%
6 100% 97.56% 98.76%
7 95.34% 100% 97.61%
8 97.61% 100% 98.79%
9 97.50% 95.12% 96.29%
10 92.50% 92.50% 92.50%
Average 96.06% 96.51% 96.27%
Fig. 4. Confusion matrix of the validation testing for the multi-class classifier.
the precision, recall, and F1 score of each fold for the binary
classification considering the COVID-19 class. From the analysis of
Table 3 the authors identified the best model, and consequently
choose the model trained in the 10-fold for validation using
samples that have not been including in the training process.
On the other hand, Table 6 presents the precision, recall, and F1
score of each fold for the multi-class classification considering
the COVID-19 class. Based on the results of Table 6 the authors
selected the model trained in the 6-fold for validation using
samples that have not been included in the training process. The
output of validation testing is detailed in Sections 4.2 and 4.4 for
binary and multi-class, respectively. Fig. 2 presents the confusion
matrix and Fig. 3 shows the AUC-ROC curve concerning binary
classification. The results presented an accuracy of 99.49% on the
validation process is similar to the presented testing accuracy of
99.62%, which proves the model is not overfitted. Fig. 4 shows
Fig. 5. ROC curve of the validation testing for the multi-class classifier. the confusion matrix, and Fig. 5 presents the AUC-ROC curve. The
achieved accuracy during the validation process of multi-class is
96.62%, that is similar to the 96.70% accuracy reported during
testing and proves that the proposed model is not overfitted.
In summary, the results of the proposed CNN model for auto-
To the best of the authors’ knowledge, there is no automated
mated medical diagnostics support are promising. The reported
system for COVID-19 diagnosis in the literature that includes
average accuracy values for binary and multi-class are 99.62%
the combination of the following features. On the one hand,
and 96.70%, respectively. On the one hand, the proposed CNN
the proposed model uses EfficientNet for transfer learning. On
model using EfficientNet architecture presents an average recall
the other hand, the proposed methods are evaluated using 10-
value of 99.63% and 96.69% concerning binary and multi-class,
fold stratified cross-validation method. This method is used for
respectively. On the other hand, the average precision of 99.64%
selecting the images for testing and training. It reduces bias
is reported by binary classification, and 97.54% is presented in
and ensures that all the images are used 9 times for training
multi-class. Finally, the average F1-score value for multi-class is
97.11%, and 99.62% is presented for binary classification. and 1 time for testing. On the other hand, the proposed model
includes validation using an external dataset of images. In total,
5. Discussion 296 images that have been not used during the training of the
model are used for cross-validation. The validation results state
The proposed model is compared with the related work. In- similar accuracy as expected, which proves the model is not
deed, it is crucial to mention this comparison is limited concern- overfitted. This is detailed in Sections 4.2 and 4.4. Furthermore,
ing the differences in the samples used and the parameters of the Albumentation library is used pre-processing images during
the machine learning methods. Furthermore, since most of the each fold. This library is not used to increase the size of the
studies do not provide the software, it is not possible to compare datasets as presented in the related work. Instead, Albumentation
the methods applied by the studies available in the literature is implemented to increase transfer learning performance, reduce
using the samples included in our research or vice-versa. overfitting, and improve execution time. Finally, the source code
The number of new methods proposed in the literature by of all the experiments is available as supplementary files to allow
computer science researchers increases every day. Currently, the the readers to reproduce the experiments. The proposed models
focus on these systems to support health professionals is a trend- have been developed and testing using Google Colab. Therefore,
ing topic. Different architectures such as DarkCovidNET, VGG19, the software can be executed on the Google’s cloud servers.
MobileNet v2, CapsNet and nCOVnet have been proposed for The proposed study includes ADAM optimization method.
automated medical diagnosis of COVID-19. This section aims to ADAM benefits of AdaGrad and RMSProp methods. Most of the
compare the proposed model results with similar studies avail- similar studies also used ADAM optimizer. Furthermore, Adam
able in state of the art. Table 11 presents the results presented is currently recommended as the default algorithm as it usually
for binary classification studies available in the literature. presents better results than RMSProp. Nevertheless, it is often
The proposed method outperforms all the works concerning also worth trying SGDNesterov Momentum as an alternative.
binary classification. Nevertheless, the authors of [45] proposed The authors aim to integrate ADAM optimizer with SOM (Self-
an architecture with a relevant accuracy of 98.75%. Moreover, Organization Map) and PCA (Principal Component Analysis) to
the method proposed in [43] also presents a significant accuracy improve performance as proposed by the authors of [55].
of 98.08%. Consequently, the implementation of the EfficientNet In the proposed work the authors have used 10-fold stratified
architecture presents promising results for automated medical cross-validation and an Albumentation library for performing
diagnosis of COVID-19 concerning binary classification. Different augmentations to pre-processing images in each fold, and not for
methods for multi-class classification of COVID-19 patients are increasing the size of training datasets as proposed the authors
presented in Table 12. of [48,49]. The authors are able to use all the images at least
The accuracy levels proposed by the methods in Table 11 once for training and testing both and also increasing the learning
range from 84.22% [47] to 97.14% [48]. When compared with of the model. Augmentation is used for two purposes such as
the system proposed by the authors of [48] our method provides increasing the training set size, and another is the k-fold cross
less accuracy but higher recall and F1-score. The authors of [48] validation for pre-processing images in each fold to improve the
present a recall of 94.61% and an F1-score of 95.75%. The pro- performance of the model as proposed in [53].
posed method provides 96.69% and 97.11% concerning recall and In summary, the authors state the promising results of the
F1-score, respectively. EfficientNet architecture for automated diagnosis of COVID-19
The authors have used stratified 10-fold cross-validation to for binary and multi-class classification. Moreover, the authors
evaluate the proposed models. On the one hand, Table 3 contains recommend the use of Albumentation and ImageDataAugmentor
G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691 9
Table 11
Comparison of the state-of-art models for binary classification.
Ref. Architecture Accuracy Recall Specificity Precision F1-Score
[43] DarkCovidNet 98.08% 95.13% 95.3% 98.03% 96.51%
[44] VGG19 96.33% 97.05% 96.0% 91.6% 94.24%
[45] VGG19 98.75% 92.85% 98.75% – –
[45] MobileNet v2 97.40% 99.10% 97.09% – –
[46] nCOVnet 88.10% 82.00% 97.06% 97.62% 89.13%
[47] CapsNet 97.24% 97.42% 97.04% 97.08% 97.24%
[49] RestNet 18 99.4% 100% 98.6% 99.00% 99.5%
[50] Semi-supervised model 93.1% 83.5% – 89.0% 82.6%
Proposed EfficientNet 99.62% 99.63% - 99.64% 99.62%
Table 12
Comparison of the state-of-art models for multi-class classification.
Ref. Architecture Accuracy Recall Specificity Precision F1-Score
[43] DarkCovidNet 87.02% 85.35% 92.18% 89.96% 87.37%
[45] VGG19 93.48% 92.85% 98.75% – –
[45] MobileNet v2 92.85% 99.10% 97.09% – –
[47] CapsNet 84.22% 84.22% 91.79% 84.61% 84.21%
[48] Proposed CNN 97.14% 94.61% 98.29% – 95.75%
Proposed EfficientNet 96.70% 96.69% - 97.59% 97.11%
modules. The presented work contributes to the actual body of Declaration of competing interest
knowledge since it provides an effective solution for automated
diagnosis of COVID-19. Systems such as the proposed will never The authors declare that they have no known competing finan-
aim to replace the medical professionals. Instead, these methods cial interests or personal relationships that could have appeared
will support them and also reduce their exposure during the to influence the work reported in this paper.
current pandemic scenario.
Acknowledgments
6. Conclusion
This research has been partially supported by European Com-
This paper has presented an automated system to support mission and the Ministry of Industry, Energy and Tourism un-
the diagnosis of COVID-19 patients. The proposed method imple-
der the project AAL-20125036 named ‘‘Wetake Care: ICT-based
ments EfficientNet architecture and has been tested using 10-fold
Solution for (Self-) Management of Daily Living’’.
cross-validation. Furthermore, an external dataset has been used
for validation. On the one hand, the average accuracy, recall, pre-
Appendix A. Supplementary data
cision and F1-score for binary classification is of 99.62%, 99.63%,
99.64% and 99.62%, respectively. On the one hand, the proposed
Supplementary material related to this article can be found
model presents an average accuracy of 96.70% for multi-class. The
online at https://doi.org/10.1016/j.asoc.2020.106691.
average recall, precision and F1-score reported by our method is
96.69%, 97.59% and 97.11%. To the best of the authors’ knowledge,
there is no similar study that proposes an automated method References
to the detected COVID-19 using EfficientNet. The authors state
[1] C. Lodigiani, G. Iapichino, L. Carenzo, M. Cecconi, P. Ferrazzi, T. Sebastian,
relevant limitations transversal to all the methods available in N. Kucher, J.-D. Studt, C. Sacco, B. Alexia, M.T. Sandri, S. Barco, Venous and
the literature. First, although the critical number of individuals arterial thromboembolic complications in COVID-19 patients admitted to
that have been affected by COVID-19, the datasets available are an academic hospital in Milan, Italy, Thrombosis Res. 191 (2020) 9–14,
not yet robust. Nevertheless, the empirical knowledge regarding http://dx.doi.org/10.1016/j.thromres.2020.04.024.
CNN applications state that the performance will increase with [2] C. Huang, Y. Wang, X. Li, L. Ren, J. Zhao, Y. Hu, L. Zhang, G. Fan, J.
Xu, X. Gu, Z. Cheng, T. Yu, J. Xia, Y. Wei, W. Wu, X. Xie, W. Yin, H.
the proliferation of the number of samples available and used for
Li, M. Liu, Y. Xiao, H. Gao, L. Guo, J. Xie, G. Wang, R. Jiang, Z. Gao,
training. Moreover, it is critical to study in detail the evaluation of Q. Jin, J. Wang, B. Cao, Clinical features of patients infected with 2019
this type of methods considering the evolution of the disease in novel coronavirus in Wuhan, China, Lancet 395 (2020) 497–506, http:
the patient. The proposed methods can be effective to detect the //dx.doi.org/10.1016/S0140-6736(20)30183-5.
disease in the advanced stage, but it is essential to focus on the [3] Y.-Y. Zheng, Y.-T. Ma, J.-Y. Zhang, X. Xie, COVID-19 and the cardiovascular
system, Nat. Rev. Cardiol. 17 (2020) 259–260, http://dx.doi.org/10.1038/
initial stage where the methods can have lower performance. The
s41569-020-0360-5.
coding scripts are provided as supplementary files. It is crucial [4] Imperial College COVID-19 Response Team, S. Flaxman, S. Mishra, A. Gandy,
to share all the methods and materials to allow the readers to H.J.T. Unwin, T.A. Mellan, H. Coupland, C. Whittaker, H. Zhu, T. Berah, J.W.
reproduce the results. Moreover, in this way, it is possible to Eaton, M. Monod, A.C. Ghani, C.A. Donnelly, S. Riley, M.A.C. Vollmer, N.M.
support future research activates. The readers can consult, update Ferguson, L.C. Okell, S. Bhatt, Estimating the effects of non-pharmaceutical
interventions on COVID-19 in Europe, Nature 584 (2020) 257–261, http:
and implement different parameters to improve the results.
//dx.doi.org/10.1038/s41586-020-2405-7.
[5] F.-X. Lescure, L. Bouadma, D. Nguyen, M. Parisey, P.-H. Wicky, S. Behillil, A.
CRediT authorship contribution statement Gaymard, M. Bouscambert-Duchamp, F. Donati, Q. Le Hingrat, V. Enouf, N.
Houhou-Fidouh, M. Valette, A. Mailles, J.-C. Lucet, F. Mentre, X. Duval, D.
Gonçalo Marques: Literature analysis, Interpretation of re- Descamps, D. Malvy, J.-F. Timsit, B. Lina, S. van-der Werf, Y. Yazdanpanah,
sults, Preparation of the manuscript. Deevyankar Agarwal: Liter- Clinical and virological data of the first cases of COVID-19 in Europe: A
case series, Lancet Infectious Dis. 20 (2020) 697–706, http://dx.doi.org/10.
ature analysis, Interpretation of results, Preparation of the
1016/S1473-3099(20)30200-0.
manuscript. Isabel de la Torre Díez: Literature analysis, Interpre- [6] A.P. Kansagra, M.S. Goyal, S. Hamilton, G.W. Albers, Collateral effect of
tation of results, Preparation of the manuscript, Project adminis- Covid-19 on stroke evaluation in the United States, New Engl. J. Med. 383
tration. (2020) 400–401, http://dx.doi.org/10.1056/NEJMc2014816.
10 G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691
[7] A.J. Rodriguez-Morales, V. Gallego, J.P. Escalera-Antezana, C.A. Méndez, [28] F.H.D. Araújo, A.M. Santana, P. de A. Santos Neto, Using machine learning to
L.I. Zambrano, C. Franco-Paredes, J.A. Suárez, H.D. Rodriguez-Enciso, G.J. support healthcare professionals in making preauthorisation decisions, Int.
Balbin-Ramon, E. Savio-Larriera, A. Risquez, S. Cimerman, COVID-19 in J. Med. Inform. 94 (2016) 1–7, http://dx.doi.org/10.1016/j.ijmedinf.2016.06.
Latin America: The implications of the first confirmed case in Brazil, Travel 007.
Med. Infectious Dis. 35 (2020) 101613, http://dx.doi.org/10.1016/j.tmaid. [29] L. Luo, J. Li, C. Liu, W. Shen, Using machine-learning methods to support
2020.101613. health-care professionals in making admission decisions, Int. J. Health
[8] D. Roy, S. Tripathy, S.K. Kar, N. Sharma, S.K. Verma, V. Kaushal, Study Plann. Mgmt. 34 (2019) http://dx.doi.org/10.1002/hpm.2769.
of knowledge, attitude, anxiety & perceived mental healthcare need in [30] the Northwell COVID-19 Research Consortium, S. Debnath, D.P. Barnaby,
Indian population during COVID-19 pandemic, Asian J. Psychiatry 51 (2020) K. Coppa, A. Makhnevich, E.J. Kim, S. Chatterjee, V. Tóth, T.J. Levy, M. d.
102083, http://dx.doi.org/10.1016/j.ajp.2020.102083. Paradis, S.L. Cohen, J.S. Hirsch, T.P. Zanos, Machine learning to assist
[9] A. Remuzzi, G. Remuzzi, COVID-19 and Italy: What Next? Lancet 395 clinical decision-making during the COVID-19 pandemic, Bioelectron Med.
(2020) 1225–1228, http://dx.doi.org/10.1016/S0140-6736(20)30627-9. 6 (2020) 14, http://dx.doi.org/10.1186/s42234-020-00050-8.
[10] H. Harapan, N. Itoh, A. Yufika, W. Winardi, S. Keam, H. Te, D. Megawati, [31] C. Dong, C.C. Loy, X. Tang, Accelerating the super-resolution convolutional
Z. Hayati, A.L. Wagner, M. Mudatsir, Coronavirus disease 2019 (COVID- neural network, in: B. Leibe, J. Matas, N. Sebe, M. Welling (Eds.), Computer
19): A literature review, J. Infection Public Health 13 (2020) 667–673, Vision – ECCV, Springer International Publishing, Cham, 2016, pp. 391–407,
http://dx.doi.org/10.1016/j.jiph.2020.03.019. http://dx.doi.org/10.1007/978-3-319-46475-6_25.
[11] X. Cao, COVID-19: Immunopathology And its implications for therapy, Nat. [32] S. Lawrence, C.L. Giles, A.D. Ah Chung Tsoi, Face recognition: a convolu-
Rev. Immunol. 20 (2020) 269–270, http://dx.doi.org/10.1038/s41577-020- tional neural-network approach, IEEE Trans. Neural Netw. 8 (1997) 98–113,
0308-3. http://dx.doi.org/10.1109/72.554195.
[12] P. Mehta, D.F. McAuley, M. Brown, E. Sanchez, R.S. Tattersall, J.J. Manson, [33] D.K. Das, S. Koley, S. Bose, A.K. Maiti, B. Mitra, G. Mukherjee, P.K. Dutta,
COVID-19: consider cytokine storm syndromes and immunosuppression, Computer aided tool for automatic detection and delineation of nucleus
Lancet 395 (2020) 1033–1034, http://dx.doi.org/10.1016/S0140-6736(20) from oral histopathology images for OSCC screening, Appl. Soft Comput.
30628-0. 83 (2019) 105642, http://dx.doi.org/10.1016/j.asoc.2019.105642.
[13] S. Al Kasab, E. Almallouhi, C.A. Holmstedt, Optimizing the Use Of Teleneu- [34] X. Ji, Q. Yu, Y. Liu, S. Kong, A recognition method for Italian alpha-
rology During the COVID-19 Pandemic. http://dx.doi.org/10.1089/tmj.2020. bet gestures based on convolutional neural network, in: D.-S. Huang,
0109. V. Bevilacqua, P. Premaratne (Eds.), Intelligent Computing Theories and
[14] A. Doshi, Y. Platt, J.R. Dressen, B.K. Mathews, J.C. Siy, Keep calm and log on: Application, Springer International Publishing, Cham, 2019, pp. 653–664,
Telemedicine for COVID-19 pandemic response, 2020, http://dx.doi.org/10. http://dx.doi.org/10.1007/978-3-030-26763-6_63.
12788/jhm.3419. [35] A. Tavanaei, M. Ghodrati, S.R. Kheradpisheh, T. Masquelier, A. Maida, Deep
[15] S. Omboni, Telemedicine During The COVID-19 in Italy: A Missed learning in spiking neural networks, Neural Netw. 111 (2019) 47–63,
Opportunity? http://dx.doi.org/10.1089/tmj.2020.0106. http://dx.doi.org/10.1016/j.neunet.2018.12.002.
[16] A. Tashiro, R. Shaw, COVID-19 pandemic response in Japan: What is
[36] R. Karthik, M. Hariharan, Sundar Anand, Priyanka Mathikshara, Annie John-
behind the initial flattening of the curve?, Sustainability 12 (2020) 5250,
son, R. Menaka, Attention embedded residual CNN for disease detection in
http://dx.doi.org/10.3390/su12135250.
tomato leaves, Appl. Soft Comput. 86 (2020) 105933, http://dx.doi.org/10.
[17] D. Lee, K. Heo, Y. Seo, COVID-19 in South Korea: Lessons for develop-
1016/j.asoc.2019.105933.
ing countries, World Dev. 135 (2020) 105057, http://dx.doi.org/10.1016/j.
[37] Y. Wang, X. Wei, H. Shen, L. Ding, J. Wan, Robust fusion for RGB-D
worlddev.2020.105057.
tracking using CNN features, Appl. Soft Comput. 92 (2020) 106302, http:
[18] M. Ostaszewski, A. Mazein, M.E. Gillespie, I. Kuperstein, A. Niarakis, H.
//dx.doi.org/10.1016/j.asoc.2020.106302.
Hermjakob, A.R. Pico, E.L. Willighagen, C.T. Evelo, J. Hasenauer, F. Schreiber,
[38] J.C. Rangel, J. Martínez-Gómez, C. Romero-González, I. García-Varea, M.
A. Dräger, E. Demir, O. Wolkenhauer, L.I. Furlong, E. Barillot, J. Dopazo, A.
Cazorla, Semi-supervised 3D object recognition through CNN labeling,
Orta-Resendiz, F. Messina, A. Valencia, A. Funahashi, H. Kitano, C. Auffray,
Appl. Soft Comput. 65 (2018) 603–613, http://dx.doi.org/10.1016/j.asoc.
R. Balling, R. Schneider, COVID-19 disease map, building a computational
2018.02.005.
repository of SARS-CoV-2 virus–host interaction mechanisms, Sci Data. 7
[39] C. Wang, Z. Zhao, Q. Ren, Y. Xu, Y. Yu, A novel multi-focus image fusion by
(2020) 136, http://dx.doi.org/10.1038/s41597-020-0477-8.
combining simplified very deep convolutional networks and patch-based
[19] N. Muralidharan, R. Sakthivel, D. Velmurugan, M.M. Gromiha, Computa-
sequential reconstruction strategy, Appl. Soft Comput. 91 (2020) 106253,
tional studies of drug repurposing and synergism of lopinavir, oseltamivir
http://dx.doi.org/10.1016/j.asoc.2020.106253.
and ritonavir binding with SARS-CoV-2 protease against COVID-19, J.
[40] L.T. Duong, P.T. Nguyen, C. Di Sipio, D. Di Ruscio, Automated fruit
Biomol. Struct. Dyn. (2020) 1–6, http://dx.doi.org/10.1080/07391102.2020.
recognition using efficientnet and mixnet, Comput. Electron. Agric. 171
1752802.
[20] Z. Li, J. Ge, M. Yang, J. Feng, M. Qiao, R. Jiang, J. Bi, G. Zhan, X. Xu, (2020) 105326, http://dx.doi.org/10.1016/j.compag.2020.105326.
L. Wang, Q. Zhou, C. Zhou, Y. Pan, S. Liu, H. Zhang, J. Yang, B. Zhu, Y. [41] M. Tan, Q.V. Le, Efficientnet: Rethinking model scaling for convolutional
Hu, K. Hashimoto, Y. Jia, H. Wang, R. Wang, C. Liu, C. Yang, Vicarious neural networks, 2019, arXiv:1905.11946 [cs, stat].
traumatization in the general public, members, and non-members of [42] Efficientnet: Improving accuracy and efficiency through automl and model
medical teams aiding in COVID-19 control, Brain, Behav. Immunity 88 scaling, 2020, http://ai.googleblog.com/2019/05/efficientnet-improving-
(2020) 916–919, http://dx.doi.org/10.1016/j.bbi.2020.03.007. accuracy-and.html, (Accessed 15 August 2020).
[21] D.A. Drew, L.H. Nguyen, C.J. Steves, C. Menni, M. Freydin, T. Varsavsky, [43] T. Ozturk, M. Talo, E.A. Yildirim, U.B. Baloglu, O. Yildirim, U. Rajen-
C.H. Sudre, M.J. Cardoso, S. Ourselin, J. Wolf, T.D. Spector, A.T. Chan, dra Acharya, Automated detection of COVID-19 cases using deep neural
COPE Consortium§: Rapid implementation of mobile technology for real- networks with X-ray images, Comput. Biol. Med. 121 (2020) 103792,
time epidemiology of COVID-19, Science. 368 (2020) 1362–1367, http: http://dx.doi.org/10.1016/j.compbiomed.2020.103792.
//dx.doi.org/10.1126/science.abc0473. [44] S. Vaid, R. Kalantar, M. Bhandari, Deep learning COVID-19 detection
[22] F. Shi, J. Wang, J. Shi, Z. Wu, Q. Wang, Z. Tang, K. He, Y. Shi, D. Shen, bias: accuracy through artificial intelligence, Int. Orthopaedics 44 (2020)
Review of artificial intelligence techniques in imaging data acquisition, 1539–1542, http://dx.doi.org/10.1007/s00264-020-04609-7.
segmentation and diagnosis for COVID-19, IEEE Rev. Biomed. Eng. 1 (2020) [45] I.D. Apostolopoulos, T.A. Mpesiana, Covid-19: automatic detection from X-
http://dx.doi.org/10.1109/RBME.2020.2987975. ray images utilizing transfer learning with convolutional neural networks,
[23] A. Alimadadi, S. Aryal, I. Manandhar, P.B. Munroe, B. Joe, X. Cheng, Artificial Phys. Eng. Sci. Med. 43 (2020) 635–640, http://dx.doi.org/10.1007/s13246-
intelligence and machine learning to fight COVID-19, Physiol. Genomics 52 020-00865-4.
(2020) 200–202, http://dx.doi.org/10.1152/physiolgenomics.00029.2020. [46] H. Panwar, P.K. Gupta, M.K. Siddiqui, R. Morales-Menendez, V. Singh,
[24] A.S.R. Srinivasa Rao, J.A. Vazquez, Identification of COVID-19 can be quicker Application of deep learning for fast detection of COVID-19 in X-rays using
through artificial intelligence framework using a mobile phone–based ncovnet, Chaos, Solitons Fractals 138 (2020) 109944, http://dx.doi.org/10.
survey when cities and towns are under quarantine, Infect. Control Hosp. 1016/j.chaos.2020.109944.
Epidemiol. 41 (2020) 826–830, http://dx.doi.org/10.1017/ice.2020.61. [47] S. Toraman, T.B. Alakus, I. Turkoglu, Convolutional capsnet: A novel
[25] R. Vaishya, M. Javaid, I.H. Khan, A. Haleem, Artificial intelligence (AI) artificial neural network approach to detect COVID-19 disease from X-
applications for COVID-19 pandemic, Diabetes Metab. Syndr: Clin. Res. Rev. ray images using capsule networks, Chaos, Solitons Fractals 140 (2020)
14 (2020) 337–339, http://dx.doi.org/10.1016/j.dsx.2020.04.012. 110122, http://dx.doi.org/10.1016/j.chaos.2020.110122.
[26] E.J. Topol, High-performance medicine: the convergence of human and [48] M. Nour, Z. Cömert, K. Polat, A novel medical diagnosis model for COVID-
artificial intelligence, Nat. Med. 25 (2019) 44–56, http://dx.doi.org/10.1038/ 19 infection detection based on deep features and Bayesian optimization,
s41591-018-0300-7. Appl. Soft Comput. 106580 (2020) http://dx.doi.org/10.1016/j.asoc.2020.
[27] I.V. Buzaev, V.V. Plechev, I.E. Nikolaeva, R.M. Galimova, Artificial intelli- 106580.
gence: Neural network model as the multidisciplinary team member in [49] S. Ahuja, B.K. Panigrahi, N. Dey, T. Gandhi, V. Rajinikanth, Deep transfer
clinical decision support to avoid medical mistakes, Chronic Dis. Transl. learning - based automated detection of COVID-19 from lung CT scan
Med. 2 (2016) 166–172, http://dx.doi.org/10.1016/j.cdtm.2016.09.007. slices, 2020, http://dx.doi.org/10.36227/techrxiv.12334265.v1.
G. Marques, D. Agarwal and I. de la Torre Díez / Applied Soft Computing Journal 96 (2020) 106691 11
[50] D. Konar, B.K. Panigrahi, S. Bhattacharyya, N. Dey, Auto-diagnosis of COVID- [53] A. Buslaev, V.I. Iglovikov, E. Khvedchenya, A. Parinov, M. Druzhinin,
19 using lung CT images with semi-supervised shallow learning network, A.A. Kalinin, Albumentations: Fast and flexible image augmentations,
in: Review, 2020, http://dx.doi.org/10.21203/rs.3.rs-34596/v1. Information 11 (2020) 125, http://dx.doi.org/10.3390/info11020125.
[51] Chest X-ray images (pneumonia), 2020, https://kaggle.com/ [54] M. Tukiainen, ImageDataAugmentor. GitHub, GitHub repository.
paultimothymooney/chest-xray-pneumonia. (Accessed 15 August 2020). [55] Md.N.Y. Ali, Md.G. Sarowar, Md.L. Rahman, J. Chaki, N. Dey, J.M.R.S. Tavares,
[52] J.P. Cohen, P. Morrison, L. Dao, K. Roth, T.Q. Duong, M. Ghassemi, COVID- Adam deep learning with SOM for human sentiment classification, Int. J.
19 image data collection: Prospective predictions are the future, 2020, Ambient Comput. Intell. 10 (2019) 92–116, http://dx.doi.org/10.4018/IJACI.
arXiv:2006.11988 [cs, eess, q-bio]. 2019070106.