Bioengineering 10 00981
Bioengineering 10 00981
Article
Lung Tumor Image Segmentation from Computer Tomography
Images Using MobileNetV2 and Transfer Learning
Zainab Riaz 1 , Bangul Khan 1,2 , Saad Abdullah 3, * , Samiullah Khan 4 and Md Shohidul Islam 1
1 Hong Kong Center for Cerebro-Cardiovascular Health Engineering (COCHE), Hong Kong SAR, China;
zriaz@hkcoche.org (Z.R.); bangukhan2-c@my.cityu.edu.hk (B.K.); mdsislam@hkcoche.org (M.S.I.)
2 Department of Biomedical Engineering, City University Hongkong, Hong Kong SAR, China
3 Division of Intelligent Future Technologies, School of Innovation, Design and Engineering,
Mälardalen University, P.O. Box 883, 721 23 Västerås, Sweden
4 Center for Eye & Vision Research, 17W Science Park, Hong Kong SAR, China; samikhan@cevr.hk
* Correspondence: saad.abdullah@mdu.se
Abstract: Background: Lung cancer is one of the most fatal cancers worldwide, and malignant tumors
are characterized by the growth of abnormal cells in the tissues of lungs. Usually, symptoms of lung
cancer do not appear until it is already at an advanced stage. The proper segmentation of cancerous
lesions in CT images is the primary method of detection towards achieving a completely automated
diagnostic system. Method: In this work, we developed an improved hybrid neural network via the
fusion of two architectures, MobileNetV2 and UNET, for the semantic segmentation of malignant
lung tumors from CT images. The transfer learning technique was employed and the pre-trained
MobileNetV2 was utilized as an encoder of a conventional UNET model for feature extraction.
The proposed network is an efficient segmentation approach that performs lightweight filtering
to reduce computation and pointwise convolution for building more features. Skip connections
were established with the Relu activation function for improving model convergence to connect the
encoder layers of MobileNetv2 to decoder layers in UNET that allow the concatenation of feature
maps with different resolutions from the encoder to decoder. Furthermore, the model was trained
Citation: Riaz, Z.; Khan, B.;
and fine-tuned on the training dataset acquired from the Medical Segmentation Decathlon (MSD)
Abdullah, S.; Khan, S.; Islam, M.S.
2018 Challenge. Results: The proposed network was tested and evaluated on 25% of the dataset
Lung Tumor Image Segmentation
obtained from the MSD, and it achieved a dice score of 0.8793, recall of 0.8602 and precision of 0.93. It
from Computer Tomography Images
Using MobileNetV2 and Transfer
is pertinent to mention that our technique outperforms the current available networks, which have
Learning. Bioengineering 2023, 10, 981. several phases of training and testing.
https://doi.org/10.3390/
bioengineering10080981 Keywords: deep learning; medical imaging; CT; UNET; MobileNetV2; lung cancer; pulmonary nodule
Qureshi et al. [5] reviewed the semantic-based segmentation methods, existing chal-
lenges and their emerging trends. The authors of that study offer insights into the devel-
opment of machine learning and deep learning mechanisms, along with their strengths
and weaknesses. The paper provides a comprehensive overview of recent advancements
in semantic segmentation techniques; additionally, it presents a thorough investigation
into the effectiveness of different architectures for medical image segmentation. Moreover,
it helped the research community by highlighting the benefits, existing challenges and
potential future directions.
Traditional methods generally demand handcrafted features, for instance, pixel thresh-
olding, voxel clustering and morphological features [6]. These approaches to medical
image segmentation also revolve around edge detection, active contours and template
matching techniques [7]. Therefore, deep-learning-based classifiers (DLCs) have changed
the research objectives from traditional image processing techniques for feature engineering
to network architecture design for obtaining high accuracy. Moreover, transfer learning [8]
has established the most practical paradigms in the field of semantic segmentation [9]
and image classification [10]. It is a way of utilizing knowledge acquired from a source
domain while solving one supervised learning task and employing it to another related
target domain. A. A. Mukhlif et al. [11] discussed the significance of accurately segmenting
and evaluating the region of interest in medical imaging for disease screening and decision
making. The research specifically explored the lung section segmentation from chest X-ray
images, training the UNET with one-fold and two-fold training processes. The investigation
concluded that the proposed approach achieved superior results in the two-fold training
compared to other methods considered in this study. In another study, the authors [9]
highlighted the limitation of a CNN to efficiently handle irregular image orientations. To
address this problem, a new hybrid deep learning architecture known as the STNCNN was
proposed by integrating the space transformer network (STN) with a CNN. The developed
model was implemented on a dataset from the Kaggle repository and achieved promising
accuracy, outperforming vanilla grey, vanilla RGB and the hybrid CNN.
Due to the heterogeneity of tumors in terms of size, shape and appearance, tumor
detection remains a challenge. The automated segmentation of lung tumors from CT scan
images can assist medical practitioners in providing an early diagnosis for the further
monitoring of disease progression. Classical methods of automatic tumor segmentation
mainly depend on feature engineering, which requires the extraction of features from input
images for further training of the classifier [10]. However, U-NET, a convolutional neural
network, set a new benchmark in biomedical image segmentation and is considered as
one of the most advanced techniques for the accurate pattern classification of tumors, as
it automatically learns the relevant features [12]. Z Kong et al. [11] presented the hybrid
model of MobileNetv2 and UNET for the precise segmentation of liver regions from a
liver CT scan dataset. The approach involved introducing random noise to the generator’s
input and replacing the fully connected layer with a probability matrix to enhance the
discriminator’s sensitivity. The proposed algorithm achieved a dice similarity coefficient of
88.7, surpassing the performance of the standard UNET algorithm.
In this work, we present a deep-learning-based architecture for the semantic segmenta-
tion of malignant lung tumors from computed tomographic (CT) images. In our proposed
technique, we made the following contributions:
• We utilized a pre-trained MobileNetV2, retaining the convolutional layers, as the
encoder of the classical UNET for generating more stable segmentation maps. The
decoder part consists of up-sample layers and convolutional layers that recover the
spatial resolution and refine the segmentation results.
• Skip connections were established with the Relu activation function for improving the
model convergence to connect the encoder layers of MobileNetV2 to the decoder layers
in UNet, which allows the concatenation of feature maps with different resolutions
from the encoder to decoder. Thus, the decoder leverages both low-level and high-level
features for accurate segmentation.
Bioengineering 2023, 10, 981 3 of 13
• Finally, we added a 1 × 1 convolution layer at the end of the decoder to reduce the
number of channels and to obtain the number of output classes, such as tumor and
background.
• The devised network was further trained and fine-tuned with optimized hyper-
parameters on the training dataset obtained from the Medical Segmentation Decathlon
(MSD) 2018 Challenge.
• The results indicate that the proposed approach is robust and significantly improved
the segmentation accuracy.
The rest of the paper is organized as follows: Section 2 of this paper indicates the
literature covering the machine learning and deep learning techniques used in this domain.
Section 3 elaborates a detailed explanation of the proposed methods and Section 4 focuses
on the results and discussion, in which the obtained results from the suggested algorithm
are discussed and presented.
2. Background
The precise assessment of a lung tumor is essential to scrutinize its malignancy and
the probability of lung cancer. Wang et al. [13] proposed a support vector machine (SVM)
based on the three-dimensional matrix pattern method to avoid the loss of local and structural
information. The three-dimensional volume of tumors took the whole region of interest
(ROI) for analysis and fed it as an input image for the training of the algorithm, and the
model was not able to classify between benign and malignant tumor. However, the lung
parenchyma segmentation technique using the fast marching method was adopted in [14] to
extract candidate nodules from segmented lung parenchyma. Afterwards, a random forest
(RF) algorithm was employed for the classification between benign and malignant tumors.
Mukhlif et al. [15] highlighted the need for smart systems to aid clinicians in the early
detection of breast cancer, where the authors aimed to address the non-medical nature of
ImageNet features by incorporating unclassified medical images of the same disease to
mitigate the reliance on ImageNet. Therefore, the proposed approach employed a modified
Xception model to classify histological images of breast cancer into four categories, and
achieved high performance compared to previous studies in this field. On the other hand,
S Lu et al. [16] aimed to develop a system for automatically identifying COVID-19 in chest
CT images using artificial intelligence. The researchers utilized transfer learning to obtain
image-level representations (ILRs) based on a deep CNN. They proposed a neighboring aware
representation (NAR) to capture neighboring relationships between ILR vectors. Based on
such representations, they introduced a novel COVID-19 classification architecture known as
NAGNN that outperformed the state-of-the-art methods in terms of generalizability.
S M Naqi et al. opted for a strategy of employing multiple ML techniques for the
detection of lung cancer and compared the obtained results. Geometric texture and 3D
component connectivity was analyzed by novel hybrid 3D nodule detection, and based
upon the extracted feature, classification was performed by K-Nearest Neighbors (KNN),
SVM and AdaBoost. The evaluation of AdaBoost was performed using a dataset acquired
from the Lung Image Database Consortium (LIDC) [17].
W. Choi and T. Choi [18] suggested an automatic approach for the identification of a
lung tumor on the basis of a feature descriptor which then differentiated by the 3D shape
of the tumor. Multi-scale dot enhancement filtering is a technique utilized for segmenting
lung volume. Afterwards, potential nodule candidates were extracted and refined by
using an iterative edge elimination algorithm. Finally, an SVM classifier was trained to
differentiate nodules and non-nodules. M. Usman et al. [19] devised an approach that
consists of two stages: the first stage provides an initial estimation of a tumor by performing
patch-wise exploration along the axial axis using an adaptive ROI algorithm. In the second
stage, the extracted region is further investigated for the existence of a malignant tumor
along the coronal and sagittal axes.
The algorithm proposed by A. Setio et al. [20] was composed of three candidate de-
tectors specially designed for the detection of cancerous lesions to enhance the detection
Bioengineering 2023, 10, 981 4 of 13
sensitivity of lesions. The nodule candidates were computed and processed by ConvNets
by averaging the position of the tumor and its probability. U Kamal et al. proposed the
recurrent 3D-DenseUNet, an architecture for the segmentation of the volume of interest
from lung CT scans. The suggested approach comprised a 3D encoder block and recurrent
block of ConvLSTM layers to bring out fine-grained spatio-temporal details and later recon-
struct the volumetric segmentation mask by introducing a 3D decoder block. S Lu et al. [21]
proposed a novel method for detecting abnormal brain regions in MRI images using a
pre-trained AlexNet model. The authors modified a pre-trained model by adding batch
normalization layers and replaced the last layers with an extreme learning machine. Fur-
thermore, the extreme learning machine was optimized utilizing a chaotic bat algorithm
to enhance the classification performance, which demonstrated state-of-the-art results in
abnormal brain region detection.
Random transformation induces deliberate changes and can be used to create varied
images from available images to enhance the size of a dataset for training the classifier.
Deep convolutional neural networks (CNNs) have performed exceptionally well on com-
puter vision tasks. Overfitting happens when a network understands a function with
high variance. However, data augmentation increases the data size, along with the class-
preserving transformation and standards of the training dataset, ultimately strengthening
the generalization ability of deep learning models [22].
Tri Dao et al. [23] established a theoretical framework for understanding data aug-
mentation schemes. The Markov process is a general model of augmentation where
kernels appear spontaneously in the model. Data augmentations can be approximated
by first-order feature averaging and second-order variance regularization components.
They also analyzed the methods of augmentation that modify the models’ learning ability.
Nonetheless, data augmentation enhances the training dataset size by geometric and color
transformations and adversarial training.
leveraging the learned features and representations; therefore, A A Mukhlif et al. [30] discussed
the applications of transfer learning in various domains, particularly image processing and
interpretation. They also revealed the prevalent use of pre-trained models from the ImageNet
dataset in applications such as skin cancer, breast cancer and diabetic retinopathy classification.
Along with that, the authors further investigated the problems in melanoma and breast cancer
datasets, and potential solutions were suggested. In another study, A A Mukhlif et al. [31]
discussed the limitations of transfer learning in the medical domain due to the mismatch
between the source and target problem. To address this issue, the study proposed a novel
approach known as dual transfer learning (DTL) that focused on the convergence of patterns
between two domains. The proposed approach employed four pre-trained models utilizing
two datasets: skin cancer images and breast cancer images. The final layers of the models
were fine-tuned on enough unclassified images of the same disease and a small number
of classified images from the target task. The experimental results demonstrated that the
proposed approach improved the performance of all models.
Figure
Figure 1. Normalized
1. (A–C) imageimage
Normalized slices of CT scan
slices ofscan
of CT sameofpatient
same with growing
patient tumor in MSD-2018
with growing tumor in MSD-
training
2018 set,set,
training along withwith
along annotation overlaid
annotation on theon
overlaid image.
the image.
3.2. Methodology
In this section, we begin by describing the architecture that we employed. Mo-
bileNetV2 is usually adapted for resource-constrained environments to accurately solve
the problem of semantic segmentation and has the advantage of improving segmentation
Bioengineering 2023, 10, 981 6 of 13
3.2. Methodology
In this section, we begin by describing the architecture that we employed. MobileNetV2
is usually adapted for resource-constrained environments to accurately solve the problem of
semantic segmentation and has the advantage of improving segmentation results. We propose
a computationally lightweight network with fewer trainable parameters, and it achieves a
perfect balance between performance results and implementation efficiency. A 2D image
containing the nodule was provided as an input to detect the presence of lesions using an
algorithm. The output of the network was a segmentation map, from which a dice score
coefficient was calculated. We provide further details on the pipeline that has several phases
in the subsequent sections.
Preprocessing
Image normalization: We converted the 3D computed tomographic (CT) images to
2D and resized them to 256 × 256 to reduce the size of the CT slices owing to memory
consideration. Furthermore, the images were normalized to minimize poor contrast issues
before feeding them into the model for training. The following min–max approach rescaled
the feature in the range of 0 and 1.
I − Imin
Inormalized = (1)
Imax − Imin
Data Augmentation: When training the neural network with limited training data,
special attention must be paid to minimize overfitting. Augmentations induce deliberate
changes and hence can be used to create varied images from the available image dataset.
Greater variation in training data ensures model generalization. Images are randomly
augmented, which reduces the possibility of modeling to learn inherent patterns in data.
Augmentations as illustrated in Figure 2. such as CLAHE, rotate, blur, random contrast,
Bioengineering 2023, 10, x FOR PEER REVIEW 7 of 14
random sized crop and Gaussian blur are applied on data during runtime to circumvent
overfitting and to enhance the segmentation accuracy.
Figure 2.
Figure 2. A
A comprehensive
comprehensiveset
setofoften
tenrandom
randomaugmentations,
augmentations, denoted
denoted as as (A–J),
(A–J), that
that were
were strategi-
strategically
cally employed to enhance the dataset size and elevate the model’s generalizability during the train-
employed to enhance the dataset size and elevate the model’s generalizability during the training
ing phase.
phase.
3.3. Network
3.3. Network Architecture
Architecture
The encoder–decoder-based
The encoder–decoder-based architecture
architecture is
is aa classical
classical U-NET
U-NET with
with MobileNetV2
MobileNetV2 asas
the pre-trained encoder; however, U-NET is a fundamental convolutional neural
the pre-trained encoder; however, U-NET is a fundamental convolutional neural network network
(CNN), initially developed by Olaf Ronneberger et al. [29] for biomedical image analysis,
and has received appreciation in the medical imaging community. On the other side, Mo-
bileNetV2 [33,34] introduced lightweight convolutions in the encoder part of the network
and achieved highly accurate results with much fewer parameters. Additionally, skip con-
nections were established with the Relu activation function to increase the model’s con-
Bioengineering 2023, 10, 981 7 of 13
(CNN), initially developed by Olaf Ronneberger et al. [29] for biomedical image analysis,
and has received appreciation in the medical imaging community. On the other side, Mo-
bileNetV2 [33,34] introduced lightweight convolutions in the encoder part of the network
and achieved highly accurate results with much fewer parameters. Additionally, skip
connections were established with the Relu activation function to increase the model’s
convergence to connect encoder layers to decoder layers, which further allowed the concate-
nation of feature maps with different spatial resolutions. The encoder takes an image as the
input of the model and extracts necessary features and relevant information, whereas the
decoder learns to generate the corresponding predictions (probability maps). Furthermore,
skip connections in the down-sampling path are concatenated with feature maps in the
up-sampling path to provide local information to global information.
Figure 3. A structural visualization of the network architecture, where the encoder exhibited on the
Figure 3. A structural visualization of the network architecture, where the encoder exhibited on the
left side is MobileNetV2 and the U-NET decoder is shown on the right side. Input of patch size 256
left side is MobileNetV2 and the U-NET decoder is shown on the right side. Input of patch size
x 256 was given into the model. Convolutional units were used with batch normalization and Relu
256 × 256activations.
function was given Up-sampling
into the model. Convolutional
along units were
with concatenated usedchannels
feature with batch normalization
were employed toand
ob-
Relu function activations. Up-sampling along with concatenated
tain the output of the same spatial size as that of the input. feature channels were employed to
obtain the output of the same spatial size as that of the input.
3.5. Evaluation Parameters
We used the following performance evaluation matrices to measure the robustness
of the classifier.
Name Value
Input size 255 × 255
Batch size 8
Learning rate 1 × 10–4
Epoch 90
Activation head sigmoid
Optimizer Adam
Loss function Ldice
2TP
DSC = (2)
2TP + FP + FN
4. Results
We present the prediction results from our devised segmentation model, evaluated
using the MSD-2018 lung tumor segmentation dataset. We used U-NET architecture by
integrating the down-sampling path of the U-NET with a pre-trained MobileNetV2 encoder
that was trained on a large ImageNet dataset. The prediction maps generated from the
proposed network are shown in Figure 4. The dice score achieved by the network is
0.8793 and the recall and precision of model are 0.8602 and 0.9322, respectively. Moreover,
the distribution of the dice score coefficient of each patient is illustrated in a histogram
Bioengineering 2023, 10, 981 9 of 13
(shown in Figure 5), and the average dice score that we achieved is 0.8793. Therefore,
the proposed method trained the deep neural network and validated it with10the
Bioengineering 2023, 10, x FOR PEER REVIEW
Medical
of 14
Segmentation Decathlon (MSD) lung CT scan dataset, showing competitive results as
Bioengineering 2023, 10, x FOR PEER REVIEW 10 of 14
compared with the state-of-the-art methods.
Figure 4. Example CT scans of different patients are exhibited in the form of rows. First row indi-
Figure 4. Example
cates the
CT scans
actual images, middle
of different
row is the
patients are exhibited
visualization of truein
in theand
labels
form ofrowrows.the First row indicates
Figure 4. Example CT scans of different patients are exhibited the form last
of rows.isFirst segmenta-
row indi-
thetion
actual images, middle row is the visualization of true labels and last row is the segmentation
cates predictions, wherein
the actual images, most row
middle of theis prediction results of
the visualization aretrue
correctly
labels segmented
and last rowasisvisualized in (a–
the segmenta-
predictions,
d) and
tion verywherein mostare
few wherein
predictions, of them of omitted
most the prediction
of the results
by the model
prediction asare
results correctly
depicted
are segmented
in (e).
correctly segmented asas visualized
visualized in in(a–
(a–d) and
very fewvery
d) and of them
few ofare omitted
them by thebymodel
are omitted as depicted
the model in (e).
as depicted in (e).
Figure 5. DSC distribution of the test dataset, wherein histogram shows number of tumors that
Figure DSC distribution
5. particular
achieve dice scoreof the test dataset, wherein histogram showsthenumber of tumors
with that
Figure 5. DSC distribution of coefficient. The histogram
the test dataset, graphically
wherein histogram illustrates
shows frequency
number of tumors that
which
achieve different tumor
particular diceinstances
score achieve particular
coefficient. The DSC values.
achieve particular dice score coefficient. The histogram graphically illustrates the frequency with with
histogram graphically illustrates the frequency
which
whichdifferent
different tumor instancesachieve
tumor instances achieve particular
particular DSC
DSC values.
values.
Bioengineering 2023, 10, 981 10 of 13
Author Contributions: Conceptualization, Z.R. and S.A.; Methodology, Z.R. and M.S.I.; Data Cura-
tion, Z.R. and S.A.; Writing—Original Draft Preparation, Z.R. and B.K.; Writing—Review and Editing,
B.K., S.A. and S.K.; Visualization, S.A., B.K. and S.K.; Supervision, S.A. and M.S.I. All authors have
read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: All data are available in the paper.
Acknowledgments: The authors would like to thank Health @ InnoHK (Hong Kong Centre for
Cerebro-Cardiovascular Health Engineering (COCHE)), the Center for Eye and Vision Research,
Shatin, Hong Kong, SAR, China, and Mälardalen University, Sweden, for providing a feasible
environment to perform the experiments and document the data.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Manikandan, T.; Devi, B.; Helanvidhya, T. A Computer-Aided Diagnosis System for Lung Cancer Detection with Automatic
Region Growing, Multistage Feature Selection and Neural Network Classifier. Int. J. Innov. Technol. Explor. Eng. 2019, 9, 409–413.
2. Drozdzal, M.; Chartrand, G.; Vorontsov, E.; Shakeri, M.; Di Jorio, L.; Tang, A.; Romero, A.; Bengio, Y.; Pal, C.; Kadoury, S. Learning
normalized inputs for iterative estimation in medical image segmentation. Med. Image Anal. 2018, 44, 1–13. [CrossRef] [PubMed]
3. Kamal, U.; Rafi, A.M.; Hoque, R.; Wu, J.; Hasan, M.K. Lung cancer tumor region segmentation using recurrent 3D-denseunet. In
Proceedings of the Thoracic Image Analysis: Second International Workshop, TIA 2020, Held in Conjunction with MICCAI 2020,
Lima, Peru, 8 October 2020; Springer International Publishing: Berlin/Heidelberg, Germany; pp. 36–47.
4. Niranjana, G.; Ponnavaikko, M. A review on image processing methods in detecting lung cancer using CT images. In Proceedings
of the 2017 International Conference on Technical Advancements in Computers and Communications (ICTACC), Melmaurvathur,
India, 10–11 April 2017; pp. 18–25.
5. Qureshi, I.; Yan, J.; Abbas, Q.; Shaheed, K.; Riaz, A.B.; Wahid, A.; Khan, M.W.J.; Szczuko, P. Medical image segmentation using
deep semantic-based methods: A review of techniques, applications and emerging trends. Inf. Fusion 2022, 90, 316–352.
6. Wang, R.; Lei, T.; Cui, R.; Zhang, B.; Meng, H.; Nandi, A.K. Medical image segmentation using deep learning: A survey. IET
Image Process. 2022, 16, 1243–1267. [CrossRef]
7. Zhou, Z.; Sodha, V.; Rahman Siddiquee, M.M.; Feng, R.; Tajbakhsh, N.; Gotway, M.B.; Liang, J. Models genesis: Generic autodidactic
models for 3D medical image analysis. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI
2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Proceedings Part IV 22. pp. 384–393.
Bioengineering 2023, 10, 981 12 of 13
8. Van Opbroek, A.; Achterberg, H.C.; Vernooij, M.W.; De Bruijne, M. Transfer learning for image segmentation by combining image
weighting and kernel learning. IEEE Trans. Med. Imaging 2018, 38, 213–224. [CrossRef]
9. Singh, G.A.P.; Gupta, P. Performance analysis of various machine learning-based approaches for detection and classification of
lung cancer in humans. Neural Comput. Appl. 2019, 31, 6863–6877. [CrossRef]
10. Anthimopoulos, M.; Christodoulidis, S.; Ebner, L.; Christe, A.; Mougiakakou, S. Lung pattern classification for interstitial lung
diseases using a deep convolutional neural network. IEEE Trans. Med. Imaging 2016, 35, 1207–1216. [CrossRef]
11. Kong, Z.; Zhang, M.; Zhu, W.; Yi, Y.; Wang, T.; Zhang, B. Data enhancement based on M2-Unet for liver segmentation in
Computed Tomography. Biomed. Signal Process. Control 2023, 79, 104032. [CrossRef]
12. Wang, Q.; Kang, W.; Wu, C.; Wang, B. Computer-aided detection of lung nodules by SVM based on 3D matrix patterns. Clin.
Imaging 2013, 37, 62–69. [CrossRef]
13. Hossain, S.; Najeeb, S.; Shahriyar, A.; Abdullah, Z.R.; Haque, M.A. A pipeline for lung tumor detection and segmentation from
CT scans using dilated convolutional neural networks. In Proceedings of the 2019 IEEE International Conference on Acoustics,
Speech and Signal Processing (ICASSP 2019), Brighton, UK, 12–17 May 2019; pp. 1348–1352.
14. Valente, I.R.S.; Cortez, P.C.; Neto, E.C.; Soares, J.M.; de Albuquerque, V.H.C.; Tavares, J.M.R. Automatic 3D pulmonary nodule
detection in CT images: A survey. Comput. Methods Programs Biomed. 2016, 124, 91–107. [CrossRef]
15. Mukhlif, A.A.; Al-Khateeb, B.; Mohammed, M. Classification of breast cancer images using new transfer learning techniques.
Iraqi J. Comput. Sci. Math. 2023, 4, 167–180.
16. Lu, S.; Zhu, Z.; Gorriz, J.M.; Wang, S.H.; Zhang, Y.D. NAGNN: Classification of COVID-19 based on neighboring aware
representation from deep graph neural network. Int. J. Intell. Syst. 2022, 37, 1572–1598. [CrossRef]
17. Naqi, S.M.; Sharif, M.; Lali, I.U. A 3D nodule candidate detection method supported by hybrid features to reduce false positives
in lung nodule detection. Multimed. Tools Appl. 2019, 78, 26287–26311. [CrossRef]
18. Choi, W.-J.; Choi, T.-S. Automated pulmonary nodule detection based on three-dimensional shape-based feature descriptor.
Comput. Methods Programs Biomed. 2014, 113, 37–54. [CrossRef]
19. Usman, M.; Lee, B.-D.; Byon, S.-S.; Kim, S.-H.; Lee, B.-I.; Shin, Y.-G. Volumetric lung nodule segmentation using adaptive roi with
multi-view residual learning. Sci. Rep. 2020, 10, 12839. [CrossRef] [PubMed]
20. Setio, A.A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; Van Riel, S.J.; Wille, M.M.W.; Naqibullah, M.; Sánchez, C.I.; Van
Ginneken, B. Pulmonary nodule detection in CT images: False positive reduction using multi-view convolutional networks. IEEE
Trans. Med. Imaging 2016, 35, 1160–1169. [CrossRef]
21. Lu, S.; Wang, S.-H.; Zhang, Y.-D. Detection of abnormal brain in MRI via improved AlexNet and ELM optimized by chaotic bat
algorithm. Neural Comput. Appl. 2021, 33, 10799–10811. [CrossRef]
22. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48.
23. Dao, T.; Gu, A.; Ratner, A.; Smith, V.; De Sa, C.; Ré, C. A kernel theory of modern data augmentation. In Proceedings of the
International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 1528–1537.
24. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor
segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [CrossRef]
25. Brosch, T.; Tang, L.Y.; Yoo, Y.; Li, D.K.; Traboulsee, A.; Tam, R. Deep 3D convolutional encoder networks with shortcuts for
multiscale feature integration applied to multiple sclerosis lesion segmentation. IEEE Trans. Med. Imaging 2016, 35, 1229–1239.
[CrossRef]
26. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.-W.; Heng, P.-A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor
segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [PubMed]
27. Isensee, F.; Petersen, J.; Klein, A.; Zimmerer, D.; Jaeger, P.F.; Kohl, S.A.A.; Wasserthal, J.; Koehler, G.; Norajitra, T.; Wirkert, S.J.;
et al. nnU-Net: Self-adapting Framework for U-Net-Based Medical Image Segmentation. arXiv 2018, arXiv:1809.10486.
28. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from
sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016: 19th
International Conference, Athens, Greece, 17–21 October 2016; Proceedings Part II 19. pp. 424–432.
29. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of
the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich,
Germany, 5–9 October 2015; Proceedings Part III 18. pp. 234–241.
30. Mukhlif, A.A.; Al-Khateeb, B.; Mohammed, M.A. An extensive review of state-of-the-art transfer learning techniques used in
medical imaging: Open issues and challenges. J. Intell. Syst. 2022, 31, 1085–1111.
31. Mukhlif, A.A.; Al-Khateeb, B.; Mohammed, M.A. Incorporating a Novel Dual Transfer Learning Approach for Medical Images.
Sensors 2023, 23, 570. [PubMed]
32. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; Van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens,
G.; Menze, B. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv
2019, arXiv:1902.09063.
33. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 4510–4520.
34. Littman, A.J.; Thornquist, M.D.; White, E.; Jackson, L.A.; Goodman, G.E.; Vaughan, T.L. Prior lung disease and risk of lung cancer
in a large prospective study. Cancer Causes Control 2004, 15, 819–827. [CrossRef]
Bioengineering 2023, 10, 981 13 of 13
35. Wang, S.; Zhou, M.; Liu, Z.; Liu, Z.; Gu, D.; Zang, Y.; Dong, D.; Gevaert, O.; Tian, J. Central focused convolutional neural networks:
Developing a data-driven model for lung nodule segmentation. Med. Image Anal. 2017, 40, 172–183.
36. Sun, W.; Zheng, B.; Qian, W. Automatic feature learning using multichannel ROI based on deep structured algorithms for
computerized lung cancer diagnosis. Comput. Biol. Med. 2017, 89, 530–539.
37. Shen, W.; Zhou, M.; Yang, F.; Yu, D.; Dong, D.; Yang, C.; Zang, Y.; Tian, J. Multi-crop convolutional neural networks for lung
nodule malignancy suspiciousness classification. Pattern Recognit. 2017, 61, 663–673.
38. Wang, S.; Zhou, M.; Gevaert, O.; Tang, Z.; Dong, D.; Liu, Z.; Jie, T. A multi-view deep convolutional neural networks for lung
nodule segmentation. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and
Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 1752–1755.
39. Liu, H.; Cao, H.; Song, E.; Ma, G.; Xu, X.; Jin, R.; Jin, Y.; Hung, C.-C. A cascaded dual-pathway residual network for lung nodule
segmentation in CT images. Phys. Medica 2019, 63, 112–121.
40. Shakibapour, E.; Cunha, A.; Aresta, G.; Mendonça, A.M.; Campilho, A. An unsupervised metaheuristic search approach for
segmentation and volume measurement of pulmonary nodules in lung CT scans. Expert Syst. Appl. 2019, 119, 415–428. [CrossRef]
41. Mahbod, A.; Tschandl, P.; Langs, G.; Ecker, R.; Ellinger, I. The effects of skin lesion segmentation on the performance of
dermatoscopic image classification. Comput. Methods Programs Biomed. 2020, 197, 105725. [PubMed]
42. Monteiro, M.; Figueiredo, M.A.; Oliveira, A.L. Conditional random fields as recurrent neural networks for 3d medical imaging
segmentation. arXiv 2018, arXiv:1807.07464.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.