Applsci 13 11548 v3
Applsci 13 11548 v3
sciences
Article
Fire Detection and Geo-Localization Using UAV’s Aerial Images
and Yolo-Based Models
Kheireddine Choutri 1, *, Mohand Lagha 1 , Souham Meshoul 2, * , Mohamed Batouche 2 , Farah Bouzidi 1
and Wided Charef 1
1 Aeronautical Sciences Laboratory, Aeronautical and Spatial Studies Institute, Blida 1 University,
Blida 0900, Algeria; laghamohand@univ-blida.dz (M.L.)
2 Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint
Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia; mabatouche@pnu.edu.sa
* Correspondence: choutri.kheireddine@univ-blida.dz (K.C.); sbmeshoul@pnu.edu.sa (S.M.)
Abstract: The past decade has witnessed a growing demand for drone-based fire detection systems,
driven by escalating concerns about wildfires exacerbated by climate change, as corroborated by
environmental studies. However, deploying existing drone-based fire detection systems in real-
world operational conditions poses practical challenges, notably the intricate and unstructured
environments and the dynamic nature of UAV-mounted cameras, often leading to false alarms and
inaccurate detections. In this paper, we describe a two-stage framework for fire detection and geo-
localization. The key features of the proposed work included the compilation of a large dataset
from several sources to capture various visual contexts related to fire scenes. The bounding boxes of
the regions of interest were labeled using three target levels, namely fire, non-fire, and smoke. The
second feature was the investigation of YOLO models to undertake the detection and localization
tasks. YOLO-NAS was retained as the best performing model using the compiled dataset with an
average mAP50 of 0.71 and an F1_score of 0.68. Additionally, a fire localization scheme based on
stereo vision was introduced, and the hardware implementation was executed on a drone equipped
with a Pixhawk microcontroller. The test results were very promising and showed the ability of the
proposed approach to contribute to a comprehensive and effective fire detection system.
- We defined the problem of wildfire observation using UAVs and developed a localiza-
tion algorithm that uses a stereo vision system and camera calibration.
- Wd designed and controlled a quadcopter based on Pixhawk technology, enabling its
real-time testing and validation.
This paper is structured as follows. In Section 2, we survey the existing literature
on fire detection using UAVs, fire localization, and vision-based detection methods. In
Section 3, we provide an outline of the proposed two-stage framework for fire detection
and localization. In Section 4, we describe the methods used in the first stage devoted to
offline fire detection using YOLO models. In Section 5, we present the experimental study
and results of the first phase. In Section 6, we describe the camera calibration and pose
estimation processes for fire detection and localization in a real scenario. The results of
the experiments are reported in Section 7. Finally, in Section 8, we draw conclusions and
outline plans for future work.
2. Related Work
Our work spans over three main areas, namely using UAVs for fire detection and
monitoring, object localization using stereovision, and vision-based models for fire detection
and localization. In the following sections, we review the related work in each of these aspects.
to establish the 3D coordinates of the fire’s location or, more accurately, its relative position
concerning the camera’s perspective. Achieving an accurate fire location stands as a critical
prerequisite for facilitating swift firefighting responses and precise water injection.
Predominantly, a stereo vision-based system centered on generating a disparity map
was introduced. By computing the camera’s meticulous calibration data, the fire’s 3D real-
world coordinates could be derived. Consequently, an increasing amount of research has
delved into fire positioning and its three-dimensional modeling using stereo vision sensors.
For instance, the work described in [23] used a stereo vision sensor system for 3D fire loca-
tion, successfully applying it within a coal chemical company setting. The authors in [24]
harnessed a combination of a stereo infrared camera and laser radar to capture fire imagery
in smoky environments. This fusion sensor achieved accurate fire positioning within clean
and intricate settings, although its effective operational distance remained limited to under
10 m. Similar investigations were conducted in [25], constrained by the infrared camera’s
operational range and the stereo vision system’s base distance. The system proved suitable
only for identifying and locating fires at short distances. The authors in [26] established
a stereo vision system with a 100 mm base distance for 3D fire modeling and geometric
analysis. Outdoor experiments exhibited reliable fire localization when the depth distance
was within 20 m. However, the stereo vision system for fire positioning encountered
challenges, as discussed in [27,28]. The calibration accuracy significantly influenced the
light positioning outcomes, with the positioning precision declining as distance increased.
Moreover, the system’s adaptability to diverse light positioning distances remained limited.
Thus, multiple techniques and solutions have been proposed and adopted to enhance
the outcomes.
3. Outline of thedecrease
Proposedthe Framework
prediction time, the model achieved a 97.29% accuracy in discerning fire from
non-fire images, focusing on unmonitored forest observations.
In this paper, we proposed a fire detection and geo-localization scheme, as shown in
Figure 1, where a3.two-stage
Outline of the Proposed is
framework Framework
shown. The ultimate goal of the fire detection
In this paper,
and localization system using UAV images waswe proposed a firethe
detection
timelyand geo-localization
identification andscheme,
preciseasspa-shown in
Figure 1, where a two-stage framework is shown. The ultimate
tial localization of fire-related incidents. The whole process started at the first stage with goal of the fire detection
and localization system using UAV images was the timely identification and precise spatial
data collection and integration. Many data sources were used to compile a large dataset
localization of fire-related incidents. The whole process started at the first stage with data
of images that represented
collection and various visual Many
integration. contexts.
dataThe second
sources werekeyusedcomponent
to compile of the pro-
a large dataset of
posed frameworkimageswas the data preparation, which included data augmentation, labeling,
that represented various visual contexts. The second key component of the proposed
and splitting. Then, a thirdwas
framework critical component
the data preparation, ofwhich
the system
included entailed the utilization
data augmentation, of and
labeling,
splitting. Then, a third critical component of the system
innovative YOLO models for object detection, specifically geared towards recognizing entailed the utilization of innovative
YOLO fire,
three primary classes: models for object
non-fire, detection,
and smoke.specifically
These YOLO geared towardswere
models recognizing
trainedthree
using primary
classes: fire, non-fire, and smoke. These YOLO models were trained using the labeled
the labeled images from the data preparation phase in order to identify fire-related objects
images from the data preparation phase in order to identify fire-related objects amidst
amidst complex visual
complexcontexts. The best
visual contexts. Theperforming
best performing YOLOYOLO model
modelwaswasthen
then used
used ininthe
thesecond
second stage of the proposed
stage framework
of the proposed to detect
framework andand
to detect localize
localizevarious areasininthe
various areas the images
images captured
captured by UAVs by equipped
UAVs equipped withwithhigh-resolution
high-resolution cameras.
cameras.
The finalmost
The final and arguably and arguably most vital component
vital component of the
of the system system revolved
revolved aroundaround
preciseprecise
fire localization, achieved using advanced stereo vision techniques. This stage comprised
fire localization, achieved using advanced stereo vision techniques. This stage comprised
several interrelated processes. Initially, a meticulous calibration process was conducted to
several interrelated processes.
precisely Initially,
align the UAV’s acameras
meticulous calibration
and establish theirprocess was conducted
relative positions. to
This calibration
precisely align thewasUAV’s cameras
foundational, and establish
serving their relative
as a cornerstone for the positions.
subsequent This
depthcalibration
estimation. The
depth estimation, achieved through stereo vision, leveraged the disparities between the
corresponding points in the stereo images to calculate the distance to objects within the
scene. This depth information, in turn, fed into the critical position estimation step, allow-
ing the system to calculate the 3D coordinates of the fire’s location relative to the UAV’s
perspective with a high degree of accuracy. Collectively, this integrated system offered a
comprehensive and sophisticated approach for fire detection and localization. By harmo-
nizing the data collection from UAVs, YOLO-based object detection, and advanced stereo
vision techniques, it not only identified fire, smoke, and non-fire objects but also precisely
pinpointed their spatial coordinates. This precision is paramount for orchestrating swift
Appl. Sci. 2023, 13, 11548 6 of 19
Figure 2.
Figure 2. Representative
Representative samples
samples from
from the
the compiled
compiled dataset
datasetwith
withoriginal
originallabels.
labels.
• Data augmentation: In order to increase the size of the dataset, amplify the dataset’s
diversity, and subject the machine learning model to an array of visual modifications,
data augmentation was performed. This was conducted by applying geometric trans-
formations such as scaling, rotations, and various affine adjustments to the images
in the dataset. This approach enhanced the probability of the model in identifying
objects in diverse configurations and contours. We ultimately assembled a dataset
containing 12,000 images. Upon assembling the dataset, the images were uniformly
resized to dimensions of 416 × 416.
• Data labeling: The crucial labeling task was carried out manually, facilitated by the
MATLAB R2021b Image Labeler app. This task was time and effort consuming as each
of the images in the dataset underwent meticulous labeling, classifying the regions in
them into one of the three categories: fire, non-fire, or smoke. Completing this labeling
procedure entailed creating ground truth data that included details about the image
filenames and their respective bounding box coordinates.
• Data splitting: The dataset containing all the incorporated features was primed for
integration into the machine learning algorithms. Yet, prior to embarking on the
algorithmic application, it was recommended to undertake data partitioning. A hold-
out sampling technique was used for this purpose where 80% of the images were for
training and 20% for testing.
4.2.1. YOLOv8
Ultralytics introduced YOLOv8 in January 2023, expanding its capabilities to encom-
pass various vision tasks like object detection, segmentation, pose estimation, tracking, and
classification. This version retained the foundational structure of YOLOv5 while modifying
the CSPLayer, now termed the C2f module. The C2f module, integrating cross-stage partial
bottlenecks with two convolutions, merges high-level features with contextual information,
enhancing detection precision. Employing an anchor-free model with a disentangled head,
YOLOv8 processes object detection and location (termed as objectness in YOLOv8), classifi-
cation, and regression tasks independently. This approach hones each branch’s focus on
its designated task, subsequently enhancing the model’s overall accuracy. The objectness
score in the output layer employs the sigmoid function, indicating the likelihood of an
object within the bounding box. For class probabilities, the SoftMax function is utilized,
representing the object’s likelihood of belonging to specific classes. The bounding box loss
is calculated using the CIoU and DFL loss functions, while the classification loss utilizes
binary cross-entropy. These losses prove particularly beneficial for detecting smaller objects,
boosting the overall object detection performance. Furthermore, YOLOv8 introduced a
semantic segmentation counterpart called the YOLOv8-Seg model. This model features a
CSPDarknet-53 feature extractor as its backbone, replaced by the C2f module instead of the
conventional YOLO neck architecture. This module was succeeded by two segmentation
heads, responsible for predicting semantic segmentation masks [39]. More details on the
YOLOv8 architecture can be found in [39].
4.2.2. YOLO-NAS
Deci AI introduced YOLO-NAS in May 2023. YOLO-NAS was designed to address the de-
tection of small objects, augment the localization accuracy, and improve the performance–com-
putation ratio, thus rendering it suitable for real-time applications on edge devices. Its
open-source framework is also accessible for research purposes. The distinctive elements
of YOLO-NAS encompass the following.
• Quantization aware modules named QSP and QCI, integrating re-parameterization
for 8-bit quantization to minimize accuracy loss during post-training quantization.
• Automatic architecture design via AutoNAC, Deci’s proprietary NAS technology.
• A hybrid quantization approach that selectively quantizes specific segments of a model
to strike a balance between latency and accuracy, deviating from the conventional
standard quantization affecting all layers.
• A pre-training regimen incorporating automatically labeled data, self-distillation, and
extensive datasets.
AutoNAC, which played a pivotal role in creating YOLO-NAS, is an adaptable system
capable of tailoring itself to diverse tasks, data specifics, inference environments, and
performance objectives. This technology assists users in identifying an optimal structure
that offers a precise blend of accuracy and inference speed for their specific use cases.
AutoNAC accounts for the data, hardware, and other factors influencing the inference
process, such as compilers and quantization. During the NAS process, RepVGG blocks
were integrated into the model architecture to ensure compatibility with post-training
quantization (PTQ). The outcome was the generation of three architectures with varying
depths and placements of the QSP and QCI blocks: YOLO-NASS, YOLO-NASM, and
YOLO-NASL (denoting small, medium, and large). The model underwent pre-training on
Objects365, encompassing two million images and 365 categories. Subsequently, pseudo-
labels were generated using the COCO dataset, followed by training with the original
118 k training images from the COCO dataset. At present, three YOLO-NAS models have
been released in FP32, FP16, and INT8 precisions. These models achieved an average
precision (AP) of 52.2% on the MS COCO dataset using 16-bit precision [40].
Appl. Sci. 2023, 13, 11548 9 of 19
TP(l )
Precision(l ) = (1)
TP(l ) + FP(l )
The precision indicated how confident we could be that a detected region predicted to
have the positive target level (fire, non-fire, smoke) actually had the positive target level.
Recall, also known as sensitivity or the true positive rate (TPR), indicates how confident
we could be that all the detected regions with the positive target level (fire, non-fire, smoke)
were found. It was defined as the ratio of TP to the sum of TP and FN, as shown in
Equation (2).
TP(l )
Recall (l ) = (2)
TP(l ) + FN (l )
The mean average precision at an intersection over union (IOU) threshold of
0.5 (mAP50) was also used to assess the performance of the detection. Using mAP50
meant that model’s predictions were considered correct if they had at least a 50% overlap
with the ground truth bounding boxes.
For the overall performance of the model, three metrics were considered, namely
the F1_score, the arithmetic average class accuracy (arithmetic mean for short), and the
harmonic average class accuracy (harmonic mean for short). They were calculated as given
by the following equations.
Precision × Recall
F1score = 2 × (3)
Precision + Recall
1 3
3 l∑
arithmetic_average_class_accuracy = Recall (l ) (4)
=1
1
harmonic_average_class_accuracy = 1 3 1
(5)
3 ∑l =1 recall (l )
Figure 4. Example of the confidence score diversity using the YOLOv8 detector.
Figure 4. Example of the confidence score diversity using the YOLOv8 detector.
5.3. Comprehensive Comparison of the YOLO Models
Table 1 presents a comparative analysis of the performance metrics for the various
models in fire detection, namely YOLOv4, YOLOv5, YOLOv8, and YOLO-NAS. The met-
rics included the precision (P), recall (R), and mean average precision at IoU 0.5 (mAP50)
for the three distinct classes: fire, non-fire, and smoke. The results indicated variations in
the models’ abilities to accurately detect these fire-related classes.
At the class level, YOLOv4 demonstrated reasonable precision for fire (0.58), fol-
Appl. Sci. 2023, 13, 11548 11 of 19
At the class level, YOLOv4 demonstrated reasonable precision for fire (0.58), followed
by YOLOv5 (0.62), YOLOv8 (0.64), and YOLO-NAS (0.67). In terms of recall, YOLO-NAS
exhibited the highest performance (0.71), closely followed by YOLOv8 (0.67), YOLOv5
(0.66), and YOLOv4 (0.61) for the fire class. Regarding the non-fire class, YOLOv8 out-
performed the other models with the highest precision (0.76) and recall (0.76). However,
YOLO-NAS achieved the highest mAP50 (0.87), indicating a superior overall performance.
YOLOv4, YOLOv5, and YOLOv8 also yielded competitive mAP50 scores (0.66, 0.68, and
0.70, respectively). For the smoke class, YOLOv8 achieved the highest precision (0.60),
while YOLOv4 yielded the highest recall (0.50). YOLO-NAS maintained a balanced mAP50
(0.53), suggesting a satisfactory performance in the presence of smoke. Furthermore, Table 1
shows that YOLO-NAS achieved the highest macro-average precision, recall, and F1-score
among the models. These results highlight the trade-offs between precision and recall, with
YOLO-NAS demonstrating a balanced performance across the classes, making it a notable
choice for comprehensive fire detection.
6. Geo-Localization
After identifying YOLO-NAS as the best performing model for fire identification, we
now describe the material related to the second stage of the proposed framework where
YOLO-NAS was used as the object detector. We first explain how the camera calibration
and depth estimation were performed.
using the OpenCV library, renowned for its potent computer vision capabilities, encom-
passing a range of functions pertaining to calibration procedures and a suite of tools that
expedite development.
In this work, the MATLAB toolboxes [41,42] were harnessed. These toolboxes offer
intuitive and user-friendly applications designed to enhance the efficiency of both intrinsic
and extrinsic calibration processes. The outcomes derived from the MATLAB camera
calibration toolbox are presented in subsequent sections. The stereo setup consisted of two
identical cameras with uniform specifications, positioned at a fixed distance from each
other. As depicted in Figure 5, the chessboard square dimensions, serving as the input
for the camera calibration, needed to be known (in our instance, it was 28 mm). Upon
x FOR PEER REVIEW image selection, the chessboard origin and X, Y directions were automatically 13 defined.
of 20
Subsequently, exporting the camera parameters to the MATLAB workspace was essential
for their utilization in object localization.
The outcomes of the simulations yielded the essential camera parameters, such as
Table 2. Camera parameters.
the focal length, principal point, radial distortion, mean projection error, and the intrinsic
Focal Length Principal matrix.
parameters Point The focal length Intrinsic Matrix point values
and principal Radial
were Distortion
stored within a
2 × 1 vector, while the radial distortion was contained within a 3 × 1 vector. The intrinsic
- 923.1819 0 0 [0.0509,
parameters, along with the mean projection error, were incorporated into −1.0153,
a 3 × 3 matrix.
[932.1819 , 929.1559] [338.5335, 246.8962] 0
These parameters are provided in Table 2.
29.1559 0
20.1072]
338.5335 246.8962 1
Table 2. Camera parameters.
- 1.1652 × 10 0 0 [0.0240, −0.1004,
[1.1652×10Camera Principal Point 0 1.1760 × Matrix
10 0
3, 1.1760×103] [355.8581, 330.2815]
Focal Length Intrinsic Radial Distortion
2.4647]
335.8581
330.28150
923.1819 10
Camera parameters 1 [932.1819, 929.1559] [338.5335, 246.8962] 0 29.1559 0 [0.0509, −1.0153, 20.1072]
338.5335 246.8962 1
the camera extrinsic parameters. In our scenario, the calibration pattern was in motion
while the camera remained stationary. This perspective offered insights into the inter-
camera separation distance, the relative positions of the cameras, and the distance be-
tween the camera and the calibration images. This distance information contributed to
assessing the accuracy of our calibration procedure.
The re-projection errors served as a qualitative indicator of the accuracy. Such errors
Appl. Sci. 2023, 13, 11548 13 of 19
The perspective either centered on the camera or the pattern was designated as camera
centric or pattern centric, respectively. This input choice governed the presentation of the
camera extrinsic parameters. In our scenario, the calibration pattern was in motion while
the camera remained stationary. This perspective offered insights into the inter-camera
separation distance, the relative positions of the cameras, and the distance between the
camera and the calibration images. This distance information contributed to assessing the
accuracy of our calibration procedure.
The re-projection errors served as a qualitative indicator of the accuracy. Such errors
represent the disparity between a pattern’s key point (corner points) detected in a calibration
image and the corresponding world point projected onto the same image. The calibration
application presented an informative display of the average re-projection error within each
calibration image. When the overall mean re-projection error surpassed an acceptable
threshold, a crucial measure for mitigating this was to exclude the images exhibiting the
highest error and then proceed with the recalibration. Re-projection errors are influenced
by camera resolution and lenses. Notably, a higher resolution combined with wider lenses
can lead to increased errors, and conversely, narrower lenses with lower resolution can
help minimize them. Typically, a mean re-projection error of less than one pixel is deemed
Appl. Sci. 2023, 13, x FOR PEER REVIEW
satisfactory. Figure 6 below illustrates both the mean re-projection error per image in pixels
and the overall mean error of the selected images.
Figure6. 6.
Figure Re-projection
Re-projection error.error.
Figure 7.
Figure 7. Fire
Fire detected
detected in
in the
the images
images adding
adding bounding
bounding boxes.
boxes.
This
This procedure
procedureensured
ensuredthetheextraction
extractionofofthe
thedepth
depthinformation
information in in
a structured manner.
a structured man-
As
ner. As depicted in Figure 8 the depth measurement outcome was recorded as 0.56m.
depicted in Figure 8 the depth measurement outcome was recorded as 0.56 m. In
In
comparison,
comparison, the actual measured distance stood at 0.65 m, resulting in an error of 9 cm.
Appl. Sci. 2023, 13, x FOR PEER REVIEW 15
This
This deviation was within
deviation was withinan anacceptable
acceptablerange,
range, attributable
attributable to factors
to factors likelike
the the camera
camera res-
resolution, potential
olution, potential calibration
calibration discrepancies,
discrepancies, and and camera
camera orientation.
orientation.
Figure8.8.Distance
Figure Distance extraction.
extraction.
where
- (u, v) are the 2D pixel coordinates in the image.
- (cx , cy ) are the principal point coordinates (intrinsic parameters).
- (fx , fy ) are the focal lengths along the x and y axes (intrinsic parameters).
- Z is the depth or distance of the point from the camera (the required value).
To obtain the depth or distance (Z) of the point from the camera, the triangulation
equations were rearranged and solved for Z using the known values of (u, v), (cx , cy ),
(fx , fy ), and the calculated values of (X, Y).
Once the depth Z was determined, the world coordinates (X, Y, Z) of the point were
derived. These coordinates represented the position of the point in the world coordinate system.
7. Experimental Results
In this section we report the results we obtained during the second stage of the
proposed framework. During this stage, a custom-built UAV was used to acquire images
of fire scenes and then YOLO-NAS was used for detecting and locating the region in the
images. The depth was then estimated, as explained in the above section.
Figure 9.
Figure 9. Final
Final UAV build.
UAV build.
Table 3. Cont.
Figure10.
Figure 10.Detection
Detectiontest
test1.1.
Table4.4.Object
Table Objectrelative
relativeposition
positiontests.
tests.
8. Conclusions
A two-stage framework for end-to-end fire detection and geo-localization using
UAVs was described in this article. The initial phase was dedicated entirely to offline fire
Appl. Sci. 2023, 13, 11548 17 of 19
8. Conclusions
A two-stage framework for end-to-end fire detection and geo-localization using UAVs
was described in this article. The initial phase was dedicated entirely to offline fire detection
and utilized four YOLO models, including the two most recent models, YOLO-NAS
and YOLO8, to determine which one was most suitable for fire detection. The models
underwent training and evaluation using a compiled dataset comprising 12,530 images,
in which regions delineating fire, non-fire, and smoke were manually annotated. The
labeling required considerable time and effort. YOLO-NAS emerged as the best performing
model among the four under consideration, exhibiting a modest superiority over YOLOv8
in each of the following metrics: precision, recall, F1_score, mAP50, and average class
accuracy. YOLO-NAS was implemented in the second stage, which incorporated the
analysis of real-life scenarios. In this stage, the images captured by a custom-built UAV
were supplied to YOLO-NAS for the purposes of object detection and localization. Geo-
localization was also considered by employing accurate camera calibration and depth
estimation techniques. The test results were extremely encouraging and demonstrated the
overall process’s viability, although it could be enhanced in numerous ways. In the future,
we intend to use optimization algorithms to fine-tune the hyper-parameters of the YOLO
models, specifically YOLO8 and YOLO-NAS, in order to further enhance their performance.
Furthermore, more advanced UAVs need to be used to evaluate the system in real-world
forest fire settings.
Author Contributions: Conceptualization, K.C. and M.L.; data curation, F.B. and W.C.; methodology,
S.M. and M.B.; software, K.C.; supervision, S.M., M.L. and M.B.; validation, M.B., K.C. and S.M.;
writing—original draft, K.C.; writing—review and editing, M.B., M.L. and K.C. All authors have read
and agreed to the published version of the manuscript.
Funding: This work was supported by the Princess Nourah bint Abdulrahman University Re-
searchers Supporting Project number (PNURSP2023R196), Princess Nourah bint Abdulrahman
University, Riyadh, Saudi Arabia.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The datasets are available from the corresponding authors upon request.
Acknowledgments: The authors would like to acknowledge the Princess Nourah bint Abdulrah-
man University Researchers Supporting Project number (PNURSP2023R196), Princess Nourah bint
Abdulrahman University, Riyadh, Saudi Arabia.
Conflicts of Interest: The authors declare no conflict of interest.
Appl. Sci. 2023, 13, 11548 18 of 19
References
1. Choutri, K.; Fadloun, S.; Lagha, M.; Bouzidi, F.; Charef, W. Forest Fire Detection Using IoT Enabled UAV And Computer
Vision. In Proceedings of the 2022 International Conference on Artificial Intelligence of Things (ICAIoT), Istanbul, Turkey,
29–30 December 2022; pp. 1–6.
2. Choutri, K.; Mohand, L.; Dala, L. Design of search and rescue system using autonomous Multi-UAVs. Intell. Decis. Technol. 2020,
14, 553–564. [CrossRef]
3. Choutri, K.; Lagha, M.; Dala, L. A fully autonomous search and rescue system using quadrotor UAV. Int. J. Comput. Digit. Syst.
2021, 10, 403–414. [CrossRef] [PubMed]
4. Khan, F.; Xu, Z.; Sun, J.; Khan, F.M.; Ahmed, A.; Zhao, Y. Recent advances in sensors for fire detection. Sensors 2022, 22, 3310.
[CrossRef] [PubMed]
5. Yu, J.; He, Y.; Zhang, F.; Sun, G.; Hou, Y.; Liu, H.; Li, J.; Yang, R.; Wang, H. An Infrared Image Stitching Method for Wind Turbine
Blade Using UAV Flight Data and U-Net. IEEE Sens. J. 2023, 23, 8727–8736. [CrossRef]
6. Hao, K.; Wang, J. Design of FPGA-based TDLAS aircraft fire detection system. In Proceedings of the Third International
Conference on Sensors and Information Technology (ICSI 2023), Xiamen, China, 15 May 2023; Volume 12699, pp. 54–61.
7. Hong, Z.; Tang, Z.; Pan, H.; Zhang, Y.; Zheng, Z.; Zhou, R.; Ma, Z.; Zhang, Y.; Han, Y.; Wang, J.; et al. Active fire detection using a
novel convolutional neural network based on Himawari-8 satellite images. Front. Environ. Sci. 2022, 10, 794028. [CrossRef]
8. Jijitha, R.; Shabin, P. A review on forest fire detection. Res. Appl. Embed. Syst. 2019, 2, 1–8.
9. Zhang, L.; Wang, M.; Fu, Y.; Ding, Y. A Forest Fire Recognition Method Using UAV Images Based on Transfer Learning. Forests
2022, 13, 975. [CrossRef]
10. Partheepan, S.; Sanati, F.; Hassan, J. Autonomous Unmanned Aerial Vehicles in Bushfire Management: Challenges and Opportu-
nities. Drones 2023, 7, 47. [CrossRef]
11. Ahmed, H.; Bakr, M.; Talib, M.A.; Abbas, S.; Nasir, Q. Unmanned aerial vehicles (UAVs) and artificial intelligence (AI) in fire
related disaster recovery: Analytical survey study. In Proceedings of the 2022 International Conference on Business Analytics for
Technology and Security (ICBATS), Dubai, United Arab Emirates, 16–17 February 2022; pp. 1–6.
12. Chino, D.Y.; Avalhais, L.P.; Rodrigues, J.F.; Traina, A.J. Bowfire: Detection of fire in still images by integrating pixel color and
texture analysis. In Proceedings of the 2015 28th SIBGRAPI Conference on Graphics, Patterns and Images, Salvador, Brazil,
26–29 August 2015; pp. 95–102.
13. Cazzolato, M.T.; Avalhais, L.; Chino, D.; Ramos, J.S.; de Souza, J.A.; Rodrigues, J.F., Jr.; Traina, A. Fismo: A compilation of datasets
from emergency situations for fire and smoke analysis. In Proceedings of the Brazilian Symposium on Databases-SBBD. SBC
Uberlândia, Minas Gerais, Brazil, 2–5 October 2017; pp. 213–223.
14. Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P.Z.; Blasch, E. Aerial imagery pile burn detection using deep learning:
The FLAME dataset. Comput. Netw. 2021, 193, 108001. [CrossRef]
15. Novo, A.; Fariñas-Álvarez, N.; Martínez-Sánchez, J.; González-Jorge, H.; Fernández-Alonso, J.M.; Lorenzo, H. Mapping Forest
Fire Risk—A Case Study in Galicia (Spain). Remote Sens. 2020, 12, 3705. [CrossRef]
16. Bouguettaya, A.; Zarzour, H.; Taberkit, A.M.; Kechida, A. A review on early wildfire detection from unmanned aerial vehicles
using deep learning-based computer vision algorithms. Signal Process. 2022, 190, 108309. [CrossRef]
17. Twidwell, D.; Allen, C.R.; Detweiler, C.; Higgins, J.; Laney, C.; Elbaum, S. Smokey comes of age: Unmanned aerial systems for fire
management. Front. Ecol. Environ. 2016, 14, 333–339. [CrossRef]
18. Skeele, R.C.; Hollinger, G.A. Aerial vehicle path planning for monitoring wildfire frontiers. In Field and Service Robotics; Springer:
Berlin/Heidelberg, Germany, 2016; pp. 455–467.
19. Beachly, E.; Detweiler, C.; Elbaum, S.; Duncan, B.; Hildebrandt, C.; Twidwell, D.; Allen, C. Fire-aware planning of aerial trajectories
and ignitions. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid,
Spain, 1–5 October 2018; pp. 685–692.
20. Kasyap, V.L.; Sumathi, D.; Alluri, K.; Ch, P.R.; Thilakarathne, N.; Shafi, R.M. Early detection of forest fire using mixed learning
techniques and UAV. Comput. Intell. Neurosci. 2022, 2022, 3170244. [CrossRef] [PubMed]
21. Weslya, U.J.; Chaitanyab, R.V.S.; Kumarc, P.L.; Kumard, N.S.; Devie, B. A Detailed Investigation on Forest Monitoring Sys-
tem for Wildfire Using IoT. In Proceedings of the First International Conference on Recent Developments in Electronics and
Communication Systems (RDECS-2022), Surampalem, India, 22–23 July 2022; Volume 32, p. 315.
22. Kustu, T.; Taskin, A. Deep learning and stereo vision-based detection of post-earthquake fire geolocation for smart cities within
the scope of disaster management: Istanbul case. Int. J. Disaster Risk Reduct. 2023, 96, 103906. [CrossRef]
23. Song, T.; Tang, B.; Zhao, M.; Deng, L. An accurate 3-D fire location method based on sub-pixel edge detection and non-parametric
545 stereo matching. Measurement 2014, 50, 160–171. [CrossRef]
24. Tsai, P.F.; Liao, C.H.; Yuan, S.M. Using deep learning with thermal imaging for human detection in heavy smoke scenarios.
Sensors 2022, 22, 5351. [CrossRef]
25. Zhu, J.; Li, W.; Lin, D.; Zhao, G. Study on water jet trajectory model of fire monitor based on simulation and experiment. Fire
Technol. 2019, 55, 773–787. [CrossRef]
26. Toulouse, T.; Rossi, L.; Akhloufi, M.; Pieri, A.; Maldague, X. A multimodal 3D framework for fire characteristics estimation. Meas.
Sci. Technol. 2018, 29, 025404. [CrossRef]
Appl. Sci. 2023, 13, 11548 19 of 19
27. McNeil, J.G.; Lattimer, B.Y. Robotic fire suppression through autonomous feedback control. Fire Technol. 2017, 53, 1171–1199.
[CrossRef]
28. Wu, B.; Zhang, F.; Xue, T. Monocular-vision-based method for online measurement of pose parameters of weld stud. Measurement
2015, 61, 263–269. [CrossRef]
29. Yuan, C.; Zhang, Y.; Liu, Z. A survey on technologies for automatic forest fire monitoring, detection, and fighting using unmanned
556 aerial vehicles and remote sensing techniques. Can. J. For. Res. 2015, 45, 783–792. [CrossRef]
30. Talaat, F.M.; ZainEldin, H. An improved fire detection approach based on YOLO-v8 for smart cities. Neural Comput. Appl. 2023,
35, 20939–20954. [CrossRef]
31. Diwan, T.; Anirudh, G.; Tembhurne, J.V. Object detection using YOLO: Challenges, architectural successors, datasets and
applications. Multimed. Tools Appl. 2023, 82, 9243–9275. [CrossRef] [PubMed]
32. Li, J.; Xu, R.; Liu, Y. An Improved Forest Fire and Smoke Detection Model Based on YOLOv5. Forests 2023, 14, 833. [CrossRef]
33. Abdusalomov, A.B.; Islam, B.M.S.; Nasimov, R.; Mukhiddinov, M.; Whangbo, T.K. An improved forest fire detection method
based on the detectron2 model and a deep learning approach. Sensors 2023, 23, 1512. [CrossRef]
34. Lu, K.; Xu, R.; Li, J.; Lv, Y.; Lin, H.; Liu, Y. A Vision-Based Detection and Spatial Localization Scheme for Forest Fire Inspection
from UAV. Forests 2022, 13, 383. [CrossRef]
35. Harjoko, A.; Dharmawan, A.; Adhinata, F.D.; Kosala, G.; Jo, K.H. Real-time Forest fire detection framework based on artificial
intelligence using color probability model and motion feature analysis. Fire 2022, 5, 23.
36. Sudhakar, S.; Vijayakumar, V.; Kumar, C.S.; Priya, V.; Ravi, L.; Subramaniyaswamy, V. Unmanned Aerial Vehicle (UAV) based
Forest Fire Detection and monitoring for reducing false alarms in forest-fires. Comput. Commun. 2020, 149, 1–16. [CrossRef]
37. Khan, A.; Hassan, B.; Khan, S.; Ahmed, R.; Abuassba, A. DeepFire: A Novel Dataset and Deep Transfer Learning Benchmark for
Forest Fire Detection. Mob. Inf. Syst. 2022, 2022, 5358359. [CrossRef]
38. Chopde, A.; Magon, A.; Bhatkar, S. Forest Fire Detection and Prediction from Image Processing Using RCNN. In Proceedings of
the 7th World Congress on Civil, Structural, and Environmental Engineering, Virtual, 10–12 April 2022.
39. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLOv8 2023; 2023. Available online: https://github.com/ultralytics/ultralytics
(accessed on 1 June 2023).
40. Aharon, S.; Dupont, L.; Masad, O.; Yurkova, K.; Fridman, L.; Lkdci; Khvedchenya, E.; Rubin, R.; Bagrov, N.; Tymchenko, B.; et al.
Super-Gradients. GitHub Repos. 2021. [CrossRef]
41. Mathwork. Computer Vision Toolbox. 2021. Available online: https://www.mathworks.com/products/computer-vision.html
(accessed on 1 May 2021).
42. Mathwork. Image Processing Toolbox. 2021. Available online: https://www.mathworks.com/products/image.html (accessed on
1 May 2021).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.