Next Article in Journal
Sensor-Based Gait and Balance Assessment in Healthy Adults: Analysis of Short-Term Training and Sensor Placement Effects
Previous Article in Journal
A Zero-Shot Learning Approach for Blockage Detection and Identification Based on the Stacking Ensemble Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Performance Evaluation of YOLO-Type Detectors on a New Open Fire and Smoke Dataset

by
Constantin Catargiu
,
Nicolae Cleju
and
Iulian B. Ciocoiu
*
Faculty of Electronics, Telecommunications and Information Technology, Gheorghe Asachi Technical University of Iasi, Bd. Carol I 11A, 700506 Iasi, Romania
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5597; https://doi.org/10.3390/s24175597 (registering DOI)
Submission received: 26 July 2024 / Revised: 22 August 2024 / Accepted: 28 August 2024 / Published: 29 August 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
The paper introduces a new FireAndSmoke open dataset comprising over 22,000 images and 93,000 distinct instances compiled from 1200 YouTube videos and public Internet resources. The scenes include separate and combined fire and smoke scenarios and a curated set of difficult cases representing real-life circumstances when specific image patches may be erroneously detected as fire/smoke presence. The dataset has been constructed using both static pictures and video sequences, covering day/night, indoor/outdoor, urban/industrial/forest, low/high resolution, and single/multiple instance cases. A rigorous selection, preprocessing, and labeling procedure has been applied, adhering to the findability, accessibility, interoperability, and reusability specifications described in the literature. The performances of the YOLO-type family of object detectors have been compared in terms of class-wise Precision, Recall, Mean Average Precision (mAP), and speed. Experimental results indicate the recently introduced YOLO10 model as the top performer, with 89% accuracy and a mAP@50 larger than 91%.

1. Introduction

Fire is one of the most hazardous and potentially fatal calamities, threatening the environment, properties, and human lives [1]. According to the most recent report of the Center for Fire Statistics affiliated with the International Association of Fire and Rescue Services, 3.7 million fire-related calls were registered in 2022 in 55 countries surveyed. About 19,600 people died and 55,600 were injured during these incidents, including a significant number of on-duty firefighters [2].
The adverse effects on the environment and the global community of both forest/wildfires and those occurring within urban areas are significant and cannot be overstated. Given the escalating pace of population concentration and an increased emphasis on safety, early fire detection has become a primary concern in developing smart cities. Classical sensor-based detection systems [3] may not be suitable in certain scenarios, such as vast coverage areas or untamed locations, as they can generate numerous false alerts and sometimes fail to trigger. While effective for identifying small flames, they often prove ineffective for larger, rapidly spreading fires. On the other hand, smoke detectors possess a limited ability to differentiate between different types of smoke, such as those generated by pollution, cigarettes, cooking fumes, and actual fire [4]. For example, according to the Duisburg Fire Department, only 84% of fire detectors’ automatically triggered fire alarm calls correspond to real fire situations [5]. Additionally, 44% of false positive alarms are triggered by aerosols, with nearly 20% being dust and water aerosols [6].
A viable alternative to traditional fire alarm systems is offered by modern AI/computer vision solutions [7]. More specifically, deep learning fire/smoke detection models can automatically learn relevant features, effectively overcoming the redundancy and interference associated with hand-crafted engineering. They enable the real-time identification and precise localization of fire-related incidents, providing clear performance metrics regarding confidence scores and speed while minimizing false alarm rates.
The present paper introduces a new open fire/smoke dataset with more than 22,000 images and 93,000 distinct instances compiled from static pictures and video sequences. The dataset was generated based on a rigorous selection, preprocessing, and labeling procedure that complies with the FAIR (findability, accessibility, interoperability, and reusability) specifications described in [8]. Providing data in multiple formats, the use of a multiple-step annotation procedure for images covering a broad range of realistic situations, along with a curated set of challenging cases exhibiting real-life circumstances when specific image patches may be erroneously detected as fire/smoke presence represent the main characteristics of this challenging dataset. Data are available on the GitHub platform [9], along with the Python code for downloading and a statistical evaluation of the distribution of the various image categories.
We comparatively evaluated the performances of several recent members of the YOLO (You Only Look Once) family of efficient real-time object detectors [10] on the proposed dataset. YOLOv5–v10 and YOLO-NAS models were compared in terms of class-wise Precision, Recall, Mean Average Precision (mAP), and speed. The performances of the models were evaluated through extensive experimentation on realistic case studies, showcasing the practical deployment of the system in real-world settings. The various sources of variability include indoor/outdoor environments, urban/industrial/ship/forest, low/high resolution, and single/multiple instance scenarios. The GitHub project page also includes Python scripts that generate metrics containing valuable information for the inference tasks. Those metrics provide useful info such as the total number of detections for each class, the average confidence of these detections per class, and the number of incorrect predictions made by each model. Additionally, we computed the average speed (fps), the total inference time, and the total number of detections across all classes.
The main contributions of the paper are as follows:
  • The generation of a novel large open dataset that includes challenging images that may (but should not) be detected as fire instances;
  • A two-step annotation procedure that combines an initial automatic solution based on the use of the YOLO8 detection model with a second manual correction and bounding box adjustment phase that also removes labeling errors;
  • A performance comparison of seven YOLO models, including the most recent YOLO10 member of this family of object detectors;
  • A comparative analysis of the detection efficiency of the analyzed models for daylight vs. night-time images.
The following sections describe existing datasets and machine learning models presented in the literature for accurately identifying fire and/or smoke, along with the protocol used to generate the new FireAndSmoke dataset. The experimental results reveal comparative performances of seven distinct members of the YOLO family of object detector models, while conclusions and suggestions for further work are finally outlined.

2. Datasets and Models for Fire/Smoke Identification

2.1. Fire and Smoke Datasets

Existing public datasets exhibiting fire and/or smoke scenes vary to a large extent in terms of the number of images/instances, content (fire and/or smoke, non-fire/smoke), resolution, and weather conditions. Additional sources of variability include the type of sensors used (visible spectrum/infrared cameras), the actual acquisition setup (space-borne/air-borne/ground-borne cameras), or the specific environment scenario (wildfire/forest/urban/industrial/ship). Table 1 illustrates a list of public datasets largely used in the literature.
Some papers have considered fire or smoke scenarios only, others aim at discriminating fire/smoke scenes from general non-fire/smoke situations, while a relatively short list of datasets includes images affected by poor weather conditions (haze, fog, or rain). Moreover, specific environments are only considered in many cases, such as forest fires or flames bursting on ships.
The present FireAndSmoke dataset contains a broad range of fire and smoke scenarios, such as indoor fires, building fires, car fires, forest/wildfires, and ship fires. The actual images largely vary in terms of resolution, lighting, and weather conditions. The new dataset additionally contains challenging situations when image patches could mistakenly be classified as fire instances. Those are marked as the “other” category and include natural and artificial light sources such as sunsets/sunrises, different types of clouds, streetlamps, bonfire glows, reflections, vehicle headlights, emergency lights of various vehicles, or lights from public electronic screens.

2.2. Machine Learning Models for Fire/Smoke Identification

Fire/smoke identification strategies typically follow one of three approaches, namely classification, detection, or segmentation, while combinations of those have also been considered. Classification methods process the images as a whole, aiming to separate the ones containing fire and/or smoke instances. Detection-based solutions take a step further by additionally indicating the actual localization and extension of the regions of interest within an image. At the same time, segmentation accurately delineates the exact set of pixels that belong to fire or smoke patches. From a practical perspective, detection- and segmentation-oriented approaches are preferred since they may offer invaluable information regarding the spread, dynamics, and possible fast expansion toward critical areas (houses, fuel storage facilities, etc.).
Classical machine learning algorithms targeting fire/smoke identification tasks typically rely on feature extraction techniques such as flame-specific chromatograms, shapes, textures, and flame motion. For instance, in [18], the authors selected the YCbCr color space to differentiate the luminance from the chrominance information better, yielding superior results compared to RGB-based solutions. Poobalan et al. [19] employed RGB and HSL filters along with a bag-of-features (BoF) model to detect fire color, improving classification performances. The use of the Lucas–Kanade optical flow algorithm was suggested in [20], with potential smoke areas determined using a background estimate and a color-based decision procedure.
In the realm of flame-texture feature extraction, Cui et al. [21] employed wavelet analysis and gray-level co-occurrence matrices (GLCMs) to analyze fire smoke texture. Yu et al. [22] proposed a real-time fire smoke detection approach based on gray-level co-occurrence matrices (GLCMs), incorporating a neural network for texture classification. Chino et al. [23] innovated by combining color features and texture classification in superpixel areas for static image fire detection. Ye et al. [24] introduced a dynamic texture descriptor using surface transform and a hidden Markov tree (HTM) model.
Through their innovative approach, Toulouse et al. [25] identified various geometrical features of flames, including location, rate of spread, length, and surface area. They categorized pixels representing fire based on color, while non-refractive pixels were utilized to detect smoke, sorted by their average intensity. Avgeris et al. [26] developed an edge computing framework for early fire detection, streamlining border identification significantly. However, these computer-vision-based frameworks were primarily employed on static fire images. Recently, researchers have employed fast Fourier transform (FFT) and wave variation techniques to analyze wildfire boundaries in videos [27], albeit with effectiveness limited to specific scenarios.
Researchers in [28] developed a method for fire detection by analyzing smoke and flame motion with linear dynamic systems. By incorporating color, motion, and spatial-temporal features into their model, they achieved high detection rates and a significant reductions in false alarms. Researchers also analyzed forest fire locations by examining the spatial and temporal dynamics of fire textures [29]. Hybrid surface descriptors were employed in a static texture analysis to generate a feature vector capable of differentiating flames or distortions without relying on conventional texture descriptors. However, challenges such as poor picture/video quality or adverse weather conditions necessitate implementing modern supplementary methods to enhance current approaches.
Handcrafted feature extraction, as described above, can become complex and labor-intensive. Learning-based methods that automatically extract relevant details from images, mainly based on deep learning architectures, have consistently demonstrated improved accuracy performances and lower false alarm rates than conventional approaches. Ba et al. [30] introduced Smoke Net, a novel CNN model incorporating spatial and flow attention mechanisms for improved feature representations in visual classifications. Luo et al. [31] proposed a CNN-based approach for flame identification, leveraging the kinetic characteristics of smoke. They employed dynamic frame references from the background and foreground and a CNN architecture to extract candidate pixel highlights automatically. Zhang et al. [32] used a Faster R-CNN approach to detect wildland forest fires while additionally developing a novel technique for generating synthetic smoke images. Chaoxia et al. [33] introduced an enhanced Faster R-CNN method for flame detection by incorporating a color-guided anchoring strategy and embedded global information-guided flame detection.
Recent novel approaches target reducing the total number of model parameters by employing innovative pruning techniques. As such, the Distillation for Knowledge Transfer to Pruned Model (DK-TPM) [34] uses feature distillation with smoothing (FDS) to transfer flame features from the teacher model into the student model.
The Dynamic Deformable Adaptive Framework (DDAF) [35] dynamically detects the flame region using the so-called Deformable Convolution Network v2 to cope with deformations and scale variations of the flames.
A distinct line of research has considered transformed-based architectures and associated attention mechanisms to implement both detection and combined segmentation+classification approaches. Proposed solutions include the Swin Transformer [17], TransFire, and TransUNet architectures [36], although the computational cost is significantly higher than existing CNN variants.
You Only Look Once (YOLO) models stand as state-of-the-art object detectors due to their demonstrated speed, precision, and adaptability [10]. Built around CNNs, the first versions of the model operate by dividing an image into a grid of cells, each representing a distinct area of the image. Within this grid, a set of bounding boxes is utilized to detect objects, while a predefined set of classes is associated with these regions. Upon receiving an input image, YOLO analyzes the grid cells, categorizes them based on the objects detected within, and constructs bounding boxes around these objects, assigning them a corresponding class. Based on the PyTorch framework, the YOLOv5 model improves feature extraction by using CSPDarknet53 as the backbone and additionally offering multi-scale feature fusion capabilities. The YOLOv7 foundation layer is represented by the Extended Efficient Layer Aggregation Network (E-ELAN), which enables faster and more precise detection performances. YOLOv8 introduces additional enhancements such as anchor-free detection, larger multi-scale feature maps, and Distribution Focal Loss. YOLOv9 builds on the solid foundation of previous models by introducing two extra features, namely Programmable Gradient Information (PGI) and a Generalized Efficient Layer Aggregation Network (GELAN) [37]. YOLOv10, the most recent member of the YOLO family, introduces the innovative idea of Non-Maximum-Suppression (NMS)-free training by implementing a consistent dual assignments training method to increase speed [38]. Finally, the YOLO-NAS model [39] is based on a Neural Architecture Search (NAS) procedure that looks for optimal architectural choices regarding the number of layers, layer types, kernel sizes, and connectivity patterns.
Each YOLO model offers various tradeoffs between accuracy and speed and comes with different sizes such as nano (-n), small (-s), medium (-m), balanced (-b), large (-l), and extra-large (-x).
Within the fire/smoke detection framework, Park et al. proposed a fire detection method for urban areas utilizing static Elastic-YOLOv3 during night-time [40]. Mukhiddinov et al. [41] proposed an early wildfire smoke detection system utilizing enhanced YOLOv5 photographs captured by unmanned aerial vehicles (UAVs). Dou et al. [42] introduced the Convolutional Block Attention Module (CBAM), BiFPN, and transposed convolution into YOLOv5, significantly enhancing the algorithm’s accuracy in fire detection. Subsequently, Du et al. [43] integrated ConvNeXt into YOLOv7, creating a fire detection model that effectively balances the detection speed and accuracy. Chen et al. [44] employed Ghost Shuffle Convolution to construct lightweight modules, reducing the model’s parameter count and accelerating the convergence speed in YOLOv7. Talaat et al. [45] proposed an enhanced method for smart city fire detection based on the YOLOv8 algorithm, capitalizing on deep learning to detect specific fire features in real time. Zhang et al. [46] proposed a change in the YOLOv8 architecture, where GhostnetV2-C2F is integrated into the algorithm’s backbone to facilitate long-range attention with minimal computational resources. The Detectron2 model has also been used for (forest) fire detection [47], showing robust performances even for small fires situated at long distances for both daylight and night-time scenarios.

3. Protocol for Dataset Generation

Figure 1 illustrates the protocol for generating the new FireAndSmoke dataset. It includes the image collection and composition, the preprocessing steps, and the annotation procedure. Finally, the efficiency of the YOLO-type object detector family is evaluated on the new set using classical performance indicators.

3.1. Dataset Composition

We collected a wide range of images from various real-world scenes, including buildings on fire, cars on fire, vegetation fires, wildfires, trash fires, and interior fires. The scenes include aerial views from drones, close-up and distant shots, as well as images captured from below, where the fire is at a higher altitude. The fire instances vary in magnitude and size and may be obscured by vegetation or other objects. The images were captured at different times throughout the day, from early morning to late night. Moreover, the images were taken under various weather conditions, including sunny, cloudy, rainy, and foggy scenarios. Overall, the comprehensive nature of the FireAndSmoke dataset, with its wide range of scenes, perspectives, sizes, and environment scenarios, makes it highly suitable for training models that can accurately and reliably detect and analyze fires and smoke presence.
The dataset includes images from over 1200 videos (manually downloaded from YouTube) and static images manually downloaded from Google, totaling 22,970 images classified into three distinct classes: fire, smoke, and other. Including the “other” class serves the purpose of labeling objects that may resemble fire or smoke but are not related to fire incidents. Instances marked as “other” include natural and artificial light sources such as sunsets/sunrises, different types of clouds, streetlamps, bonfire glows, reflections, vehicle headlights, emergency lights of various vehicles, or lights from public electronic screens, all of which could be mistaken for fire. This approach enhances the model’s ability to distinguish between relevant and irrelevant features in images, ultimately improving the overall accuracy and reliability of the fire detection system.
Details on the actual dataset composition are given below:
(a)
Daylight scenes (15,812 images):
-
2121 smoke-only scenes (thin or dense, small/medium/large size, far/medium/close distance)
-
2475 fire-only scenes (small/medium/large size, far/medium/close distance, car/building/trash/wood burning)
-
645 other-only scenes (different types of clouds, sunrises/sunsets, car headlights, or other types of shiny surfaces)
-
4312 fire-and-smoke scenes (small/medium/large size, far/medium/close distance, car/building/trash/wood burning)
-
3138 drone-captured scenes containing fire-and-smoke or just smoke-only images (small/medium/large size, far/medium/close distance, car/building/trash/wood burning)
-
3121 vegetation fires (forest fires, grass fires, wildfires)
(b)
Night-time scenes (7158 images):
-
2657 fire-and-smoke scenes (small/medium/large size, far/medium/close distance, car/building/wood burning)
-
1852 fire, smoke, and other scenes (small/medium/large size, far/medium/close distance, car/building burning, car lights, streetlamps, illuminated advertisement panels)
-
1228 fire-only scenes (small/medium/large size, far/medium/close distance, car/building burning)
-
1421 other-only scenes (small/medium/large size, far/medium/close distance, car lights, streetlamps, illuminated advertisement panels)
Images were normalized to the dimensions required by the models, and histogram equalization was employed to compensate for the illumination variability. Augmentation techniques were applied based on horizontal flip and rotation with ±20°. Examples demonstrating the diversity of the actual images are given in Figure 2, Figure 3 and Figure 4.

3.2. Data Labeling Procedure

To avoid manually annotating all collected images, we adopted a multi-step approach. Initially, we annotated 1000 images containing fire, smoke, and other instances, creating a small dataset hosted on the Roboflow platform [48]. We then trained a YOLOv8m model using this dataset to identify fire and smoke instances. Subsequently, we used this trained model to run inferences on the remaining images. The bounding box coordinates and labels generated by the model during this prediction process were saved as. txt files (using the YOLOv8 format). The images in the dataset and the corresponding labels were uploaded on the Roboflow platform, where we reviewed all the annotated images to ensure the accuracy of bounding boxes and labels. Any incorrectly annotated images were removed, and adjustments were made to bounding box dimensions or locations as needed. Additionally, any labeling errors were corrected to guarantee the robustness, quality, and reliability of the dataset.
We built on the availability of a broad range of tools that Roboflow provides for data conversion and preprocessing, further streamlining the training process. The images were resized to a consistent size of 640 × 640 pixels to ensure uniformity across the dataset. Once the dataset was successfully reviewed, we exported it in various formats, such as YOLOv5–v8 PyTorch. The same dataset format as YOLOv8 can be used for YOLO-NAS training, ensuring compatibility across different YOLO models.

4. Fire/Smoke Detection Evaluation

Detection performances were assessed in terms of speed (expressed as the number of frames per second), class-wise Precision (P), Recall (R), and Mean Average Precision (mAP@50) evaluated at both 50% and in the 50–95% IoU (intersection-over-union) range for the three classes under study. The IoU value (also known as the Jaccard index) quantifies the degree of overlap between two bounding boxes (e.g., the ground truth and the one generated by an object detection algorithm) as the ratio between their corresponding intersection and reunion areas, respectively.
Precision and Recall parameters are defined as follows:
P   =   T P T P   +   F P ;   R   =   T P T P   +   F N
where TN is the true negative rate, FN is the false negative rate, TP is the true positive rate, and FP is the false positive rate.
When the overlap between the predicted box and the ground truth bounding box exceeds a given threshold value (e.g., IoU > 50%), the detection is considered correct, otherwise it is counted as a false positive. If the predicted box around an actual fire/smoke instance is missing, we get a false negative case. The F1-score, defined as the harmonic mean of Precision and Recall, may offer valuable information in cases exhibiting a significant imbalance between the number of examples available in the categories under study.
The Precision–Recall curve (computed for a given IoU threshold value) illustrates the relationship between these two metrics for different confidence scores associated with the instance detections of a specific object provided by a given model, while the area under the curve (denoted as AP) quantifies the performance of the model on this task. The Mean Average Precision (mAP) calculates the average precision across multiple classes or categories and provides a single scalar value that summarizes the overall model performance.
All experiments were conducted on an RTX 4090/24GB GPU running PyTorch models. Details on the experimental setup parameters are given in Table 2 and Table 3. The data split is indicated in Table 4.
Table 5 presents comparative performance results on the test set images for the seven models considered, showing a mild superiority of YOLOv8 and YOLOv10 variants. mAP@50 values are situated in the range exhibited by other models evaluated on different datasets [7] and speed performances indicate real-time operation.
Examples of correct fire/smoke detections are given in Figure 5, Figure 6, Figure 7 and Figure 8. The figures illustrate the best performances and false detections made by the models under study.
To better understand the comparative performances of the models under study, we generated two test video sequences containing challenging scenarios that are difficult to evaluate even by human subjects. Those are called VideoDayMixt.mp4 (18,680 frames) and VideoNightMixt.mp4 (10,645 frames) and are hosted on the GitHub page of the project [9]. We additionally provide scripts implementing detailed performance metrics measuring the total number of detections per class, the average confidence score of the detections per class within specific value ranges (e.g., the number of detections for a specific class with a confidence score above 0.8), and the total number of incorrect predictions made by the models. The performances are reported in Table 6 and Table 7.

5. Discussion

It is difficult to appropriately compare the performances of different models introduced in the literature since they are not typically evaluated on the same dataset, whose composition may vary significantly in terms of dimensionality, diversity, resolution, or background complexity. Nevertheless, the experimental results indicated previously revealed that all models achieved high detection performances, with mAP@50 values reaching 91% for the “fire” and “smoke” classes. As shown in Table 8, those compare favorably with existing results, although some papers only consider fire or smoke detection separately and seldom use “other” class instances.
All the evaluated models successfully coped with realistic situations where other light sources resembling fire may translate into false positive predictions. Moreover, all models exhibited real-time operation, with inference speeds above 100 fps. Overall, the YOLOv5 model performed fastest on inference tasks, while the YOLOv10 model provided the most robust results, reliably detecting fire and smoke instances in images affected by significant variations in scale and illumination conditions. It performed best even for very difficult smoke detection situations, where the background or the sky raises serious challenges. Despite being slower than the other variants, YOLO-NAS excelled in predicting fire instances at high distances, confirming its known capacity for identifying small objects.
Regarding the training time, YOLOv7 required the longest duration (≈8 h) due to its high number of parameters (about 36 million). YOLO-NAS trained for about 6 h, although its number of parameters is much smaller (19 million). This may be attributed to the expansive search space that covers various architectural configurations and hyperparameter settings.
The performance evaluation of the models on the challenging daylight and night-time videos revealed that YOLOv5 and YOLOv10 offered around 0.8 precision scores. Moreover, both models exhibited high-confidence detection rates, making them ideal for critical applications such as monitoring residential areas or industrial settings where timely and accurate detection can prevent disasters. The results suggest that YOLOv10 yields the highest precision on detected scenes, a much-desired trait for cases when false positives must be minimized, such as deploying first responders at distant or dangerous locations.
In terms of detection volume, YOLOv6 and YOLOv7 were notable. On the two test videos, YOLOv6 recorded the highest number of detections, with 14,635 for fire and 22,145 for smoke, albeit at lower precision scores of 0.730 and 0.766, respectively. YOLOv7 logged 14,761 fire detections and 21,895 smoke detections with slightly better precision scores of 0.770 and 0.758, respectively. These models are suited for applications prioritizing coverage over accuracy, such as vast industrial complexes or forest areas where missing a single event could have severe consequences.
YOLOv6 and YOLOv8 provided a balanced approach with a good number of detections and acceptable precision levels. These models are suitable for general surveillance and monitoring, where a balance between detection volume and accuracy is needed.
Explainable artificial intelligence (XAI) may increase the transparency and reveal the decision-making operation of various learning-based models (including the source of incorrect behavior), thus paving the way for their broad application in real-life tasks. While the literature on this topic related to classification tasks is quite large, the references focusing on object detection are relatively scarce.
This is mainly due to the essential differences between classification and object detection tasks. While the former uses (single) categorical output objectives, the second employs regression-type targets that consider the object class probabilities, the localization information, and the objectness score, directly affecting defining proper similarity metrics that account for all those components.
Significant XAI approaches for object detection include model agnostic solutions such as Detector-Randomized Input Sampling for Explanation of Black-box Models (D-RISE) [49] and the Gaussian Class Activation Mapping Explainer (G-CAME) [50].
Identified as a random masking perturbation-based approach, the D-RISE model builds on the success of the previously introduced RISE [51] technique to yield saliency maps that would explain detection decisions. The method uses masks to separate particular regions of an input image, passes the masked images through the analyzed model, and assesses the importance of each area based on a weighting score associated with each mask that combines both the object class probabilities and the localization information.
G-CAME is a Class Activation Mapping (CAM)-based method that also yields a saliency map for each detected object. The algorithm operates by adding a Gaussian kernel as the weight for each pixel in the activation maps of specific layers in the model’s architecture and is more performant and computationally faster than D-RISE.
Table 8. Performance comparison with other fire/smoke detection systems.
Table 8. Performance comparison with other fire/smoke detection systems.
SystemModelPrecisionRecallF1mAP@50Dataset Size
Tian [52]YOLOv5n0.920.910.910.94>10,000
Avazov [53]YOLOv70.970.940.960.974622
Talaat [45] YOLOv80.930.940.940.97>25,000
Luo [31]YOLOX0.860.720.780.845000
Geng [54] YOLOFM0.940.940.940.9718,644
This workYOLO100.860.850.850.91>22,000

6. Conclusions

The paper introduces a new annotated image dataset including various fire and smoke scenarios derived from real-world events. These encompass diverse settings such as indoor and outdoor fires and both small and large-scale flames/smoke, under largely varying lighting conditions. In order to minimize false positive detections, the dataset incorporates “other class” examples of non-fire or non-smoke scenarios, forcing the models to accurately discriminate between genuine and false instances of fire and smoke.
The comparative performance results of the YOLO family of object detectors reported in the previous sections revealed that each model might find utility for specific operational requirements, ensuring optimal performance in diverse surveillance and safety environments. The analysis concluded that YOLOv5 and YOLO10 are most suitable for applications requiring high accuracy and reliability due to their superior precision and confidence detection rates in both fire and smoke scenarios. While YOLOv10 is recommended when minimizing false positives is paramount, YOLOv7 offers advantages in scenarios requiring many detections.
Integrating deep learning models within a video processing pipeline may facilitate the automatic analysis of surveillance footage, enabling timely response and mitigation actions. We have performed preliminary tests with an end-to-end solution that integrates our YOLO-based fire detection models within the DeepStream framework from Nvidia [55]. Images captured by (visible spectrum) cameras are transmitted through a 5G communication network to a processing server that yields a dashboard that includes valuable insights for the authorized users (firefighters, emergency services). The application will report fire incidents, the actual location of the incident, and the real-time video flow from the location.
Additional comparative experiments will be conducted considering transformer-based alternatives and ensemble learning methods. To effectively use examples of images that may resemble (but not actually represent) fire instances, we will consider a contrastive learning approach (e.g., by means of Siamese-type networks) that has proved effective in classification and object tracking tasks. One essential subject to be addressed is a comparative analysis of the effect of explainability methods on the models under study. From a practical perspective, an exciting research topic could consider the efficiency of the various fire and smoke detection approaches in adverse weather conditions (rain, fog, and snow).

Author Contributions

Conceptualization and methodology, equal contribution of all authors; software and validation, C.C.; formal analysis, N.C. and I.B.C.; data curation, C.C.; writing—review and editing, equal contribution of all authors. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data and accompanying Python code for downloading are freely available on the GitHub platform: https://github.com/CostiCatargiu/NEWFireSmokeDataset_YoloModels (accessed on 26 July 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kumar, A. Preserving life on earth. In Adaptation, Ecosystem-Based; Elsevier: Amsterdam, The Netherlands, 2022; pp. 503–602. [Google Scholar]
  2. International Association of Fire and Rescue Services. Available online: www.ctif.org/news/world-fire-statistics-report-no-29-comprehensive-analysis-fire-statistics-2022 (accessed on 2 July 2024).
  3. Zhang, F.; Zhao, P.; Xu, S.; Wu, Y.; Yang, X.; Zhang, Y. Integrating multiple factors to optimize watchtower deployment for wildfire detection. Sci. Total Environ. 2020, 737, 139561. [Google Scholar] [CrossRef]
  4. Chagger, R.; Smith, D. The Causes of False Fire Alarms in Buildings; Tech. Rep. BC2982, no. 1; BRE Trust: Hertfordshire, UK, 2014. [Google Scholar]
  5. Statistics 2008; Duisburg Fire Brigade: Duisburg, Germany, 2009.
  6. Kruell, W.; Schultze, T.; Tobera, R.; Willms, I. Analysis of dust properties to solve the complex problem of non-fire sensitivity testing of optical smoke detectors. Procedia. Eng. 2013, 62, 859–867. [Google Scholar] [CrossRef]
  7. Saleh, A.; Zulkifley, M.A.; Harun, H.H.; Gaudreault, F.; Davison, I.; Spraggon, M. Forest fire surveillance systems: A review of deep learning methods. Heliyon 2024, 10, E23127. [Google Scholar] [CrossRef] [PubMed]
  8. Wilkinson, M.D.; Dumontier, M.; Jan Aalbersberg, I.J.; Appleton, G.; Axton, M.; Baak, A.; Blomberg, N.; Boiten, J.-W.; da Silva Santos, L.B.; Bourne, P.E.; et al. The FAIR guiding principles for scientific data management and stewardship. Sci. Data 2016, 3, 160018. [Google Scholar] [CrossRef]
  9. Fire and Smoke Dataset. Available online: https://thub.com/CostiCatargiu/NEWFireSmokeDataset_YoloModels (accessed on 26 July 2024).
  10. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of YOLO architectures in computer vision: From YOLOv1 to YOLOv8 and YOLO-NAS. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  11. Foggia, P.; Saggese, A.; Vento, M. Real-Time fire detection for video-surveillance applications using a combination of experts based on color, shape, and motion. IEEE Trans. Circ. Syst. Video Technol. 2015, 25, 1545–1556. [Google Scholar] [CrossRef]
  12. Li, S.; Yan, Q.; Liu, P. An Efficient fire detection method based on multiscale feature extraction, implicit deep supervision and channel attention mechanism. IEEE Trans. Image Process. 2020, 29, 8467–8475. [Google Scholar] [CrossRef] [PubMed]
  13. Yar, H.; Hussain, T.; Agarwal, M.; Khan, Z.A.; Gupta, S.K.; Wook, S. Baik Optimized dual fire attention network and medium-scale fire classification benchmark. IEEE Trans. Image Process. 2022, 31, 6331–6343. [Google Scholar] [CrossRef]
  14. Dilshad, N.; Khan, S.U.; Alghamdi, N.S.; Taleb, T.; Song, J. Toward efficient fire detection in IoT environment: A modified attention network and large-scale data set. IEEE Internet Things J. 2024, 11, 13467–13481. [Google Scholar] [CrossRef]
  15. Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P.Z.; Blasch, E. Aerial imagery pile burn detection using deep learning: The FLAME dataset. Comput. Netw. 2021, 193, 108001. [Google Scholar] [CrossRef]
  16. Yar, H.; Ullah, W.; Khan, Z.A.; Baik, S.W. An effective attention-based CNN model for fire detection in adverse weather conditions. ISPRS J. Photogramm. Remote Sens. 2023, 206, 335–346. [Google Scholar] [CrossRef]
  17. Wang, M.; Yue, P.; Jiang, L.; Yu, D.; Tuo, T.; Li, J. An open flame and smoke detection dataset for deep learning in remote sensing based fire detection. Geo-Spat. Inf. Sci. 2024, 1–16. [Google Scholar] [CrossRef]
  18. Çelik, T.; Demirel, H. Fire detection in video sequences using a generic color model. Fire Saf. J. 2009, 44, 147–158. [Google Scholar] [CrossRef]
  19. Poobalan, K.; Liew, S.-C. Fire detection based on color filters and Bag-of-Features classification. In Proceedings of the 2015 IEEE Student Conference on Research and Development, Kuala Lumpur, Malaysia, 13–14 December 2015. [Google Scholar]
  20. Yu, C.; Fang, J.; Wang, J.; Zhang, Y. Video fire smoke detection using motion and color features. Fire Technol. 2010, 46, 651–663. [Google Scholar]
  21. Cui, Y.; Dong, H.; Zhou, E. An early fire detection method based on smoke texture analysis and discrimination. In Proceedings of the 2008 Congress on Image and Signal Processing, Sanya, China, 27–30 May 2008. [Google Scholar]
  22. Yu, C.; Zhang, Y.; Fang, J.; Wang, J. Texture analysis of smoke for real-time fire detection. In Proceedings of the 2009 Second Internat Workshop on Computer Science and Engineering, Qingdao, China, 28–30 October 2009. [Google Scholar]
  23. Chino, D.Y.T.; Avalhais, L.P.S.; Rodrigues, J.F.; Traina, A.J.M. Bowfire: Detection of fire in still images by integrating pixel color and texture analysis. In Proceedings of the 2015 28th SIBGRAPI Conference on Graphics, Patterns and Images, Salvador, Brazil, 26–29 August 2015. [Google Scholar]
  24. Ye, W.; Zhao, J.; Wang, S.; Wang, Y.; Zhang, D.; Yuan, Z. Dynamic texture based smoke detection using Surfacelet transform and HMT model. Fire Saf. J. 2015, 73, 91–101. [Google Scholar] [CrossRef]
  25. Toulouse, T.; Rossi, L.; Celik, T.; Akhloufi, M. Automatic fire pixel detection using image processing: A comparative analysis of rule-based and machine learning-based methods. Signal Image Video Process. 2016, 10, 647–654. [Google Scholar] [CrossRef]
  26. Avgeris, M.; Spatharakis, D.; Dechouniotis, D.; Kalatzis, N.; Roussaki, I.; Papavassiliou, S. Where there is fire there is SMOKE: A scalable edge computing framework for early fire detection. Sensors 2019, 19, 639. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, Z.; Zhao, J.; Zhang, D.; Qu, C.; Ke, Y.; Cai, B. Contour based forest fire detection using FFT and wavelet. In Proceedings of the 2008 International Conference on Computer Science and Software Engineering, Wuhan, China, 12–14 December 2008. [Google Scholar]
  28. Dimitropoulos, K.; Barmpoutis, P.; Grammalidis, N. Spatio-temporal flame modeling and dynamic texture analysis for automatic video-based fire detection. IEEE Trans. Circuits Syst. Video Technol. 2015, 25, 339–351. [Google Scholar] [CrossRef]
  29. Prema, C.E.; Vinsley, S.S.; Suresh, S. Efficient flame detection based on static and dynamic texture analysis in forest fire detection. Fire Technol. 2018, 54, 255–288. [Google Scholar] [CrossRef]
  30. Ba, R.; Chen, C.; Yuan, J.; Song, W.; Lo, S. SmokeNet: Satellite smoke scene detection using convolutional neural network with spatial and channel-wise attention. Remote Sens. 2019, 11, 1702. [Google Scholar] [CrossRef]
  31. Luo, M.; Xu, L.; Yang, Y.; Cao, M.; Yang, J. Laboratory flame smoke detection based on an improved YOLOX algorithm. Appl. Sci. 2022, 12, 12876. [Google Scholar] [CrossRef]
  32. Zhang, Q.X.; Lin, G.H.; Zhang, Y.M.; Xu, G.; Wang, J.J. Wildland forest fire smoke detection based on Faster R-CNN using synthetic smoke images. Procedia Eng. 2018, 211, 441–446. [Google Scholar] [CrossRef]
  33. Chaoxia, C.; Shang, W.; Shang, W.; Zhang, F. Information-guided flame detection based on Faster R-CNN. IEEE Access 2020, 8, 58923–58932. [Google Scholar] [CrossRef]
  34. Tao, H.; Liu, J.; Yang, Z.; Wang, G.; Shang, J.; Qiu, H.; Gao, L. Revolutionizing flame detection: Novelization in flame detection through transferring distillation for knowledge to pruned model. Expert Syst. Appl. 2024, 249, 123787. [Google Scholar] [CrossRef]
  35. Tao, H.; Wang, G.; Liu, J.; Yang, Z. A deep learning-based dynamic deformable adaptive framework for locating the root region of the dynamic flames. PLoS ONE 2024, 19, e0301839. [Google Scholar] [CrossRef]
  36. Ghali, R.; Akhloufi, M.A.; Mseddi, W.S. Deep learning and transformer approaches for UAV-based wildfire detection and segmentation. Sensors 2022, 22, 1977. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
  38. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar] [CrossRef]
  39. YOLO-NAS Documentation. Available online: https://github.com/Deci-AI/super-gradients (accessed on 2 July 2024).
  40. Park, M.; Ko, B.C. Two-step real-time night-time fire detection in an urban environment using static ELASTIC-YOLOv3 and temporal fire-tube. Sensors 2020, 20, 2202. [Google Scholar] [CrossRef]
  41. Mukhiddinov, M.; Abdusalomov, A.B.; Cho, J. A wildfire smoke detection system using unmanned aerial vehicle images based on the optimized YOLOv5. Sensors 2022, 22, 9384. [Google Scholar] [CrossRef]
  42. Dou, Z.; Zhou, H.; Liu, Z.; Hu, Y.; Wang, P.; Zhang, J.; Wang, Q.; Chen, L.; Diao, X.; Li, J. An improved YOLOv5s fire detection model. Fire Technol. 2023, 60, 135–166. [Google Scholar] [CrossRef]
  43. Du, H.; Zhu, W.; Peng, K.; Li, W. Improved high speed flame detection method based on YOLOv7. Open J. Appl. Sci. 2022, 12, 2004–2018. [Google Scholar] [CrossRef]
  44. Chen, G.; Cheng, R.; Lin, X.; Jiao, W.; Bai, D.; Lin, H. LMDFS: A lightweight model for detecting forest fire smoke in UAV images based on YOLOv7. Remote Sens. 2023, 15, 3790. [Google Scholar] [CrossRef]
  45. Talaat, F.M.; Eldin, H.Z. An improved fire detection approach based on YOLO-v8 for smart cities. Neural. Comp. Appl. 2023, 35, 20939–20954. [Google Scholar] [CrossRef]
  46. Zhang, Z.; Tan, L.; Tiong, R.L.K. Ship-Fire Net: An improved YOLOv8 algorithm for ship fire detection. Sensors 2024, 24, 727. [Google Scholar] [CrossRef] [PubMed]
  47. Abdusalomov, A.B.; Islam, B.M.S.; Nasimov, R.; Mukhiddinov, M.; Whangbo, T.K. An Improved Forest Fire Detection Method Based on the Detectron2 Model and a Deep Learning Approach. Sensors 2023, 23, 1512. [Google Scholar] [CrossRef]
  48. Fire and Smoke Dataset Health Check. Available online: https://universe.roboflow.com/catargiuconstantin2/firesmokenewdataset/health (accessed on 2 July 2024).
  49. Petsiuk, V.; Jain, R.; Manjunatha, V.; Morariu, V.I.; Mehra, A.; Ordonez, V.; Saenko, K. Black-box explanation of object detectors via saliency maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11443–11452. [Google Scholar]
  50. Nguyen, K.; Nguyen, T.T.H.; Nguyen, V.T.K.; Truong, V.B.; Phan, T.; Cao, H. Efficient and Concise Explanations for Object Detection with Gaussian-Class Activation Mapping Explainer. In Proceedings of the 37th Canadian Conference on Artificial Intelligence, Guelph, ON, Canada, 27–31 May 2024. [Google Scholar]
  51. Petsiuk, V.; Das, A.; Saenko, K. RISE: Randomized input sampling for explanation of black-box models. In Proceedings of the British Machine Vision Conference, Northumbria University, Newcastle, UK, 3–6 September 2018. [Google Scholar]
  52. Tian, H.; Zheng, Y.; Jin, Z. Improved RetinaNet model for the application of small target detection in the aerial images. In IOP Conference Series: Earth and Environmental Science; IOP Publishing: Bristol, UK, 2020; Volume 585, p. 012142. [Google Scholar]
  53. Avazov, K.; Jamil, M.K.; Muminov, B.; Abdusalomov, A.B.; Cho, Y.I. Fire detection and notifcation method in ship areas using deep learning and computer vision approaches. Sensors 2023, 23, 7078. [Google Scholar] [CrossRef]
  54. Geng, X.; Su, Y.; Cao, X.; Li, H.; Liu, L. YOLOFM: An improved fire and smoke object detection algorithm based onYOLOv5n. Sci. Rep. 2024, 14, 4543. [Google Scholar] [CrossRef]
  55. Nvidia DeepStream Documentation. Available online: https://developer.nvidia.com/deepstream-sdk (accessed on 2 July 2024).
Figure 1. Overview of the dataset generation protocol.
Figure 1. Overview of the dataset generation protocol.
Sensors 24 05597 g001
Figure 2. Examples of images containing fire, smoke, and combined instances in daylight images.
Figure 2. Examples of images containing fire, smoke, and combined instances in daylight images.
Sensors 24 05597 g002
Figure 3. Examples of images containing fire, smoke, and combined instances in night-time images.
Figure 3. Examples of images containing fire, smoke, and combined instances in night-time images.
Sensors 24 05597 g003
Figure 4. Examples of images containing “other” class instances.
Figure 4. Examples of images containing “other” class instances.
Sensors 24 05597 g004
Figure 5. Examples of correct “smoke” detection by YOLO-type models and false positive “fire” detection by YOLOv5–v7 models.
Figure 5. Examples of correct “smoke” detection by YOLO-type models and false positive “fire” detection by YOLOv5–v7 models.
Sensors 24 05597 g005aSensors 24 05597 g005b
Figure 6. Examples of correct “fire” and “smoke” detections by YOLO-type models, except for YOLOv8 which misses the fire instance detection.
Figure 6. Examples of correct “fire” and “smoke” detections by YOLO-type models, except for YOLOv8 which misses the fire instance detection.
Sensors 24 05597 g006aSensors 24 05597 g006b
Figure 7. Examples of correct “fire”, “smoke”, and “other” instances detection by the YOLO-type models, except for YOLOv8 which generates a false positive “fire” detection.
Figure 7. Examples of correct “fire”, “smoke”, and “other” instances detection by the YOLO-type models, except for YOLOv8 which generates a false positive “fire” detection.
Sensors 24 05597 g007aSensors 24 05597 g007b
Figure 8. Challenging night-time scenes showing correct “smoke” and “other” class detection by YOLO-type models, while YOLO-NAS misses detecting the smoke instance.
Figure 8. Challenging night-time scenes showing correct “smoke” and “other” class detection by YOLO-type models, while YOLO-NAS misses detecting the smoke instance.
Sensors 24 05597 g008aSensors 24 05597 g008b
Table 1. Fire/smoke open datasets.
Table 1. Fire/smoke open datasets.
DatasetClassesNum ImagesSourceContent
Foggia et al. [11]Fire/Non-fire62,690Ground cameras’ video frames Indoor/outdoor scenes
Li et al. [12]Fire/Non-fire50,000Ground cameras’ video frames and static images Indoor/outdoor scenes
Yar et al. [13]Fire/Non-fire3804Ground cameras’ video framesOutdoor scenes
Dilshad et al. [14]Fire/Non-fire47,124Ground cameras’
images
Foggy, low-light outdoor scenes
Shamsoshoara et al. [15]Fire/Smoke/Other39,375Air-borne visible spectrum/infrared camerasForest fires
Yar et al. [16]Fire/Non-fire6000Air/space-borne
cameras
Haze, fog, night outdoor scenes
Wang et al. [17]Fire/Smoke/Other122,634Ground/air-borne camerasIndoor/outdoor scenes
Table 2. Experimental setup parameters.
Table 2. Experimental setup parameters.
Parameter NameType
Operating systemUbuntu 22.04.3 LTS
CPUIntel Core 9 Gen i9-13900KF (5.8 GHz)
GPURTX 4090 24 GB
RAMDDR5 64 GB (6000 MHz)
Deployment environmentPython 3.10.2
Deep learning frameworkPyTorch 2.1.2
Accelerated computing
architecture
CUDA 12.1
Table 3. Training parameters of the deep learning models.
Table 3. Training parameters of the deep learning models.
YOLOv5YOLOv6YOLOv7YOLOv8YOLOv9YOLOv10YOLO-NAS
Epochs100100100100100100100
Batch size64443264163264
Optim. alg.SGDSGDSGDSGDSGDSGDAdam
Model weightsyolov5myolov6myolov7yolov8myolov9-cyolov10myolo_nas_s
Num. params.
(million)
21.234.935.925.950.719.119
Table 4. The train/validation/test dataset split for the experiments.
Table 4. The train/validation/test dataset split for the experiments.
TrainValidationTest
Nr of images20,10847572417
Fire instances19,86444902353
Smoke instances18,01742822249
Other instances10,64826911332
Table 5. The detection performances of the deep learning models on the test set.
Table 5. The detection performances of the deep learning models on the test set.
ModelFire Smoke Other Speed
(fps)
PRmAP
@50
mAP
@50–95
F1PRmAP
@50
mAP
@50–95
F1PRmAP
@50
mAP
@50–95
F1
YOLOv50.860.860.910.690.860.880.820.890.660.850.800.780.840.690.79210
YOLOv60.840.860.910.640.850.890.810.910.630.840.780.730.810.560.75215
YOLOv70.830.880.910.640.850.870.840.900.620.860.760.790.840.580.77230
YOLOv80.860.850.910.690.850.880.840.910.680.850.820.770.850.630.79230
YOLOv90.850.850.910.690.850.880.850.910.680.850.760.680.840.630.72115
YOLOv100.860.850.910.690.850.890.830.910.680.860.820.740.840.630.78210
YOLO-NAS--0.84----0.80----0.70---
Table 6. Comparative performances of deep learning models on VideoDayMixt.mp4 video.
Table 6. Comparative performances of deep learning models on VideoDayMixt.mp4 video.
YOLOv5YOLOv6YOLOv7YOLOv8YOLOv9YOLOv10YOLO-NAS
Confidence score > 0.5
Number
of detections
(Average confidence score)
Fire13,025
(0.799)
14,635 (0.730)14,761
(0.770)
12,703 (0.743)12,874
(0.719)
12,968
(0.769)
15,414
(0.717)
Smoke21,694 (0.792)22,145 (0.766)21,895
(0.758)
20,948 (0.768)20,204
(0.759)
20,733
(0.796)
17,439
(0.722)
Confidence score > 0.8
Number
of detections
(Average confidence score)
Fire8937 (0.882)5570 (0.885)7949 (0.883)4762 (0.865)2733 (0.876)6715
(0.885)
4447
(0.891)
Smoke13,033 (0.892)11,530 (0.890)10,303
(0.885)
11,954 (0.880)10,227
(0.887)
13,389
(0.895)
5405
(0.885)
Speed (fps)20721417919711621031
Table 7. Comparative performances of deep learning models on VideoNightMixt.mp4 video.
Table 7. Comparative performances of deep learning models on VideoNightMixt.mp4 video.
YOLOv5YOLOv6YOLOv7YOLOv8YOLOv9YOLOv10YOLO-NAS
Confidence score > 0.5
Number
of detections
(Average confidence score)
Fire7053
(0.836)
7472
(0.772)
7533 (0.796)7114 (0.786)6358
(0.778)
6606
(0.835)
5408
(0.781)
Smoke6105 (0.744)5777 (0.673)6152
(0.693)
4752 (0.722)5181
(0.721)
4621
(0.743)
3020
(0.720)
Confidence score > 0.8
Number
of detections
(Average confidence score)
Fire5540 (0.891)4216 (0.889)4981
(0.889)
4762 (0.865)3716
(0.883)
5086
(0.896)
2968
(0.890)
Smoke2699
(0.889)
1705 (0.887)1867 (0.881)1959 (0.880)2107
(0.890)
2214
(0.893)
940
(0.892)
Speed (fps)18021924524011421631
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Catargiu, C.; Cleju, N.; Ciocoiu, I.B. A Comparative Performance Evaluation of YOLO-Type Detectors on a New Open Fire and Smoke Dataset. Sensors 2024, 24, 5597. https://doi.org/10.3390/s24175597

AMA Style

Catargiu C, Cleju N, Ciocoiu IB. A Comparative Performance Evaluation of YOLO-Type Detectors on a New Open Fire and Smoke Dataset. Sensors. 2024; 24(17):5597. https://doi.org/10.3390/s24175597

Chicago/Turabian Style

Catargiu, Constantin, Nicolae Cleju, and Iulian B. Ciocoiu. 2024. "A Comparative Performance Evaluation of YOLO-Type Detectors on a New Open Fire and Smoke Dataset" Sensors 24, no. 17: 5597. https://doi.org/10.3390/s24175597

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop