Abstract

Forest conservation is crucial for the maintenance of a healthy and thriving ecosystem. The field of remote sensing (RS) has been integral with the wide adoption of computer vision and sensor technologies for forest land observation. One critical area of interest is the detection of active forest fires. A forest fire, which occurs naturally or manually induced, can quickly sweep through vast amounts of land, leaving behind unfathomable damage and loss of lives. Automatic detection of active forest fires (and burning biomass) is hence an important area to pursue to avoid unwanted catastrophes. Early fire detection can also be useful for decision makers to plan mitigation strategies as well as extinguishing efforts. In this paper, we present a deep learning framework called Fire-Net, that is trained on Landsat-8 imagery for the detection of active fires and burning biomass. Specifically, we fuse the optical (Red, Green, and Blue) and thermal modalities from the images for a more effective representation. In addition, our network leverages the residual convolution and separable convolution blocks, enabling deeper features from coarse datasets to be extracted. Experimental results show an overall accuracy of 97.35%, while also being able to robustly detect small active fires. The imagery for this study is taken from Australian and North American forests regions, the Amazon rainforest, Central Africa and Chernobyl (Ukraine), where forest fires are actively reported.

1. Introduction

Forests, lands dominated by trees, cover approximately 4 billion hectares of the earth’s land area [1]. This is equivalent to around 29% of the earth. Forest management and conservation are therefore essential tasks for sustaining biodiversity at global scale [2]. Forest fires, which can be manmade or naturally occurring, are events that can threaten our planet leaving behind catastrophic circumstances such as damages and losses [37]. Forest fires have long-term devastating effects on ecosystems such as destroying vegetation dynamics, emission of greenhouse gases, loss of wildlife habit, and also devastation of land covers [811]. Therefore, accurate and timely detection of active forest fires is critical to minimize and/or prevent such hazards.

In recent years, advanced remote sensing (RS) technologies have been adopted to monitor and observe the earth and land covers, which can be on a large scale [1215]. Availability of high spatio-temporal resolution data and multispectral imagery allow tremendous applications of forest monitoring such as burned area mapping [7], active fire detection [11, 16], burning biomass detection [17, 18], and forest disturbance monitoring [19, 20]. It is worth noting that one widely and freely available RS data source can be used for all the tasks, which are imagery form Landsat satellites.

Active fire detection is an important application of RS in forest monitoring. The active fire detection and analysis can be used in many applications such as source point pollution in air quality analysis, initial seed point in burned area mapping, and prediction of growth and spread of active fires [21]. Until now, many researchers detect active fires using multispectral satellite imagery, assisted by computer algorithms. For instance, Wooster and Nightingale [4] presented an active fire detection method based on Sentinel-3 imagery using Land Surface Temperature Radiometer (LSTR) near nadir scans and fire characterization (potential fire pixel classification) through middle infrared radiance calculation method. Cruz et al. [22] proposed a fire detection method according to color index. Their method was based on classification of vegetation to detect the smoke and fire flames. They claimed their method could be applied in real-time using unnamed aerial vehicle (UAV) dataset. Schroeder et al. [23] studied active fire detection by exploiting Visible Infrared Imaging Radiometer Suite (VIIRS) sensor data 375 (m). Their method deployed a contextual algorithm built on the heritage MODIS (MODerate Resolution Imaging Spectroradiometer) fire and thermal anomaly product. Jang et al. [24] designed forest fire detection algorithm based on Himawari-8 geostationary satellite data for detection of fire candidate pixels with thresholding based on multi-temporal analysis. Jiao et al. [25] proposed a deep learning based forest detection algorithm based on UAV dataset. They used YOLOV3 (You Only Look Once version 3), RGB (Red, Green, and Blue) imagery. They reported the potential of YOLOV3 in detection of location of active fires. Yuan et al. [26] designed an active fire detection framework based on UAV imagery for forest areas. Their method was deployed in three stages: (1) color detection for forest candidate pixels, (2) motion detection for reducing false alarm pixels, and (3) classification fires pixels based on extracted features (direction and magnitude). Schroeder et al. [21] detected the active fire by using Landsat-8 imagery for both daytime and nighttime fires. They applied thresholding short-wave infrared, optical and near infrared bands. Amraoui et al. [27] presented the active fire detection based on Meteosat-8/SEVIRI (Spinning Enhanced Visible and Infrared Imager) dataset. The fire detection in this framework as following: (1) mask non-target object (water, desert, urban), (2) identification high potential pixels based on thresholding brightness temperature and solar zenith angle, (3) removing contaminated pixels (cloud pixels, sun glint, highly reflective surfaces) by thresholding spectral and thermal bands, and (4) making decision on obtained potential fire pixels by obtained contextual information. Chen et al. [28] proposed a Convolutional Neural Network (CNN) based forest fire detection method using UAV dataset. In their framework after dataset capturing, some pre-processing such as histogram equalization, low-pass filtering was conducted and then, the CNN network with seven convolution layers and some hidden layers were utilized for binary classification. Gargiulo et al. [29] designed a fusion framework for active fire detection using Sentinel-2 imagery and CNN to improve the spatial resolution Short Wave Infrared (SWIR) bands toward 10 m resolution. They enhanced the spatial resolution of SWIR bands and improved the performance of active fire detection. Lin et al. [30] proposed an active fire detection framework using FengYun-2G dataset. The active fire detection procedure in this framework was based on comparing predicted value and observed value. That algorithm deployed image analysis for reducing false alarm pixels and temporal analysis for confirming fire pixels. Zhang et al. [31] investigated the active fire detection based on VIIRS dataset. This method was applied in 5 steps: (1) initial data screening by thresholding spectral and thermal bands, (2) data partitioning, (3) the detection of thermal anomaly pixels, (4) contextual analysis, (5) confirmed thermal anomaly pixels based on thresholding of the view zenith and solar zenith angles.

Similarly, there are many active fire detection frameworks exploiting RS datasets. In general, active fire detection methods based on traditional fire detection methods can be applied based three basic principles [32]: (1) detection of hot-temperature pixels using thresholding methods, (2) context-based methods where hot-temperature pixels are compared with the background, and (3) detection of fire pixels based on generated smoke and moving fire plumes. Generally, these active detection methods using RS imagery face their own respective challenges. Thresholding methods tend to fail due to tremendously varying environmental conditions and air temperature. Air temperature variations originating from various factors such as shadow, clouds, wind, illumination angle, seasonal variation, and climate change, can also be a problem for contextual methods. Then, small fires make smoke and non-fire objects detection problematic, especially when dealing with medium and low-resolution RS dataset (i.e. VIRIIS, and MODIS, Sentinel-3). It is also worth mentioning that most of the aforementioned methods rely on traditional machine learning classification, which requires feature engineering.

As one of the widely used RS data, Landsat-8 satellite data was selected for this research. The dataset has higher spatial resolution compared to other modalities such as MODIS, VIRIIS, and Sentinel-3. Landsat-8 also allows improved affordability in detecting small fires as the high-resolution imagery allows for a potentially more improved discrimination capabilities. To minimize the challenges mentioned above (pertaining to datasets), it is necessary to design an advance method for active fire detection. Therefore, this research presents a novel active fire detection method based on deep learning. This method combines the RGB and Thermal images from Landsat-8 satellite imagery. The proposed deep learning framework has two-stream channels and can detect deep features by exploiting the prowess of its convolutional neural network architecture. The key contributions of this research are presented as follows: (a)Presenting a framework for active fire detection based on deep learning called active-fire-net (Fire-Net).(b)High efficiency against detection of small fires using freely available data(c)Taking advantages of deep features that combined spatial and spectral features instead of single spectral features(d)Accurate and automatic detection of active fires with the potential of being applied as real time(e)Taking advantages of residual and depth-wise and separable convolution block(f)Introducing novel loss function for imbalancing problems(g)Testing the transferability of training datasets

2. Study area

Table 1 describes the five study areas selected for active fire detection. Generally, the areas cover North America, Australia, the Amazon rainforest, Central Africa and Chernobyl (Ukraine). Different areas were selected due to the varying climate conditions, temperature affecting the thermal bands and also forest density. This allows for a variety of data to be analyzed with the hope of producing a more generalized model. The Fire-Net model utilizes training data from the active fires in North America and Australia. Model testing is done on the Central Africa, Brazil, Chernobyl and also data from Australia (the parts were not involved in training). [33]. The North American’s forest includes 4 regions (Figure 1(a)) where the active fires were reported in 2013-2014. The Australian forests, which had average temperature of 19°C and humidity of 62%, were obtained in June 2019 where the southern and eastern parts of Australia exhibited intense wildfire. The area includes 6 sub-regions with active fires in 2019-2020 (Figure 1(b)). The third study area is part of the Amazon rainforest, whose average humidity and temperature are 76% and 18.5°C, respectively. The fourth study area is that of north-central Africa wildfires (occurred in December 2018), which was caused by human activity. The average of humidity and temperature of this study area are 72% and 25°C, respectively. The fifth study area belongs to Chernobyl, Ukraine. The wildfires began in April 2020, which was spread quickly by wind. At the time of wildfire, the scene was mostly cloudy. Due to variations temperature and presence of cloud, this case study was chosen to evaluate the performance of proposed methodology against mentioned factors. The average of humidity and temperature are 46% and 9°C, respectively.

2.1. Landsat-8 Dataset

The Landsat-8 satellite sensor was designed by NASA (National Aeronautics and Space Administration) and was launched on into orbit in February 2013. This sensor is able to capture data in 11 spectral bands with spatial resolutions of 15 meters (panchromatic), 30 meters (visible, NIR, and SWIR), and 100 meters (thermal). Furthermore, this sensor collects dataset form Earth with a temporal resolution of 16-days. Since this study requires 4 spectral bands (R, G, B, and thermal), we chose the standard terrain corrected Level 1 T datasets.

2.2. Inventory data

The inventory (sample) data is meant as training images (maps and images) for the proposed supervised Fire-net network. The quality and quantity of sample data are the most important factors to obtain promising result. Therefore, we carefully identified the location of active fires from related resources such as [34, 35]. This was mainly done through visual inspection guided by expert knowledge to generate the reference data. The corresponding active fire locations were then extracted from Landsat-8 datasets for training (65% of the samples), validating (15% of the samples), and testing (20% of the samples). In this research 722 patches were generated with the size of 256 × 256 pixels representing the training, validation, and testing datasets by 469, 109, and 144 patches, respectively.

3. Methodology

Out of five study areas, two regions were selected as the training and validation samples for the Fire-Net deep learning algorithm and multi-scale residual learning networks (MSR-U-Net). This two-streamed pixel classification deep learning-based (Fire-Net) method was proposed to detect active fires from Landsat-8 imagery. This network uses the encoder-decoder architecture for active fire detection. Due to the unique structure of active fires and spatial resolution of the Landsat-8 dataset, a two-stream deep feature extractor architecture is proposed. This means that the Fire-Net framework consists of two deep feature extractor channels. The first channel is meant to detect the active fires whereas the second channel for background and non-fire objects detection. The general overview of Fire-Net is presented in Figure 2.

For dataset preparation, the Landsat-8 images are preprocessed and converted into image patches. Image patches can suppress speckle noise and optimize segmentation. Basically, we divide each image into a grid of patches, specifically, in size. Since we are considering 4-bands, the total number of images are 4. As with any other machine learning algorithm, we divide the dataset into three parts: (1) training dataset, (2) validation dataset, and (3) testing dataset. The training and validation datasets (which form what is called the development set) were used to discover the optimal weights for the entire network. For this work, we defined the stopping condition as 250-epochs. The testing dataset was used to evaluate the trained-and-validated network where seven metrics were calculated. Finally, the trained network is deployed to detect the active fires through the second (Australian Wildfire, the parts were not involved in training), third (Amazon rainforest), forth (Central Africa), and fifth study areas (Chernobyl). The next subsection presents the details of Fire-Net deep learning architectures.

3.1. Image Pre-processing

For the landsat-8 L1-T product some pre-processing as radiometric correction and orthorectification were conducted. The atmospheric correction was also done by Fast Line-of-sight Atmospheric Analysis of Hypercubes (FLAASH) modules. Besides, the thermal bands need to convert into radiance and then to brightness temperature. The final step was splitting data to small patches with size of 256 × 256 pixels to be fed into Fire-Net.

3.2. Proposed Deep Learning Architecture (Fire-Net)

A two-stream feature deep learning method was proposed for forest fire detection. This architecture is based on multiscale-residual convolution layers where it takes the advantage of the point/depth-wise convolution, residual, and multiscale convolution blocks. The multiscale block increases the robustness of Fire-Net against size variations [36]. This allows Fire-Net to improve detection performance of small fires. The residual blocks also has been shown to prevent the vanishing gradient problem and also affords efficiency equal to that of a deeper network [37]. The depth/point-wise convolutions furthermore are cheaper operators that effectively reduce the number of model parameters and computational cost [38]. The Fire-Net architecture is presented by Figure 3 in two-streams for deep feature extraction. Firstly, the shallow features are extracted by the multiscale 2D convolutional layers. Next, the extracted deep features are fed to each deep feature extractor channels. Since small fires can cover small areas, we used up-sampling with a rate of two for the first channel before extracting the deep features. This stream combined residual block and multiscale residual block for extracting deep features. Simultaneously, the second channel was discovered by the deep features from the original resolution dataset by combing the residual block and multiscale residual block that are presented by Figure 4. Then, the extracted deep features from two layers were fused by the summation operator. Finally, the extracted features were fed to 2D-convlotution layer as a single feature map to decide on the final class label (active fire or not). The details of Fire-Net architecture are explained in next sub-sections.

3.2.1. Convolution Layers

In CNN, the main task of the convolution layers is to extract high level deep features from the input dataset [39, 40]. For a convolutional layer in the l-th layer, the computation is expressed according to equation (1) [41]. where x is the input data from layer ; is the activation function; is the weighted template; and is the bias vector.

In 2D convolution layer, the output of the feature map () in layer at the spatial location can be computed using equation (2) [41].

where is the feature cube connected to the current feature cube in the layer; is the value of the kernel connected to the feature cube in the preceding layer; and , and are the length and width of the convolution kernel size, respectively.

The Fire-Net architecture employs three strategies: (1) consistent with [36], we make use of multiscale kernel convolution (i.e. different kernel size convolutions) to ensure robustness against variations in scale [42], (2) we prevent the problem of vanishing or exploding gradient by utilizing residual blocks (or skip connection) layers, which allows the gradient to be directly back-propagated to earlier layers [43], and (3) Depth/Point-wise convolution block: this kind of convolution layer considers only a single filter for each input feature [44, 45]. Figure 5 represents the main differences of standard and depth/point-wise convolution layers.

The loss function used for measuring the training error is based on predicted value of true value [46]. This research is used a hybrid loss function (Equation (3)) is known weight-binary-cross-entropy-dice (WBCED) loss function that it is more efficient for small targets [47]. The WBCED loss function is combined dice loss (Equation (4)) and weighted cross entropy loss (Equation (5)) that can be defined between predicted value () of true value () as following: where is obtained by overage pooling of reference map furthermore, , are width and length of Mask.

The Fire-Net architecture is trained by an Adaptive Moment Estimation (Adam) optimizer [48] through backpropagation where weights were initialized using the Golorot initializer [49]. The hyperparameters for training Fire-Net is as follows: (i)Batch-size =7-patches(ii)Learning rate = (iii)Number of Epochs =250

The shuffle technique was used during training process. Fire-Net was implemented on Tensorflow v2.4.1 and Keras v2.4.3.

3.2.2. Evaluation Indices/methods or accuracy assessment

We performed two types of evaluation for Fire-Net. Firstly, visual analysis was performed where the results were compared with reference data. Next, we calculated seven quality indices namely overall accuracy (OA), recall, precision false positive rate (FPR), miss-detection (MD), F1-score, and kappa coefficient (KC). More details of these indices can be found in [41]. We also performed a comparison between Fire-Net with an established CNN, namely the MSR-U-Net. The main reason is due to MSR-U-Net’s similarity to Fire-Net in combining multi-scale kernel convolution filters with residual blocks deep features extraction [50, 51]. MSR-U-Net also uses an encoder-decoder structure for exploring deep features. For a detailed evaluation of Fire-Net’s efficiency, 5 machine learning methods (K-Nears-Neighbor (KNN), Support Vector Machine (SVM), Multi-Layer Perceptron (MLP), Random Forest (RF), and Extreme Gradient Boosting (XGBoost)) were used for comparing fire detection results. Brief explanations of each algorithm are provided in the following: (1)KNN: This is a simple non-parametric classifier. Without any assumption about the dataset, KNN classifies the data based on each data point’s neighbors. This research used the Euclidian metric and the numbers of neighbors were set to 4(2)SVM: The SVM is a supervised machine learning classifier that is widely used in RS. The main idea behind SVM is to find a hyperplane that best separates a dataset into two-classes. Although in its original form, the SVM is meant for binary classification, it can still be used for multi-class classification tasks. In general, a kernel function is used to generate a nonlinear decision boundary in the feature space. In this work, we empirically tuned two of the SVM parameters, which are the penalty coefficient and the (Radial Basis) kernel parameter. They were set to 101 and 103, respectively(3)MLP: This classifier is a supplement of a feed forward neural network. The MLP has several hidden layers where each has several neurons. The neurons are trained using the back propagation learning algorithm. We constructed a two hidden-layer MLP with 40 and 30 neurons in the first and second layers, respectively(4)RF: RF is a supervised learning classifier consisting of an ensemble of decision trees where the final decision is obtained by majority voting. RF is trained by a bagging method where the gist of bagging is the combination of learning models that improves overall classification results. This algorithm has two main parameters i.e. number of trees and number of features to split. Each were set to 93 and 3, respectively(5)XGBoost: XGBoost is a decision-tree-based ensemble supervised learning method, based on the gradient boosting framework. Several parameters need to be tuned for this classifier, namely number of estimators (set to 105), max-depth (set to 4), eta (set to 0.03), min-child-weight (set to 1), subsample (set to 0.8) and colsample-bytree (set to 0.8).

4. Results

This section explains the experimental setup and results of the proposed active fire detection algorithm (Fire-Net). The dataset we used contains 22 tiles of Landsat-8 images covering the first and second study areas (i.e. North USA and Australia). From the 22 tiles, 14 tiles were used for training, 3 tiles for validation, and 5 tiles for testing. For the ease of processing, the 22 tiles were split into small patches in size of 256 × 256. At the end, to check the transferability of the Fire-Net method it was tested on other study areas (i.e. the Amazon rainforest, Central Africa and Chernobyl).

4.1. Australian Forest

The results of the Australian’s forest fire detection, by methods, are presented by Figures 6(b) and 6(c). Visually, both deep learning methods managed to detect fire areas. However, misclassification occurred in the patch boundaries for MSR-U-Net (Figure 6(b)). Based on the evaluation metrics, the main difference between the two networks were in detection of active fires with low MD and FPR rates. Many false positive pixels were detected using MSR-U-Net, while Fire-Net had lower false positives.

The numerical results of active fire detection for the first test are presented in Table 2(a). Fire-Net outperformed MSR-U-Net with a higher overall accuracy of 99.95%, precision of 97.94%, F1-Score of 97.57%, and KC of 0.975 indices. MSR-U-Net exhibited a lower rate of MD (near 0%), while considerably showing a higher FPR compared to Fire-Net. There is however a trade-off between detecting fire pixels and non-fire pixels. The Fire-Net model has lower error rates for active fire detection whereas MSR-U-Net has lower MD rates but higher FPR rates. The ideal situation is when a method can effectively detect both fire and non-fire pixels with lowest error. The numerical results for active fire detection by the other machine learning algorithms are presented in Table 3(a). Although the SVM had the most accurate OA (99.03%), its performances against the other indices of Precision, Recall and F1-Score were low. The MD index (99.98%) for SVM revealed that it could not effectively detect active fires. RF and XGBOOST on the other hand showed good performance compared to the other algorithms. In all however, Fire-Net outperformed all these algorithms where they lacked the robustness, correctness, and certainty against the 7 evaluation metrics.

4.2. Central Africa Forest Fire

The result of active fire detection for Central Africa is presented in Figures 6(f) and 6(g). It shows background from foreground segmentation for each scene. For this case study, the active fire areas were small compared to the north of USA and Australia. Both methods were able to detect active fires with cases of misclassifications.

Fire pixels can be seen when zooming into Figures 7(b) and 7(c). Compared to MSR-U-Net, Fire-Net provides more fitting segmentations, covering more relevant areas of fire areas. This indicates that Fire-Net detects almost all relevant fire pixels that were not detected by MSR-U-Net.

Table 2(b) shows the accuracy assessments for both networks. The results indicate that both methods performed well in detecting non-fire pixels, while their accuracies decreased when detecting fire pixels. For this dataset, the MD rate is high and both networks focused on non-fire pixels. The low values of KC indicate the decrease of performance reliability of both networks. Table 3(b) shows the quantitative evaluations for 5 machine learning methods to detect active fires. Accordingly, most of them were successful in detection of non-active fires as indicated by FPR values lower than 1%. Moreover, the lower Precision and F1-Scores values show that consistency and balance among all 7 evaluation metrics were not seen amongst those 5 methods compared with Fire-Net.

4.3. Brazilian’s Forest Fire

The active fire maps for Brazil’s forest produced by the different methods are illustrated in Figures 6(j) and 6(k). The southern part of the region was mainly classified as having active fires. Seemingly, both methods generated active fires map with lowest error.

Figures 7(g) to 7(i) presents a closer look of the active fire areas. Figure 7(i) is the active fire map generated by Fire-Net. It is comparable and as complete as the one produced by MSR-U-NET. The Accuracy assessment for the generated fires maps are presented in Table 2(c). It can be seen that both methods successfully detected the fires, with Fire-Net outperforming MSR-U-Net. Fire-Net’s precision is 95.98%, recall 98.04%, near zero-FPR, and an F1-Score of 97.0%. The numerical assessments for the 5 machine learning algorithms are shown in Table 3(c). OA was mostly equal for all at approximately 99%. Precision on the other hand were lower than 60% with low FPR rates (below 1%) and higher MD rates. These indicate weak active fire detection. Therefore, the 5 machine learning methods were not comparable with Fire-Net in this region.

4.4. Chernobyl’s Forest Fire

Figures 6(n) and 6(o) shows both network’s fire detection performance in the Chernobyl forest. Note that fire detection in this region is challenging due to the presence of high clouds and fires of small sizes. Nonetheless, both networks performed well with MSR-U-Net failing to detect a small number of active fires (Figures 7(j)–7(m)). As with the other regions, Fire-Net showed higher precision, recall, and f1-Score. However, we posit that the diversity of the terrain might be the cause for higher MD rates compared to FPR. The results of active fire detection using the 5 machine learning methods are presented in Table 3(d). OAs exceeding 99% for KNN, RF, and XBGOOST, with an added low FPR (below 1%) indicates accurate active fire detection. However, SVM and MLP were successful in the mapping of non-active fires. Again, based on the 7 evaluation metrics, the Fire-Net method showed its superiority.

5. Discussion

5.1. Accuracy

Timely detection of active fires is critical for disaster management. This research proposes a deep network called Fire-Net for active forest fire detection based on optical RGB and thermal dataset. The accuracy of the network was evaluated based on 4 different case study areas. Furthermore, Fire-Net’s results were also compared against the established MSR-U-Net deep neural network. The evaluation metrics and visual inspection of active fire detection results showed that the proposed method outperformed the MSR-U-Net and exhibited higher accuracy in detection of fire and non-fire pixels. For a more detailed analysis, Fire-Net was compared against 5 other popular machine learning algorithms. It was discovered that MLP, RF, SVM, KNN, and XGBOOST required a relatively small number of samples, with the exception that several parameters must be tuned for optimal classification. Despite good performance, all these methods require manual handcrafted features which is laborious and time-consuming.

As for the evaluation metrics, Recall is an evaluation metric that finds true positive (TP) pixels. The mean of recall of the Fire-Net algorithm was 92.64% while MSR-U-Net was 86.25%. This indicates Fire-Net’s superiority in detecting more TP pixels. On the other hand, precision is a metric that describes a model’s ability to identify positive pixels. In our study, Fire-Net had a precision of 93%, indicating high-confidence in detection of active fire pixels. MSR-U-Net on the other hand only had a precision of under 67%. Although Fire-Net clearly has higher precision and recall scores, the F1-Score is important to measure as well since it provides a harmonic mean of precision and recall. With regards to this metric, Fire-Net had a score of more than 93% compared to MSR-U-Net at 67%. The high rate of F1-Scores expresses the model has either the high believable and high ability in prediction of TP pixels. The OA index is one of the most common factors in quantity assessment. This index associates the relation among TN and TP pixels with all pixels. The OA proposed Fire-Net algorithm is more than 99% in all datasets. This subject shows the proposed method has high performance in prediction TP and true negative (TN) pixels. The KC measures the reliability of model in classification that mainly, the proposed method was provided the moderated level.

MD and FPR are two metrics that should be low for effective active fire detection. This is especially true for MD because undetected fire pixels are more costly than labeling false detections. Our results show that Fire-Net’s average MD is at 8% with an FPR of 0.0001%. MSR-U-Net on the other hand had an average of MD rate more than 14% with a mean FPR of 1%. Essentially, precision, recall, and the F1-Scores significantly improved by the Fire-Net algorithm within all study areas. From the valuations, the 5 machine learning algorithms showed high performance, mainly in mapping non-fire areas. They however failed to map active fires. In comparison to the deep networks (especially to Fire-Net), all 5 machine learning methods had lower success rates. Fire-Net showed the highest precision for the active fire detection task. Recently, more research has been done for active fire detection based on RS imagery. Table 4 presents the obtained results by other active fire detection methods.

There are some products for active fire detection as globally that they mainly are generated by MODIS, VIIRIS (Visible Infrared Imaging Radiometer Suite), and NOAA (National Oceanic and Atmospheric Administration) sensors. Oliva and Schroeder [52] investigated the active fire product based on VIIRS dataset by comparing the Landsat-8 dataset. The result of mentioned study shows the accuracy of this product depend on size fires as the accuracy of product covers ranges 7% through 100%. Furthermore, Giglio et al. [53] had investigated the performance of collection six MODIS active fire products. The numerical results had shown the FPR rate of this product is nearby 2.4%.

One of most important issue of deep learning method is balancing dataset. The fire areas cover very small areas while the non-fire areas are dominant in the scene. Therefore, for solving this issue we introduced novel loss function. This loss function combines dice-loss and weighted-binary cross-entropy.

5.2. Feature Extraction

More active fire detection methods focus on spectral features and contextual information [30]. The ability of deep learning methods for feature extraction have been proven by many researches [12, 40, 42, 43, 54]. The deep learning methods are able to detect deep features containing various spectral/spatial features. The quality and quantity of active fire methods based on manual hand crafted features is the main challenge in traditional active fire detection methods. The deep learning methods can extract deep informative features in an automatic manner. It is worth noting that the proposed method used only four coarse bands (Red, Green, Blue, and thermal). These bands are the most common bands and widely available data in most RS satellite imagery. Therefore, this methodology can be deployed by other satellites with higher temporal resolution in active fire detection and forest management for real time and continuous fire monitoring.

5.3. Transferability

Most of the proposed active fire detection methods are based on thresholding [21, 24, 55]. Due to different environmental conditions and diversity in forest layouts, it is impractical to fix constant threshold values for decision making. From deep learning point of view, both the methods are based on deep learning, which does not require hardcoding threshold values. Instead, during the training step, each network discovers the best weights connecting each node through an optimization process. The transferability is clearly shown with high accuracy for all regions.

6. Conclusion

Early detection of active forest fires is critical to determine the starting point of the fire for effective emergency responses. In this study, active fire detection was performed on a medium spatial resolution dataset (Landsat-8 imagery) where the extent of active fires was very low. A deep CNN (Fire-Net) was proposed in this work to detect active forest fires in various regions. Specifically, the USA and Australia regions were used to train the network whereas testing was done for Africa, Brazil, Ukraine, Australia (the parts was not involved in training) regions. The results depicted a high transferability of the proposed method. Then, the Fire-Net was compared with another state-of-the-art deep network, i.e., MSR-U-Net and other common machine learning algorithms. The results for active fire detection were qualitatively and quantitatively assessed. The performance evaluations showed there was a trade-off between active fire and non-active fires detection. Due to the extent of active fires in small areas, most of the machine learning algorithms could not detect active fires. The high accuracies measured by OA index for these algorithms, were mostly for non-fire zones. The other indices namely precision, KC, and F1-score for these models were low for the detection of active fires. In contrast, the proposed Fire-Net method showed high efficacy for both active and non-active fire detection. Here, small active fires were detected with high accuracy and low miss detection rates. The efficiency of Fire-Net originated in its architecture and convolution layers structure enabling high level and informative features extraction. Experimental results indicate that Fire-Net: (1) has higher accuracy, (2) obtains higher sensitivity to small active fires, (3) the proposed method can be applied as real time processing due to high transferability. Future work will be based on using geostationary satellite imagery for rapid monitoring of active fires to provides high temporal resolution for fast monitoring in a larger scale.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The author(s) declare(s) that they have no conflicts of interest.