Abstract
Wildfires are a natural disaster that results in significant harm and catastrophic destruction. Forest areas tend to be more prone to the devastating effects of wildfires. Global warming causes wildfires to occur more frequently and with severe effects, forcing them to spread across wide amount of land areas, causing unimaginable harm and even claiming lives. In this paper, we propose a novel methodology to analyze the effects of wildfire and estimating its probability to spread using satellite data. The severity of wildfire is determined through fire and smoke detection via deep learning approach Modified-Residual Unet. To categorize areas based on their susceptibility to wildfires, NDVI imagery is given to the ZFNet classifier which determines the region's risk of being prone to wildfire. It achieves an impressive accuracy of 98.3% proving its ability in classifying wildfire risk. A novel Deep Probabilistic (P) Learning along with Cellular Automaton and Diffusion Limited Aggregation Algorithm is used to simulate the spread of wildfires and estimates are made by Anisotropic Generalized Regression Neural Network for the impacted areas. Thus, the efficiency of this novel approach has been tested with various datasets and our approach proves to have notable merits with greater accuracy and substantially lesser time when compared to other methods.
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig1_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig2_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig3_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig4_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig5_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig6_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig7_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig8_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig9_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig10_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig11_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig12_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig13_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig14_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig15_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig16_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig17_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig18_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig19_HTML.png)
![](http://media.springernature.com/m312/springer-static/image/art%3A10.1007%2Fs12524-024-01888-0/MediaObjects/12524_2024_1888_Fig20_HTML.png)
Similar content being viewed by others
Data Availability
The datasets used and/or analyzed during the current study are available from the corresponding author on request.
References
Ajith, M., & Martínez-Ramón, M. (2019). Unsupervised segmentation of fire and smoke from infra-red videos. IEEE Access, 16(7), 182381–182394.
Akbari Asanjan, A., Memarzadeh, M., Lott, P. A., Rieffel, E., & Grabbe, S. (2023). Probabilistic wildfire segmentation using supervised deep generative model from satellite imagery. Remote Sensing, 15(11), 2718. https://doi.org/10.3390/rs15112718
Bo, W., Liu, J., Fan, X., Tjahjadi, T., Ye, Q., & Fu, L. (2022). BASNet: Burned area segmentation network for real-time detection of damage maps in remote sensing images. IEEE Transactions on Geoscience and Remote Sensing., 8(60), 1–3.
Celik, T., & Demirel, H. (2009). Fire detection in video sequences using a generic color model. Fire Safety Journal, 44(2), 147–158.
Chaoxia, C., Shang, W., & Zhang, F. (2020). Information-guided flame detection based on faster R-CNN. IEEE Access, 24(8), 58923–58932.
Chaturvedi, S., Khanna, P. & Ojha, A. (2021). Comparative Analysis of Traditional and Deep Learning Techniques for Industrial and Wildfire Smoke Segmentation. In Proceedings of the Sixth International Conference on Image Information Processing (ICIIP) (pp. 326–331).
Chen, B., Bai, D., Lin, H., & Jiao, W. (2023). FlameTransNet: Advancing forest flame segmentation with fusion and augmentation techniques. Forests, 14, 1887. https://doi.org/10.3390/f14091887
Collumeau, J. F., Laurent H., Hafiane A. & Chetehouna, K. (2011). Fire scene segmentations for forest fire characterization: A comparative study. In: Proceedings of the 18th IEEE International Conference on Image Processing (pp. 2973–2976)
Dunnings, A.J., & Breckon, T.P. (2018). Experimentally defined convolutional neural network architecture variants for non- temporal real-time fire detection. In: Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP) (pp. 1558–1562).
Frizzi, S., Bouchouicha, M. & Moreau, E. (2021). Comparison of two semantic segmentation databases for smoke detection. In: Proceedings of the 22nd IEEE International Conference on Industrial Technology (ICIT) (pp. 856–863).
Ghali, R., Akhloufi, M. A., Jmal, M., Mseddi, W.S., & Attia, R. (2021). Forest Fires Segmentation using Deep Convolutional Neural Networks. In: Proceedings of the IEEE International Conference on Systems, Man and Cybernetics (SMC) (pp. 2109–2114).
Hossain, F. M. A., & Zhang, Y. (2023). MsFireD-Net: A lightweight and efficient convolutional neural network for flame and smoke segmentation. Journal of Automation and Intelligence, 2(3), 130–138. https://doi.org/10.1016/j.jai.2023.08.003.ISSN2949-8554
Hu, X., Jiang, F., Qin, X., Huang, S., Yang, X., & Meng, F. (2024). An optimized smoke segmentation method for forest and grassland fire based on the UNet framework. Fire, 7, 68. https://doi.org/10.3390/fire7030068
Huot, F., Hu, R. L., Goyal, N., Sankar, T., Ihme, M., & Chen, Y. F. (2022). Next day wildfire spread: a machine learning dataset to predict wildfire spreading from remote-sensing data. IEEE Transactions on Geoscience and Remote Sensing, 26(60), 1–3.
Khryashchev, V. & Larionov, R. (2020). Wildfire Segmentation on Satellite Images using Deep Learning. Moscow Workshop on Electronic and Networking Technologies. (MWENT)
Kolarić, D., Skala, K., & Dubravić, A. (2008). Integrated system for forest fire early detection and management. Periodicum Biologorum, 110(2), 205–211.
Krizhevsky, A., Sutskever, I. & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems (pp. 1097- 1105).
Lasaponara, R., & Tucci, B. (2019). Identification of burned areas and severity using SAR sentinel-1. IEEE Geoscience and Remote Sensing Letters, 16(6), 917–921.
Lee, W. Kim, S. Lee, Y. T., Hyun-Woo Lee and Choi, M., (2017). Deep neural networks for wild fire detection with unmanned aerial vehicle. In: IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, (pp. 252–253).
Naderpour, M., Rizeei, H. M., & Ramezani, F. (2021). Forest fire risk prediction: A spatial deep neural network-based framework. Remote Sensing, 13, 2513. https://doi.org/10.3390/rs13132513
Nguyen, T. T., Phan,T. C., Nguyen, Q. D. and Nguyen, Q. V. H. (2021). Realtime Bushfire Detection with Spatial-based Complex Event Processing. In: 15th International Conference on Advanced Computing and Applications (ACOMP), Ho Chi Minh City, Vietnam, (pp. 1–8).
Pinto, M. M., Libonati, R., Trigo, R. M., Trigo, I. F., & DaCamara, C. C. (2020). A deep learning approach for mapping and dating burned areas using temporal sequences of satellite images. ISPRS Journal of Photogrammetry and Remote Sensing, 160, 260–274.
Shahid, M., Chen, S.-F., Hsu, Y.-L., Chen, Y.-Y., Chen, Y.-L., & Hua, K.-L. (2023). Forest fire segmentation via temporal transformer from aerial images. Forests, 14, 563. https://doi.org/10.3390/f14030563
Wang, Z., Tao, P., & Zhaoyou, Lu. (2022). Comparative Research on forest fire image segmentation algorithms based on fully convolutional neural networks. Forests, 13(7), 1133.
Wang, G., Wang, F., Zhou, H., & Lin, H. (2024a). Fire in focus: Advancing wildfire image segmentation by focusing on fire edges. Forests, 15(1), 217. https://doi.org/10.3390/f15010217
Wang, G., Bai, D., Lin, H., Zhou, H., & Qian, J. (2024b). FireViTNet: A hybrid model integrating ViT and CNNs for forest fire segmentation. Computers and Electronics in Agriculture, 218, 108722. https://doi.org/10.1016/j.compag.2024.108722.ISSN0168-1699
Wuebbles, D. J., Fahey, D. W., & Hibbard, K. A. (2017) Climate science special report: fourth national climate assessment, volume i. US Global Change Research Program.
Yuan, C., Zhang, Y., & Liu, Z. (2015). A survey on technologies for automatic forest fire monitoring, detection, and fighting using unmanned aerial vehicles and remote sensing techniques. Canadian Journal of Forest Research, 45(7), 783–792.
Zhang, J., Zhu, H., Wang, P., & Ling, X. (2021). ATT squeeze U-Net: A lightweight network for forest fire detection and recognition. IEEE Access, 11(9), 10858–10870.
Zhang, L., Wang, M., Ding, Y., Wan, T., Qi, B., & Pang, Y. (2023). FBC-ANet: a semantic segmentation model for UAV forest fire images combining boundary enhancement and context awareness. Drones, 7, 456. https://doi.org/10.3390/drones7070456
Acknowledgements
We would like to thank the NASA FIRMS, the EarthExplorer of USGS and the One Soil Application for providing accessible data products. The current research was made possible by equally scientific involvement of all the concerned authors.
Funding
The author(s) received no financial support for the research, author ship, and/or publication of this article.
Author information
Authors and Affiliations
Contributions
Conceptualization, R.S.P. (R. Shanmuga Priya) and K.V. (K. Vani); methodology, R.S.P.; validation, R.S.P. and K.V.; formal analysis, R.S.P. and K.V.; resources and curation, R.S.P; writing—original draft preparation, R.S.P.; writing—review and editing, R.S.P. and K.V.; visualization, R.S.P.; supervision, K.V.; project administration, R.S.P. and K.V. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Ethical Approval
Not applicable.
Consent for publications
Not applicable.
Consent to participate
Not applicable.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Priya, R.S., Vani, K. Wildfire Impact Analysis and Spread Dynamics Estimation on Satellite Images Using Deep Learning. J Indian Soc Remote Sens 52, 1385–1403 (2024). https://doi.org/10.1007/s12524-024-01888-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12524-024-01888-0