Skip to main content
Log in

Recurrent Convolutional Deep Neural Networks for Modeling Time-Resolved Wildfire Spread Behavior

  • Published:
Fire Technology Aims and scope Submit manuscript

Abstract

The increasing incidence and severity of wildfires underscores the necessity of accurately predicting their behavior. While high-fidelity models derived from first principles offer physical accuracy, they are too computationally expensive for use in real-time fire response. Low-fidelity models sacrifice some physical accuracy and generalizability via the integration of empirical measurements, but enable real-time simulations for operational use in fire response. Machine learning techniques have demonstrated the ability to bridge these objectives by learning first-principles physics while achieving computational speedups. While deep learning approaches have demonstrated the ability to predict wildfire propagation over large time periods, time-resolved fire-spread predictions are needed for active fire management. In this work, we evaluate the ability of deep learning approaches in accurately modeling the time-resolved dynamics of wildfires. We use an autoregressive process in which a convolutional recurrent deep learning model makes predictions that propagate a wildfire over 15 min increments. We apply the model to four simulated datasets of increasing complexity, containing both field fires with homogeneous fuel distribution as well as real-world topologies sampled from the California region of the United States. We show that even after 100 autoregressive predictions representing more than 24 h of simulated fire spread, the resulting models generate stable and realistic propagation dynamics, achieving a Jaccard score between 0.89 and 0.94 when predicting the resulting fire scar. The inference time of the deep learning models are examined and compared, and directions for future work are discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1
Figure 2
Figure 3
Figure 4
Figure 5
Figure 6
Figure 7
Figure 8
Figure 9
Figure 10
Figure 11
Figure 12
Figure 13
Figure 14

Similar content being viewed by others

References

  1. Covington WW, Moore MM (1994) Southwestern ponderosa forest structure: changes since Euro-American settlement. J Forest 92:39–47

    Google Scholar 

  2. Westerling AL, Hidalgo HG, Cayan DR, Swetnam TW (2006) Warming and earlier spring increase Western U.S. forest wildfire activity. Science 313:940–943

    Article  Google Scholar 

  3. Mueller SE, Thode AE, Margolis EQ, Yocom LL, Young JD, Iniguez JM (2020) Climate relationships with increasing wildfire in the southwestern US from 1984 to 2015. For Ecol Manage 460:117861

    Article  Google Scholar 

  4. Thomas D, Butry D, Gilbert S, Webb D, Fung J (2017) The costs and losses of wildfires: a literature survey. NIST Special Publication 1215. National Institute of Standards and Technology, Gaithersburg

  5. Kollanus V, Prank M, Gens A, Soares J, Vira J, Kukkonen J, Sofiev M, Salonen RO, Lanki T (2017) Mortality due to vegetation fire-originated PM2.5 exposure in Europe-Assessment for the years 2005 and 2008. Environ Health Perspect 125:30–37

    Article  Google Scholar 

  6. van der Werf GR, Randerson JT, Giglio L, van Leeuwen TT, Chen Y, Rogers BM, Mu M, van Marle MJE, Morton DC, Collatz GJ, Yokelson RJ, Kasibhatla PS (2017) Global fire emissions estimates during 1997–2016. Earth Syst Sci Data 9:697–720

    Article  Google Scholar 

  7. Bakhshaii A, Johnson EA (2019) A review of a new generation of wildfire-atmosphere modeling. Can J For Res 49:565–574

    Article  Google Scholar 

  8. Sullivan AL (2009) Wildland surface fire spread modelling, 1990–2007. 1. Physical and quasi-physical models. Int J Wildl Fire 18:349–368

    Article  Google Scholar 

  9. Sullivan AL (2009) Wildland surface fire spread modelling, 1990–2007. 2. Empirical and quasi-empirical models. Int J Wildl Fire 18:369–386

    Article  Google Scholar 

  10. Sullivan AL (2009) Wildland surface fire spread modelling, 199–2007. 3. Simulation and mathematical analogue models. Int J Wildl Fire 18:387–403

    Article  Google Scholar 

  11. Linn RR, Harlow FH (1997) FIRETEC: a transport description of wildfire behavior, LA-UR-97-3920. Los Alamos National Laboratory, Los Alamos

  12. Larini M, Giroud F, Porterie B, Loraud JC (1998) A multiphase formulation for fire propagation in heterogeneous combustible media. Int J Heat Mass Transf 41:881–897

    Article  MATH  Google Scholar 

  13. Mell W, Jenkins MA, Gould J, Cheney P (2007) A physics-based approach to modelling grassland fires. Int J Wildl Fire 16:1–22

    Article  Google Scholar 

  14. Rothermel RC (1972) A mathematical model for predicting fire spread in wildland fuels. Research Paper INT-115. USDA Forest Service, Intermountain Forest and Range Experiment Station, Ogden

  15. P. L. Andrews PL (1986) BEHAVE: fire behavior prediction and fuel modeling system–BURN Subsystem, part 1, General Technical Report INT-194. USDA Forest Service, Intermountain Research Station, Ogden

  16. Finney MA (1998) FARSITE: fire area simulator–model development and evaluation. Research Paper RMRS-RP-4. USDA Forest Service, Rocky Mountain Research Station. Revised 2004

  17. Jain P, Coogan SCP, Subramanian SG, Crowley M, Taylor S, Flannigan MD (2020) A review of machine learning applications in wildfire science and management. Environ Rev 28:478–505

    Article  Google Scholar 

  18. Ihme M, Chung WT, Mishra AA (2022) Combustion machine learning: principles, progress and prospects. Prog Energy Combust Sci 91:101010

    Article  Google Scholar 

  19. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444

    Article  Google Scholar 

  20. Hodges JL, Lattimer BY (2019) Wildland fire spread modeling using convolutional neural networks. Fire Technol 55:2115–2142

    Article  Google Scholar 

  21. Radke D, Hessler A, Ellsworth D (2019) Proceedings of the twenty-eighth international joint conference on artificial intelligence (IJCAI-19). International Joint Conferences on Artificial Intelligence Organization, pp 4575–4581

  22. Burge J, Bonanni M, Ihme M, Hu L (2020) Convolutional LSTM neural networks for modeling wildland fire dynamics. arXiv Preprint. arXiv:2012.06679

  23. Bolt A, Huston C, Kuhnert P, Dabrowski JJ, Hilton J, Sanderson C (2022) A spatio-temporal neural network forecasting approach for emulation of firefront models. arXiv Preprint. arXiv:2206.08523

  24. Kulkarni TD, Whitney WF, Kohli P, Tenenbaum J (2015) In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates, Montreal

  25. Sanchez-Gonzalez A, Godwin J, Pfaff T, Ying R, Leskovec, Battaglia P (2020) In: Daumé III H, Singh A (eds) Proceedings of the 37th international conference on machine learning, vol 119. Proceedings of Machine Learning Research, pp 8459–8468

  26. Kochkov D, Smith JA, Alieva A, Wang Q, Brenner MP, Hoyer S (2021) Machine learning accelerated computational fluid dynamics. Proc Natl Acad Sci USA 118:e2101784118

    Article  MathSciNet  Google Scholar 

  27. Rollins MG, Rollins MG (2009) LANDFIRE: A nationally consistent vegetation, wildland fire, and fuel assessment. Int J Wildl Fire 18:235–249

    Article  Google Scholar 

  28. Scott JH, Burgan RE (2005) Standard fire behavior fuel models: a comprehensive set for use with Rothermel’s surface fire spread model. General Technical Report RMRS-GTR-153. USDA Forest Service Rocky Mountain Research Station, Fort Collins

  29. Forthofer JM, Butler BW, Wagenbrenner NS, Forthofer JM, Butler BW, Wagenbrenner NS (2014) A comparison of three approaches for simulating fine-scale surface winds in support of wildland fire management. Part I. Model formulation and comparison against measurements. Int J Wildl Fire 23:969–981

    Article  Google Scholar 

  30. Foster D (2019) Generative deep learning. O’Reilly, Sebastopol

    Google Scholar 

  31. He K, Zhang X, Ren S, Sun J (2016) IEEE conference on computer vision and pattern recognition (CVPR), Las Vegas, pp 770–778. https://doi.org/10.1109/CVPR.2016.90

  32. Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jozefowicz YR, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X (2015) TensorFlow: large-scale machine learning on heterogeneous systems. TensorFlow: large-scale machine learning on heterogeneous systems 2015. https://iwww.tensorflow.org. Accessed Aug 4, 2023.

  33. Chollet F et al (2015) Keras. https://keras.io. Accessed Aug 4, 2023.

  34. Kingma DP, Ba J (2014) Proceedings of the international conference for learning representations. arXiv Preprint. arXiv:1412.6980

  35. Filippi J-B, Mallet V, Nader B (2014) Representation and evaluation of wildfire propagation simulations. Int J Wildl Fire 23:46

    Article  Google Scholar 

  36. Tibshirani RJ, Efron B (1993) An introduction to the bootstrap. Monographs on statistics and applied probability, vol 57. Chapman and Hall, New York

  37. Shi X, Chen Z, Wang H, Yeung D-Y, Wong W, Woo W (2015) In: Cortes C, Lawrence N, Lee D, Sugiyama M, Garnett R (eds) Advances in neural information processing systems, vol 28. Curran Associates, Montreal

Download references

Acknowledgements

We thank the reviewers for their constructive feedback. MB acknowledges financial support through the NASA Fellowship Activity, #80NSSC21K2054, and the Stanford University School of Engineering Departmental Fellowship. We thank the members of Google Research’s AI For Weather team: Jason Hickey, Cenk Gazen and Shreya Agrawal, for their support in this project. All of the training, validation and test data, and each of the four EPD and EPD-ConvLSTM models, as well code that will facilitate performing inferences, is available through Google’s research GitHub repository at https://github.com/google-research/google-research/tree/master/wildfire_conv_lstm.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to John Burge.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix 1: Deep Neural Network Transformations

Appendix 1: Deep Neural Network Transformations

While a full summary of deep neural networks is outside the scope of this paper, we provide a brief description of the transformations used in the DNNs described in this work. For a deeper background on DNNs, we refer the interested reader to [30].

The fundamental unit processed by a DNN is a tensor. A tensor is an n dimensional collection of scalar values, \(n\ge 0\). For example, a single field specifying how much fuel exists in a 2D patch of ground could be stored in a 2D tensor with shape (HW) where H is the height of the field and W is the width. That field represents a channel of information and if there are additional channels, they can be stored in a tensor of shape (HWC) where C is the number of channels. A time series of fields can be stored in a tensor of shape (THWC) where T is the number time points in the series. Training a DNN is usually done in batches, so a DNN that works on (THWC) data will actually take (BTHWC) data where B is the number of data points in a single batch.

DNNs are often said to predict some outcome based on a given data point though in reality, the output is merely the input after its been passed through a potentially large number of transformations. We sometimes refer to sets of transformations as a layer. Some layers contain parameters such that the process of training a DNN attempts to find the optimal set of values that result in transforming as much of the training input into as correct a set of output as possible. These are the primary transformations used in this work:

  • fully connected: Every value in the output is a parameterized combination of every value in the input. Given an input shape (BM) and output shape (BN), there are \(M \times N\) total parameters. This layer can be useful when many output values need to consider many input values, but can be particularly prone to overfitting and parameter bloat.

  • convolutional (2D): This layer is appropriate for inputs that contain 2D fields. A 2D convolutional kernel is swept across both dimensions of the input field. At each location, the dot product between the kernel and the 2D field is computed, which generates a new field that is the convolution of the input field with the kernel. There can be multiple independent kernels, resulting in multiple field outputs. Given an input shape of (BHWC), the output shape is (BHWK) where K is the number of kernels.

  • recurrent: A layer which explicitly considers temporal relationships in the input data when generating the output. Unlike spatial relationships, temporal relationships often cannot effectively be modeled by simply considering neighboring input. Temporal dynamics often require considering events that occurred at more distant times in the past. Recurrent layers build up a memory by iterating over individual time points one at a time. Cells in the output depend on the memory instead of actual time points in the past. If the input contains fields, the layers will also leverage convolutions such that an output cell can depend on the memory of itself, and the memory of neighboring cells. If the convolutions have K kernels and the input shape is (BTHWC), then the output shape will be (BTHWK). Given its wide-spread success at modeling image-to-image type tasks, we use the Convolutional Long-Short Term Memory layer (ConvLSTM) [37].

  • activation: Passes each value in the input through an often non-linear and potentially parameterized function. The output shape is equal to the input shape.

  • skip link: Taking the input for some transformation and concatenating (or adding) it with the output of the transformation. This effectively gives a path for the gradients of the model’s parameters to skip the transformation during training, significantly increasing the efficacy of training large networks.

  • max pooling (2D): A field transformation that downsamples the spatial dimensions of the input and while typically doubling the number of channels. If the input has shape (BHWC), then the output will have shape \((B, H/2, W/2, C \times 2)\).

  • zero padding: This layer adds padding around the field. If the input shape is (BHWC) and the padding size is P then the output shape will be \((B, H+2P, W+2P, C)\).

  • one hot encoding: This layer transforms integer values in the input into a list of all zeros with the exception of a single element in the list corresponding to the integer value being given a value of one. If the input shape is (BHWC), the output will be (BHWCE) where E is the maximum value the input can take.

  • embedding: This layer converts a field that contains discrete valued cells into an field where each cell is replaced by a set of floating-point values. Values in the input that result in similar predictions are placed close to each other in the embedding space, which subsequent layers in the model can leverage more effectively than the discrete values alone. If the input has shape (BHW, 1) and the embedding space has rank E, then the output shape will be (BHWE).

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Burge, J., Bonanni, M.R., Hu, R.L. et al. Recurrent Convolutional Deep Neural Networks for Modeling Time-Resolved Wildfire Spread Behavior. Fire Technol 59, 3327–3354 (2023). https://doi.org/10.1007/s10694-023-01469-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10694-023-01469-6

Keywords

Navigation