Hostname: page-component-77f85d65b8-2tv5m Total loading time: 0 Render date: 2026-04-18T18:31:49.288Z Has data issue: false hasContentIssue false

Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts

Published online by Cambridge University Press:  12 February 2025

Kyleen Liao*
Affiliation:
Department of Computer Science, Stanford University, Stanford, CA, USA
Jatan Buch
Affiliation:
Department of Earth and Environmental Engineering, Columbia University, New York City, NY, USA
Kara D. Lamb
Affiliation:
Department of Earth and Environmental Engineering, Columbia University, New York City, NY, USA
Pierre Gentine
Affiliation:
Department of Earth and Environmental Engineering, Columbia University, New York City, NY, USA
*
Corresponding author: Kyleen Liao; Email: kyleenliao@stanford.edu

Abstract

The increasing size and severity of wildfires across the western United States have generated dangerous levels of PM2.5 concentrations in recent years. In a changing climate, expanding the use of prescribed fires is widely considered to be the most robust fire mitigation strategy. However, reliably forecasting the potential air quality impact from prescribed fires, which is critical in planning the prescribed fires’ location and time, at hourly to daily time scales remains a challenging problem. In this paper, we introduce a spatio-temporal graph neural network (GNN)-based forecasting model for hourly PM2.5 predictions across California. Utilizing a two-step approach, we use our forecasting model to predict the net and ambient PM2.5 concentrations, which are used to estimate wildfire contributions. Integrating the GNN-based PM2.5 forecasting model with simulations of historically prescribed fires, we propose a novel framework to forecast their air quality impact. This framework determines that March is the optimal month for implementing prescribed fires in California and quantifies the potential air quality trade-offs involved in conducting more prescribed fires outside the peak of the fire season.

Information

Type
Application Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Table 1. Dataset predictors and PM2.5-GNN node features

Figure 1

Figure 1. FRP aggregated around a given radius for each PM$ {}_{2.5} $ monitor location using wind and distance information.

Figure 2

Figure 2. Graph neural network (GNN) used in our PM2.5 forecasting model considers PM2.5 monitors as nodes in the graph and produces node-level predictions.

Figure 3

Table 2. Training/validation/testing split

Figure 4

Figure 3. Heat maps illustrating the observed PM2.5 concentrations versus the PM2.5-GNN model predictions for 1-, 12-, and 48-hour forecasts, with log-transformed axes and color scale. Also indicated is the identity line (dotted black) and $ {R}^2 $ values of the best-fit linear model.

Figure 5

Table 3. Results of the PM2.5-GNN, RF, LSTM, and MLP models

Figure 6

Figure 4. PM2.5 predictions 1 hour into the future from a temporal subset of testing results for example sites. The PM2.5-GNN (Column 1), Random Forest, LSTM, and MLP all use the WIDW FRP predictor, while the PM2.5-GNN with IDW FRP (Column 2) uses the IDW FRP predictor.

Figure 7

Table 4. Results of the PM2.5-GNN with WIDW FRP and IDW FRP

Figure 8

Figure 5. Conceptual diagram of the methodology for distinguishing fire-specific and ambient PM2.5 concentrations.

Figure 9

Figure 6. Ambient and fire-specific PM2.5 predictions 1 hour into the future from a temporal subset of testing results for example sites.

Figure 10

Figure 7. Schematic illustration of the methodology used in Experiment 1 to generate PM$ {}_{2.5} $ predictions based on simulated prescribed burns and observed fire events throughout 2021.

Figure 11

Table 5. Comparing the results of PM2.5 predictions based on simulated prescribed fires in Experiment 1 (see text for more details) for each month of 2021

Figure 12

Figure 8. Schematic illustration of the methodology used in Experiment 2 for simulating prescribed burns during spring and the absence of the Caldor Fire during the wildfire season.

Figure 13

Table 6. Comparing the predicted PM2.5 effect of simulated prescribed burns in Experiment 2 (see text for more details) with baseline PM2.5 predictions

Figure 14

Figure 9. Maximum PM2.5 predictions per site from 8/14/21–12/31/21 under conditions (a) with prescribed burns at the Caldor Fire location during the spring and without Caldor Fire during the wildfire season and (b) without prescribed burns at the Caldor Fire location and with the Caldor Fire during the wildfire season.

Figure 15

Figure 10. Heat maps illustrating the observed versus predicted PM2.5 concentrations for 1-hour, 12-hour, and 48-hour forecasts, with log-transformed axes and color scale, for the Random Forest, LSTM, and MLP models. Also indicated are the identity line (dotted black) and $ {R}^2 $ values of the best-fit linear model.

Author comment: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R0/PR1

Comments

With climate change driving increasingly large and severe wildfires, dangerous levels of PM<sub>2.5</sub> pollution are being produced, posing significant health risks and causing millions of deaths annually. Our research applies machine learning to forecast PM<sub>2.5</sub> concentrations in California and estimate fire-specific PM<sub>2.5</sub> contributions.

Furthermore, our work quantifies the impact of prescribed burns, which while preventing wildfires also generate PM<sub>2.5</sub> pollution, leading to air quality concerns in nearby communities. To the best of our knowledge, our work is the first to use machine learning to predict the PM<sub>2.5</sub> concentrations from prescribed fires. We developed a novel framework and conducted experiments, which can assist the fire services in better understanding and minimizing the pollution exposure from prescribed fires.

We believe that our research is well-suited for publication in the Environmental Data Science (EDS) Special Collection for Tackling Climate Change with Machine Learning. Our work leverages machine learning to forecast PM<sub>2.5</sub> pollution during wildfires and prescribed burns, addressing both a critical environmental issue and a mechanism for its mitigation in a warmer, drier future.

Review: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R0/PR2

Conflict of interest statement

Reviewer declares none.

Comments

The paper introduces a novel application of spatial-temporal Graph Neural Networks (GNNs) to forecast PM2.5 concentrations resulting from prescribed fires in California. Initially, a GNN model forecasts total PM2.5 using meteorological, fire, and satellite data. Subsequently, a second GNN predicts ambient PM2.5 based solely on environmental data. By subtracting the predictions of the second model from those of the first, the method isolates wildfire-specific PM2.5, enhancing air quality assessments and supporting policy-making.

**Strengths:**

The paper pioneers the application of GNNs for predicting PM2.5 concentrations from prescribed fires, effectively addressing a significant environmental challenge and outperforming traditional models. The study presents a comprehensive methodology, featuring a detailed GNN architecture and a large dataset. The authors enhance understanding by meticulously detailing algorithm implementations, including model data features and schematic illustrations. Additionally, the open-sourcing of the code facilitates easy deployment in practical scenarios, improving the accessibility and applicability of the research.

**Weaknesses:**

The study introduces a promising application of Graph Neural Networks (GNN) for forecasting PM2.5 from prescribed burns but lacks extensive comparisons with established methods like XGBoost or Random Forests, which could benchmark the GNN’s performance improvements. Furthermore, expanding the comparative analysis to demonstrate how traditional methods have been applied in managing prescribed burns and air quality monitoring could provide quantitative insights into the GNN’s advantages in predictive accuracy and operational efficiency. This analysis would be highly valuable for practitioners and policymakers to understand the benefits of integrating advanced machine learning into environmental management practices.

Review: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R0/PR3

Conflict of interest statement

Reviewer declares none.

Comments

Liao et al. present an interesting study of using a GNN framework to simulate prescribed (Rx) fires under different environmental scenarios. Overall, the paper has potential to be of interest to the wildfire/Rx fire community and can be an innovative application of ML to this field. However, I have many issues and comments that needs to be addressed before recommending for publication.

Intro:

“Furthermore, the extensive calculations in CTMs make it challenging to explore a large range of parameters for simulating prescribed burns”

Have there been any CTM studies of Rx fires in the western US? Seems like the emissions and the area treated are so small that there would not be much of a precedent for CTMs to simulate this given that their grid boxes are >10km.

“In air quality predictions, machine learning models have been shown to outperform

CTMs in terms of accuracy and computational burden [13]. While several studies have used machine learning to forecast air quality [6, 14], this is the first research paper, to the best of our knowledge, that utilizes machine learning to predict the PM2.5 concentration from simulated prescribed fires.”

I think the authors should provide more of a background of ML applications of wildfire smoke air quality. There exist multiple data products looking at wildfire smoke PM2.5:

Marissa L Childs, Jessica Li, Jeff Wen, Sam Heft-Neal, Anne Driscoll, Sherrie Wang, Carlos F Gould, Minghao Qiu, Jennifer A Burney, and Marshall Burke. 2022. “Daily local-level estimates of ambient wildfire smoke PM2.5 for the contiguous US.” Environmental Science & Technology.

Aguilera R, Luo N, Basu R et al. 2023. A novel ensemble-based statistical approach to estimate daily wildfire-specific PM2.5 in California (2006–2020). Environ. Int. 171:107719

And characterizing their accuracies among methods:

Considine, E. M., Hao, J., de Souza, P., Braun, D., Reid, C. E., & Nethery, R. C. (2023). Evaluation of model-based PM2.5 estimates for exposure assessment during wildfire smoke episodes in the western U.S. Environmental Science and Technology, 57(5), 2031–2041. https://doi.org/10.1021/acs.est.2c06288

Qiu M, Kelp M, Heft - Neal S, Jin X, Gould CF, Tong DQ, Burke M (2024) Evaluating estimation methods for wildfire smoke and their implications for assessing health effects. Environmental Science & Technology In review,. doi:10.31223/X59M59.

Dataset:

“the Julian date and hour of the day are also included as predictors to provide the model with additional context.”

Doesn’t this reduce generalizability of the ML model? How much less accurate is the model without using these variables as predictors? Assuming they are ~top 3 important if you are including them.

Do you regrid any of your data to a common mesh? All these data products have different scales and ERA5 data has a strict grid. Otherwise, do you extract the information of each FRP observation as vector pixels?

Overall, this data section is way too short. There is no sense of how much data is involved with each variable. Most prescribed fire treatments in California are small, on the order of less than 1 km, while your IDW goes up to 500 km. How would the signal of a Rx fire not be drowned out with such differences in scales. The CAL FIRE Rx database (assuming this is ds397? Citation was not exact) contains shapefiles of Rx fire perimeters. How does this overlap/compare with FRP observations?

3. PM2.5 forecasts:

“For instance, the graph explicitly includes wind direction information and considers geographical elevation differences.”

How is elevation explicitly encoded? It doesn’t seem to be a variable unless you are making assumptions about the surface pressure?

“Furthermore, domain knowledge is also incorporated in the graph representation”

This just seems like a generic throwaway statement. How are you using domain knowledge other than just throwing in a series of variables that you think are important? I find the issue and lack of discussion of scale worrying. Rx fires are tiny and can often not be captured from satellites without some kind of post-processing or statistical downscaling. Your assumptions about using stationary sensors via a GNN + LSTM may be appropriate for wildfire smoke or general regional air pollution where the relationships can be modeled by connected PDEs (e.g., neighbor 1 looks like neighbor 2 and neighbor 2 to 3, etc.). However, Rx fires are so small and transient in time I am not convinced that such a modeling framework would be able to forecast these dynamics.

Results:

“PM2.5 predictions one hour into the future from a temporal subset of testing results for example sites.”

What is the time step of the model? There really doesn’t seem to be enough information about how the model works. Are these cherry-picked case studies? Is it predicting one time step/hour into the future? What utility does it have for land mangers if that is the case. How close are far are these case studies from the reference monitors? There just doesn’t seem to be much background or justification in this section.

“This GNN model is trained on all data variables, including meteorological and fire-related data. The second GNN model, on the other hand, focuses only on predicting the ambient PM2.5 and is thus trained only on the meteorological data.”

Is the first GNN model also trained on PM2.5 data? It’s not in the table.

“All the data during fire events are also excluded because including time points during fires would allow the model to learn the influence of fires from meteorological variables like temperature”

How is this determined? Based on FRP observations? Does this mean most of this data occurs in winter? Typically, Rx fires happen in the spring and fall, while wildfires happen in the summer and fall. Would this result in a biased training set? Need more information on all data and methods.

“However, selecting the time points without fire events is challenging, as PM2.5 particles emitted by a fire can persist in the air for weeks [22].”

Perhaps, but that timescale usually deals with long-range transport. And if you’re only looking at the Bay Area in California, then this shouldn’t be a huge issue.

“The GNN model trained to predict only ambient PM2.5 had an MAE of 6.30 g/m3 and RMSE of 7.80 g/m3 for its predictions on time points without fire influence (times not within 10 days of WIDWFRP 500km values greater than 0.15).”

Also provide a relative error metric please.

“For the fire-specific estimate, there is no ground truth value for the PM2.5 concentration

produced by ambient versus wildfire sources, and thus no metric describing the accuracy of the forecast”

This seems weird to say. Can you describe how this method compares to other wildfire smoke estimation methods used from ambient data like in Childs et al. (2022) and Qiu et al (2024)?

“However, the fire-specific predictions can be assumed to have a comparable accuracy as the GNN model results for the net pollution and the ambient pollution.”

How can you make this assumption? Fire-specific predictions (which so far, I guess this is mostly wildfires based on the results) are incredibly heavy-tailed distributions which would be presumably much more difficult to optimize than general ambient PM2.5?

Figure 6 can be improved; it is visually very boring and doesn’t say much beyond what the text describes.

“The Cal Fire [20] latitude, longitude, and duration data for the prescribed burns are matched with the VIIRS FRP data. The transposed prescribed fire FRP information is combined with the observed meteorological data at the target times and inputted into the GNN model, which produces the PM2.5 predictions.”

Are there any references that suggest that VIIRS FRP is even able to capture the signal from Rx fires? An easy way to test this is to overlap the CAL FIRE Rx perimeters with the VIIRS FRP data at the same time/location. This itself would be a nice contribution to the literature.

Experiment 1:

To rephrase, would you say that this is basically using an ML model framework to input some kind of historic Rx fire information (which still needs more discussion + validation) and simulate what would happen under different weather conditions? If so, please say so in a more straightforward manner and upfront in the abstract and introduction.

Can you add some kind of error range or confidence interval to Table 5? How much overlap is there between MAM versus other months? Why is October lower than the other months in the latter half of the year? There should be more of a scientific interpretation of these results as well, as presumably weather patterns drive these behaviors. (E.g., Diablo winds and Santa Ana winds dissipate later in the year driving more stagnation, or perhaps hotter temperatures drive these results). Is there a way to determine the competing effects of meteorological variables on the output? That would be a great result.

Final thoughts:

I understand that this project was undertaken by a high school student, and I suggest many edits to the manuscript. I do believe that this is an interesting application, and that ML has a potentially important and large role to play for the field of wildfires, Rx fires, and air quality. However, science is being inundated with a lot of computer science applications of ML to different domains in a style where the data is not fully discussed, the history or background of the topic is not fully understood, and the metrics of success rely on simple RMSE or training/testing statistics. I provide several examples throughout my comments to make this paper both innovative with its application of ML to an important, understudied field, but also useful for many in the fire community who would like to understand better the mitigative role of Rx fires.

Recommendation: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R0/PR4

Comments

Having received both reviewer reports, I am recommending major revisions. Specifically, I would like a more detailed method section as detailed by Reviewer #2, and a wider synthesis of the results which is recommended by both reviewers.

Decision: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R0/PR5

Comments

No accompanying comment.

Author comment: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R1/PR6

Comments

With climate change driving increasingly large and severe wildfires, dangerous levels of PM<sub>2.5</sub> pollution are being produced, posing significant health risks and causing millions of deaths annually. Our research applies machine learning to forecast PM<sub>2.5</sub> concentrations in California and estimate fire-specific PM<sub>2.5</sub> contributions.

Furthermore, our work quantifies the impact of prescribed burns, which while preventing wildfires also generate PM<sub>2.5</sub> pollution, leading to air quality concerns in nearby communities. To the best of our knowledge, our work is the first to use machine learning to predict the PM<sub>2.5</sub> concentrations from prescribed fires. We developed a novel framework and conducted experiments, which can assist the fire services in better understanding and minimizing the pollution exposure from prescribed fires.

We believe that our research is well-suited for publication in the Environmental Data Science (EDS) Special Collection for Tackling Climate Change with Machine Learning. Our work leverages machine learning to forecast PM<sub>2.5</sub> pollution during wildfires and prescribed burns, addressing both a critical environmental issue and a mechanism for its mitigation in a warmer, drier future.

Review: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R1/PR7

Conflict of interest statement

Reviewer declares none.

Comments

Thank you to the authors for overhauling the manuscript. I think it reads better. I am still recommending a second round of major revisions because I believe the Caldor Fire case study to be too contrived to include as a main result in this paper, but enjoyed other aspects of the text.

I find having a methods section for each results section very weird for a scientific paper. Would prefer if you just have one results section as there is a lot of redundant information.

Figure 4 labels and graphs are too small to see performance.

“These sites were selected due to their unhealthy observed PM2.5 levels, and demonstrate typical performance of the models for elevated PM2.5.”

--> not sure what unhealthy means here, please be more quantitative

“major contribution of this work is the novel simulation of the effect of prescribed fires”

--> This is not a novelty. It has been done before:

Kiely et al., 2024

https://pubs.acs.org/doi/10.1021/acs.est.3c06421

Highlighting the simulation of historic prescribed fires takes away from the work as these are past or hypothetical simulations. Need to focus on the empirical aspect of this field rather than pure modeling

“Thus, when creating the firerelated input predictors, the FRP values from the prescribed fires are artificially increased by a factor of 100 and transposed together to 2021, thereby simulating large prescribed fires from 3/21/21- 5/31/21.”

This seems odd. Why focus on a very large wildfire when Rx burning is limited in California and most of them are less than 100 acres in size. Would have preferred a smaller case study

“2018 PM2.5 data is chosen because, in comparison to the other years, the 2018 fire season most closely resembles the 2021 fire activity without having fires at the Caldor Fire location.”

How did you determine this? Fire-weather indices are comparable? There is not physical justification here even if it is true

I find the Caldor Fire case study too contrived. You’re just artificially scaling up Rx fires and then saying that it entirely prevents the Caldor Fire. This does not make sense for a large fire like Caldor but would be more reasonable for smaller fires. I would not have this as a main result in this paper

Recommendation: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R1/PR8

Comments

I have received the reviewer report for the revision of “Simulating the Air Quality Impact of Prescribed Fires Using Graph Neural Network-Based PM2.5 Forecasts”, and I am recommending further revisions to this manuscript, specifically regarding the Caldor Fire case study. The reviewer points out that some of the methodological decisions are not supported with evidence. I suggest therefore the inclusion of further evidence to support e.g. the RFP scaling values used in this case study.

Decision: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R1/PR9

Comments

No accompanying comment.

Author comment: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R2/PR10

Comments

With climate change driving increasingly large and severe wildfires, dangerous levels of PM<sub>2.5</sub> pollution are being produced, posing significant health risks and causing millions of deaths annually. Our research applies machine learning to forecast PM<sub>2.5</sub> concentrations in California and estimate fire-specific PM<sub>2.5</sub> contributions.

Furthermore, our work quantifies the impact of prescribed burns, which while preventing wildfires also generate PM<sub>2.5</sub> pollution, leading to air quality concerns in nearby communities. To the best of our knowledge, our work is the first to use machine learning to predict the PM<sub>2.5</sub> concentrations from prescribed fires. We developed a novel framework and conducted experiments, which can assist the fire services in better understanding and minimizing the pollution exposure from prescribed fires.

We believe that our research is well-suited for publication in the Environmental Data Science (EDS) Special Collection for Tackling Climate Change with Machine Learning. Our work leverages machine learning to forecast PM<sub>2.5</sub> pollution during wildfires and prescribed burns, addressing both a critical environmental issue and a mechanism for its mitigation in a warmer, drier future.

Review: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R2/PR11

Conflict of interest statement

Reviewer declares none.

Comments

Thank you for making the suggested changes throughout the review process. Although I still believe the Caldor Fire case study to be a bit tenuous, the authors caveat it well and there is novelty in the methods to warrant publication. Thank you

Recommendation: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R2/PR12

Comments

Thank you for addressing the reviewer’s comments on the revision. I am pleased to accept the article for publication in Environmental Data Science.

Decision: Simulating the air quality impact of prescribed fires using graph neural network-based PM2.5 forecasts — R2/PR13

Comments

No accompanying comment.