Hostname: page-component-77f85d65b8-6c7dr Total loading time: 0 Render date: 2026-03-28T06:34:16.205Z Has data issue: false hasContentIssue false

The influence of correlated features on neural network attribution methods in geoscience

Published online by Cambridge University Press:  13 May 2025

Evan Krell*
Affiliation:
Department of Computer Science, Texas A&M University - Corpus Christi, Corpus Christi, TX, USA innovation in COmputer REsearch Lab (iCORE), Texas A&M University - Corpus Christi, Corpus Christi, TX, USA Conrad Blucher Institute for Surveying and Science, Texas A&M University - Corpus Christi, Corpus Christi, TX, USA NSF AI Institute for Research on Trustworthy AI in Weather, Climate and Coastal Oceanography, University of Oklahoma, Norman, OK, USA
Antonios Mamalakis
Affiliation:
NSF AI Institute for Research on Trustworthy AI in Weather, Climate and Coastal Oceanography, University of Oklahoma, Norman, OK, USA Department of Environmental Sciences, University of Virginia, Charlottesville, VA, USA School of Data Science, University of Virginia, Charlottesville, VA, USA
Scott A. King
Affiliation:
Department of Computer Science, Texas A&M University - Corpus Christi, Corpus Christi, TX, USA innovation in COmputer REsearch Lab (iCORE), Texas A&M University - Corpus Christi, Corpus Christi, TX, USA NSF AI Institute for Research on Trustworthy AI in Weather, Climate and Coastal Oceanography, University of Oklahoma, Norman, OK, USA
Philippe Tissot
Affiliation:
Conrad Blucher Institute for Surveying and Science, Texas A&M University - Corpus Christi, Corpus Christi, TX, USA NSF AI Institute for Research on Trustworthy AI in Weather, Climate and Coastal Oceanography, University of Oklahoma, Norman, OK, USA
Imme Ebert-Uphoff
Affiliation:
NSF AI Institute for Research on Trustworthy AI in Weather, Climate and Coastal Oceanography, University of Oklahoma, Norman, OK, USA Department of Electrical and Computer Engineering, Colorado State University, Fort Collins, CO, USA Cooperative Institute for Research in the Atmosphere, Colorado State University, Fort Collins, CO, USA
*
Corresponding author: Evan Krell; Email: evankrell730@gmail.com

Abstract

Artificial neural networks are increasingly used for geophysical modeling to extract complex nonlinear patterns from geospatial data. However, it is difficult to understand how networks make predictions, limiting trust in the model, debugging capacity, and physical insights. EXplainable Artificial Intelligence (XAI) techniques expose how models make predictions, but XAI results may be influenced by correlated features. Geospatial data typically exhibit substantial autocorrelation. With correlated input features, learning methods can produce many networks that achieve very similar performance (e.g., arising from different initializations). Since the networks capture different relationships, their attributions can vary. Correlated features may also cause inaccurate attributions because XAI methods typically evaluate isolated features, whereas networks learn multifeature patterns. Few studies have quantitatively analyzed the influence of correlated features on XAI attributions. We use a benchmark framework of synthetic data with increasingly strong correlation, for which the ground truth attribution is known. For each dataset, we train multiple networks and compare XAI-derived attributions to the ground truth. We show that correlation may dramatically increase the variance of the derived attributions, and investigate the cause of the high variance: is it because different trained networks learn highly different functions or because XAI methods become less faithful in the presence of correlation? Finally, we show XAI applied to superpixels, instead of single grid cells, substantially decreases attribution variance. Our study is the first to quantify the effects of strong correlation on XAI, to investigate the reasons that underlie these effects, and to offer a promising way to address them.

Information

Type
Methods Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Figure 1. Methodology for creating a suite of synthetic benchmarks and using it to analyze the influence of feature correlation on NN attribution methods.

Figure 1

Figure 2. Three randomly selected synthetic SST anomaly samples generated using covariance matrix estimated calculated from the COBE-SST 2 dataset (Japanese Meteorological Center, 2024). Red values are positive and blue values are negative. The black regions represent land; these regions are masked out and are not used as network inputs.

Figure 2

Figure 3. Nine correlation matrices are generated based on the SST anomaly dataset. The pair-wise correlation values are positively shifted to increase overall correlation. The distributions of correlation values are shown in (a) and the absolute values in (b), to confirm that the overall correlation increases even the magnitude of negative correlations are reduced. In (c), the correlation matrices are shown as heatmaps to make it clear that the original relationships are preserved, but their magnitudes shifted.

Figure 3

Figure 4. The four benchmarks suites (a–d) each consist of nine datasets generated using nine different covariance matrices (Figure 3). Points represent the mean performance ($ {R}^2 $) for 10 trained networks.

Figure 4

Figure 5. An example of a synthetic SST anomaly sample and its ground truth attribution, along with XAI attributions from three methods. Each sample is initially a 18$ \times $32 raster of synthetic SST anomalies, but the actual input to the network is a flattened vector with the non-ocean pixels (shown here in black) removed. After filtering the non-ocean pixels, each input sample contains $ D $ = 460 features. Ten trained networks are used to generate explanations, and we show samples from four of them here. Below each XAI, attribution is the Pearson correlation between it and the ground truth. This sample is generated from covariance matrix $ {\varSigma}_1 $, and the networks are trained with $ {10}^6 $ samples.

Figure 5

Figure 6. Input × Gradient summary for four benchmark suits (a–d). These correspond to the four sets of synthetic samples, from 103 (a) to 106 (d). For each, three subpanels are provided to analyze how increasing correlation ($ {\boldsymbol{\varSigma}}_0 $$ {\boldsymbol{\varSigma}}_8 $) influences the agreement between the attribution maps. The top-left panel ($ \alpha $) shows the distribution of correlation between XAI-based and ground truth attributions. The bottom-left panel ($ \beta $) shows the distribution of correlation between XAI attributions between trained model repetitions. The left panel ($ \gamma $) compares the alignment in $ \alpha $ and $ \beta $: for each input sample, the mean correlation between XAI and ground truth is plotted against the mean correlation between the model’s XAI results. The results show that increasing the correlation can substantially increase the variance between XAI attributions, but this is greatly limited with a sufficiently large training set.

Figure 6

Figure 7. Input × Gradient: XAI evaluation metrics. Faithfulness Correlation and Monotonicity Correlation measure how faithful attributions are to the network, and Sparseness measures attributions’ complexity.

Figure 7

Figure 8. SHAP summary for four benchmark suits (a–d). For each, three subpanels are provided to analyze how increasing correlation ($ {\boldsymbol{\varSigma}}_0 $$ {\boldsymbol{\varSigma}}_8 $) influences the agreement between attribution maps. The top-left panel ($ \alpha $) shows the distribution of correlation between XAI-based and ground truth attributions. The bottom-left panel ($ \beta $) shows the distribution of correlation between XAI attributions between trained model repetitions. The left panel ($ \gamma $) compares the alignment in $ \alpha $ and $ \beta $: for each input sample, the mean correlation between XAI and ground truth is plotted against the mean correlation between the model’s XAI results. The results show that increasing correlation can substantially increase the variance between XAI attributions, but this is greatly limited with a sufficiently large training set.

Figure 8

Figure 9. SHAP: XAI evaluation metrics. Faithfulness Correlation and Monotonicity Correlation measure how faithful attributions are to the network, and Sparseness measures their complexity. Each faithfulness metric makes different assumptions to perturb the data (e.g., replacement with 0). Since the results do not vary widely across the two methods, it strengthens our confidence in using the results for our analysis.

Figure 9

Figure 10. Superpixel XAI for $ {10}^3 $ (a) and $ {10}^6 $ (b) training samples, from 1 × 1 to 8 × 8 patch sizes: the Pearson correlation between SHAP results from NNs $ {\hat{\mathcal{F}}}_1 $$ {\hat{\mathcal{F}}}_4 $ with the attribution from $ \mathcal{F} $. The left-hand plots show correlation when SHAP is applied to each superpixel. The right-hand plots show correlation when superpixels attributions are made by simply summing the 1 × 1 SHAP values.

Figure 10

Figure 11. SHAP superpixel attributions (sizes 1 × 1 … 8 × 8) from trained networks $ {\hat{\mathcal{F}}}_1 $$ {\hat{\mathcal{F}}}_4 $ are compared to the ground truth attribution from $ \mathcal{F} $. The networks were trained with $ {10}^4 $ synthetic samples that were generated using covariance matrix $ {\boldsymbol{\varSigma}}_1 $. The input sample (index 9900 in the supplemental results) was not used during network training. For each superpixel size, we calculate the mean correlation between each SHAP attribution and the ground truth attribution. For the ground truth attributions, superpixel values are calculated by summing the 1 × 1 values. The SHAP attributions are calculated by running SHAP directly on the grouped grid cells.

Author comment: The influence of correlated features on neural network attribution methods in geoscience — R0/PR1

Comments

Dear editors of EDS,

Thank you for considering our original research paper “The Influence of Correlated Features on Neural Network Attribution Methods in Geoscience” for publication in Environmental Data Science.

In this research, we investigated the potential issues that arise when using eXplainable AI (XAI) methods to investigate AI models that rely on geospatial input data. XAI methods tend to struggle with correlated features, and gridded spatial data tends to have extensive correlation. As XAI is increasingly used by the environmental science community to investigate AI models, we believe this work provides very relevant observations and recommended strategies to guide practitioners.

This research builds off of two papers previously published in Environmental Data Science. In 2023, Krell et al. published a study that highlighted potential pitfalls that may arise when using XAI methods to investigate geospatial models. In 2022, Mamalakis et al. proposed using synthetic benchmarks for quantitative assessment of XAI techniques. In this research, the authors of these two papers have collaborated to use synthetic benchmarks for a quantitative investigation of how correlated features in geospatial data may influence XAI methods and the user’s interpretation of the model.

Thanks again for the consideration.

Note that we provided both the Latex source files and the compiled PDF. The automatic compilation system was complaining about one of the packages, even though that package is used in the template provided by EDS. Please contact us if there is any need to modify the submission.

Evan Krell

ekrell@islander.tamucc.edu

Review: The influence of correlated features on neural network attribution methods in geoscience — R0/PR2

Conflict of interest statement

Reviewer declares none.

Comments

Review of EDS-2024-0083

Summary:

This article explores how correlations in a training dataset can impact model explainability. By adopting and building upon a previous method for creating synthetic data with “known truths”, the authors can create synthetic datasets with a controlled amount of correlations. The paper is generally well-written, but I have many little comments. After careful consideration, I recommend that the paper undergo minor revisions.

References:

Flora, M. L., C. K. Potvin, A. McGovern, and S. Handler, 2024: A Machine Learning Explainability Tutorial for Atmospheric Sciences. Artif. Intell. Earth Syst., 3, e230018, https://doi.org/10.1175/AIES-D-23-0018.1.

Flora, M., Potvin, C., McGovern, A., & Handler, S. (2022). Comparing explanation methods for traditional machine learning models Part 1: An overview of current methods and quantifying their disagreement. arXiv. https://arxiv.org/abs/2211.08943

Hooker, G., Mentch, L. & Zhou, S. (2021). Unrestricted permutation forces extrapolation: variable importance requires at least one more model, or there is no free variable importance. Statistics and Computing, 31(6), 82. https://doi.org/10.1007/s11222-021-10057-z

Bilodeau, Blair, Natasha Jaques, Pang Wei Koh, and Been Kim. “Impossibility Theorems for Feature Attribution.” Proceedings of the National Academy of Sciences 121, no. 2 (January 2024). https://doi.org/10.1073/pnas.2304406120.

Bommer, P. L., M. Kretschmer, A. Hedström, D. Bareeva, and M. M. Höhne, 2024: Finding the Right XAI Method—A Guide for the Evaluation and Ranking of Explainable AI Methods in Climate Science. Artif. Intell. Earth Syst., 3, e230074, https://doi.org/10.1175/AIES-D-23-0074.1.

Broad Minor Comments:

Flora et al. (2024), a recent AMS AIES paper, has covered many topics in explainability and established some definitions of key terms; one such distinction is the difference between feature relevance (contribution of an input to the model’s output) and feature importance (contribution of an input to the model performance/accuracy). I’ve highlighted some text (non-exhaustive) where the language should be updated to reflect these important (no pun intended) distinctions.

Line 31, pg 3: Remove “importance” from influence/importance; Most attribution methods are feature relevance methods. Shapely Additive Global Importance (SAGE) is an importance attribution method, but I don’t believe the authors are trying to make such a distinction in this study.

Line 52; pg 8: Replace “important” with “relevant”

The authors point out that their benchmark attributions (“groud truth”) as based on their near-perfect approximation function. The authors used R^2 to convey that the model is a good approximation. However, I ask the authors: Can we be sure the model is always a good “local” approximation of the truth? Yes, it has good global predictions, but could the model be deviating from the truth locally? If not, then I think briefly discussing that could strengthen your approach.

A related concept is the Impossible Theorem introduced in Bilodeau et al. 2024; the idea is that local attribution methods are no better than random guesses of model behavior when extrapolated beyond the exact local sample they are meant to “explain”. I’d be curious to see the authors discuss this in the revised manuscript.

I’d be curious to know the author’s thoughts on the Bommer et al. (2023) study that explored methods for exploring different aspects of XAI methods and found LRP was a good performing method. If the authors want to exclude LRP, I think more justification is warranted.

Shapel Additive Explanation section (starting on line 21, pg 9):

Not all SHAP approaches ignore feature dependencies. It is possible to create feature coalitions that respect feature dependences; this extension becomes known as the Owen values (see Flora et al. 2024 for more details). It may not be possible to easily extend this for the DeepShap algorithm. Please clarify that in the revised manuscript.

Excluding LIME:

Perhaps extending it to image-based deep learning models is difficult, but I’ve found that this implementation of LIME (https://github.com/seansaito/Faster-LIME) greatly improves upon the original LIME approach. This was the LIME method used in Flora et al. (2022) and the one available in scikit-explain (https://github.com/monte-flora/scikit-explain). Flora et al. (2022) found that this version of LIME was often comparable to SHAP. The authors do not have to provide LIME in the revised manuscript but could ease their critique of the method and recommend it be explored further.

The Faithfulness Correlation:

This faithfulness metric seems inappropriate. Usually, when permuting data, you do not want to change the marginal distribution. That’s why the SHAP method draws samples from a background dataset or the permutation importance method shuffles the existing dataset. Replacing feature values with the mean or zero it distorts the marginal distribution and the conditional distribution amongst those sets of features. Especially when features are correlated, it is unclear whether this metric measures something useful about the model’s learned relationship or its ability to extrapolate (see Hooker et al. 2021). Ultimately, I don’t think it’s a useful metric, and it could be misleading. Strong evidence or support is warranted if the authors wish to keep the metric in the revised manuscript.

Number of models and their differences:

Line 42, pg 12: So are their 4 models trained for each training sample size? If so, why such a low sample and also what makes the models distinct from one another? This part of the methods was unclear and needs clarification in the revised manuscript.

Specific Minor Comments/Grammar:

In the abstract, the sentence “a network may learn many different relationships so that attributions vary among different trained networks.” You talk about a single network and then talk about multiple networks towards the end of the sentence. I think this could be rephrased for clarity.

For the paragraph starting on line 30, pg 3, Flora et al. (2024) would be a great citation for future readers on all of these explainability terms.

For the citations at the end of line 45-46, Hooker et al. 2021 is another example of limitations of permutation importance, a well-known XAI method.

Line 26, pg 4: Perhaps change the word “learning” to “understanding”; a future reader may interpret the XAI method as “learning” something; akin to an ML method.

The Rashomon set could be introduced in the section starting at the top of page 5 (look for papers by Cynthia Rudin and her research group). It is the statistical idea that for a given dataset, a range of disparate models (based on the relationship they learn) can achieve near-identical (and near-optimal) performance. This idea also leads to the distinction between model-specific versus model-agnostic feature importance discussed in Flora et al. (2024); is this feature only important for this model, or is it important to all models in the Rashomon set? Though this study covers feature relevance and not feature importance, I still think the idea is being explored.

Line 50, pg 5: What is D for the dataset used in this study? I couldn’t find that information.

For example, for the description of the covariance matrices, is this for different input features, or is there only one? That part was unclear and should be clarified in the revised manuscript.

Figure 5: Is there only 1 input feature? Shouldn’t there be an attribution map for each feature? Also if the sample of an input SST? Or is the SST meant to be predicted?

Also, how sensitive are the results in this paper to the size of D? Is it possible that correlated features' “impacts” increase with the increase of D? I think the authors could speculate on that in the conclusion section.

Line 2; pg 6: Is Anomalies suppose to be captialized? It is also capitalized in other locations.

Line 41; pg 8: Are you allowed to reference figures out of order?

Line 9/10, pg 9: This is super nitpicky, but what about using “ * “ instead of a capital X for multiply? Will most readers interpret capital X as multiply in this context?

Line 46. Pg 12: Instead of “The Figure” say Figure 4 for clarity.

Line 35, pg 13: You say “114” models, but on line 43 on the previous page it was 144.

Figure 6 is a really dense figure! Some kind of sub-panel labelling would be helpful. I found it difficult to follow the text and which sub-panels I was supposed to focus on.

Line 46-48,pg 16: Interesting advice! I’ve advocated that XAI is only meaningful once the model performance is great!

Line 41, pg 17: I think this is a great takeaway and something I’ve long considered is likely true for explainability methods in general.

Line 49, pg 17: a comma instead of a period for “1,000”

Line 37, pg 20: In most cases, does applying XAI to a single pixel make sense? We care about spatial features like mesocyclones in supercells, a cold front on synoptic system, etc. So its great to see that XAI method become more useful in the types of applications users would actually like to see them used!

Line 46-47, pg 20: It is exciting to see that SHAP is not significantly influence by correlations. SHAP is a really powerful method and has great theoretical backing.

Line 26-28, pg 21: But it’s not the fault of the XAI; ultimately, the underlying AI model has learned undesirable behaviour, which apparently can be alleviated with more training data! That is a great takeaway.

Review: The influence of correlated features on neural network attribution methods in geoscience — R0/PR3

Conflict of interest statement

Reviewer declares none.

Comments

Dear authors and editors,

I am very glad to review this paper entitled “The Influence of Correlated Features on Neural Network Attribution Methods in Geoscience”, which is quite related to my background. It is a very interesting topic especially under the background of huge progress of XAI in Earth system related fields. There is a mainstream that researchers are using the XAI to gain some insight from a well trained model with good performance. However, due to the randomness and the preconditions of independent distribution in machine learning, correlated features for model to predict or gain insights are potentially risk. This paper is very applicable which provide some evidence to identify the correlated features’ effect the AI and XAI’s behaviors. After carefully reviewing on the manuscript, my decision for this paper to Environmental Science Data is “Minor Revision”. I have several major and minor comments on the paper as follows:

Major comments:

1. Writing formation: The paper presents a thorough and well-documented manuscript. The methodologies are clearly articulated, making the study’s replication feasible. But the writing is in a way of technical report instead of a scientific writing way. Specifically, there are some necessary references should be presented. Here is an example:

“In other cases, a network could learn to exploit novel relationships in the dataset of scientific interest, and exposing these relationships could lead to novel hypotheses about the geophysical phenomena at hand.”

The audiences should have the information on what the “other cases” are and what can prove such a statement. And there more cases in the manuscript the authors should carefully check. In particular, the authors should show if the other researchers have studied this question in the introduction section. On the other hand, authors provide some interesting results and conclusion but they did not explain the reason making it more technical, which lack of relevance of Environmental science data.

2. Clarity in Methodology: The methods section is particularly well-written, providing clear and concise descriptions of the procedures and techniques used. This clarity enhances the overall readability and comprehension of the study.

3. Relevance and Applicability: This research is highly relevant, especially considering the significant advancements in XAI within Earth system-related fields. The study addresses a critical aspect, highlighting the mainstream practice where researchers leverage XAI to extract insights from well-trained models with good performance. However, the paper aptly points out the potential risks associated with correlated features due to the inherent randomness and the preconditions of independent distribution in machine learning. This aspect is crucial for practitioners who seek to utilize XAI but face challenges related to dataset correlations and XAI methodologies.

4. Significance and Conclusions: The paper provides excellent guidance for those encountering difficulties with XAI, datasets, and correlation issues. The conclusions drawn are practical and useful, offering valuable insights for future research and applications in this domain. However, I insist that one case study on SST anomaly is not enough for the geoscience fields as the title. At first, SST is very different from the land variables which are more related to geoscience. As the climate changing and intervene by human activities, the other variables have more variability leading the variables with more difficulties in modelling. Here, I strongly recommend authors consider more variables in the experiments, which are more representative in the geoscience field.

Minor comments and questions:

1. Keywords might be misleading. Please clarify what “regression problems” is referred to.

2. Impact statement section should announce the main function and significance of this paper. For example, I think authors should highlight the methods in this section which is referred to how the methods work and how the method help in this field.

3. “In prior work, Krell et al. demonstrated that geophysical models can be greatly influenced by highly correlated features in high-dimensional gridded datasets [16]. Applying several XAI methods to explain FogNet, a 3D Convolutional Neural Network (CNN) for coastal fog prediction [16], Krell et al. showed that explanations were highly sensitive to how correlated features were grouped before applying XAI. Based on the scale at which features were investigated, explanations could yield completely opposite descriptions of model behavior.”

The conclusion of this study is not align with the findings in [16]. What do the authors think are the possible reasons? It needs clarification.

4. Broadly, there are two main ways in which correlated features may influence XAI results and compromise its utility. First, correlated features can negatively affect the XAI’s faithfulness to the network, i.e., make it difficult for XAI to accurately identify the important features that the network used to predict. Second, correlated features can increase the chances for a network to learn many, possibly infinite, different combinations of the features to achieve approximately the same performance.

In my opinion, there might be some misleading here. At first, there is a little confusion on the concept of XAI faithfulness. Commonly, faithfulness measures the extent to which an XAI method accurately identifies important features that genuinely influence a model’s prediction; if an XAI method assigns high relevance, it should actually change the outcome (Alvarez-Melis & Jaakkola, 2018; Bommer et al., 2023). Thus, through my consideration, “correlated features can negatively affect the XAI’s faithfulness to the network” maybe not very appropriate. Because the faithfulness is related to the nature of XAI method. In other words, the correlated features probably cause the network can not recognize the important features, but it is not refer to faithfulness. On the other hand, I agree on “correlated features can increase the chances for a network to learn many, possibly infinite, different combinations of the features to achieve approximately the same performance.”, which is the reason why XAI is hard to identify the important features. In my theory, the correlated features cause trouble in training a network which makes it difficult when the model need to choose a more important feature from correlated features. It would bring randomness that might lead unmeasurable uncertainty in both AI and XAI.

5. All experiments are varying N ∈ [103, 104, 105, 106]. And the R2 of larger dataset is unsurprisingly higher. However, larger dataset might cause overfitting. How did the authors avoid such a problem? I think more test should be done to find the effect of overfitting.

6. The hidden layers contain 512, 256, 128, 64, 32, and 16 neurons, respectively, and are connected between ReLU activation functions. The final output is a single neuron using a linear activation function. The network is trained using a mean squared error loss function.

All the models for different dataset are using the same model setting? I don’t think it is appropriate. With smaller dataset, they can not fit such a complex model structure. I think the model should be fine-tuned with different structure to achieve a similarly high performance instead of using the same structure to test the difference of performance. Because authors want to detect the number of dataset’s impact on the explanation. At first, the same structure can not ensure the same model. Second, with the same structure, model with smaller dataset can not perform as well as those with larger dataset. it indicates that the comparison is not fair. The model with smaller dataset have little ability to detect the real relationships between input variables.

7. The case study is on the SST. The land pixels in the input are masked but they identified importance/attributions as my previous research. How to deal with these pixel importances?

8. In each case, the last 10% of the data is reserved for validation.

What does “last” mean? Is it time series? Please clarify.

9. Figure4: For the case of using 103 samples, possible disagreement between the XAI-based and ground truth attributions may be likely due to the networks not capturing the known function as well. Why S2-5 in 103 have a relatively good r2? As I expected, more correlation lead lower r2. Because more correlation means less useful information fed to the model.

10. “LIME consistently failed to provide consistent attributions, either among the set of trained networks or with the ground truth. Our findings corroborate the observations of other researchers that LIME fails to produce stable or accurate explanations in many situations. Since the LIME results were not useful from the beginning, we deemed them not useful for analysing how they are influenced by correlation.”

Would you provide some examples in other studies? If it fails, it is unnecessary to introduce in the methods.

11. “Gradient results are used mainly as a sanity check. We expect the other methods to consistently have a stronger match with the ground truth since Gradient alone is not a true attribution method.”

Would you provide some examples in other studies? If it fails, it is unnecessary to introduce in the methods.

12. “While correlations do not degrade mean attribution performance, the variance of attribution increases considerably. Even the second covariance matrix widens the distribution around the mean slightly. By the sixth covariance matrix, there is a considerable increase in variance. For covariance matrices 5 - 8, many explanations have a strong alignment with the ground truth (correlation ∼ 0.9), but others approach a correlation of zero: practically no relationship to either the ground truth or among the training runs.”

Can you offer the figs to illustrate it? I can not find any evidence to support it or I miss some points.

13. “With 1,000 training samples (Figure 6a), the mean correlation among attributions does not decrease with increased correlation structure in the input. Initially, we expected to see a decline in mean correlation as the correlation strength grows because more relationships should be available for the network to learn. However, since all correlations are strengthened equally, the original relationships between the inputs and target are preserved (and strengthened as well), which we suspect allows a high performing network to still identify the best relationships. In fact, the mean correlation initially increases. We suspect this is because the increased correlation smooths out noise in the original dataset by constraining the values to be similar to each other, making it an easier problem for the network to learn.”

How the model distinguish the noise (useless information) and sparse information? I mean if correlation increases, the original relationship has been reserved by the model, then the real useful information of sparse would be masked. The hypothesis would be hold. However, the author modify the input according to the extent of correlation by adding noise (useless information) which leads to this conclusion. In this way, I think, they should consider some real useful information to enhance the sparseness of dataset but not only add some noise.

14. “As the number of training samples increases (Figures 6b, 6c, and 6d), the mean correlations tend to increase. This meets expectations since the additional training samples improves network performance, as shown in Figure 4. With a sufficient number of samples, the true relationships is consistently found. That is, the network learns to approximate the designed function, as evidenced by consistently strong correlation between learned attributions and the ground truth (Figure 6d).” This is interesting, more data, the correlation is not important. But in fig6d, the different models would give different explanation even the big dataset. Can you explain why?

15. “First, mean agreement between XAI and the ground truth may improve. We expect this is because the complexity of the training dataset is reduced by enforcing a consistent structure across the samples and the network’s prediction performance increases.”

For the fairness of comparison, the R2 should be similar even when the dataset is smaller. This can make sure the models have the same capacity to find the real relationships. In this way, the XAI’s performance can be compared. Procedures: different dataset -> different models -> XAIs: models have the same performance, then smaller dataset choose the wrong way to relationships due to correlation. But bigger dataset won’t.

16. “SHAP results (Figure 9) show very similar characteristics as Input X Gradient.”

I believe authors can only compare the SHAP which is more popular in the earth fields and acknowledged the effective methods. Other XAI can be presented in the supplementary.

17. “Here, we cannot conclude that correlations degrade XAI accuracy. Thus, we argue that indeed, the increased variance of the correlation between XAI-derived and ground truth attributions is not a result for decreasing XAI faithfulness but rather that the networks are learning different functions than the true one to achieve equally high performance”

I agreed. XAI is not affected by the dataset or correlation. Those metric measure the ability of XAIs which is determined by the XAI design. For example, some assumptions such as Gaussian distribution and liner relationships in the LIME.

18. “Each SHAP attribution map was generated using 10,000 evaluations.”

Settings of SHAP would affect the explanation. Did the authors test it?

19. Fig.8d is better than fig.6d, can you explain why?

20. “Based on Figure 10, we make two main observations. First, the correlation between SHAP attributions and the ground truth attribution increases substantially when going from 1x1 to 2x2 superpixels. This suggests that very localized autocorrelation is a strong driver in attribution differences. That is, the differences between attributions are mostly in small neighborhoods. With autocorrelation, the optimal grid cell used in the synthetic function is surrounded by other, very similar grid cells. So the trained networks can distribute the attribution among that cell’s neighborhood.”

I partially agree. Because the complexity of explanation increase, which enhance the difficulty to find the ground truth.

Recommendation: The influence of correlated features on neural network attribution methods in geoscience — R0/PR4

Comments

The paper examines the explainability of features in AI models for geoscience applications. The problem is extremely important and timely. Both reviews find the work of high-quality and the analysis results are interesting. The authors are encouraged to address the minor issues raised in the reviews for a stronger resubmission.

Decision: The influence of correlated features on neural network attribution methods in geoscience — R0/PR5

Comments

No accompanying comment.

Author comment: The influence of correlated features on neural network attribution methods in geoscience — R1/PR6

Comments

Dear editors of EDS,

Thank you for considering our manuscript “The Influence of Correlated Features on Neural Network Attribution Methods in Geoscience” for publication in Environmental Data Science. We have carefully taken into account all comments from the reviewers. We believe that the revised manuscript has several improvements over the original, as described in our Responses to Reviewers document. We also thank the editors for allowing us to extend the submission deadline.

Please note that the two latex “.tex” files we uploaded failed to compile. We provided our compiled PDF as well, so that the created document looks correct. The Latex error log reported being unable to find “etoc.sty”, which is not part of the provided template we accessed through Overleaf. If there are additional steps we need to do related to the upload, we are happy to do so!

Thanks again,

Evan Krell

Review: The influence of correlated features on neural network attribution methods in geoscience — R1/PR7

Conflict of interest statement

Reviewer declares none.

Comments

The authors have adequately addressed my comments and I recommend accept. Great job!

Review: The influence of correlated features on neural network attribution methods in geoscience — R1/PR8

Conflict of interest statement

no

Comments

the authors provided an applicable experiment results which solved a problem for the XAI users in Earth system science. i think it is qualified to published in this journal.

Recommendation: The influence of correlated features on neural network attribution methods in geoscience — R1/PR9

Comments

The current version has addressed all concerns from the reviewers. The paper is ready for publication.

Decision: The influence of correlated features on neural network attribution methods in geoscience — R1/PR10

Comments

No accompanying comment.