Select to view content in your preferred language

Propose renaming "Confidence" via OHSA to "Confidence Int." or "Confidence Interval"

78
1
Monday
CStandley_1
New Contributor

Would Esri be open to renaming “Confidence” to “Confident Interval” or “Confidence Int.” for OHSA results?

I realize this could be rectified quickly by the user by simply renaming it within the "Contents" pane. However, I think it's important Esri considers the potential misinterpretation of the OHSA results by designating "Confidence" alongside hot/cold spots (see first image below). We certainly want to report Confidence Intervals (CI) alongside any designated hot or cold spots since we are using a statistical test with hypotheses. However, use of “Confidence” as the default output designated in the Contents pane (and shown in the below graphic) could result in user statements like “…we demonstrate 90%, 95%, and/or 99% Confidence that a hot/cold spot exists...”. That’s not entirely true. Adding “Interval” or “Int.” could avoid this misinterpretation and lead to more accurate statements like:

  • “We demonstrate statistically significant hot/cold spot(s) (90, 95, 99% Confidence Intervals)…” or
  • “The project contains statistically significant hot/cold spots with 90, 95, 99% Confidence Intervals”

CStandley_1_3-1759158055711.png

It’s a minor suggestion but it’s actually really important statistical vernacular since Confidence Intervals indicate the level of error we are willing to accept for rejecting the Null Hypothesis. In the case of OHSA, it’s a two-tailed test via the following hypotheses:

Null Hypothesis (HO): Points adhere to Complete Spatial Randomness.

Alternate Hypothesis (HA): Points are not randomly distributed, the exhibit spatial patterning – clustering (hot spot) or dispersion (cold spot).

If we have a statistically significant hot/cold spot with a 90% CI, we are admitting there is a 10% chance that we reject the Null Hypothesis when it is true (Type I error). See error information below.

Our p-value provides the probability that the observed result was created by some random process. A small p-value means it is very unlikely that our observed spatial pattern was derived from a random occurrence. We use both the p-value and the Test Statistic (z-score) as inputs to decide whether to accept/reject the Null Hypothesis (above) and at what level of acceptable error via CI. See the graphic directly from ESRI below for p-values and z-scores and the table that relates them to CIs.

 

 

CStandley_1_1-1759157579759.png

 

CStandley_1_2-1759157579759.png

In statistics there are two types of errors:

Type I – False positive – you reject the null hypothesis when you shouldn’t have (it’s true).

  • Related to the probability alpha (α)

Type II – False negative – you fail to reject the null hypothesis when you should have (it’s false).

  • Related to the probability beta (β); 1-β (inverse) is related to your statistical power.
  • Statistical power is the probability that a hypothesis test correctly rejects a false null hypothesis.
    • In essence, will the test detect a true effect?

Thank you,

Chris

1 Reply
EricKrause
Esri Regular Contributor

Hi @CStandley_1 ,

There seems to be a bit of confusion in that the word "confidence" isn't exclusively used for confidence intervals, and here it is referring to something slightly different.  Here, it is referring to a confidence level, not a confidence interval.

The confidence level is tied to the significance (alpha) level.  There are several ways to say exactly the same thing:

  1. A feature is statistically significant at significance level 0.05.
  2. The p-value of a feature is less than 0.05.
  3. A feature is statistically significant with greater than 95% confidence. (This is the context that the label uses).
  4. A feature is not within a 95% confidence interval constructed assuming the null hypothesis is true.

So the symbology of the features is based on the confidence level (or equivalently, the significance level).  While this is related to confidence intervals, to make the labels be about confidence intervals, they would have to change to something like: "Cold Spot (Not Within 90% Confidence Interval)".  People can certainly disagree, but I don't think this would help with understanding the output.

Further, actually constructing confidence intervals for hot spot analysis results is much harder than you might suspect.  If you've done much work with confidence intervals, you're probably familiar with estimating the standard error, then creating a 95% confidence interval using a formula that looks like: Value +/- 1.96 * StandardError.

However, this isn't really valid for hot spot analysis. This is honestly too complicated to try to explain in a forum environment, but there are nuances about when you can and cannot construct confidence intervals of this type (distinctions between sampling distributions and raw value distributions).  For hot spot analysis specifically, if you want to create a confidence interval for the level of "hotness" of a feature, you'd need to use something more sophisticated (like resampling), but the confidence intervals that you'd produce would be different than the confidence level of the symbology (because they would be estimating the variability of very subtly different things).  You're not expected to understand that, but just know that there is a reason that the hot spot analysis tools do not produce confidence intervals and only refer to the confidence level.

0 Kudos