The spatial statistical analysis history is fraught with spatial autocorrelation (SA) ponderings, mostly questioning the nature and degree of this observational correlation type’s impacts. Tacit awareness of its impending complications emerged in the early 1900s, with time series methodology spuriously guiding thought and practice. Initial debates cast SA as a nuisance, encouraging its expulsion from geospatial data. The first spatial filtering models strove to do this. However, value ultimately bestowed upon SA by such procedures as spatial interpolation (e.g., geostatistical kriging) moderated this excessive action, with a suite of spatial autoregressive models emerging that fostered spatial filtering engendering isolation-but-retention of global SA in data analyses. Next, Getis refocused this effort on local SA statistics to devise an alternative spatial filtering model whose capabilities include response variable and covariate decompositions into disjunct spatial and aspatial components, suggestive of the spatial Durbin specification. One weakness of these models is their strict normal curve theory reliance. More recent MESF model formulation and articulation transcends this drawback. Consequently, today, spatial analysts can tap a wide variety of spatial filtering conceptualizations, the subject matter this article reviews. In doing so, it presents an original spatially autocorrelated gamma variate empirical example, itself a novel literature contribution.
Griffith, D. A. (2024). Spatial Filtering Models. The Geographic Information Science & Technology Body of Knowledge (2024 Edition). John P. Wilson (Ed.). DOI: 10.22224/gistbok/2024.1.10
Two correlation types can burden empirical data with impediments, namely attribute and observational. Conventional multivariate statistics addresses the former within a multicollinearity context that univariate statistics naturally spawned as it matured. The latter has a more nuanced and complicated history (e.g., Griffith, 2020). The statistical filtering notion closely relates to its chronicle trajectory, in a shift from specialized independent and identically distributed (iid) to more general correlated/dependent data analytic frameworks beleaguered by redundant information. Laplace inaugurated envisaging this generic data category in the early 1800s, investigating daily barometric pressure serial correlation (i.e., time series). Next, in the early 1900s, Student (aka Gossett) enlarged its domain to embrace spatial autocorrelation (SA) impacts on the Pearson correlation coefficient, an issue preoccupying some statisticians into the end of the last millennium. Unfortunately, mathematical statistical theory for both varieties lacked momentum for a half-century or more, a period in which Hotelling conceptualized paired observations correlated data (e.g., repeated measures, dependent samples with before-after formats being time series predecessors), a form more akin to multicollinearity, an innovative idea that immediately flourished (e.g., pre-post intervention matched data). Recognizing analytical and inferential complications attributable to observational correlation, statisticians began devising techniques to filter it from their data in a quest to return their analyses to ones at least mimicking iid, the first and simplest being differencing partnered observation duos to compute equality-of-two-means tests, a straightforward arithmetic operation that filters out correlation between paired observations (halving a sample size in doing so). Thus, they initiated a subtraction strategy that eventually extended to time series and SA situations.
Although Yule and Walker began applying autoregressive specifications to time series data in the 1920s and 1930s, filtering serial correlation from such dependent data had to await the 1949 arrival of the Cochran-Orcutt tactic, which again involved the existing differencing protocol, but modified so that subtraction is of a time-lag term. Next, Cliff and Ord’s popularizing transferal of temporal data analytics to spatial series (circa 1970), promoting the formative Whittle-Bartlet-Besag spatial statistics tradition, inspired Griffith to spatialize this Cochran-Orcutt autoregressive-based approach to filter georeferenced data (circa 1980), conserving the differencing protocol by substituting a spatial- for a time-lag subtraction. Eventually, its inflexibilities prompted him to craft Moran eigenvector spatial filtering methodology (MESF; circa 1996), further adapting the differencing practice to subtract a SA component in a way that transports it to its variable’s mean, a breakthrough shortly thereafter (circa 2002) independently replicated by Legendre and his colleagues in statistical ecology that loosely links to geostatistics. In tandem, Getis (circa 1995) autonomously concocted a spatial filtering apparatus whose differencing routine subtraction exploits his Getis-Ord local SA hot/cold spot index, and also is reminiscent of a geostatistics viewpoint. In conclusion, the key message this introductory narrative advocates is that spatial filtering is part of a long, robust, pioneering praxis encompassing cognate disciplines that transforms data ladened with unorthodox correlation into synthetic numerical content imitating the classical iid scenario for inferential and interpretation purposes through differencing processes.
Spatial filtering’s goal is an informative model-based management of SA in geospatial data that accounts for it while enabling legitimate statistical analysis undertakings. Accordingly, by contextualizing it within its boundaries and extent, targeted disciplinary domains, given application utilities, pertinent georeferenced attribute variables, spatial statistical methodologies, and limitations and vulnerabilities, it may be defined as follows (e.g., Getis, 1995; Borcard et al., 2002; Griffith, 2008, 2010; Griffith and Getis, 2016; Griffith, Chun, and Li, 2019; Griffith and Chun, 2022):
Spatial filtering’s overarching objective is to enable a creative and unhindered use of classical statistical techniques, particularly linear and generalized linear model (GLM) regression, when analyzing geotagged data, a brand of correlated/dependent information, which constitutes its importance. Its mechanics are of two types: removal or transferal/isolation. The Cochrane-Orcutt (1949) mechanism genre purges SA. In contrast, its autoregressive counterparts sequester and preserve its response variable constituent, often via its regression residual manifestations, with the autoregressive response (AR) specification ignoring SA in covariates, the simultaneous autoregressive (SAR) specification artificially filtering its exact same nature and degree from all covariates (i.e., most likely just partial filterings), and the spatial Durbin specification—somewhat emulating Cochrane-Orcutt—estimating an individual variable-specific SA term for each covariate (i.e., spatially lagged explanatory variates), frequently without discarding any of them. MESF (e.g., Borcard and Legendre, 2002; Griffith and Peres-Neto, 2006; Griffith et al., 2019), in its elementary mode, extracts SA from regression residuals (i.e., essentially focusing on SA in Y), and shifts it to a (usually regression) equation’s intercept, converting it from a constant to a varying term. In some instances, this algebraic manipulation constructively substitutes SA surrogates/proxies for potentially relevant omitted variables sharing their same map patterns (Griffith and Chun, 2016; Paez, 2019). More generally, it can furnish a portal for the usage of instrumental variables (Le Gallo and Paez, 2013), a practice capable of addressing bias arising from not only omitted variables, but also measurement error (i.e., errors in variables) and simultaneous causality. Meanwhile, duplicating the Cochrane-Orcutt manoeuvre, each covariate also can undergo this decomposition. Getis (e.g., 1995) does a bit of both, basically removing SA from covariates in order to better segregate its response variable Y component. He argues that one always should retain the response variable spatial factor, and exercise discretion with regard to including or excluding each of the covariate spatialized elements. Getis and Griffith (2002), Griffith and Peres-Neto (2006), and Thayn and Simanis (2013) furnish enlightening side-by-side comparisons of a number of these formulations.
A second invaluable spatial filtering advantage eclipses its staunchest competitor, spatial autoregression, and Besag’s family of auto- models, which, as already noted, separate but usually keep at least some SA in an equation, again concentrating on such correlation latent in Y. Regardless, these are the basis of the Cochrane-Orcutt SA elimination scheme. Besag’s auto-Poisson, negative binomial (NB), exponential, and gamma creations are incapable of accommodating positive SA, generally (although perhaps erroneously) viewed as the sole SA nature in practice. Consequently, their publicized successful treatment utilizes Box-Cox transformation inputs into normal approximations, and hence frequently confront serious specification error obstacles (e.g., a la back-transformations). In contrast, MESF negotiates this possibility with ease, whereas Getis’s method does so strictly for normal/Gaussian geospatial random variables.
Besag dismisses auto-exponential and gamma probability model inventions as being unwieldly and lacking any intuitive appeal to spatial analysts, even though their minimal aim is to seclude and then deposit into a spatial lag term SA latent in Y. However, exponential distributions provide elucidatory descriptions of wait times until a certain event occurs, such as Puerto Rican municipio urban area electricity restoration after Hurricane Maria devastated the island in 2017. This case study encompasses positive SA, a condition defied by auto-gamma models, via its geographic diffusion of recovery efforts pursuing a contiguous expansion rebuilding of a dramatically damaged electric grid from 22 coastal power plants, coupled with the United States Federal Emergency Management Agency’s (FEMA’s) aid entrance through an eastern seaport (i.e., the defunct Roosevelt Roads naval base) to all island neighborhoods (the last one being reconnected 328 days after the hurricane). This empirical example furnishes a novel and insightful spatial filtering exemplification involving a spatially autocorrelated (positive SA) gamma variate.
Covariation exists between re-electrification wait times for Puerto Rican urban places (see https://earthobservatory.nasa.gov/images/144371/night-lights-show-slow-recovery-from-maria) and several salient substantive attributes (Figure 1): population density (i.e., intensely inhabited coastal lowland built-up areas received high, whereas sparsely inhabited interior mountainous rural regions received low, priority); elevation (i.e., coastal generating station locations privileged nearby cities, towns, and other human settlements); and, an east-west gradient (i.e., the island’s elongated/elliptical shape conflated with FEMA’s entry point) channeled relief aid movements (e.g., circumnavigating la Cordillera Central, the island’s interior mountain range). One question spatial filtering models help answer here asks about conspicuous SA effects on measured associations for this battery of correlates. The Cochrane-Orcutt device, requiring estimation of, for example, multiple spatial AR models, supplies the simplest reply, producing a contrast between outcomes rendered by the presence and absence of SA in a linear regression analysis. It begins with a Box-Cox power transformation of variables (i.e., attaching a exponent to measurements):
= ‒0.43 for wait times (Y);
= ‒0.17 for 2020 population density (
);
= 0.27 for mean elevation [
; extracted from a digital elevation model (DEM)]; and,
= 1 (i.e., no Box-Cox transformation; but converted to z-scores to ensure numerical stability) for longitude (
). These three covariates account for roughly 50% of the geographic variation (a la
, adjusted-
, predicted-
and GLM pseudo-
indices) across the island in urban wait times, and yield extremely favorable residual diagnostics.
Tables 1-4 respectively tabulate conventional linear and gamma regression (always using a log-link function—capturing positive-only wait time data with positively-skewed errors, sustaining some degree of consistency with the normal approximations, and depicting an underlying multiplicative process) results ignoring the presence of SA in any of the georeferenced variables (i.e., no spatial filtering), isolating and retaining SA in the response variable only (i.e., implicit spatial filtering of Y), in the response variable plus the covariates (i.e., implicit spatial filtering of Y as well as the covariates), and SA removal from all variates (i.e., spatial filtering as though SA is a nuisance, expunging it from all georeferenced variates). One principal finding concerns null hypothesis Type I errors (i.e., false positives, or decision errors of commission). Another entails multicollinearity affiliated with non-intercept (i.e., covariate) parameter estimates.
The remarkable unambiguous revelations documented by Tables 1-4 for this exclusive case study are that the detected weak-to-moderate positive SA encumbering the Puerto Rican geospatial urban wait times data—disregarded by Besag’s auto-gamma model—can mask and/or distort true geospatial relationships. Ignoring SA (Table 1) implies that all three covariates are statistically significant, multicollinearity is not troublesome, a normal approximation (prior to its back-transformation) and a more appropriate gamma probability model furnish equally good data descriptions, and the three covariates together account for roughly half of the geographic variation in urban electricity restoration wait times. Completely eliminating SA in the data (Table 4) tells a different story: population density is the single important covariate—with evidence also signalling distance from the federal relief entrance point to the island being prominent—the normal approximation and gamma specifications do not perform equivalently, multicollinearity remains undisruptive, and the electricity restoration process potentially is more random than map patterned. Preserving SA latent in wait times (Tables 2 and 3), which is both sensible and satisfying given the incremental contiguous expansion of the destroyed electric grid’s restoration, not only maintains, but also bolsters a spatial statistical data description. Elevation appears to be a questionable covariate. In addition, explicitly exposing SA in covariates suggests that it can exacerbate multicollinearity issues (e.g., the spatial Durbin and possibly the Getis filtering cases in Table 3). One overall implication is that normal approximations can be misleading; their results also may be overly optimistic because they still require back-transforming for final decision-making purposes. Therefore, this empirical example nicely illustrates the entire spectrum of spatial filtering perspectives and options currently available to spatial analysts.
This narrative synthesizes a definition of contemporary spatial filtering models, lesser known but very powerful instruments for handling SA that already are in the spatial analysis toolbox, casting them in terms of spatial autoregressive, Getis-Ord statistic decomposition, and MESF models. As such, this construct spans the full range of spatial statistics/econometrics, from Cliff-Ord/Besag type spatial auto-, through Getis’s geostatistically informed, to Griffith’s Moran eigenvector-based, conceptualizations, all supporting full eradication from, or sundry degrees of partial isolation of SA in, geospatial data. As more and more present-day quantitative spatial analysts become convinced that SA is a fundamental, rather than menacing, georeferenced data property, partial solutions that clarifyingly shift SA around in an equation will replace full SA jettisonings as the modus operandi norm. MESF offers a clear advantage in this arena because it not only is capable of efficiently and effectively handling positive-negative SA mixtures, a theme aligned with this article whose widespread existence increasingly is becoming commonly recognized, but also easily integrates with any non-normal probability model having a regression implementation, avoiding the historically dominant normal curve theory approximations and their attendant specification error.