Skip to main content icon/video/no-internet

Geostatistics has been defined broadly as the study of phenomena that vary over space. Developed originally to address problems of spatial prediction in the mining industry, the generality of the approach has led subsequently to its application in a diverse range of settings across the geographical, geological, atmospheric, environmental, and epidemiological sciences, among others. Geostatistics offers a collection of primarily probabilistic tools that have been developed to aid the understanding and modeling of spatial variability, with the principal motivation of predicting unsampled values dispersed in space (also termed interpolation). The origins of the discipline are generally traced back to the work of Daniel Krige and others on methods for gold and uranium reserve valuation in the Witwatersrand, South Africa, during the 1950s. These ideas were extended and formalized during the 1960s by the French statistician Georges Matheron, who coined the term geostatistics and, with the publication of a series of seminal texts throughout the 1960s and 1970s, laid the foundations for the modern, and still evolving, discipline practiced today by scientists and engineers worldwide.

Theoretical Background

Deterministic and Probabilistic Modeling

The problem motivating the early pioneers of geostatistics, and still the most common goal of contemporary geostatistical analyses, can be stated in the simplest terms as the prediction of a variable (e.g., the concentration of an ore grade) at locations where it has not been sampled (e.g., across a potentially minable deposit). Such predictions require, at least implicitly, the use of a model describing in some way how the phenomenon of interest behaves at these unsampled locations. Various conceptual approaches exist for the formulation of such a model, and a useful categorization is between deterministic and probabilistic models. In a deterministic model, each unknown value is predicted as a single value with no associated prediction error. Such models are best employed when the physical mechanisms that govern the variable of interest are well understood and established physical equations exist that allow calculation of the unknown value with negligible or no error. In fields such as environmental, epidemiological, and public health sciences, however, the systems of interest are often of such complexity and scale that they retain an inherent unpredictability, even when many of the constituent processes are understood in detail, meaning that a deterministic model may be neither feasible nor appropriate. Probabilistic models represent an alternative paradigm to deterministic approaches by considering the underlying mechanism that generates observed data and, by extension, that determines values at unsampled locations, as a random process. Although the mechanism in question is rarely, if ever, entirely random, the adoption of a probabilistic model provides a framework that can prove extremely useful in both predicting unsampled values and assessing the uncertainty of those predictions. Instead of predicting a single value for each unsampled location with assumed zero error, probabilistic models allow the prediction of a set of possible values with corresponding probabilities of occurrence. Unlike deterministic models, probabilistic models do not necessarily require knowledge of the physical process that generated sample data. Rather, most of the information used is derived from the data.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading