Skip to main content icon/video/no-internet

Nonparametric methods are a class of statistical techniques that use minimal assumptions for both testing and estimation. Here, standard statistical assumptions are often replaced with computationally intensive calculations. Nonparametric methods provide a valuable alternative to classical parametric techniques. They are often called weak assumption statistics because the assumptions required for validity are quite general compared with classical parametric techniques. In many cases, even the weak assumptions made can be further relaxed. As such, conclusions based on nonparametric techniques need not be tempered by qualifying statements about the underlying assumptions. Other advantages are that they (a) are often easy to understand and apply, (b) are especially good for small samples, (c) are frequently appropriate for discrete data, and (d) may be robust with incomplete or imprecise data. Nonparametric techniques date back to 1710 when John Arbuthnott introduced the sign test. More widespread development of nonparametric techniques did not occur until the 1940s when Frank Wilcoxon developed rank-based tests. Nonparametric statistical techniques come in many different forms. Below, three different types of nonparametric techniques are highlighted. The first is the bootstrap, a method for replacing distributional assumptions in statistical tests. The second is an example of a classical nonparametric test. The final technique is that of nonparametric regression.

Bootstrap

The bootstrap relies on resampling. While it has many applications, it is often used when errors in a regression model are nonnormal. Assume that one estimates the following simple regression model:

None

In this model, we assume that the errors, ε, follow a normal distribution. We can relax this assumption using the bootstrap. The bootstrap simulates the sampling distribution for b through sampling from the original data with replacement. In short, we treat the sample as the population and then sample from it. For example, assume that the sample size for this regression model is 100 cases. For the bootstrap, we would take a random sample of size 100 with replacement from the original data. Since we sample with replacement, some data points will appear in this sample more than once. We then reestimate the original model using this new sample and save the new estimate of β. We repeat this process a large number of times. Typically, we would resample and estimate the model from 1,000 to 5,000 times. This results in 1,000 to 5,000 estimates of β. This new set of β estimates serves as an empirical sampling distribution for this parameter. Percentiles of this distribution can serve as confidence intervals for the parameter. With minor adjustments, one can also calculate a p value to test hypotheses about β. The bootstrap is nonparametric since we use an empirical estimate for the sampling distribution instead of assuming that the test statistic is from a t distribution. One strength of the bootstrap stems from the fact that this basic algorithm can be used to provide inferences for a wide class of statistics.

Classical Nonparametric Test

Another common nonparametric technique is the rank-sum test, which is an alternative to the classical two-sample t test of location. Consider an experiment where we select seven students to play a divide-the-dollar game with a computer program. Three of the students are randomly chosen to receive the treatment, a prime expected to make them more altruistic. If the treatment is effective, we would expect that the students who receive it would give away more of their money. If the treatment is without effect, we would expect no such difference across the two sets of students. Relying on the classical parametric techniques, we might translate these expectations into a t test, comparing the mean difference we observe with a null of no mean difference. We would then calculate a p value using the critical value from a t distribution with five degrees of freedom. For this inference, we must assume that the test statistic follows a parametric t distribution—a distribution that can be characterized with parameters for the mean, variance, and other higher moments. As an alternative, we can use the nonparametric rank-sum test to derive the test from the randomization of treatment assignment and avoid the parametric assumption.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading