# Outlier

(in statistics)

Any observation in a set of data that is inconsistent with the remainder of the observations in that data set. The outlier is inconsistent in the sense that it is not indicative of possible future behaviour of data sets coming from the same source. Outliers sometimes go by the name of contaminants, spurious or rogue observations, or discordant values.

Valid inferences concerning a particular data set can only be made once one first determines which observations, if any, are potential outliers, and how these observations should be treated in the subsequent analysis. These considerations lead to three main issues associated with outliers: outlier testing, efficient or accommodative estimation, and robust estimation.

Before these issues can be discussed, however, one needs to formulate a mathematical model that describes outliers. One of the most commonly used models, which was originally introduced in its most general form in [a4], is called the slippage model. In this model it is assumed that, out of $n$ observations, $n - p$ arise from some probability distribution with probability density function $f ( x )$( cf. also Density of a probability distribution), while the remaining observations (the outliers) come from some modified form of the original distribution, usually denoted by $g ( x )$. For example, $f ( x )$ could be the normal $N ( \mu, \sigma ^ {2} )$ probability density function (cf. Normal distribution), while $g ( x )$ could be $N ( \mu + a \sigma, \sigma ^ {2} )$. Another model, called the mixture model, assumes that, for each observation, there is a certain probability $\lambda$ that the observation comes from the probability density function $f ( x )$ and a probability $( 1 - \lambda )$ that it comes from the probability density function $g ( x )$. There are many other outlier models; they can all be found in [a1] or [a6], which are both general reference books on outliers.

The first step in effectively dealing with outliers is to perform a statistical test to determine which observations are potential outliers. In this regard, many test statistics are available for a variety of distributions, many of which have been derived using the maximum-likelihood ratio principle with one of the above-mentioned outlier models as the alternative model. For example, the test statistic $T = { {( Y _ {n:n } - {\overline{Y}\; } _ {n} ) } / S }$( where $Y _ {n:n }$ is the largest observation, ${\overline{Y}\; } _ {n}$ is the average of the observations, and $S ^ {2}$ is the sample variance) is the maximum-likelihood ratio test (cf. also Likelihood-ratio test) for the normal model when the alternative model is the slippage outlier model given in the example above with $p = 1$ and $a > 0$. Unusually large values of $T$( given in [a1]) will lead to declaring the largest observation to be an outlier.

Once one has determined that there are potential outliers in the sample, there are two ways to proceed. The first way involves the use of robust methods, that is, methods of analysis that will be minimally affected by the presence of outliers. For example, to estimate the population mean $\mu$ based on a sample of observations, one might use a trimmed mean,

$$T _ {r,n } = { \frac{1}{n - 2r } } ( Y _ {r + 1:n } + \dots + Y _ {n - r:n } )$$

for $r \geq 1$, where $Y _ {1:n } \leq \dots \leq Y _ {n:n }$ are the ordered observations, rather than the full sample mean (which is the trimmed mean with $r = 0$). The trimmed mean excludes entirely the $r$ largest and $r$ smallest observations (which are most likely to be the outliers under the slippage model) from the analysis, and can therefore be expected to perform better than the full sample mean when outliers are indeed present. The performance of an estimator can be assessed based on its bias and mean-square error (cf. also Biased estimator; Error). Such comparisons can be found in [a4] and [a5] for the single outlier ( $p = 1$) normal model. For example, using the slippage model described above when $n = 10$, the full sample mean $T _ {0,10 }$ has mean-square error $0.10 \sigma ^ {2}$ when no outliers are present and $0.26 \sigma ^ {2}$ when there is a single outlier in the sample with $a = 4.0$. On the other hand, the mean-square error of $T _ {1,10 }$ is $0.105 \sigma ^ {2}$ when no outliers are present, and only increases to $0.149 \sigma ^ {2}$ when there is a single outlier with $a = 4.0$. Thus, the robust estimator $T _ {1,10 }$ provides protection (considerably lower mean-square error) against the presence of outliers, for a premium (slightly higher mean-square error when there are no outliers).

Recent advances in this area, which deal with the multiple outlier situation ( $p \geq 1$) for several different parametric models, can be found in [a2] and [a3].

The second method for dealing with outliers, known as efficient estimation, involves looking for a specific estimator that is optimal for the given data set. In this method, the estimator to be used will vary for different data sets, depending on the size of the sample, the number of potential outliers and how pronounced the outliers are. In fact, when outliers are present, the trimmed mean is not always the best estimator. The reason for this is the fact that the trimmed mean entirely excludes observations that may still contain some useful information about the parameter. A less drastic alternative to the trimmed mean involves the use of an estimator that includes the possible outliers, but gives them less weight. One such estimator is the linearly weighted mean,

$$L = { \frac{1}{c} } ( Y _ {1:n } + 3Y _ {2:n } + 5Y _ {3:n } + \dots + 3Y _ {n - 1:n } + Y _ {n:n } ) ,$$

which can be used as a robust estimator of the mean $\mu$ for the normal model. An example of efficient estimation would be to choose between the trimmed mean and linearly weighted mean, depending on which estimator performs the best (lowest mean-square error) for the given data set. Tables to facilitate such a decision for the exponential model, $f ( x ) = ( {1 / \theta } ) e ^ {- {x / \theta } }$, can be found in [a2]. There it is shown that when the outliers are not very pronounced, a weighted type of estimator of $\theta$ is actually more efficient than a trimmed mean, while the reverse is true when the outliers are far removed from the remainder of the data.

#### References

 [a1] V. Barnett, T. Lewis, "Outliers in statistical data" , Wiley (1994) (Edition: Third) [a2] A. Childs, N. Balakrishnan, "Relations for single moments of order statistics from non-identical logistic random variables and robust estimation of the location and scale parameters in the presence of multiple outliers" C.R. Rao (ed.) G.S. Maddala (ed.) , Handbook of Statistics , Elsevier Sci. (1997) [a3] A. Childs, N. Balakrishnan, "Some extensions in the robust estimation of parameters of exponential and double exponential distributions in the presence of multiple outliers" C.R. Rao (ed.) G.S. Maddala (ed.) , Handbook of Statistics , Elsevier Sci. (1997) [a4] H.A. David, "Robust estimation in the presence of outliers" R.L. Launer (ed.) G.N. Wilkinson (ed.) , Robustness in Statistics , Acad. Press (1979) pp. 61–74 [a5] H.A. David, V.S. Shu, "Robustness of location estimators in the presence of an outlier" H.A. David (ed.) , Contributions to Survey Sampling and Applied Statistics: Papers in Honour of H.O. Hartley , Acad. Press (1978) pp. 235–250 [a6] D.M. Hawkins, "Identification of outliers" , Chapman and Hall (1980)
How to Cite This Entry:
Outlier. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Outlier&oldid=48089
This article was adapted from an original article by N. BalakrishnanA. Childs (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article