Nuclear regression

from Wikipedia, the free encyclopedia

Under kernel regression ( English kernel regression , hence kernel regression ) refers to a series of non-parametric statistical methods , in which the function of a random variable output data by means of core density estimation is estimated. In contrast to linear regression, the type of dependency, represented by the regression curve, is not defined as linear. The advantage is a better adaptation to the data in the case of non-linear relationships. Depending on whether the initial data itself is random or not, a distinction is made between random design and fixed design approaches. The basic procedure was proposed independently in 1964 by Geoffrey Watson and Elisbar Nadaraia (English transcription : Elizbar Nadaraya).

One-dimensional nuclear regression

Kernel density estimator

A kernel density estimator for bandwidth is an estimate of the unknown density function of a variable. Is a sample , a core, the core density estimate is defined as:

.

As the graphic on the right shows, the choice of bandwidth is decisive for the quality of the approximation.

Typical kernels with
unlimited carrier carrier
core core
Gaussian kernel Uniform distribution or rectangular core
Cauchy core Triangle core
Picard core Cosine kernel

Epanechnikov core (p = 1)
quartic core (p = 2)
triweight core (p = 3)



Nadaraya-Watson estimator

Linear regression (black) and Nadaraya-Watson estimators with different bandwidths (red: medium, green: large and blue: small)

The Nadaraya-Watson estimator estimates an unknown regression function from the observation data as

with and a core and a bandwidth . The function is a function that assigns observations close to a large weight and observations far from a small weight. The bandwidth defines in which area the observations are of great importance.

While the choice of the core can usually be made quite freely, the choice of the bandwidth has a great influence on the smoothness of the estimator. The graph on the right shows that a large bandwidth (green) leads to a smoother estimate than choosing a small bandwidth (blue).

Derivation

The idea of ​​the Nadaraya-Watson estimator is based on the fact that the unknown regression function

with the help of the conditional expectation value is represented by the common density and the edge density .

The unknown densities and are estimated using a kernel density estimate. A bivariate kernel density estimator with product kernel and bandwidths is used to calculate the common density from the observations :

.

It follows

and using kernel density estimation for the Nadaraya-Watson estimator.

properties

Weights for different , and belt widths .

1. As in the case of linear regression, the Nadaraya-Watson estimator can also be written as a linear combination of those with weight functions:

.

The Nadaraya-Watson estimator is thus the (locally) weighted mean of the observed values , it applies

.

The graphic on the right shows the weights for different values ​​of (blue :, green:, red:) . The scatter plot below zero shows the data of the explanatory variable. The larger the bandwidth (solid line vs. dashed line), the more observations um have a non-zero weight. The less data are available (right), the more the available observations have to be weighted.

2. The mean square deviation results approximately as

with and independent of and . This means that the convergence is slower than with linear regression, i.e. H. with the same number of observations, the predictive value can be estimated more precisely in the linear regression than in the Nadaraya-Watson estimator.

It is the squared bias ( English bias ) of Nadaraya-Watson estimator

with and the first and second derivative of the unknown regression function, the first derivative of the density and .

And the variance of the estimator

with and .

Bandwidth selection

Resubstitution and leave-one-out cross-validation for the bandwidth of the Nadaraya-Watson estimator for the above example. The "optimal" bandwidth results for approx .

The main problem with nuclear regression is choosing an appropriate bandwidth . The minimization of the mean square deviation serves as the basis

or their approximation. However, the approximation contains the second derivative of the unknown regression function as well as the unknown density function and its derivative. Instead, it uses the data-based mean square deviation

minimized. Since the value of is used to estimate, a bandwidth leads to a (resubstitution estimate). Therefore a leave-one-out cross-validation is performed, i.e. H. all observations except the i-th are used to calculate the estimated value . This is used to calculate the for different bandwidths. The bandwidth that gives a minimum ASE is then used to estimate the unknown regression function.

Confidence bands

After estimating the regression function , the question arises as to how far it deviates from the true function . The work by Bickel and Rosenblatt (1973) provides two theorems for point-wise confidence bands and uniform confidence bands .

In addition to information about the deviation between and , the confidence bands provide an indication of whether a possible parametric regression model, e.g. B. a linear regression, fits the data. If the estimated course of the regression function of the parametric regression model lies outside the confidence bands, this is an indication that the parametric regression model does not fit the data. A formal test is possible with the help of bootstrapping procedures .

Linear regression (black) and Nadaraya-Watson estimator (red) with optimal bandwidth and point-wise 95% confidence band.

Point-wise confidence bands : Under certain conditions converges in distribution

with , and .

If the bandwidth is small enough, then the asymptotic bias can be neglected against the asymptotic variance . This allows approximate confidence bands to be calculated

with the quantile of the standard normal distribution . The unknown density is estimated with a kernel density estimate and with

.

The graph on the right shows the Nadaraya-Watson estimator with a 95% confidence band (red lines). The black linear regression line lies clearly outside the confidence band in various areas. This is an indication that a linear regression model is not appropriate here.

Uniform confidence bands : Under somewhat stronger conditions than before and with , with and for cores with carriers in converged

With

.

The condition is not a restriction, as the data can only be transformed to the interval . The confidence band is then calculated and transformed back to the original data.

Gasser-Müller estimator

In the fixed design case with , the density is known and therefore does not have to be estimated. This simplifies both the calculations and the mathematical treatment of the estimator. In this case, the Gasser-Müller estimator was defined as

With

and , and .

properties

1. Like the Nadaraya-Watson estimator, the Gasser-Müller estimator is a linear estimator and the sum of the weight functions is one.

2. The following applies to the mean square deviation:

.

Local nuclear polynomial regression

Local approximations for the Nadaraya-Watson estimator (locally constant) and the locally linear estimator at selected data points. The graphic is restricted to the range of the x-values ​​(i.e. left margin of the data), but the calculations were carried out with all data.

The Nadaraya-Watson estimator can be written as a solution to the following local minimization problem:

,

i.e. for each a locally constant value is determined which is equal to the value of the Nadaraya-Watson estimator at that point .

A polynomial can also be used instead of a local constant:

,

d. H. the unknown regression value is approximated by a local polynomial . The local polynomial nuclear regression results at every point from

.

The graphic on the right shows the local polynomials used at selected points . The Nadaraya-Watson estimator (red) uses locally constant functions . The locally linear nuclear regression (blue) uses locally linear functions at the point . The selected positions are identical in the graphic with data points. The vertical gray lines connect the local polynomials with the associated x-value (data point). The intersection with the red or blue polynomial gives the estimated value at the corresponding point for the Nadaraya-Watson estimator and the locally linear nuclear regression.

Advantages and features

The local polynomial regression offers several advantages over the Nadaraya-Watson estimator:

  • In general, the local constant is influenced by observation values ​​that are both left and right of the value . At the edges of which, however, does not work and this leads to edge effects ( English boundary effects ). However, the local polynomial nuclear regression approximates locally with a polynomial and can avoid this problem.
  • To appreciate the te derivation, one could simply derive the Nadaraya-Watson correspondingly often. With the local polynomial nuclear regression, however, there is a much more elegant way:
Mostly is used or . Odd orders are better than even orders.
  • As in the case of the linear regression and the Nadaraya-Watson estimator, the locally polynomial nuclear regression can also be written as a linear combination of the weight functions:
.

Estimation of the regression parameters

Define the following matrices:

,

and

the regression parameters are estimated as

.

The coefficients necessary for the derivation are automatically calculated in the estimation process!

To make the estimate practical, one calculates

and calculated

Locally linear nuclear regression

Different local regression methods: Nadaraya-Watson (red), local-linear (blue) and LOWESS (green) and linear regression (black).

One of the best-known locally linear regression models ( ) is the locally weighted regression scatter plot smoother , abbreviated as LOESS or obsolete LOWESS ( English for locally weighted scatterplot smoothing , German locally weighted scatter plot smoothing ). However, the LOWESS is not a local-linear nuclear regression, because

  • the regression weights are estimated robustly and
  • the bandwidth varies with .

The graphic on the right shows two different methods of nuclear regression: locally constant (red, Nadaraya-Watson) and locally linear (blue). The local linear core regression approximates the data somewhat better, especially at the edges.

The locally linear nuclear regression results as

.

The mean square deviation of the locally linear regression results, as with the Nadaraya-Watson estimator, as

With

and the variance is identical to the variance of the Nadaraya-Watson estimator . The simpler form of the distortion makes the local core linear regression more attractive for practical purposes.

Individual evidence

  1. Elizbar A. Nadaraya: On estimating regression . In: Theory of Probability and its Applications . tape 9 , no. 1 , 1964, pp. 141-142 , doi : 10.1137 / 1109020 .
  2. ^ Geoffrey S. Watson: Smooth Regression Analysis . In: Sankhya: The Indian Journal of Statistics, Series A . tape 26 , no. 4 , December 1964, p. 359-372 .
  3. Bickel, Rosenblatt (1973) On some global measures of the deviations of density function estimators , Annals of Statistics 1, pp. 1071-1095
  4. ^ Theo Gasser, Hans-Georg Müller: Estimating Regression Functions and Their Derivatives by the Kernel Method . In: Scandinavian Journal of Statistics . tape 11 , no. 3 , 1984, pp. 171-185 .
  5. ^ WS Cleveland: Robust Locally Weighted Regression and Smoothing Scatterplots . In: Journal of the American Statistical Association . tape 74 , no. 368 , December 1979, p. 829-836 , JSTOR : 2286407 .

literature

  • Jianqing Fan, Irene Gijbels: Local Polynomial Modeling and Its Applications . Chapman and Hall / CRC, 1996, ISBN 978-0-412-98321-4 .
  • Wolfgang Härdle , Marlene Müller, Stefan Sperlich, Axel Werwatz: Nonparametric and Semiparametric Models . Springer Verlag, Berlin, Heidelberg 2004, ISBN 978-3-540-20722-1 ( hu-berlin.de ).
  • Tristen Hayfield, Jeffrey S. Racine: Nonparametric Econometrics: The np Package . In: Journal of Statistical Software . tape 27 , no. 5 , 2008 ( jstatsoft.org ).
  • MP Wand, MC Jones: Kernel Smoothing . Chapman and Hall / CRC, 1994, ISBN 978-0-412-55270-0 .