Coefficient of Determination

Jump to: navigation, search

A Coefficient of Determination is a statistic for the amount of variability of a continuous dependent variable that can be accounted for by a regression model (typically a linear regression model) on one or more regressors



    • QUOTE:In statistics, the coefficient of determination R2 is used in the context of statistical models whose main purpose is the prediction of future outcomes on the basis of other related information. It is the proportion of variability in a data set that is accounted for by the statistical model.[1] It provides a measure of how well future outcomes are likely to be predicted by the model.

      There are several different definitions of R2 which are only sometimes equivalent. One class of such cases includes that of linear regression. In this case, if an intercept is included then R2 is simply the square of the sample correlation coefficient between the outcomes and their predicted values, or in the case of simple linear regression, between the outcomes and the values of the single regressor being used for prediction. In such cases, the coefficient of determination ranges from 0 to 1. Important cases where the computational definition of R2 can yield negative values, depending on the definition used, arise where the predictions which are being compared to the corresponding outcomes have not been derived from a model-fitting procedure using those data, and where linear regression is conducted without including an intercept. Additionally, negative values of R2 may occur when fitting non-linear trends to data.[2] In these instances, the mean of the data provides a fit to the data that is superior to that of the trend under this goodness of fit analysis.

    • QUOTE:The better the linear regression (on the right) fits the data in comparison to the simple average (on the left graph), the closer the value of [math]R^2[/math] is to one. The areas of the blue squares represent the squared residuals with respect to the linear regression. The areas of the red squares represent the squared residuals with respect to the average value.]] A data set has values yi, each of which has an associated modelled value fi (also sometimes referred to as ŷi). Here, the values yi are called the observed values and the modelled values fi are sometimes called the predicted values.

      The "variability" of the data set is measured through different sums of squares: :[math]SS_\text{tot}=\sum_i (y_i-\bar{y})^2,[/math] the total sum of squares (proportional to the sample variance); :[math]SS_\text{reg}=\sum_i (f_i -\bar{y})^2,[/math] the regression sum of squares, also called the explained sum of squares. :[math]SS_\text{err}=\sum_i (y_i - f_i)^2\,[/math], the sum of squares of residuals, also called the residual sum of squares. In the above [math]\bar{y}[/math] is the mean of the observed data: [math]\bar{y}=\frac{1}{n}\sum_i^n y_i [/math] where n is the number of observations.

      The notations [math]SS_{R}[/math] and [math]SS_{E}[/math] should be avoided, since in some texts their meaning is reversed to Residual sum of squares and Explained sum of squares, respectively. The most general definition of the coefficient of determination is :[math]R^2 \equiv 1 - {SS_{\rm err}\over SS_{\rm tot}}.\,[/math]

  1. Steel, R. G. D. and Torrie, J. H., Principles and Procedures of Statistics, New York: McGraw-Hill, 1960, pp. 187, 287.
  2. Cameron, A.C., Windmeijer, F.A.G., (1997). “An R-squared measure of goodness of fit for some common nonlinear regression models.” In: Journal of Econometrics, Volume 77, Issue 2, April 1997, Pages 329-342.