digplanet beta 1: Athena
Share digplanet:

Agriculture

Applied sciences

Arts

Belief

Business

Chronology

Culture

Education

Environment

Geography

Health

History

Humanities

Language

Law

Life

Mathematics

Nature

People

Politics

Science

Society

Technology

Deming regression. The red lines show the error in both x and y. This is different from the traditional least squares method which measures error parallel to the y axis. The case shown, with deviations measured perpendicularly, arises when errors in x and y have equal variances.

In statistics, Deming regression, named after W. Edwards Deming, is an errors-in-variables model which tries to find the line of best fit for a two-dimensional dataset. It differs from the simple linear regression in that it accounts for errors in observations on both the x- and the y- axis. It is a special case of total least squares, which allows for any number of predictors and a more complicated error structure.

Deming regression is equivalent to the maximum likelihood estimation of an errors-in-variables model in which the errors for the two variables are assumed to be independent and normally distributed, and the ratio of their variances, denoted δ, is known.[1] In practice, this ratio might be estimated from related data-sources; however the regression procedure takes no account for possible errors in estimating this ratio.

The Deming regression is only slightly more difficult to compute compared to the simple linear regression. Many software packages used in clinical chemistry, such as Analyse-it, EP Evaluator, MedCalc, R, S-PLUS and StatsDirect offer Deming regression.

The model was originally introduced by Adcock (1878) who considered the case δ = 1, and then more generally by Kummell (1879) with arbitrary δ. However their ideas remained largely unnoticed for more than 50 years, until they were revived by Koopmans (1937) and later propagated even more by Deming (1943). The latter book became so popular in clinical chemistry and related fields that the method was even dubbed Deming regression in those fields.[2]

Specification[edit]

Assume that the available data (yi, xi) are measured observations of the "true" values (yi*, xi*):

\begin{align}
  y_i &= y^*_i + \varepsilon_i, \\
  x_i &= x^*_i + \eta_i,
  \end{align}

where errors ε and η are independent and the ratio of their variances is assumed to be known:

 \delta = \frac{\sigma_\varepsilon^2}{\sigma_\eta^2}.

In practice the variance of the x and y parameters is often unknown which complicates the estimate of  \delta but where the measurement method for x and y is the same they are likely to be equal so that  \delta = 1 for this case.

We seek to find the line of "best fit"

y^* = \beta_0 + \beta_1 x^*,

such that the weighted sum of squared residuals of the model is minimized:[3]

SSR = \sum_{i=1}^n\bigg(\frac{\varepsilon_i^2}{\sigma_\varepsilon^2} + \frac{\eta_i^2}{\sigma_\eta^2}\bigg) = \frac{1}{\sigma_\varepsilon^2} \sum_{i=1}^n\Big((y_i-\beta_0-\beta_1x^*_i)^2 + \delta(x_i-x^*_i)^2\Big) \ \to\ \min_{\beta_0,\beta_1,x_1^*,\ldots,x_n^*} SSR

Solution[edit]

The solution can be expressed in terms of the second-degree sample moments. That is, we first calculate the following quantities (all sums go from i = 1 to n):

\begin{align}
  & \overline{x} = \frac{1}{n}\sum x_i, \quad \overline{y} = \frac{1}{n}\sum y_i, \\
  & s_{xx} = \tfrac{1}{n-1}\sum (x_i-\overline{x})^2, \\
  & s_{xy} = \tfrac{1}{n-1}\sum (x_i-\overline{x})(y_i-\overline{y}), \\
  & s_{yy} = \tfrac{1}{n-1}\sum (y_i-\overline{y})^2.
  \end{align}

Finally, the least-squares estimates of model's parameters will be[4]

\begin{align}
  & \hat\beta_1 = \frac{s_{yy}-\delta s_{xx} + \sqrt{(s_{yy}-\delta s_{xx})^2 + 4\delta s_{xy}^2}}{2s_{xy}}, \\
  & \hat\beta_0 = \overline{y} - \hat\beta_1\overline{x}, \\
  & \hat{x}_i^* = x_i + \frac{\hat\beta_1}{\hat\beta_1^2+\delta}(y_i-\hat\beta_0-\hat\beta_1x_i).
  \end{align}

The case of equal error variances[edit]

When \delta=1, Deming regression becomes orthogonal regression: it minimizes the sum of squared perpendicular distances from the data points to the regression line. In this case, denote each observation as a point zj in the complex plane (i.e., the point (xj, yj) is written as zj = xj + iyj where i is the imaginary unit). Denote as Z the sum of the squared differences of the data points from the centroid (also denoted in complex coordinates), which is the point whose horizontal and vertical locations are the averages of those of the data points. Then:[5]

  • If Z = 0, then every line through the centroid is a line of best orthogonal fit.
  • If Z ≠ 0, the orthogonal regression line goes through the centroid and is parallel to the vector from the origin to \sqrt{Z}.

A trigonometric representation of the orthogonal regression line was given by Coolidge in 1913.[6]

Application[edit]

In the case of three non-collinear points in the plane, the triangle with these points as its vertices has a unique Steiner inellipse that is tangent to the triangle's sides at their midpoints. The major axis of this ellipse falls on the orthogonal regression line for the three vertices.[7]

Notes[edit]

  1. ^ (Linnet 1993)
  2. ^ Cornbleet, Gochman (1979)
  3. ^ Fuller, ch.1.3.3
  4. ^ Glaister (2001)
  5. ^ Minda and Phelps (2008), Theorem 2.3.
  6. ^ Coolidge, J. L. (1913).
  7. ^ Minda and Phelps (2008), Corollary 2.4.

References[edit]


Original courtesy of Wikipedia: http://en.wikipedia.org/wiki/Deming_regression — Please support Wikipedia.
This page uses Creative Commons Licensed content from Wikipedia. A portion of the proceeds from advertising on Digplanet goes to supporting Wikipedia.
94 videos foundNext > 

Tutorial for MedCalc Deming Regression: 教學

Tutorial for MedCalc Deming Regression: 教學MedCalc 授權經銷商SoftHome International ; Software for Science 13F, NO. 55, SEC.1, CHIEN KUO N-ROAD, TAIPEI, 10491,TA...

SigmaPlot Deming Regression

Deming Regression は、測定誤差を含む 2変数間の回帰直線を求めます。 測定誤差を変数毎に定数で指定するか、測定毎に異なる値で指定します。

免費統計教學範例48 Deming Regression

Tutorial for SigmaPlot V.12 Deming Regression SigmaPlot 科學繪圖統計軟體教學範例V.12 二維及三維科學繪圖軟體The technical graphing software choice for more than 100000 professio...

Polynomial Regression

You can get XLStatistics from the author here: http://www.deakin.edu.au/software/course.php?anchor=xlstatistics.

correlation and regression analysis npart 2

statistical significance of r and b.

MedCalc 12.7.7.0 (X86-X64)

Downlaod Full and Free 100% Click this Link: http://www.full-software-downloads.info/2013/11/medcalc-12-7-7-0-x86-x64.html Software sales Download: http://so...

The Deming Funnel Minitab Demonstration

Thomas Pyzdek shows how to use a Minitab macro to demonstrate Deming's funnel rules.

Mike Tveite on Enumerative and Analytic Studies

Excerpt from 2012 Deming Institute Fall Conference See our blog post on this topic: http://blog.deming.org/2013/05/enumerative-and-analytic-studies/ "An enum...

Partial least squares regression

Partial least squares regression (PLS regression) is a statistical method that bears some relation to principal components regression; instead of finding hyp...

Tutorial for MedCalc Passing-Bablok Regression: 教學

Tutorial for MedCalc: 教學MedCalc 授權經銷商SoftHome International ; Software for Science 13F, NO. 55, SEC.1, CHIEN KUO N-ROAD, TAIPEI, 10491,TAIWAN 全傑科技股份有限公司科學...

94 videos foundNext > 

1 news items

 
Analytik-News (Pressemitteilung)
Wed, 18 Aug 2010 08:41:29 -0700

Einige der neuen "Wizard-gesteuerten" Statistikfunktionen sind der One-Sample Signed Rank-Test und die Deming-Regression zum Methodenvergleich. Dazu bietet die Normalverteilungs-Vergleichsfunktion die Darstellung der Verteilung von bis zu vier ...
Loading

Oops, we seem to be having trouble contacting Twitter

Support Wikipedia

A portion of the proceeds from advertising on Digplanet goes to supporting Wikipedia. Please add your support for Wikipedia!

Searchlight Group

Digplanet also receives support from Searchlight Group. Visit Searchlight