Content deleted Content added
ce |
Owen Reich (talk | contribs) Link suggestions feature: 3 links added. |
||
(34 intermediate revisions by 21 users not shown) | |||
Line 1:
{{Short description|Algorithm for the line of best fit for a two-dimensional dataset}}
[[Image:Total least squares.svg|thumb|Deming regression. The red lines show the error in both ''x'' and ''y''. This is different from the traditional least squares method, which measures error parallel to the ''y'' axis. The case shown, with deviations measured perpendicularly, arises when errors in ''x'' and ''y'' have equal variances.]]
In [[statistics]], '''Deming regression''', named after [[W. Edwards Deming]], is an [[errors-in-variables model]]
Deming regression is equivalent to the [[maximum likelihood]] estimation of an [[errors-in-variables model]] in which the errors for the two variables are assumed to be independent and [[normal distribution|normally distributed]], and the ratio of their variances, denoted ''δ'', is known.
The Deming regression is only slightly more difficult to compute than the [[simple linear regression]]. Most statistical software packages used in clinical chemistry offer Deming regression.
The model was originally introduced by {{harvtxt|Adcock|1878}} who considered the case ''δ'' = 1, and then more generally by {{harvtxt|Kummell|1879}} with arbitrary ''δ''. However their ideas remained largely unnoticed for more than 50 years, until they were revived by {{harvtxt|Koopmans|
== Specification ==
Line 23 ⟶ 24:
We seek to find the line of "best fit"
: <math>y^* = \beta_0 + \beta_1 x^*,</math>
such that the weighted sum of squared residuals of the model is minimized:
: <math>SSR = \sum_{i=1}^n\bigg(\frac{\varepsilon_i^2}{\sigma_\varepsilon^2} + \frac{\eta_i^2}{\sigma_\eta^2}\bigg) = \frac{1}{\sigma_\
See {{harvtxt|Jensen
== Solution ==
The solution can be expressed in terms of the second-degree sample moments. That is, we first calculate the following quantities (all sums go from ''i'' = 1 to ''n''):
:
Finally, the least-squares estimates of model's parameters will be
: <math>\begin{align}
& \hat\beta_1 = \frac{s_{yy}-\delta s_{xx} + \sqrt{(s_{yy}-\delta s_{xx})^2 + 4\delta s_{xy}^2}}{2s_{xy}}, \\
Line 44 ⟶ 45:
==Orthogonal regression==
For the case of equal error variances, i.e., when <math>\delta=1</math>, Deming regression becomes '''orthogonal regression''': it minimizes the sum of squared [[distance from a point to a line|perpendicular distances from the data points to the regression line]]. In this case, denote each observation as a point
*If
*If
A [[trigonometry|trigonometric]] representation of the orthogonal regression line was given by Coolidge in 1913.
===Application===
In the case of three [[Line (geometry)|non-collinear]] points in the plane, the [[triangle]] with these points as its [[vertex (geometry)|vertices]] has a unique [[Steiner inellipse]] that is tangent to the triangle's sides at their midpoints. The [[Ellipse#Elements of an ellipse|major axis of this ellipse]] falls on the orthogonal regression line for the three vertices.
When humans are asked to draw a linear regression on a scatterplot by guessing, their answers are closer to orthogonal regression than to [[ordinary least squares]] regression.<ref>{{cite journal |last1=Ciccione |first1=Lorenzo |last2=Dehaene |first2=Stanislas |title=Can humans perform mental regression on a graph? Accuracy and bias in the perception of scatterplots |journal=Cognitive Psychology |date=August 2021 |volume=128 |pages=101406 |doi=10.1016/j.cogpsych.2021.101406|doi-access=free }}</ref>
== York regression ==
The York regression extends Deming regression by allowing correlated errors in x and y.<ref>York, D., Evensen, N. M., Martınez, M. L., and Delgado, J. D. B.: Unified equations for the slope, intercept, and standard errors of the best straight line, Am. J. Phys., 72, 367–375, https://doi.org/10.1119/1.1632486, 2004.</ref>
==See also==
* [[Line fitting]]
* [[Regression dilution]]
==Notes==▼
{{Reflist}}▼
==References==
* {{cite journal|last=Adcock|first=R. J.|year=1878|title=A problem in least squares|journal=The Analyst|volume=5|issue=2|pages=53–54|doi=10.2307/2635758|jstor=2635758}}▼
▲{{Reflist|30em}}
* {{cite journal|author=Coolidge, J. L.|year=1913|title=Two geometrical applications of the mathematics of least squares|journal=The American Mathematical Monthly|volume=20|issue= 6|pages=187–190|doi=10.2307/2973072|jstor=2973072}}▼
* {{cite journal|author=Cornbleet, P.J.|author2=Gochman, N.|year=1979|title=Incorrect Least–Squares Regression Coefficients|journal=Clinical Chemistry |volume=25|issue=3|pages=432–438|doi=10.1093/clinchem/25.3.432|pmid=262186}}▼
;Bibliography
▲* {{cite journal|last=Adcock|first=R. J.|year=1878|title=A problem in least squares|journal=The Analyst|volume=5|issue=2|pages=53–54|doi=10.2307/2635758|doi-access=free|jstor=2635758|jstor-access=free}}
▲* {{cite journal|author=Coolidge
▲* {{cite journal|author=Cornbleet
* {{cite book|last=Deming|first=W. E.|author-link=W. Edwards Deming|year=1943|title=Statistical adjustment of data|publisher=Wiley, NY (Dover Publications edition, 1985)|isbn=0-486-64685-8}}
* {{cite book|last=Fuller|first=Wayne A.|year=1987|title=Measurement error models|publisher=John Wiley & Sons, Inc|isbn=0-471-86187-1}}
* {{cite journal |last1 = Glaister | first1 = P. | year = 2001 | title = Least squares revisited | journal = [[The Mathematical Gazette]] | volume = 85 | pages = 104–107 | doi=10.2307/3620485| jstor = 3620485 | s2cid = 125949467 }}
* {{cite web |last=Jensen |first=Anders Christian |year=2007 |title=Deming regression, MethComp package |url=
* {{cite book|last=Koopmans|first=T. C.|year=
* {{cite journal
| last = Kummell
Line 79 ⟶ 87:
| issue = 4
| pages = 97–105
| doi = 10.2307/2635646|doi-access=free
| jstor = 2635646|jstor-access=free
}}
* {{cite journal
Line 94 ⟶ 102:
| url = http://www.clinchem.org/cgi/reprint/39/3/424
| pmid = 8448852
| doi-access = free
}}
*{{cite journal
Line 106 ⟶ 115:
| pages = 679–689
| title = Triangles, ellipses, and cubic polynomials
| volume = 115
| year = 2008
| doi = 10.1080/00029890.2008.11920581
| s2cid = 15049234
}}
* {{cite journal|last=Quarton|first=T. G.|year=2020|title=Uncoupling gene expression noise along the central dogma using genome engineered human cell lines|journal=Nucleic Acids Research|volume=48|issue=16|pages=9406–9413|doi=10.1093/nar/gkaa668|doi-access=free|pmid=32810265|pmc=7498316}}
{{DEFAULTSORT:Deming Regression}}
[[Category:Curve fitting]]
[[Category:Regression analysis]]
|