Jump to content

Frisch–Waugh–Lovell theorem: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
m clear ref errors
Citation bot (talk | contribs)
Alter: url. URLs might have been anonymized. Add: isbn. | Use this bot. Report bugs. | #UCB_CommandLine
 
(39 intermediate revisions by 16 users not shown)
Line 1: Line 1:
{{Short description|Theorem in statistics and econometrics}}
In [[econometrics]], the '''Frisch–Waugh–Lovell (FWL) theorem''' is named after the econometricians [[Ragnar Frisch]], [[Frederick V. Waugh]], and [[Michael C. Lovell]].<ref>{{cite journal |first=Ragnar |last=Frisch |first2=Frederick V. |last2=Waugh |title=Partial Time Regressions as Compared with Individual Trends |journal=[[Econometrica]] |volume=1 |issue=4 |year=1933 |pages=387–401 |jstor=1907330 }}</ref><ref>{{cite journal |last=Lovell |first=M. |year=1963 |title=Seasonal Adjustment of Economic Time Series and Multiple Regression Analysis |journal=[[Journal of the American Statistical Association]] |volume=58 |issue=304 |pages=993–1010 |doi=10.1080/01621459.1963.10480682 }}</ref><ref>{{cite journal |last=Lovell |first=M. |year=2008 |title=A Simple Proof of the FWL Theorem |journal=[[Journal of Economic Education]] |volume=39 |issue=1 |pages=88–91 |doi=10.3200/JECE.39.1.88-91 }}</ref>
In [[econometrics]], the '''Frisch–Waugh–Lovell (FWL) theorem''' is named after the econometricians [[Ragnar Frisch]], [[Frederick V. Waugh]], and [[Michael C. Lovell]].<ref>{{cite journal |first1=Ragnar |last1=Frisch |first2=Frederick V. |last2=Waugh |title=Partial Time Regressions as Compared with Individual Trends |journal=[[Econometrica]] |volume=1 |issue=4 |year=1933 |pages=387–401 |doi=10.2307/1907330 |jstor=1907330 }}</ref><ref>{{cite journal |last=Lovell |first=M. |year=1963 |title=Seasonal Adjustment of Economic Time Series and Multiple Regression Analysis |journal=[[Journal of the American Statistical Association]] |volume=58 |issue=304 |pages=993–1010 |doi=10.1080/01621459.1963.10480682 }}</ref><ref>{{cite journal |last=Lovell |first=M. |year=2008 |title=A Simple Proof of the FWL Theorem |journal=[[Journal of Economic Education]] |volume=39 |issue=1 |pages=88–91 |doi=10.3200/JECE.39.1.88-91 |s2cid=154907484 }}</ref>


The Frisch–Waugh–Lovell theorem states that if the [[linear regression|regression]] we are concerned with is:
The Frisch–Waugh–Lovell theorem states that if the [[linear regression|regression]] we are concerned with is expressed in terms of two separate sets of predictor variables:


:<math> Y = X_1 \beta_1 + X_2 \beta_2 + u </math>
:<math> Y = X_1 \beta_1 + X_2 \beta_2 + u </math>


where <math>X_1</math> and <math>X_2</math> are <math>n \times k_1</math> and <math>n \times k_2</math> [[Matrix (mathematics)|matrices]] respectively and where <math> \beta_1 </math> and <math> \beta_2 </math> are [[conformable matrix|conformable]], then the estimate of <math> \beta_2 </math> will be the same as the estimate of it from a modified regression of the form:
where <math>X_1</math> and <math>X_2</math> are [[Matrix (mathematics)|matrices]], <math> \beta_1 </math> and <math> \beta_2 </math> are vectors (and <math>u</math> is the error term), then the estimate of <math> \beta_2 </math> will be the same as the estimate of it from a modified regression of the form:


:<math> M_{X_1} Y = M_{X_1} X_2 \beta_2 + M_{X_1} u, </math>
:<math> M_{X_1} Y = M_{X_1} X_2 \beta_2 + M_{X_1} u, </math>
Line 13: Line 14:
:<math> M_{X_1} = I - X_1(X_1^{\mathsf{T}}X_1)^{-1}X_1^{\mathsf{T}}, </math>
:<math> M_{X_1} = I - X_1(X_1^{\mathsf{T}}X_1)^{-1}X_1^{\mathsf{T}}, </math>


and this particular orthogonal projection matrix is known as the [[annihilator matrix]].<ref>{{cite book |first=Fumio |last=Hayashi |author-link=Fumio Hayashi |title=Econometrics |location=Princeton |publisher=Princeton University Press |year=2000 |isbn=0-691-01018-8 |pages=18–19 |url=https://books.google.com/books?id=QyIW8WUIyzcC&pg=PA18 }}</ref><ref>{{cite book |first=James |last=Davidson |title=Econometric Theory |location=Malden |publisher=Blackwell |year=2000 |isbn=0-631-21584-0
and this particular orthogonal projection matrix is known as the [[annihilator matrix|residual maker matrix or annihilator matrix]].<ref>{{cite book |first=Fumio |last=Hayashi |author-link=Fumio Hayashi |title=Econometrics |location=Princeton |publisher=Princeton University Press |year=2000 |isbn=0-691-01018-8 |pages=18–19 |url=https://books.google.com/books?id=QyIW8WUIyzcC&pg=PA18 }}</ref><ref>{{cite book |first=James |last=Davidson |title=Econometric Theory |location=Malden |publisher=Blackwell |year=2000 |isbn=0-631-21584-0
|page=7 |url=https://books.google.com/books?id=shWtvsFbxlkC&pg=PA7 }}</ref>
|page=7 |url=https://books.google.com/books?id=shWtvsFbxlkC&pg=PA7 }}</ref>


The vector <math display="inline"> M_{X_1} Y </math> is the vector of residuals from regression of <math display="inline"> Y </math> on the columns of <math display="inline"> X_1</math>.
The vector <math display="inline"> M_{X_1} Y </math> is the vector of residuals from regression of <math display="inline"> Y </math> on the columns of <math display="inline"> X_1</math>.


The most relevant consequence of the theorem is that the parameters in <math display="inline"> \beta_2 </math> do not apply to <math display="inline"> X_2 </math> but to <math display="inline"> M_{X_1} X_2 </math>, that is: the part of <math display="inline"> X_2 </math> uncorrelated with <math display="inline"> X_1 </math>. This is the basis for understanding the contribution of each single variable to a multivariate regression (see, for instance, Ch. 13 in <ref>{{cite book |first1=F. |last1=Mosteller |first2=J. W.| last2=Tukey |title=Data Analysis and Regression a Second Course in Statistics |publisher=Addison-Wesley |year=1977 }}</ref>).
The theorem implies that the secondary regression used for obtaining <math> M_{X_1}</math> is unnecessary: using projection matrices to make the explanatory variables orthogonal to each other will lead to the same results as running the regression with all non-orthogonal explanators included.

The theorem also implies that the secondary regression used for obtaining <math> M_{X_1}</math> is unnecessary when the predictor variables are uncorrelated: using projection matrices to make the explanatory variables orthogonal to each other will lead to the same results as running the regression with all non-orthogonal explanators included.

Moreover, the standard errors from the partial regression equal those from the full regression.<ref>{{Cite journal |last=Peng |first=Ding |date=2021 |title=The Frisch--Waugh--Lovell theorem for standard errors |url=https://www.sciencedirect.com/science/article/pii/S0167715220302480 |journal=Statistics and Probability Letters |volume=168 |pages=108945}}</ref>

== History ==
The origin of the theorem is uncertain, but it was well-established in the realm of linear regression before the Frisch and Waugh paper. [[George Udny Yule]]'s comprehensive analysis of partial regressions, published in 1907, included the theorem in section 9 on page 184.<ref name=":0">{{cite journal |last=Yule |first=George Udny |year=1907 |title=On the Theory of Correlation for any Number of Variables, Treated by a New System of Notation |url=https://royalsocietypublishing.org/doi/abs/10.1098/rspa.1907.0028 |journal=Proceedings of the Royal Society A |volume=79 |issue=529 |pages=182–193 |doi=10.1098/rspa.1907.0028|hdl=2027/coo.31924081088423 |hdl-access=free }}</ref> Yule emphasized the theorem's importance for understanding multiple and partial regression and correlation coefficients, as mentioned in section 10 of the same paper.<ref name=":0" />

By 1933, Yule's findings were generally recognized{{Weasel inline|date=June 2023}}, thanks in part to the detailed discussion of partial correlation and the introduction of his innovative notation in 1907.{{Cn|date=June 2023}} The theorem, later associated with Frisch, Waugh, and Lovell, was also included in chapter 10 of Yule's successful statistics textbook, first published in 1911. The book reached its tenth edition by 1932.<ref>{{cite book |last=Yule |first=George Udny |url=https://archive.org/embed/introductiontoth00yule_0 |title=An Introduction to the Theory of Statistics 10th edition |publisher=Charles Griffin &Co |year=1932 |location=London}}</ref>

In a 1931 paper co-authored with Mudgett, Frisch cited Yule's results.<ref name=":1">{{cite journal |last1=Frisch |first1=Ragnar |last2=Mudgett |first2=B. D. |year=1931 |title=Statistical Correlation and the Theory of Cluster Types |url=https://www.sv.uio.no/econ/om/tall-og-fakta/nobelprisvinnere/ragnar-frisch/published-scientific-work/rf-published-scientific-works/rf1931g.pdf |journal=Journal of the American Statistical Association |volume=21 |issue=176 |pages=375–392|doi=10.1080/01621459.1931.10502225 }}</ref> Yule's formulas for partial regressions were quoted and explicitly attributed to him in order to rectify a misquotation by another author.<ref name=":1" /> Although Yule was not explicitly mentioned in the 1933 paper by Frisch and Waugh, they utilized the notation for partial regression coefficients initially introduced by Yule in 1907, which was widely accepted by 1933{{Original research inline|date=June 2023}}.

In 1963, Lovell published a proof<ref>{{cite journal |last=Lovell |first=M. |year=1963 |title=Seasonal Adjustment of Economic Time Series and Multiple Regression Analysis |journal=[[Journal of the American Statistical Association]] |volume=58 |issue=304 |pages=993–1010 |doi=10.1080/01621459.1963.10480682}}</ref> considered more straightforward and intuitive. In recognition, people generally add his name to the theorem name.


==References==
==References==
Line 24: Line 38:


==Further reading==
==Further reading==
* {{cite book |first=Russell |last=Davidson |first2=James G. |last2=MacKinnon |title=Estimation and Inference in Econometrics |location=New York |publisher=Oxford University Press |year=1993 |isbn=0-19-506011-3 |pages=19–24 |url=https://books.google.com/books?id=Ot6DByCF6osC&pg=PA19|ref=none }}
* {{cite book |first1=Russell |last1=Davidson |first2=James G. |last2=MacKinnon |title=Estimation and Inference in Econometrics |location=New York |publisher=Oxford University Press |year=1993 |isbn=0-19-506011-3 |pages=19–24 |url=https://books.google.com/books?id=Ot6DByCF6osC&pg=PA19|ref=none }}
* {{cite book |first=Russell |last=Davidson |first2=James G. |last2=MacKinnon |title=Econometric Theory and Methods |url=https://archive.org/details/econometrictheor00davi |url-access=limited |location=New York |publisher=Oxford University Press |year=2004 |pages=[https://archive.org/details/econometrictheor00davi/page/n59 62]–75 |isbn=0-19-512372-7|ref=none }}
* {{cite book |first1=Russell |last1=Davidson |first2=James G. |last2=MacKinnon |title=Econometric Theory and Methods |url=https://archive.org/details/econometrictheor00davi |url-access=limited |location=New York |publisher=Oxford University Press |year=2004 |pages=[https://archive.org/details/econometrictheor00davi/page/n59 62]–75 |isbn=0-19-512372-7|ref=none }}
* {{cite book |first=Trevor |last=Hastie |author-link=Trevor Hastie |first2=Robert |last2=Tibshirani |author-link2=Robert Tibshirani |first3=Jerome |last3=Friedman |author-link3=Jerome H. Friedman |title=The Elements of Statistical Learning : Data Mining, Inference, and Prediction |location=New York |publisher=Springer |edition=2nd |year=2017 |isbn=978-0-387-84857-0 |chapter=Multiple Regression from Simple Univariate Regression |pages=52–55 |chapter-url=https://web.stanford.edu/~hastie/Papers/ESLII.pdf#page=71|ref=none }}
* {{cite book |first1=Trevor |last1=Hastie |author-link=Trevor Hastie |first2=Robert |last2=Tibshirani |author-link2=Robert Tibshirani |first3=Jerome |last3=Friedman |author-link3=Jerome H. Friedman |title=The Elements of Statistical Learning : Data Mining, Inference, and Prediction |location=New York |publisher=Springer |edition=2nd |year=2017 |isbn=978-0-387-84857-0 |chapter=Multiple Regression from Simple Univariate Regression |pages=52–55 |chapter-url=https://web.stanford.edu/~hastie/Papers/ESLII.pdf#page=71|ref=none }}
* {{cite book |last=Ruud |first=P. A. |year=2000 |title=An Introduction to Classical Econometric Theory |location=New York |publisher=Oxford University Press |isbn=0-19-511164-8 |pages=54–60 |url=https://books.google.com/books?id=PnVCEZOOFr0C&pg=PA54|ref=none }}
* {{cite book |last=Ruud |first=P. A. |year=2000 |title=An Introduction to Classical Econometric Theory |location=New York |publisher=Oxford University Press |isbn=0-19-511164-8 |pages=54–60 |url=https://books.google.com/books?id=PnVCEZOOFr0C&pg=PA54|ref=none }}
* {{cite book |first=John |last=Stachurski |title=A Primer in Econometric Theory |publisher=MIT Press |year=2016 |pages=311–314 |url=https://books.google.com/books?id=DrbrDAAAQBAJ&pg=PA311|ref=none }}
* {{cite book |first=John |last=Stachurski |title=A Primer in Econometric Theory |publisher=MIT Press |year=2016 |pages=311–314 |isbn=9780262337465 |url=https://books.google.com/books?id=DrbrDAAAQBAJ&pg=PA311|ref=none }}


{{Least squares and regression analysis}}
{{Least squares and regression analysis}}

Latest revision as of 17:38, 2 December 2023

In econometrics, the Frisch–Waugh–Lovell (FWL) theorem is named after the econometricians Ragnar Frisch, Frederick V. Waugh, and Michael C. Lovell.[1][2][3]

The Frisch–Waugh–Lovell theorem states that if the regression we are concerned with is expressed in terms of two separate sets of predictor variables:

where and are matrices, and are vectors (and is the error term), then the estimate of will be the same as the estimate of it from a modified regression of the form:

where projects onto the orthogonal complement of the image of the projection matrix . Equivalently, MX1 projects onto the orthogonal complement of the column space of X1. Specifically,

and this particular orthogonal projection matrix is known as the residual maker matrix or annihilator matrix.[4][5]

The vector is the vector of residuals from regression of on the columns of .

The most relevant consequence of the theorem is that the parameters in do not apply to but to , that is: the part of uncorrelated with . This is the basis for understanding the contribution of each single variable to a multivariate regression (see, for instance, Ch. 13 in [6]).

The theorem also implies that the secondary regression used for obtaining is unnecessary when the predictor variables are uncorrelated: using projection matrices to make the explanatory variables orthogonal to each other will lead to the same results as running the regression with all non-orthogonal explanators included.

Moreover, the standard errors from the partial regression equal those from the full regression.[7]

History

[edit]

The origin of the theorem is uncertain, but it was well-established in the realm of linear regression before the Frisch and Waugh paper. George Udny Yule's comprehensive analysis of partial regressions, published in 1907, included the theorem in section 9 on page 184.[8] Yule emphasized the theorem's importance for understanding multiple and partial regression and correlation coefficients, as mentioned in section 10 of the same paper.[8]

By 1933, Yule's findings were generally recognized[weasel words], thanks in part to the detailed discussion of partial correlation and the introduction of his innovative notation in 1907.[citation needed] The theorem, later associated with Frisch, Waugh, and Lovell, was also included in chapter 10 of Yule's successful statistics textbook, first published in 1911. The book reached its tenth edition by 1932.[9]

In a 1931 paper co-authored with Mudgett, Frisch cited Yule's results.[10] Yule's formulas for partial regressions were quoted and explicitly attributed to him in order to rectify a misquotation by another author.[10] Although Yule was not explicitly mentioned in the 1933 paper by Frisch and Waugh, they utilized the notation for partial regression coefficients initially introduced by Yule in 1907, which was widely accepted by 1933[original research?].

In 1963, Lovell published a proof[11] considered more straightforward and intuitive. In recognition, people generally add his name to the theorem name.

References

[edit]
  1. ^ Frisch, Ragnar; Waugh, Frederick V. (1933). "Partial Time Regressions as Compared with Individual Trends". Econometrica. 1 (4): 387–401. doi:10.2307/1907330. JSTOR 1907330.
  2. ^ Lovell, M. (1963). "Seasonal Adjustment of Economic Time Series and Multiple Regression Analysis". Journal of the American Statistical Association. 58 (304): 993–1010. doi:10.1080/01621459.1963.10480682.
  3. ^ Lovell, M. (2008). "A Simple Proof of the FWL Theorem". Journal of Economic Education. 39 (1): 88–91. doi:10.3200/JECE.39.1.88-91. S2CID 154907484.
  4. ^ Hayashi, Fumio (2000). Econometrics. Princeton: Princeton University Press. pp. 18–19. ISBN 0-691-01018-8.
  5. ^ Davidson, James (2000). Econometric Theory. Malden: Blackwell. p. 7. ISBN 0-631-21584-0.
  6. ^ Mosteller, F.; Tukey, J. W. (1977). Data Analysis and Regression a Second Course in Statistics. Addison-Wesley.
  7. ^ Peng, Ding (2021). "The Frisch--Waugh--Lovell theorem for standard errors". Statistics and Probability Letters. 168: 108945.
  8. ^ a b Yule, George Udny (1907). "On the Theory of Correlation for any Number of Variables, Treated by a New System of Notation". Proceedings of the Royal Society A. 79 (529): 182–193. doi:10.1098/rspa.1907.0028. hdl:2027/coo.31924081088423.
  9. ^ Yule, George Udny (1932). An Introduction to the Theory of Statistics 10th edition. London: Charles Griffin &Co.
  10. ^ a b Frisch, Ragnar; Mudgett, B. D. (1931). "Statistical Correlation and the Theory of Cluster Types" (PDF). Journal of the American Statistical Association. 21 (176): 375–392. doi:10.1080/01621459.1931.10502225.
  11. ^ Lovell, M. (1963). "Seasonal Adjustment of Economic Time Series and Multiple Regression Analysis". Journal of the American Statistical Association. 58 (304): 993–1010. doi:10.1080/01621459.1963.10480682.

Further reading

[edit]