Who developed the statistical technique known as regression?
Answers
The earliest form of regression was the method of least squares, which was published by Legendre in 1805,[7] and by Gauss in 1809.[8] Legendre and Gauss both applied the method to the problem of determining, from astronomical observations, the orbits of bodies about the Sun (mostly comets, but also later the then newly discovered minor planets). Gauss published a further development of the theory of least squares in 1821,[9] including a version of the Gauss–Markov theorem.
The term "regression" was coined by Francis Galton in the nineteenth century to describe a biological phenomenon. The phenomenon was that the heights of descendants of tall ancestors tend to regress down towards a normal average (a phenomenon also known as regression toward the mean).[10][11] For Galton, regression had only this biological meaning,[12][13] but his work was later extended by Udny Yule and Karl Pearson to a more general statistical context.[14][15] In the work of Yule and Pearson, the joint distribution of the response and explanatory variables is assumed to be Gaussian. This assumption was weakened by R.A. Fisher in his works of 1922 and 1925.[16][17][18] Fisher assumed that the conditional distribution of the response variable is Gaussian, but the joint distribution need not be. In this respect, Fisher's assumption is closer to Gauss's formulation of 1821.
In the 1950s and 1960s, economists used electromechanical desk calculators to calculate regressions. Before 1970, it sometimes took up to 24 hours to receive the result from one regression.[19]
Regression methods continue to be an area of active research. In recent decades, new methods have been developed for robust regression, regression involving correlated responses such as time series and growth curves, regression in which the predictor (independent variable) or response variables are curves, images, graphs, or other complex data objects, regression methods accommodating various types of missing data, nonparametric regression, Bayesian methods for regression, regression in which the predictor variables are measured with error, regression with more predictor variables than observations, and causal inference with regression.
The word "regression" was invented by Francis Galton in the 19th century to define a biological event.
It incorporates many methods for modeling and investigating various variables when the center is on the association within a second variable and one or more autonomous variables. More particularly, regression study supports one to understand how the average value of the second variable or criterion variable shifts when any one of the autonomous variables is different, while the other autonomous variables are fixed.