How Do OLS and LAD Regression Methods Differ?

In summary: However, least squares remains the most common choice and is still the estimation technique of choice for linear models with continuous responses.
  • #1
oleandora
2
0
Hi
I'mwondering what's the difference between least squares method with least absolute deviation method.
Assume we have y=ax+b+s where s isdeviation.
Is the step to calculate even a and b is different.
I read that those two methods are almost the same but hardly found a real good explanation about LAD.
Thank you
 
Physics news on Phys.org
  • #2
oleandora said:
I'mwondering what's the difference between least squares method with least absolute deviation method.
Assume we have y=ax+b+s where s isdeviation.
Is the step to calculate even a and b is different.
I read that those two methods are almost the same but hardly found a real good explanation about LAD.

I assume that you are referring specifically to linear regressions. The difference between least squares and least absolute deviation is what is being optimized when the line is fit to the data. Yes, the mechanics of the LS and LAD (also called "L-1") fitting procedures are quite different.

While regression procedures which optimize different error functions sometimes produce similar results on a given set of data, they can also yield substantially different results. You can see an example of this in my posting, http://matlabdatamining.blogspot.com/2007/10/l-1-linear-regression.html" .


-Will Dwinnell
http://matlabdatamining.blogspot.com/"
 
Last edited by a moderator:
  • #3
The two methods are quite different in concept. In least squares the estimates are obtained by minimizing the sum of squared differences between the data and the fit (also described as minimizing the sum of the squares of the residuals)

If we call the estimates [tex] \widehat{\alpha} [/tex] and [tex] \widehat{\beta} [/tex], then for least squares

[tex]
S(\widehat{\alpha}, \widehat{\beta}) = \min_{(\alpha,\beta) \in R^2} \sum (y-(\alpha + \beta x))^2
[/tex]

while for L1

[tex]
S(\widehat{\alpha}, \widehat{\beta}) = \min_{(\alpha, \beta) \in R^2} \sum |y - (\alpha + \beta) }
[/tex]

Two benefits, not the most important, of least squares:
- the underlying calculations are easier to show with pencil and paper (than they are for L1)
- it is possible to write down formulas for the two estimates obtained from least squares - it isn't for L1

Estimates from both methods have asymptotic normal distributions under fairly general conditions.

The least squares estimates are the classical estimates when normality of the error distribution is assumed: they have certain optimality properties in that case, and, if you are interested in looking only at certain types of estimates, they are BLUE (Best Linear Unbiased Estimates) of the underlying parameters.

Least squares is so widely used because people are familiar with it. Its biggest downside is that fits from least squares are incredibly non-robust (sensitive to outliers and leverage points). L1 fits also suffer from this, but not quite as seriously as least squares.

Regression based on ranks, as well as regression based on Huber's M-estimates, are more robust and, with the ongoing combination of computing power increase and lower cost, are
ever-more reasonable alternatives.
 

Related to How Do OLS and LAD Regression Methods Differ?

What is the difference between OLS and LAD?

OLS (Ordinary Least Squares) and LAD (Least Absolute Deviations) are two commonly used methods for estimating the parameters of a linear regression model. The main difference between them lies in the way they handle the errors or residuals in the model.

Which method is more robust to outliers?

LAD is considered to be more robust to outliers compared to OLS. This is because LAD minimizes the sum of the absolute residuals, while OLS minimizes the sum of squared residuals. The absolute value function used in LAD gives less weight to extreme values, making it less sensitive to outliers.

How do OLS and LAD differ in terms of assumptions?

OLS has strict assumptions about the distribution of errors, such as normality and homoscedasticity, while LAD has fewer assumptions and is more robust to violations of these assumptions. LAD only assumes that the errors have a symmetric distribution.

Which method is more efficient?

OLS is considered to be more efficient than LAD in terms of producing estimates with lower variance. This is because OLS takes into account all the data points in the model, while LAD only considers the ranks of the data points. However, if the assumptions of OLS are violated, LAD may produce more accurate estimates.

When should one use OLS over LAD?

OLS is commonly used when the assumptions of normality and homoscedasticity are met, and the data does not have significant outliers. It also provides more precise estimates when the data is normally distributed. LAD, on the other hand, can be used when the data has outliers or is not normally distributed, as it is more robust to these issues.

Similar threads

  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
649
  • Set Theory, Logic, Probability, Statistics
Replies
3
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
6
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
23
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
654
  • Set Theory, Logic, Probability, Statistics
Replies
4
Views
1K
  • Set Theory, Logic, Probability, Statistics
Replies
8
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
5
Views
1K
Replies
1
Views
2K
  • Set Theory, Logic, Probability, Statistics
Replies
7
Views
1K
Back
Top