Can anyone give me some reasons as to why you would use least squares instead of gradient descent? Whenever I've come across one being used instead of the other its primarily for compute purposes. But is there any advantage / disadvantage one would have over the other?

asked Sep 30 '12 at 05:58

John%20Brown's gravatar image

John Brown
1111

edited Sep 30 '12 at 07:26

Lucian%20Sasu's gravatar image

Lucian Sasu
513172634

1

The least square is the loss function which measures the goodness of fit for your model, compared to actual output values. The gradient descent is a technique to modify model's parameters in order to minimize the squared error. Thus, they are different subjects.

For different loss functions one can use, see http://slipguru.disi.unige.it/Downloads/publications/loss.pdf. For the gradient descent algorithm see the wikipedia article or the presentation Andrew Ng gives on the Machine learning course, on coursera.org. The answer at http://metaoptimize.com/qa/questions/11116/what-are-the-best-resourses-to-learn-optimization-for-machine-learning might also be of help for you.

(Sep 30 '12 at 07:24) Lucian Sasu
Be the first one to answer this question!
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.