I have read lda papers for a while, i'm so much confused right now. LDA is used for topic model, but these topic reprented by words can't be used directly, they are ususlly meaning-less words. So, LDA is nothing different from other document clustering methods, but with more complex model and inference, maybe, more interpretable, then , why is lda so hot?? what's its benefit?

asked Mar 03 '11 at 04:00

Fischer%20Yu's gravatar image

Fischer Yu
16223


2 Answers:

LDA works pretty good when you compare it to other document clustering methods.

It is a pretty decent generative process overall, and in his original paper, Blei clearly explained differences and advantages over different models.

It is far from perfect, and that is the reasion there are so many extension concerning its inference and the overall structure of the model.

It is a great way to model topics not only in documents, but it can also be applied to a whole different set of different problems.

answered Mar 03 '11 at 05:39

Leon%20Palafox's gravatar image

Leon Palafox ♦
40857194128

LDA is not exactly a document clustering method, it's easier to think of it as a discrete matrix decomposition method, something like PCA or LSA but regularized and dealing directly with probabilities that works well for discrete data. Some big uses of LDA, I think, are:

  1. browsing: if you have a huge document collection it can get a lot easier to navigate using topics than it would be using words. There are far less topics than words, and yet the topics retain a lot of the meaning of the documents, usually, and it ends up being very intuitive to alternate looking at the more representative documents of a topic then the more representative topics for a document.

  2. dimensionality reduction: just like PCA and LSA, LDA can give you a lower-dimensional representation for the documents in your corpus, and this can be useful for many things. For example, if you're learning a classifier but have very little training data you can use LDA to learn features for your classifier, and if the classes happen close to the topic lines (things like subject, genre, etc, are usually better captured by topics than things like sentiment) you will probably see an improvement when using topic features as well as word features.

  3. as a part of more complex models: This is, I think, where a lot of the attention to LDA comes from. LDA is a rather simple building block that makes sense to use in all sorts of discrete bayesian models. For example, summarization can be done better with the help of topics, bilingual word alignment, as a part of deeper analysis of history of ideas, exploring the impact of papers, the votes of congressmen, and many other things. Topic models are interesting for these applications because they can help you do a focused dimensionality reduction, that simplifies the rest of the model reducing the number of parameters you need to deal with.

answered Mar 03 '11 at 05:42

Alexandre%20Passos's gravatar image

Alexandre Passos ♦
2554154278421

3ks a lot! besides your 3 points listed above, in blei's original paper, it seems that lda benefit a lot from its "generative", so what's the benefit of the generative or it's just used for inference?

(Mar 03 '11 at 09:20) Fischer Yu

Since it's generative (i.e. a joint density estimation problem), it can be trained easily on unlabeled data. Discriminative/conditional models typically need quite a lot of labeled data.

(Mar 03 '11 at 09:43) Oscar Täckström
Your answer
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.