Latent dirichlet allocation
Citations
176 citations
Cites background from "Latent dirichlet allocation"
...Standard topic models [2, 4] assume a document is generated from a mixture of topics, where a topic is a probabilistic distribution of words....
[...]
...As one of the most typical probabilistic topic model, LDA [2] has achieved great success in modeling text collections like news articles, research papers and blogs....
[...]
...Being one of the most classical topic models, LDA [2] can induce sparsity as its Dirichlet prior approaches zero....
[...]
176 citations
Cites background or methods or result from "Latent dirichlet allocation"
...where [i]1 1 , [j] k2 1 , [u] n1 1 , [v] n2 1 , φ1ui is the i th component of φ1 for row u, φ2vj is the j th component of φ2 for column v, and similarly for γ1ui and γ2vj , and Ψ(·) is the digamma function [7]....
[...]
...As compared to the variational approximation used in BNB [5] and LDA [7], where the cluster assignment z for every single feature has a variational discrete distribution, in our approximation there is only one variational discrete distribution for an entire row/column....
[...]
...To estimate the Dirichlet parameters (α1, α2), one can use an efficient Newton update as shown in [7, 5] for LDA and BNB....
[...]
...Variational inference [7] and Gibbs sampling [11] are two most popular approaches proposed to address the problem....
[...]
...Recall that the perplexity [7] of a dataset X is defined as: perp(X) = exp(− log p(X)/N), where N is the number of non-missing entries....
[...]
176 citations
Cites background from "Latent dirichlet allocation"
...It can be thought of as the “infinite” topic version of latent Dirichlet allocation (LDA) [13]....
[...]
176 citations
176 citations
Cites background or methods from "Latent dirichlet allocation"
...To address this issue, Latent Dirichlet Allocation (LDA) (Blei et al., 2003) is then proposed....
[...]
...DATA AND EXPERIMENTAL SETTINGS The experimental settings in this work are basically the same as those in (Blei et al., 2003)....
[...]
...However, the number of parameters of PLSA grows linearly with the number of documents, which suggests that PLSA is prone to over.tting (Blei et al., 2003)....
[...]
...Two of the most well known topic modeling algorithms include Probabilistic Latent Semantic Analysis (PLSA) (Hofmann, 2001) and Latent Dirichlet Allocation (LDA) (Blei et al., 2003)....
[...]
...The linear growth in parameters suggests that the model is prone to over.tting (Blei et al., 2003)....
[...]
References
17,608 citations
16,079 citations
"Latent dirichlet allocation" refers background in this paper
...Finally, Griffiths and Steyvers (2002) have presented a Markov chain Monte Carlo algorithm for LDA....
[...]
...Structures similar to that shown in Figure 1 are often studied in Bayesian statistical modeling, where they are referred to ashierarchical models(Gelman et al., 1995), or more precisely asconditionally independent hierarchical models(Kass and Steffey, 1989)....
[...]
...Structures similar to that shown in Figure 1 are often studied in Bayesian statistical modeling, where they are referred to as hierarchical models (Gelman et al., 1995), or more precisely as conditionally independent hierarchical models (Kass and Steffey, 1989)....
[...]
12,443 citations
"Latent dirichlet allocation" refers methods in this paper
...To address these shortcomings, IR researchers have proposed several other dimensionality reduction techniques, most notably latent semantic indexing (LSI) (Deerwester et al., 1990)....
[...]
...To address these shortcomings, IR researchers have proposed several other dimensionality reduction techniques, most notablylatent semantic indexing (LSI)(Deerwester et al., 1990)....
[...]
12,059 citations
"Latent dirichlet allocation" refers background or methods in this paper
...In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabulary of “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of the number of occurrences of each word....
[...]
...We report results in document modeling, text classification, and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSI model....
[...]
7,086 citations