Latent dirichlet allocation
Citations
471 citations
Cites background or methods from "Latent dirichlet allocation"
...Block-LDA: Jointly modeling entity-annotated text and entity-entity links....
[...]
...LDA associates each document in a corpus d ....
[...]
...We use topic models [4] to discover topics from product reviews and other sources of text....
[...]
...Topic models are a fundamental building block of text modeling [3, 4, 5] and form the cornerstone of our model....
[...]
...As with LDA, we assign each word to a topic (an integer between 1 and K) randomly, with probability proportional to the likelihood of that topic occurring with that word....
[...]
469 citations
468 citations
Cites background or methods from "Latent dirichlet allocation"
..., Latent Dirichlet Allocation (LDA) [11] and Aspect and Sentiment Unification Model (ASUM) [30] (adopted in [22]) in our experiments....
[...]
...We vary the number of topics (denoted as K) and choose the appropriate K values according to (i) the perplexity scores [11] on 20% held-out data (should be small); and (ii) the results themselves (should be reasonable)....
[...]
...First, it cannot discover app-specific topics by using Latent Dirichlet Allocation (LDA) [11], since it links all the user reviews from the same app together as a document....
[...]
464 citations
Cites methods from "Latent dirichlet allocation"
...These include, for example, the probabilistic latent semantic indexing (Hofmann 1999) or latent Dirichlet allocation models (Blei et al. 2003) along with algorithms such as singular value decomposition or nonnegative matrix factorization (Blei 2011)....
[...]
459 citations
Cites background or methods from "Latent dirichlet allocation"
..., PLSA [10] and LDA [6]), GSDMM can also obtain the representative words of each cluster....
[...]
...As a result, GSDMM can obtain the representative words of each cluster like Topic Models (e.g., PLSA [10] and LDA [6])....
[...]
...We find that GSDMM has the following nice properties: 1) GSDMM can infer the number of clusters automatically; 2) GSDMM has a clear way to balance the completeness and homogeneity of the clustering results; 3) GSDMM is fast to converge; 4) Unlike the Vector Space Model (VSM)-based approaches, GSDMM can cope with the sparse and highdimensional problem of short texts; 5) Like Topic Models (e.g., PLSA [10] and LDA [6]), GSDMM can also obtain the representative words of each cluster....
[...]
...They compared DMAFP with other four clustering models: EM-DMM [20], K-means [13], LDA [6], and EDCM [7]....
[...]
References
17,608 citations
16,079 citations
"Latent dirichlet allocation" refers background in this paper
...Finally, Griffiths and Steyvers (2002) have presented a Markov chain Monte Carlo algorithm for LDA....
[...]
...Structures similar to that shown in Figure 1 are often studied in Bayesian statistical modeling, where they are referred to ashierarchical models(Gelman et al., 1995), or more precisely asconditionally independent hierarchical models(Kass and Steffey, 1989)....
[...]
...Structures similar to that shown in Figure 1 are often studied in Bayesian statistical modeling, where they are referred to as hierarchical models (Gelman et al., 1995), or more precisely as conditionally independent hierarchical models (Kass and Steffey, 1989)....
[...]
12,443 citations
"Latent dirichlet allocation" refers methods in this paper
...To address these shortcomings, IR researchers have proposed several other dimensionality reduction techniques, most notably latent semantic indexing (LSI) (Deerwester et al., 1990)....
[...]
...To address these shortcomings, IR researchers have proposed several other dimensionality reduction techniques, most notablylatent semantic indexing (LSI)(Deerwester et al., 1990)....
[...]
12,059 citations
"Latent dirichlet allocation" refers background or methods in this paper
...In the populartf-idf scheme (Salton and McGill, 1983), a basic vocabulary of “words” or “terms” is chosen, and, for each document in the corpus, a count is formed of the number of occurrences of each word....
[...]
...We report results in document modeling, text classification, and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSI model....
[...]
7,086 citations