scispace - formally typeset
Search or ask a question
Topic

Latent Dirichlet allocation

About: Latent Dirichlet allocation is a research topic. Over the lifetime, 5351 publications have been published within this topic receiving 212555 citations. The topic is also known as: LDA.


Papers
More filters
Journal ArticleDOI
01 Jan 1995-Genetica
TL;DR: The Dirichlet distribution provides a convenient conjugate prior for Bayesian analyses involving multinomial proportions and can be employed to model the contributions from different ancestral populations in computing forensic match probabilities.
Abstract: The Dirichlet distribution provides a convenient conjugate prior for Bayesian analyses involving multinomial proportions. In particular, allele frequency estimation can be carried out with a Dirichlet prior. If data from several distinct populations are available, then the parameters characterizing the Dirichlet prior can be estimated by maximum likelihood and then used for allele frequency estimation in each of the separate populations. This empirical Bayes procedure tends to moderate extreme multinomial estimates based on sample proportions. The Dirichlet distribution can also be employed to model the contributions from different ancestral populations in computing forensic match probabilities. If the ancestral populations are in genetic equilibrium, then the product rule for computing match probabilities is valid conditional on the ancestral contributions to a typical person of the reference population. This fact facilitates computation of match probabilities and tight upper bounds to match probabilities.

62 citations

Journal ArticleDOI
TL;DR: The experimental results reported for both synthetic data and real-world challenging applications involving image categorization, automatic semantic annotation and retrieval show the ability of the approach to provide accurate models by distinguishing between relevant and irrelevant features without over- or under-fitting the data.

62 citations

Journal ArticleDOI
TL;DR: In this article, a vector lattice ordering is used to represent textual entailment, inspired by a strengthened form of the distributional hypothesis, and a degree of entailment is defined in the form of a conditional probability.
Abstract: Formalizing “meaning as context” mathematically leads to a new, algebraic theory of meaning, in which composition is bilinear and associative These properties are shared by other methods that have been proposed in the literature, including the tensor product, vector addition, point-wise multiplication, and matrix multiplicationEntailment can be represented by a vector lattice ordering, inspired by a strengthened form of the distributional hypothesis, and a degree of entailment is defined in the form of a conditional probability Approaches to the task of recognizing textual entailment, including the use of subsequence matching, lexical entailment probability, and latent Dirichlet allocation, can be described within our framework

61 citations

Journal ArticleDOI
TL;DR: 10 topics were identified in which the main security issues are malware, cybersecurity attacks, data storing vulnerabilities, the use of testing software in IoT, and possible leaks due to the lack of user experience.

61 citations

Journal ArticleDOI
TL;DR: This study adopts the topic model approach, which automatically discovers topics that pervade a large and unstructured collection of documents, to uncover research topics in TIM research.
Abstract: The study of technology and innovation management (TIM) has continued to evolve and expand with great speed over the last three decades. This research aims to identify core topics in TIM studies and explore their dynamic changes. The conventional approach, based on discrete assignments by subjective judgment with predetermined categories, cannot effectively capture latent topics from large volumes of scholarly data. Hence, this study adopts the topic model approach, which automatically discovers topics that pervade a large and unstructured collection of documents, to uncover research topics in TIM research. The 50 topics of TIM research are identified through the Latent Dirichlet Allocation model from 11,693 articles published from 1997 to 2016 in 11 TIM journals, and top 10 most popular topics in TIM research are briefly reviewed. We then explore topic trends by examining the changes in topics rankings over different time periods and identifying hot and cold topics of TIM research over the last two decades. For each of the 11 TIM journals, the areas of subspecialty and the effects of editor changes on topic portfolios are also investigated. The findings of this study are expected to provide implications for researchers, journal editors, and policy makers in the field of TIM.

61 citations


Network Information
Related Topics (5)
Cluster analysis
146.5K papers, 2.9M citations
86% related
Support vector machine
73.6K papers, 1.7M citations
86% related
Deep learning
79.8K papers, 2.1M citations
85% related
Feature extraction
111.8K papers, 2.1M citations
84% related
Convolutional neural network
74.7K papers, 2M citations
83% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
2023323
2022842
2021418
2020429
2019473
2018446