scispace - formally typeset
Open AccessJournal ArticleDOI

Self-Taught convolutional neural networks for short text clustering.

TLDR
A flexible Self-Taught Convolutional neural network framework for Short Text Clustering, which can flexibly and successfully incorporate more useful semantic features and learn non-biased deep text representation in an unsupervised manner is proposed.
About
This article is published in Neural Networks.The article was published on 2017-04-01 and is currently open access. It has received 166 citations till now. The article focuses on the topics: Cluster analysis & Convolutional neural network.

read more

Citations
More filters
Journal ArticleDOI

Unsupervised text feature selection technique based on hybrid particle swarm optimization algorithm with genetic operators for the text clustering

TL;DR: The results show that the proposed algorithm hybrid algorithm (H-FSPSOTC) improved the performance of the clustering algorithm by generating a new subset of more informative features, and is compared with the other comparative algorithms published in the literature.
Proceedings ArticleDOI

Supporting Clustering with Contrastive Learning

TL;DR: This work proposes Supporting Clustering with Contrastive Learning (SCCL) – a novel framework to leverage contrastive learning to promote better separation in distance-based clustering and demonstrates the effectiveness of SCCL in leveraging the strengths of both bottom-up instance discrimination and top-down clustering to achieve better intra-clusters and inter-cluster distances.
Journal ArticleDOI

Hybridizing Gray Wolf Optimization (GWO) with Grasshopper Optimization Algorithm (GOA) for text feature selection and clustering

TL;DR: The functional selection technique has two aims: maximize text clustering algorithm reliability and minimize the number of uninformative traits and the proposed technique produces a mature convergence rate and requires minimal computational time and is trapped in local minima in a low dimensional space.
Proceedings ArticleDOI

A Self-Training Approach for Short Text Clustering

TL;DR: The method is proposed, which learns discriminative features from both an autoencoder and a sentence embedding, then uses assignments from a clustering algorithm as supervision to update weights of the encoder network.
Book ChapterDOI

Learning Convolutional Ranking-Score Function by Query Preference Regularization

TL;DR: A new ranking scoring function based on the convolutional neural network (CNN) is proposed, which has a structure of CNN, and its parameters are adjusted to both queries and query preferences.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Journal Article

Visualizing Data using t-SNE

TL;DR: A new technique called t-SNE that visualizes high-dimensional data by giving each datapoint a location in a two or three-dimensional map, a variation of Stochastic Neighbor Embedding that is much easier to optimize, and produces significantly better visualizations by reducing the tendency to crowd points together in the center of the map.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Related Papers (5)