scispace - formally typeset
Open AccessProceedings Article

Better Word Representations with Recursive Neural Networks for Morphology

Reads0
Chats0
TLDR
This paper combines recursive neural networks, where each morpheme is a basic unit, with neural language models to consider contextual information in learning morphologicallyaware word representations and proposes a novel model capable of building representations for morphologically complex words from their morphemes.
Abstract
Vector-space word representations have been very successful in recent years at improving performance across a variety of NLP tasks. However, common to most existing work, words are regarded as independent entities without any explicit relationship among morphologically related words being modeled. As a result, rare and complex words are often poorly estimated, and all unknown words are represented in a rather crude way using only one or a few vectors. This paper addresses this shortcoming by proposing a novel model that is capable of building representations for morphologically complex words from their morphemes. We combine recursive neural networks (RNNs), where each morpheme is a basic unit, with neural language models (NLMs) to consider contextual information in learning morphologicallyaware word representations. Our learned models outperform existing word representations by a good margin on word similarity tasks across many datasets, including a new dataset we introduce focused on rare words to complement existing ones in an interesting way.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Language with Vision: a Study on Grounded Word and Sentence Embeddings

TL;DR: A series of evaluations on word similarity benchmarks shows that visual grounding is beneficial not only for concrete words, but also for abstract words, as well as for contextualized embeddings trained on corpora of relatively modest size.
Journal ArticleDOI

Learning Fair Representations via Rate-Distortion Maximization

TL;DR: A novel debiasing technique, Fairness-aware Rate Maximization (FaRM), that removes protected information by making representations of instances belonging to the same protected attribute class uncorrelated, using the rate-distortion function.
Proceedings ArticleDOI

Text Classification Algorithm Based on TF-IDF and BERT

TL;DR: The results suggest that the using BERT model based method for technology information text auto-Categoriz has significantly improved accuracy, recall and fl_score, and has a good Chinese text classification effect.
Posted Content

Empirical Study of Diachronic Word Embeddings for Scarce Data.

TL;DR: The authors compare three models to learn diachronic word embeddings on scarce data: incremental updating of a Skip-gram from Kim et al. (2014), dynamic filtering from Bamler and Mandt (2017), and dynamic Bernoulli embedding from Rudolph and Blei (2018).
Book ChapterDOI

Learning Distributed Representations of Uyghur Words and Morphemes

TL;DR: An approach to learn distributed representations of Uyghur words and morphemes from unlabeled data is proposed and it is shown that this approach achieves significant improvements over CBOW, a state-of-the-art model for computing vector representations of words.
References
More filters
Journal ArticleDOI

WordNet: a lexical database for English

TL;DR: WordNet1 provides a more effective combination of traditional lexicographic information and modern computing, and is an online lexical database designed for use under program control.
Journal ArticleDOI

A neural probabilistic language model

TL;DR: The authors propose to learn a distributed representation for words which allows each training sentence to inform the model about an exponential number of semantically neighboring sentences, which can be expressed in terms of these representations.
Journal Article

Natural Language Processing (Almost) from Scratch

TL;DR: A unified neural network architecture and learning algorithm that can be applied to various natural language processing tasks including part-of-speech tagging, chunking, named entity recognition, and semantic role labeling is proposed.
Proceedings ArticleDOI

A unified architecture for natural language processing: deep neural networks with multitask learning

TL;DR: This work describes a single convolutional neural network architecture that, given a sentence, outputs a host of language processing predictions: part-of-speech tags, chunks, named entity tags, semantic roles, semantically similar words and the likelihood that the sentence makes sense using a language model.
Proceedings Article

Recurrent neural network based language model

TL;DR: Results indicate that it is possible to obtain around 50% reduction of perplexity by using mixture of several RNN LMs, compared to a state of the art backoff language model.
Related Papers (5)