scispace - formally typeset
Open AccessPosted Content

Neural Entity Linking: A Survey of Models Based on Deep Learning

TLDR
This work distills a generic architecture of a neural EL system and discusses its components, such as candidate generation, mention-context encoding, and entity ranking, summarizing prominent methods for each of them.
Abstract
In this survey, we provide a comprehensive description of recent neural entity linking (EL) systems developed since 2015 as a result of the "deep learning revolution" in NLP. Our goal is to systemize design features of neural entity linking systems and compare their performance to the prominent classic methods on common benchmarks. We distill generic architectural components of a neural EL system, like candidate generation and entity ranking, and summarize prominent methods for each of them. The vast variety of modifications of this general neural entity linking architecture are grouped by several common themes: joint entity recognition and linking, models for global linking, domain-independent techniques including zero-shot and distant supervision methods, and cross-lingual approaches. Since many neural models take advantage of entity and mention/context embeddings to catch semantic meaning of them, we provide an overview of popular embedding techniques. Finally, we briefly discuss applications of entity linking, focusing on the recently emerged use-case of enhancing deep pre-trained masked language models based on the transformer architecture.

read more

Citations
More filters
Posted Content

Machine Knowledge: Creation and Curation of Comprehensive Knowledge Bases

TL;DR: In this article, the authors survey fundamental concepts and practical methods for creating and curating large-scale knowledge bases, including methods for discovering and canonicalizing entities and their semantic types and organizing them into clean taxonomies.
Proceedings ArticleDOI

Extended Overview of CLEF HIPE 2020: Named Entity Processing on Historical Newspapers

TL;DR: This paper presents an extended overview of the first edition of HIPE (Identifying Historical People, Places and other Entities), a pioneering shared task dedicated to the evaluation of named entity processing on historical newspapers in French, German and English.
Journal ArticleDOI

Reddit entity linking dataset

TL;DR: An entity linking dataset from Reddit is introduced that contains 17,316 linked entities, each annotated by three human annotators and then grouped into Gold, Silver, and Bronze to indicate inter-annotator agreement, which indicates the need for better entity linking models that can be applied to the enormous amount of social media text.
Journal ArticleDOI

Medical concept normalization in French using multilingual terminologies and contextual embeddings

TL;DR: In this paper, a system for concept normalization in French is presented, which takes advantage of the multilingual nature of available terminologies and embedding models to improve concept normalisation in French without translation nor direct supervision.
References
More filters
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings Article

Attention is All you Need

TL;DR: This paper proposed a simple network architecture based solely on an attention mechanism, dispensing with recurrence and convolutions entirely and achieved state-of-the-art performance on English-to-French translation.
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Proceedings Article

Distributed Representations of Words and Phrases and their Compositionality

TL;DR: This paper presents a simple method for finding phrases in text, and shows that learning good vector representations for millions of phrases is possible and describes a simple alternative to the hierarchical softmax called negative sampling.
Related Papers (5)