scispace - formally typeset
Journal ArticleDOI

Named Entity Recognition by Using XLNet-BiLSTM-CRF

Reads0
Chats0
TLDR
A new neural network model is proposed to improve the effectiveness of the NER by using a pre-trained XLNet, bi-directional long-short term memory (Bi-LSTM) and conditional random field (CRF) and the superiority of XLNet in NER tasks is demonstrated.
Abstract
Named entity recognition (NER) is the basis for many natural language processing (NLP) tasks such as information extraction and question answering. The accuracy of the NER directly affects the results of downstream tasks. Most of the relevant methods are implemented using neural networks, however, the word vectors obtained from a small data set cannot describe unusual, previously-unseen entities accurately and the results are not sufficiently accurate. Recently, the use of XLNet as a new pre-trained model has yielded satisfactory results in many NLP tasks, integration of XLNet embeddings in existent NLP tasks is not straightforward. In this paper, a new neural network model is proposed to improve the effectiveness of the NER by using a pre-trained XLNet, bi-directional long-short term memory (Bi-LSTM) and conditional random field (CRF). Pre-trained XLNet model is used to extract sentence features, then the classic NER neural network model is combined with the obtained features. In addition, the superiority of XLNet in NER tasks is demonstrated. We evaluate our model on the CoNLL-2003 English dataset and WNUT-2017 and show that the XLNet-BiLSTM-CRF obtains state-of-the-art results.

read more

Citations
More filters
Journal ArticleDOI

ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope

TL;DR: A comprehensive review of the background, applications, key challenges, and future directions of ChatGPT can be found in this article , highlighting the importance of striking a balance between AI-assisted innovation and human expertise.
Journal ArticleDOI

Deep learning-based methods for natural hazard named entity recognition

TL;DR: In this article , a natural hazard named entity recognition method based on deep learning is proposed, namely XLNet-BiLSTM-CRF model, which can automatically mine text features and reduce the dependence on manual rules.

ChemNLP: A Natural Language Processing based Library for Materials Chemistry Text Data

TL;DR: The ChemNLP library and an accompany-ing web-app that can be used to analyze important materials chemistry information are presented and the overlap between density functional theory and text-based databases for superconductors is determined.
Journal ArticleDOI

Transformer-Based Named Entity Recognition for French Using Adversarial Adaptation to Similar Domain Corpora

TL;DR: The authors proposed a transformer-based NER approach for French using adversarial adaptation to similar domain or general corpora for improved feature extraction and better generalization, which outperforms the corresponding non-adaptive models.
References
More filters
Proceedings ArticleDOI

Glove: Global Vectors for Word Representation

TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Proceedings Article

Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data

TL;DR: This work presents iterative parameter estimation algorithms for conditional random fields and compares the performance of the resulting models to HMMs and MEMMs on synthetic and natural-language data.
Related Papers (5)
Trending Questions (1)
What features are used for BILSTM named entity recognition?

The features used for BiLSTM named entity recognition include pre-trained XLNet embeddings and sentence features extracted from the XLNet model.