Journal ArticleDOI
Named Entity Recognition by Using XLNet-BiLSTM-CRF
Rongen Yan,Xue Jiang,Depeng Dang +2 more
Reads0
Chats0
TLDR
A new neural network model is proposed to improve the effectiveness of the NER by using a pre-trained XLNet, bi-directional long-short term memory (Bi-LSTM) and conditional random field (CRF) and the superiority of XLNet in NER tasks is demonstrated.Abstract:
Named entity recognition (NER) is the basis for many natural language processing (NLP) tasks such as information extraction and question answering. The accuracy of the NER directly affects the results of downstream tasks. Most of the relevant methods are implemented using neural networks, however, the word vectors obtained from a small data set cannot describe unusual, previously-unseen entities accurately and the results are not sufficiently accurate. Recently, the use of XLNet as a new pre-trained model has yielded satisfactory results in many NLP tasks, integration of XLNet embeddings in existent NLP tasks is not straightforward. In this paper, a new neural network model is proposed to improve the effectiveness of the NER by using a pre-trained XLNet, bi-directional long-short term memory (Bi-LSTM) and conditional random field (CRF). Pre-trained XLNet model is used to extract sentence features, then the classic NER neural network model is combined with the obtained features. In addition, the superiority of XLNet in NER tasks is demonstrated. We evaluate our model on the CoNLL-2003 English dataset and WNUT-2017 and show that the XLNet-BiLSTM-CRF obtains state-of-the-art results.read more
Citations
More filters
Journal ArticleDOI
ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope
TL;DR: A comprehensive review of the background, applications, key challenges, and future directions of ChatGPT can be found in this article , highlighting the importance of striking a balance between AI-assisted innovation and human expertise.
Journal ArticleDOI
Deep learning-based methods for natural hazard named entity recognition
TL;DR: In this article , a natural hazard named entity recognition method based on deep learning is proposed, namely XLNet-BiLSTM-CRF model, which can automatically mine text features and reduce the dependence on manual rules.
ChemNLP: A Natural Language Processing based Library for Materials Chemistry Text Data
Kamal Choudhary,Mathew L. Kelley +1 more
TL;DR: The ChemNLP library and an accompany-ing web-app that can be used to analyze important materials chemistry information are presented and the overlap between density functional theory and text-based databases for superconductors is determined.
Journal ArticleDOI
TFM: A Triple Fusion Module for Integrating Lexicon Information in Chinese Named Entity Recognition
Journal ArticleDOI
Transformer-Based Named Entity Recognition for French Using Adversarial Adaptation to Similar Domain Corpora
Arjun Choudhry,Pankaj Gupta,Inder Khatri,Aaryan Gupta,Maxime Nicol,Marie-Jean Meurs,Dinesh Kumar Vishwakarma +6 more
TL;DR: The authors proposed a transformer-based NER approach for French using adversarial adaptation to similar domain or general corpora for improved feature extraction and better generalization, which outperforms the corresponding non-adaptive models.
References
More filters
Proceedings ArticleDOI
Glove: Global Vectors for Word Representation
TL;DR: A new global logbilinear regression model that combines the advantages of the two major model families in the literature: global matrix factorization and local context window methods and produces a vector space with meaningful substructure.
Proceedings ArticleDOI
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Posted Content
Efficient Estimation of Word Representations in Vector Space
TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Proceedings Article
Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data
TL;DR: This work presents iterative parameter estimation algorithms for conditional random fields and compares the performance of the resulting models to HMMs and MEMMs on synthetic and natural-language data.
Related Papers (5)
Effect of Character and Word Features in Bidirectional LSTM-CRF for NER
Chirawan Ronran,Seungwoo Lee +1 more