scispace - formally typeset
Open AccessJournal ArticleDOI

A Comparative Analysis of Active Learning for Biomedical Text Mining

Reads0
Chats0
TLDR
Experiments show that AL has the potential to significantly reducing the cost of manual labelling, and AL-assisted pre-annotations accelerates the de novo annotation process with less annotation time required.
Abstract: 
An enormous amount of clinical free-text information, such as pathology reports, progress reports, clinical notes and discharge summaries have been collected at hospitals and medical care clinics. These data provide an opportunity of developing many useful machine learning applications if the data could be transferred into a learn-able structure with appropriate labels for supervised learning. The annotation of this data has to be performed by qualified clinical experts, hence, limiting the use of this data due to the high cost of annotation. An underutilised technique of machine learning that can label new data called active learning (AL) is a promising candidate to address the high cost of the label the data. AL has been successfully applied to labelling speech recognition and text classification, however, there is a lack of literature investigating its use for clinical purposes. We performed a comparative investigation of various AL techniques using ML and deep learning (DL)-based strategies on three unique biomedical datasets. We investigated random sampling (RS), least confidence (LC), informative diversity and density (IDD), margin and maximum representativeness-diversity (MRD) AL query strategies. Our experiments show that AL has the potential to significantly reducing the cost of manual labelling. Furthermore, pre-labelling performed using AL expediates the labelling process by reducing the time required for labelling.

read more

Citations
More filters
Proceedings Article

Interactive Span Recommendation for Biomedical Text

TL;DR: In this article , a few-shot span recommendation task is addressed by leveraging the Unified Medical Language Systems (UMLS) ontology, and a supervised algorithm is proposed to identify spans that are similar to a given span of interest.
Journal ArticleDOI

Conditional random field-recurrent neural network segmentation with optimized deep learning for brain tumour classification using magnetic resonance imaging

TL;DR: In this paper , a deep learning-based approach for segmenting and classifying brain tumours from the brain MRI was proposed, whose weights were adapted with the help of the Chronological Artificial Hummingbird Algorithm (CAHA).
Proceedings ArticleDOI

A Comparative Analysis of word embedding techniques and text similarity Measures

TL;DR: In this article , a comparative analysis of word embedding techniques and text similarity measures is presented to determine how similar two bits of text are in respective lexical, semantic characteristics, and closeness.
Journal ArticleDOI

Early stage autism detection using ANFIS and extreme learning machine algorithm

TL;DR: In this article , the authors used the data mining technique in the process of autism detection, which provides multiple beneficial impacts with high accuracy as it identifies the essential genes and gene sequences in a gene expression microarray dataset.
Proceedings Article

Reducing Knowledge Noise for Improved Semantic Analysis in Biomedical Natural Language Processing Applications

TL;DR: In this paper , a knowledge graph-based language model was combined with graph neural networks (GNNs) to leverage feature information from neighboring nodes in the graph, which showed significant performance improvements on relation extraction (RE) and classification tasks.
References
More filters
Book

The Nature of Statistical Learning Theory

TL;DR: Setting of the learning problem consistency of learning processes bounds on the rate of convergence ofLearning processes controlling the generalization ability of learning process constructing learning algorithms what is important in learning theory?
Posted Content

Efficient Estimation of Word Representations in Vector Space

TL;DR: This paper proposed two novel model architectures for computing continuous vector representations of words from very large data sets, and the quality of these representations is measured in a word similarity task and the results are compared to the previously best performing techniques based on different types of neural networks.
Posted Content

Neural Machine Translation by Jointly Learning to Align and Translate

TL;DR: In this paper, the authors propose to use a soft-searching model to find the parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly.
Proceedings Article

Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data

TL;DR: This work presents iterative parameter estimation algorithms for conditional random fields and compares the performance of the resulting models to HMMs and MEMMs on synthetic and natural-language data.
Journal ArticleDOI

Representation Learning: A Review and New Perspectives

TL;DR: Recent work in the area of unsupervised feature learning and deep learning is reviewed, covering advances in probabilistic models, autoencoders, manifold learning, and deep networks.
Related Papers (5)