scispace - formally typeset
Book ChapterDOI

Improving Short Answer Grading Using Transformer-Based Pre-training.

TLDR
This work experiments with fine-tuning a pre-trained self-attention language model, namely Bidirectional Encoder Representations from Transformers (BERT) applying it to short answer grading, and shows that it produces superior results across multiple domains.
Abstract
Dialogue-based tutoring platforms have shown great promise in helping individual students improve mastery. Short answer grading is a crucial component of such platforms. However, generative short answer grading using the same platform for diverse disciplines and titles is a crucial challenge due to data distribution variations across domains and a frequent occurrence of non-sentential answers. Recent NLP research has introduced novel deep learning architectures such as the Transformer, which merely uses self-attention mechanisms. Pre-trained models based on the Transformer architecture have been used to produce impressive results across a range of NLP tasks. In this work, we experiment with fine-tuning a pre-trained self-attention language model, namely Bidirectional Encoder Representations from Transformers (BERT) applying it to short answer grading, and show that it produces superior results across multiple domains. On the benchmarking dataset of SemEval-2013, we report up to 10% absolute improvement in macro-average-F1 over state-of-the-art results. On our two psychology domain datasets, the fine-tuned model yields classification almost up to the human-agreement levels. Moreover, we study the effectiveness of fine-tuning as a function of the size of the task-specific labeled data, the number of training epochs, and its generalizability to cross-domain and join-domain scenarios.

read more

Citations
More filters
Proceedings ArticleDOI

Pre-Training BERT on Domain Resources for Short Answer Grading.

TL;DR: It is shown that the pre-trained BERT model can be improved by augmenting data from the domain-specific resources like textbooks like textbooks, and a new approach to use labeled short answering grading data for further enhancement of the language model.
Book ChapterDOI

Investigating Transformers for Automatic Short Answer Grading

TL;DR: This work trains the newest and most powerful, according to the glue benchmark, transformers on the SemEval-2013 dataset, and shows that models trained with knowledge distillation are feasible for use in short answer grading.
Proceedings ArticleDOI

Neural Automated Essay Scoring Incorporating Handcrafted Features

TL;DR: This method concatenates handcrafted essay-level features to a distributed essay representation vector, which is obtained from an intermediate layer of a DNN-AES model, which significantly improves scoring accuracy.
Journal ArticleDOI

A review of deep-neural automated essay scoring models

TL;DR: A comprehensive survey of deep neural network AES models is presented, describing the main idea and detailed architecture of each model and introducing existing DNN-AES models according to this classification.
Book ChapterDOI

Robust Neural Automated Essay Scoring Using Item Response Theory

TL;DR: A new DNN-AES framework that integrates IRT models to deal with rater bias within training data is proposed, a first attempt at addressing rating bias effects in training data, which is a crucial but overlooked problem.
References
More filters
Proceedings ArticleDOI

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

TL;DR: BERT as mentioned in this paper pre-trains deep bidirectional representations from unlabeled text by jointly conditioning on both left and right context in all layers, which can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks.
Journal ArticleDOI

A Survey on Transfer Learning

TL;DR: The relationship between transfer learning and other related machine learning techniques such as domain adaptation, multitask learning and sample selection bias, as well as covariate shift are discussed.
Posted Content

Attention Is All You Need

TL;DR: A new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely is proposed, which generalizes well to other tasks by applying it successfully to English constituency parsing both with large and limited training data.
Posted Content

Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

TL;DR: GNMT, Google's Neural Machine Translation system, is presented, which attempts to address many of the weaknesses of conventional phrase-based translation systems and provides a good balance between the flexibility of "character"-delimited models and the efficiency of "word"-delicited models.
Proceedings ArticleDOI

Self-taught learning: transfer learning from unlabeled data

TL;DR: An approach to self-taught learning that uses sparse coding to construct higher-level features using the unlabeled data to form a succinct input representation and significantly improve classification performance.
Related Papers (5)