scispace - formally typeset
Open AccessPosted Content

FeTaQA: Free-form Table Question Answering.

Reads0
Chats0
TLDR
FeTaQA as discussed by the authors is a table question answering dataset with 10k Wikipedia-based pairs, where answers are human-generated explanations involving entities and their high-level relations.
Abstract
Existing table question answering datasets contain abundant factual questions that primarily evaluate the query and schema comprehension capability of a system, but they fail to include questions that require complex reasoning and integration of information due to the constraint of the associated short-form answers. To address these issues and to demonstrate the full challenge of table question answering, we introduce FeTaQA, a new dataset with 10K Wikipedia-based {table, question, free-form answer, supporting table cells} pairs. FeTaQA yields a more challenging table question answering setting because it requires generating free-form text answers after retrieval, inference, and integration of multiple discontinuous facts from a structured knowledge source. Unlike datasets of generative QA over text in which answers are prevalent with copies of short text spans from the source, answers in our dataset are human-generated explanations involving entities and their high-level relations. We provide two benchmark methods for the proposed task: a pipeline method based on semantic-parsing-based QA systems and an end-to-end method based on large pretrained text generation models, and show that FeTaQA poses a challenge for both methods.

read more

Citations
More filters
Posted Content

Data-QuestEval: A Referenceless Metric for Data to Text Semantic Evaluation

TL;DR: In this paper, the authors propose a method to build synthetic multimodal corpora for data-QuestEval, which is a reference-less metric used in text-to-text tasks, that compares the generated summaries directly to the source text by automatically asking and answering questions.
Posted Content

Turning Tables: Generating Examples from Semi-structured Tables for Endowing Language Models with Reasoning Skills

TL;DR: The authors propose to leverage semi-structured tables, and automatically generate at scale question-paragraph pairs, where answering the question requires reasoning over multiple facts in the paragraph, such as number comparison, conjunction, and fact composition.
Proceedings ArticleDOI

Turning Tables: Generating Examples from Semi-structured Tables for Endowing Language Models with Reasoning Skills

TL;DR: The authors propose to leverage semi-structured tables, and automatically generate at scale question-paragraph pairs, where answering the question requires reasoning over multiple facts in the paragraph and add a pre-training step over this synthetic data, which includes examples that require 16 different reasoning skills such as number comparison, conjunction and fact composition.
Proceedings Article

Data-QuestEval: A Referenceless Metric for Data-to-Text Semantic Evaluation

TL;DR: In this article, the authors propose a method to build synthetic multimodal corpora for data-QuestEval, which is a reference-less metric used in text-to-text tasks, that compares the generated summaries directly to the source text by automatically asking and answering questions.
References
More filters
Proceedings Article

ROUGE: A Package for Automatic Evaluation of Summaries

TL;DR: Four different RouGE measures are introduced: ROUGE-N, ROUge-L, R OUGE-W, and ROUAGE-S included in the Rouge summarization evaluation package and their evaluations.
Proceedings ArticleDOI

Transformers: State-of-the-Art Natural Language Processing

TL;DR: Transformers is an open-source library that consists of carefully engineered state-of-the art Transformer architectures under a unified API and a curated collection of pretrained models made by and available for the community.
Proceedings Article

METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments

TL;DR: METEOR is described, an automatic metric for machine translation evaluation that is based on a generalized concept of unigram matching between the machineproduced translation and human-produced reference translations and can be easily extended to include more advanced matching strategies.
Proceedings ArticleDOI

SQuAD: 100,000+ Questions for Machine Comprehension of Text

TL;DR: The Stanford Question Answering Dataset (SQuAD) as mentioned in this paper is a reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corresponding reading passage.
Journal Article

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

TL;DR: This article introduced a unified framework that converts all text-based language problems into a text-to-text format and compared pre-training objectives, architectures, unlabeled data sets, transfer approaches, and other factors on dozens of language understanding tasks.
Related Papers (5)