scispace - formally typeset
K

Kathleen R. McKeown

Researcher at Columbia University

Publications -  383
Citations -  20869

Kathleen R. McKeown is an academic researcher from Columbia University. The author has contributed to research in topics: Automatic summarization & Natural language. The author has an hindex of 67, co-authored 355 publications receiving 19242 citations. Previous affiliations of Kathleen R. McKeown include New York University & Amazon.com.

Papers
More filters
Journal ArticleDOI

Learning methods to combine linguistic indicators: improving aspectual classification and revealing linguistic insights

TL;DR: This article compares three supervised machine learning methods for combining multiple linguistic indicators for aspectual classification: decision trees, genetic programming, and logistic regression with an unsupervised method for this task.
Journal ArticleDOI

Customization in a unified framework for summarizing medical literature

TL;DR: The research shows that customization is feasible in a medical digital library and employs a unified user model to create a tailored summary of relevant documents for either a physician or lay person.
Journal ArticleDOI

Generating concise natural language summaries

TL;DR: An approach to summary generation that opportunistically folds information from multiple facts into a single sentence using concise linguistic constructions, which allows the construction of concise summaries, containing complex sentences that pack in information.
Posted Content

Aligning Noisy Parallel Corpora Across Language Groups : Word Pair Feature Matching by Dynamic Time Warping

TL;DR: The authors proposed a new algorithm called DK-vec for aligning pairs of Asian/Indo-European noisy parallel texts without sentence boundaries, which uses frequency, position and recency information as features for pattern matching.
Posted Content

WikiLingua: A New Benchmark Dataset for Cross-Lingual Abstractive Summarization

TL;DR: A method for direct crosslingual summarization without requiring translation at inference time is proposed by leveraging synthetic data and Neural Machine Translation as a pre-training step, which significantly outperforms the baseline approaches, while being more cost efficient during inference.