scispace - formally typeset
M

Mark Johnson

Researcher at Oracle Corporation

Publications -  325
Citations -  24160

Mark Johnson is an academic researcher from Oracle Corporation. The author has contributed to research in topics: Parsing & Language model. The author has an hindex of 62, co-authored 323 publications receiving 20142 citations. Previous affiliations of Mark Johnson include Brown University & Carnegie Mellon University.

Papers
More filters
Proceedings ArticleDOI

Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering

TL;DR: In this paper, a bottom-up and top-down attention mechanism was proposed to enable attention to be calculated at the level of objects and other salient image regions, which achieved state-of-the-art results on the MSCOCO test server.
Posted Content

Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering.

TL;DR: A combined bottom-up and top-down attention mechanism that enables attention to be calculated at the level of objects and other salient image regions is proposed, demonstrating the broad applicability of this approach to VQA.
Proceedings ArticleDOI

Coarse-to-Fine n-Best Parsing and MaxEnt Discriminative Reranking

TL;DR: This paper describes a simple yet novel method for constructing sets of 50- best parses based on a coarse-to-fine generative parser that generates 50-best lists that are of substantially higher quality than previously obtainable.
Book ChapterDOI

SPICE: Semantic Propositional Image Caption Evaluation

TL;DR: This paper proposed a new automated caption evaluation metric defined over scene graphs coined SPICE, which captures human judgments over model-generated captions better than other automatic metrics (e.g., system-level correlation of 0.88 with human judgments on the MS COCO dataset, versus 0.43 for CIDEr and 0.53 for METEOR).
Posted Content

SPICE: Semantic Propositional Image Caption Evaluation

TL;DR: It is hypothesized that semantic propositional content is an important component of human caption evaluation, and a new automated caption evaluation metric defined over scene graphs coined SPICE is proposed, which can answer questions such as which caption-generator best understands colors?