scispace - formally typeset
R

Ranjay Krishna

Researcher at Stanford University

Publications -  70
Citations -  10873

Ranjay Krishna is an academic researcher from Stanford University. The author has contributed to research in topics: Computer science & Scene graph. The author has an hindex of 24, co-authored 47 publications receiving 7403 citations.

Papers
More filters
Journal ArticleDOI

Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations

TL;DR: The Visual Genome dataset as mentioned in this paper contains over 108k images where each image has an average of $35$35 objects, $26$26 attributes, and $21$21 pairwise relationships between objects.
Posted Content

Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations

TL;DR: The Visual Genome dataset is presented, which contains over 108K images where each image has an average of $$35$$35 objects, $$26$$26 attributes, and $$21$$21 pairwise relationships between objects, and represents the densest and largest dataset of image descriptions, objects, attributes, relationships, and question answer pairs.
Proceedings ArticleDOI

Image retrieval using scene graphs

TL;DR: A conditional random field model that reasons about possible groundings of scene graphs to test images and shows that the full model can be used to improve object localization compared to baseline methods and outperforms retrieval methods that use only objects or low-level image features.
Book ChapterDOI

Visual Relationship Detection with Language Priors

TL;DR: In this article, the authors propose a model that uses this insight to train visual models for objects and predicates individually and later combines them together to predict multiple relationships per image and localize the objects in the predicted relationships as bounding boxes in the image.
Proceedings ArticleDOI

Dense-Captioning Events in Videos

TL;DR: In this article, the authors introduce a new model that is able to identify all events in a single pass of the video while simultaneously describing the detected events with natural language, which is called ActivityNet Captions.