scispace - formally typeset
A

Andrew Zisserman

Researcher at University of Oxford

Publications -  808
Citations -  312028

Andrew Zisserman is an academic researcher from University of Oxford. The author has contributed to research in topics: Convolutional neural network & Real image. The author has an hindex of 167, co-authored 808 publications receiving 261717 citations. Previous affiliations of Andrew Zisserman include University of Edinburgh & Microsoft.

Papers
More filters
Proceedings ArticleDOI

Get Out of my Picture! Internet-based Inpainting.

TL;DR: This paper uses recent advances in viewpoint invariant image search to find other images of the same scene on the Internet to replace large occlusions in photographs, and uses a Markov random field formulation to combine the proposals into a single, occlusion-free result.
Proceedings ArticleDOI

A Compact and Discriminative Face Track Descriptor

TL;DR: This work proposes a novel face track descriptor, based on the Fisher Vector representation, and demonstrates that it has a number of favourable properties, including compact size and fast computation, which render it very suitable for large scale visual repositories.
Book ChapterDOI

DisLocation: Scalable Descriptor Distinctiveness for Location Recognition

TL;DR: A novel method for efficiently estimating distinctiveness of all database descriptors, based on estimating local descriptor density everywhere in the descriptor space is proposed, which takes only a 100 s on a single CPU core for a 1 million image database.
Proceedings ArticleDOI

Extracting projective structure from single perspective views of 3D point sets

TL;DR: It is proved by example that although the general statement is true, invariants do exist for structured three-dimensional point sets and they can be used to compute projective structure.
Journal ArticleDOI

You Said That?: Synthesising Talking Faces from Audio

TL;DR: An encoder–decoder convolutional neural network model is developed that uses a joint embedding of the face and audio to generate synthesised talking face video frames and proposed methods to re-dub videos by visually blending the generated face into the source video frame using a multi-stream CNN model.