scispace - formally typeset
A

Andrew Zisserman

Researcher at University of Oxford

Publications -  808
Citations -  312028

Andrew Zisserman is an academic researcher from University of Oxford. The author has contributed to research in topics: Convolutional neural network & Real image. The author has an hindex of 167, co-authored 808 publications receiving 261717 citations. Previous affiliations of Andrew Zisserman include University of Edinburgh & Microsoft.

Papers
More filters
Proceedings ArticleDOI

The Visual Centrifuge: Model-Free Layered Video Representations

TL;DR: In this paper, uncertainty-capturing 3D convolutional architectures are used to separate blended videos and then generalize to single videos, where they exhibit interesting abilities: color constancy, factoring out shadows and separating reflections.
Proceedings ArticleDOI

Linear auto-calibration for ground plane motion

TL;DR: This work shows that when there is some control over the motion of the camera, a fast linear solution is available without these restrictions, and shows the algorithm to be simple, fast, and accurate.
Book ChapterDOI

Recognising rotationally symmetric surfaces from their outlines

TL;DR: This paper shows techniques for recognising a significant class of surfaces from a single perspective view that uses geometrical facts about bitangencies, creases, and inflections to compute descriptions of the surface's shape from its image outline, unaffected by the viewpoint or the camera parameters.
Proceedings ArticleDOI

Co-Attention for Conditioned Image Matching

TL;DR: In this paper, a spatial attention mechanism (a co-attention module, CoAM) is proposed to determine correspondences between image pairs in the wild under large changes in illumination, viewpoint, context, and material.
Posted Content

Self-Supervised MultiModal Versatile Networks

TL;DR: This work learns representations using self-supervision by leveraging three modalities naturally present in videos: vision, audio and language by incorporating a novel process of deflation, so that the networks can be effortlessly applied to the visual data in the form of video or a static image.