scispace - formally typeset
Search or ask a question

Showing papers by "Jamie Shotton published in 2006"


Book ChapterDOI
07 May 2006
TL;DR: A new approach to learning a discriminative model of object classes, incorporating appearance, shape and context information efficiently, is proposed, which is used for automatic visual recognition and semantic segmentation of photographs.
Abstract: This paper proposes a new approach to learning a discriminative model of object classes, incorporating appearance, shape and context information efficiently. The learned model is used for automatic visual recognition and semantic segmentation of photographs. Our discriminative model exploits novel features, based on textons, which jointly model shape and texture. Unary classification and feature selection is achieved using shared boosting to give an efficient classifier which can be applied to a large number of classes. Accurate image segmentation is achieved by incorporating these classifiers in a conditional random field. Efficient training of the model on very large datasets is achieved by exploiting both random feature selection and piecewise training methods. High classification and segmentation accuracy are demonstrated on three different databases: i) our own 21-object class database of photographs of real objects viewed under general lighting conditions, poses and viewpoints, ii) the 7-class Corel subset and iii) the 7-class Sowerby database used in [1]. The proposed algorithm gives competitive results both for highly textured (e.g. grass, trees), highly structured (e.g. cars, faces, bikes, aeroplanes) and articulated objects (e.g. body, cow).

1,343 citations


Proceedings ArticleDOI
17 Jun 2006
TL;DR: This paper addresses the problem of detecting and segmenting partially occluded objects of a known category by defining a part labelling which densely covers the object and imposing asymmetric local spatial constraints on these labels to ensure the consistent layout of parts whilst allowing for object deformation.
Abstract: This paper addresses the problem of detecting and segmenting partially occluded objects of a known category. We first define a part labelling which densely covers the object. Our Layout Consistent Random Field (LayoutCRF) model then imposes asymmetric local spatial constraints on these labels to ensure the consistent layout of parts whilst allowing for object deformation. Arbitrary occlusions of the object are handled by avoiding the assumption that the whole object is visible. The resulting system is both efficient to train and to apply to novel images, due to a novel annealed layout-consistent expansion move algorithm paired with a randomised decision tree classifier. We apply our technique to images of cars and faces and demonstrate state-of-the-art detection and segmentation performance even in the presence of partial occlusion.

318 citations


Patent
John Winn1, Jamie Shotton1
21 Sep 2006
TL;DR: In this article, a conditional random field is used to force a global part labeling which is substantially layout-consistent and a part label map is inferred from this, which can be used to estimate belief distributions over parts for each image element of a test image.
Abstract: During a training phase we learn parts of images which assist in the object detection and recognition task. A part is a densely represented area of an image of an object to which we assign a unique label. Parts contiguously cover an image of an object to give a part label map for that object. The parts do not necessarily correspond to semantic object parts. During the training phase a classifier is learnt which can be used to estimate belief distributions over parts for each image element of a test image. A conditional random field is used to force a global part labeling which is substantially layout-consistent and a part label map is inferred from this. By recognizing parts we enable object detection and recognition even for partially occluded objects, for multiple-objects of different classes in the same scene, for unstructured and structured objects and allowing for object deformation.

59 citations


Journal ArticleDOI
TL;DR: A method that generates a composite image when a user types in nouns, such as “boat” and “sand” is presented, and a combined algorithm for automatically building an image library with semantic annotations from any photo collection is presented.
Abstract: Composite images are synthesized from existing photographs by artists who make concept art, e.g., storyboards for movies or architectural planning. Current techniques allow an artist to fabricate such an image by digitally splicing parts of stock photographs. While these images serve mainly to “quickly”convey how a scene should look, their production is laborious. We propose a technique that allows a person to design a new photograph with substantially less effort. This paper presents a method that generates a composite image when a user types in nouns, such as “boat”and “sand.”The artist can optionally design an intended image by specifying other constraints. Our algorithm formulates the constraints as queries to search an automatically annotated image database. The desired photograph, not a collage, is then synthesized using graph-cut optimization, optionally allowing for further user interaction to edit or choose among alternative generated photos. An implementation of our approach, shown in the associated video, demonstrates our contributions of (1) a method for creating specific images with minimal human effort, and (2) a combined algorithm for automatically building an image library with semantic annotations from any photo collection.

58 citations


Patent
21 Sep 2006
TL;DR: In this article, a shape filter comprises one or more regions of arbitrary shape, size and position within a bounding area of an image, paired with a specified texton, which describes the texture of a patch of surface of an object.
Abstract: Given an image of structured and/or unstructured objects we automatically partition it into semantically meaningful areas each labeled with a specific object class We use a novel type of feature which we refer to as a shape filter Shape filters enable us to capture some or all of shape, texture and appearance context information A shape filter comprises one or more regions of arbitrary shape, size and position within a bounding area of an image, paired with a specified texton A texton comprises information describing the texture of a patch of surface of an object In a training process we select a sub-set of possible shape filters and incorporate those into a conditional random field model of object classes That model is then used for object detection and recognition

46 citations