scispace - formally typeset
Search or ask a question

Showing papers by "Jimei Yang published in 2015"


Proceedings Article
07 Dec 2015
TL;DR: A novel recurrent convolutional encoder-decoder network that is trained end-to-end on the task of rendering rotated objects starting from a single image and allows the model to capture long-term dependencies along a sequence of transformations.
Abstract: An important problem for both graphics and vision is to synthesize novel views of a 3D object from a single image. This is particularly challenging due to the partial observability inherent in projecting a 3D object onto the image space, and the ill-posedness of inferring object shape and pose. However, we can train a neural network to address the problem if we restrict our attention to specific object categories (in our case faces and chairs) for which we can gather ample training data. In this paper, we propose a novel recurrent convolutional encoder-decoder network that is trained end-to-end on the task of rendering rotated objects starting from a single image. The recurrent structure allows our model to capture long-term dependencies along a sequence of transformations. We demonstrate the quality of its predictions for human faces on the Multi-PIE dataset and for a dataset of 3D chair models, and also show its ability to disentangle latent factors of variation (e.g., identity and pose) without using full supervision.

346 citations


Proceedings ArticleDOI
07 Jun 2015
TL;DR: A novel multi-objective learning method that optimizes a single unified deep convolutional network with two distinct non-structured loss functions: one encoding the unary label likelihoods and the other encoding the pairwise label dependencies.
Abstract: This paper formulates face labeling as a conditional random field with unary and pairwise classifiers. We develop a novel multi-objective learning method that optimizes a single unified deep convolutional network with two distinct non-structured loss functions: one encoding the unary label likelihoods and the other encoding the pairwise label dependencies. Moreover, we regularize the network by using a nonparametric prior as new input channels in addition to the RGB image, and show that significant performance improvements can be achieved with a much smaller network size. Experiments on both the LFW and Helen datasets demonstrate state-of-the-art results of the proposed algorithm, and accurate labeling results on challenging images can be obtained by the proposed algorithm for real-world applications.

151 citations


Posted Content
TL;DR: In this paper, a variational auto-encoder is used to generate images from visual attributes, where the image is modeled as a composite of foreground and background and a layered generative model with disentangled latent variables is developed.
Abstract: This paper investigates a novel problem of generating images from visual attributes. We model the image as a composite of foreground and background and develop a layered generative model with disentangled latent variables that can be learned end-to-end using a variational auto-encoder. We experiment with natural images of faces and birds and demonstrate that the proposed models are capable of generating realistic and diverse samples with disentangled latent representations. We use a general energy minimization algorithm for posterior inference of latent variables given novel images. Therefore, the learned generative models show excellent quantitative and visual results in the tasks of attribute-conditioned image reconstruction and completion.

131 citations


Proceedings ArticleDOI
05 Jan 2015
TL;DR: A novel ensemble model that combines different color descriptors in the decision level through metric learning is proposed that significantly outperforms state-of-the-art algorithms on two challenging datasets (VIPeR and PRID 450S).
Abstract: Appearance-based human re-identification is challenging due to different camera characteristics, varying lighting conditions, pose variations across camera views, etc. Recent studies have revealed that color information plays a critical role on performance. However, two problems remain unclear: (1) how do different color descriptors perform under the same scene in re-identification problem? and (2) how can we combine these descriptors without losing their invariance property and distinctiveness power? In this paper, we propose a novel ensemble model that combines different color descriptors in the decision level through metric learning. Experiments show that the proposed system significantly outperforms state-of-the-art algorithms on two challenging datasets (VIPeR and PRID 450S). We have improved the Rank 1 recognition rate on VIPeR dataset by 8.7%.

54 citations


Proceedings ArticleDOI
07 Jun 2015
TL;DR: This paper develops a novel cascade algorithm, PatchCut, for coarse-to-fine object segmentation, which matches query image patches with example images in multiscale to enable local shape transfer.
Abstract: Object segmentation is highly desirable for image understanding and editing. Current interactive tools require a great deal of user effort while automatic methods are usually limited to images of special object categories or with high color contrast. In this paper, we propose a data-driven algorithm that uses examples to break through these limits. As similar objects tend to share similar local shapes, we match query image patches with example images in multiscale to enable local shape transfer. The transferred local shape masks constitute a patch-level segmentation solution space and we thus develop a novel cascade algorithm, PatchCut, for coarse-to-fine object segmentation. In each stage of the cascade, local shape mask candidates are selected to refine the estimated segmentation of the previous stage iteratively with color models. Experimental results on various datasets (Weizmann Horse, Fashionista, Object Discovery and PASCAL) demonstrate the effectiveness and robustness of our algorithm.

25 citations


Patent
04 Aug 2015
TL;DR: In this paper, a technique for automatically segmenting an object in an image by referencing a dataset of already-segmented images is described. But this technique is not suitable for image classification.
Abstract: Systems and methods are disclosed herein for using one or more computing devices to automatically segment an object in an image by referencing a dataset of already-segmented images. The technique generally involves identifying a patch of an already-segmented image in the dataset based on the patch of the already-segmented image being similar to an area of the image including a patch of the image. The technique further involves identifying a mask of the patch of the already-segmented image, the mask representing a segmentation in the already-segmented image. The technique also involves segmenting the object in the image based on at least a portion of the mask of the patch of the already-segmented image.

9 citations


Proceedings ArticleDOI
05 Jan 2015
TL;DR: This work exploits an efficient algorithm for structured edge prediction to extract regions from structured edge detection and develops a novel algorithm to link the structured edge and gPb hierarchical image segmentation framework with steerable filters.
Abstract: We present two approaches to extract regions from structured edge detection. While the state-of-the-art algorithm based on globalized probability of boundary (gPb) generates a hierarchical region tree, it entails significant computational load. In this work, we exploit an efficient algorithm for structured edge prediction to extract regions. To generate high quality regions, we develop a novel algorithm to link the structured edge and gPb hierarchical image segmentation framework with steerable filters. The extracted regions are grouped by the proposed hierarchical grouping method to generate object proposals for effective detection and recognition problems. We demonstrate the effectiveness of our region generation for image segmentation on the BSDS500 database, and region generation for object proposals on the PASCAL VOC 2007 benchmark database. Experimental results show that the proposed algorithm achieves the comparable or superior quality to the state-of-the-art methods.

5 citations