scispace - formally typeset
Proceedings ArticleDOI

Paper Doll Parsing: Retrieving Similar Styles to Parse Clothing Items

Reads0
Chats0
TLDR
This paper tackles the clothing parsing problem using a retrieval based approach that combines parsing from: pre-trained global clothing models, local clothing models learned on the fly from retrieved examples, and transferred parse masks (paper doll item transfer) from retrieved example.
Abstract
Clothing recognition is an extremely challenging problem due to wide variation in clothing item appearance, layering, and style. In this paper, we tackle the clothing parsing problem using a retrieval based approach. For a query image, we find similar styles from a large database of tagged fashion images and use these examples to parse the query. Our approach combines parsing from: pre-trained global clothing models, local clothing models learned on the fly from retrieved examples, and transferred parse masks (paper doll item transfer) from retrieved examples. Experimental evaluation shows that our approach significantly outperforms state of the art in parsing accuracy.

read more

Content maybe subject to copyright    Report

Citations
More filters
Proceedings ArticleDOI

Image-Based Recommendations on Styles and Substitutes

TL;DR: The approach is not based on fine-grained modeling of user annotations but rather on capturing the largest dataset possible and developing a scalable method for uncovering human notions of the visual relationships within.
Proceedings ArticleDOI

DeepFashion: Powering Robust Clothes Recognition and Retrieval with Rich Annotations

TL;DR: This work introduces DeepFashion1, a large-scale clothes dataset with comprehensive annotations, and proposes a new deep model, namely FashionNet, which learns clothing features by jointly predicting clothing attributes and landmarks.
Proceedings ArticleDOI

Hypercolumns for object segmentation and fine-grained localization

TL;DR: In this paper, the authors define the hypercolumn at a pixel as the vector of activations of all CNN units above that pixel, and use hypercolumns as pixel descriptors.
Posted Content

Hypercolumns for Object Segmentation and Fine-grained Localization

TL;DR: Using hypercolumns as pixel descriptors, this work defines the hypercolumn at a pixel as the vector of activations of all CNN units above that pixel, and shows results on three fine-grained localization tasks: simultaneous detection and segmentation, and keypoint localization.
Proceedings ArticleDOI

Look into Person: Self-Supervised Structure-Sensitive Learning and a New Benchmark for Human Parsing

TL;DR: A new benchmark Look into Person (LIP) is introduced that makes a significant advance in terms of scalability, diversity and difficulty, and a novel self-supervised structure-sensitive learning approach, which imposes human pose structures into parsing results without resorting to extra supervision.
References
More filters
Journal Article

LIBLINEAR: A Library for Large Linear Classification

TL;DR: LIBLINEAR is an open source library for large-scale linear classification that supports logistic regression and linear support vector machines and provides easy-to-use command-line tools and library calls for users and developers.
Journal ArticleDOI

Fast approximate energy minimization via graph cuts

TL;DR: This work presents two algorithms based on graph cuts that efficiently find a local minimum with respect to two types of large moves, namely expansion moves and swap moves that allow important cases of discontinuity preserving energies.
Journal ArticleDOI

Efficient Graph-Based Image Segmentation

TL;DR: An efficient segmentation algorithm is developed based on a predicate for measuring the evidence for a boundary between two regions using a graph-based representation of the image and it is shown that although this algorithm makes greedy decisions it produces segmentations that satisfy global properties.
Journal ArticleDOI

An experimental comparison of min-cut/max- flow algorithms for energy minimization in vision

TL;DR: This paper compares the running times of several standard algorithms, as well as a new algorithm that is recently developed that works several times faster than any of the other methods, making near real-time performance possible.
Proceedings ArticleDOI

Vlfeat: an open and portable library of computer vision algorithms

TL;DR: VLFeat is an open and portable library of computer vision algorithms that includes rigorous implementations of common building blocks such as feature detectors, feature extractors, (hierarchical) k-means clustering, randomized kd-tree matching, and super-pixelization.
Related Papers (5)