scispace - formally typeset
Proceedings ArticleDOI

Quadruplet Networks for Sketch-Based Image Retrieval

TLDR
This paper uses deep convolutional neural networks (ConvNets) to address sketch-based image retrieval (SBIR) and proposes a new architecture the quadruplet networks which enhance ConvNet features for SBIR, enabling ConvNets to extract more robust global and local features.
Abstract
Freehand sketches are a simple and powerful tool for communication. They are easily recognized across cultures and suitable for various applications. In this paper, we use deep convolutional neural networks (ConvNets) to address sketch-based image retrieval (SBIR). We first train our ConvNets on sketch and image object recognition in a large scale benchmark for SBIR (the sketchy database). We then conduct a comprehensive study of ConvNets features for SBIR, using a kNN similarity search paradigm in the ConvNet feature space. In contrast to recent SBIR works, we propose a new architecture the quadruplet networks which enhance ConvNet features for SBIR. This new architecture enables ConvNets to extract more robust global and local features. We evaluate our approach on three large scale datasets. Our quadruplet networks outperform previous state-of-the-art on two of them by a significant margin and gives competitive results on the third. Our system achieves a recall of 42.16% (at k=1) for the sketchy database (more than 5% improvement), a Kendal score of 43.28Τb on the TU-Berlin SBIR benchmark (close to 6Τb improvement) and a mean average precision (MAP) of 32.16% on Flickr15k (a category level SBIR benchmark).

read more

Citations
More filters

IEEE transactions on pattern analysis and machine intelligence

Ieee Xplore
TL;DR: This special issue aims at gathering the recent advances in learning with shared information methods and their applications in computer vision and multimedia analysis and addressing interesting real-world computer Vision and multimedia applications.
Journal ArticleDOI

Enhancing Sketch-Based Image Retrieval by CNN Semantic Re-ranking

TL;DR: A convolutional neural network semantic re-ranking system to enhance the performance of sketch-based image retrieval (SBIR) and achieves significantly higher precision in the top ten different SBIR methods and datasets.
Journal ArticleDOI

Sketching out the details: Sketch-based image retrieval using convolutional neural networks with multi-stage regression

TL;DR: A hybrid multi-stage training network is described that exploits both contrastive and triplet networks to exceed state of the art performance on several SBIR benchmarks by a significant margin.
Book ChapterDOI

Deep Shape Matching

TL;DR: A network is trained with edge maps of landmark images, which are automatically obtained by a structure-from-motion pipeline, providing improvements on challenging cases of domain generalization, generic sketch-based image retrieval or its fine-grained counterpart.
Proceedings ArticleDOI

Ensemble Deep Manifold Similarity Learning Using Hard Proxies

TL;DR: A new time- and memory-efficient method for estimating the manifold similarities by using a closed-form convergence solution of the Random Walk algorithm that outperforms the state of the art in both image retrieval and clustering on the benchmark CUB-200-2011, Cars196, and Stanford Online Products datasets.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI

Distinctive Image Features from Scale-Invariant Keypoints

TL;DR: This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene and can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Proceedings ArticleDOI

Going deeper with convolutions

TL;DR: Inception as mentioned in this paper is a deep convolutional neural network architecture that achieves the new state of the art for classification and detection in the ImageNet Large-Scale Visual Recognition Challenge 2014 (ILSVRC14).
Journal ArticleDOI

ImageNet classification with deep convolutional neural networks

TL;DR: A large, deep convolutional neural network was trained to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes and employed a recently developed regularization method called "dropout" that proved to be very effective.