scispace - formally typeset
Search or ask a question
Topic

Distance transform

About: Distance transform is a research topic. Over the lifetime, 2886 publications have been published within this topic receiving 59481 citations.


Papers
More filters
Posted Content
TL;DR: In this article, an object mask network (OMNOMN) is proposed to predict masks that go beyond the scope of the bounding boxes and are thus robust to inaccurate object candidates.
Abstract: We address the problem of instance-level semantic segmentation, which aims at jointly detecting, segmenting and classifying every individual object in an image. In this context, existing methods typically propose candidate objects, usually as bounding boxes, and directly predict a binary mask within each such proposal. As a consequence, they cannot recover from errors in the object candidate generation process, such as too small or shifted boxes. In this paper, we introduce a novel object segment representation based on the distance transform of the object masks. We then design an object mask network (OMN) with a new residual-deconvolution architecture that infers such a representation and decodes it into the final binary object mask. This allows us to predict masks that go beyond the scope of the bounding boxes and are thus robust to inaccurate object candidates. We integrate our OMN into a Multitask Network Cascade framework, and learn the resulting boundary-aware instance segmentation (BAIS) network in an end-to-end manner. Our experiments on the PASCAL VOC 2012 and the Cityscapes datasets demonstrate the benefits of our approach, which outperforms the state-of-the-art in both object proposal generation and instance segmentation.

79 citations

DOI
27 May 2002
TL;DR: A robust and efficient method in 2D and 3D for the calculation of skeletons for arbitrary objects is presented, combined with a new indicator to identify the skeleton, which coincides with the singularity set of the distance map.
Abstract: A robust and efficient method in 2D and 3D for the calculation of skeletons for arbitrary objects is presented. The method is based on the calculation of the distance function with respect to the object boundary. This is combined, in a post processing step, with a new indicator to identify the skeleton, which coincides with the singularity set of the distance map. The indicator is defined as a suitable function of certain local momenta of this distance map and allows a robust and accurate computation of the distance from the skeleton set. This distance is then extended, again via the level set method, onto the whole space. Several applications in 2D and 3D are presented.

78 citations

Journal ArticleDOI
TL;DR: Two new features based on distance information are proposed which contains rich information encoding both the black/white and directional distance distributions and a new concept of map tiling is introduced and applied to the DDD feature to improve its discriminative power.
Abstract: Features play an important role in OCR systems. In this paper, we propose two new features which are based on distance information. In the first feature (called DT, Distance Transformation), each white pixel has a distance value to the nearest black pixel. The second feature is called DDD (Directional Distance Distribution) which contains rich information encoding both the black/white and directional distance distributions. A new concept of map tiling is introduced and applied to the DDD feature to improve its discriminative power. For an objective evaluation and comparison of the proposed and conventional features, three distinct sets of characters (i.e., numerals, English capital letters, and Hangul initial sounds) have been tested using standard databases. Based on the results, three propositions can be derived to confirm the superiority of both the DDD feature and the map tilings.

78 citations

Patent
25 Jan 2000
TL;DR: In this paper, the authors calculate the position of a view point V to be set behind the current position of the car and up in the air, based on the position calculated by the position-measuring part and the position correcting part.
Abstract: The view point-setting part calculates the position of the view point V to be set behind the current position of the car and up in the air, based on the current position calculated by the current position-measuring part and the position correcting part. The data reading part reads out map data M about map elements such as buildings, which are located around the current position. The representing method-changing part calculates the distance or level difference between the viewpoint and each of the map elements, and, for example, changes the representation shapes of ones of the map elements near the view point, that is, in the region within a predetermined distance from the view point, to detailed shapes, and the rest ones far from the view point, that is, outside the region within a predetermined distance from the view point, to simple shapes, respectively. Finally, the display processing part composes an image of each of the map element according to the displaying mode of the map element, which is sent from the representation method-changing part, and displays the image.

78 citations

Proceedings ArticleDOI
01 Dec 2008
TL;DR: This work proposes the use of 3D (2D+time) shape context to recognize the spatial and temporal details inherent in human actions and introduces a non-uniform sampling method that gives preference to fast moving body parts using a Euclidean 3D distance transform.
Abstract: We propose the use of 3D (2D+time) shape context to recognize the spatial and temporal details inherent in human actions. We represent an action in a video sequence by a 3D point cloud extracted by sampling 2D silhouettes over time. A non-uniform sampling method is introduced that gives preference to fast moving body parts using a Euclidean 3D distance transform. Actions are then classified by matching the extracted point clouds. Our proposed approach is based on a global matching and does not require specific training to learn the model. We test the approach thoroughly on two publicly available datasets and compare to several state-of-the-art methods. The achieved classification accuracy is on par with or superior to the best results reported to date.

77 citations


Network Information
Related Topics (5)
Image segmentation
79.6K papers, 1.8M citations
91% related
Image processing
229.9K papers, 3.5M citations
91% related
Feature (computer vision)
128.2K papers, 1.7M citations
90% related
Convolutional neural network
74.7K papers, 2M citations
89% related
Feature extraction
111.8K papers, 2.1M citations
88% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20235
202217
202161
202099
2019112
201881