scispace - formally typeset
Search or ask a question
Topic

Epipolar geometry

About: Epipolar geometry is a research topic. Over the lifetime, 4224 publications have been published within this topic receiving 135847 citations.


Papers
More filters
Proceedings ArticleDOI
01 Dec 2013
TL;DR: A novel simple and principled algorithm is presented that computes dense depth estimation by combining both defocus and correspondence depth cues, and shows how to combine the two cues into a high quality depth map, suitable for computer vision applications such as matting, full control of depth-of-field, and surface reconstruction.
Abstract: Light-field cameras have recently become available to the consumer market. An array of micro-lenses captures enough information that one can refocus images after acquisition, as well as shift one's viewpoint within the sub-apertures of the main lens, effectively obtaining multiple views. Thus, depth cues from both defocus and correspondence are available simultaneously in a single capture. Previously, defocus could be achieved only through multiple image exposures focused at different depths, while correspondence cues needed multiple exposures at different viewpoints or multiple cameras, moreover, both cues could not easily be obtained together. In this paper, we present a novel simple and principled algorithm that computes dense depth estimation by combining both defocus and correspondence depth cues. We analyze the x-u 2D epipolar image (EPI), where by convention we assume the spatial x coordinate is horizontal and the angular u coordinate is vertical (our final algorithm uses the full 4D EPI). We show that defocus depth cues are obtained by computing the horizontal (spatial) variance after vertical (angular) integration, and correspondence depth cues by computing the vertical (angular) variance. We then show how to combine the two cues into a high quality depth map, suitable for computer vision applications such as matting, full control of depth-of-field, and surface reconstruction.

582 citations

Journal ArticleDOI
TL;DR: The problem of view synthesis is formulated as a continuous inverse problem, which allows us to correctly take into account foreshortening effects caused by scene geometry transformations, and all optimization problems are solved with state-of-the-art convex relaxation techniques.
Abstract: We develop a continuous framework for the analysis of 4D light fields, and describe novel variational methods for disparity reconstruction as well as spatial and angular super-resolution. Disparity maps are estimated locally using epipolar plane image analysis without the need for expensive matching cost minimization. The method works fast and with inherent subpixel accuracy since no discretization of the disparity space is necessary. In a variational framework, we employ the disparity maps to generate super-resolved novel views of a scene, which corresponds to increasing the sampling rate of the 4D light field in spatial as well as angular direction. In contrast to previous work, we formulate the problem of view synthesis as a continuous inverse problem, which allows us to correctly take into account foreshortening effects caused by scene geometry transformations. All optimization problems are solved with state-of-the-art convex relaxation techniques. We test our algorithms on a number of real-world examples as well as our new benchmark data set for light fields, and compare results to a multiview stereo method. The proposed method is both faster as well as more accurate. Data sets and source code are provided online for additional evaluation.

575 citations

Journal ArticleDOI
TL;DR: To increase the robustness of the system, two semi-local constraints on combinations of region correspondences are derived (one geometric, the other photometric) allow to test the consistency of correspondences and hence to reject falsely matched regions.
Abstract: ‘Invariant regions’ are self-adaptive image patches that automatically deform with changing viewpoint as to keep on covering identical physical parts of a scene. Such regions can be extracted directly from a single image. They are then described by a set of invariant features, which makes it relatively easy to match them between views, even under wide baseline conditions. In this contribution, two methods to extract invariant regions are presented. The first one starts from corners and uses the nearby edges, while the second one is purely intensity-based. As a matter of fact, the goal is to build an opportunistic system that exploits several types of invariant regions as it sees fit. This yields more correspondences and a system that can deal with a wider range of images. To increase the robustness of the system, two semi-local constraints on combinations of region correspondences are derived (one geometric, the other photometric). They allow to test the consistency of correspondences and hence to reject falsely matched regions. Experiments on images of real-world scenes taken from substantially different viewpoints demonstrate the feasibility of the approach.

568 citations

Journal ArticleDOI
TL;DR: Several variants of the algorithm are developed that avoid classical regularization while imposing several global cohesiveness constraints, and this is a novel approach that has the advantage of guaranteeing that solutions minimize the original cost function and preserve discontinuities.

533 citations

Journal ArticleDOI
TL;DR: A method is designed, based on intersecting epipolar constraints, for providing ground truth correspondence automatically, which is based purely on geometric information, and does not rely on the choice of a specific feature appearance descriptor.
Abstract: We explore the performance of a number of popular feature detectors and descriptors in matching 3D object features across viewpoints and lighting conditions. To this end we design a method, based on intersecting epipolar constraints, for providing ground truth correspondence automatically. These correspondences are based purely on geometric information, and do not rely on the choice of a specific feature appearance descriptor. We test detector-descriptor combinations on a database of 100 objects viewed from 144 calibrated viewpoints under three different lighting conditions. We find that the combination of Hessian-affine feature finder and SIFT features is most robust to viewpoint change. Harris-affine combined with SIFT and Hessian-affine combined with shape context descriptors were best respectively for lighting change and change in camera focal length. We also find that no detector-descriptor combination performs well with viewpoint changes of more than 25---30?.

497 citations


Network Information
Related Topics (5)
Image segmentation
79.6K papers, 1.8M citations
90% related
Feature (computer vision)
128.2K papers, 1.7M citations
87% related
Convolutional neural network
74.7K papers, 2M citations
87% related
Feature extraction
111.8K papers, 2.1M citations
86% related
Pixel
136.5K papers, 1.5M citations
85% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202373
2022180
2021114
2020191
2019168
2018183