scispace - formally typeset
Search or ask a question

Showing papers by "Andrés Bruhn published in 2018"


Proceedings Article
03 Jun 2018
TL;DR: In this paper, an unsupervised online learning approach based on a convolutional neural network (CNN) is proposed to estimate a motion model individually for each frame by relating forward and backward motion, which not only allow to infer valuable motion information based on the backward flow, but also help to improve the performance at occlusions.
Abstract: Temporal coherence is a valuable source of information in the context of optical flow estimation. However, finding a suitable motion model to leverage this information is a non-trivial task. In this paper we propose an unsupervised online learning approach based on a convolutional neural network (CNN) that estimates such a motion model individually for each frame. By relating forward and backward motion these learned models not only allow to infer valuable motion information based on the backward flow, they also help to improve the performance at occlusions, where a reliable prediction is particularly useful. Moreover, our learned models are spatially variant and hence allow to estimate non-rigid motion per construction. This, in turns, allows to overcome the major limitation of recent rigidity-based approaches that seek to improve the estimation by incorporating additional stereo/SfM constraints. Experiments demonstrate the usefulness of our new approach. They not only show a consistent improvement of up to 27% for all major benchmarks (KITTI 2012, KITTI 2015, MPI Sintel) compared to a baseline without prediction, they also show top results for the MPI Sintel benchmark -- the one of the three benchmarks that contains the largest amount of non-rigid motion.

36 citations


Journal ArticleDOI
TL;DR: This paper fuse a Lambertian SfS approach with a robust stereo model and supplement the resulting energy functional with a detail-preserving anisotropic second-order smoothness term, and extends the resulting model in such a way that it jointly estimates depth, albedo and illumination.
Abstract: Shape from shading (SfS) and stereo are two fundamentally different strategies for image-based 3-D reconstruction. While approaches for SfS infer the depth solely from pixel intensities, methods for stereo are based on a matching process that establishes correspondences across images. This difference in approaching the reconstruction problem yields complementary advantages that are worthwhile being combined. So far, however, most “joint” approaches are based on an initial stereo mesh that is subsequently refined using shading information. In this paper we follow a completely different approach. We propose a joint variational method that combines both cues within a single minimisation framework. To this end, we fuse a Lambertian SfS approach with a robust stereo model and supplement the resulting energy functional with a detail-preserving anisotropic second-order smoothness term. Moreover, we extend the resulting model in such a way that it jointly estimates depth, albedo and illumination. This in turn makes the approach applicable to objects with non-uniform albedo as well as to scenes with unknown illumination. Experiments for synthetic and real-world images demonstrate the benefits of our combined approach: They not only show that our method is capable of generating very detailed reconstructions, but also that joint approaches are feasible in practice.

22 citations


Book ChapterDOI
08 Sep 2018
TL;DR: This paper proposes a novel structure-from-motion-aware PatchMatch approach that, in contrast to existing matching techniques, combines two hierarchical feature matching methods: a recent two-framePatchMatch approach for optical flow estimation (general motion) and a specifically tailored three-frame Patch match approach for rigid scene reconstruction (SfM).
Abstract: Many recent energy-based methods for optical flow estimation rely on a good initialization that is typically provided by some kind of feature matching. So far, however, these initial matching approaches are rather general: They do not incorporate any additional information that could help to improve the accuracy or the robustness of the estimation. In particular, they do not exploit potential cues on the camera poses and the thereby induced rigid motion of the scene. In the present paper, we tackle this problem. To this end, we propose a novel structure-from-motion-aware PatchMatch approach that, in contrast to existing matching techniques, combines two hierarchical feature matching methods: a recent two-frame PatchMatch approach for optical flow estimation (general motion) and a specifically tailored three-frame PatchMatch approach for rigid scene reconstruction (SfM). While the motion PatchMatch serves as baseline with good accuracy, the SfM counterpart takes over at occlusions and other regions with insufficient information. Experiments with our novel SfM-aware PatchMatch approach demonstrate its usefulness. They not only show excellent results for all major benchmarks (KITTI 2012/2015, MPI Sintel), but also improvements up to 50% compared to a PatchMatch approach without structure information.

11 citations


Posted Content
TL;DR: An unsupervised online learning approach based on a convolutional neural network that estimates such a motion model individually for each frame, allowing to overcome the major limitation of recent rigidity-based approaches that seek to improve the estimation by incorporating additional stereo/SfM constraints.
Abstract: Temporal coherence is a valuable source of information in the context of optical flow estimation. However, finding a suitable motion model to leverage this information is a non-trivial task. In this paper we propose an unsupervised online learning approach based on a convolutional neural network (CNN) that estimates such a motion model individually for each frame. By relating forward and backward motion these learned models not only allow to infer valuable motion information based on the backward flow, they also help to improve the performance at occlusions, where a reliable prediction is particularly useful. Moreover, our learned models are spatially variant and hence allow to estimate non-rigid motion per construction. This, in turns, allows to overcome the major limitation of recent rigidity-based approaches that seek to improve the estimation by incorporating additional stereo/SfM constraints. Experiments demonstrate the usefulness of our new approach. They not only show a consistent improvement of up to 27% for all major benchmarks (KITTI 2012, KITTI 2015, MPI Sintel) compared to a baseline without prediction, they also show top results for the MPI Sintel benchmark -- the one of the three benchmarks that contains the largest amount of non-rigid motion.

11 citations


Proceedings Article
01 Jan 2018
TL;DR: This paper proposes a trajectorial refinement method that lifts successful concepts of recent variational two-frame methods to the multi-frame domain and demonstrates the clear benefits of using multiple frames and of imposing directional constraints on the prefiltering step and the refinement.
Abstract: Pipeline approaches that interpolate and refine an initial set of point correspondences have recently shown a good performance in the field of optical flow estimation. However, so far, these methods are typically restricted to two frames which makes exploiting temporal information difficult. In this paper, we show how such pipeline approaches can be extended to the temporal domain and how directional constraints can be incorporated to further improve the estimation. To this end, we not only suggest to exploit temporal information in the prefiltering step, we also propose a trajectorial refinement method that lifts successful concepts of recent variational two-frame methods to the multi-frame domain. Experiments demonstrate the usefulness of our pipeline approach. They do not only show good results in general, they also demonstrate the clear benefits of using multiple frames and of imposing directional constraints on the prefiltering step and the refinement.

2 citations