scispace - formally typeset
Search or ask a question
Topic

View synthesis

About: View synthesis is a research topic. Over the lifetime, 1701 publications have been published within this topic receiving 42333 citations.


Papers
More filters
Proceedings ArticleDOI
01 Oct 2017
TL;DR: A deep learning approach is applied to the domain of foot scanning, and a method to reconstruct a 3D point cloud from a single input depth map is presented.
Abstract: In footwear, fit is highly dependent on foot shape, which is not fully captured by shoe size. Scanners can be used to acquire better sizing information and allow for more personalized footwear matching, however when scanning an object, many images are usually needed for reconstruction. Semantics such as knowing the kind of object in view can be leveraged to determine the full 3D shape given only one input view. Deep learning methods have been shown to be able to reconstruct 3D shape from limited inputs in highly symmetrical objects such as furniture and vehicles. We apply a deep learning approach to the domain of foot scanning, and present a method to reconstruct a 3D point cloud from a single input depth map. Anthropomorphic body parts can be challenging due to their irregular shapes, difficulty for parameterizing and limited symmetries. We train a view synthesis based network and show that our method can produce foot scans with accuracies of 1.55 mm from a single input depth map.

7 citations

Proceedings ArticleDOI
01 Dec 2014
TL;DR: The main idea of the method is to exploit disocclusion properties to limit and guide the search for candidate patches (labels) and to minimize efficiently the resulting MRF energy.
Abstract: In this paper, we propose a novel patch-based disocclusion filling method for view synthesis from video-plus-depth data. The proposed method treats disocclusion filling as a global optimization problem, where global (spatial) consistency among the patches is enforced via a Markov random field (MRF) model. The main idea of our method is to exploit disocclusion properties to limit and guide the search for candidate patches (labels) and to minimize efficiently the resulting MRF energy. In particular, we propose to constrain the label selection to local background regions in order to ensure that the disocclusions are filled with background information. Background is determined based on the locally estimated hard threshold on the depth values. The efficient minimization approach represents an extension of our previous method for general inpainting, where we propose to visit the MRF nodes from the background to the foreground disocclusion border and discard unnecessary labels. In this way, the number of labels is further reduced and the propagation of background information is additionally enforced. Finally, efficient inference is performed to obtain the final inpainting result. The proposed disocclusion filling method represents one step of the complete view synthesis framework that we also introduce in this paper. Experimental results show improvement of the proposed approach over related state-of-the-art methods both for small and big disocclusions.

7 citations

Proceedings ArticleDOI
01 Oct 2014
TL;DR: This work studies the luminance compression under GBR, which is not well considered in existing literature and shows that the graph-based transform can be applied on the GBR paradigm, hence better extracting the correlation among pixels along graph connections.
Abstract: Multi-view video transmission poses great challenges because of its data size and dimension. Therefore, how to design efficient 3D scene representations and coding (of luminance and geometry) has become a critical research topic. Recently, the graph-based representation (GBR) is introduced, which provides a lossless compression of multi-view geometry by connecting informative pixels among views. This representation has been shown as a promising alternative to the classical depth-based representation, where the view synthesis accuracy is hard to control. In this work, we study the luminance compression under GBR, which is not well considered in existing literature. With a proper structural reformulation, we show that the graph-based transform can be applied on the GBR paradigm, hence better extracting the correlation among pixels along graph connections. Moreover, we extend the popular SPIHT coding scheme to further improve coding efficiency. The experimental results show that our method leads to better RD coding performance as compared the classical luminance coding algorithms.

7 citations

01 Jan 2008
TL;DR: A technique to project a synthesized texture onto 3D textureless real objects without contact, which can be applied for example, in medical or archaeological applications, and successfully determines both 3D geometry and the appropriate texture to be projected onto the 3D model.
Abstract: We propose a technique to project a synthesized texture onto 3D textureless real objects without contact, which can be applied for example, in medical or archaeological applications. Our system uses two cameras and one projector. By projecting a suitable set of light patterns onto the surface of a 3D real object and by capturing images with one or several cameras, a large number of correspondences can be determined and the 3D geometry can be estimated. Since we deal with textureless objects, traditional corner detection algorithms can not be applied, but our approach successfully determines both 3D geometry and the appropriate texture to be projected onto the 3D model.

7 citations

Proceedings ArticleDOI
12 Jul 2015
TL;DR: A set of color and depth cameras are designed to reduce occlusion regions and improve the depth precision in the less-detailed region and view synthesis is performed to create the gaze corrected image from the obtained depth information.
Abstract: In this paper, we propose a gaze correction method using 3D video processing techniques including depth estimation and virtual view synthesis. We design a set of color and depth cameras to reduce occlusion regions and improve the depth precision in the less-detailed region. The proposed algorithm deals with fully unsolved problems by fusing the depth data from both depth sensors. Furthermore, view synthesis is performed to create the gaze corrected image from the obtained depth information. Experimental results show our contribution is useful for videoconferencing.

7 citations


Network Information
Related Topics (5)
Image segmentation
79.6K papers, 1.8M citations
86% related
Feature (computer vision)
128.2K papers, 1.7M citations
86% related
Object detection
46.1K papers, 1.3M citations
85% related
Convolutional neural network
74.7K papers, 2M citations
85% related
Feature extraction
111.8K papers, 2.1M citations
84% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202354
2022117
2021189
2020158
2019114
2018102