scispace - formally typeset
Proceedings ArticleDOI

A Unified Deep Learning Approach for Foveated Rendering & Novel View Synthesis from Sparse RGB-D Light Fields

Reads0
Chats0
TLDR
In this paper, an end-to-end convolutional neural network was designed to perform both foveated reconstruction and view synthesis using only 1.2% of the total light field data.
Abstract
Near-eye light field displays provide a solution to visual discomfort when using head mounted displays by presenting accurate depth and focal cues. However, light field HMDs require rendering the scene from a large number of viewpoints. This computational challenge of rendering sharp imagery of the foveal region and reproduce retinal defocus blur that correctly drives accommodation is tackled in this paper. We designed a novel end-to-end convolutional neural network that leverages human vision to perform both foveated reconstruction and view synthesis using only 1.2% of the total light field data. The proposed architecture comprises of log-polar sampling scheme followed by an interpolation stage and a convolutional neural network. To the best of our knowledge, this is the first attempt that synthesizes the entire light field from sparse RGB-D inputs and simultaneously addresses foveation rendering for computational displays. Our algorithm achieves fidelity in the fovea without any perceptible artifacts in the peripheral regions. The performance in fovea is comparable to the state-of-the-art view synthesis methods, despite using around 10x less light field data.

read more

Citations
More filters
Journal ArticleDOI

An integrative view of foveated rendering

TL;DR: Foveated rendering as mentioned in this paper adapts the image synthesis process to the user's gaze by exploiting the human visual system's limitations, in particular in terms of reduced acuity in peripheral vision, it strives to deliver high-quality visual experiences at very reduced computational, storage and transmission costs.
Journal ArticleDOI

2T-UNET: A Two-Tower UNet with Depth Clues for Robust Stereo Depth Estimation

TL;DR: The depth estimation problem is revisits, avoiding the explicit stereo matching step using a simple two-tower convolutional neural network, and the proposed algorithm is entitled 2T-UNet, which surpasses state-of-the-art monocular and stereo depth estimation methods on the challenging Scene dataset.
References
More filters
Journal ArticleDOI

Perceptually-guided foveation for light field displays

TL;DR: A content-adaptive importance model in the 4D ray space is formulated based on psychophysical experiments and theoretical analysis on visual and display bandwidths and verified by building a prototype light field display that can render only 16% -- 30% rays without compromising perceptual quality.
Journal ArticleDOI

Fast gaze-contingent optimal decompositions for multifocal displays

TL;DR: An efficient algorithm for optimal decompositions is presented, incorporating insights from vision science, and eye tracking can be used for adequate plane alignment with efficient image-based deformations, adjusting for both eye rotation and head movement relative to the display.
Proceedings ArticleDOI

DeepFocus: learned image synthesis for computational display

TL;DR: Deep-Focus is introduced, a generic, end-to-end trainable convolutional neural network designed to efficiently solve the full range of computational tasks for accommodation-supporting HMDs and is demonstrated to accurately synthesize defocus blur, focal stacks, multilayer decompositions, and multiview imagery using commonly available RGB-D images.
Proceedings ArticleDOI

An introduction to the log-polar mapping [image sampling]

TL;DR: In this article, a log-polar mapping of the eye's retinal image is proposed. But the mapping is not suitable for the human visual system and its main properties are unknown.
Related Papers (5)