Design, Implementation, and Evaluation of a Point Cloud Codec for Tele-Immersive Video
Summary (2 min read)
Introduction
- AVALANCHE photodiodes (APDs) and single-photonavalanche diodes , or Geiger-mode APDs, are widely used in optical telecommunications, imaging, and medical diagnostics, where high sensitivity to light in the visible or near-infrared (NIR) ranges is needed [1].
- Ge is attractive because it has been possible to develop processes allowing integration in CMOS, and low-quality Ge photodiodes have been demonstrated in telecom circuits [5], [6].
II. DEVICE DESIGN AND FABRICATION
- Fig. 1 shows a schematic of the fabrication process of the PureGaB Ge-on-Si arrays.
- First, a 30-nm thermal SiO2 is grown on the Si surface followed by a low-pressure CVD SiO2 layer deposition with a thickness of ∼1 μm.
GEOMETRICAL PARAMETERS OF THE THREE DIFFERENT PHOTODIODE PIXELS
- Etched on the Si surface by a mask that defined the Ge-on-Si photodiode areas.
- Since the diodes in the multidiode devices were separated by a 1-μm-wide oxide region, the actual Ge area decreases with the number of diodes while the Ge perimeter increases.
- Ge deposited on surrounding oxide regions from reaching the windows designed for the deposition of the As-doped Ge islands.
- Next, 800-nm Al/Si1% was sputtered and then removed over the photosensitive junctions by means of selective plasma etching of the Al layer to the oxide covering the PureGaB.
- This phenomenon was studied in detail in [7].
III. ELECTRICAL CHARACTERIZATION
- A continuous-flow cryostat system was used for achieving low temperatures.
- Vacuum conditions were maintained to be free from ice and frosting on the chip surface even for temperatures as low as 77 K achieved with liquid nitrogen cooling.
- Typical I–V characteristics are shown in Fig. 4 for both room temperature and 77 K measurements of all three device types.
- The perimeter of the hexa device was 3 times longer than that of the single-diode pixel, i.e., it became less likely that a perfect device would be found.
- The depletion width in reverse will quickly become larger than the 0.55 μm thickness of the Ge-islands and transverse the nonperfect interface with the Si.
IV. OPTICAL CHARACTERIZATION
- The Ge photodiodes were mounted in the vacuum chamber and cooled to 77 K.
- The incident optical power on the devices was attenuated, resulting in 9.8, 0.75, and 0.78 μW for the respective laser sources.
- The diameter of the incident beam spot was 0.3 cm.
- The photocurrent was measured using a computer-controlled HP Semiconductor Parameter Analyzer model 4145B.
- Neutral density filters were used to prevent light saturation and pileup.
SUMMARY OF OVERALL BEST PERFORMANCE COMPARED WITH
- The responsivity R as a function of reverse voltage Vx was calculated as R(Vx ) = (Iph,Vx − Idark,Vx )/P (2) where the photocurrent Iph,Vx and the dark current Idark,Vx are measured at Vx and P is the incident power at the surface of the pixel.
- Ge crystal but the leakage current is not yet influenced by impact ionization.
- The responsivity at 660 nm is mainly not reported for Ge photodiodes.
- This trend is corroborated by Fig. 10, which plots the maximum optical gain as a function of Vbd.
SUMMARY OF APD PERFORMANCE PARAMETERS FOR THE THREE PIXEL TYPES
- Inner ellipse for σ and the outer ellipse for 3σ spread.
- The mean values of maximum optical gain and breakdown, as well as the breakdown standard deviation, are listed in Table III.
- Just as the lower spread around high current levels of the reverse I–V characteristics of the hexa devices can be explained by the higher probability of perimeter imperfections, this also explains the lower spread with perimeter of the maximum gain to breakdown voltage relationship.
V. CONCLUSION
- The PureGaB Ge-on-Si photodiodes integrated in 300 × 1 pixel arrays were characterized at cryogenic temperatures for operation as proportional APDs.
- The main differentiating factor is a very different area and perimeter, while the electrical/optical performance is comparable with very high optical gain of up to 106 measured at 77 K, where the low-voltage dark current is lower than the measurement limit of 2.5 × 10−2 μA/cm2.
- Ge area but the largest perimeter with an average increase in the Ge thickness of ∼30% due mainly to V-groove formation.
- All in all, the PureGaB Ge-on-Si offers a very low-complexity CMOS-compatible means of fabricating uniform arrays of the NIR sensitive photodetectors that are operational in linear, avalanche, and even Geiger modes.
Did you find this useful? Give us your feedback
Citations
470 citations
Cites background or methods from "Design, Implementation, and Evaluat..."
...Previous compression solutions for volumetric visual representations either focused on computer-generated content [1], [4] or suffered from low spatial and temporal compression performance [5], [6] when...
[...]
...To have a baseline for determining target bitrates and distortions, a recent hybrid octree-image point cloud codec for tele-immersive video [6] was chosen as anchor....
[...]
...In [6], an extension to this framework was introduced, combining the octree-based codec with a common image codec for color attribute coding....
[...]
118 citations
95 citations
Cites background or methods or result from "Design, Implementation, and Evaluat..."
...Here, the rigid-body motion estimation is based on the Iterative Closest Points (ICP) algorithm [Paul J. Besl 1992], which has been also used for similar purpose in [Mekuria et al. 2017] 6⃝....
[...]
...Authors show that the method achieves significant gains in compression ratio when compared to the JPEGbased compression approach introduced in [Mekuria et al. 2017] 6⃝....
[...]
...…on the multi-view video coding algorithms introduced in [Merkle et al. 2007], which exploit both temporal and inter-view statistical dependencies under a prediction framework, several coding schemes targeting real-time compression, are proposed in [Mekuria et al. 2017] 6⃝ and [Lien et al. 2009] 7⃝....
[...]
...Besl 1992], which has been also used for similar purpose in [Mekuria et al. 2017] 6 ⃝....
[...]
...2007], which exploit both temporal and inter-view statistical dependencies under a prediction framework, several coding schemes targeting real-time compression, are proposed in [Mekuria et al. 2017] 6 ⃝ and [Lien et al....
[...]
95 citations
95 citations
Additional excerpts
...the efficient compression of this representation [1]–[5]....
[...]
References
4,501 citations
"Design, Implementation, and Evaluat..." refers methods in this paper
...The vnn_deg is efficiently computed via a K -d tree in L2 distance norm based on algorithms available in [4]...
[...]
683 citations
"Design, Implementation, and Evaluat..." refers methods in this paper
...3) Multiview Plus Depth Compression: Multiview plus depth representation was considered for storing video and depth maps from multiple cameras [7], [8]....
[...]
...Existing video coding standards, such as Advanced Video Coding (AVC) Multi View Video (MVV) [7] and MVV-D [8], can support these functionalities via techniques from (depth) image-based rendering (DIBR)....
[...]
406 citations
"Design, Implementation, and Evaluat..." refers background or methods in this paper
...Note that we made comparisons with the available real-time point cloud codec in [3] with octree composition and DPCM coding....
[...]
...Surprisingly, the JPEG color coding method does not introduce significant subjective distortion, as this method codes at up to ten times lower bitrates compared with the 8-b DPCM color data....
[...]
...The real-time results for intra encoding for different LoDs and both proposed (JPEG) and original DPCM-based color coding are shown in Fig....
[...]
...to [11] and [12], we develop a position coder that uses an...
[...]
...In Section V-E, we will show in the subjective studies that the color quality degradation introduced by the color coding method is negligible even compared with 8-b DPCM-based coding....
[...]
365 citations
"Design, Implementation, and Evaluat..." refers methods in this paper
...3) Multiview Plus Depth Compression: Multiview plus depth representation was considered for storing video and depth maps from multiple cameras [7], [8]....
[...]
...Existing video coding standards, such as Advanced Video Coding (AVC) Multi View Video (MVV) [7] and MVV-D [8], can support these functionalities via techniques from (depth) image-based rendering (DIBR)....
[...]
341 citations
"Design, Implementation, and Evaluat..." refers background or methods in this paper
...we made comparisons with the available real-time point cloud codec in [3] with octree composition and DPCM coding....
[...]
...The work in [3] introduced a real-time octree-based codec that can also exploit temporal redundancies by XOR operations on the octree byte stream....
[...]
...Compression of 3D point clouds has received significant attention in recent years [3]–[6]....
[...]
...Instead, we follow a modified approach, taken from [3] (based on a carry-less-byte-based range coder), which we applied to the different decodable LoDs....
[...]
Related Papers (5)
Frequently Asked Questions (14)
Q2. What is the common method of compression for 3D objects?
Mesh Compression: 3D objects are often coded as 3DMeshes, for which a significant number of compression methods have been developed.
Q3. What is the metric used to evaluate the point cloud quality?
To evaluate the point cloud quality, the authors deploy a full reference quality metric that combines common practices from 3D mesh and Video Compression: a PSNR metric based on point to point symmetric root mean square distances.
Q4. What is the hardware used in the experiments?
The hardware used in the experiments are a Dell Precision M6800 PC with intel core i7-4810MQ 2,8 MHz CPU and 16.0 GB of Ram running 64 win7 Operating system, A Dell Precision T3210 (Xeon 3.7 Ghz), and a custom built system with i7 3.2 GhZ running Ubuntu Linux.
Q5. What is the main aim of the experiments?
The main aim of their experiments is to check that the developed codec does not introduce significant extra subjective distortions.
Q6. How did the authors speed up the inter-predictive encoding?
In addition, the authors have been able to speed up the inter-predictive encoding significantly by parallelizing its execution based on OpenMP for multi core intel architectures (we measured up to 20 % improvement on the windows platforms).
Q7. what is the value of point clouds in augmented reality?
The results show that the degradation introduced by the codecs is negligible and that the point cloud has an added value in terms of “feeling together” compared to simple avatar representations, highlighting the importance of point clouds in these applications and the importance of work on compression of point clouds and its standardization for immersive and augmented reality systems.
Q8. How is the color quality degradation in the subjective studies?
In V.E the authors will show in the subjective studies that the color quality degradation introduced by the color coding method is negligible even compared to 8-bits DPCM based coding.
Q9. What can happen in the traversal of macroblocks in M_p?
in the traversal of macroblocks in M_p (2), operations (5) and especially (6) are most computationally intensive but can happen using parallel computation.
Q10. What is the method for coding 3D objects?
While these methods are promising, it seems that methods based on 3D Point cloudscan result in coding with even less overhead and more flexibleprogressive rendering capabilities, as the format is simpler to acquire and process.
Q11. What is the percentage that is shared and converged?
The percentage thatis shared and converged is shown in red, and relates largely to the bitrate savings (which is over 30% for the Dimitrios dataset that has the largest percentage of shared blocks).
Q12. What are the quality evaluation metrics for point cloud codecs?
These quality evaluation metrics are well aligned with existing practice in mesh and video compression and are recommended for evaluation of point cloud codecs [25].
Q13. What is the difference between 3Dimmersive and augmented 3D video coding?
In 3Dimmersive and augmented 3D Video coding, having both lowencoder and decoder complexity are important (analogous to video coding in video conferencing systems compared tovideo on demand).
Q14. What is the coding algorithm used to compute the bounding box?
As shown in Fig 5 the bounding box of an intra frame is expanded from the BB_IE (x_min – bb_exp, y_min – bb_exp, z_min –bb_exp) to upper corner (x_max+ bb_exp, y_max+ bb_exp,z_max+bb_exp), where bb_exp was computed from δ and the ranges of the original bounding box.