scispace - formally typeset
Search or ask a question

Showing papers by "Qualcomm published in 2016"


Proceedings Article
19 Jun 2016
TL;DR: This paper proposes a quantizer design for fixed point implementation of DCNs, formulate and solve an optimization problem to identify optimal fixed point bit-width allocation across DCN layers, and demonstrates that fine-tuning can further enhance the accuracy of fixed point DCNs beyond that of the original floating point model.
Abstract: In recent years increasingly complex architectures for deep convolution networks (DCNs) have been proposed to boost the performance on image recognition tasks. However, the gains in performance have come at a cost of substantial increase in computation and model storage resources. Fixed point implementation of DCNs has the potential to alleviate some of these complexities and facilitate potential deployment on embedded hardware. In this paper, we propose a quantizer design for fixed point implementation of DCNs. We formulate and solve an optimization problem to identify optimal fixed point bit-width allocation across DCN layers. Our experiments show that in comparison to equal bitwidth settings, the fixed point DCNs with optimized bit width allocation offer > 20%reduction in the model size without any loss in accuracy on CIFAR-10 benchmark. We also demonstrate that fine-tuning can further enhance the accuracy of fixed point DCNs beyond that of the original floating point model. In doing so, we report a new state-of-the-art fixed point performance of 6.78% error-rate on CIFAR-10 benchmark.

619 citations


Proceedings ArticleDOI
05 Jun 2016
TL;DR: This work proposes Pinatubo, a Processing In Non-volatile memory ArchiTecture for bUlk Bitwise Operations, which redesigns the read circuitry so that it can compute the bitwise logic of two or more memory rows very efficiently, and support one-step multi-row operations.
Abstract: Processing-in-memory (PIM) provides high bandwidth, massive parallelism, and high energy efficiency by implementing computations in main memory, therefore eliminating the overhead of data movement between CPU and memory. While most of the recent work focused on PIM in DRAM memory with 3D die-stacking technology, we propose to leverage the unique features of emerging non-volatile memory (NVM), such as resistance-based storage and current sensing, to enable efficient PIM design in NVM. We propose Pinatubo1, a Processing In Non-volatile memory ArchiTecture for bUlk Bitwise Operations. Instead of integrating complex logic inside the cost-sensitive memory, Pinatubo redesigns the read circuitry so that it can compute the bitwise logic of two or more memory rows very efficiently, and support one-step multi-row operations. The experimental results on data intensive graph processing and database applications show that Pinatubo achieves a ∼500 x speedup, ∼28000x energy saving on bitwise operations, and 1.12× overall speedup, 1.11× overall energy saving over the conventional processor.

389 citations


Journal ArticleDOI
TL;DR: The more advanced 3D video extension, 3D-HEVC, targets a coded representation consisting of multiple views and associated depth maps, as required for generating additional intermediate views inAdvanced 3D displays.
Abstract: The High Efficiency Video Coding (HEVC) standard has recently been extended to support efficient representation of multiview video and depth-based 3D video formats. The multiview extension, MV-HEVC, allows efficient coding of multiple camera views and associated auxiliary pictures, and can be implemented by reusing single-layer decoders without changing the block-level processing modules since block-level syntax and decoding processes remain unchanged. Bit rate savings compared with HEVC simulcast are achieved by enabling the use of inter-view references in motion-compensated prediction. The more advanced 3D video extension, 3D-HEVC, targets a coded representation consisting of multiple views and associated depth maps, as required for generating additional intermediate views in advanced 3D displays. Additional bit rate reduction compared with MV-HEVC is achieved by specifying new block-level video coding tools, which explicitly exploit statistical dependencies between video texture and depth and specifically adapt to the properties of depth maps. The technical concepts and features of both extensions are presented in this paper.

385 citations


Journal ArticleDOI
TL;DR: Scalable High efficiency Video Coding further provides spatial, signal-to-noise ratio, bit depth, and color gamut scalability functionalities, as well as combinations of any of these.
Abstract: This paper provides an overview of Scalable High efficiency Video Coding (SHVC), the scalable extensions of the High Efficiency Video Coding (HEVC) standard, published in the second version of HEVC. In addition to the temporal scalability already provided by the first version of HEVC, SHVC further provides spatial, signal-to-noise ratio, bit depth, and color gamut scalability functionalities, as well as combinations of any of these. The SHVC architecture design enables SHVC implementations to be built using multiple repurposed single-layer HEVC codec cores, with the addition of interlayer reference picture processing modules. The general multilayer high-level syntax design common to all multilayer HEVC extensions, including SHVC, MV-HEVC, and 3D HEVC, is described. The interlayer reference picture processing modules, including texture and motion resampling and color mapping, are also described. Performance comparisons are provided for SHVC versus simulcast HEVC and versus the scalable video coding extension to H.264/advanced video coding.

334 citations


Journal ArticleDOI
TL;DR: This paper compares three candidate large-scale propagation path loss models for use over the entire microwave and millimeter-wave (mmWave) radio spectrum and shows the CI model with a 1-m reference distance is suitable for outdoor environments, while the CIF model is more appropriate for indoor modeling.
Abstract: This paper compares three candidate large-scale propagation path loss models for use over the entire microwave and millimeter-wave (mmWave) radio spectrum: the alpha–beta–gamma (ABG) model, the close-in (CI) free-space reference distance model, and the CI model with a frequency-weighted path loss exponent (CIF). Each of these models has been recently studied for use in standards bodies such as 3rd Generation Partnership Project (3GPP) and for use in the design of fifth-generation wireless systems in urban macrocell, urban microcell, and indoor office and shopping mall scenarios. Here, we compare the accuracy and sensitivity of these models using measured data from 30 propagation measurement data sets from 2 to 73 GHz over distances ranging from 4 to 1238 m. A series of sensitivity analyses of the three models shows that the four-parameter ABG model underpredicts path loss when relatively close to the transmitter, and overpredicts path loss far from the transmitter, and that the physically based two-parameter CI model and three-parameter CIF model offer computational simplicity, have very similar goodness of fit (i.e., the shadow fading standard deviation), exhibit more stable model parameter behavior across frequencies and distances, and yield smaller prediction error in sensitivity tests across distances and frequencies, when compared to the four-parameter ABG model. Results show the CI model with a 1-m reference distance is suitable for outdoor environments, while the CIF model is more appropriate for indoor modeling. The CI and CIF models are easily implemented in existing 3GPP models by making a very subtle modification—by replacing a floating non-physically based constant with a frequency-dependent constant that represents free-space path loss in the first meter of propagation. This paper shows this subtle change does not change the mathematical form of existing ITU/3GPP models and offers much easier analysis, intuitive appeal, better model parameter stability, and better accuracy in sensitivity tests over a vast range of microwave and mmWave frequencies, scenarios, and distances, while using a simpler model with fewer parameters.

288 citations


Proceedings ArticleDOI
15 May 2016
TL;DR: This document describes an initial 3D channel model which includes a baseline model for incorporating path loss, shadow fading, line of sight probability, penetration and blockage models for the typical scenarios of 5G channel models for bands up to 100 GHz.
Abstract: For the development of new 5G systems to operate in bands up to 100 GHz, there is a need for accurate radio propagation models at these bands that currently are not addressed by existing channel models developed for bands below 6 GHz. This document presents a preliminary overview of 5G channel models for bands up to 100 GHz. These have been derived based on extensive measurement and ray tracing results across a multitude of frequencies from 6 GHz to 100 GHz, and this document describes an initial 3D channel model which includes: 1) typical deployment scenarios for urban microcells (UMi) and urban macrocells (UMa), and 2) a baseline model for incorporating path loss, shadow fading, line of sight probability, penetration and blockage models for the typical scenarios. Various processing methodologies such as clustering and antenna decoupling algorithms are also presented.

281 citations


Proceedings ArticleDOI
15 May 2016
TL;DR: This paper presents and compares two candidate large-scale propagation path loss models, the alpha-beta-gamma (ABG) model and the close-in (CI) free space reference distance model, for the design of fifth generation (5G) wireless communication systems in urban micro- and macro-cellular scenarios.
Abstract: This paper presents and compares two candidate large-scale propagation path loss models, the alpha-beta-gamma (ABG) model and the close-in (CI) free space reference distance model, for the design of fifth generation (5G) wireless communication systems in urban micro- and macro-cellular scenarios. Comparisons are made using the data obtained from 20 propagation measurement campaigns or ray- tracing studies from 2 GHz to 73.5 GHz over distances ranging from 5 m to 1429 m. The results show that the one-parameter CI model has a very similar goodness of fit (i.e., the shadow fading standard deviation) in both line-of-sight and non-line-of-sight environments, while offering substantial simplicity and more stable behavior across frequencies and distances, as compared to the three-parameter ABG model. Additionally, the CI model needs only one very subtle and simple modification to the existing 3GPP floating-intercept path loss model (replacing a constant with a close-in free space reference value) in order to provide greater simulation accuracy, more simplicity, better repeatability across experiments, and higher stability across a vast range of frequencies.

269 citations


Posted Content
TL;DR: In this article, the authors compared three large-scale propagation path loss models for use over the entire microwave and millimeter-wave (mmWave) radio spectrum: the alpha-beta-gamma (ABG) model, the close-in (CI) free space reference distance model, and the CI model with a frequency-weighted path loss exponent (CIF).
Abstract: This paper compares three candidate large-scale propagation path loss models for use over the entire microwave and millimeter-wave (mmWave) radio spectrum: the alpha-beta-gamma (ABG) model, the close-in (CI) free space reference distance model, and the CI model with a frequency-weighted path loss exponent (CIF). Each of these models have been recently studied for use in standards bodies such as 3GPP, and for use in the design of fifth generation (5G) wireless systems in urban macrocell, urban microcell, and indoor office and shopping mall scenarios. Here we compare the accuracy and sensitivity of these models using measured data from 30 propagation measurement datasets from 2 GHz to 73 GHz over distances ranging from 4 m to 1238 m. A series of sensitivity analyses of the three models show that the physically-based two-parameter CI model and three-parameter CIF model offer computational simplicity, have very similar goodness of fit (i.e., the shadow fading standard deviation), exhibit more stable model parameter behavior across frequencies and distances, and yield smaller prediction error in sensitivity testing across distances and frequencies, when compared to the four-parameter ABG model. Results show the CI model with a 1 m close-in reference distance is suitable for outdoor environments, while the CIF model is more appropriate for indoor modeling. The CI and CIF models are easily implemented in existing 3GPP models by making a very subtle modification -- by replacing a floating non-physically based constant with a frequency-dependent constant that represents free space path loss in the first meter of propagation.

255 citations


Journal ArticleDOI
TL;DR: An overview of the technical features and characteristics of the current HEVC-SCC test model and related coding tools, including intra-block copy, palette mode, adaptive color transform, and adaptive motion vector resolution are provided.
Abstract: A screen content coding (SCC) extension to High Efficiency Video Coding (HEVC) is currently under development by the Joint Collaborative Team on Video Coding, which is a joint effort from the ITU-T Video Coding Experts Group and the ISO/IEC Moving Picture Experts Group. The main goal of the HEVC-SCC standardization effort is to enable significantly improved compression performance for videos containing a substantial amount of still or moving rendered graphics, text, and animation rather than, or in addition to, camera-captured content. This paper provides an overview of the technical features and characteristics of the current HEVC-SCC test model and related coding tools, including intra-block copy, palette mode, adaptive color transform, and adaptive motion vector resolution. The performance of the SCC extension is compared against existing standards in terms of bitrate savings at equal distortion.

247 citations


Journal ArticleDOI
TL;DR: An overview of several features included in 3GPP to accommodate the needs of M2M communications, including changes in the physical layer such as enhanced machine type communications, and new MAC and higher-layer procedures provided by extended discontinuous reception are presented.
Abstract: The broad connection of devices to the Internet, known as the IoT or M2M, requires lowcost power-efficient global connectivity services. New physical layer solutions, MAC procedures, and network architectures are needed to evolve the current LTE cellular systems to meet the demands of IoT services. Several steps have been taken under the 3GPP to accomplish these objectives and are included in the upcoming 3GPP LTE standards release (3GPP Release 13). In this tutorial article, we present an overview of several features included in 3GPP to accommodate the needs of M2M communications, including changes in the physical layer such as enhanced machine type communications, and new MAC and higher-layer procedures provided by extended discontinuous reception. We also briefly discuss the narrowband IoT, which is in the development stage with a target completion date of June 2016.

186 citations


Patent
04 Jan 2016
TL;DR: In this article, a software/firmware mechanism can read the various hardware counters once the committed instruction counter reaches a threshold value and divide the value of the first hardware counter by a value of second hardware counter to measure a stall fraction during a current program execution phase.
Abstract: The disclosure generally relates to dynamic clock and voltage scaling (DCVS) based on program phase. For example, during each program phase, a first hardware counter may count each cycle where a dispatch stall occurs and an oldest instruction in a load queue is a last-level cache miss, a second hardware counter may count total cycles, and a third hardware counter may count committed instructions. Accordingly, a software/firmware mechanism may read the various hardware counters once the committed instruction counter reaches a threshold value and divide a value of the first hardware counter by a value of the second hardware counter to measure a stall fraction during a current program execution phase. The measured stall fraction can then be used to predict a stall fraction in a next program execution phase such that optimal voltage and frequency settings can be applied in the next phase based on the predicted stall fraction.

Posted Content
Se Rim Park, Jinwon Lee1
TL;DR: The proposed network, Redundant Convolutional Encoder Decoder (R-CED), demonstrates that a convolutional network can be 12 times smaller than a recurrent network and yet achieves better performance, which shows its applicability for an embedded system: the hearing aids.
Abstract: In hearing aids, the presence of babble noise degrades hearing intelligibility of human speech greatly. However, removing the babble without creating artifacts in human speech is a challenging task in a low SNR environment. Here, we sought to solve the problem by finding a `mapping' between noisy speech spectra and clean speech spectra via supervised learning. Specifically, we propose using fully Convolutional Neural Networks, which consist of lesser number of parameters than fully connected networks. The proposed network, Redundant Convolutional Encoder Decoder (R-CED), demonstrates that a convolutional network can be 12 times smaller than a recurrent network and yet achieves better performance, which shows its applicability for an embedded system: the hearing aids.

Journal ArticleDOI
TL;DR: A physical interpretation between the RSV-type beamformer structures and the angles of departure/arrival of the dominant path(s) capturing the scattering environment provides a theoretical underpinning to the emerging interest on directional beamforming approaches that are less sensitive to small path length changes.
Abstract: Millimeter-wave (mmW) multi-input multi-output (MIMO) systems have gained increasing traction toward the goal of meeting the high data-rate requirements in next-generation wireless systems. The focus of this work is on low-complexity beamforming approaches for initial user equipment (UE) discovery in such systems. Toward this goal, we first note the structure of the optimal beamformer with per-antenna gain and phase control and establish the structure of good beamformers with per-antenna phase-only control. Learning these right singular vector (RSV)-type beamforming structures in mmW systems is fraught with considerable complexities such as the need for a non-broadcast system design, the sensitivity of the beamformer approximants to small path length changes, inefficiencies due to power amplifier backoff, etc. To overcome these issues, we establish a physical interpretation between the RSV-type beamformer structures and the angles of departure/arrival (AoD/AoA) of the dominant path(s) capturing the scattering environment. This physical interpretation provides a theoretical underpinning to the emerging interest on directional beamforming approaches that are less sensitive to small path length changes. While classical approaches for direction learning such as MUltiple SIgnal Classification (MUSIC) have been well-understood, they suffer from many practical difficulties in a mmW context such as a non-broadcast system design and high computational complexity. A simpler broadcast-based solution for mmW systems is the adaptation of limited feedback-type directional codebooks for beamforming at the two ends. We establish fundamental limits for the best beam broadening codebooks and propose a construction motivated by a virtual subarray architecture that is within a couple of dB of the best tradeoff curve at all useful beam broadening factors. We finally provide the received ${\text{SNR}}$ loss-UE discovery latency tradeoff with the proposed beam broadening constructions. Our results show that users with a reasonable link margin can be quickly discovered by the proposed design with a smooth roll-off in performance as the link margin deteriorates. While these designs are poorer in performance than the RSV learning approaches or MUSIC for cell-edge users, their low-complexity that leads to a broadcast system design makes them a useful candidate for practical mmW systems.

Journal ArticleDOI
TL;DR: This paper presents the first linear bulk CMOS power amplifier (PA) targeting low-power fifth-generation (5G) mobile user equipment integrated phased array transceivers and achieves the highest measured PAE values among published K-and Ka-band CMOS PAs.
Abstract: This paper presents the first linear bulk CMOS power amplifier (PA) targeting low-power fifth-generation (5G) mobile user equipment integrated phased array transceivers. The output stage of the PA is first optimized for power-added efficiency (PAE) at a desired error vector magnitude (EVM) and range given a challenging 5G uplink use case scenario. Then, inductive source degeneration in the optimized output stage is shown to enable its embedding into a two-stage transformer-coupled PA; by broadening interstage impedance matching bandwidth and helping to reduce distortion. Designed and fabricated in 1P7M 28 nm bulk CMOS and using a 1 V supply, the PA achieves +4.2 dBm/9% measured $P_{\text {out}}$ /PAE at −25 dBc EVM for a 250 MHz-wide 64-quadrature amplitude modulation orthogonal frequency division multiplexing signal with 9.6 dB peak-to-average power ratio. The PA also achieves 35.5%/10% PAE for continuous wave signals at saturation/9.6 dB back-off from saturation. To the best of the authors’ knowledge, these are the highest measured PAE values among published ${K}$ - and Ka -band CMOS PAs.

Journal ArticleDOI
TL;DR: IBC, or intra picture block compensation, also named current picture referencing (CPR) in HEVC SCC, is introduced and discussed, and results show that up to 50% BD rate reduction in all intra coding can be achieved with intra block copy enabled, compared to the HEVC reference encoder without this tool.
Abstract: With the emerging applications such as online gaming and Wi-Fi display, screen content video, including computer generated text, graphics and animations, becomes more popular than ever. Traditional video coding technologies typically were developed based on models that fit into natural, camera-captured video. The distinct characteristics exhibited between these two types of contents necessitate the exploration of coding efficiency improvement given that new tools can be developed specially for screen content video. The HEVC Screen Content Coding Extensions (HEVC SCC) have been developed to incorporate such new coding tools in order to achieve better compression efficiency. In this paper, intra block copy (IBC, or intra picture block compensation), also named current picture referencing (CPR) in HEVC SCC, is introduced and discussed. This tool is very efficient for coding of screen content video in that repeated patterns in text and graphics rich content occur frequently within the same picture. Having a previously reconstructed block with equal or similar pattern as a predictor can effectively reduce the prediction error and therefore improve coding efficiency. Simulation results show that up to 50% BD rate reduction in all intra coding can be achieved with intra block copy enabled, compared to the HEVC reference encoder without this tool. Significant BD rate reductions for other coding configurations are also observed.

Posted Content
TL;DR: The authors propose a character composition model, tweet2vec, which finds vector-space representations of whole tweets by learning complex, non-local dependencies in character sequences, which outperforms a word-level baseline at predicting user-annotated hashtags associated with the posts, doing significantly better when the input contains many out-of-vocabulary words or unusual character sequences.
Abstract: Text from social media provides a set of challenges that can cause traditional NLP approaches to fail. Informal language, spelling errors, abbreviations, and special characters are all commonplace in these posts, leading to a prohibitively large vocabulary size for word-level approaches. We propose a character composition model, tweet2vec, which finds vector-space representations of whole tweets by learning complex, non-local dependencies in character sequences. The proposed model outperforms a word-level baseline at predicting user-annotated hashtags associated with the posts, doing significantly better when the input contains many out-of-vocabulary words or unusual character sequences. Our tweet2vec encoder is publicly available.

Journal ArticleDOI
TL;DR: A novel mathematical model based on queuing theory and stochastic geometry is proposed, which extends the Matérn hard-core type-II process with a discrete and nonuniform distribution, which is used to derive the temporal states of backoff counters, leading to a more accurate approximation to real network dynamics.
Abstract: Vehicle-to-vehicle safety communications based on the dedicated short-range communication technology have the potential to enable a set of applications that help avoid traffic accidents. The performance of these applications, largely affected by the reliability of communication links, stringently ties back to the MAC and PHY layer design, which has been standardized as IEEE 802.11p. The link reliabilities depend on the signal-to-interference-plus-noise ratio (SINR), which, in turn, depends on the locations and transmit power values of the transmitting nodes. Hence, an accurate network model needs to take into account the network geometry. For such geometric models, however, there is a lack of mathematical understanding of the characteristics and performance of IEEE 802.11p. Important questions such as the scalability performance of IEEE 802.11p have to be answered by simulations, which can be very time consuming and provide limited insights to future protocol design. In this paper, we investigate the performance of IEEE 802.11p by proposing a novel mathematical model based on queuing theory and stochastic geometry. In particular, we extend the Matern hard-core type-II process with a discrete and nonuniform distribution, which is used to derive the temporal states of backoff counters. By doing so, concurrent transmissions from nodes within the carrier sensing ranges of each other are taken into account, leading to a more accurate approximation to real network dynamics. A comparison with Network Simulator 2 (ns2) simulations shows that our model achieves a good approximation in networks with different densities.

Journal ArticleDOI
TL;DR: A Ka-band highly linear power amplifier (PA) is implemented in 28-nm bulk CMOS technology using a deep class-AB PA topology with appropriate harmonic control circuit, highly linear and efficient PAs are designed at millimeter-wave band.
Abstract: A Ka-band highly linear power amplifier (PA) is implemented in 28-nm bulk CMOS technology. Using a deep class-AB PA topology with appropriate harmonic control circuit, highly linear and efficient PAs are designed at millimeter-wave band. This PA architecture provides a linear PA operation close to the saturated power. Also elaborated harmonic tuning and neutralization techniques are used to further improve the transistor gain and stability. A two-stack PA is designed for higher gain and output power than a common source (CS) PA. Additionally, average power tracking (APT) is applied to further reduce the power consumption at a low power operation and, hence, extend battery life. Both the PAs are tested with two different signals at 28.5 GHz; they are fully loaded long-term evolution (LTE) signal with 16-quadrature amplitude modulation (QAM), a 7.5-dB peak-to-average power ratio (PAPR), and a 20-MHz bandwidth (BW), and a wireless LAN (WLAN) signal with 64-QAM, a 10.8-dB PAPR, and an 80-MHz BW. The CS/two-stack PAs achieve power-added efficiency (PAE) of 27%/25%, error vector magnitude (EVM) of 5.17%/3.19%, and adjacent channel leakage ratio (ACLR $_{\mathrm{ E-UTRA}}$ ) of −33/−33 dBc, respectively, with an average output power of 11/14.6 dBm for the LTE signal. For the WLAN signal, the CS/2-stack PAs achieve the PAE of 16.5%/17.3%, and an EVM of 4.27%/4.21%, respectively, at an average output power of 6.8/11 dBm.

Proceedings ArticleDOI
11 May 2016
TL;DR: A character composition model, tweet2vec, is proposed, which finds vector-space representations of whole tweets by learning complex, non-local dependencies in character sequences.
Abstract: Text from social media provides a set of challenges that can cause traditional NLP approaches to fail. Informal language, spelling errors, abbreviations, and special characters are all commonplace in these posts, leading to a prohibitively large vocabulary size for word-level approaches. We propose a character composition model, tweet2vec, which finds vector-space representations of whole tweets by learning complex, non-local dependencies in character sequences. The proposed model outperforms a word-level baseline at predicting user-annotated hashtags associated with the posts, doing significantly better when the input contains many out-of-vocabulary words or unusual character sequences. Our tweet2vec encoder is publicly available.

Journal ArticleDOI
TL;DR: The technical aspects of HEVC RExt are presented, including a discussion of RExt profiles, tools, applications, and experimental results for a performance comparison with previous relevant coding technology are provided.
Abstract: The Range Extensions (RExt) of the High Efficiency Video Coding (HEVC) standard have recently been approved by both ITU-T and ISO/IEC. This set of extensions targets video coding applications in areas including content acquisition, postproduction, contribution, distribution, archiving, medical imaging, still imaging, and screen content. In addition to the functionality of HEVC Version 1, RExt provide support for monochrome, 4:2:2, and 4:4:4 chroma sampling formats as well as increased sample bit depths beyond 10 bits per sample. This extended functionality includes new coding tools with a view to provide additional coding efficiency, greater flexibility, and throughput at high bit depths/rates. Improved lossless, near-lossless, and very high bit-rate coding is also a part of the RExt scope. This paper presents the technical aspects of HEVC RExt, including a discussion of RExt profiles, tools, applications, and provides experimental results for a performance comparison with previous relevant coding technology. When compared with the High 4:4:4 Predictive Profile of H.264/Advanced Video Coding (AVC), the corresponding HEVC 4:4:4 RExt profile provides up to $\sim 25$ %, $\sim 32$ %, and $\sim 36$ % average bit-rate reduction at the same PSNR quality level for intra, random access, and low delay configurations, respectively.

Patent
08 Feb 2016
TL;DR: In this paper, an apparatus includes a substrate package and a three-dimensional antenna structure formed in the substrate package, and the 3D antenna structure includes multiple substructures to enable the antenna structure to operate as a beamforming antenna.
Abstract: An apparatus includes a substrate package and a three dimensional (3D) antenna structure formed in the substrate package. The 3D antenna structure includes multiple substructures to enable the 3D antenna structure to operate as a beam-forming antenna. Each of the multiple substructures has a slanted-plate configuration or a slanted-loop configuration.

Proceedings ArticleDOI
23 May 2016
TL;DR: The measurements indicate that the smaller wavelengths introduce an increased sensitivity of the propagation models to the scale of the environment and show some frequency dependence of the path loss as well as increased occurrence of blockage.
Abstract: Future mobile communications systems are likely to be very different to those of today with new service innovations driven by increasing data traffic demand, increasing processing power of smart devices and new innovative applications. To meet these service demands the telecommunications industry is converging on a common set of 5G requirements which includes network speeds as high as 10 Gbps, cell edge rate greater than 100 Mbps, and latency of less than 1 msec. To reach these 5G requirements the industry is looking at new spectrum bands in the range up to 100 GHz where there is spectrum availability for wide bandwidth channels. For the development of new 5G systems to operate in bands up to 100 GHz there is a need for accurate radio propagation models which are not addressed by existing channel models developed for bands below 6 GHz. This paper presents a preliminary overview of the 5G channel models for bands up to 100 GHz in indoor offices and shopping malls, derived from extensive measurements across a multitude of bands. These studies have found some extensibility of the existing 3GPP models (e.g. 3GPP TR36.873) to the higher frequency bands up to 100 GHz. The measurements indicate that the smaller wavelengths introduce an increased sensitivity of the propagation models to the scale of the environment and show some frequency dependence of the path loss as well as increased occurrence of blockage. Further, the penetration loss is highly dependent on the material and tends to increase with frequency. The small-scale characteristics of the channel such as delay spread and angular spread and the multipath richness is somewhat similar over the frequency range, which is encouraging for extending the existing 3GPP models to the wider frequency range. Further work will be carried out to complete these models, but this paper presents the first steps for an initial basis for the model development.

Journal ArticleDOI
TL;DR: An overhearing-based protocol that utilizes the overheard interference as useful side information in the receiver design is developed and the optimum relay beamforming matrix maximizing the bottleneck of the achievable rates of the two information flows is developed subject to a finite power constraint.
Abstract: The cellular communications scenario involving the coexistence of the one-hop direct transmission and the two-hop relaying is studied in this paper. In contrast to conventional cellular systems featuring orthogonal information flows served by decoupled channel resources via, e.g., time division, we propose a novel protocol in which two information flows are served simultaneously via the shared channel resource, thus constituting a spectral-efficient solution for cellular communications. On the other hand, however, an inevitable issue associated with the proposed protocol is the interflow interference, which may lead to serious deterioration on both information flows. To tackle this issue, we develop an overhearing-based protocol that utilizes the overheard interference as useful side information in the receiver design. Specifically, depending on the interference levels, an adaptive linear minimum mean squared error (MMSE) and nonlinear MMSE successive interference cancellation (SIC) receiver exploiting the overheard interference at the direct mobile terminal is developed. To balance between the two information flows, we develop the asymptotically optimum superposition coding at the base station (BS) in the high-power regime. Furthermore, the optimum relay beamforming matrix maximizing the bottleneck of the achievable rates of the two information flows is developed subject to a finite power constraint. Finally, simulations demonstrate a remarkable throughput gain over the conventional cellular systems.

Journal ArticleDOI
TL;DR: Both the fundamental performance limits of cache networks and the practical challenges that need to be overcome in real-life scenarios are discussed.
Abstract: Caching is an essential technique to improve throughput and latency in a vast variety of applications. The core idea is to duplicate content in memories distributed across the network, which can then be exploited to deliver requested content with less congestion and delay. The traditional role of cache memories is to deliver the maximal amount of requested content locally rather than from a remote server. While this approach is optimal for single-cache systems, it has recently been shown to be significantly suboptimal for systems with multiple caches (i.e., cache networks). Instead, cache memories should be used to enable a coded multicasting gain. In this article, we survey these recent developments. We discuss both the fundamental performance limits of cache networks and the practical challenges that need to be overcome in real-life scenarios.

Journal ArticleDOI
TL;DR: Numerical results show that the performance of the proposed algorithm is near-optimal and the power consumption is significantly reduced compared with both the conventional OFDM-based frequency division multiple access and the static NOMA resource allocation schemes.
Abstract: In this letter, resource allocation in a downlink orthogonal frequency division multiplexing (OFDM)-based non-orthogonal multiple access (NOMA) system is studied. We investigate an optimization problem of minimizing the total transmit power under quality of service requirements, by jointly searching for optimal subcarrier assignments of each user and power allocations over subcarriers. We first present an optimal solution of power allocation with fixed subcarrier assignments, based on which we next propose a low-complexity algorithm of jointly optimizing the subcarrier assignments and power allocations. Numerical results show that the performance of the proposed algorithm is near-optimal and the power consumption is significantly reduced compared with both the conventional OFDM-based frequency division multiple access and the static NOMA resource allocation schemes.

Journal ArticleDOI
TL;DR: The 3-D imaging capability combined with the sensor's sensitivity to the acoustic properties of the tissue translates into excellent robustness against spoofing attacks.
Abstract: A fully integrated 3-D ultrasonic fingerprint sensor-on-a-chip is presented. The device consists of a $110\times 56$ piezoelectric micromachined ultrasonic transducer (PMUT) array bonded at the wafer level to custom readout electronics fabricated in a 180-nm CMOS process with a HV (24 V) transistor option. With the 24 V driving signal strength, the sensor consumes 280 $\mu \text{J}$ to image a 4.73 mm $\times3.24$ mm section of a fingerprint at a rate of 380 fps. A wakeup mode that detects the presence of a finger at 4 fps and dissipates 10 $\mu \text{W}$ allows the proposed sensor to double as a power switch. The sensor is capable of imaging both the surface epidermal and subsurface dermal fingerprints and is insensitive to contaminations, including perspiration or oil. The 3-D imaging capability combined with the sensor’s sensitivity to the acoustic properties of the tissue translates into excellent robustness against spoofing attacks.

Proceedings ArticleDOI
Xin Zhao1, Jianle Chen1, Marta Karczewicz1, Li Zhang1, Xiang Li1, Chien Wei-Jung1 
01 Mar 2016
TL;DR: To further improve the transform efficiency, an Enhanced Multiple Transform (EMT) scheme is proposed and has been implemented on top of High-Efficiency Video Coding (HEVC) reference software, and significant coding gain has been verified.
Abstract: The Discrete Cosine Transform (DCT), and in particular the DCT type II, has been widely used for image and video compression. Although DCT efficiently approximates the optimal Karhunen–Loeve transform under first-order Markov conditions with low complexity, the energy packing efficiency is still limited since a fixed transform cannot always capture the highly dynamic statistics of natural video content. In this paper, to further improve the transform efficiency, an Enhanced Multiple Transform (EMT) scheme is proposed. In the proposed EMT, a few sinusoidal transforms, other than DCT, have also been utilized for coding both Intra and Inter prediction residuals. The best transform, as selected from a pre-defined transform subset specified by prediction mode, is explicitly signaled in a joint coding block level manner. Moreover, to accelerate encoding process, fast methods have also been proposed by skipping unnecessary transform rate-distortion evaluations using previously encoding statistics. The proposed method has been implemented on top of High-Efficiency Video Coding (HEVC) reference software, and significant coding gain has been verified.

Journal ArticleDOI
TL;DR: A fully-integrated 18.5 kHz RC time-constant-based oscillator is designed in 65 nm CMOS for sleep-mode timers in wireless sensors and provides timing noise suppression, leading to 10× reduction in long-term Allan deviation.
Abstract: A fully-integrated 18.5 kHz RC time-constant-based oscillator is designed in 65 nm CMOS for sleep-mode timers in wireless sensors. A comparator offset cancellation scheme achieves $4\times $ to $25\times $ temperature stability improvement, leading to an accuracy of ±0.18% to ±0.55% over −40 to 90 °C. Sub-threshold operation and low-swing oscillations result in ultra-low power consumption of 130 nW. The architecture also provides timing noise suppression, leading to $10\times $ reduction in long-term Allan deviation. It is measured to have a stability of 20 ppm or better for measurement intervals over 0.5 s. The oscillator also has a fast startup-time, with the period settling in 4 cycles.

Journal ArticleDOI
TL;DR: In this article, the impact of phase noise on the downlink performance of a multiuser multiple-input-multiple-output (MIMO) system was studied. But the authors focused on the impact on the quality of the channel state information (CSI) available at the BS when compared with a system without phase noise.
Abstract: We study the impact of phase noise on the downlink performance of a multiuser multiple-input–multiple-output (MIMO) system, where the base station (BS) employs a large number of transmit antennas $M$ . We consider a setup where the BS employs $M_{\mathrm{osc}}$ free-running oscillators, and $M/M_{\mathrm{osc}}$ antennas are connected to each oscillator. For this configuration, we analyze the impact of phase noise on the performance of zero forcing (ZF), regularized ZF, and matched filter precoders when $M$ and the number of users $K$ are asymptotically large, whereas the ratio $M/K=\beta$ is fixed. We analytically show that the impact of phase noise on the signal-to-interference-plus-noise ratio (SINR) can be quantified as an effective reduction in the quality of the channel state information (CSI) available at the BS when compared with a system without phase noise. As a consequence, we observe that as $M_{\mathrm{osc}}$ increases, the SINR performance of all considered precoders degrades. On the other hand, the variance of the random phase variations caused by the BS oscillators reduces with increasing $M_{\mathrm{osc}}$ . Through Monte Carlo simulations, we verify our analytical results and compare the performance of the precoders for different phase noise and channel noise variances. For all considered precoders, we show that when $\beta$ is small, the performance of the setup where all BS antennas are connected to a single oscillator is superior to that of the setup where each BS antenna has its own oscillator. However, the opposite is true when $\beta$ is large and the signal-to-noise ratio (SNR) at the users is low.

Posted Content
TL;DR: The role of caches in wireless interference networks is studied, and a constant-factor approximation of the system’s degrees of freedom is provided, for arbitrary number of transmitters, number of receivers, content library size, receiver cache size, and transmitter cache size.
Abstract: We study the role of caches in wireless interference networks. We focus on content caching and delivery across a Gaussian interference network, where both transmitters and receivers are equipped with caches. We provide a constant-factor approximation of the system's degrees of freedom (DoF), for arbitrary number of transmitters, number of receivers, content library size, receiver cache size, and transmitter cache size (as long as the transmitters combined can store the entire content library among them). We demonstrate approximate optimality with respect to information-theoretic bounds that do not impose any restrictions on the caching and delivery strategies. Our characterization reveals three key insights. First, the approximate DoF is achieved using a strategy that separates the physical and network layers. This separation architecture is thus approximately optimal. Second, we show that increasing transmitter cache memory beyond what is needed to exactly store the entire library between all transmitters does not provide more than a constant-factor benefit to the DoF. A consequence is that transmit zero-forcing is not needed for approximate optimality. Third, we derive an interesting trade-off between the receiver memory and the number of transmitters needed for approximately maximal performance. In particular, if each receiver can store a constant fraction of the content library, then only a constant number of transmitters are needed. Our solution to the caching problem requires formulating and solving a new communication problem, the symmetric multiple multicast X-channel, for which we provide an exact DoF characterization.