scispace - formally typeset
Search or ask a question

Showing papers in "IEEE Transactions on Wireless Communications in 2020"


Journal ArticleDOI
TL;DR: This paper proposes to invoke an IRS at the cell boundary of multiple cells to assist the downlink transmission to cell-edge users, whilst mitigating the inter-cell interference, which is a crucial issue in multicell communication systems.
Abstract: Intelligent reflecting surfaces (IRSs) constitute a disruptive wireless communication technique capable of creating a controllable propagation environment. In this paper, we propose to invoke an IRS at the cell boundary of multiple cells to assist the downlink transmission to cell-edge users, whilst mitigating the inter-cell interference, which is a crucial issue in multicell communication systems. We aim for maximizing the weighted sum rate (WSR) of all users through jointly optimizing the active precoding matrices at the base stations (BSs) and the phase shifts at the IRS subject to each BS’s power constraint and unit modulus constraint. Both the BSs and the users are equipped with multiple antennas, which enhances the spectral efficiency by exploiting the spatial multiplexing gain. Due to the non-convexity of the problem, we first reformulate it into an equivalent one, which is solved by using the block coordinate descent (BCD) algorithm, where the precoding matrices and phase shifts are alternately optimized. The optimal precoding matrices can be obtained in closed form, when fixing the phase shifts. A pair of efficient algorithms are proposed for solving the phase shift optimization problem, namely the Majorization-Minimization (MM) Algorithm and the Complex Circle Manifold (CCM) Method. Both algorithms are guaranteed to converge to at least locally optimal solutions. We also extend the proposed algorithms to the more general multiple-IRS and network MIMO scenarios. Finally, our simulation results confirm the advantages of introducing IRSs in enhancing the cell-edge user performance.

865 citations


Journal ArticleDOI
TL;DR: A novel over-the-air computation based approach for fast global model aggregation via exploring the superposition property of a wireless multiple-access channel and providing a difference-of-convex-functions (DC) representation for the sparse and low-rank function to enhance sparsity and accurately detect the fixed-rank constraint in the procedure of device selection.
Abstract: The stringent requirements for low-latency and privacy of the emerging high-stake applications with intelligent devices such as drones and smart vehicles make the cloud computing inapplicable in these scenarios. Instead, edge machine learning becomes increasingly attractive for performing training and inference directly at network edges without sending data to a centralized data center. This stimulates a nascent field termed as federated learning for training a machine learning model on computation, storage, energy and bandwidth limited mobile devices in a distributed manner. To preserve data privacy and address the issues of unbalanced and non-IID data points across different devices, the federated averaging algorithm has been proposed for global model aggregation by computing the weighted average of locally updated model at each selected device. However, the limited communication bandwidth becomes the main bottleneck for aggregating the locally computed updates. We thus propose a novel over-the-air computation based approach for fast global model aggregation via exploring the superposition property of a wireless multiple-access channel. This is achieved by joint device selection and beamforming design, which is modeled as a sparse and low-rank optimization problem to support efficient algorithms design. To achieve this goal, we provide a difference-of-convex-functions (DC) representation for the sparse and low-rank function to enhance sparsity and accurately detect the fixed-rank constraint in the procedure of device selection. A DC algorithm is further developed to solve the resulting DC program with global convergence guarantees. The algorithmic advantages and admirable performance of the proposed methodologies are demonstrated through extensive numerical results.

579 citations


Journal ArticleDOI
TL;DR: A low-complexity algorithm is proposed to obtain the stationary solution for the joint design problem by utilizing the fractional programming technique and extended to the scenario wherein the CSI is imperfect.
Abstract: Reconfigurable intelligent surfaces (RIS) is a promising solution to build a programmable wireless environment via steering the incident signal in fully customizable ways with reconfigurable passive elements. In this paper, we consider a RIS-aided multiuser multiple-input single-output (MISO) downlink communication system. Our objective is to maximize the weighted sum-rate (WSR) of all users by joint designing the beamforming at the access point (AP) and the phase vector of the RIS elements, while both the perfect channel state information (CSI) setup and the imperfect CSI setup are investigated. For perfect CSI setup, a low-complexity algorithm is proposed to obtain the stationary solution for the joint design problem by utilizing the fractional programming technique. Then, we resort to the stochastic successive convex approximation technique and extend the proposed algorithm to the scenario wherein the CSI is imperfect. The validity of the proposed methods is confirmed by numerical results. In particular, the proposed algorithm performs quite well when the channel uncertainty is smaller than 10%.

576 citations


Journal ArticleDOI
TL;DR: In this paper, the authors proposed a three-phase pilot-based channel estimation framework for IRS-assisted uplink multiuser communications, in which the user-BS direct channels and the users-IRS-BS reflected channels of a typical user were estimated in Phase I and Phase II, respectively, while the users reflected channels were estimated with low overhead in Phase III via leveraging their strong correlation with those of the typical user under the case without receiver noise at the BS.
Abstract: In intelligent reflecting surface (IRS) assisted communication systems, the acquisition of channel state information is a crucial impediment for achieving the beamforming gain of IRS because of the considerable overhead required for channel estimation Specifically, under the current beamforming design for IRS-assisted communications, in total $KMN+KM$ channel coefficients should be estimated, where $K$ , $N$ and $M$ denote the numbers of users, IRS reflecting elements, and antennas at the base station (BS), respectively For the first time in the literature, this paper points out that despite the vast number of channel coefficients that should be estimated, significant redundancy exists in the user-IRS-BS reflected channels of different users arising from the fact that each IRS element reflects the signals from all the users to the BS via the same channel To utilize this redundancy for reducing the channel estimation time, we propose a novel three-phase pilot-based channel estimation framework for IRS-assisted uplink multiuser communications, in which the user-BS direct channels and the user-IRS-BS reflected channels of a typical user are estimated in Phase I and Phase II, respectively, while the user-IRS-BS reflected channels of the other users are estimated with low overhead in Phase III via leveraging their strong correlation with those of the typical user Under this framework, we analytically prove that a time duration consisting of $K+N+\max (K-1,\lceil (K-1)N/M \rceil)$ pilot symbols is sufficient for perfectly recovering all the $KMN+KM$ channel coefficients under the case without receiver noise at the BS Further, under the case with receiver noise, the user pilot sequences, IRS reflecting coefficients, and BS linear minimum mean-squared error channel estimators are characterized in closed-form

571 citations


Journal ArticleDOI
TL;DR: In this paper, the uplink spectral efficiencies of four different cell-free implementations are analyzed, with spatially correlated fading and arbitrary linear processing, and it is shown that a centralized implementation with optimal minimum mean-square error (MMSE) processing not only maximizes the SE but largely reduces the fronthaul signaling compared to the standard distributed approach.
Abstract: Cell-free Massive MIMO is considered as a promising technology for satisfying the increasing number of users and high rate expectations in beyond-5G networks. The key idea is to let many distributed access points (APs) communicate with all users in the network, possibly by using joint coherent signal processing. The aim of this paper is to provide the first comprehensive analysis of this technology under different degrees of cooperation among the APs. Particularly, the uplink spectral efficiencies of four different cell-free implementations are analyzed, with spatially correlated fading and arbitrary linear processing. It turns out that it is possible to outperform conventional Cellular Massive MIMO and small cell networks by a wide margin, but only using global or local minimum mean-square error (MMSE) combining. This is in sharp contrast to the existing literature, which advocates for maximum-ratio combining. Also, we show that a centralized implementation with optimal MMSE processing not only maximizes the SE but largely reduces the fronthaul signaling compared to the standard distributed approach. This makes it the preferred way to operate Cell-free Massive MIMO networks. Non-linear decoding is also investigated and shown to bring negligible improvements.

546 citations


Journal ArticleDOI
TL;DR: In this article, the authors proposed a distributed stochastic gradient descent (DSGD) over a shared noisy wireless channel for federated machine learning at the wireless network edge, where limited power wireless devices, each with its own dataset, build a joint model with the help of a remote parameter server.
Abstract: We study federated machine learning at the wireless network edge, where limited power wireless devices, each with its own dataset, build a joint model with the help of a remote parameter server (PS). We consider a bandwidth-limited fading multiple access channel (MAC) from the wireless devices to the PS, and propose various techniques to implement distributed stochastic gradient descent (DSGD) over this shared noisy wireless channel. We first propose a digital DSGD (D-DSGD) scheme, in which one device is selected opportunistically for transmission at each iteration based on the channel conditions; the scheduled device quantizes its gradient estimate to a finite number of bits imposed by the channel condition, and transmits these bits to the PS in a reliable manner. Next, motivated by the additive nature of the wireless MAC, we propose a novel analog communication scheme, referred to as the compressed analog DSGD (CA-DSGD), where the devices first sparsify their gradient estimates while accumulating error from previous iterations, and project the resultant sparse vector into a low-dimensional vector for bandwidth reduction. We also design a power allocation scheme to align the received gradient vectors at the PS in an efficient manner. Numerical results show that D-DSGD outperforms other digital approaches in the literature; however, in general the proposed CA-DSGD algorithm converges faster than the D-DSGD scheme, and reaches a higher level of accuracy. We have observed that the gap between the analog and digital schemes increases when the datasets of devices are not independent and identically distributed (i.i.d.). Furthermore, the performance of the CA-DSGD scheme is shown to be robust against imperfect channel state information (CSI) at the devices. Overall these results show clear advantages for the proposed analog over-the-air DSGD scheme, which suggests that learning and communication algorithms should be designed jointly to achieve the best end-to-end performance in machine learning applications at the wireless edge.

343 citations


Journal ArticleDOI
TL;DR: In this article, the authors proposed Edgent, a framework that leverages edge computing for DNN collaborative inference through device-edge synergy, which adaptively partitions computation between device and edge for purpose of coordinating the powerful cloud resource and the proximal edge resource for real-time DNN inference.
Abstract: As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep Neural Networks (DNNs) have quickly attracted widespread attention. However, it is challenging to run computation-intensive DNN-based tasks on mobile devices due to the limited computation resources. What’s worse, traditional cloud-assisted DNN inference is heavily hindered by the significant wide-area network latency, leading to poor real-time performance as well as low quality of user experience. To address these challenges, in this paper, we propose Edgent , a framework that leverages edge computing for DNN collaborative inference through device-edge synergy. Edgent exploits two design knobs: (1) DNN partitioning that adaptively partitions computation between device and edge for purpose of coordinating the powerful cloud resource and the proximal edge resource for real-time DNN inference; (2) DNN right-sizing that further reduces computing latency via early exiting inference at an appropriate intermediate DNN layer. In addition, considering the potential network fluctuation in real-world deployment, Edgent is properly design to specialize for both static and dynamic network environment. Specifically, in a static environment where the bandwidth changes slowly, Edgent derives the best configurations with the assist of regression-based prediction models, while in a dynamic environment where the bandwidth varies dramatically, Edgent generates the best execution plan through the online change point detection algorithm that maps the current bandwidth state to the optimal configuration. We implement Edgent prototype based on the Raspberry Pi and the desktop PC and the extensive experimental evaluations demonstrate Edgent ’s effectiveness in enabling on-demand low-latency edge intelligence.

329 citations


Journal ArticleDOI
TL;DR: This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards and proposes an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning.
Abstract: Unmanned aerial vehicles (UAVs) are capable of serving as aerial base stations (BSs) for providing both cost-effective and on-demand wireless communications. This article investigates dynamic resource allocation of multiple UAVs enabled communication networks with the goal of maximizing long-term rewards. More particularly, each UAV communicates with a ground user by automatically selecting its communicating user, power level and subchannel without any information exchange among UAVs. To model the dynamics and uncertainty in environments, we formulate the long-term resource allocation problem as a stochastic game for maximizing the expected rewards, where each UAV becomes a learning agent and each resource allocation solution corresponds to an action taken by the UAVs. Afterwards, we develop a multi-agent reinforcement learning (MARL) framework that each agent discovers its best strategy according to its local observations using learning. More specifically, we propose an agent-independent method, for which all agents conduct a decision algorithm independently but share a common structure based on Q-learning. Finally, simulation results reveal that: 1) appropriate parameters for exploitation and exploration are capable of enhancing the performance of the proposed MARL based resource allocation algorithm; 2) the proposed MARL algorithm provides acceptable performance compared to the case with complete information exchanges among UAVs. By doing so, it strikes a good tradeoff between performance gains and information exchange overheads.

327 citations


Journal ArticleDOI
TL;DR: In this article, a downlink multiple-input single-output intelligent reflecting surface (IRS) aided non-orthogonal multiple access (NOMA) system is investigated, where a base station (BS) serves multiple users with the aid of RISs.
Abstract: This paper investigates a downlink multiple-input single-output intelligent reflecting surface (IRS) aided non-orthogonal multiple access (NOMA) system, where a base station (BS) serves multiple users with the aid of IRSs. Our goal is to maximize the sum rate of all users by jointly optimizing the active beamforming at the BS and the passive beamforming at the IRS, subject to successive interference cancellation decoding rate conditions and IRS reflecting elements constraints. In term of the characteristics of reflection amplitudes and phase shifts, we consider ideal and non-ideal IRS assumptions. To tackle the formulated non-convex problems, we propose efficient algorithms by invoking alternating optimization, which design the active beamforming and passive beamforming alternately. For the ideal IRS scenario, the two subproblems are solved by invoking the successive convex approximation technique. For the non-ideal IRS scenario, constant modulus IRS elements are further divided into continuous phase shifts and discrete phase shifts. To tackle the passive beamforming problem with continuous phase shifts, a novel algorithm is developed by utilizing the sequential rank-one constraint relaxation approach, which is guaranteed to find a locally optimal rank-one solution. Then, a quantization-based scheme is proposed for discrete phase shifts. Finally, numerical results illustrate that: i) the system sum rate can be significantly improved by deploying the IRS with the proposed algorithms; ii) 3-bit phase shifters are capable of achieving almost the same performance as the ideal IRS; iii) the proposed IRS-aided NOMA systems achieve higher system sum rate than the IRS-aided orthogonal multiple access system.

325 citations


Journal ArticleDOI
TL;DR: In this paper, the authors studied the minimum signal-to-interference-plus-noise ratio (SINR) achieved by the optimal linear precoder (OLP) that maximizes the minimum SINR subject to a given power constraint for any given RIS phase matrix.
Abstract: This work focuses on the downlink of a single-cell multi-user system in which a base station (BS) equipped with $M$ antennas communicates with $K$ single-antenna users through a reconfigurable intelligent surface (RIS) installed in the line-of-sight (LoS) of the BS. RIS is envisioned to offer unprecedented spectral efficiency gains by utilizing $N$ passive reflecting elements that induce phase shifts on the impinging electromagnetic waves to smartly reconfigure the signal propagation environment. We study the minimum signal-to-interference-plus-noise ratio (SINR) achieved by the optimal linear precoder (OLP), that maximizes the minimum SINR subject to a given power constraint for any given RIS phase matrix, for the cases where the LoS channel matrix between the BS and the RIS is of rank-one and of full-rank. In the former scenario, the minimum SINR achieved by the RIS-assisted link is bounded by a quantity that goes to zero with $K$ . For the high-rank scenario, we develop accurate deterministic approximations for the parameters of the asymptotically OLP, which are then utilized to optimize the RIS phase matrix. Simulation results show that RISs can outperform half-duplex relays with a small number of passive reflecting elements while large RISs are needed to outperform full-duplex relays.

320 citations


Journal ArticleDOI
TL;DR: In this paper, a low-latency multi-access scheme for edge learning is proposed, where the updates simultaneously transmitted by devices over broadband channels should be analog aggregated "over-the-air" by exploiting the waveform-superposition property of a multiaccess channel.
Abstract: To leverage rich data distributed at the network edge, a new machine-learning paradigm, called edge learning, has emerged where learning algorithms are deployed at the edge for providing intelligent services to mobile users. While computing speeds are advancing rapidly, the communication latency is becoming the bottleneck of fast edge learning. To address this issue, this work is focused on designing a low-latency multi-access scheme for edge learning. To this end, we consider a popular privacy-preserving framework, federated edge learning (FEEL), where a global AI-model at an edge-server is updated by aggregating (averaging) local models trained at edge devices. It is proposed that the updates simultaneously transmitted by devices over broadband channels should be analog aggregated “over-the-air” by exploiting the waveform-superposition property of a multi-access channel. Such broadband analog aggregation (BAA) results in dramatical communication-latency reduction compared with the conventional orthogonal access (i.e., OFDMA). In this work, the effects of BAA on learning performance are quantified targeting a single-cell random network. First, we derive two tradeoffs between communication-and-learning metrics, which are useful for network planning and optimization. The power control (“truncated channel inversion”) required for BAA results in a tradeoff between the update-reliability [as measured by the receive signal-to-noise ratio (SNR)] and the expected update-truncation ratio. Consider the scheduling of cell-interior devices to constrain path loss. This gives rise to the other tradeoff between the receive SNR and fraction of data exploited in learning. Next, the latency-reduction ratio of the proposed BAA with respect to the traditional OFDMA scheme is proved to scale almost linearly with the device population. Experiments based on a neural network and a real dataset are conducted for corroborating the theoretical results.

Journal ArticleDOI
TL;DR: This work jointly optimize the trajectory of a UAV and the radio resource allocation to maximize the number of served IoT devices, where each device has its own target data upload deadline.
Abstract: The global evolution of wireless technologies and intelligent sensing devices are transforming the realization of smart cities. Among the myriad of use cases, there is a need to support applications whereby low-resource IoT devices need to upload their sensor data to a remote control centre by target hard deadlines; otherwise, the data becomes outdated and loses its value, for example, in emergency or industrial control scenarios. In addition, the IoT devices can be either located in remote areas with limited wireless coverage or in dense areas with relatively low quality of service. This motivates the utilization of UAVs to offload traffic from existing wireless networks by collecting data from time-constrained IoT devices with performance guarantees. To this end, we jointly optimize the trajectory of a UAV and the radio resource allocation to maximize the number of served IoT devices, where each device has its own target data upload deadline. The formulated optimization problem is shown to be mixed integer non-convex and generally NP-hard. To solve it, we first propose the high-complexity branch, reduce and bound (BRB) algorithm to find the global optimal solution for relatively small scale scenarios. Then, we develop an effective sub-optimal algorithm based on successive convex approximation in order to obtain results for larger networks. Next, we propose an extension algorithm to further minimize the UAV’s flight distance for cases where the initial and final UAV locations are known a priori. We demonstrate the favourable characteristics of the algorithms via extensive simulations and analysis as a function of various system parameters, with benchmarking against two greedy algorithms based on distance and deadline metrics.

Journal ArticleDOI
TL;DR: An end-to-end wireless communication system using deep neural networks using DNNs is developed, where a conditional generative adversarial net (GAN) is employed to model the channel effects in a data-driven way, where the received signal corresponding to the pilot symbols is added as a part of the conditioning information of the GAN.
Abstract: In this article, we develop an end-to-end wireless communication system using deep neural networks (DNNs), where DNNs are employed to perform several key functions, including encoding, decoding, modulation, and demodulation. However, an accurate estimation of instantaneous channel transfer function, i.e. , channel state information (CSI), is needed in order for the transmitter DNN to learn to optimize the receiver gain in decoding. This is very much a challenge since CSI varies with time and location in wireless communications and is hard to obtain when designing transceivers. We propose to use a conditional generative adversarial net (GAN) to represent channel effects and to bridge the transmitter DNN and the receiver DNN so that the gradient of the transmitter DNN can be back-propagated from the receiver DNN. In particular, a conditional GAN is employed to model the channel effects in a data-driven way, where the received signal corresponding to the pilot symbols is added as a part of the conditioning information of the GAN. To address the curse of dimensionality when the transmit symbol sequence is long, convolutional layers are utilized. From the simulation results, the proposed method is effective on additive white Gaussian noise (AWGN) channels, Rayleigh fading channels, and frequency-selective channels, which opens a new door for building data-driven DNNs for end-to-end communication systems.

Journal ArticleDOI
TL;DR: It is proved that the OP at high signal-to-noise ratios (SNRs) is a function of threshold, distortion noises, estimation errors and fading parameters, which results in 0 diversity order, and it is demonstrated that the outage performance of cooperative NOMA scenario exceeds the non-cooperative NomA scenario in the high SNR regime.
Abstract: This paper investigates the impact of residual transceiver hardware impairments (RTHIs) on cooperative non-orthogonal multiple access (NOMA) networks, where generic $\alpha -\mu $ fading channel is considered. To be practical, imperfect channel state information (CSI) and imperfect successive interference cancellation (SIC) are taken into account. More particularly, two representative NOMA scenarios are proposed, namely non-cooperative NOMA and cooperative NOMA. For the non-cooperative NOMA, the base station (BS) directly performs NOMA with all users. For the cooperative NOMA, the BS communicates with NOMA users with the aid of an amplify-and-forward (AF) relay, and the direct links between BS and users are existent. To characterize the performance of the proposed networks, new closed-form and asymptotic expressions for the outage probability (OP), ergodic capacity (EC) and energy efficiency (EE) are derived, respectively. Specifically, we also design the relay location optimization algorithms from the perspectives of minimize the asymptotic OP. For non-cooperative NOMA, it is proved that the OP at high signal-to-noise ratios (SNRs) is a function of threshold, distortion noises, estimation errors and fading parameters, which results in 0 diversity order. In addition, high SNR slopes and high SNR power offsets achieved by users are studied. It is shown that there are rate ceilings for the EC at high SNRs due to estimation error and distortion noise, which cause 0 high SNR slopes and $\infty $ high SNR power offsets . For cooperative NOMA, similar results can be obtained, and it also demonstrates that the outage performance of cooperative NOMA scenario exceeds the non-cooperative NOMA scenario in the high SNR regime.

Journal ArticleDOI
TL;DR: In this paper, an asymptotic analysis of the uplink data rate in an LIS-based large antenna-array system is presented, where the estimated channel on LIS is subject to estimation errors, interference channels are spatially correlated Rician fading channels, and the LIS experiences hardware impairments.
Abstract: The concept of a large intelligent surface (LIS) has recently emerged as a promising wireless communication paradigm that can exploit the entire surface of man-made structures for transmitting and receiving information. An LIS is expected to go beyond massive multiple-input multiple-output (MIMO) system, insofar as the desired channel can be modeled as a perfect line-of-sight. To understand the fundamental performance benefits, it is imperative to analyze its achievable data rate, under practical LIS environments and limitations. In this paper, an asymptotic analysis of the uplink data rate in an LIS-based large antenna-array system is presented. In particular, the asymptotic LIS rate is derived in a practical wireless environment where the estimated channel on LIS is subject to estimation errors, interference channels are spatially correlated Rician fading channels, and the LIS experiences hardware impairments. Moreover, the occurrence of the channel hardening effect is analyzed and the performance bound is asymptotically derived for the considered LIS system. The analytical asymptotic results are then shown to be in close agreement with the exact mutual information as the number of antennas and devices increase without bounds. Moreover, the derived ergodic rates show that hardware impairments, noise, and interference from estimation errors and the non-line-of-sight path become negligible as the number of antennas increases. Simulation results show that an LIS can achieve a performance that is comparable to conventional massive MIMO with improved reliability and a significantly reduced area for antenna deployment.

Journal ArticleDOI
TL;DR: Results show that the UC approach, which requires smaller backhaul overhead and is more scalable that the CF deployment, also achieves generally better performance than the CF approach for the vast majority of the users, especially on the uplink.
Abstract: Recently, the so-called cell-free (CF) massive multiple-input multiple-output (MIMO) architecture has been introduced, wherein a very large number of distributed access points (APs) simultaneously and jointly serve a much smaller number of mobile stations (MSs). The paper extends the CF approach to the case in which both the APs and the MSs are equipped with multiple antennas, proposing a beamfoming scheme that, relying on the zero-forcing strategy, does not require channel estimation at the MSs. We contrast the originally proposed formulation of CF massive MIMO with a user-centric (UC) approach wherein each MS is served only by a limited number of APs. Exploiting the framework of successive lower-bound maximization, the paper also proposes and analyzes power allocation strategies aimed at either sum-rate maximization or minimum-rate maximization, both for the uplink and downlink. Results show that the UC approach, which requires smaller backhaul overhead and is more scalable that the CF deployment, also achieves generally better performance than the CF approach for the vast majority of the users, especially on the uplink.

Journal ArticleDOI
TL;DR: A single edge server that assists a mobile user in executing a sequence of computation tasks is considered, and a mixed integer non-linear programming (MINLP) is formulated that jointly optimizes the service caching placement, computation offloading decisions, and system resource allocation.
Abstract: In mobile edge computing (MEC) systems, edge service caching refers to pre-storing the necessary programs for executing computation tasks at MEC servers. Service caching effectively reduces the real-time delay/bandwidth cost on acquiring and initializing service applications when computation tasks are offloaded to the MEC servers. The limited caching space at resource-constrained edge servers calls for careful design of caching placement to determine which programs to cache over time. This is in general a complicated problem that highly correlates to the computation offloading decisions of computation tasks, i.e., whether or not to offload a task for edge execution. In this paper, we consider a single edge server that assists a mobile user (MU) in executing a sequence of computation tasks. In particular, the MU can upload and run its customized programs at the edge server, while the server can selectively cache the previously generated programs for future reuse. To minimize the computation delay and energy consumption of the MU, we formulate a mixed integer non-linear programming (MINLP) that jointly optimizes the service caching placement, computation offloading decisions, and system resource allocation (e.g., CPU processing frequency and transmit power of MU). To tackle the problem, we first derive the closed-form expressions of the optimal resource allocation solutions, and subsequently transform the MINLP into an equivalent pure 0-1 integer linear programming (ILP) that is much simpler to solve. To further reduce the complexity in solving the ILP, we exploit the underlying structures of caching causality and task dependency models, and accordingly devise a reduced-complexity alternating minimization technique to update the caching placement and offloading decision alternately. Extensive simulations show that the proposed joint optimization techniques achieve substantial resource savings of the MU compared to other representative benchmark methods considered.

Journal ArticleDOI
TL;DR: MMNet is a deep learning MIMO detection scheme that significantly outperforms existing approaches on realistic channels with the same or lower computational complexity, and is 4–8dB better overall than a classic linear scheme like the minimum mean square error (MMSE) detector.
Abstract: Traditional symbol detection algorithms either perform poorly or are impractical to implement for Massive Multiple-Input Multiple-Output (MIMO) systems. Recently, several learning-based approaches have achieved promising results on simple channel models (e.g., i.i.d. Gaussian channel coefficients), but as we show, their performance degrades on real-world channels with spatial correlation. We propose MMNet, a deep learning MIMO detection scheme that significantly outperforms existing approaches on realistic channels with the same or lower computational complexity. MMNet ’s design builds on the theory of iterative soft-thresholding algorithms, and uses a novel training algorithm that leverages temporal and spectral correlation in real channels to accelerate training. These innovations make it practical to train MMNet online for every realization of the channel. On i.i.d. Gaussian channels, MMNet requires two orders of magnitude fewer operations than existing deep learning schemes but achieves near-optimal performance. On spatially-correlated channels, it achieves the same error rate as the next-best learning scheme (OAMPNet) at 2.5dB lower signal-to-noise ratio (SNR), and with at least $10\times $ less computational complexity. MMNet is also 4–8dB better overall than a classic linear scheme like the minimum mean square error (MMSE) detector.

Journal ArticleDOI
TL;DR: This paper considers a two-user MEC network, where each WD has a sequence of tasks to execute and proves that the optimal offloading decisions follow an one-climb policy, based on which a reduced-complexity Gibbs Sampling algorithm is proposed to obtain the optimalOffloading decisions.
Abstract: Mobile-edge computing (MEC) has recently emerged as a cost-effective paradigm to enhance the computing capability of hardware-constrained wireless devices (WDs). In this paper, we first consider a two-user MEC network, where each WD has a sequence of tasks to execute. In particular, we consider task dependency between the two WDs, where the input of a task at one WD requires the final task output at the other WD. Under the considered task-dependency model, we study the optimal task offloading policy and resource allocation (e.g., on offloading transmit power and local CPU frequencies) that minimize the weighted sum of the WDs’ energy consumption and task execution time. The problem is challenging due to the combinatorial nature of the offloading decisions among all tasks and the strong coupling with resource allocation. To tackle this problem, we first assume that the offloading decisions are given and derive the closed-form expressions of the optimal offloading transmit power and local CPU frequencies. Then, an efficient bi-section search method is proposed to obtain the optimal solutions. Furthermore, we prove that the optimal offloading decisions follow an one-climb policy, based on which a reduced-complexity Gibbs Sampling algorithm is proposed to obtain the optimal offloading decisions. We then extend the investigation to a general multi-user scenario, where the input of a task at one WD requires the final task outputs from multiple other WDs. Numerical results show that the proposed method can significantly outperform the other representative benchmarks and efficiently achieve low complexity with respect to the call graph size.

Journal ArticleDOI
TL;DR: Overall, this work shows that a suitable digitally modulated waveform enables to efficiently operate joint radar parameter estimation and communication by achieving full information rate of the modulation and near-optimal radar estimation performance.
Abstract: We consider a joint radar parameter estimation and communication system using orthogonal time frequency space (OTFS) modulation. The scenario is motivated by vehicular applications where a vehicle (or the infrastructure) equipped with a mono-static radar wishes to communicate data to its target receiver, while estimating parameters of interest related to this receiver. Provided that the radar-equipped transmitter is ready to send data to its target receiver, this setting naturally assumes that the receiver has been already detected. In a point-to-point communication setting over multipath time-frequency selective channels, we study the joint radar and communication system from two perspectives, i.e., the radar parameter estimation at the transmitter as well as the data detection at the receiver. For the radar parameter estimation part, we derive an efficient approximated Maximum Likelihood algorithm and the corresponding Cramer-Rao lower bound for range and velocity estimation. Numerical examples demonstrate that multi-carrier digital formats such as OTFS can achieve as accurate radar estimation as state-of-the-art radar waveforms such as frequency-modulated continuous wave (FMCW). For the data detection part, we focus on separate detection and decoding and consider a soft-output detector that exploits efficiently the channel sparsity in the Doppler-delay domain. We quantify the detector performance in terms of its pragmatic capacity , i.e., the achievable rate of the channel induced by the signal constellation and the detector soft-output. Simulations show that the proposed scheme outperforms concurrent state-of-the-art solutions. Overall, our work shows that a suitable digitally modulated waveform enables to efficiently operate joint radar parameter estimation and communication by achieving full information rate of the modulation and near-optimal radar estimation performance. Furthermore, OTFS appears to be particularly suited to the scope.

Journal ArticleDOI
TL;DR: In this article, a multiple-rate compressive sensing neural network framework was proposed to compress and quantize the channel state information (CSI) in massive MIMO networks, which not only improves reconstruction accuracy but also decreases storage space at the UE.
Abstract: Massive multiple-input multiple-output (MIMO) is a promising technology to increase link capacity and energy efficiency. However, these benefits are based on available channel state information (CSI) at the base station (BS). Therefore, user equipment (UE) needs to keep on feeding CSI back to the BS, thereby consuming precious bandwidth resource. Large-scale antennas at the BS for massive MIMO seriously increase this overhead. In this paper, we propose a multiple-rate compressive sensing neural network framework to compress and quantize the CSI. This framework not only improves reconstruction accuracy but also decreases storage space at the UE, thus enhancing the system feasibility. Specifically, we establish two network design principles for CSI feedback, propose a new network architecture, CsiNet+, according to these principles, and develop a novel quantization framework and training strategy. Next, we further introduce two different variable-rate approaches, namely, SM-CsiNet+ and PM-CsiNet+, which decrease the parameter number at the UE by 38.0% and 46.7%, respectively. Experimental results show that CsiNet+ outperforms the state-of-the-art network by a margin but only slightly increases the parameter number. We also investigate the compression and reconstruction mechanism behind deep learning-based CSI feedback methods via parameter visualization, which provides a guideline for subsequent research.

Journal ArticleDOI
TL;DR: This paper considers an optimization problem to minimize the computation mean-squared error of the K sensors’ signals at the receiver by optimizing the transmitting-receiving (Tx-Rx) policy, under the peak power constraint of each sensor.
Abstract: For future Internet-of-Things based Big Data applications, data collection from ubiquitous smart sensors with limited spectrum bandwidth is very challenging. On the other hand, to interpret the meaning behind the collected data, it is also challenging for an edge fusion center running computing tasks over large data sets with a limited computation capacity. To tackle these challenges, by exploiting the superposition property of multiple-access channel and the functional decomposition, the recently proposed technique, over-the-air computation (AirComp), enables an effective joint data collection and computation from concurrent sensor transmissions. In this paper, we focus on a single-antenna AirComp system consisting of K sensors and one receiver. We consider an optimization problem to minimize the computation mean-squared error (MSE) of the K sensors’ signals at the receiver by optimizing the transmitting-receiving (Tx-Rx) policy, under the peak power constraint of each sensor. Although the problem is not convex, we derive the computation-optimal policy in closed form. Also, we comprehensively investigate the ergodic performance of the AirComp system, and the scaling laws of the average computation MSE (ACM) and the average power consumption (APC) of different Tx-Rx policies with respect to K . For the computation-optimal policy, we show that the policy has a vanishing ACM and a vanishing APC with the increasing K .

Journal ArticleDOI
Siqi Luo1, Xu Chen1, Qiong Wu1, Zhi Zhou1, Shuai Yu1 
TL;DR: A novel Hierarchical Federated Edge Learning (HFEL) framework is introduced in which model aggregation is partially migrated to edge servers from the cloud and achieves better training performance compared to conventional federated learning.
Abstract: Federated Learning (FL) has been proposed as an appealing approach to handle data privacy issue of mobile devices compared to conventional machine learning at the remote cloud with raw user data uploading By leveraging edge servers as intermediaries to perform partial model aggregation in proximity and relieve core network transmission overhead, it enables great potentials in low-latency and energy-efficient FL Hence we introduce a novel Hierarchical Federated Edge Learning (HFEL) framework in which model aggregation is partially migrated to edge servers from the cloud We further formulate a joint computation and communication resource allocation and edge association problem for device users under HFEL framework to achieve global cost minimization To solve the problem, we propose an efficient resource scheduling algorithm in the HFEL framework It can be decomposed into two subproblems: resource allocation given a scheduled set of devices for each edge server and edge association of device users across all the edge servers With the optimal policy of the convex resource allocation subproblem for a set of devices under a single edge server, an efficient edge association strategy can be achieved through iterative global cost reduction adjustment process, which is shown to converge to a stable system point Extensive performance evaluations demonstrate that our HFEL framework outperforms the proposed benchmarks in global cost saving and achieves better training performance compared to conventional federated learning

Journal ArticleDOI
TL;DR: ViterbiNet, which is a data-driven symbol detector that does not require channel state information (CSI), is obtained by integrating deep neural networks (DNNs) into the Viterbi algorithm and demonstrates the conceptual benefit of designing communication systems that integrate DNNs into established algorithms.
Abstract: Symbol detection plays an important role in the implementation of digital receivers. In this work, we propose ViterbiNet, which is a data-driven symbol detector that does not require channel state information (CSI). ViterbiNet is obtained by integrating deep neural networks (DNNs) into the Viterbi algorithm. We identify the specific parts of the Viterbi algorithm that depend on the channel model, and design a DNN to implement only those computations, leaving the rest of the algorithm structure intact. We then propose a meta-learning based approach to train ViterbiNet online based on recent decisions, allowing the receiver to track dynamic channel conditions without requiring new training samples for every coherence block. Our numerical evaluations demonstrate that the performance of ViterbiNet, which is ignorant of the CSI, approaches that of the CSI-based Viterbi algorithm, and is capable of tracking time-varying channels without needing instantaneous CSI or additional training data. Moreover, unlike conventional Viterbi detection, ViterbiNet is robust to CSI uncertainty, and it can be reliably implemented in complex channel models with constrained computational burden. More broadly, our results demonstrate the conceptual benefit of designing communication systems that integrate DNNs into established algorithms.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a novel scheme for cell-free massive MIMO networks to support any federated learning (FL) framework, which allows each instead of all the iterations of the FL framework to happen in a large-scale coherence time to guarantee a stable operation of an FL process.
Abstract: This paper proposes a novel scheme for cell-free massive multiple-input multiple-output (CFmMIMO) networks to support any federated learning (FL) framework This scheme allows each instead of all the iterations of the FL framework to happen in a large-scale coherence time to guarantee a stable operation of an FL process To show how to optimize the FL performance using this proposed scheme, we consider an existing FL framework as an example and target FL training time minimization for this framework An optimization problem is then formulated to jointly optimize the local accuracy, transmit power, data rate, and users’ processing frequency This mixed-timescale stochastic nonconvex problem captures the complex interactions among the training time, and transmission and computation of training updates of one FL process By employing the online successive convex approximation approach, we develop a new algorithm to solve the formulated problem with proven convergence to the neighbourhood of its stationary points Our numerical results confirm that the presented joint design reduces the training time by up to 55% over baseline approaches They also show that CFmMIMO here requires the lowest training time for FL processes compared with cell-free time-division multiple access massive MIMO and collocated massive MIMO

Journal ArticleDOI
TL;DR: In this paper, the authors considered a scenario where the central controller transmits different packets to a robot and an actuator, where the actuator is located far from the controller, and the robot can move between the controller and the actuators.
Abstract: Ultra-reliable and low-latency communication (URLLC) is one of three pillar applications defined in the fifth generation new radio (5G NR), and its research is still in its infancy due to the difficulties in guaranteeing extremely high reliability (say 10−9 packet loss probability) and low latency (say 1 ms) simultaneously. In URLLC, short packet transmission is adopted to reduce latency, such that conventional Shannon’s capacity formula is no longer applicable, and the achievable data rate in finite blocklength becomes a complex expression with respect to the decoding error probability and the blocklength. To provide URLLC service in a factory automation scenario, we consider that the central controller transmits different packets to a robot and an actuator, where the actuator is located far from the controller, and the robot can move between the controller and the actuator. In this scenario, we consider four fundamental downlink transmission schemes, including orthogonal multiple access (OMA), non-orthogonal multiple access (NOMA), relay-assisted, and cooperative NOMA (C-NOMA) schemes. For all these transmission schemes, we aim for jointly optimizing the blocklength and power allocation to minimize the decoding error probability of the actuator subject to the reliability requirement of the robot, the total energy constraints, as well as the latency constraints. We further develop low-complexity algorithms to address the optimization problems for each transmission scheme. For the general case with more than two devices, we also develop a low-complexity efficient algorithm for the OMA scheme. Our results show that the relay-assisted transmission significantly outperforms the OMA scheme, while the NOMA scheme performs well when the blocklength is very limited. We further show that the relay-assisted transmission has superior performance over the C-NOMA scheme due to larger feasible region of the former scheme.

Journal ArticleDOI
TL;DR: In this paper, a dynamic downlink power control problem is considered for maximizing the sum-rate in a multi-user wireless cellular network, and a mathematical analysis is presented for the top-level design of the near-static problem.
Abstract: The model-based power allocation has been investigated for decades, but this approach requires mathematical models to be analytically tractable and it has high computational complexity. Recently, the data-driven model-free approaches have been rapidly developed to achieve near-optimal performance with affordable computational complexity, and deep reinforcement learning (DRL) is regarded as one such approach having great potential for future intelligent networks. In this paper, a dynamic downlink power control problem is considered for maximizing the sum-rate in a multi-user wireless cellular network. Using cross-cell coordinations, the proposed multi-agent DRL framework includes off-line and on-line centralized training and distributed execution, and a mathematical analysis is presented for the top-level design of the near-static problem. Policy-based REINFORCE , value-based deep Q-learning (DQL), actor-critic deep deterministic policy gradient (DDPG) algorithms are proposed for this sum-rate problem. Simulation results show that the data-driven approaches outperform the state-of-art model-based methods on sum-rate performance. Furthermore, the DDPG outperforms the REINFORCE and DQL in terms of both sum-rate performance and robustness.

Journal ArticleDOI
TL;DR: In this article, the authors investigated the use of a reconfigurable intelligent surface (RIS) to aid point-to-point multi-data-stream multiple-input multiple-output (MIMO) wireless communications.
Abstract: This paper investigates the use of a reconfigurable intelligent surface (RIS) to aid point-to-point multi-data-stream multiple-input multiple-output (MIMO) wireless communications. With practical finite alphabet input, the reflecting elements at the RIS and the precoder at the transmitter are alternatively optimized to minimize the symbol error rate (MSER). In the reflecting optimization with a fixed precoder, two reflecting design methods are developed, referred as eMSER-Reflecting and vMSER-Reflecting. In the optimization of the precoding matrix with a fixed reflecting pattern, the matrix optimization is transformed to be a vector optimization problem and two methods are proposed to solve it, which are referred as MSER-Precoding and MMED-Precoding. The superiority of the proposed designs is investigated by simulations. Simulation results demonstrate that the proposed reflecting and precoding designs can offer a lower SER than existing designs with the assumption of complex Gaussian input. Moreover, we compare RIS with a full-duplex Amplify-and-Forward (AF) relay system in terms of SER to show the advantage of RIS.

Journal ArticleDOI
TL;DR: A deep reinforcement learning (DRL)-based algorithm, named as EEFC-TDBA (energy-efficient fair communication through trajectory design and band allocation) that chooses the state-of-the-art DRL algorithm, deep deterministic policy gradient (DDPG), as its basis is proposed.
Abstract: Unmanned Aerial Vehicle (UAV)-assisted communication has drawn increasing attention recently. In this paper, we investigate 3D UAV trajectory design and band allocation problem considering both the UAV’s energy consumption and the fairness among the ground users (GUs). Specifically, we first formulate the energy consumption model of a quad-rotor UAV as a function of the UAV’s 3D movement. Then, based on the fairness and the total throughput, the fair throughput is defined and maximized within limited energy. We propose a deep reinforcement learning (DRL)-based algorithm, named as EEFC-TDBA (energy-efficient fair communication through trajectory design and band allocation) that chooses the state-of-the-art DRL algorithm, deep deterministic policy gradient (DDPG), as its basis. EEFC-TDBA allows the UAV to: 1) adjust the flight speed and direction so as to enhance the energy efficiency and reach the destination before the energy is exhausted; and 2) allocate frequency band to achieve fair communication service. Simulation results are provided to demonstrate that EEFC-TDBA outperforms the baseline methods in terms of the fairness, the total throughput, as well as the minimum throughput.

Journal ArticleDOI
TL;DR: A proactive algorithm based on long short-term memory and deep reinforcement learning techniques to address the partial observability and the curse of high dimensionality in local network state space faced by each VUE-pair is proposed.
Abstract: In this paper, we investigate the problem of age of information (AoI)-aware radio resource management for expected long-term performance optimization in a Manhattan grid vehicle-to-vehicle network. With the observation of global network state at each scheduling slot, the roadside unit (RSU) allocates the frequency bands and schedules packet transmissions for all vehicle user equipment-pairs (VUE-pairs). We model the stochastic decision-making procedure as a discrete-time single-agent Markov decision process (MDP). The technical challenges in solving the optimal control policy originate from high spatial mobility and temporally varying traffic information arrivals of the VUE-pairs. To make the problem solving tractable, we first decompose the original MDP into a series of per-VUE-pair MDPs. Then we propose a proactive algorithm based on long short-term memory and deep reinforcement learning techniques to address the partial observability and the curse of high dimensionality in local network state space faced by each VUE-pair. With the proposed algorithm, the RSU makes the optimal frequency band allocation and packet scheduling decision at each scheduling slot in a decentralized way in accordance with the partial observations of the global network state at the VUE-pairs. Numerical experiments validate the theoretical analysis and demonstrate the significant performance improvements from the proposed algorithm.