scispace - formally typeset
Open AccessJournal ArticleDOI

A Deep Learning Framework for Optimization of MISO Downlink Beamforming

TLDR
A deep learning framework for the optimization of downlink beamforming is proposed based on convolutional neural networks and exploitation of expert knowledge, such as the uplink-downlink duality and the known structure of optimal solutions, paving the way for fast realization of optimal beamforming in multiuser MISO systems.

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Quantum Machine Learning for 6G Communication Networks: State-of-the-Art and Vision for the Future

TL;DR: A novel QC-assisted and QML-based framework for 6G communication networks is proposed while articulating its challenges and potential enabling technologies at the network infrastructure, network edge, air interface, and user end.
Journal ArticleDOI

Fast Beamforming Design via Deep Learning

TL;DR: This work proposes a deep learning based fast beamforming design method which separates the problem into power allocation and virtual uplink beamforming (VUB) design and designs a heuristic solution structure of the downlink beamforming through the virtual equivalent uplink channel based on optimum MMSE receiver.
Journal ArticleDOI

Deep Reinforcement Learning for 5G Networks: Joint Beamforming, Power Control, and Interference Coordination

TL;DR: This paper formulate the joint design of beamforming, power control, and interference coordination as a non-convex optimization problem to maximize the signal to interference plus noise ratio (SINR) and solve this problem using deep reinforcement learning.
Posted Content

6G for Vehicle-to-Everything (V2X) Communications: Enabling Technologies, Challenges, and Opportunities

TL;DR: A series of key enabling technologies from a range of domains, such as new materials, algorithms, and system architectures are outlined, envisioning that machine learning will play an instrumental role for advanced vehicular communication and networking.
References
More filters
Proceedings Article

Adam: A Method for Stochastic Optimization

TL;DR: This work introduces Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments, and provides a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework.
Proceedings Article

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

TL;DR: Applied to a state-of-the-art image classification model, Batch Normalization achieves the same accuracy with 14 times fewer training steps, and beats the original model by a significant margin.
Journal ArticleDOI

Multilayer feedforward networks are universal approximators

TL;DR: It is rigorously established that standard multilayer feedforward networks with as few as one hidden layer using arbitrary squashing functions are capable of approximating any Borel measurable function from one finite dimensional space to another to any desired degree of accuracy, provided sufficiently many hidden units are available.
Proceedings Article

Understanding the difficulty of training deep feedforward neural networks

TL;DR: The objective here is to understand better why standard gradient descent from random initialization is doing so poorly with deep neural networks, to better understand these recent relative successes and help design better algorithms in the future.
Related Papers (5)