scispace - formally typeset
Search or ask a question

Showing papers on "Approximation algorithm published in 2020"


Journal ArticleDOI
Yu. A. Malkov1, D. A. Yashunin
TL;DR: Hierarchical Navigable Small World (HNSW) as mentioned in this paper is a fully graph-based approach for approximate K-nearest neighbor search without any need for additional search structures (typically used at the coarse search stage of most proximity graph techniques).
Abstract: We present a new approach for the approximate K-nearest neighbor search based on navigable small world graphs with controllable hierarchy (Hierarchical NSW, HNSW). The proposed solution is fully graph-based, without any need for additional search structures (typically used at the coarse search stage of the most proximity graph techniques). Hierarchical NSW incrementally builds a multi-layer structure consisting of a hierarchical set of proximity graphs (layers) for nested subsets of the stored elements. The maximum layer in which an element is present is selected randomly with an exponentially decaying probability distribution. This allows producing graphs similar to the previously studied Navigable Small World (NSW) structures while additionally having the links separated by their characteristic distance scales. Starting the search from the upper layer together with utilizing the scale separation boosts the performance compared to NSW and allows a logarithmic complexity scaling. Additional employment of a heuristic for selecting proximity graph neighbors significantly increases performance at high recall and in case of highly clustered data. Performance evaluation has demonstrated that the proposed general metric space search index is able to strongly outperform previous opensource state-of-the-art vector-only approaches. Similarity of the algorithm to the skip list structure allows straightforward balanced distributed implementation.

776 citations


Journal ArticleDOI
TL;DR: A low-complexity algorithm is proposed to obtain the stationary solution for the joint design problem by utilizing the fractional programming technique and extended to the scenario wherein the CSI is imperfect.
Abstract: Reconfigurable intelligent surfaces (RIS) is a promising solution to build a programmable wireless environment via steering the incident signal in fully customizable ways with reconfigurable passive elements. In this paper, we consider a RIS-aided multiuser multiple-input single-output (MISO) downlink communication system. Our objective is to maximize the weighted sum-rate (WSR) of all users by joint designing the beamforming at the access point (AP) and the phase vector of the RIS elements, while both the perfect channel state information (CSI) setup and the imperfect CSI setup are investigated. For perfect CSI setup, a low-complexity algorithm is proposed to obtain the stationary solution for the joint design problem by utilizing the fractional programming technique. Then, we resort to the stochastic successive convex approximation technique and extend the proposed algorithm to the scenario wherein the CSI is imperfect. The validity of the proposed methods is confirmed by numerical results. In particular, the proposed algorithm performs quite well when the channel uncertainty is smaller than 10%.

576 citations


Journal ArticleDOI
TL;DR: In this paper, a hybrid representative selection strategy and a fast approximation method for $K$K -nearest representatives are proposed for the construction of a sparse affinity sub-matrix.
Abstract: This paper focuses on scalability and robustness of spectral clustering for extremely large-scale datasets with limited resources. Two novel algorithms are proposed, namely, ultra-scalable spectral clustering (U-SPEC) and ultra-scalable ensemble clustering (U-SENC). In U-SPEC, a hybrid representative selection strategy and a fast approximation method for $K$ K -nearest representatives are proposed for the construction of a sparse affinity sub-matrix. By interpreting the sparse sub-matrix as a bipartite graph, the transfer cut is then utilized to efficiently partition the graph and obtain the clustering result. In U-SENC, multiple U-SPEC clusterers are further integrated into an ensemble clustering framework to enhance the robustness of U-SPEC while maintaining high efficiency. Based on the ensemble generation via multiple U-SEPC's, a new bipartite graph is constructed between objects and base clusters and then efficiently partitioned to achieve the consensus clustering result. It is noteworthy that both U-SPEC and U-SENC have nearly linear time and space complexity, and are capable of robustly and efficiently partitioning 10-million-level nonlinearly-separable datasets on a PC with 64 GB memory. Experiments on various large-scale datasets have demonstrated the scalability and robustness of our algorithms. The MATLAB code and experimental data are available at https://www.researchgate.net/publication/330760669 .

220 citations


Journal ArticleDOI
TL;DR: In this article, an alternating optimization (AO) algorithm is proposed to jointly optimize the transmit covariance R at transmitter and phase shift coefficient Q at IRS by fixing the other as a constant.
Abstract: In this letter, we consider an intelligent reflecting surface (IRS) assisted Guassian multiple-input multiple-output (MIMO) wiretap channel in which a multi-antenna transmitter communicates with a multi-antenna receiver in the presence of a multi-antenna eavesdropper. To maximize the secrecy rate of this channel, an alternating optimization (AO) algorithm is proposed to jointly optimize the transmit covariance R at transmitter and phase shift coefficient Q at IRS by fixing the other as a constant. When Q is fixed, existing numerical algorithm is used to search for global optimal R . When R is fixed, three sucessive approximation to the objective function to surrogate lower bound is applied and minorization-maximization (MM) algorithm is proposed to optimize the local optimal Q . Simulation results have be provided to validate the convergence and performance of the proposed AO algorithm.

171 citations


Journal ArticleDOI
TL;DR: In this paper, the authors investigated the passive beamforming and information transfer (PBIT) technique for multiuser multiple-input multiple-output (Mu-MIMO) systems with the aid of reconfigurable intelligent surfaces (RISs).
Abstract: This paper investigates the passive beamforming and information transfer (PBIT) technique for multiuser multiple-input multiple-output (Mu-MIMO) systems with the aid of reconfigurable intelligent surfaces (RISs), where the RISs enhance the primary communication via passive beamforming (P-BF) and at the same time deliver additional information by the on-off reflecting modulation (in which the RIS information is carried by the on/off state of each reflecting element). For the P-BF design, we propose to maximize the achievable user sum rate of the RIS-aided Mu-MIMO channel and formulate the problem as a two-step stochastic program. A sample average approximation (SAA) based iterative algorithm is developed for the efficient P-BF design of the considered scheme. To strike a balance between complexity and performance, we further propose a simplified P-BF algorithm by approximating the stochastic program as a deterministic alternating optimization problem. For the receiver design, the signal detection at the receiver is a bilinear estimation problem since the RIS information is multiplicatively modulated onto the reflected signals of the reflecting elements. To solve this bilinear estimation problem, we develop a turbo message passing (TMP) algorithm in which the factor graph associated with the problem is divided into two modules: one for the estimation of the user signals and the other for the estimation of the on-off state of each RIS element. The two modules are executed iteratively to yield a near-optimal low-complexity solution. Furthermore, we extend the design of the Mu-MIMO PBIT scheme from single-RIS to multi-RIS, by leveraging the similarity between the single-RIS and multi-RIS system models. Extensive simulation results are provided to demonstrate the advantages of our P-BF and receiver designs.

151 citations


Journal ArticleDOI
TL;DR: This work proposes a safe off-policy deep reinforcement learning algorithm to solve Volt-VAR control problems in a model-free manner, and outperforms the existing reinforcement learning algorithms and conventional optimization-based approaches on a large feeder.
Abstract: Volt-VAR control is critical to keeping distribution network voltages within allowable range, minimizing losses, and reducing wear and tear of voltage regulating devices. To deal with incomplete and inaccurate distribution network models, we propose a safe off-policy deep reinforcement learning algorithm to solve Volt-VAR control problems in a model-free manner. The Volt-VAR control problem is formulated as a constrained Markov decision process with discrete action space, and solved by our proposed constrained soft actor-critic algorithm. Our proposed reinforcement learning algorithm achieves scalability, sample efficiency, and constraint satisfaction by synergistically combining the merits of the maximum-entropy framework, the method of multiplier, a device-decoupled neural network structure, and an ordinal encoding scheme. Comprehensive numerical studies with the IEEE distribution test feeders show that our proposed algorithm outperforms the existing reinforcement learning algorithms and conventional optimization-based approaches on a large feeder.

150 citations


Journal ArticleDOI
TL;DR: In this article, an approximate variant of the gradient coding problem is introduced, in which they settle for approximate gradient computation instead of the exact one, which enables graceful degradation, i.e., the approximation error of the approximate gradient is a decreasing function of the number of straggglers.
Abstract: Gradient coding is a technique for straggler mitigation in distributed learning. In this paper we design novel gradient codes using tools from classical coding theory, namely, cyclic MDS codes, which compare favorably with existing solutions, both in the applicable range of parameters and in the complexity of the involved algorithms. Second, we introduce an approximate variant of the gradient coding problem, in which we settle for approximate gradient computation instead of the exact one. This approach enables graceful degradation, i.e., the $\ell _{2}$ error of the approximate gradient is a decreasing function of the number of stragglers. Our main result is that normalized adjacency matrices of expander graphs yield excellent approximate gradient codes, which enable significantly less computation compared to exact gradient coding, and guarantee faster convergence than trivial solutions under standard assumptions. We experimentally test our approach on Amazon EC2, and show that the generalization error of approximate gradient coding is very close to the full gradient while requiring significantly less computation from the workers.

149 citations


Journal ArticleDOI
12 Aug 2020
TL;DR: A comprehensive survey and a comparative evaluation of recently developed approximate arithmetic circuits under different design constraints, synthesized and characterized under optimizations for performance and area.
Abstract: Approximate computing has emerged as a new paradigm for high-performance and energy-efficient design of circuits and systems. For the many approximate arithmetic circuits proposed, it has become critical to understand a design or approximation technique for a specific application to improve performance and energy efficiency with a minimal loss in accuracy. This article aims to provide a comprehensive survey and a comparative evaluation of recently developed approximate arithmetic circuits under different design constraints. Specifically, approximate adders, multipliers, and dividers are synthesized and characterized under optimizations for performance and area. The error and circuit characteristics are then generalized for different classes of designs. The applications of these circuits in image processing and deep neural networks indicate that the circuits with lower error rates or error biases perform better in simple computations, such as the sum of products, whereas more complex accumulative computations that involve multiple matrix multiplications and convolutions are vulnerable to single-sided errors that lead to a large error bias in the computed result. Such complex computations are more sensitive to errors in addition than those in multiplication, so a larger approximation can be tolerated in multipliers than in adders. The use of approximate arithmetic circuits can improve the quality of image processing and deep learning in addition to the benefits in performance and power consumption for these applications.

143 citations


Posted Content
TL;DR: It is proved that the random features enable GNNs to learn almost optimal polynomial-time approximation algorithms for the minimum dominating set problem and maximum matching problem in terms of the approximation ratio.
Abstract: Graph neural networks (GNNs) are powerful machine learning models for various graph learning tasks. Recently, the limitations of the expressive power of various GNN models have been revealed. For example, GNNs cannot distinguish some non-isomorphic graphs and they cannot learn efficient graph algorithms. In this paper, we demonstrate that GNNs become powerful just by adding a random feature to each node. We prove that the random features enable GNNs to learn almost optimal polynomial-time approximation algorithms for the minimum dominating set problem and maximum matching problem in terms of approximation ratios. The main advantage of our method is that it can be combined with off-the-shelf GNN models with slight modifications. Through experiments, we show that the addition of random features enables GNNs to solve various problems that normal GNNs, including the graph convolutional networks (GCNs) and graph isomorphism networks (GINs), cannot solve.

136 citations


Journal ArticleDOI
TL;DR: A block-wise one-sided non-convex min-max problem, in which the minimization problem consists of multiple blocks and is non- Convex, while the maximization problem is (strongly) concave is considered.
Abstract: The min-max problem, also known as the saddle point problem, is a class of optimization problems which minimizes and maximizes two subsets of variables simultaneously. This class of problems can be used to formulate a wide range of signal processing and communication (SPCOM) problems. Despite its popularity, most existing theory for this class has been mainly developed for problems with certain special convex-concave structure. Therefore, it cannot be used to guide the algorithm design for many interesting problems in SPCOM, where various kinds of non-convexity arise. In this work, we consider a block-wise one-sided non-convex min-max problem, in which the minimization problem consists of multiple blocks and is non-convex, while the maximization problem is (strongly) concave. We propose a class of simple algorithms named Hybrid Block Successive Approximation (HiBSA), which alternatingly performs gradient descent-type steps for the minimization blocks and gradient ascent-type steps for the maximization problem. A key element in the proposed algorithm is the use of certain regularization and penalty sequences, which stabilize the algorithm and ensure convergence. We show that HiBSA converges to some properly defined first-order stationary solutions with quantifiable global rates. To validate the efficiency of the proposed algorithms, we conduct numerical tests on a number of problems, including the robust learning problem, the non-convex min-utility maximization problems, and certain wireless jamming problem arising in interfering channels.

130 citations


Journal ArticleDOI
TL;DR: A large-scale multi-object tracker based on the generalised labeled multi-Bernoulli (GLMB) filter is proposed and a new method of applying the optimal sub-pattern assignment (OSPA) metric to determine a meaningful distance between two sets of tracks is introduced.
Abstract: A large-scale multi-object tracker based on the generalised labeled multi-Bernoulli (GLMB) filter is proposed. The algorithm is capable of tracking a very large, unknown and time-varying number of objects simultaneously, in the presence of a high number of false alarms, as well as missed detections and measurement origin uncertainty due to closely spaced objects. The algorithm is demonstrated on a simulated tracking scenario, where the peak number objects appearing simultaneously exceeds one million. Additionally, we introduce a new method of applying the optimal sub-pattern assignment (OSPA) metric to determine a meaningful distance between two sets of tracks. We also develop an efficient strategy for its exact computation in large-scale scenarios to evaluate the performance of the proposed tracker.

Journal ArticleDOI
TL;DR: A reinforcement learning approach with value function approximation and feature learning is proposed for autonomous decision making of intelligent vehicles on highways and uses data-driven feature representation for value and policy approximation so that better learning efficiency can be achieved.
Abstract: Autonomous decision making is a critical and difficult task for intelligent vehicles in dynamic transportation environments. In this paper, a reinforcement learning approach with value function approximation and feature learning is proposed for autonomous decision making of intelligent vehicles on highways. In the proposed approach, the sequential decision making problem for lane changing and overtaking is modeled as a Markov decision process with multiple goals, including safety, speediness, smoothness, etc. In order to learn optimized policies for autonomous decision-making, a multiobjective approximate policy iteration (MO-API) algorithm is presented. The features for value function approximation are learned in a data-driven way, where sparse kernel-based features or manifold-based features can be constructed based on data samples. Compared with previous RL algorithms such as multiobjective Q-learning, the MO-API approach uses data-driven feature representation for value and policy approximation so that better learning efficiency can be achieved. A highway simulation environment using a 14 degree-of-freedom vehicle dynamics model was established to generate training data and test the performance of different decision-making methods for intelligent vehicles on highways. The results illustrate the advantages of the proposed MO-API method under different traffic conditions. Furthermore, we also tested the learned decision policy on a real autonomous vehicle to implement overtaking decision and control under normal traffic on highways. The experimental results also demonstrate the effectiveness of the proposed method.

Journal ArticleDOI
TL;DR: It is shown that the proposed fuzzy system and associated learning algorithm result in better approximation performance in comparison with the other well-known approaches.
Abstract: The main reason of the extensive usage of the fuzzy systems in many branches of science is their approximation ability. In this paper, an interval type-3 fuzzy system (IT3FS) is proposed. The uncertainty modeling capability of the proposed IT3FS is improved in contrast to type-1 and type-2 fuzzy systems (T1FS and T2FS). Because in the proposed IT3FS, the membership is defined as an interval type-2 fuzzy set, whereas in T1FS and T2FS, the membership is crisp value and type-1 fuzzy set, respectively. An online fractional-order learning algorithm is given to optimize the consequent parameters of the IT3FS. The stability of the learning algorithm is proved by utilizing the Lyapunov method. The validity of the proposed fuzzy system is illustrated by both simulation and the experimental studies. It is shown that the proposed fuzzy system and associated learning algorithm result in better approximation performance in comparison with the other well-known approaches.

Journal ArticleDOI
TL;DR: The UAV is utilized as a computing node as well as a relay node to improve the average user latency in the UAV-aided MEC (UAV-MEC) network and a proposed approximation algorithm is proposed that is superior to three baseline algorithms to minimize the average latency of all UEs.
Abstract: Advances in wireless communications are empowering the emerging Internet-of-Things (IoT) applications and services with billions of connected devices Mobile-edge computing (MEC) has been proposed to reduce the round-trip delay of these applications as IoT devices may have limited computing resources and the resource-rich mobile cloud may be far away On the other aspect, unmanned aerial vehicles (UAVs) may potentially be employed to improve the quality of service and the channel conditions of users We thus propose to utilize the UAV as a computing node as well as a relay node to improve the average user latency in the UAV-aided MEC (UAV-MEC) network and formulate the UAV-MEC problem with the objective to minimize the average latency of all UEs As the UAV-MEC problem is NP-hard, we decompose it into three subproblems We propose an approximation algorithm with low complexity to solve the first subproblem and then we obtain the optimal solutions of the remaining two subproblems, upon which another proposed approximation algorithm employs these solutions to finally solve the UAV-MEC problem The evaluation results demonstrate that the proposed algorithm is superior to three baseline algorithms

Journal ArticleDOI
TL;DR: A survey of the known approximation properties of the outputs of neural networks with the aim of uncovering the properties that are not present in the more traditional methods of approximation used in numerical analysis, such as approximations using polynomials, wavelets, rational functions and splines is presented in this paper.
Abstract: Neural networks (NNs) are the method of choice for building learning algorithms. They are now being investigated for other numerical tasks such as solving high-dimensional partial differential equations. Their popularity stems from their empirical success on several challenging learning problems (computer chess/Go, autonomous navigation, face recognition). However, most scholars agree that a convincing theoretical explanation for this success is still lacking. Since these applications revolve around approximating an unknown function from data observations, part of the answer must involve the ability of NNs to produce accurate approximations. This article surveys the known approximation properties of the outputs of NNs with the aim of uncovering the properties that are not present in the more traditional methods of approximation used in numerical analysis, such as approximations using polynomials, wavelets, rational functions and splines. Comparisons are made with traditional approximation methods from the viewpoint of rate distortion, i.e. error versus the number of parameters used to create the approximant. Another major component in the analysis of numerical approximation is the computational time needed to construct the approximation, and this in turn is intimately connected with the stability of the approximation algorithm. So the stability of numerical approximation using NNs is a large part of the analysis put forward. The survey, for the most part, is concerned with NNs using the popular ReLU activation function. In this case the outputs of the NNs are piecewise linear functions on rather complicated partitions of the domain of f into cells that are convex polytopes. When the architecture of the NN is fixed and the parameters are allowed to vary, the set of output functions of the NN is a parametrized nonlinear manifold. It is shown that this manifold has certain space-filling properties leading to an increased ability to approximate (better rate distortion) but at the expense of numerical stability. The space filling creates the challenge to the numerical method of finding best or good parameter choices when trying to approximate.

Journal ArticleDOI
TL;DR: JGraphT as discussed by the authors is a programming library that contains very efficient and generic graph data structures along with a large collection of state-of-the-art algorithms, such as shortest paths, spanning trees, graph and subgraph isomorphism, matching and flow problems, approximation algorithms for NP-hard problems such as independent set and the traveling salesman problem.
Abstract: Mathematical software and graph-theoretical algorithmic packages to efficiently model, analyze, and query graphs are crucial in an era where large-scale spatial, societal, and economic network data are abundantly available. One such package is JGraphT, a programming library that contains very efficient and generic graph data structures along with a large collection of state-of-the-art algorithms. The library is written in Java with stability, interoperability, and performance in mind. A distinctive feature of this library is its ability to model vertices and edges as arbitrary objects, thereby permitting natural representations of many common networks, including transportation, social, and biological networks. Besides classic graph algorithms such as shortest-paths and spanning-tree algorithms, the library contains numerous advanced algorithms: graph and subgraph isomorphism, matching and flow problems, approximation algorithms for NP-hard problems such as independent set and the traveling salesman problem, and several more exotic algorithms such as Berge graph detection. Due to its versatility and generic design, JGraphT is currently used in large-scale commercial products, as well as noncommercial and academic research projects. In this work, we describe in detail the design and underlying structure of the library, and discuss its most important features and algorithms. A computational study is conducted to evaluate the performance of JGraphT versus several similar libraries. Experiments on a large number of graphs over a variety of popular algorithms show that JGraphT is highly competitive with other established libraries such as NetworkX or the BGL.

Journal ArticleDOI
TL;DR: This paper presents a novel accelerated exact k-means using the ball to describe each cluster, which focus on reducing the point-centroid distance computation, which attains both higher performance and performs fewer distance calculations, especially for large-k problems.
Abstract: This paper presents a novel accelerated exact k-means called as "Ball k-means" by using the ball to describe each cluster, which focus on reducing the point-centroid distance computation. It can exactly find its neighbor clusters for each cluster, resulting distance computations only between a point and its neighbor clusters' centroids instead of all centroids. What's more, each cluster can be divided into "stable area" and "active area", and the latter one is further divided into some exact "annular area". The assignment of the points in the "stable area" is not changed while the points in each "annular area" will be adjusted within a few neighbor clusters. There are no upper or lower bounds in the whole process. Moreover, ball k-means uses ball clusters and neighbor searching along with multiple novel stratagems for reducing centroid distance computations. In comparison with the current state-of-the art accelerated exact bounded methods, the Yinyang algorithm and the Exponion algorithm, as well as other top-of-the-line tree-based and bounded methods, the ball k-means attains both higher performance and performs fewer distance calculations, especially for large-k problems. The faster speed, no extra parameters and simpler design of "Ball k-means" make it an all-around replacement of the naive k-means.

Proceedings ArticleDOI
01 May 2020
TL;DR: This work takes inspiration from the iterative linear-quadratic regulator (ILQR), which solves repeated approximations with linear dynamics and quadratic costs, and proposes an algorithm that solves repeated linear- quadratic games.
Abstract: Many problems in robotics involve multiple decision making agents. To operate efficiently in such settings, a robot must reason about the impact of its decisions on the behavior of other agents. Differential games offer an expressive theoretical framework for formulating these types of multi-agent problems. Unfortunately, most numerical solution techniques scale poorly with state dimension and are rarely used in real-time applications. For this reason, it is common to predict the future decisions of other agents and solve the resulting decoupled, i.e., single-agent, optimal control problem. This decoupling neglects the underlying interactive nature of the problem; however, efficient solution techniques do exist for broad classes of optimal control problems. We take inspiration from one such technique, the iterative linear-quadratic regulator (ILQR), which solves repeated approximations with linear dynamics and quadratic costs. Similarly, our proposed algorithm solves repeated linear-quadratic games. We experimentally benchmark our algorithm in several examples with a variety of initial conditions and show that the resulting strategies exhibit complex interactive behavior. Our results indicate that our algorithm converges reliably and runs in real-time. In a three-player, 14-state simulated intersection problem, our algorithm initially converges in < 0.25 s. Receding horizon invocations converge in < 50 ms in a hardware collision-avoidance test.

Journal ArticleDOI
TL;DR: The proposed robust optimal control algorithm tunes the parameters of critic-only neural network by event-triggering condition and runs in a plug-n-play framework without system functions, where fewer transmissions and less computation are required as all the measurements received simultaneously.
Abstract: In this paper, a novel event-sampled robust optimal controller is proposed for a class of continuous-time constrained-input nonlinear systems with unknown dynamics. In order to solve the robust optimal control problem, an online data-driven identifier is established to construct the system dynamics, and an event-sampled critic-only adaptive dynamic programming method is developed to replace the conventional time-driven actor–critic structure. The designed online identification method runs during the solving process and is not applied as a priori part for the solutions, which simplifies the architecture and reduces computational load. The proposed robust optimal control algorithm tunes the parameters of critic-only neural network (NN) by event-triggering condition and runs in a plug-n-play framework without system functions, where fewer transmissions and less computation are required as all the measurements received simultaneously. Based on the novel design, the stability of system and the convergence of critic NN are demonstrated by Lyapunov theory, where the state is asymptotically stable and weight error is guaranteed to be uniformly ultimately bounded. Finally, the applications in a basic nonlinear system and the complex rotational–translational actuator problem demonstrate the effectiveness of the proposed method.

Journal ArticleDOI
TL;DR: A randomized approximation algorithm which is provably superior to the state-of-the art methods with respect to running time is presented.
Abstract: Social networks allow rapid spread of ideas and innovations while negative information can also propagate widely. When a user receives two opposing opinions, they tend to believe the one arrives first. Therefore, once misinformation or rumor is detected, one containment method is to introduce a positive cascade competing against the rumor. Given a budget $k$ , the rumor blocking problem asks for $k$ seed users to trigger the spread of a positive cascade such that the number of the users who are not influenced by rumor can be maximized. The prior works have shown that the rumor blocking problem can be approximated within a factor of $(1-1/e)$ by a classic greedy algorithm combined with Monte Carlo simulation. Unfortunately, the Monte Carlo simulation based methods are time consuming and the existing algorithms either trade performance guarantees for practical efficiency or vice versa. In this paper, we present a randomized approximation algorithm which is provably superior to the state-of-the art methods with respect to running time. The superiority of the proposed algorithm is demonstrated by experiments done on both the real-world and synthetic social networks.

Journal ArticleDOI
TL;DR: This article develops a joint optimization problem model of partition deployment and resource allocation in MECSs (JPDRA) and designs a CRA algorithm based on Markov approximation and a low-complexity DPD algorithm to obtain the near-optimal solution in the polynomial time.
Abstract: Nowadays, the widely used Internet-of-Things (IoT) mobile devices (MDs) generate huge volumes of data, which need analyzing and extracting accurate information in real time by compute-intensive deep learning (DL) inference tasks. Due to its multilayer structure, the deep neural network (DNN) is appropriate for the mobile-edge computing (MEC) environment, and the DL tasks can be offloaded to DNN partitions deployed in MEC servers (MECSs) for speed-up inference. In this article, we first assume the arrival process of DL tasks as Poisson distribution and develop a tandem queueing model to evaluate the end-to-end (E2E) inference delay of DL tasks in multiple DNN partitions. To minimize the E2E delay, we develop a joint optimization problem model of partition deployment and resource allocation in MECSs (JPDRA). Since the JPDRA is a mixed-integer nonlinear programming (MINLP) problem, we decompose the original problem into a computing resource allocation (CRA) problem with fixed partition deployment decision and a DNN partition deployment (DPD) problem that optimizes the optimal-delay function related to the CRA problem. Next, we design a CRA algorithm based on Markov approximation and a low-complexity DPD algorithm to obtain the near-optimal solution in the polynomial time. The simulation results demonstrate that the proposed algorithms are more efficient and can reduce the average E2E delay by 25.7% with better convergence performance.

Journal ArticleDOI
TL;DR: A new decomposition approach with two mechanisms (static and dynamic) based on multiple reference points under the multiobjective evolutionary algorithm based on decomposition (MOEA/D) framework to address the above-mentioned difficulties of feature selection.
Abstract: Feature selection is an important task in machine learning that has two main objectives: 1) reducing dimensionality and 2) improving learning performance. Feature selection can be considered a multiobjective problem. However, it has its problematic characteristics, such as a highly discontinuous Pareto front, imbalance preferences, and partially conflicting objectives. These characteristics are not easy for existing evolutionary multiobjective optimization (EMO) algorithms. We propose a new decomposition approach with two mechanisms (static and dynamic) based on multiple reference points under the multiobjective evolutionary algorithm based on decomposition (MOEA/D) framework to address the above-mentioned difficulties of feature selection. The static mechanism alleviates the dependence of the decomposition on the Pareto front shape and the effect of the discontinuity. The dynamic one is able to detect regions in which the objectives are mostly conflicting, and allocates more computational resources to the detected regions. In comparison with other EMO algorithms on 12 different classification datasets, the proposed decomposition approach finds more diverse feature subsets with better performance in terms of hypervolume and inverted generational distance. The dynamic mechanism successfully identifies conflicting regions and further improves the approximation quality for the Pareto fronts.

Journal ArticleDOI
TL;DR: This paper presents a new approach to clustering called "supervised reinforcement learning" that automates the very labor-intensive and therefore time-heavy and therefore computationally-heavy process of discrete-time reinforcement learning.
Abstract: Clustering is a classic topic in optimization with $k$-means being one of the most fundamental such problems. In the absence of any restrictions on the input, the best-known algorithm for $k$-means...

Journal ArticleDOI
TL;DR: This paper presents a meta-modelling framework that automates the very labor-intensive and therefore time-heavy and therefore expensive process of manually selecting a subset of items to offer to a buyer.
Abstract: Assortment optimization is an important problem that arises in many practical applications such as retailing and online advertising. The fundamental goal is to select a subset of items to offer fro...

Journal ArticleDOI
TL;DR: In this article, the problem of jointly optimizing the precoders, message split, time slot allocation, and relaying user scheduling with the objective of maximizing the minimum rate among users subject to a transmit power constraint at the base station is studied.
Abstract: Cooperative Rate-Splitting (CRS) strategy, relying on linearly precoded rate-splitting at the transmitter and opportunistic transmission of the common message by the relaying user, has recently been shown to outperform typical Non-cooperative Rate-Splitting (NRS), Cooperative Non-Orthogonal Multiple Access (C-NOMA) and Space Division Multiple Access (SDMA) in a two-user Multiple Input Single Output (MISO) Broadcast Channel (BC) with user relaying In this work, the existing two-user CRS transmission strategy is generalized to the $K$ -user case We study the problem of jointly optimizing the precoders, message split, time slot allocation, and relaying user scheduling with the objective of maximizing the minimum rate among users subject to a transmit power constraint at the base station As the user scheduling problem is discrete and the entire problem is non-convex, we propose a two-stage low-complexity algorithm to solve the problem Both centralized and decentralized relaying protocols based on selecting $K_{1}$ ( $K_{1} ) strongest users are first proposed followed by a Successive Convex Approximation (SCA)-based algorithm to jointly optimize the time slot, precoders and message split Numerical results show that by applying the proposed two-stage algorithm, the worst-case achievable rate achieved by CRS is significantly increased over that of NRS and SDMA in a wide range of network loads (underloaded and overloaded regimes) and user deployments (with a diversity of channel strengths) Importantly, the proposed SCA-based algorithm dramatically reduces the computational complexity without any rate loss compared with the conventional algorithm in the literature of CRS Therefore, we conclude that the proposed $K$ -user CRS combined with the two-stage algorithm is more powerful than the existing transmission schemes

Journal ArticleDOI
TL;DR: This work develops a game theoretical model of the problem of coordinating the offloading decisions of wireless devices, proves the existence of pure strategy Nash equilibria, and proposes a polynomial complexity algorithm for computing an equilibrium.
Abstract: Motivated by various delay sensitive applications, we address the problem of coordinating the offloading decisions of wireless devices that periodically generate computationally intensive tasks. We consider autonomous devices that aim at minimizing their own cost by choosing when to perform their tasks and whether or not to offload their tasks to an edge cloud through one of the multiple wireless links. We develop a game theoretical model of the problem, prove the existence of pure strategy Nash equilibria and propose a polynomial complexity algorithm for computing an equilibrium. Furthermore, we characterize the structure of the equilibria, and by providing an upper bound on the price of anarchy of the game we establish an asymptotically tight bound on the approximation ratio of the proposed algorithm. Our simulation results show that the proposed algorithm achieves significant performance gain compared to uncoordinated computation offloading at a computational complexity that is on average linear in the number of devices.

Journal ArticleDOI
TL;DR: This paper designs joint charging and scheduling schemes to maximize the Quality of Monitoring (QoM) for stochastic events, which arrive and depart according to known probability distributions of time.
Abstract: In this paper, we consider the scenario in which a mobile charger (MC) periodically travels within a sensor network to recharge the sensors wirelessly. We design joint charging and scheduling schemes to maximize the Quality of Monitoring (QoM) for stochastic events, which arrive and depart according to known probability distributions of time. Information is considered captured if it is sensed by at least one sensor. We focus on two closely related research issues, i.e., how to choose the sensors for charging and decide the charging time for each of them, and how to schedule the sensors’ activation schedules according to their received energy. We formulate our problem as the maximum QoM CHA rging and S ch E duling problem (CHASE). We first ignore the MC's travel time and study the resulting relaxed version of the problem, which we call CHASE-R. We show that both CHASE and CHASE-R are NP-hard. For CHASE-R, we prove that it can be formulated as a submodular function maximization problem, which allows two algorithms to achieve $1/6$ 1 / 6 - and $1/(4 + \epsilon)$ 1 / ( 4 + e ) -approximation ratios. Then, for CHASE, we propose approximation algorithms to solve it by extending the CHASE-R results. We conduct simulations to validate our algorithm design.

Posted Content
TL;DR: For some > 10−36 the authors give a randomized 3/2− approximation algorithm for metric TSP, which is equivalent to a randomized 2/3− approximation for standard TSP.
Abstract: For some $\epsilon > 10^{-36}$ we give a $3/2-\epsilon$ approximation algorithm for metric TSP.

Journal ArticleDOI
TL;DR: In this article, a differentially private ADMM-based distributed learning algorithm called DP-ADMM is proposed, which combines an approximate augmented Lagrangian function with time-varying Gaussian noise addition in the iterative process.
Abstract: Alternating direction method of multipliers (ADMM) is a widely used tool for machine learning in distri-buted settings where a machine learning model is trained over distributed data sources through an interactive process of local computation and message passing. Such an iterative process could cause privacy concerns of data owners. The goal of this paper is to provide differential privacy for ADMM-based distributed machine learning. Prior approaches on differentially private ADMM exhibit low utility under high privacy guarantee and assume the objective functions of the learning problems to be smooth and strongly convex. To address these concerns, we propose a novel differentially private ADMM-based distributed learning algorithm called DP-ADMM, which combines an approximate augmented Lagrangian function with time-varying Gaussian noise addition in the iterative process to achieve higher utility for general objective functions under the same differential privacy guarantee. We also apply the moments accountant method to analyze the end-to-end privacy loss. The theoretical analysis shows that the DP-ADMM can be applied to a wider class of distributed learning problems, is provably convergent, and offers an explicit utility-privacy tradeoff. To our knowledge, this is the first paper to provide explicit convergence and utility properties for differentially private ADMM-based distributed learning algorithms. The evaluation results demonstrate that our approach can achieve good convergence and model accuracy under high end-to-end differential privacy guarantee.

Journal ArticleDOI
TL;DR: The proposed task allocation algorithm (Energy-aware Task Allocation in Multi-Cloud Networks (ETAMCN) minimizes the overall energy consumption and also reduces the makespan), which improves the average energy consumption through ETAMCN.
Abstract: In recent years, the growth rate of Cloud computing technology is increasing exponentially, mainly for its extraordinary services with expanding computation power, the possibility of massive storage, and all other services with the maintained quality of services (QoSs). The task allocation is one of the best solutions to improve different performance parameters in the cloud, but when multiple heterogeneous clouds come into the picture, the allocation problem becomes more challenging. This research work proposed a resource-based task allocation algorithm. The same is implemented and analyzed to understand the improved performance of the heterogeneous multi-cloud network. The proposed task allocation algorithm (Energy-aware Task Allocation in Multi-Cloud Networks ( ETAMCN )) minimizes the overall energy consumption and also reduces the makespan. The results show that the makespan is approximately overlapped for different tasks and does not show a significant difference. However, the average energy consumption improved through ETAMCN is approximately 14%, 6.3%, and 2.8% in opposed to the random allocation algorithm, Cloud Z-Score Normalization ( CZSN ) algorithm, and multi-objective scheduling algorithm with Fuzzy resource utilization (FR-MOS), respectively. An observation of the average SLA-violation of ETAMCN for different scenarios is performed.