scispace - formally typeset
Search or ask a question

Showing papers on "Greedy algorithm published in 2020"


Journal ArticleDOI
TL;DR: This work jointly optimize the trajectory of a UAV and the radio resource allocation to maximize the number of served IoT devices, where each device has its own target data upload deadline.
Abstract: The global evolution of wireless technologies and intelligent sensing devices are transforming the realization of smart cities. Among the myriad of use cases, there is a need to support applications whereby low-resource IoT devices need to upload their sensor data to a remote control centre by target hard deadlines; otherwise, the data becomes outdated and loses its value, for example, in emergency or industrial control scenarios. In addition, the IoT devices can be either located in remote areas with limited wireless coverage or in dense areas with relatively low quality of service. This motivates the utilization of UAVs to offload traffic from existing wireless networks by collecting data from time-constrained IoT devices with performance guarantees. To this end, we jointly optimize the trajectory of a UAV and the radio resource allocation to maximize the number of served IoT devices, where each device has its own target data upload deadline. The formulated optimization problem is shown to be mixed integer non-convex and generally NP-hard. To solve it, we first propose the high-complexity branch, reduce and bound (BRB) algorithm to find the global optimal solution for relatively small scale scenarios. Then, we develop an effective sub-optimal algorithm based on successive convex approximation in order to obtain results for larger networks. Next, we propose an extension algorithm to further minimize the UAV’s flight distance for cases where the initial and final UAV locations are known a priori. We demonstrate the favourable characteristics of the algorithms via extensive simulations and analysis as a function of various system parameters, with benchmarking against two greedy algorithms based on distance and deadline metrics.

259 citations


Journal ArticleDOI
TL;DR: A selective model aggregation approach is proposed, where “fine” local DNN models are selected and sent to the central server by evaluating the local image quality and computation capability, and demonstrated to outperform the original federated averaging approach in terms of accuracy and efficiency.
Abstract: Federated learning is a newly emerged distributed machine learning paradigm, where the clients are allowed to individually train local deep neural network (DNN) models with local data and then jointly aggregate a global DNN model at the central server. Vehicular edge computing (VEC) aims at exploiting the computation and communication resources at the edge of vehicular networks. Federated learning in VEC is promising to meet the ever-increasing demands of artificial intelligence (AI) applications in intelligent connected vehicles (ICV). Considering image classification as a typical AI application in VEC, the diversity of image quality and computation capability in vehicular clients potentially affects the accuracy and efficiency of federated learning. Accordingly, we propose a selective model aggregation approach, where “fine” local DNN models are selected and sent to the central server by evaluating the local image quality and computation capability. Regarding the implementation of model selection, the central server is not aware of the image quality and computation capability in the vehicular clients, whose privacy is protected under such a federated learning framework. To overcome this information asymmetry, we employ two-dimension contract theory as a distributed framework to facilitate the interactions between the central server and vehicular clients. The formulated problem is then transformed into a tractable problem through successively relaxing and simplifying the constraints, and eventually solved by a greedy algorithm. Using two datasets, i.e., MNIST and BelgiumTSC, our selective model aggregation approach is demonstrated to outperform the original federated averaging (FedAvg) approach in terms of accuracy and efficiency. Meanwhile, our approach also achieves higher utility at the central server compared with the baseline approaches.

235 citations


Journal ArticleDOI
TL;DR: A two-layer optimization method for jointly optimizing the deployment of UAVs and task scheduling and an efficient greedy algorithm is presented to obtain the near-optimal solution with much less time with the aim of minimizing system energy consumption.
Abstract: This article establishes a new multiunmanned aerial vehicle (multi-UAV)-enabled mobile edge computing (MEC) system, where a number of unmanned aerial vehicles (UAVs) are deployed as flying edge clouds for large-scale mobile users. In this system, we need to optimize the deployment of UAVs, by considering their number and locations. At the same time, to provide good services for all mobile users, it is necessary to optimize task scheduling. Specifically, for each mobile user, we need to determine whether its task is executed locally or on a UAV (i.e., offloading decision), and how many resources should be allocated (i.e., resource allocation). This article presents a two-layer optimization method for jointly optimizing the deployment of UAVs and task scheduling, with the aim of minimizing system energy consumption. By analyzing this system, we obtain the following property: the number of UAVs should be as small as possible under the condition that all tasks can be completed. Based on this property, in the upper layer, we propose a differential evolution algorithm with an elimination operator to optimize the deployment of UAVs, in which each individual represents a UAV’s location and the entire population represents an entire deployment of UAVs. During the evolution, we first determine the maximum number of UAVs. Subsequently, the elimination operator gradually reduces the number of UAVs until at least one task cannot be executed under delay constraints. This process achieves an adaptive adjustment of the number of UAVs. In the lower layer, based on the given deployment of UAVs, we transform the task scheduling into a 0-1 integer programming problem. Due to the large-scale characteristic of this 0-1 integer programming problem, we propose an efficient greedy algorithm to obtain the near-optimal solution with much less time. The effectiveness of the proposed two-layer optimization method and the established multi-UAV-enabled MEC system is demonstrated on ten instances with up to 1000 mobile users.

156 citations


Journal ArticleDOI
TL;DR: It is proven that the fourth measure, called relative neighborhood self-information, is better for feature selection than the other measures, because not only does it consider both the lower and the upper approximations but also the change of its magnitude is largest with the variation of feature subsets.
Abstract: The concept of dependency in a neighborhood rough set model is an important evaluation function for the feature selection. This function considers only the classification information contained in the lower approximation of the decision while ignoring the upper approximation. In this paper, we construct a class of uncertainty measures: decision self-information for the feature selection. These measures take into account the uncertainty information in the lower and the upper approximations. The relationships between these measures and their properties are discussed in detail. It is proven that the fourth measure, called relative neighborhood self-information, is better for feature selection than the other measures, because not only does it consider both the lower and the upper approximations but also the change of its magnitude is largest with the variation of feature subsets. This helps to facilitate the selection of optimal feature subsets. Finally, a greedy algorithm for feature selection has been designed and a series of numerical experiments was carried out to verify the effectiveness of the proposed algorithm. The experimental results show that the proposed algorithm often chooses fewer features and improves the classification accuracy in most cases.

147 citations


Journal ArticleDOI
TL;DR: A complete coverage path planning model trained using deep blackreinforcement learning (RL) for the tetromino based reconfigurable robot platform called hTetro to simultaneously generate the optimal set of shapes for any pretrained arbitrary environment shape with a trajectory that has the least overall cost.

97 citations


Proceedings ArticleDOI
06 Jul 2020
TL;DR: This work designs a novel two-stage latency-aware VNF deployment scheme, highlighted by a constrained depth-first search algorithm (CDFSA) for selecting paths, and a path-based greedy algorithm (PGA) for assigning V NFs by reusing as many VNFs as possible.
Abstract: With the increasing demand of low-latency network services, mobile edge computing (MEC) emerges as a new paradigm, which provides server resources and processing capacities in close proximity to end users. Based on network function virtualization (NFV), network services can be flexibly provisioned as virtual network function (VNF) chains deployed at edge servers. However, due to the resource shortage at the network edge, how to efficiently deploy VNF chains with latency guarantees and resource efficiency remains as a challenging problem. In this work, we focus on jointly optimizing the resource utilization of both edge servers and physical links under the latency limitations. Specifically, we formulate the VNF chain deployment problem as a mixed integer linear programming (MILP) to minimize the total resource consumption. We design a novel two-stage latency-aware VNF deployment scheme: highlighted by a constrained depth-first search algorithm (CDFSA) for selecting paths, and a path-based greedy algorithm (PGA) for assigning VNFs by reusing as many VNFs as possible. We demonstrate that our proposed algorithm can efficiently achieve a near-optimal solution with a theoretically-proved worstcase performance bound. Extensive simulation results show that the proposed algorithm outperforms three previous heuristic algorithms.

96 citations


Journal ArticleDOI
TL;DR: A novel resource provisioning mechanism and a workflow scheduling algorithm, named Greedy Resource Provisioning and modified HEFT (GRP-HEFT), for minimizing the makespan of a given workflow subject to a budget constraint for the hourly-based cost model of modern IaaS clouds.
Abstract: In Infrastructure as a Service (IaaS) Clouds, users are charged to utilize cloud services according to a pay-per-use model. If users intend to run their workflow applications on cloud resources within a specific budget, they have to adjust their demands for cloud resources with respect to this budget. Although several scheduling approaches have introduced solutions to optimize the makespan of workflows on a set of heterogeneous IaaS cloud resources within a certain budget, the hourly-based cost model of some well-known cloud providers (e.g., Amazon EC2 Cloud) can easily lead to a higher makespan and some schedulers may not find any feasible solution. In this article, we propose a novel resource provisioning mechanism and a workflow scheduling algorithm, named Greedy Resource Provisioning and modified HEFT (GRP-HEFT), for minimizing the makespan of a given workflow subject to a budget constraint for the hourly-based cost model of modern IaaS clouds. As a resource provisioning mechanism, we propose a greedy algorithm which lists the instance types according to their efficiency rate. For our scheduler, we modified the HEFT algorithm to consider a budget limit. GRP-HEFT is compared against state-of-the-art workflow scheduling techniques, including MOACS (Multi-Objective Ant Colony System), PSO (Particle Swarm Optimization), and GA (Genetic Algorithm). The experimental results demonstrate that GRP-HEFT outperforms GA, PSO, and MOACS for several well-known scientific workflow applications for different problem sizes on average by 13.64, 19.77, and 11.69 percent, respectively. Also in terms of time complexity, GRP-HEFT outperforms GA, PSO and MOACS.

95 citations


Journal ArticleDOI
TL;DR: An autonomous underwater vehicle-assisted data gathering scheme based on clustering and matrix completion (ACMC) to improve the data gathering efficiency in the underwater wireless sensor network (UWSN) and presents an in-cluster data collection mechanism based on matrix completion.
Abstract: The oceans cover more than 71% of the Earth’s surface and have a surging amount of data. It is of great significance to seek energy-effective and ultrareliable communication and transmission mechanism for effectively gathering abundant maritime data. In this article, we propose an autonomous underwater vehicle (AUV)-assisted data gathering scheme based on clustering and matrix completion (ACMC) to improve the data gathering efficiency in the underwater wireless sensor network (UWSN). Specifically, we first improve the $K$ -means algorithm by adopting the Elbow method to determine the optimal $K$ and setting a distance threshold to select the separate initial cluster centers. Then, we introduce a two-phase AUV trajectory optimization mechanism to effectively reduce the trajectory length of the AUV. In the first phase, the optimized trajectory of the AUV is planned by adopting the greedy algorithm. In the second phase, the ordinary nodes close to the AUV trajectory are selected as secondary cluster heads to share the workload of cluster heads. Finally, we present an in-cluster data collection mechanism based on matrix completion. An extensive experiment validates the effectiveness of our proposed scheme in terms of energy and data collection delay.

88 citations


Journal ArticleDOI
TL;DR: This paper considers the cellular Internet of unmanned aerial vehicles, where UAVs sense data with onboard sensors for multiple sensing tasks and transmit the data to the base station, and proposes an iterative algorithm to optimize the sensing time, transmission time, and UAV velocity for completing a specific task.
Abstract: In this paper, we consider the cellular Internet of unmanned aerial vehicles (UAVs), where UAVs sense data with onboard sensors for multiple sensing tasks and transmit the data to the base station (BS). To quantify the “freshness” of the data at the BS, we bring in the concept of the age of information (AoI). The AoI is determined by the time for UAV sensing and that for UAV transmission, which gives rise to a trade-off within a given period. To minimize the AoI, we formulate a joint sensing time, transmission time, UAV trajectory, and task scheduling optimization problem. This NP-hard problem can be decoupled into two subproblems. We first propose an iterative algorithm to optimize the sensing time, transmission time, and UAV velocity for completing a specific task. Afterwards, we design the order in which the UAV performs data updates for multiple sensing tasks. The convergence and complexity of the proposed algorithm, together with the trade-off between UAV sensing and UAV transmission, are analyzed. Simulation results show that the AoI with the proposed algorithm is about 15% lower than that of the greedy algorithm, and over 40% lower than that of the random algorithm.

80 citations


Journal ArticleDOI
TL;DR: This paper proposes a reconfigurable service provisioning framework based on service function chaining (SFC) for SAGIN, and forms the SFC planning problem as an integer non-linear programming problem, which is NP-hard.
Abstract: Space-air-ground integrated networks (SAGIN) extend the capability of wireless networks and will be the essential building block for many advanced applications, like autonomous driving, earth monitoring, and etc. However, coordinating heterogeneous physical resources is very challenging in such a large-scale dynamic network. In this paper, we propose a reconfigurable service provisioning framework based on service function chaining (SFC) for SAGIN. In SFC, the network functions are virtualized and the service data needs to flow through specific network functions in a predefined sequence. The inherent issue is how to plan the service function chains over large-scale heterogeneous networks, subject to the resource limitations of both communication and computation. Specifically, we must jointly consider the virtual network functions (VNFs) embedding and service data routing. We formulate the SFC planning problem as an integer non-linear programming problem, which is NP-hard. Then, a heuristic greedy algorithm is proposed, which concentrates on leveraging different features of aerial and ground nodes and balancing the resource consumptions. Furthermore, a new metric, aggregation ratio (AR) is proposed to elaborate the communication-computation tradeoff. Extensive simulations shows that our proposed algorithm achieves near-optimal performance. We also find that the SAGIN significantly reduces the service blockage probability and improves the efficiency of resource utilization. Finally, a case study on multiple intersection traffic scheduling is provided to demonstrate the effectiveness of our proposed SFC-based service provisioning framework.

75 citations


Journal ArticleDOI
TL;DR: A randomized approximation algorithm which is provably superior to the state-of-the art methods with respect to running time is presented.
Abstract: Social networks allow rapid spread of ideas and innovations while negative information can also propagate widely. When a user receives two opposing opinions, they tend to believe the one arrives first. Therefore, once misinformation or rumor is detected, one containment method is to introduce a positive cascade competing against the rumor. Given a budget $k$ , the rumor blocking problem asks for $k$ seed users to trigger the spread of a positive cascade such that the number of the users who are not influenced by rumor can be maximized. The prior works have shown that the rumor blocking problem can be approximated within a factor of $(1-1/e)$ by a classic greedy algorithm combined with Monte Carlo simulation. Unfortunately, the Monte Carlo simulation based methods are time consuming and the existing algorithms either trade performance guarantees for practical efficiency or vice versa. In this paper, we present a randomized approximation algorithm which is provably superior to the state-of-the art methods with respect to running time. The superiority of the proposed algorithm is demonstrated by experiments done on both the real-world and synthetic social networks.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a neural network architecture that utilizes the structure of the power grid to reduce the number of coefficients needed to parameterize the mapping from the measurements to the network state by exploiting separability of the estimation problem.
Abstract: The distribution system state estimation problem seeks to determine the network state from available measurements. Widely used Gauss-Newton approaches are very sensitive to the initialization and often not suitable for real-time estimation. Learning approaches are very promising for real-time estimation, as they shift the computational burden to an offline training stage. Prior machine learning approaches to power system state estimation have been electrical model-agnostic, in the sense that they did not exploit the topology and physical laws governing the power grid to design the architecture of the learning model. In this paper, we propose a novel learning model that utilizes the structure of the power grid. The proposed neural network architecture reduces the number of coefficients needed to parameterize the mapping from the measurements to the network state by exploiting the separability of the estimation problem. This prevents overfitting and reduces the complexity of the training stage. We also propose a greedy algorithm for phasor measuring units placement that aims at minimizing the complexity of the neural network required for realizing the state estimation mapping. Simulation results show superior performance of the proposed method over the Gauss-Newton approach.

Journal ArticleDOI
TL;DR: Comprehensive evaluations demonstrate that the Lagreedy algorithm is able to obtain the shortest delay with a high power consumption, while the branch-and-bound algorithm can achieve both shorter delay and lower power consumption with reliability guarantees.
Abstract: Computation offloading over fog computing has the potential to improve reliability and reduce latency in future networks. This paper considers a scenario where roadside units (RSUs) are installed for offloading tasks to the computation nodes including nearby fog nodes and a cloud center. To guarantee the reliable communication, we formulate the first subproblem of power allocation, and leverage the conditional value-at-risk approach to analyze the successful transmission probability in the worse-case channel condition. To complete computation tasks with low latency, we formulate the second subproblem of task allocation into a multi-period generalized assignment problem (MPGAP), which aims at minimizing the total delay by offloading tasks to the ‘right’ fog nodes at ‘right’ period. Then, we propose a modified branch-and-bound algorithm to derive the optimal solution and a heuristic greedy algorithm to obtain approximate performance. In addition, the master problem is proposed as a non-convex optimization problem, which considers both the reliability-guaranteed and delay-sensitive requirements. We design the Lagreedy algorithm by combining the subgradient algorithm and the heuristic algorithm. Comprehensive evaluations demonstrate that the Lagreedy is able to obtain the shortest delay with a high power consumption, while the branch-and-bound algorithm can achieve both shorter delay and lower power consumption with reliability guarantees.

Journal ArticleDOI
TL;DR: This work presents a deep reinforcement learning method for solving the global routing problem in a simulated environment and indicates that the approach can outperform the benchmark method of a sequential A* method, suggesting a promising potential forDeep reinforcement learning for global routing and other routing or path planning problems in general.
Abstract: Global routing has been a historically challenging problem in electronic circuit design, where the challenge is to connect a large and arbitrary number of circuit components with wires without violating the design rules for the printed circuit boards or integrated circuits. Similar routing problems also exist in the design of complex hydraulic systems, pipe systems and logistic networks. Existing solutions typically consist of greedy algorithms and hard-coded heuristics. As such, existing approaches suffer from a lack of model flexibility and non-optimum solutions. As an alternative approach, this work presents a deep reinforcement learning method for solving the global routing problem in a simulated environment. At the heart of the proposed method is deep reinforcement learning that enables an agent to produce an optimal policy for routing based on the variety of problems it is presented with leveraging the conjoint optimization mechanism of deep reinforcement learning. Conjoint optimization mechanism is explained and demonstrated in details; the best network structure and the parameters of the learned model are explored. Based on the fine-tuned model, routing solutions and rewards are presented and analyzed. The results indicate that the approach can outperform the benchmark method of a sequential A* method, suggesting a promising potential for deep reinforcement learning for global routing and other routing or path planning problems in general. Another major contribution of this work is the development of a global routing problem sets generator with the ability to generate parameterized global routing problem sets with different size and constraints, enabling evaluation of different routing algorithms and the generation of training datasets for future data-driven routing approaches.

Journal ArticleDOI
TL;DR: Experiments conducted on a set of real-world online social networks confirm that the proposed biobjective optimization model and the developed MOEA/D-ADACO are promising for the pollutant spreading control.
Abstract: The rapid development of online social networks not only enables prompt and convenient dissemination of desirable information but also incurs fast and wide propagation of undesirable information. A common way to control the spread of pollutants is to block some nodes, but such a strategy may affect the service quality of a social network and leads to a high control cost if too many nodes are blocked. This paper considers the node selection problem as a biobjective optimization problem to find a subset of nodes to be blocked so that the effect of the control is maximized while the cost of the control is minimized. To solve this problem, we design an ant colony optimization algorithm with an adaptive dimension size selection under the multiobjective evolutionary algorithm framework based on decomposition (MOEA/D-ADACO). The proposed algorithm divides the biobjective problem into a set of single-objective subproblems and each ant takes charge of optimizing one subproblem. Moreover, two types of pheromone and heuristic information are incorporated into MOEA/D-ADACO, that is, pheromone and heuristic information of dimension size selection and that of node selection. While constructing solutions, the ants first determine the dimension size according to the former type of pheromone and heuristic information. Then, the ants select a specific number of nodes to build solutions according to the latter type of pheromone and heuristic information. Experiments conducted on a set of real-world online social networks confirm that the proposed biobjective optimization model and the developed MOEA/D-ADACO are promising for the pollutant spreading control.

Journal ArticleDOI
TL;DR: A dynamic greedy search strategy was developed to avoid blind searching in traditional strategy and weighted iteration update of the Q function, including the weighted mean of the maximum fuzzy earning, was designed to improve the speed and accuracy of the improved learning algorithm.
Abstract: Given the dynamic and uncertain production environment of job shops, a scheduling strategy with adaptive features must be developed to fit variational production factors. Therefore, a dynamic scheduling system model based on multi-agent technology, including machine, buffer, state, and job agents, was built. A weighted Q-learning algorithm based on clustering and dynamic search was used to determine the most suitable operation and to optimize production. To address the large state space problem caused by changes in the system state, four state features were extracted. The dimension of the system state was decreased through the clustering method. To reduce the error between the actual system states and clustering ones, the state difference degree was defined and integrated with the iteration formula of the Q function. To select the optimal state-action pair, improved search and iteration update strategies were proposed. Convergence analysis of the proposed algorithm and simulation experiments indicated that the proposed adaptive strategy is well adaptable and effective in different scheduling environments, and shows better performance in complex environments. The two contributions of this research are as follows: (1) a dynamic greedy search strategy was developed to avoid blind searching in traditional strategy. (2) Weighted iteration update of the Q function, including the weighted mean of the maximum fuzzy earning, was designed to improve the speed and accuracy of the improved learning algorithm.

Journal ArticleDOI
TL;DR: Experimental results show that the proposed scheme, Soft-VAN, is capable of enhancing the performance approximately by 30%, 45%, and 50% in terms of delay compared to state-of-the-art schemes — Detour, DAGP, and SD2O, respectively.
Abstract: In this paper, we propose a mobility-aware task offloading scheme, named as Soft-VAN , with an aim to minimize task computation delay in a software-defined vehicular network. The proposed scheme consists of two phases — fog node selection and task offloading. In the first phase, we formulate an integer linear program (ILP), and solve the problem to get optimal number of fog nodes required for a given network. In the task offloading phase, we formulate an optimization problem to minimize overall delay in task computation, while considering associated constraints. As finding optimal solution to the problem is NP-hard, we propose a greedy heuristic approach in two phases — task offloading and computed task downloading — to solve it in polynomial time. The greedy solution for offloading takes into account network delay, flow-rule capacity, and link utilization. On the other hand, the solution for computed task downloading considers vehicle's mobility in addition to the parameters associated with the offloading decisions. Experimental results show that the proposed scheme, Soft-VAN, is capable of enhancing the performance approximately by 30%, 45%, and 50% in terms of delay compared to state-of-the-art schemes — Detour, DAGP, and SD2O, respectively.

Posted Content
TL;DR: LA-MCTS as discussed by the authors adopts a nonlinear decision boundary and learns a local model to pick good candidates to reduce the sample complexity empirically, and it achieves good performance in neural architecture search.
Abstract: High dimensional black-box optimization has broad applications but remains a challenging problem to solve. Given a set of samples $\{\vx_i, y_i\}$, building a global model (like Bayesian Optimization (BO)) suffers from the curse of dimensionality in the high-dimensional search space, while a greedy search may lead to sub-optimality. By recursively splitting the search space into regions with high/low function values, recent works like LaNAS shows good performance in Neural Architecture Search (NAS), reducing the sample complexity empirically. In this paper, we coin LA-MCTS that extends LaNAS to other domains. Unlike previous approaches, LA-MCTS learns the partition of the search space using a few samples and their function values in an online fashion. While LaNAS uses linear partition and performs uniform sampling in each region, our LA-MCTS adopts a nonlinear decision boundary and learns a local model to pick good candidates. If the nonlinear partition function and the local model fits well with ground-truth black-box function, then good partitions and candidates can be reached with much fewer samples. LA-MCTS serves as a \emph{meta-algorithm} by using existing black-box optimizers (e.g., BO, TuRBO) as its local models, achieving strong performance in general black-box optimization and reinforcement learning benchmarks, in particular for high-dimensional problems.

Proceedings ArticleDOI
11 May 2020
TL;DR: A genetic algorithm is proposed to solve the optimization problem of UAV path planning as a traveling salesman problem to minimize the energy consumption for the UAV to complete a task.
Abstract: Unmanned Aerial Vehicles (UAVs) have been increasingly used in environmental sensing and surveying applications. Coverage path planning to survey an area while following a set of waypoints is required to complete a task. Due to the battery capacity, the UAV flight time is often limited. In this paper, we formulate the UAV path planning problem as a traveling salesman problem in order to optimize UAV energy. We propose a genetic algorithm to solve the optimization problem i.e. to minimize the energy consumption for the UAV to complete a task. We also consider reducing the number of turns to allow the UAV to optimize the flight path and to minimize its energy consumption. We compare the energy consumption of the proposed genetic algorithm to the greedy algorithm with different number of waypoints. Results show that our proposed algorithm consumes 2–5 times less energy than that of the greedy algorithm by reducing the number of turns while covering all the waypoints.

Proceedings ArticleDOI
20 Apr 2020
TL;DR: This work studies a novel spatial crowdsourcing problem, namely Predictive Task Assignment (PTA), which aims to maximize the number of assigned tasks by taking into account both current and future workers/tasks that enter the system dynamically with location unknown in advance and proposes a two-phase data-driven framework.
Abstract: With the rapid development of mobile networks and the widespread usage of mobile devices, spatial crowdsourcing, which refers to assigning location-based tasks to moving workers, has drawn increasing attention. One of the major issues in spatial crowdsourcing is task assignment, which allocates tasks to appropriate workers. However, existing works generally assume the static offline scenarios, where the spatio-temporal information of all the workers and tasks is determined and known a priori. Ignorance of the dynamic spatio-temporal distributions of workers and tasks can often lead to poor assignment results. In this work we study a novel spatial crowdsourcing problem, namely Predictive Task Assignment (PTA), which aims to maximize the number of assigned tasks by taking into account both current and future workers/tasks that enter the system dynamically with location unknown in advance. We propose a two-phase data-driven framework. The prediction phase hybrids different learning models to predict the locations and routes of future workers and designs a graph embedding approach to estimate the distribution of future tasks. In the assignment component, we propose both greedy algorithm for large-scale applications and optimal algorithm with graph partition based decomposition. Extensive experiments on two real datasets demonstrate the effectiveness of our framework.

Journal ArticleDOI
TL;DR: This work presents a hybrid algorithm called parallel simulated annealing with a greedy algorithm (PSAGA) to learn Bayesian network structures and demonstrates that the proposed PSAGA shows better performance than the alternatives in terms of computational time and accuracy.
Abstract: We present a hybrid algorithm called parallel simulated annealing with a greedy algorithm (PSAGA) to learn Bayesian network structures. This work focuses on simulated annealing and its parallelization with memoization to accelerate the search process. At each step of the local search, a hybrid search method combining simulated annealing with a greedy algorithm was adopted. The proposed PSAGA aims to achieve both the efficiency of parallel search and the effectiveness of a more exhaustive search. The Bayesian Dirichlet equivalence metric was used to determine an optimal structure for PSAGA. The proposed PSAGA was evaluated on seven well-known Bayesian network benchmarks generated at random. We first conducted experiments to evaluate the computational time performance of the proposed parallel search. We then compared PSAGA with existing variants of simulated annealing-based algorithms to evaluate the quality of the learned structure. Overall, the experimental results demonstrate that the proposed PSAGA shows better performance than the alternatives in terms of computational time and accuracy.

Journal ArticleDOI
TL;DR: A Restarted Iterated Pareto Greedy algorithm is designed to optimize both objectives of efficient task and worker assignment and a reduction in ergonomic risks in U-shaped assembly lines to simultaneously minimize cycle times and ergonomic risk.

Journal ArticleDOI
TL;DR: It is demonstrated that the greedy-choice property applies, which means that a globally optimal solution can be achieved by making locally optimal greedy choices, whereas it does not apply to Objective 2, and a non-myopic charging strategy accounting for future demands to achieve global optimality is proposed.
Abstract: Electric vehicles (EVs) endow great potentials for future transportation systems, while efficient charge scheduling strategies are crucial for improving profits and mass adoption of EVs. Two critical and open issues concerning EV charging are how to minimize the total charging cost (Objective 1) and how to minimize the peak load (Objective 2). Although extensive efforts have been made to model EV charging problems, little information is available about model properties and efficient algorithms for dynamic charging problems. This paper aims to fill these gaps. For Objective 1, we demonstrate that the greedy-choice property applies, which means that a globally optimal solution can be achieved by making locally optimal greedy choices, whereas it does not apply to Objective 2. We propose a non-myopic charging strategy accounting for future demands to achieve global optimality for Objective 2. The problem is addressed by a heuristic algorithm combining a multi-commodity network flow model with customized bisection search algorithm in a rolling horizon framework. To expedite the solution efficiency, we derive the upper bound and lower bound in the bisection search based on the relationship between charging volume and parking time. We also explore the impact of demand levels and peak arrival ratios on the system performance. Results show that with prediction, the peak load can converge to a globally optimal solution, and that an optimal look-ahead time exists beyond which any prediction is ineffective. The proposed algorithm outperforms the state-of-the-art algorithms, and is robust to the variations of demand and peak arrival ratios.

Proceedings Article
30 Apr 2020
TL;DR: A novel score-based approach to learning a directed acyclic graph (DAG) from observational data that outperforms current continuous methods on most tasks, while being competitive with existing greedy search methods on important metrics for causal inference.
Abstract: We propose a novel score-based approach to learning a directed acyclic graph (DAG) from observational data. We adapt a recently proposed continuous constrained optimization formulation to allow for nonlinear relationships between variables using neural networks. This extension allows to model complex interactions while avoiding the combinatorial nature of the problem. In addition to comparing our method to existing continuous optimization methods, we provide missing empirical comparisons to nonlinear greedy search methods. On both synthetic and real-world data sets, this new method outperforms current continuous methods on most tasks while being competitive with existing greedy search methods on important metrics for causal inference.

Journal ArticleDOI
TL;DR: An intelligent Task Offloading Algorithm (iTOA) for UAV edge computing network that is able to perceive the network’s environment intelligently to decide the offloading action based on deep Monte Calor Tree Search (MCTS), the core algorithm of Alpha Go.

Journal ArticleDOI
TL;DR: A novel framework for compressing trajectory data, REST (Reference-based Spatio-temporal trajectory compression), by which a raw trajectory is represented by concatenation of a series of historical (sub-)trajectories (called reference trajectories) that form the compressed trajectory within a given spatio-tem temporal deviation threshold.
Abstract: The pervasiveness of GPS-enabled devices and wireless communication technologies results in massive trajectory data, incurring expensive cost for storage, transmission, and query processing. To relieve this problem, in this paper we propose a novel framework for compressing trajectory data, REST ( Re ference-based S patio- t emporal trajectory compression), by which a raw trajectory is represented by concatenation of a series of historical (sub-)trajectories (called reference trajectories) that form the compressed trajectory within a given spatio-temporal deviation threshold. In order to construct a reference trajectory set that can most benefit the subsequent compression, we propose three kinds of techniques to select reference trajectories wisely from a large dataset such that the resulting reference set is more compact yet covering most footprints of trajectories in the area of interest. To address the computational issue caused by the large number of combinations of reference trajectories that may exist for resembling a given trajectory, we propose efficient greedy algorithms that run in the blink of an eye and dynamic programming algorithms that can achieve the optimal compression ratio. Compared to existing work on trajectory compression, our framework has few assumptions about data such as moving within a road network or moving with constant direction and speed, and better compression performance with fairly small spatio-temporal loss. In addition, by indexing the reference trajectories directly with an in-memory R-tree and building connections to the raw trajectories with inverted index, we develop an extremely efficient algorithm that can answer spatio-temporal range queries over trajectories in their compressed form. Extensive experiments on a real taxi trajectory dataset demonstrate the superiority of our framework over existing representative approaches in terms of both compression ratio and efficiency.

Journal ArticleDOI
01 Mar 2020
TL;DR: This paper investigates the Maximum Coverage Sets Scheduling (MCSS) problem, and proposes a greedy algorithm, called Greedy-MCSS, based on which an approximation algorithm, MCSSA, is proposed for solving the MCSS problem, which gives the theoretical performance guarantee.
Abstract: In a Wireless Sensor Network (WSN), when a large amount of sensors are randomly deployed into a detection area, an efficient sleep/active scheduling for sensors to maximize the network lifetime of target (or detection area) coverage, which is called the coverage problem, is an important issue. The problem was proved NP-complete. Recently, many methods were proposed for solving the coverage problem, each of which can be divided into two phases: the first is to find as many as possible coverage sets from the sensors and the other is to schedule the coverage sets got from the first phase. Therefore, all coverage problems involve the scheduling process of the coverage sets to maximize the network lifetime. In this paper, we investigate the Maximum Coverage Sets Scheduling (MCSS) problem: given a coverage set collection in which each coverage set covers all targets (or the whole detection area) in WSN, the problem is to find a feasible scheduling for the coverage set collection to maximize the network lifetime. Firstly, we prove the MCSS problem is NP-hard. Secondly, we formulate the problem as an integer linear programming problem. Thirdly, we first propose a greedy algorithm, called Greedy-MCSS, to solve the MCSS problem. Then based on the Greedy-MCSS algorithm, we propose an approximation algorithm, MCSS Algorithm (MCSSA) for solving the MCSS problem, which gives the theoretical performance guarantee. Finally, extensive simulation results are shown to further verify the performance of our algorithms.

Journal ArticleDOI
TL;DR: Experimental results on both synthetic and real datasets show that the proposed TI-SC algorithm significantly outperforms the state-of-the-art algorithms in terms of efficiency in both small and large-scale datasets.
Abstract: Influence maximization is a classic optimization problem to find a subset of seed nodes in a social network that has a maximum influence with respect to a propagation model. This problem suffers from the overlap of seed nodes and the lack of optimal selection of seed nodes. Kempe et al. have shown that this problem is an NP-hard problem, and the objective function is submodular. Therefore, some heuristic and greedy algorithms have been proposed to find a near-optimal solution. However, the greedy algorithm may not satisfy the accuracy of a given solution and high time-consuming problem. To overcome these problems, the TI-SC algorithm is proposed for the problem of influence maximization. The TI-SC algorithm selects the influential nodes by examining the relationships between the core nodes and the scoring ability of other nodes. After selecting each seed node, the scores are updated to reduce the overlap in selecting the seed nodes. This algorithm has efficient performance in high Rich-Club networks. The Rich-Club phenomenon causes overlapping of the influence spread among the seed nodes in most of the other methods so that the TI-SC algorithm reduces this overlapping. Furthermore, the discovered communities with low expansion are not considered in the seed node selection phase, and this is useful for reducing computational overhead. Experimental results on both synthetic and real datasets show that the proposed TI-SC algorithm significantly outperforms the state-of-the-art algorithms in terms of efficiency in both small and large-scale datasets.

Journal ArticleDOI
TL;DR: This work tackles the classical traveling salesman problem (TSP) by combining a graph neural network and Monte Carlo Tree Search and adopts a greedy algorithm framework to derive a promising tour by adding the vertices successively.
Abstract: We tackle the classical traveling salesman problem (TSP) by combining a graph neural network and Monte Carlo Tree Search. We adopt a greedy algorithm framework to derive a promising tour by adding the vertices successively. A graph neural network is trained to capture graph motifs and interactions between vertices, and then to give the prior probability of selecting a vertex at every step. Instead of making decisions directly based on the output of graph neural networks, we combine the graph neural network with Monte Carlo Tree Search to provide a more reliable policy as the output of the latter is the feedback information by fusing the prior probability with the scouting exploration. Without much heuristic designing, our approach outperforms recent state-of-the-art learning-based methods on the TSP. Experimental results demonstrate that the proposed method can be generalized to instances with more vertices than those used during the training.

Book
18 Jun 2020
TL;DR: In this paper, the classical and modern time-dependent scheduling theories are discussed. But the classical scheduling theory does not consider the problem of time dependent scheduling under precedence constraints.
Abstract: Part I, Fundamentals.- Fundamentals.- Preliminaries.- Problems and Algorithms.- NP-Complete Problems.- Part II, Scheduling Models.- The Classical Scheduling Theory.- The Modern Scheduling Theory.- The Time-Dependent Scheduling.- Part III, Polynomial Problems.- Polynomial Single Machine Problems.- Polynomial Parallel Machine Problems.- Polynomial Dedicated Machine Problems.- Part IV, NP-Hard Problems.- NP-Hard Single Machine Problems.- NP-Hard Parallel Machine Problems.- NP-Hard Dedicated Machine Problems.- Part V, Algorithms.- Exact Algorithms.- Approximation Algorithms and Schemes.- Greedy Algorithms Based on Signatures.- Heuristic Algorithms.- Local Search and Meta-heuristic Algorithms.- Part VI, Advanced Topics.- Time-Dependent Scheduling Under Precedence Constraints.- Matrix Methods in Time-Dependent Scheduling.- Bi-criteria Time-Dependent Scheduling.- New Topics in Time-Dependent Scheduling.- App. A, Open Time-Dependent Scheduling Problems.- List of Algorithms.- List of Figures.- List of Tables.- Symbol Index.- Subject Index.