scispace - formally typeset
Search or ask a question

Showing papers on "Heuristic published in 2022"


Journal ArticleDOI
TL;DR: Particle swarm optimization (PSO) is one of the most well-regarded swarm-based algorithms in the literature as mentioned in this paper , and many researchers have been modifying it resulting in a large number of PSO variants with either slightly or significantly better performance.
Abstract: Particle swarm optimization (PSO) is one of the most well-regarded swarm-based algorithms in the literature. Although the original PSO has shown good optimization performance, it still severely suffers from premature convergence. As a result, many researchers have been modifying it resulting in a large number of PSO variants with either slightly or significantly better performance. Mainly, the standard PSO has been modified by four main strategies: modification of the PSO controlling parameters, hybridizing PSO with other well-known meta-heuristic algorithms such as genetic algorithm (GA) and differential evolution (DE), cooperation and multi-swarm techniques. This paper attempts to provide a comprehensive review of PSO, including the basic concepts of PSO, binary PSO, neighborhood topologies in PSO, recent and historical PSO variants, remarkable engineering applications of PSO, and its drawbacks. Moreover, this paper reviews recent studies that utilize PSO to solve feature selection problems. Finally, eight potential research directions that can help researchers further enhance the performance of PSO are provided.

99 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a new bio-inspired meta-heuristic algorithm, named artificial rabbits optimization (ARO), which is applied to the fault diagnosis of a rolling bearing, in which the back-propagation (BP) network optimized by ARO is developed.

89 citations


Journal ArticleDOI
01 Mar 2022
TL;DR: In this paper , an ant colony system (ACS)-based algorithm was proposed to obtain good enough paths for UAVs and fully cover all regions efficiently, inspired by the foraging behavior of ants that they can obtain the shortest path between their nest and food.
Abstract: Unmanned aerial vehicle (UAV) has been extensively studied and widely adopted in practical systems owing to its effectiveness and flexibility. Although heterogeneous UAVs have an enormous advantage in improving performance and conserving energy with respect to homogeneous ones, they give rise to a complex path planning problem. Especially in large-scale cooperative search systems with multiple separated regions, coverage path planning which seeks optimal paths for UAVs to completely visit and search all of regions of interest, has a NP-hard computation complexity and is difficult to settle. In this work, we focus on the coverage path planning problem of heterogeneous UAVs, and present an ant colony system (ACS)-based algorithm to obtain good enough paths for UAVs and fully cover all regions efficiently. First, models of UAVs and regions are built, and a linear programming-based formulation is presented to exactly provide the best point-to-point flight path for each UAV. Then, inspired by the foraging behaviour of ants that they can obtain the shortest path between their nest and food, an ACS-based heuristic is presented to seek approximately optimal solutions and minimize the time consumption of tasks in the cooperative search system. Experiments on randomly generated regions have been organized to evaluate the performance of the new heuristic in terms of execution time, task completion time and deviation ratio.

63 citations


Journal ArticleDOI
01 Mar 2022-Energy
TL;DR: In this paper , the authors proposed an optimized forecasting model-an extreme learning machine (ELM) model coupled with the heuristic Kalman filter (HKF) algorithm to forecast the capacity of supercapacitors.

59 citations


Journal ArticleDOI
01 Jul 2022
TL;DR: In this paper , a hierarchical multiobjective heuristic (HMOH) is proposed to optimize printed-circuit board assembly (PCBA) in a single beam-head surface mounter.
Abstract: This article proposes a hierarchical multiobjective heuristic (HMOH) to optimize printed-circuit board assembly (PCBA) in a single beam-head surface mounter. The beam-head surface mounter is the core facility in a high-mix and low-volume PCBA line. However, as a large-scale, complex, and multiobjective combinatorial optimization problem, the PCBA optimization of the beam-head surface mounter is still a challenge. This article provides a framework for optimizing all the interrelated objectives, which has not been achieved in the existing studies. A novel decomposition strategy is applied. This helps to closely model the real-world problem as the head task assignment problem (HTAP) and the pickup-and-place sequencing problem (PAPSP). These two models consider all the factors affecting the assembly time, including the number of pickup-and-place (PAP) cycles, nozzle changes, simultaneous pickups, and the PAP distances. Specifically, HTAP consists of the nozzle assignment and component allocation, while PAPSP comprises place allocation, feeder set assignment, and place sequencing problems. Adhering strictly to the lexicographic method, the HMOH solves these subproblems in a descending order of importance of their involved objectives. Exploiting the expert knowledge, each subproblem is solved by an elaborately designed heuristic. Finally, the proposed HMOH realizes the complete and optimal PCBA decision making in real time. Using industrial PCB datasets, the superiority of HMOH is elucidated through comparison with the built-in optimizer of the widely used Samsung SM482.

52 citations


Journal ArticleDOI
TL;DR: In this paper , an enhanced version of the Black Widow Optimization Algorithm called SDABWO is proposed to solve the feature selection problem, which has faster convergence speed and higher accuracy.
Abstract: Feature selection is an important data processing method to reduce dimension of the raw datasets while preserving the information as much as possible. In this paper, an enhanced version of Black Widow Optimization Algorithm called SDABWO is proposed to solve the feature selection problem. The Black Widow Optimization Algorithm (BWO) is a new population-based meta-heuristic algorithm inspired by the evolution process of spider population. Three main improvements were included into the BWO to overcome the shortcoming of low accuracy, slow convergence speed and being easy to fall into local optima. Firstly, a novel strategy for selecting spouses by calculating the weight of female spiders and the distance between spiders is proposed. By applying the strategy to the original algorithm, it has faster convergence speed and higher accuracy. The second improvement includes the use of mutation operator of differential evolution at mutation phase of BWO which helps the algorithm escape from the local optima. And then, three key parameters are set to adjust adaptively with the increase of iteration times. To confirm and validate the performance of the improved BWO, other 10 algorithms are used to compared with the SDABWO on 25 benchmark functions. The results show that the proposed algorithm enhances the exploitation ability, improves the convergence speed and is more stable when solving optimization problems. Furthermore, the proposed SDABWO algorithm is employed for feature selection. Twelve standard datasets from UCI repository prove that SDABWO-based method has stronger search ability in the search space of feature selection than the other five popular feature selection methods. These results confirm the capability of the proposed method simultaneously improve the classification accuracy while reducing the dimensions of the original datasets. Therefore, SDABWO-based method was found to be one of the most promising for feature selection problem over other approaches that are currently used in the literature.

51 citations


Journal ArticleDOI
TL;DR: In this paper, an enhanced version of the Black Widow Optimization Algorithm called SDABWO is proposed to solve the feature selection problem, which has faster convergence speed and higher accuracy.
Abstract: Feature selection is an important data processing method to reduce dimension of the raw datasets while preserving the information as much as possible. In this paper, an enhanced version of Black Widow Optimization Algorithm called SDABWO is proposed to solve the feature selection problem. The Black Widow Optimization Algorithm (BWO) is a new population-based meta-heuristic algorithm inspired by the evolution process of spider population. Three main improvements were included into the BWO to overcome the shortcoming of low accuracy, slow convergence speed and being easy to fall into local optima. Firstly, a novel strategy for selecting spouses by calculating the weight of female spiders and the distance between spiders is proposed. By applying the strategy to the original algorithm, it has faster convergence speed and higher accuracy. The second improvement includes the use of mutation operator of differential evolution at mutation phase of BWO which helps the algorithm escape from the local optima. And then, three key parameters are set to adjust adaptively with the increase of iteration times. To confirm and validate the performance of the improved BWO, other 10 algorithms are used to compared with the SDABWO on 25 benchmark functions. The results show that the proposed algorithm enhances the exploitation ability, improves the convergence speed and is more stable when solving optimization problems. Furthermore, the proposed SDABWO algorithm is employed for feature selection. Twelve standard datasets from UCI repository prove that SDABWO-based method has stronger search ability in the search space of feature selection than the other five popular feature selection methods. These results confirm the capability of the proposed method simultaneously improve the classification accuracy while reducing the dimensions of the original datasets. Therefore, SDABWO-based method was found to be one of the most promising for feature selection problem over other approaches that are currently used in the literature.

51 citations


Journal ArticleDOI
TL;DR: In this paper , the authors reviewed the sparrow search algorithm (SSA), one of the new and robust algorithms for solving optimization problems, and covered all the SSA literature on variants, improvement, hybridization, and optimization.
Abstract: Mathematical programming and meta-heuristics are two types of optimization methods. Meta-heuristic algorithms can identify optimal/near-optimal solutions by mimicking natural behaviours or occurrences and provide benefits such as simplicity of execution, a few parameters, avoidance of local optimization, and flexibility. Many meta-heuristic algorithms have been introduced to solve optimization issues, each of which has advantages and disadvantages. Studies and research on presented meta-heuristic algorithms in prestigious journals showed they had good performance in solving hybrid, improved and mutated problems. This paper reviews the sparrow search algorithm (SSA), one of the new and robust algorithms for solving optimization problems. This paper covers all the SSA literature on variants, improvement, hybridization, and optimization. According to studies, the use of SSA in the mentioned areas has been equal to 32%, 36%, 4%, and 28%, respectively. The highest percentage belongs to Improved, which has been analyzed by three subsections: Meat-Heuristics, artificial neural networks, and Deep Learning.

46 citations


Journal ArticleDOI
TL;DR: In this article , the authors proposed an efficient IoT service placement solution based on the autonomic methodology for deploying IoT applications on the fog infrastructure using the whale optimization algorithm (WOA) meta-heuristic technique.
Abstract: The rapid development of Internet of Things (IoT)-based applications and the era of 5G networks has led to an exponential increase in the amount of data required for processing the IoT services. The fog computing paradigm has emerged as a distributed computing solution for serving these applications using available fog nodes near the IoT devices. Since the IoT applications are developed in the form of several IoT services with various quality of service (QoS) requirements that can be deployed on the fog nodes with different resource capabilities in the fog ecosystem, finding an efficient service placement plan is one of the challenging issues to be considered. In this paper, we propose an efficient IoT service placement solution based on the autonomic methodology for deploying IoT applications on the fog infrastructure. Our proposed solution monitors the QoS requirements of IoT services and capabilities of available fog nodes to determine an efficient service placement plan using the whale optimization algorithm (WOA) meta-heuristic technique. Besides, our evolutionary-based mechanism utilized the throughput and the energy consumption as objective functions for finding desirable IoT service placement plan while meeting the QoS requirements of each IoT service. Also, we develop an autonomous service placement framework according to a three-tier architecture of the fog ecosystem to show the interaction between the main components of the IoT device and fog layers for deploying IoT applications. The simulation results demonstrate that the proposed solution increases the resource usage and service acceptance ratio and reduces the service delay and the energy consumption compared with the other metaheuristic-based mechanisms.

41 citations


Journal ArticleDOI
01 Jan 2022-Energy
TL;DR: In this paper , a real-time dynamic optimal energy management (OEM) based on deep reinforcement learning (DRL) algorithm is proposed to help the EMS make optimal schedule decisions, and the case study demonstrates the effectiveness and the computation efficiency of the proposed method.

39 citations


Journal ArticleDOI
TL;DR: Zhang et al. as discussed by the authors proposed attributed influence maximization based on the crowd emotion, aiming to apply the user emotion and group features to study the influence of multi-dimensional characteristics on information propagation.
Abstract: Most research on influence maximization focuses on the network structure features of the diffusion process but lacks the consideration of multi-dimensional characteristics. This paper proposes the attributed influence maximization based on the crowd emotion, aiming to apply the user’s emotion and group features to study the influence of multi-dimensional characteristics on information propagation. To measure the interaction effects of individual emotions, we define the user emotion power and the cluster credibility, and propose a potential influence user discovery algorithm based on the emotion aggregation mechanism to locate seed candidate sets. A two-factor information propagation model is then introduced, which considers the complexity of real networks. Experiments on real-world datasets demonstrate the effectiveness of the proposed algorithm. The results outperform the heuristic methods and are almost consistent with the greedy methods yet with improved time performance.

Journal ArticleDOI
TL;DR: A thorough literature review in three main streams of research lines: Feature selection problem, optimization algorithms, particularly, meta-heuristic algorithms, and modifications applied to NIAs to tackle the FS problem is identified.
Abstract: This survey is an effort to provide a research repository and a useful reference for researchers to guide them when planning to develop new Nature-inspired Algorithms tailored to solve Feature Selection problems (NIAs-FS). We identified and performed a thorough literature review in three main streams of research lines: Feature selection problem, optimization algorithms, particularly, meta-heuristic algorithms, and modifications applied to NIAs to tackle the FS problem. We provide a detailed overview of 156 different articles about NIAs modifications for tackling FS. We support our discussions by analytical views, visualized statistics, applied examples, open-source software systems, and discuss open issues related to FS and NIAs. Finally, the survey summarizes the main foundations of NIAs-FS with approximately 34 different operators investigated. The most popular operator is chaotic maps. Hybridization is the most widely used modification technique. There are three types of hybridization: Integrating NIA with another NIA, integrating NIA with a classifier, and integrating NIA with a classifier. The most widely used hybridization is the one that integrates a classifier with the NIA. Microarray and medical applications are the dominated applications where most of the NIA-FS are modified and used. Despite the popularity of the NIAs-FS, there are still many areas that need further investigation.

Journal ArticleDOI
TL;DR: SCOT is presented, an unsupervised algorithm that uses the Gromov-Wasserstein optimal transport to align single-cell multi-omics data sets and performs on par with the current state-of-the-art un supervised alignment methods, is faster, and requires tuning of fewer hyperparameters.
Abstract: Recent advances in sequencing technologies have allowed us to capture various aspects of the genome at single-cell resolution. However, with the exception of a few of co-assaying technologies, it is not possible to simultaneously apply different sequencing assays on the same single cell. In this scenario, computational integration of multi-omic measurements is crucial to enable joint analyses. This integration task is particularly challenging due to the lack of sample-wise or feature-wise correspondences. We present single-cell alignment with optimal transport (SCOT), an unsupervised algorithm that uses the Gromov-Wasserstein optimal transport to align single-cell multi-omics data sets. SCOT performs on par with the current state-of-the-art unsupervised alignment methods, is faster, and requires tuning of fewer hyperparameters. More importantly, SCOT uses a self-tuning heuristic to guide hyperparameter selection based on the Gromov-Wasserstein distance. Thus, in the fully unsupervised setting, SCOT aligns single-cell data sets better than the existing methods without requiring any orthogonal correspondence information.

Journal ArticleDOI
TL;DR: In this article , a mixed-objective mixed-integer linear programming model is developed for responsive, resilient and sustainable mixed open and closed-loop supply chain network design (SCND) problem and the uncertainty of the problem is handled with a hybrid robust-stochastic optimization approach.


Journal ArticleDOI
TL;DR: In this paper , a min-max cost-optimal problem to guarantee the convergence rate of federated learning in terms of cost in wireless edge networks is explored, which minimizes the cost of the worst-case participant subject to the delay, local CPU-cycle frequency, power allocation, local accuracy, and subcarrier assignment constraints.
Abstract: Federated learning is a distributed machine learning technology that can protect users’ data privacy, so it has attracted more and more attention in the industry and academia. Nonetheless, most of the existing works focused on the cost optimization of the entire process, while the cost of individual participants cannot be considered. In this article, we explore a min-max cost-optimal problem to guarantee the convergence rate of federated learning in terms of cost in wireless edge networks. In particular, we minimize the cost of the worst-case participant subject to the delay, local CPU-cycle frequency, power allocation, local accuracy, and subcarrier assignment constraints. Considering that the formulated problem is a mixed-integer nonlinear programming problem, we decompose it into several sub-problems to derive its solutions, in which the subcarrier assignment and power allocation are obtained by utilizing the Lagrangian dual decomposition method, the CPU-cycle frequency is obtained by a heuristic algorithm, and the local accuracy is obtained by an iteration algorithm. Simulation results show the convergence of the proposed algorithm and reveal that the proposed scheme can accomplish a tradeoff between the cost and fairness by comparing the proposed scheme with the existing schemes.

Journal ArticleDOI
TL;DR: In this article , two sub-models are developed using the vehicle routing problem concept, the first sub-model uses modern traceability IoT-based devices to obtain data in real-time, making it possible to identify the threshold waste level (TWL) parameter.
Abstract: In the history of sustainable development, particular attention has been paid to waste management systems (WMS), especially in smart cities. Putting real-world assumptions into practice with real-time waste bins' fill levels has not been considered in previous literature and can be investigated by utilizing the Internet of Things (IoT) concept. In this paper, two sub-models are developed using the vehicle routing problem concept. The first sub-model uses modern traceability IoT-based devices to obtain data in real-time, making it possible to identify the threshold waste level (TWL) parameter. The importance of the first sub-model is not only to determine an effective and innovative collection route for achieving sustainable being social and environmental impacts on WMS, but also to consider the priority of visiting bins based on their significance. Both waste separation and transferring them into the recovery value center are considered in the second model, both to maximize the recovery value and minimize visual pollution. The recent and capable meta-heuristic algorithms are utilized and probed to test the proposed problem's accuracy and find the best algorithm for this problem. Finally, in the sensitivity analysis section, different parameters of the problem were investigated. Besides, by defining and analyzing two indices, the best efficiency in using the transport fleet, the optimal amount of traveled distance, and the amount of collected waste can be achieved by setting the TWL between 70 and 75%.

Journal ArticleDOI
TL;DR: In this paper , the design and optimization of the sizing of hybrid renewable energy systems (HRESs) with power sharing capabilities in conjunction with electric vehicles (EVs) were proposed in two case studies.
Abstract: Currently, the ideal sizing of hybrid technologies is one of the vital aspects of power system design. In this article, the design and optimization of the sizing of hybrid renewable energy systems (HRESs) with power‐sharing capabilities in conjunction with electric vehicles (EVs) were proposed in two case studies. Two algorithms, namely, multi‐objective particle swarm optimization (MOPSO) and multi‐objective crow search (MOCS), have been formulated and were used to solve the problem being investigated. In case study 1 (CS1), four different HRESs are designed in the presence of EVs, meaning that for each HRES an EV and the power‐sharing capability is employed. And also, the stochastic behavior of the EV using Monte Carlo simulation (MCS) is modeled. In case study 2 (CS2), four HRESs are designed with power‐sharing capabilities, but in this case, for any of the HRESs, EV is not considered. This idea can be considered a novel breakthrough for the potential of power‐sharing has been incorporated with the integration of EVs and HRESs. This approach improves the life cycle cost and loss of power supply probability indices. In summary, both cases in the presence and absence of EVs were compared with the simulation results. The results show that the use of the proposed EV significantly reduces the total cost of the engineered system. Furthermore, two meta‐heuristic techniques were compared, and it was concluded that MOPSO had performed better than MOCS.

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a heuristic method to build a recommender engine in IoT environment exploiting swarm intelligence techniques, where smart objects are represented using real-valued vectors obtained through the Doc2Vec model, a word embedding technique able to capture the semantic context representing documents and sentences with dense vectors.
Abstract: In smart environments, traditional information management approaches are often unsuitable to tackle with the needed elaborations due to the amount and the high dynamicity of entities involved. Smart objects (enhanced devices or IoT services belonging to a smart system) interact and maintain relations which need of effective and efficient selection/filtering mechanisms to better meet users’ requirements. Recommender systems provide useful and customized information, properly selected and filtered, for users and services. This paper proposes a heuristic method to build a recommender engine in IoT environment exploiting swarm intelligence techniques. Smart objects are represented using real-valued vectors obtained through the Doc2Vec model, a word embedding technique able to capture the semantic context representing documents and sentences with dense vectors. The vectors are associated to mobile agents that move in a virtual 2D space following a bio-inspired model - the flocking model - in which agents perform simple and local operations autonomously obtaining a global intelligent organization. A similarity rule, based on the assigned vectors, was designed so enabling agents to discriminate among them. A closer positioning (clustering) of only similar agents is achieved. The intelligent positioning allows easy identifying of similar smart objects, thus enabling a fast and effective selection operations. Experimental evaluations have allowed to demonstrate the validity of the approach, and on how the proposed methodology allows obtaining an increasing in performance of about 50%, in terms of clustering quality and relevance, compared to other existing approaches.

Journal ArticleDOI
TL;DR: In this article , a swarm based recommendation system in Internet of Things is proposed, where real-valued vectors obtained through the Doc2Vec model are associated to mobile agents that move in a virtual 2D space following a bio-inspired model, in which agents perform simple and local operations autonomously obtaining a global intelligent organization.
Abstract: • Swarm based recommendation system in Internet of Things. • Word Embedding concepts drive bio-inspired agents’ movements and decisions. • Real-valued vectors map smart objects thus enabling agents’ organization. • Decentralization, adaptivity and self-organization features improve recommender tasks. In smart environments, traditional information management approaches are often unsuitable to tackle with the needed elaborations due to the amount and the high dynamicity of entities involved. Smart objects (enhanced devices or IoT services belonging to a smart system) interact and maintain relations which need of effective and efficient selection/filtering mechanisms to better meet users’ requirements. Recommender systems provide useful and customized information, properly selected and filtered, for users and services. This paper proposes a heuristic method to build a recommender engine in IoT environment exploiting swarm intelligence techniques. Smart objects are represented using real-valued vectors obtained through the Doc2Vec model, a word embedding technique able to capture the semantic context representing documents and sentences with dense vectors. The vectors are associated to mobile agents that move in a virtual 2D space following a bio-inspired model - the flocking model - in which agents perform simple and local operations autonomously obtaining a global intelligent organization. A similarity rule, based on the assigned vectors, was designed so enabling agents to discriminate among them. A closer positioning (clustering) of only similar agents is achieved. The intelligent positioning allows easy identifying of similar smart objects, thus enabling a fast and effective selection operations. Experimental evaluations have allowed to demonstrate the validity of the approach, and on how the proposed methodology allows obtaining an increasing in performance of about 50%, in terms of clustering quality and relevance, compared to other existing approaches.

Journal ArticleDOI
TL;DR: Wang et al. as discussed by the authors investigated a new multi-period vaccination planning problem that simultaneously optimizes the total travel distance of vaccination recipients (service level) and the operational cost, and proposed a weighted-sum and an ϵ -constraint methods, which rely on solving many singleobjective MILPs and thus lose efficiency for practical-sized instances.
Abstract: This work investigates a new multi-period vaccination planning problem that simultaneously optimizes the total travel distance of vaccination recipients (service level) and the operational cost. An optimal plan determines, for each period, which vaccination sites to open, how many vaccination stations to launch at each site, how to assign recipients from different locations to opened sites, and the replenishment quantity of each site. We formulate this new problem as a bi-objective mixed-integer linear program (MILP). We first propose a weighted-sum and an ϵ -constraint methods, which rely on solving many single-objective MILPs and thus lose efficiency for practical-sized instances. To this end, we further develop a tailored genetic algorithm where an improved assignment strategy and a new dynamic programming method are designed to obtain good feasible solutions. Results from a case study indicate that our methods reduce the operational cost and the total travel distance by up to 9.3% and 36.6%, respectively. Managerial implications suggest enlarging the service capacity of vaccination sites can help improve the performance of the vaccination program. The enhanced performance of our heuristic is due to the newly proposed assignment strategy and dynamic programming method. Our findings demonstrate that vaccination programs during pandemics can significantly benefit from formal methods, drastically improving service levels and decreasing operational costs.

Journal ArticleDOI
TL;DR: In this article , a dynamic scale-free network-based differential evolution (DSNDE) is developed by considering the demands of convergent speed and the ability to jump out of local minima.
Abstract: Some recent research reports that a dendritic neuron model (DNM) can achieve better performance than traditional artificial neuron networks (ANNs) on classification, prediction, and other problems when its parameters are well-tuned by a learning algorithm. However, the back-propagation algorithm (BP), as a mostly used learning algorithm, intrinsically suffers from defects of slow convergence and easily dropping into local minima. Therefore, more and more research adopts non-BP learning algorithms to train ANNs. In this paper, a dynamic scale-free network-based differential evolution (DSNDE) is developed by considering the demands of convergent speed and the ability to jump out of local minima. The performance of a DSNDE trained DNM is tested on 14 benchmark datasets and a photovoltaic power forecasting problem. Nine meta-heuristic algorithms are applied into comparison, including the champion of the 2017 IEEE Congress on Evolutionary Computation (CEC2017) benchmark competition effective butterfly optimizer with covariance matrix adapted retreat phase (EBOwithCMAR). The experimental results reveal that DSNDE achieves better performance than its peers.

Proceedings ArticleDOI
24 Feb 2022
TL;DR: An effective yet efficient model PAIE for both sentence-level and document-level Event Argument Extraction (EAE), which also generalizes well when there is a lack of training data is proposed.
Abstract: In this paper, we propose an effective yet efficient model PAIE for both sentence-level and document-level Event Argument Extraction (EAE), which also generalizes well when there is a lack of training data. On the one hand, PAIE utilizes prompt tuning for extractive objectives to take the best advantages of Pre-trained Language Models (PLMs). It introduces two span selectors based on the prompt to select start/end tokens among input texts for each role. On the other hand, it captures argument interactions via multi-role prompts and conducts joint optimization with optimal span assignments via a bipartite matching loss. Also, with a flexible prompt design, PAIE can extract multiple arguments with the same role instead of conventional heuristic threshold tuning. We have conducted extensive experiments on three benchmarks, including both sentence- and document-level EAE. The results present promising improvements from PAIE (3.5% and 2.3% F1 gains in average on three benchmarks, for PAIE-base and PAIE-large respectively). Further analysis demonstrates the efficiency, generalization to few-shot settings, and effectiveness of different extractive prompt tuning strategies. Our code is available at https://github.com/mayubo2333/PAIE.

Journal ArticleDOI
TL;DR: In this paper, the authors model the entire process of data generation to processing, model updating and reliability calculation, and investigate it on a deteriorating bridge system, assuming that dynamic response data are obtained in a sequential fashion from deployed accelerometers, subsequently processed by an output-only operational modal analysis scheme for identifying the system's modal characteristics.

Proceedings ArticleDOI
17 Jul 2022
TL;DR: This paper proposes AlphaRepair, the first cloze-style APR approach to directly leveraging large pre-trained code models for APR without any fine-tuning/retraining on historical bug fixes, and implementsAlphaRepair as a practical multilingual APR tool based on the recent CodeBERT model.
Abstract: Due to the promising future of Automated Program Repair (APR), researchers have proposed various APR techniques, including heuristic-based, template-based, and constraint-based techniques. Among such classic APR techniques, template-based techniques have been widely recognized as state of the art. However, such template-based techniques require predefined templates to perform repair, and their effectiveness is thus limited. To this end, researchers have leveraged the recent advances in Deep Learning to further improve APR. Such learning-based techniques typically view APR as a Neural Machine Translation problem, using the buggy/fixed code snippets as the source/target languages for translation. In this way, such techniques heavily rely on large numbers of high-quality bug-fixing commits, which can be extremely costly/challenging to construct and may limit their edit variety and context representation. In this paper, we aim to revisit the learning-based APR problem, and propose AlphaRepair, the first cloze-style (or infilling-style) APR approach to directly leveraging large pre-trained code models for APR without any fine-tuning/retraining on historical bug fixes. Our main insight is instead of modeling what a repair edit should look like (i.e., a NMT task), we can directly predict what the correct code is based on the context information (i.e., a cloze or text infilling task). Although our approach is general and can be built on various pre-trained code models, we have implemented AlphaRepair as a practical multilingual APR tool based on the recent CodeBERT model. Our evaluation of AlphaRepair on the widely used Defects4J benchmark shows for the first time that learning-based APR without any history bug fixes can already outperform state-of-the-art APR techniques. We also studied the impact of different design choices and show that AlphaRepair performs even better on a newer version of Defects4J (2.0) with 3.3X more fixes than best performing baseline, indicating that AlphaRepair can potentially avoid the dataset-overfitting issue of existing techniques. Additionally, we demonstrate the multilingual repair ability of AlphaRepair by evaluating on the QuixBugs dataset where AlphaRepair achieved the state-of-the-art results on both Java and Python versions.

Journal ArticleDOI
TL;DR: This article designs a joint computing and caching framework by integrating deep deterministic policy gradient (DDPG) algorithm for Internet of Vehicles scenario, which needs the support of mobile network provided by MNO.
Abstract: Mobile network operators (MNOs) allocate computing and caching resources for mobile users by deploying a central control system. Existing studies mainly use programming and heuristic methods to solve the resource allocation problem, which ignores the energy cost problem that is really significant to the MNO. To solve this problem, in this article, we design a joint computing and caching framework by integrating deep deterministic policy gradient (DDPG) algorithm. Especially, we focus on the Internet of Vehicles scenario, which needs the support of mobile network provided by MNO. We first formulate an optimization problem to minimize MNO’s energy cost by considering the computation and caching energy costs jointly. Then, we turn the formulated problem into a reinforcement learning problem and utilize DDPG methods to solve this problem. The final simulation result shows that our solution can reduce energy costs by more than 15%, while ensuring the tasks can be completed on time.

Journal ArticleDOI
TL;DR: Zhang et al. as discussed by the authors proposed a user-level differential privacy (UDP) algorithm by adding artificial noise to the shared models before uploading them to servers and derived a theoretical convergence upper bound for the UDP algorithm.
Abstract: Federated learning (FL), as a type of collaborative machine learning framework, is capable of preserving private data from mobile terminals (MTs) while training the data into useful models. Nevertheless, from a viewpoint of information theory, it is still possible for a curious server to infer private information from the shared models uploaded by MTs. To address this problem, we first make use of the concept of local differential privacy (LDP), and propose a user-level differential privacy (UDP) algorithm by adding artificial noise to the shared models before uploading them to servers. According to our analysis, the UDP framework can realize $(\epsilon _{i}, \delta _{i})$ -LDP for the $i$ th MT with adjustable privacy protection levels by varying the variances of the artificial noise processes. We then derive a theoretical convergence upper-bound for the UDP algorithm. It reveals that there exists an optimal number of communication rounds to achieve the best learning performance. More importantly, we propose a communication rounds discounting (CRD) method. Compared with the heuristic search method, the proposed CRD method can achieve a much better trade-off between the computational complexity of searching and the convergence performance. Extensive experiments indicate that our UDP algorithm using the proposed CRD method can effectively improve both the training efficiency and model quality for the given privacy protection levels.

Journal ArticleDOI
TL;DR: In this paper , a responsive green-cold vaccine supply chain network during the COVID-19 pandemic is developed for the first time to enhance the accuracy, speed, and justice of vaccine injection with existing priorities.
Abstract: In this paper, a new responsive-green-cold vaccine supply chain network during the COVID-19 pandemic is developed for the first time. According to the proposed network, a new multi-objective, multi-period, multi-echelon mathematical model for the distribution-allocation-location problem is designed. Another important novelty in this paper is that it considers an Internet-of-Things application in the COVID-19 condition in the suggested model to enhance the accuracy, speed, and justice of vaccine injection with existing priorities. Waste management, environmental effects, coverage demand, and delivery time of COVID-19 vaccine simultaneously are therefore considered for the first time. The LP-metric method and meta-heuristic algorithms called Gray Wolf Optimization (GWO), and Variable Neighborhood Search (VNS) algorithms are then used to solve the developed model. The other significant contribution, based on two presented meta-heuristic algorithms, is a new heuristic method called modified GWO (MGWO), and is developed for the first time to solve the model. Therefore, a set of test problems in different sizes is provided. Hence, to evaluate the proposed algorithms, assessment metrics including (1) percentage of domination, (2) the number of Pareto solutions, (3) data envelopment analysis, and (4) diversification metrics and the performance of the convergence are considered. Moreover, the Taguchi method is used to tune the algorithm's parameters. Accordingly, to illustrate the efficiency of the model developed, a real case study in Iran is suggested. Finally, the results of this research show MGO offers higher quality and better performance than other proposed algorithms based on assessment metrics, computational time, and convergence.

Book ChapterDOI
TL;DR: In this article , the authors provide an overview of six key challenges that are promising topics for research and describe some interesting opportunities, including mining patterns in complex graph data, targeted pattern mining, repetitive sequential pattern mining (RSSM), incremental, stream, and interactive pattern mining and heuristic pattern mining.
Abstract: Pattern mining is a key subfield of data mining that aims at developing algorithms to discover interesting patterns in databases. The discovered patterns can be used to help understanding the data and also to perform other tasks such as classification and prediction. After more than two decades of research in this field, great advances have been achieved in terms of theory, algorithms, and applications. However, there still remains many important challenges to be solved and also many unexplored areas. Based on this observations, this paper provides an overview of six key challenges that are promising topics for research and describe some interesting opportunities. Those challenges were identified by researchers from the field, and are: (1) mining patterns in complex graph data, (2) targeted pattern mining, (3) repetitive sequential pattern mining, (4) incremental, stream, and interactive pattern mining, (5) heuristic pattern mining, and (6) mining interesting patterns.

Journal ArticleDOI
TL;DR: In this article , the authors model the entire process of data generation to processing, model updating and reliability calculation, and investigate it on a deteriorating bridge system, assuming that dynamic response data are obtained in a sequential fashion from deployed accelerometers, subsequently processed by an output-only operational modal analysis scheme for identifying the system's modal characteristics.