scispace - formally typeset
Search or ask a question

Showing papers in "Knowledge Based Systems in 2018"


Journal ArticleDOI
TL;DR: A comprehensive and structured analysis of various graph embedding techniques proposed in the literature, and the open-source Python library, named GEM (Graph Embedding Methods, available at https://github.com/palash1992/GEM ), which provides all presented algorithms within a unified interface to foster and facilitate research on the topic.
Abstract: Graphs, such as social networks, word co-occurrence networks, and communication networks, occur naturally in various real-world applications. Analyzing them yields insight into the structure of society, language, and different patterns of communication. Many approaches have been proposed to perform the analysis. Recently, methods which use the representation of graph nodes in vector space have gained traction from the research community. In this survey, we provide a comprehensive and structured analysis of various graph embedding techniques proposed in the literature. We first introduce the embedding task and its challenges such as scalability, choice of dimensionality, and features to be preserved, and their possible solutions. We then present three categories of approaches based on factorization methods, random walks, and deep learning, with examples of representative algorithms in each category and analysis of their performance on various tasks. We evaluate these state-of-the-art methods on a few common datasets and compare their performance against one another. Our analysis concludes by suggesting some potential applications and future directions. We finally present the open-source Python library we developed, named GEM (Graph Embedding Methods, available at https://github.com/palash1992/GEM ), which provides all presented algorithms within a unified interface to foster and facilitate research on the topic.

1,553 citations


Journal ArticleDOI
TL;DR: A novel optimization algorithm, called Emperor Penguin Optimizer (EPO), which mimics the huddling behavior of emperor penguins, which is compared with eight state-of-the-art optimization algorithms.
Abstract: This paper proposes a novel optimization algorithm, called Emperor Penguin Optimizer (EPO), which mimics the huddling behavior of emperor penguins (Aptenodytes forsteri). The main steps of EPO are to generate the huddle boundary, compute temperature around the huddle, calculate the distance, and find the effective mover. These steps are mathematically modeled and implemented on 44 well-known benchmark test functions. It is compared with eight state-of-the-art optimization algorithms. The paper also considers for solving six real-life constrained and one unconstrained engineering design problems. The convergence and computational complexity are also analyzed to ensure the applicability of proposed algorithm. The experimental results show that the proposed algorithm is able to provide better results as compared to the other well-known metaheuristic algorithms.

508 citations


Journal ArticleDOI
TL;DR: Two new wrapper FS approaches that use SSA as the search strategy are proposed and it is observed that the proposed approach significantly outperforms others on around 90% of the datasets.
Abstract: Searching for the (near) optimal subset of features is a challenging problem in the process of feature selection (FS). In the literature, Swarm Intelligence (SI) algorithms show superior performance in solving this problem. This motivated our attempts to test the performance of the newly proposed Salp Swarm Algorithm (SSA) in this area. As such, two new wrapper FS approaches that use SSA as the search strategy are proposed. In the first approach, eight transfer functions are employed to convert the continuous version of SSA to binary. In the second approach, the crossover operator is used in addition to the transfer functions to replace the average operator and enhance the exploratory behavior of the algorithm. The proposed approaches are benchmarked on 22 well-known UCI datasets and the results are compared with 5 FS methods: Binary Grey Wolf Optimizer (BGWO), Binary Gravitational Search Algorithms (BGSA), Binary Bat Algorithm (BBA), Binary Particle Swarm Optimization (BPSO), and Genetic Algorithm (GA). The paper also considers an extensive study of the parameter setting for the proposed technique. From the results, it is observed that the proposed approach significantly outperforms others on around 90% of the datasets.

476 citations


Journal ArticleDOI
TL;DR: A review of CRPs in SNGDM is provided, and as a result it classifies them into two paradigms: (i) the CRP paradigm based on trust relationships, and (ii) theCRP paradigmbased on opinion evolution.
Abstract: In social network group decision making (SNGDM), the consensus reaching process (CRP) is used to help decision makers with social relationships reach consensus. Many CRP studies have been conducted in SNGDM until now. This paper provides a review of CRPs in SNGDM, and as a result it classifies them into two paradigms: (i) the CRP paradigm based on trust relationships, and (ii) the CRP paradigm based on opinion evolution. Furthermore, identified research challenges are put forward to advance this area of research.

378 citations


Journal ArticleDOI
TL;DR: A wrapper-feature selection algorithm is proposed based on the Binary Dragonfly Algorithm based on time-varying S-shaped and V-shaped transfer functions to leverage the impact of the step vector on balancing exploration and exploitation.
Abstract: The Dragonfly Algorithm (DA) is a recently proposed heuristic search algorithm that was shown to have excellent performance for numerous optimization problems. In this paper, a wrapper-feature selection algorithm is proposed based on the Binary Dragonfly Algorithm (BDA). The key component of the BDA is the transfer function that maps a continuous search space to a discrete search space. In this study, eight transfer functions, categorized into two families (S-shaped and V-shaped functions) are integrated into the BDA and evaluated using eighteen benchmark datasets obtained from the UCI data repository. The main contribution of this paper is the proposal of time-varying S-shaped and V-shaped transfer functions to leverage the impact of the step vector on balancing exploration and exploitation. During the early stages of the optimization process, the probability of changing the position of an element is high, which facilitates the exploration of new solutions starting from the initial population. On the other hand, the probability of changing the position of an element becomes lower towards the end of the optimization process. This behavior is obtained by considering the current iteration number as a parameter of transfer functions. The performance of the proposed approaches is compared with that of other state-of-art approaches including the DA, binary grey wolf optimizer (bGWO), binary gravitational search algorithm (BGSA), binary bat algorithm (BBA), particle swarm optimization (PSO), and genetic algorithm in terms of classification accuracy, sensitivity, specificity, area under the curve, and number of selected attributes. Results show that the time-varying S-shaped BDA approach outperforms compared approaches.

304 citations


Journal ArticleDOI
TL;DR: The Content-based Journals & Conferences Recommender System on computer science, as well as its web service, is presented, which recommends suitable journals or conferences with a priority order based on the abstract of a manuscript.
Abstract: As computer science and information technology are making broad and deep impacts on our daily lives, more and more papers are being submitted to computer science journals and conferences. To help authors decide where they should submit their manuscripts, we present the Content-based Journals & Conferences Recommender System on computer science, as well as its web service at http://www.keaml.cn/prs/ . This system recommends suitable journals or conferences with a priority order based on the abstract of a manuscript. To follow the fast development of computer science and technology, a web crawler is employed to continuously update the training set and the learning model. To achieve interactive online response, we propose an efficient hybrid model based on chi-square feature selection and softmax regression. Our test results show that, the system can achieve an accuracy of 61.37% and suggest the best journals or conferences in about 5 s on average.

287 citations


Journal ArticleDOI
TL;DR: The results show that the proposed criterion outperforms MIFS in both single objective and multi-objective DE frameworks, and indicates that considering feature selection as a multi- objective problem can generally provide better performance in terms of the feature subset size and the classification accuracy.
Abstract: Feature selection is an essential step in various tasks, where filter feature selection algorithms are increasingly attractive due to their simplicity and fast speed. A common filter is to use mutual information to estimate the relationships between each feature and the class labels (mutual relevancy), and between each pair of features (mutual redundancy). This strategy has gained popularity resulting a variety of criteria based on mutual information. Other well-known strategies are to order each feature based on the nearest neighbor distance as in ReliefF, and based on the between-class variance and the within-class variance as in Fisher Score. However, each strategy comes with its own advantages and disadvantages. This paper proposes a new filter criterion inspired by the concepts of mutual information, ReliefF and Fisher Score. Instead of using mutual redundancy, the proposed criterion tries to choose the highest ranked features determined by ReliefF and Fisher Score while providing the mutual relevance between features and the class labels. Based on the proposed criterion, two new differential evolution (DE) based filter approaches are developed. While the former uses the proposed criterion as a single objective problem in a weighted manner, the latter considers the proposed criterion in a multi-objective design. Moreover, a well known mutual information feature selection approach (MIFS) based on maximum-relevance and minimum-redundancy is also adopted in single-objective and multi-objective DE algorithms for feature selection. The results show that the proposed criterion outperforms MIFS in both single objective and multi-objective DE frameworks. The results also indicate that considering feature selection as a multi-objective problem can generally provide better performance in terms of the feature subset size and the classification accuracy.

256 citations


Journal ArticleDOI
TL;DR: An overview of studies on UAV path planning based on CI methods published in major journals and conference proceedings is provided and it is observed that CI methods outperform traditional methods on online and 3D problems.
Abstract: The key objective of unmanned aerial vehicle (UAV) path planning is to produce a flight path that connects a start state and a goal state while meeting the required constraints. Computational intelligence (CI) is a set of nature-inspired computational methodologies and approaches for addressing complex real-world problems for which mathematical or traditional modelling does not perform well. It has been applied in the field of UAVs since it can yield effective, accurate and rapid solutions. This article provides an overview of studies on UAV path planning based on CI methods published in major journals and conference proceedings. We survey relevant studies with respect to different CI algorithms utilized in UAV path planning, the types of time domain in UAV path planning, namely, offline and online, and the types of environment models, namely, 2D and 3D. It is observed that CI methods outperform traditional methods on online and 3D problems. The analysis is useful for identifying key results from UAV path planning research and is leveraged in this article to highlight trends and open issues.

242 citations


Journal ArticleDOI
TL;DR: A novel method called deep wavelet auto-encoder (DWAE) with extreme learning machine (ELM) is proposed for intelligent fault diagnosis of rolling bearing and the results confirm that the proposed method is superior to the traditional methods and standard deep learning methods.
Abstract: Unsupervised feature learning from the raw vibration data is a great challenge for rolling bearing intelligent fault diagnosis. In this paper, a novel method called deep wavelet auto-encoder (DWAE) with extreme learning machine (ELM) is proposed for intelligent fault diagnosis of rolling bearing. Firstly, wavelet function is employed as the nonlinear activation function to design wavelet auto-encoder (WAE), which can effectively capture the signal characteristics. Secondly, a DWAE is constructed with multiple WAEs to enhance the unsupervised feature learning ability. Finally, ELM is adopted as the classifier to accurately identify different bearing faults. The proposed method is applied to analyze the experimental bearing vibration signals, and the results confirm that the proposed method is superior to the traditional methods and standard deep learning methods.

235 citations


Journal ArticleDOI
TL;DR: This article proposed a hierarchical feature fusion strategy that fuses the modalities two in two and only then fuses all three modalities in a hierarchical fashion to improve the multimodal fusion mechanism.
Abstract: Multimodal sentiment analysis is a very actively growing field of research. A promising area of opportunity in this field is to improve the multimodal fusion mechanism. We present a novel feature fusion strategy that proceeds in a hierarchical fashion, first fusing the modalities two in two and only then fusing all three modalities. On multimodal sentiment analysis of individual utterances, our strategy outperforms conventional concatenation of features by 1%, which amounts to 5% reduction in error rate. On utterance-level multimodal sentiment analysis of multi-utterance video clips, for which current state-of-the-art techniques incorporate contextual information from other utterances of the same clip, our hierarchical fusion gives up to 2.4% (almost 10% error rate reduction) over currently used concatenation. The implementation of our method is publicly available in the form of open-source code.

205 citations


Journal ArticleDOI
TL;DR: A multi-objective version of recently developed Spotted Hyena Optimizer (SHO) is proposed called MOSHO, used to optimize the multiple objectives problems and performs better than the others and produces the Pareto optimal solutions with high convergence.
Abstract: This paper proposes a multi-objective version of recently developed Spotted Hyena Optimizer (SHO) called Multi-objective Spotted Hyena Optimizer (MOSHO). It is used to optimize the multiple objectives problems. In the proposed algorithm, a fixed-sized archive is employed for storing the non-dominated Pareto optimal solutions. The roulette wheel mechanism is used to select the effective solutions from archive to simulate the social and hunting behaviors of spotted hyenas. The proposed algorithm is tested on 24 benchmark test functions and compared with six recently developed metaheuristic algorithms. The proposed algorithm is then applied on six constrained engineering design problems to demonstrate its applicability on real-life problems. The experimental results reveal that the proposed algorithm performs better than the others and produces the Pareto optimal solutions with high convergence.

Journal ArticleDOI
TL;DR: In this article, a convolutional neural network-based hidden Markov models (CNN HMMs) are presented to classify multi-fault in mechanical systems, and the average classification accuracy ratios are 98.125% and 98% for two data series with agreeable error rate reductions.
Abstract: Vibration signals of faulty rolling element bearings usually exhibit non-linear and non-stationary characteristics caused by the complex working environment. It is difficult to develop a robust method to detect faults in bearings based on signal processing techniques. In this paper, convolutional neural network -based hidden Markov models (CNN HMMs) are presented to classify multi-faults in mechanical systems. In CNN HMMs, a CNN model is first employed to learn data features automatically from raw vibration signals. By utilizing the t-distributed stochastic neighbor embedding (t-SNE) technique, feature visualization is constructed to manifest the powerful learning ability of CNN. Then, HMMs are employed as a strong stability tool to classify faults. Both the benchmark data and experimental data are applied to the CNN HMMs. Classification results confirm the superior performance of the present combination model by comparing with CNN model alone, support vector machine (SVM) and back propagation (BP) neural network. It is shown that the average classification accuracy ratios are 98.125% and 98% for two data series with agreeable error rate reductions.

Journal ArticleDOI
TL;DR: A consensus reaching process for LSGDM with double hierarchy hesitant fuzzy linguistic preference relations is developed and the similarity degree-based clustering method, the double hierarchy information entropy-based weights-determining method and the consensus measures are proposed.
Abstract: Large-scale group decision making (LSGDM) or complex group decision making (GDM) problems are very commonly encountered in actual life, especially in the era of data. At present, double hierarchy hesitant fuzzy linguistic term set is a reasonable linguistic expression when describing some complex linguistic preference information. In this paper, we develop a consensus reaching process for LSGDM with double hierarchy hesitant fuzzy linguistic preference relations. To ensure the implementation of consensus reaching process, we also propose the similarity degree-based clustering method, the double hierarchy information entropy-based weights-determining method and the consensus measures. Finally, we apply our model to deal with a practical problem that is to evaluate Sichuan water resource management and make some comparisons with the existing approaches.

Journal ArticleDOI
TL;DR: A review of different neuro-fuzzy systems based on the classification of research articles from 2000 to 2017 is proposed to help readers have a general overview of the state-of-the-arts of neuro- fizzy systems and easily refer suitable methods according to their research interests.
Abstract: Neuro-fuzzy systems have attracted the growing interest of researchers in various scientific and engineering areas due to its effective learning and reasoning capabilities. The neuro-fuzzy systems combine the learning power of artificial neural networks and explicit knowledge representation of fuzzy inference systems. This paper proposes a review of different neuro-fuzzy systems based on the classification of research articles from 2000 to 2017. The main purpose of this survey is to help readers have a general overview of the state-of-the-arts of neuro-fuzzy systems and easily refer suitable methods according to their research interests. Different neuro-fuzzy models are compared and a table is presented summarizing the different learning structures and learning criteria with their applications.

Journal ArticleDOI
TL;DR: This article investigates the state-of-the-art multi-output Gaussian processes (MOGPs) that can transfer the knowledge across related outputs in order to improve prediction quality and gives some recommendations regarding the usage of MOGPs.
Abstract: Multi-output regression problems have extensively arisen in modern engineering community. This article investigates the state-of-the-art multi-output Gaussian processes (MOGPs) that can transfer the knowledge across related outputs in order to improve prediction quality. We classify existing MOGPs into two main categories as (1) symmetric MOGPs that improve the predictions for all the outputs, and (2) asymmetric MOGPs, particularly the multi-fidelity MOGPs, that focus on the improvement of high fidelity output via the useful information transferred from related low fidelity outputs. We review existing symmetric/asymmetric MOGPs and analyze their characteristics, e.g., the covariance functions (separable or non-separable), the modeling process (integrated or decomposed), the information transfer (bidirectional or unidirectional), and the hyperparameter inference (joint or separate). Besides, we assess the performance of ten representative MOGPs thoroughly on eight examples in symmetric/asymmetric scenarios by considering, e.g., different training data (heterotopic or isotopic), different training sizes (small, moderate and large), different output correlations (low or high), and different output sizes (up to four outputs). Based on the qualitative and quantitative analysis, we give some recommendations regarding the usage of MOGPs and highlight potential research directions.

Journal ArticleDOI
TL;DR: A new adaptive CRP model to deal with LS-GDM is presented which includes a clustering process to weight experts’ sub-groups taking into account their size and cohesion, and it uses hesitant fuzzy sets to fuse expert’s sub-group preferences to keep as much information as possible.
Abstract: Nowadays due to the social networks and the technological development, large-scale group decision making (LS-GDM) problems are fairly common and decisions that may affect to lots of people or even the society are better accepted and more appreciated if they agreed. For this reason, consensus reaching processes (CRPs) have attracted researchers attention. Although, CRPs have been usually applied to GDM problems with a few experts, they are even more important for LS-GDM, because differences among a big number of experts are higher and achieving agreed solutions is much more complex. Therefore, it is necessary to face some challenges in LS-GDM. This paper presents a new adaptive CRP model to deal with LS-GDM which includes: (i) a clustering process to weight experts’ sub-groups taking into account their size and cohesion, (ii) it uses hesitant fuzzy sets to fuse expert’s sub-group preferences to keep as much information as possible and (iii) it defines an adaptive feedback process that generates advice depending on the consensus level achieved to reduce the time and supervision costs of the CRP. Additionally, the proposed model is implemented and integrated in an intelligent CRP support system, so-called AFRYCA 2.0 to carry out this new CRP on a case study and compare it with existing models.

Journal ArticleDOI
TL;DR: A new multi-class imbalance classification algorithm, which is hereafter referred to as the Diversified Error Correcting Output Codes (DECOC) method, which finds the best classification algorithm (out of many heterogeneous classification algorithms) for each individual sub-dataset resampled from the original data.
Abstract: Class-imbalance learning is one of the most challenging problems in machine learning. As a new and important direction in this field, multi-class imbalanced data classification has attracted a great many research focus in recent years. In this paper, we first make a very comprehensive review on state-of-the-art classification algorithms for multi-class imbalanced data. Moreover, we propose a new multi-class imbalance classification algorithm, which is hereafter referred to as the Diversified Error Correcting Output Codes (DECOC) method. The main idea of DECOC is to combine the improved ECOC (Error Correcting Output Codes) method for tackling class imbalance, and the diversified ensemble learning framework, which finds the best classification algorithm (out of many heterogeneous classification algorithms) for each individual sub-dataset resampled from the original data. We conduct experiments on 19 public datasets to empirically compare the performance of DECOC with 17 state-of-the-art multi-class imbalance learning algorithms, using 4 different accuracy measures: overall accuracy, Geometric mean, F-measure, and Area Under Curve. Experimental results demonstrate that DECOC achieves significantly better accuracy performance than the other 17 algorithms on these accuracy metrics. To advance research in this field, we make all the source codes of DECOC and the above-mentioned 17 state-of-the-art algorithms for imbalanced data classification be available at GitHub: https://github.com/chongshengzhang/Multi_Imbalance .

Journal ArticleDOI
TL;DR: A framework that characterizes context-aware recommendation processes in terms of the recommendation techniques used at every stage of the process and the techniques used to incorporate context is characterized, providing a clear understanding about the integration of context into recommender systems.
Abstract: Context-aware recommender systems leverage the value of recommendations by exploiting context information that affects user preferences and situations, with the goal of recommending items that are really relevant to changing user needs. Despite the importance of context-awareness in the recommender systems realm, researchers and practitioners lack guides that help them understand the state of the art and how to exploit context information to smarten up recommender systems. This paper presents the results of a comprehensive systematic literature review we conducted to survey context-aware recommenders and their mechanisms to exploit context information. The main contribution of this paper is a framework that characterizes context-aware recommendation processes in terms of: i) the recommendation techniques used at every stage of the process, ii) the techniques used to incorporate context, and iii) the stages of the process where context is integrated into the system. This systematic literature review provides a clear understanding about the integration of context into recommender systems, including context types more frequently used in the different application domains and validation mechanismsexplained in terms of the used datasets, properties, metrics, and evaluation protocols. The paper concludes with a set of research opportunities in this field.

Journal ArticleDOI
TL;DR: This paper is, to the best of the knowledge, the first study to propose a deep learning method for detecting FOG episodes in PD patients using a novel spectral data representation strategy which considers information from both the previous and current signal windows.
Abstract: Among Parkinsons disease (PD) motor symptoms, freezing of gait (FOG) may be the most incapacitating. FOG episodes may result in falls and reduce patients quality of life. Accurate assessment of FOG would provide objective information to neurologists about the patients condition and the symptoms characteristics, while it could enable non-pharmacologic support based on rhythmic cues.This paper is, to the best of our knowledge, the first study to propose a deep learning method for detecting FOG episodes in PD patients. This model is trained using a novel spectral data representation strategy which considers information from both the previous and current signal windows. Our approach was evaluated using data collected by a waist-placed inertial measurement unit from 21 PD patients who manifested FOG episodes. These data were also employed to reproduce the state-of-the-art methodologies, which served to perform a comparative study to our FOG monitoring system.The results of this study demonstrate that our approach successfully outperforms the state-of-the-art methods for automatic FOG detection. Precisely, the deep learning model achieved 90% for the geometric mean between sensitivity and specificity, whereas the state-of-the-art methods were unable to surpass the 83% for the same metric.

Journal ArticleDOI
TL;DR: A hybrid risk evaluation model by FMEA is exploited with multi-granular linguistic distribution assessments to suit practical case and results derived from comparative and sensitivity analyses fully demonstrate the reliability and validity of the model.
Abstract: The sustainability challenge is increasingly driving the adoption of supercritical water gasification (SCWG) technology to ensure the elimination and recovery of pollution produced by sewage sludge treatment (SST). Risk evaluation by failure mode and effects analysis (FMEA) plays a crucial role in guaranteeing the reliability and safety of SCWG systems. However, some limitations in existing FMEA methods need to be ameliorated. Multiple risk factors are involved in prioritizing risk levels for failure modes in SCWG systems, it is essential a multiple criteria decision making (MCDM) process, in which overall assessments of failure modes should be provided according to their performances from several points during a system operation period. Due to differences in knowledge backgrounds and experiences, FMEA team members prefer to utilize multi-granular linguistic term sets to express their assessments of system risk. A hybrid risk evaluation model by FMEA is exploited with multi-granular linguistic distribution assessments to suit practical case. Best-worst and maximizing derivation methods are adopted to determine subjective and objective combined weights for distinguishing the importance of risk factors. Complex proportional assessment method is used to prioritize failure modes for explicitly and effectively reflecting the risk level of each failure mode. The proposed model is applied in a practical case of an SCWG system used in SST. Results derived from comparative and sensitivity analyses fully demonstrate the reliability and validity of the model.

Journal ArticleDOI
TL;DR: A two-side Cross Domain Collaborate Filtering model that can make use of both user-side and item-side shared information and infer the domain independent user and item features, which can transfer knowledge from auxiliary domains more effectively.
Abstract: Recently, Cross Domain Collaborate Filtering (CDCF) is a new way to alleviate the sparsity problem in the recommender systems. CDCF solves the sparsity problem by transferring rating knowledge from auxiliary domains. Most of previous work only uses one-side (user-side or item-side) auxiliary domain information to help the recommendation in the target domain. In this paper, we propose a two-side Cross Domain Collaborate Filtering model. We assume that there exist two auxiliary domains, i.e., user-side domain and item-side domain, where the user-side auxiliary domain shares the same aligned users with the target domain, and the item-side shares the same aligned items. Also both the two auxiliary domains contain dense rating data. In this scenario, we first employ the bi-orthogonal tri-factorization model to infer the intrinsic user and item features from the user-side and item-side auxiliary domain respectively. The inferred intrinsic features are independent on domains. Then we convert the recommendation problem into a classification problem. In detail, we use the inferred user and item features to compose the feature vector, and use the corresponding rating as the class label. Thus the user-item interactions can be represented as training samples. Finally, we employ SVMs model to solve the converted classification problem. The major advantage of our model is that it can make use of both user-side and item-side shared information. Furthermore, it can infer the domain independent user and item features. Thus it can transfer knowledge from auxiliary domains more effectively. We conduct extensive experiments to show that the proposed model performs significantly better than many state-of-the-art single domain and cross domain CF methods.

Journal ArticleDOI
TL;DR: This paper presents a hybrid multi-objective discrete artificial bee colony (HDABC) algorithm for the BLSFS scheduling problem with two conflicting criteria: the makespan and the earliness time and shows that the proposed algorithm significantly outperforms the compared ones in terms of several widely-used performance metrics.
Abstract: A blocking lot-streaming flow shop (BLSFS) scheduling problem is to schedule a number of jobs on more than one machine, where each job is split into a number of sublots while no intermediate buffers exist between adjacent machines. The BLSFS scheduling problem roots from traditional job shop scheduling problems but with additional constraints. It is more difficult to be solved than traditional job shop scheduling problems, yet very popular in real-world applications, and research on the problem has been in its infancy to date. This paper presents a hybrid multi-objective discrete artificial bee colony (HDABC) algorithm for the BLSFS scheduling problem with two conflicting criteria: the makespan and the earliness time. The main contributions of this paper include: (1) developing an initialization approach using a prior knowledge which can produce a number of promising solutions, (2) proposing two crossover operators by taking advantage of valuable information extracted from all the non-dominated solutions in the current population, and (3) presenting an efficient Pareto local search operator based on the Pareto dominance relation. The proposed algorithm is empirically compared with four state-of-the-art multi-objective evolutionary algorithms on 18 test subsets of the BLSFS scheduling problem. The experimental results show that the proposed algorithm significantly outperforms the compared ones in terms of several widely-used performance metrics.

Journal ArticleDOI
TL;DR: The aim of this paper is to personalize individual semantics in the hesitant GDM with comparative linguistic expressions to show the individual difference in understanding the meaning of words.
Abstract: In decision making problems, decision makers may prefer to use more flexible linguistic expressions instead of using only one linguistic term to express their preferences. The recent proposals of hesitant fuzzy linguistic terms sets (HFLTSs) are developed to support the elicitation of comparative linguistic expressions in hesitant decision situations. In group decision making (GDM), the statement that words mean different things for different people has been highlighted and it is natural that a word should be defined by individual semantics described by different numerical values. Considering this statement in hesitant linguistic decision making, the aim of this paper is to personalize individual semantics in the hesitant GDM with comparative linguistic expressions to show the individual difference in understanding the meaning of words. In our study, the personalized individual semantics are carried out by the fuzzy envelopes of HFLTSs based on the personalized numerical scales of linguistic term set.

Journal ArticleDOI
TL;DR: In the proposed SNA-based consensus framework, a trust propagation and aggregation mechanism to yield experts’ weights from the social trust network is presented, and the obtained weights of experts are integrated into the consensus-based MAGDM framework.
Abstract: In consensus-based multiple attribute group decision making (MAGDM) problems, it is frequent that some experts exhibit non-cooperative behaviors owing to the different areas to which they may belong and the different (sometimes conflicting) interests they might present. This may adversely affect the overall efficiency of the consensus reaching process, especially when some uncooperative behaviors by experts arise. To this end, this paper develops a novel consensus framework based on social network analysis (SNA) to deal with non-cooperative behaviors. In the proposed SNA-based consensus framework, a trust propagation and aggregation mechanism to yield experts’ weights from the social trust network is presented, and the obtained weights of experts are then integrated into the consensus-based MAGDM framework. Meanwhile, a non-cooperative behavior analysis module is designed to analyze the behaviors of experts. Based on the results of such analysis during the consensus process, each expert can express and modify the trust values pertaining other experts in the social trust network. As a result, both the social trust network and the weights of experts derived from it are dynamically updated in parallel. A simulation and comparison study is presented to demonstrate the efficiency of the SNA-based consensus framework for coping with non-cooperative behaviors.

Journal ArticleDOI
TL;DR: A hybrid incremental learning approach composed of Discrete Wavelet Transform, Empirical Mode Decomposition and Random Vector Functional Link network is presented, which can significantly improve the forecasting performance with respect to both efficiency and accuracy.
Abstract: Short-term electric load forecasting plays an important role in the management of modern power systems. Improving the accuracy and efficiency of electric load forecasting can help power utilities design reasonable operational planning which will lead to the improvement of economic and social benefits of the systems. A hybrid incremental learning approach composed of Discrete Wavelet Transform (DWT), Empirical Mode Decomposition (EMD) and Random Vector Functional Link network (RVFL) is presented in this work. RVFL network is a universal approximator with good efficiency because of the randomly generated weights between input and hidden layers and the close form solution for parameter computation. By introducing incremental learning, along with ensemble approach via DWT and EMD into RVFL network, the forecasting performance can be significantly improved with respect to both efficiency and accuracy. The electric load datasets from Australian Energy Market Operator (AEMO) were used to evaluate the effectiveness of the proposed incremental DWT-EMD based RVFL network. Moreover, the attractiveness of the proposed method can be demonstrated by the comparison with eight benchmark forecasting methods.

Journal ArticleDOI
TL;DR: This paper formalizes the problem of aspect-level sentiment analysis from a different perspective, i.e., that sentiment at aspect target level should be the main focus and proposes to explicitly model the aspect target and conduct sentiment classification directly at the aspect targets level via three granularities.
Abstract: Aspect-based sentiment analysis aims at identifying sentiment polarity towards aspect targets in a sentence. Previously, the task was modeled as a sentence-level sentiment classification problem that treated aspect targets as a hint. Such approaches oversimplify the problem by averaging word embeddings when the aspect target is a multi-word sequence. In this paper, we formalize the problem from a different perspective, i.e., that sentiment at aspect target level should be the main focus. Due to the fact that written Chinese is very rich and complex, Chinese aspect targets can be studied at three different levels of granularity: radical, character and word. Thus, we propose to explicitly model the aspect target and conduct sentiment classification directly at the aspect target level via three granularities. Moreover, we study two fusion methods for such granularities in the task of Chinese aspect-level sentiment analysis. Experimental results on a multi-word aspect target subset from SemEval2014 and four Chinese review datasets validate our claims and show the improved performance of our model over the state of the art.

Journal ArticleDOI
TL;DR: The experimental results show that, for almost all functions, the proposed chaotic dynamic weight particle swarm optimization technique has superior performance compared with other nature-inspired optimizations and well-known PSO variants.
Abstract: Particle swarm optimization (PSO), which is inspired by social behaviors of individuals in bird swarms, is a nature-inspired and global optimization algorithm. The PSO method is easy to implement and has shown good performance for many real-world optimization tasks. However, PSO has problems with premature convergence and easy trapping into local optimum solutions. In order to overcome these deficiencies, a chaotic dynamic weight particle swarm optimization (CDW-PSO) is proposed. In the CDW-PSO algorithm, a chaotic map and dynamic weight are introduced to modify the search process. The dynamic weight is defined as a function of the fitness. The search accuracy and performance of the CDW-PSO algorithm are verified on seventeen well-known classical benchmark functions. The experimental results show that, for almost all functions, the CDW-PSO technique has superior performance compared with other nature-inspired optimizations and well-known PSO variants. Namely, the proposed algorithm of CDW-PSO has better search performance.

Journal ArticleDOI
TL;DR: The simulation results show that the improved K-means algorithm based on density Canopy achieves better clustering results and is insensitive to noisy data compared to the traditional K-Means algorithm, the Canopy-based K-MEan algorithm, Semi-supervised K- means++ algorithm and K- Means-u* algorithm.
Abstract: In order to improve the accuracy and stability of K-means algorithm and solve the problem of determining the most appropriate number K of clusters and best initial seeds, an improved K-means algorithm based on density Canopy is proposed. Firstly, the density of sample data sets, the average sample distance in clusters and the distance between clusters are calculated, choosing the density maximum sampling point as the first cluster center and removing the density cluster from the data sets. Defining the product of sample density, the reciprocal of the average distance between the samples in the cluster, and the distance between the clusters as weight product, the other initial seeds is determined by the maximum weight product in the remaining data sets until the data sets is empty. The density Canopy is used as the preprocessing procedure of K-means and its result is used as the cluster number and initial clustering center of K-means algorithm. Finally, the new algorithm is tested on some well-known data sets from UCI machine learning repository and on some simulated data sets with different proportions of noise samples. The simulation results show that the improved K-means algorithm based on density Canopy achieves better clustering results and is insensitive to noisy data compared to the traditional K-means algorithm, the Canopy-based K-means algorithm, Semi-supervised K-means++ algorithm and K-means-u* algorithm. The clustering accuracy of the proposed K-means algorithm based on density Canopy is improved by 30.7%, 6.1%, 5.3% and 3.7% on average on UCI data sets, and improved by 44.3%, 3.6%, 9.6% and 8.9% on the simulated data sets with noise signal respectively. With the increase of the noise ratio, the noise immunity of the new algorithm is more obvious, when the noise ratio reached 30%, the accuracy rate is improved 50% and 6% compared to the traditional K-means algorithm and the Canopy-based K-means algorithm.

Journal ArticleDOI
TL;DR: The purpose of this paper is to develop a cloud broker architecture for cloud service selection by finding a pattern of the changing priorities of User Preferences (UPs), and it is shown that the method outperforms the Analytic Hierarchy Process (AHP).
Abstract: Due to the increasing number of cloud services, service selection has become a challenging decision for many organisations. It is even more complicated when cloud users change their preferences based on the requirements and the level of satisfaction of the experienced service. The purpose of this paper is to overcome this drawback and develop a cloud broker architecture for cloud service selection by finding a pattern of the changing priorities of User Preferences (UPs). To do that, a Markov chain is employed to find the pattern. The pattern is then connected to the Quality of Service (QoS) for the available services. A recently proposed Multi Criteria Decision Making (MCDM) method, Best Worst Method (BWM), is used to rank the services. We show that the method outperforms the Analytic Hierarchy Process (AHP). The proposed methodology provides a prioritized list of the services based on the pattern of changing UPs. The methodology is validated through a case study using real QoS performance data of Amazon Elastic Compute (Amazon EC2) cloud services.

Journal ArticleDOI
TL;DR: Different classification models for predicting student performance are created, using data collected from an Australian university, to validate the hypothesis that the models trained with instances in student sub-populations outperform those constructed using all data instances.
Abstract: The capacity to predict student academic outcomes is of value for any educational institution aiming to improve student performance and persistence. Based on the generated predictions, students identified as being at risk of academic retention or performance can be provided support in a more timely manner. This study creates different classification models for predicting student performance, using data collected from an Australian university. The data include student enrolment details as well as the activity data generated from the university learning management system (LMS). The enrolment data contain student information such as socio-demographic features, university admission basis (e.g. via entry exam or past experience) and attendance type (e.g. full-time vs. part-time). The LMS data record student engagement with their online learning activities. An important contribution of this study is the consideration of student heterogeneity in constructing the predictive models. This is based on the observation that students with different socio-demographic features or study modes may exhibit varying learning motivations. The experiments validated the hypothesis that the models trained with instances in student sub-populations outperform those constructed using all data instances. Furthermore, the experiments revealed that considering both enrolment and course activity features aids in identifying vulnerable students more precisely. The experiments determined that no individual method exhibits superior performance in all aspects. However, the rule-based and tree-based methods generate models with higher interpretability, making them more useful for designing effective student support.