scispace - formally typeset
Search or ask a question

Showing papers in "Computational Intelligence and Neuroscience in 2020"


Journal ArticleDOI
TL;DR: This paper will update on the current progress of emotion recognition using EEG signals from 2016 to 2019, focusing on the elements of emotion stimuli type and presentation approach, study size, EEG hardware, machine learning classifiers, and classification approach.
Abstract: Emotions are fundamental for human beings and play an important role in human cognition. Emotion is commonly associated with logical decision making, perception, human interaction, and to a certain extent, human intelligence itself. With the growing interest of the research community towards establishing some meaningful "emotional" interactions between humans and computers, the need for reliable and deployable solutions for the identification of human emotional states is required. Recent developments in using electroencephalography (EEG) for emotion recognition have garnered strong interest from the research community as the latest developments in consumer-grade wearable EEG solutions can provide a cheap, portable, and simple solution for identifying emotions. Since the last comprehensive review was conducted back from the years 2009 to 2016, this paper will update on the current progress of emotion recognition using EEG signals from 2016 to 2019. The focus on this state-of-the-art review focuses on the elements of emotion stimuli type and presentation approach, study size, EEG hardware, machine learning classifiers, and classification approach. From this state-of-the-art review, we suggest several future research opportunities including proposing a different approach in presenting the stimuli in the form of virtual reality (VR). To this end, an additional section devoted specifically to reviewing only VR studies within this research domain is presented as the motivation for this proposed new approach using VR as the stimuli presentation device. This review paper is intended to be useful for the research community working on emotion recognition using EEG signals as well as for those who are venturing into this field of research.

129 citations


Journal ArticleDOI
TL;DR: A modified Harris hawks optimization (HHO) algorithm based on the simulated annealing (SA) for scheduling jobs in the cloud environment, which demonstrates that HHOSA can achieve significant reductions in makespan of the job scheduling problem as compared to the standard HHO and other existing scheduling algorithms.
Abstract: In recent years, cloud computing technology has attracted extensive attention from both academia and industry. The popularity of cloud computing was originated from its ability to deliver global IT services such as core infrastructure, platforms, and applications to cloud customers over the web. Furthermore, it promises on-demand services with new forms of the pricing package. However, cloud job scheduling is still NP-complete and became more complicated due to some factors such as resource dynamicity and on-demand consumer application requirements. To fill this gap, this paper presents a modified Harris hawks optimization (HHO) algorithm based on the simulated annealing (SA) for scheduling jobs in the cloud environment. In the proposed HHOSA approach, SA is employed as a local search algorithm to improve the rate of convergence and quality of solution generated by the standard HHO algorithm. The performance of the HHOSA method is compared with that of state-of-the-art job scheduling algorithms, by having them all implemented on the CloudSim toolkit. Both standard and synthetic workloads are employed to analyze the performance of the proposed HHOSA algorithm. The obtained results demonstrate that HHOSA can achieve significant reductions in makespan of the job scheduling problem as compared to the standard HHO and other existing scheduling algorithms. Moreover, it converges faster when the search space becomes larger which makes it appropriate for large-scale scheduling problems.

76 citations


Journal ArticleDOI
TL;DR: This paper presents an in-depth survey and performance evaluation of cat swarm optimization (CSO) algorithm, a robust and powerful metaheuristic swarm-based optimization approach that has received very positive feedback since its emergence.
Abstract: This paper presents an in-depth survey and performance evaluation of cat swarm optimization (CSO) algorithm. CSO is a robust and powerful metaheuristic swarm-based optimization approach that has received very positive feedback since its emergence. It has been tackling many optimization problems, and many variants of it have been introduced. However, the literature lacks a detailed survey or a performance evaluation in this regard. Therefore, this paper is an attempt to review all these works, including its developments and applications, and group them accordingly. In addition, CSO is tested on 23 classical benchmark functions and 10 modern benchmark functions (CEC 2019). The results are then compared against three novel and powerful optimization algorithms, namely, dragonfly algorithm (DA), butterfly optimization algorithm (BOA), and fitness dependent optimizer (FDO). These algorithms are then ranked according to Friedman test, and the results show that CSO ranks first on the whole. Finally, statistical approaches are employed to further confirm the outperformance of CSO algorithm.

73 citations


Journal ArticleDOI
TL;DR: Experimental results demonstrate that the proposed EM-CNN can efficiently detect driver fatigue status using driving images, and outperforms other CNN-based methods, i.e., AlexNet, VGG-16, GoogLeNet, and ResNet50.
Abstract: With a focus on fatigue driving detection research, a fully automated driver fatigue status detection algorithm using driving images is proposed. In the proposed algorithm, the multitask cascaded convolutional network (MTCNN) architecture is employed in face detection and feature point location, and the region of interest (ROI) is extracted using feature points. A convolutional neural network, named EM-CNN, is proposed to detect the states of the eyes and mouth from the ROI images. The percentage of eyelid closure over the pupil over time (PERCLOS) and mouth opening degree (POM) are two parameters used for fatigue detection. Experimental results demonstrate that the proposed EM-CNN can efficiently detect driver fatigue status using driving images. The proposed algorithm EM-CNN outperforms other CNN-based methods, i.e., AlexNet, VGG-16, GoogLeNet, and ResNet50, showing accuracy and sensitivity rates of 93.623% and 93.643%, respectively.

49 citations


Journal ArticleDOI
TL;DR: This paper reviews the key challenges of deep learning for retail product recognition and discusses potential techniques that can be helpful for the research of the topic, and provides the details of public datasets which could be used for deep learning.
Abstract: Taking time to identify expected products and waiting for the checkout in a retail store are common scenes we all encounter in our daily lives. The realization of automatic product recognition has great significance for both economic and social progress because it is more reliable than manual operation and time-saving. Product recognition via images is a challenging task in the field of computer vision. It receives increasing consideration due to the great application prospect, such as automatic checkout, stock tracking, planogram compliance, and visually impaired assistance. In recent years, deep learning enjoys a flourishing evolution with tremendous achievements in image classification and object detection. This article aims to present a comprehensive literature review of recent research on deep learning-based retail product recognition. More specifically, this paper reviews the key challenges of deep learning for retail product recognition and discusses potential techniques that can be helpful for the research of the topic. Next, we provide the details of public datasets which could be used for deep learning. Finally, we conclude the current progress and point new perspectives to the research of related fields.

47 citations


Journal ArticleDOI
TL;DR: This review presents an overview of different stereo matching algorithms based on deep learning and provided a comprehensive coverage of the remarkable approaches in each category and summarized the strengths, weaknesses, and major challenges, respectively.
Abstract: Stereo vision is a flourishing field, attracting the attention of many researchers. Recently, leveraging on the development of deep learning, stereo matching algorithms have achieved remarkable performance far exceeding traditional approaches. This review presents an overview of different stereo matching algorithms based on deep learning. For convenience, we classified the algorithms into three categories: (1) non-end-to-end learning algorithms, (2) end-to-end learning algorithms, and (3) unsupervised learning algorithms. We have provided a comprehensive coverage of the remarkable approaches in each category and summarized the strengths, weaknesses, and major challenges, respectively. The speed, accuracy, and time consumption were adopted to compare the different algorithms.

45 citations


Journal ArticleDOI
TL;DR: The advantages and the shortcomings of these methods are discussed, providing the commonly used datasets and evaluation criteria in this field, and some open challenges in the image caption task are highlighted.
Abstract: In recent years, with the rapid development of artificial intelligence, image caption has gradually attracted the attention of many researchers in the field of artificial intelligence and has become an interesting and arduous task. Image caption, automatically generating natural language descriptions according to the content observed in an image, is an important part of scene understanding, which combines the knowledge of computer vision and natural language processing. The application of image caption is extensive and significant, for example, the realization of human-computer interaction. This paper summarizes the related methods and focuses on the attention mechanism, which plays an important role in computer vision and is recently widely used in image caption generation tasks. Furthermore, the advantages and the shortcomings of these methods are discussed, providing the commonly used datasets and evaluation criteria in this field. Finally, this paper highlights some open challenges in the image caption task.

44 citations


Journal ArticleDOI
TL;DR: This paper investigates image transformation operations and generative adversarial networks (GAN) for data augmentation and state-of-the-art deep neural networks (i.e., VGG-16, ResNet, and DenseNet) for the classification of white blood cells into the five types and finds the proposed approach outperforms other methods that rely on sophisticated image processing and manual feature engineering.
Abstract: White blood cells (leukocytes) are a very important component of the blood that forms the immune system, which is responsible for fighting foreign elements. The five types of white blood cells include neutrophils, eosinophils, lymphocytes, monocytes, and basophils, where each type constitutes a different proportion and performs specific functions. Being able to classify and, therefore, count these different constituents is critical for assessing the health of patients and infection risks. Generally, laboratory experiments are used for determining the type of a white blood cell. The staining process and manual evaluation of acquired images under the microscope are tedious and subject to human errors. Moreover, a major challenge is the unavailability of training data that cover the morphological variations of white blood cells so that trained classifiers can generalize well. As such, this paper investigates image transformation operations and generative adversarial networks (GAN) for data augmentation and state-of-the-art deep neural networks (i.e., VGG-16, ResNet, and DenseNet) for the classification of white blood cells into the five types. Furthermore, we explore initializing the DNNs' weights randomly or using weights pretrained on the CIFAR-100 dataset. In contrast to other works that require advanced image preprocessing and manual feature extraction before classification, our method works directly with the acquired images. The results of extensive experiments show that the proposed method can successfully classify white blood cells. The best DNN model, DenseNet-169, yields a validation accuracy of 98.8%. Particularly, we find that the proposed approach outperforms other methods that rely on sophisticated image processing and manual feature engineering.

43 citations


Journal ArticleDOI
TL;DR: 3 improved MobileNet models with local receptive field expansion in shallow layers, also called Dilated-MobileNet (Dilated Convolution MobileNet) models, are proposed, in which dilated convolutions are introduced into a specific convolutional layer of the MobileNet model to obtain better classification accuracy.
Abstract: Because deep neural networks (DNNs) are both memory-intensive and computation-intensive, they are difficult to apply to embedded systems with limited hardware resources. Therefore, DNN models need to be compressed and accelerated. By applying depthwise separable convolutions, MobileNet can decrease the number of parameters and computational complexity with less loss of classification precision. Based on MobileNet, 3 improved MobileNet models with local receptive field expansion in shallow layers, also called Dilated-MobileNet (Dilated Convolution MobileNet) models, are proposed, in which dilated convolutions are introduced into a specific convolutional layer of the MobileNet model. Without increasing the number of parameters, dilated convolutions are used to increase the receptive field of the convolution filters to obtain better classification accuracy. The experiments were performed on the Caltech-101, Caltech-256, and Tubingen animals with attribute datasets, respectively. The results show that Dilated-MobileNets can obtain up to 2% higher classification accuracy than MobileNet.

36 citations


Journal ArticleDOI
TL;DR: The effects of single regression and classification combined with regression are compared in terms of accuracy, and four kinds of classical networks are improved and trained using 300W-LP and AFLW datasets to propose a driver distraction method based on the head pose.
Abstract: In view of the fact that the detection of driver’s distraction is a burning issue, this study chooses the driver’s head pose as the evaluation parameter for driving distraction and proposes a driver distraction method based on the head pose. The effects of single regression and classification combined with regression are compared in terms of accuracy, and four kinds of classical networks are improved and trained using 300W-LP and AFLW datasets. The HPE_Resnet50 with the best accuracy is selected as the head pose estimator and applied to the ten-category distracted driving dataset SF3D to obtain 20,000 sets of head pose data. The differences between classes are discussed qualitatively and quantitatively. The analysis of variance shows that there is a statistically significant difference in head posture between safe driving and all kinds of distracted driving at 95% and 90% confidence levels, and the postures of all kinds of driving movements are distributed in a specific Euler angle range, which provides a characteristic basis for the design of subsequent recognition methods. In addition, according to the continuity of human movement, this paper also selects 90 drivers’ videos to analyze the difference in head pose between safe driving and distracted driving frame by frame. By calculating the spatial distance and sample statistics, the results provide the reference point, spatial range, and threshold of safe driving under this driving condition. Experimental results show that the average error of HPE_Resnet50 in AFLW2000 is 6.17° and that there is an average difference of 12.4° to 54.9° in the Euler angle between safe driving and nine kinds of distracted driving on SF3D.

35 citations


Journal ArticleDOI
TL;DR: The proposed algorithm is efficient and robust in target detection and shadow detection and removal under different scenes and embeds into C/S framework based on the HTML5 web socket protocol.
Abstract: Shadow detection and removal in real scene images are a significant problem for target detection. This work proposes an improved shadow detection and removal algorithm for urban video surveillance. First, the foreground is detected by background subtraction and the shadow is detected by HSV color space. Using local variance and OTSU method, we obtain the moving targets with texture features. According to the characteristics of shadow in HSV space and texture feature, the shadow is detected and removed to eliminate the shadow interference for the subsequent processing of moving targets. Finally, we embed our algorithm into C/S framework based on the HTML5 web socket protocol. Both the experimental and actual operation results show that the proposed algorithm is efficient and robust in target detection and shadow detection and removal under different scenes.

Journal ArticleDOI
TL;DR: A dendritic neuron model with adaptive synapses (DMASs) based on differential evolution (DE) algorithm training is proposed and results show that the DE-DMAS shows better performance in all aspects.
Abstract: A dendritic neuron model with adaptive synapses (DMASs) based on differential evolution (DE) algorithm training is proposed. According to the signal transmission order, a DNM can be divided into four parts: the synaptic layer, dendritic layer, membrane layer, and somatic cell layer. It can be converted to a logic circuit that is easily implemented on hardware by removing useless synapses and dendrites after training. This logic circuit can be designed to solve complex nonlinear problems using only four basic logical devices: comparators, AND (conjunction), OR (disjunction), and NOT (negation). To obtain a faster and better solution, we adopt the most popular DE for DMAS training. We have chosen five classification datasets from the UCI Machine Learning Repository for an experiment. We analyze and discuss the experimental results in terms of the correct rate, convergence rate, ROC curve, and the cross-validation and then compare the results with a dendritic neuron model trained by the backpropagation algorithm (BP-DNM) and a neural network trained by the backpropagation algorithm (BPNN). The analysis results show that the DE-DMAS shows better performance in all aspects.

Journal ArticleDOI
TL;DR: The experimental results demonstrate the effectiveness of the proposed deep residual network method in lung nodule classification for CT images, which improved the accuracy and the false positive rate compared with the conventional support vector machine (SVM)-based CAD system.
Abstract: The classification process of lung nodule detection in a traditional computer-aided detection (CAD) system is complex, and the classification result is heavily dependent on the performance of each step in lung nodule detection, causing low classification accuracy and high false positive rate. In order to alleviate these issues, a lung nodule classification method based on a deep residual network is proposed. Abandoning traditional image processing methods and taking the 50-layer ResNet network structure as the initial model, the deep residual network is constructed by combining residual learning and migration learning. The proposed approach is verified by conducting experiments on the lung computed tomography (CT) images from the publicly available LIDC-IDRI database. An average accuracy of 98.23% and a false positive rate of 1.65% are obtained based on the ten-fold cross-validation method. Compared with the conventional support vector machine (SVM)-based CAD system, the accuracy of our method improved by 9.96% and the false positive rate decreased by 6.95%, while the accuracy improved by 1.75% and 2.42%, respectively, and the false positive rate decreased by 2.07% and 2.22%, respectively, in contrast to the VGG19 model and InceptionV3 convolutional neural networks. The experimental results demonstrate the effectiveness of our proposed method in lung nodule classification for CT images.

Journal ArticleDOI
TL;DR: The proposed work aims to improve upon the existing approaches by incorporating a new perspective which classifies the sarcasm based on the level of harshness employed, and relating the emotional state of a person to the type of sarcasm exhibited by him/her which could provide major insights about the emotional behavior of a people.
Abstract: Sentimental analysis aims at inferring how people express their opinion over any piece of text or topic of interest. This article deals with detection of an implicit form of the sentiment, referred to as sarcasm. Sarcasm conveys the opposite of what people try to convey in order to criticize or ridicule in a humorous way. It plays a vital role in social networks since most of the tweets or posts contain sarcastic nuances. Existing approaches towards the study of sarcasm deals only with the detection of sarcasm. In this paper, in addition to detecting sarcasm from text, an approach has been proposed to identify the type of sarcasm. The main motivation behind determining the types of sarcasm is to identify the level of hurt or the true intent behind the sarcastic text. The proposed work aims to improve upon the existing approaches by incorporating a new perspective which classifies the sarcasm based on the level of harshness employed. The major application of the proposed work would be relating the emotional state of a person to the type of sarcasm exhibited by him/her which could provide major insights about the emotional behavior of a person. An ensemble-based feature selection method has been proposed for identifying the optimal set of features needed to detect sarcasm from tweets. This optimal set of features was employed to detect whether the tweet is sarcastic or not. After detecting sarcastic sentences, a multi-rule based approach has been proposed to determine the type of sarcasm. As an initial attempt, sarcasm has been classified into four types, namely, polite sarcasm, rude sarcasm, raging sarcasm, and deadpan sarcasm. The performance and efficiency of the proposed approach has been experimentally analyzed, and change in mood of a person for each sarcastic type has been modelled. The overall accuracy of the proposed ensemble feature selection algorithm for sarcasm detection is around 92.7%, and the proposed multi-rule approach for sarcastic type identification achieves an accuracy of 95.98%, 96.20%, 99.79%, and 86.61% for polite, rude, raging, and deadpan types of sarcasm, respectively.

Journal ArticleDOI
TL;DR: In this study, a new method of smooth path planning is proposed based on Bezier curves and is applied to solve the problem of redundant nodes and peak inflection points in the path planning process of traditional algorithms.
Abstract: In this study, a new method of smooth path planning is proposed based on Bezier curves and is applied to solve the problem of redundant nodes and peak inflection points in the path planning process of traditional algorithms. First, genetic operations are used to obtain the control points of the Bezier curve. Second, a shorter path is selected by an optimization criterion that the length of the Bezier curve is determined by the control points. Finally, a safe distance and adaptive penalty factor are introduced into the fitness function to ensure the safety of the walking process of the robot. Numerous experiments are implemented in two different environments and compared with the existing methods. It is proved that the proposed method is more effective to generate a shorter, smoother, and safer path compared with traditional approaches.

Journal ArticleDOI
TL;DR: The results of this study suggest that the proposed data-driven prognostic method offers a new and promising prediction approach and an efficient feature extraction scheme.
Abstract: Accurate predictions of remaining useful life (RUL) of important components play a crucial role in system reliability, which is the basis of prognostics and health management (PHM). This paper proposed an integrated deep learning approach for RUL prediction of a turbofan engine by integrating an autoencoder (AE) with a deep convolutional generative adversarial network (DCGAN). In the pretraining stage, the reconstructed data of the AE not only participate in its error reconstruction but also take part in the DCGAN parameter training as the generated data of the DCGAN. Through double-error reconstructions, the capability of feature extraction is enhanced, and high-level abstract information is obtained. In the fine-tuning stage, a long short-term memory (LSTM) network is used to extract the sequential information from the features to predict the RUL. The effectiveness of the proposed scheme is verified on the NASA commercial modular aero-propulsion system simulation (C-MAPSS) dataset. The superiority of the proposed method is demonstrated via excellent prediction performance and comparisons with other existing state-of-the-art prognostics. The results of this study suggest that the proposed data-driven prognostic method offers a new and promising prediction approach and an efficient feature extraction scheme.

Journal ArticleDOI
TL;DR: The fusion of 5 algorithms to enhance the accuracy for detection of mental stress using EEG signals using Whale Optimization Algorithm has been modified to select the optimal kernel in the SVM classifier for stress detection.
Abstract: Stress is categorized as a condition of mental strain or pressure approaches because of upsetting or requesting conditions. There are various sources of stress initiation. Researchers consider human cerebrum as the primary wellspring of stress. To study how each individual encounters stress in different forms, researchers conduct surveys and monitor it. The paper presents the fusion of 5 algorithms to enhance the accuracy for detection of mental stress using EEG signals. The Whale Optimization Algorithm has been modified to select the optimal kernel in the SVM classifier for stress detection. An integrated set of algorithms (NLM, DCT, and MBPSO) has been used for preprocessing, feature extraction, and selection. The proposed algorithm has been tested on EEG signals collected from 14 subjects to identify the stress level. The proposed approach was validated using accuracy, sensitivity, specificity, and F1 score with values of 96.36%, 96.84%, 90.8%, and 97.96% and was found to be better than the existing ones. The algorithm may be useful to psychiatrists and health consultants for diagnosing the stress level.

Journal ArticleDOI
TL;DR: The theoretical basis of GANs is reviewed and some recently developed GAN models are surveyed, in comparison with traditional G AN models, for computer vision applications.
Abstract: Computer vision is one of the hottest research fields in deep learning. The emergence of generative adversarial networks (GANs) provides a new method and model for computer vision. The idea of GANs using the game training method is superior to traditional machine learning algorithms in terms of feature learning and image generation. GANs are widely used not only in image generation and style transfer but also in the text, voice, video processing, and other fields. However, there are still some problems with GANs, such as model collapse and uncontrollable training. This paper deeply reviews the theoretical basis of GANs and surveys some recently developed GAN models, in comparison with traditional GAN models. The applications of GANs in computer vision include data enhancement, domain transfer, high-quality sample generation, and image restoration. The latest research progress of GANs in artificial intelligence (AI) based security attack and defense is introduced. The future development of GANs in computer vision is also discussed at the end of the paper with possible applications of AI in computer vision.

Journal ArticleDOI
TL;DR: This study demonstrates that DCNNs provide an effective approach to diagnose nutrient deficiencies in rice with accuracy of over 90%, with DenseNet121 performing best.
Abstract: Symptoms of nutrient deficiencies in rice plants often appear on the leaves. The leaf color and shape, therefore, can be used to diagnose nutrient deficiencies in rice. Image classification is an efficient and fast approach for this diagnosis task. Deep convolutional neural networks (DCNNs) have been proven to be effective in image classification, but their use to identify nutrient deficiencies in rice has received little attention. In the present study, we explore the accuracy of different DCNNs for diagnosis of nutrient deficiencies in rice. A total of 1818 photographs of plant leaves were obtained via hydroponic experiments to cover full nutrition and 10 classes of nutrient deficiencies. The photographs were divided into training, validation, and test sets in a 3 : 1 : 1 ratio. Fine-tuning was performed to evaluate four state-of-the-art DCNNs: Inception-v3, ResNet with 50 layers, NasNet-Large, and DenseNet with 121 layers. All the DCNNs obtained validation and test accuracies of over 90%, with DenseNet121 performing best (validation accuracy = 98.62 ± 0.57%; test accuracy = 97.44 ± 0.57%). The performance of the DCNNs was validated by comparison to color feature with support vector machine and histogram of oriented gradient with support vector machine. This study demonstrates that DCNNs provide an effective approach to diagnose nutrient deficiencies in rice.

Journal ArticleDOI
TL;DR: A case study of a wind harvesting farm is investigated in terms of wind speed collected data and results show that even though SVM delivered the most accurate predictions, ConvLSTM was chosen due to its less computational efforts as well as high prediction accuracy.
Abstract: The need for an efficient power source for operating the modern industry has been rapidly increasing in the past years. Therefore, the latest renewable power sources are difficult to be predicted. The generated power is highly dependent on fluctuated factors (such as wind bearing, pressure, wind speed, and humidity of surrounding atmosphere). Thus, accurate forecasting methods are of paramount importance to be developed and employed in practice. In this paper, a case study of a wind harvesting farm is investigated in terms of wind speed collected data. For data like the wind speed that are hard to be predicted, a well built and tested forecasting algorithm must be provided. To accomplish this goal, four neural network-based algorithms: artificial neural network (ANN), convolutional neural network (CNN), long short-term memory (LSTM), and a hybrid model convolutional LSTM (ConvLSTM) that combines LSTM with CNN, and one support vector machine (SVM) model are investigated, evaluated, and compared using different statistical and time indicators to assure that the final model meets the goal that is built for. Results show that even though SVM delivered the most accurate predictions, ConvLSTM was chosen due to its less computational efforts as well as high prediction accuracy.

Journal ArticleDOI
TL;DR: The results show that the proposed Z-shaped probability transfer function improves the convergence speed and optimization accuracy of the BPSO algorithm.
Abstract: Particle swarm optimization (PSO) algorithm is a swarm intelligent searching algorithm based on population that simulates the social behavior of birds, bees, or fish groups. The discrete binary particle swarm optimization (BPSO) algorithm maps the continuous search space to a binary space through a new transfer function, and the update process is designed to switch the position of the particles between 0 and 1 in the binary search space. Aiming at the existed BPSO algorithms which are easy to fall into the local optimum, a new Z-shaped probability transfer function is proposed to map the continuous search space to a binary space. By adopting nine typical benchmark functions, the proposed Z-probability transfer function and the V-shaped and S-shaped transfer functions are used to carry out the performance simulation experiments. The results show that the proposed Z-shaped probability transfer function improves the convergence speed and optimization accuracy of the BPSO algorithm.

Journal ArticleDOI
TL;DR: An aesthetic degree evaluation method of interface design based on Kansei engineering was put forward and showed that the proposed method was feasible and effective compared with the method of paired comparison commonly used in psychophysics.
Abstract: The current research on human-machine interaction interface layout focused on ergonomic analysis, while the research on aesthetics and aesthetic degree calculation of interface layout was insufficient. In order to objectively evaluate the aesthetic degree of interface layout, this paper put forward an aesthetic degree evaluation method of interface design based on Kansei engineering. Firstly, the perceptual image structure of interface aesthetic degree was analyzed from the perspective of aesthetic cognition. Six aesthetic image factors affecting interface aesthetic degree, including proportion, conciseness, order, rhythm, density, and equilibrium, were extracted by factor analysis method, and the variance contribution rate of each factor was taken as the weight. Secondly, according to the six aesthetic degree indexes, the calculation system of interface aesthetic degree was constructed, and the aesthetic degree value of aesthetic image factor was calculated by the corresponding aesthetic degree evaluation mathematical formula. Then, Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) method was used to analyze the order of aesthetic degree superiority of design schemes, and the comprehensive aesthetic degree evaluation was carried out. Finally, the aesthetic degree evaluation of human-machine interaction interface layout of the driller's console of an AC variable frequency drilling rig was taken as an example to verify that this method was helpful for designers to optimize the design scheme. The experimental results showed that the proposed method was feasible and effective compared with the method of paired comparison commonly used in psychophysics.

Journal ArticleDOI
TL;DR: Eye-tracking signals recorded from 30 participants were used to calculate 18 features associated with eye movements (fixations and saccades) and pupil diameter that were related to emotions and investigated the influence of luminance and the dynamics of the presented movies.
Abstract: This article reports the results of the study related to emotion recognition by using eye-tracking. Emotions were evoked by presenting a dynamic movie material in the form of 21 video fragments. Eye-tracking signals recorded from 30 participants were used to calculate 18 features associated with eye movements (fixations and saccades) and pupil diameter. To ensure that the features were related to emotions, we investigated the influence of luminance and the dynamics of the presented movies. Three classes of emotions were considered: high arousal and low valence, low arousal and moderate valence, and high arousal and high valence. A maximum of 80% classification accuracy was obtained using the support vector machine (SVM) classifier and leave-one-subject-out validation method.

Journal ArticleDOI
TL;DR: The dataset which has chest X-ray images for classification made available by Kaggle is used, with a total of 5216 train and 624 test images, with 2 classes as normal and pneumonia, and studies using five mainstream network algorithms to classify these diseases are performed and results are compared.
Abstract: As a respiratory infection, pneumonia has gained great attention from countries all over the world for its strong spreading and relatively high mortality. For pneumonia, early detection and treatment will reduce its mortality rate significantly. Currently, X-ray diagnosis is recognized as a relatively effective method. The visual analysis of a patient's X-ray chest radiograph by an experienced doctor takes about 5 to 15 minutes. When cases are concentrated, this will undoubtedly put tremendous pressure on the doctor's clinical diagnosis. Therefore, relying on the naked eye of the imaging doctor has very low efficiency. Hence, the use of artificial intelligence for clinical image diagnosis of pneumonia is a necessary thing. In addition, artificial intelligence recognition is very fast, and the convolutional neural networks (CNNs) have achieved better performance than human beings in terms of image identification. Therefore, we used the dataset which has chest X-ray images for classification made available by Kaggle with a total of 5216 train and 624 test images, with 2 classes as normal and pneumonia. We performed studies using five mainstream network algorithms to classify these diseases in the dataset and compared the results, from which we improved MobileNet's network structure and achieved a higher accuracy rate than other methods. Furthermore, the improved MobileNet's network could also extend to other areas for application.

Journal ArticleDOI
TL;DR: Two deep CNN models were proposed with an ensemble technique to detect all the stages of DR by using balanced and imbalanced datasets and perform better compared to state-of-the-art methods on the same Kaggle dataset.
Abstract: Diabetic retinopathy (DR) is an eye disease that damages the blood vessels of the eye. DR causes blurred vision or it may lead to blindness if it is not detected in early stages. DR has five stages, i.e., 0 normal, 1 mild, 2 moderate, 3 severe, and 4 PDR. Conventionally, many hand-on projects of computer vision have been applied to detect DR but cannot code the intricate underlying features. Therefore, they result in poor classification of DR stages, particularly for early stages. In this research, two deep CNN models were proposed with an ensemble technique to detect all the stages of DR by using balanced and imbalanced datasets. The models were trained with Kaggle dataset on a high-end Graphical Processing data. Balanced dataset was used to train both models, and we test these models with balanced and imbalanced datasets. The result shows that the proposed models detect all the stages of DR unlike the current methods and perform better compared to state-of-the-art methods on the same Kaggle dataset.

Journal ArticleDOI
TL;DR: The experimental results demonstrate that the proposed framework distinguishes SZ patients from NCs, with the highest classification accuracy reaching over 85%.
Abstract: Utilizing neuroimaging and machine learning (ML) to differentiate schizophrenia (SZ) patients from normal controls (NCs) and for detecting abnormal brain regions in schizophrenia has several benefits and can provide a reference for the clinical diagnosis of schizophrenia. In this study, structural magnetic resonance images (sMRIs) from SZ patients and NCs were used for discriminative analysis. This study proposed an ML framework based on coarse-to-fine feature selection. The proposed framework used two-sample t-tests to extract the differences between groups first, then further eliminated the nonrelevant and redundant features with recursive feature elimination (RFE), and finally utilized the support vector machine (SVM) to learn the decision models with selected gray matter (GM) and white matter (WM) features. Previous studies have tended to report differences at the group level instead of at the individual level and cannot be widely applied. The method proposed in this study extends the diagnosis to the individual level and has a higher recognition rate than previous methods. The experimental results of this study demonstrate that the proposed framework distinguishes SZ patients from NCs, with the highest classification accuracy reaching over 85%. The identified biomarkers are also consistent with previous literature findings. As a universal method, the proposed framework can be extended to diagnose other diseases.

Journal ArticleDOI
TL;DR: A hybrid method for fuzzy clustering that combines FCM with an improved version of PSO, called ELPSO, which adopts a new enhanced logarithmic inertia weight strategy to provide better balance between exploration and exploitation.
Abstract: Fuzzy c-means (FCM) is one of the best-known clustering methods to organize the wide variety of datasets automatically and acquire accurate classification, but it has a tendency to fall into local minima. For overcoming these weaknesses, some methods that hybridize PSO and FCM for clustering have been proposed in the literature, and it is demonstrated that these hybrid methods have an improved accuracy over traditional partition clustering approaches, whereas PSO-based clustering methods have poor execution time in comparison to partitional clustering techniques, and the current PSO algorithms require tuning a range of parameters before they are able to find good solutions. Therefore, this paper introduces a hybrid method for fuzzy clustering, named FCM-ELPSO, which aim to deal with these shortcomings. It combines FCM with an improved version of PSO, called ELPSO, which adopts a new enhanced logarithmic inertia weight strategy to provide better balance between exploration and exploitation. This new hybrid method uses PBM(F) index and the objective function value as cluster validity indexes to evaluate the clustering effect. To verify the effectiveness of the algorithm, two types of experiments are performed, including PSO clustering and hybrid clustering. Experiments show that the proposed approach significantly improves convergence speed and the clustering effect.

Journal ArticleDOI
TL;DR: The dominant result has been obtained using the committee fuzzy MCDM and ANN to select the best green supplier.
Abstract: In recent years, environmental awareness has increased considerably, and in order to decrease endangerments such as air and water pollution, and also global warming, green procurement should be employed. Therefore, in the assessment of suppliers, their environmental performance should be taken into consideration along with other criteria for supplier selection. Raising awareness of sustainability in production and conservation and protection of the environment is very important both for the whole environment and for the company itself by increasing its competitive advantage. And, one of the steps to achieve this is for the companies to try to select green suppliers. So, the purpose of this study is to raise awareness and tackle the need for green supplier selection and, using multiple-criteria decision-making models, to elaborate a case study regarding this. A survey was conducted in a manufacturing firm. The data were analysed, and fuzzy MCDM (multicriteria decision-making) methods and artificial neural networks were implemented. Fuzzy methods are the fuzzy analytic hierarchy process (fuzzy AHP), fuzzy TOPSIS, and fuzzy ELECTRE. ANN supports the result of fuzzy MCDM models from the profit side. ANN can make the best estimate of the current year based on historical data. Fuzzy MCDM methods will also find good solutions using the available data but will produce different solutions as there are different decision-making methods. It is aimed to produce a synergy from the solutions obtained here and to produce a better solution. Instead of a single method, it would be more accurate to produce a better solution than the solution provided by all of them. The dominant result has been obtained using the committee fuzzy MCDM and ANN to select the best green supplier.

Journal ArticleDOI
TL;DR: Schizophrenia EEG signal classification is performed wherein, initially, features such as Detrend Fluctuation Analysis (DFA), Hurst Exponent, Recurrence Quantification Analysis (RQA), Sample Entropy, Fractal Dimension (FD), Kolmogorov Complexity, Hjorth exponent, Lempel Ziv Complexity (LZC), and Largest Lyapunov Exponent (LLE) are extracted initially.
Abstract: One of the serious mental disorders where people interpret reality in an abnormal state is schizophrenia. A combination of extremely disordered thinking, delusion, and hallucination is caused due to schizophrenia, and the daily functions of a person are severely disturbed because of this disorder. A wide range of problems are caused due to schizophrenia such as disturbed thinking and behaviour. In the field of human neuroscience, the analysis of brain activity is quite an important research area. For general cognitive activity analysis, electroencephalography (EEG) signals are widely used as a low-resolution diagnosis tool. The EEG signals are a great boon to understand the abnormality of the brain disorders, especially schizophrenia. In this work, schizophrenia EEG signal classification is performed wherein, initially, features such as Detrend Fluctuation Analysis (DFA), Hurst Exponent, Recurrence Quantification Analysis (RQA), Sample Entropy, Fractal Dimension (FD), Kolmogorov Complexity, Hjorth exponent, Lempel Ziv Complexity (LZC), and Largest Lyapunov Exponent (LLE) are extracted initially. The extracted features are, then, optimized for selecting the best features through four types of optimization algorithms here such as Artificial Flora (AF) optimization, Glowworm Search (GS) optimization, Black Hole (BH) optimization, and Monkey Search (MS) optimization, and finally, it is classified through certain classifiers. The best results show that, for normal cases, a classification accuracy of 87.54% is obtained when BH optimization is utilized with Support Vector Machine-Radial Basis Function (SVM-RBF) kernel, and for schizophrenia cases, a classification accuracy of 92.17% is obtained when BH optimization is utilized with SVM-RBF kernel.

Journal ArticleDOI
Qianhao Zhai1, Tao Ye2, Mengxing Huang1, Siling Feng1, H. Li1 
TL;DR: The whale optimization algorithm based on FTSE100 index data is used to optimize the above multiconstraint portfolio optimization model, which significantly improves the rate of return of the simple diversified buy-and-hold strategy or the FTSe100 index.
Abstract: In the field of asset allocation, how to balance the returns of an investment portfolio and its fluctuations is the core issue. Capital asset pricing model, arbitrage pricing theory, and Fama-French three-factor model were used to quantify the price of individual stocks and portfolios. Based on the second-order stochastic dominance rule, the higher moments of return series, the Shannon entropy, and some other actual investment constraints, we construct a multiconstraint portfolio optimization model, aiming at comprehensively weighting the returns and risk of portfolios rather than blindly maximizing its returns. Furthermore, the whale optimization algorithm based on FTSE100 index data is used to optimize the above multiconstraint portfolio optimization model, which significantly improves the rate of return of the simple diversified buy-and-hold strategy or the FTSE100 index. Furthermore, extensive experiments validate the superiority of the whale optimization algorithm over the other four swarm intelligence optimization algorithms (gray wolf optimizer, fruit fly optimization algorithm, particle swarm optimization, and firefly algorithm) through various indicators of the results, especially under harsh constraints.