scispace - formally typeset
Search or ask a question

Showing papers on "Unsupervised learning published in 2022"


Journal ArticleDOI
TL;DR: Multi-task learning (MTL) as mentioned in this paper is a learning paradigm in machine learning and its aim is to leverage useful information contained in multiple related tasks to help improve the generalization performance of all the tasks.
Abstract: Multi-Task Learning (MTL) is a learning paradigm in machine learning and its aim is to leverage useful information contained in multiple related tasks to help improve the generalization performance of all the tasks. In this paper, we give a survey for MTL from the perspective of algorithmic modeling, applications and theoretical analyses. For algorithmic modeling, we give a definition of MTL and then classify different MTL algorithms into five categories, including feature learning approach, low-rank approach, task clustering approach, task relation learning approach and decomposition approach as well as discussing the characteristics of each approach. In order to improve the performance of learning tasks further, MTL can be combined with other learning paradigms including semi-supervised learning, active learning, unsupervised learning, reinforcement learning, multi-view learning and graphical models. When the number of tasks is large or the data dimensionality is high, we review online, parallel and distributed MTL models as well as dimensionality reduction and feature hashing to reveal their computational and storage advantages. Many real-world applications use MTL to boost their performance and we review representative works in this paper. Finally, we present theoretical analyses and discuss several future directions for MTL.

223 citations


Journal ArticleDOI
TL;DR: In this paper, a comprehensive survey of the most important aspects of multi-sensor applications for human activity recognition, including those recently added to the field for unsupervised learning and transfer learning, is presented.

136 citations


Journal ArticleDOI
06 Jun 2022-Scanning
TL;DR: A novel research on hyperspectral microscopic picture using deep learning and effective unsupervised learning is explored and the Kullback–Leibler divergence is used to test the objective function convergence.
Abstract: Hyperspectral microscopy in biology and minerals, unsupervised deep learning neural network denoising SRS photos: hyperspectral resolution enhancement and denoising one hyperspectral picture is enough to teach unsupervised method. An intuitive chemical species map for a lithium ore sample is produced using k-means clustering. Many researchers are now interested in biosignals. Uncertainty limits the algorithms' capacity to evaluate these signals for further information. Even while AI systems can answer puzzles, they remain limited. Deep learning is used when machine learning is inefficient. Supervised learning needs a lot of data. Deep learning is vital in modern AI. Supervised learning requires a large labeled dataset. The selection of parameters prevents over- or underfitting. Unsupervised learning is used to overcome the challenges outlined above (performed by the clustering algorithm). To accomplish this, two processing processes were used: (1) utilizing nonlinear deep learning networks to turn data into a latent feature space (Z). The Kullback–Leibler divergence is used to test the objective function convergence. This article explores a novel research on hyperspectral microscopic picture using deep learning and effective unsupervised learning.

116 citations


Journal ArticleDOI
TL;DR: A critical review of the existing internal combustion engine (ICE) modeling, optimization, diagnosis, and control challenges and the promising state-of-the-art Machine Learning (ML) solutions for them is provided in this paper.

50 citations


Journal ArticleDOI
TL;DR: Recently, many efforts have been made on training sophisticated models with few labeled data in an unsupervised and semi-supervised fashion as mentioned in this paper , and the recent progresses on these two major categories of methods are reviewed in this paper.
Abstract: Representation learning with small labeled data have emerged in many problems, since the success of deep neural networks often relies on the availability of a huge amount of labeled data that is expensive to collect. To address it, many efforts have been made on training sophisticated models with few labeled data in an unsupervised and semi-supervised fashion. In this paper, we will review the recent progresses on these two major categories of methods. A wide spectrum of models will be categorized in a big picture, where we will show how they interplay with each other to motivate explorations of new ideas. We will review the principles of learning the transformation equivariant, disentangled, self-supervised and semi-supervised representations, all of which underpin the foundation of recent progresses. Many implementations of unsupervised and semi-supervised generative models have been developed on the basis of these criteria, greatly expanding the territory of existing autoencoders, generative adversarial nets (GANs) and other deep networks by exploring the distribution of unlabeled data for more powerful representations. We will discuss emerging topics by revealing the intrinsic connections between unsupervised and semi-supervised learning, and propose in future directions to bridge the algorithmic and theoretical gap between transformation equivariance for unsupervised learning and supervised invariance for supervised learning, and unify unsupervised pretraining and supervised finetuning. We will also provide a broader outlook of future directions to unify transformation and instance equivariances for representation learning, connect unsupervised and semi-supervised augmentations, and explore the role of the self-supervised regularization for many learning problems.

44 citations


Journal ArticleDOI
TL;DR: In this paper , a novel unsupervised learning framework based on concept-based and hierarchical clustering is proposed for Twitter sentiment analysis, and two different feature representation methods including Boolean and Term frequency-inverse document frequency (TF-IDF) are investigated.

43 citations


Journal ArticleDOI
TL;DR: In this paper, a taxonomy of X-ray security imaging algorithms is presented, with a particular focus on object classification, detection, segmentation and anomaly detection tasks, and a performance benchmark is provided based on the current and future trends in deep learning.

37 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a hybrid contrastive model (HCM) for unsupervised person Re-ID, where an identity-based memory is constructed to store pedestrian features and an image based memory is established to store each image feature.
Abstract: Unsupervised person re-identification (Re-ID) aims to learn discriminative features without human-annotated labels. Recently, contrastive learning provides a new prospect for unsupervised person Re-ID, and existing methods mainly constrain the feature similarity among easy sample pairs. However, the feature similarity among hard sample pairs is neglected, which causes suboptimal performance in unsupervised person Re-ID. In the paper, we propose a novel Hybrid Contrastive Model (HCM) to perform the identity-level contrastive learning and the image-level contrastive learning for unsupervised person Re-ID, which adequately explores the feature similarity among hard sample pairs. Specifically, for the identity-level contrastive learning, an identity-based memory is constructed to store pedestrian features. Accordingly, we define the dynamic contrast loss to identify identity information with dynamic factor for distinguishing hard/easy samples. As for the image-level contrastive learning, an image-based memory is established to store each image feature. We design the sample constraint loss to explore the similarity relationship between hard positive and negative sample pairs. Furthermore, we optimize the two contrastive learning processes in one unified framework to make use of their own advantages as so to constrain the feature distribution for extracting potential information. Extensive experiments prove that the proposed HCM distinctly outperforms the state-of-the-art methods.

35 citations


Posted ContentDOI
TL;DR: This study develops explainable data-driven IFD approaches for nonlinear dynamic systems through a generalized kernel representation for system modeling and the associated fault diagnosis, and discovers the existence of a bridge between some supervised and unsupervised learning-based entities.
Abstract: The increased complexity and intelligence of automation systems require the development of intelligent fault diagnosis (IFD) methodologies. By relying on the concept of a suspected space, this study develops explainable data-driven IFD approaches for nonlinear dynamic systems. More specifically, we parameterize nonlinear systems through a generalized kernel representation for system modeling and the associated fault diagnosis. An important result obtained is a unified form of kernel representations, applicable to both unsupervised and supervised learning. More importantly, through a rigorous theoretical analysis, we discover the existence of a bridge (i.e., a bijective mapping) between some supervised and unsupervised learning-based entities. Notably, the designed IFD approaches achieve the same performance with the use of this bridge. In order to have a better understanding of the results obtained, both unsupervised and supervised neural networks are chosen as the learning tools to identify the generalized kernel representations and design the IFD schemes; an invertible neural network is then employed to build the bridge between them. This article is a perspective article, whose contribution lies in proposing and formalizing the fundamental concepts for explainable intelligent learning methods, contributing to system modeling and data-driven IFD designs for nonlinear dynamic systems.

32 citations


Journal ArticleDOI
TL;DR: In this paper , a survey focusing on four types of methods from machine learning for intrusion and anomaly detection, namely, supervised, semi-supervised, unsupervised and reinforcement learning, is presented.

30 citations


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed an unsupervised re-identification learning module and an occlusion estimation module to predict the locations where occlusions happen, which are used to estimate the positions of missed objects by the detector.

Journal ArticleDOI
01 Oct 2022-Cities
TL;DR: In this paper , the authors provide a systematic review of the use of unsupervised learning in urban studies based on 140 publications, and discuss how UL is applied in a broad range of urban topics.

Journal ArticleDOI
TL;DR: A general-purpose method for modeling and predicting the binding affinities of protein-peptide interactions (PpIs) at the structural level and examines the robustness and fault-tolerance of usPpIA predictor when applied to treat the coarse-grained PpI complex structures modeled computationally by sophisticated peptide docking and dynamics simulation.
Abstract: Cell signal networks are orchestrated directly or indirectly by various peptide-mediated protein-protein interactions, which are normally weak and transient and thus ideal for biological regulation and medicinal intervention. Here, we develop a general-purpose method for modeling and predicting the binding affinities of protein-peptide interactions (PpIs) at the structural level. The method is a hybrid strategy that employs an unsupervised approach to derive a layered PpI atom-residue interaction (ulPpI[a-r]) potential between different protein atom types and peptide residue types from thousands of solved PpI complex structures and then statistically correlates the potential descriptors with experimental affinities (KD values) over hundreds of known PpI samples in a supervised manner to create an integrated unsupervised-supervised PpI affinity (usPpIA) predictor. Although both the ulPpI[a-r] potential and usPpIA predictor can be used to calculate PpI affinities from their complex structures, the latter seems to perform much better than the former, suggesting that the unsupervised potential can be improved substantially with a further correction by supervised statistical learning. We examine the robustness and fault-tolerance of usPpIA predictor when applied to treat the coarse-grained PpI complex structures modeled computationally by sophisticated peptide docking and dynamics simulation. It is revealed that, despite developed solely based on solved structures, the integrated unsupervised-supervised method is also applicable for locally docked structures to reach a quantitative prediction but can only give a qualitative prediction on globally docked structures. The dynamics refinement seems not to change (or improve) the predictive results essentially, although it is computationally expensive and time-consuming relative to peptide docking. We also perform extrapolation of usPpIA predictor to the indirect affinity quantities of HLA-A*0201 binding epitope peptides and NHERF PDZ binding scaffold peptides, consequently resulting in a good and moderate correlation of the predicted KD with experimental IC50 and BLU on the two peptide sets, with Pearson's correlation coefficients Rp = 0.635 and 0.406, respectively.

Journal ArticleDOI
TL;DR: An overview of the autoML field with a particular emphasis on the automated methods and strategies that have been proposed for unsupervised anomaly detection can be found in this article , where the authors present an overview of autoML algorithms for anomaly detection.
Abstract: The last decade has witnessed the explosion of machine learning research studies with the inception of several algorithms proposed and successfully adopted in different application domains. However, the performance of multiple machine learning algorithms is very sensitive to multiple ingredients (e.g., hyper-parameters tuning and data cleaning) where a significant human effort is required to achieve good results. Thus, building well-performing machine learning algorithms requires domain knowledge and highly specialized data scientists. Automated machine learning (autoML) aims to make easier and more accessible the use of machine learning algorithms for researchers with varying levels of expertise. Besides, research effort to date has mainly been devoted to autoML for supervised learning, and only a few research proposals have been provided for the unsupervised learning. In this paper, we present an overview of the autoML field with a particular emphasis on the automated methods and strategies that have been proposed for unsupervised anomaly detection.

DOI
15 Mar 2022
TL;DR: This article applied a contrastive self-supervised learning method to digital histopathology by collecting and pretraining on 57 histopathological datasets without any labels and found that combining multiple multi-organ datasets with different types of staining and resolution properties improves the quality of the learned features.
Abstract: Unsupervised learning has been a long-standing goal of machine learning and is especially important for medical image analysis, where the learning can compensate for the scarcity of labeled datasets. A promising subclass of unsupervised learning is self-supervised learning, which aims to learn salient features using the raw input as the learning signal. In this work, we tackle the issue of learning domain-specific features without any supervision to improve multiple task performances that are of interest to the digital histopathology community. We apply a contrastive self-supervised learning method to digital histopathology by collecting and pretraining on 57 histopathology datasets without any labels. We find that combining multiple multi-organ datasets with different types of staining and resolution properties improves the quality of the learned features. Furthermore, we find using more images for pretraining leads to a better performance in multiple downstream tasks, albeit there are diminishing returns as more unlabeled images are incorporated into the pretraining. Linear classifiers trained on top of the learned features show that networks pretrained on digital histopathology datasets perform better than ImageNet pretrained networks, boosting task performances by more than 28% in F 1 scores on average. Interestingly, we did not observe a consistent correlation between the pretraining dataset site or the organ versus the downstream task (e.g., pretraining with only breast images does not necessarily lead to a superior downstream task performance for breast-related tasks). These findings may also be useful when applying newer contrastive techniques to histopathology data. Pretrained PyTorch models are made publicly available at https://github.com/ozanciga/self-supervised-histopathology .

Journal ArticleDOI
TL;DR: In this paper, the authors evaluated four unsupervised machine learning methods on two recent datasets and then defined their generalization strength using a novel inter-dataset evaluation strategy estimating their adaptability.
Abstract: Through the ongoing digitization of the world, the number of connected devices is continuously growing without any foreseen decline in the near future. In particular, these devices increasingly include critical systems such as power grids and medical institutions, possibly causing tremendous consequences in the case of a successful cybersecurity attack. A network intrusion detection system (NIDS) is one of the main components to detect ongoing attacks by differentiating normal from malicious traffic. Anomaly-based NIDS, more specifically unsupervised methods previously proved promising for their ability to detect known as well as zero-day attacks without the need for a labeled dataset. Despite decades of development by researchers, anomaly-based NIDS are only rarely employed in real-world applications, most possibly due to the lack of generalization power of the proposed models. This article first evaluates four unsupervised machine learning methods on two recent datasets and then defines their generalization strength using a novel inter-dataset evaluation strategy estimating their adaptability. Results show that all models can present high classification scores on an individual dataset but fail to directly transfer those to a second unseen but related dataset. Specifically, the accuracy dropped on average 25.63% in an inter-dataset setting compared to the conventional evaluation approach. This generalization challenge can be observed and tackled in future research with the help of the proposed evaluation strategy in this paper.

Journal ArticleDOI
TL;DR: In this paper , an explainable data-driven intelligent fault diagnosis (IFD) approach for nonlinear dynamic systems is proposed, which is based on the concept of a suspected space.
Abstract: The increased complexity and intelligence of automation systems require the development of intelligent fault diagnosis (IFD) methodologies. By relying on the concept of a suspected space, this study develops explainable data-driven IFD approaches for nonlinear dynamic systems. More specifically, we parameterize nonlinear systems through a generalized kernel representation for system modeling and the associated fault diagnosis. An important result obtained is a unified form of kernel representations, applicable to both unsupervised and supervised learning. More importantly, through a rigorous theoretical analysis, we discover the existence of a bridge (i.e., a bijective mapping) between some supervised and unsupervised learning-based entities. Notably, the designed IFD approaches achieve the same performance with the use of this bridge. In order to have a better understanding of the results obtained, both unsupervised and supervised neural networks are chosen as the learning tools to identify the generalized kernel representations and design the IFD schemes; an invertible neural network is then employed to build the bridge between them. This article is a perspective article, whose contribution lies in proposing and formalizing the fundamental concepts for explainable intelligent learning methods, contributing to system modeling and data-driven IFD designs for nonlinear dynamic systems.

Journal ArticleDOI
TL;DR: This article proposed an unsupervised contrastive learning framework that is motivated from the perspective of label smoothing, which uses a novel contrastive loss that naturally exploits a data augmentation scheme in which new samples are generated by mixing two data samples with a mixing component.

Journal ArticleDOI
TL;DR: Recently, deep learning has become the main research frontier for biological image reconstruction and enhancement problems thanks to their high performance and ultrafast inference times as discussed by the authors . However, due to the difficulty of obtaining matched reference data for supervised learning, there has been increasing interest in unsupervised learning approaches that do not need paired reference data.
Abstract: Recently, deep learning (DL) approaches have become the main research frontier for biological image reconstruction and enhancement problems thanks to their high performance and ultrafast inference times. However, due to the difficulty of obtaining matched reference data for supervised learning, there has been increasing interest in unsupervised learning approaches that do not need paired reference data. In particular, self-supervised learning and generative models have been successfully used for various biological imaging applications. In this article, we provide an overview of these approaches from a coherent perspective in the context of classical inverse problems and discuss their applications to biological imaging, including electron, fluorescence, deconvolution microscopy, optical diffraction tomography (ODT), and functional neuroimaging.

Journal ArticleDOI
TL;DR: This work proposes a general framework for learning node representations in a self supervised manner called Graph Constrastive Learning (GraphCL), which learns node embeddings by maximizing the similarity between the nodes representations of two randomly perturbed versions of the same graph.

Journal ArticleDOI
TL;DR: Zhang et al. as discussed by the authors proposed an unsupervised deep Random Vector Functional Link (usRVFL) based on the manifold regularization, which can increase the capability and diversity of RVFLs.

Journal ArticleDOI
TL;DR: A new framework named Spatio-Temporal Association Rule based Deep Annotation-free Clustering (STAR-DAC) which incrementally clusters the unlabeled person re-identification images based on visual features and performs cluster fine-tuning through the mined spatio-temporal association rules.

Journal ArticleDOI
TL;DR: This article is the first attempt to fine-tune SSAE in an unsupervised manner, and to propose an un supervised fault recognition framework that requires no prior knowledge or data labels at all.
Abstract: Currently, with the development of the Internet of Things (IoTs) and artificial intelligence, a new IoT structure known as the artificial Intelligence of Things (AIoTs) comes into play. With the development of AIoT, a large amount of unlabeled industrial big data has been accumulated. The analysis of large amounts of unlabeled data is labor-intensive and time-consuming for diagnostic personnel. To improve this situation, a novel two-stage unsupervised fault recognition algorithm, namely, deep adaptive fuzzy clustering algorithm (DAFC) is proposed for unsupervised fault clustering in this article. DAFC amalgamates stacked sparse autoencoder (SSAE) into adaptive weighted Gath–Geva (AWGG) clustering to form an unsupervised fault recognition framework for clustering analysis of unlabeled industrial big data. SSAE can extract the highly abstract features of the original data, and adopt different unsupervised strategies to fine-tune the network in two stages. AWGG is an improvement of Gath–Geva clustering, and can adaptively obtain optimal clustering results without presetting the number of clusters. Experimental results on two different datasets show that the proposed DAFC can stably extract fault features from unlabeled data, and automatically obtain the optimal clustering results without knowing the number of clusters in advance. To the best of our knowledge, this article is the first attempt to fine-tune SSAE in an unsupervised manner, and to propose an unsupervised fault recognition framework that requires no prior knowledge or data labels at all. DAFC can be a feasible industrial big data application for collaborative AIoT. Diagnostic personnel analyze the clustering results obtained by DAFC instead of the original unlabeled data, greatly saving time and labor costs.

Journal ArticleDOI
TL;DR: In this paper , a VAE-based 1SVM detector was proposed to detect COVID-19 infection using a blood test as an anomaly detection problem through an unsupervised deep hybrid model, which combines the features extraction capability of the variational autoencoder (VAE) and the detection sensitivity of the one-class SVM algorithm.
Abstract: A sample blood test has recently become an important tool to help identify false-positive/false-negative real-time reverse transcription polymerase chain reaction (rRT-PCR) tests. Importantly, this is mainly because it is an inexpensive and handy option to detect the potential COVID-19 patients. However, this test should be conducted by certified laboratories, expensive equipment, and trained personnel, and 3-4 h are needed to deliver results. Furthermore, it has relatively large false-negative rates around 15%-20%. Consequently, an alternative and more accessible solution, quicker and less costly, is needed. This article introduces flexible and unsupervised data-driven approaches to detect the COVID-19 infection based on blood test samples. In other words, we address the problem of COVID-19 infection detection using a blood test as an anomaly detection problem through an unsupervised deep hybrid model. Essentially, we amalgamate the features extraction capability of the variational autoencoder (VAE) and the detection sensitivity of the one-class support vector machine (1SVM) algorithm. Two sets of routine blood tests samples from the Albert Einstein Hospital, S ao Paulo, Brazil, and the San Raffaele Hospital, Milan, Italy, are used to assess the performance of the investigated deep learning models. Here, missing values have been imputed based on a random forest regressor. Compared to generative adversarial networks (GANs), deep belief network (DBN), and restricted Boltzmann machine (RBM)-based 1SVM, the traditional VAE, GAN, DBN, and RBM with softmax layer as discriminator layer, and the standalone 1SVM, the proposed VAE-based 1SVM detector offers superior discrimination performance of potential COVID-19 infections. Results also revealed that the deep learning-driven 1SVM detection approaches provide promising detection performance compared to the conventional deep learning models.

MonographDOI
30 Jun 2022
TL;DR: This book presents a unified theory of random matrices for applications in machine learning, offering a large-dimensional data vision that exploits concentration and universality phenomena, to enable a precise understanding of the core mechanisms at play in real-world machine learning algorithms.
Abstract: This book presents a unified theory of random matrices for applications in machine learning, offering a large-dimensional data vision that exploits concentration and universality phenomena. This enables a precise understanding, and possible improvements, of the core mechanisms at play in real-world machine learning algorithms. The book opens with a thorough introduction to the theoretical basics of random matrices, which serves as a support to a wide scope of applications ranging from SVMs, through semi-supervised learning, unsupervised spectral clustering, and graph methods, to neural networks and deep learning. For each application, the authors discuss small- versus large-dimensional intuitions of the problem, followed by a systematic random matrix analysis of the resulting performance and possible improvements. All concepts, applications, and variations are illustrated numerically on synthetic as well as real-world data, with MATLAB and Python code provided on the accompanying website.

Proceedings ArticleDOI
01 Jun 2022
TL;DR: Wang et al. as discussed by the authors proposed an unsupervised Generative Cooperative Learning (GCL) approach for video anomaly detection that exploits the low frequency of anomalies towards building a cross-supervision between a generator and a discriminator.
Abstract: Video anomaly detection is well investigated in weakly-supervised and one-class classification (OCC) settings. However, unsupervised video anomaly detection methods are quite sparse, likely because anomalies are less frequent in occurrence and usually not well-defined, which when coupled with the absence of ground truth supervision, could adversely affect the performance of the learning algorithms. This problem is challenging yet rewarding as it can completely eradicate the costs of obtaining laborious annotations and enable such systems to be deployed without human intervention. To this end, we propose a novel unsupervised Generative Cooperative Learning (GCL) approach for video anomaly detection that exploits the low frequency of anomalies towards building a cross-supervision between a generator and a discriminator. In essence, both networks get trained in a cooperative fashion, thereby allowing unsupervised learning. We conduct extensive experiments on two large-scale video anomaly detection datasets, UCF crime and ShanghaiTech. Consistent improvement over the existing state-of-the-art unsupervised and OCC methods corroborate the effectiveness of our approach.

Journal ArticleDOI
TL;DR: In this paper, the authors presented a techno-economic life-extension analysis for fixed offshore wind turbines for the purpose of classification and certification, which combines revenue estimates and operational expenditures, considering the lifeextension duration and appropriate discount rate.

Journal ArticleDOI
TL;DR: Wang et al. as mentioned in this paper proposed a multi-cluster feature selection based on isometric mapping (MCFS-I) algorithm to perform unsupervised feature selection adaptively for multiple clusters.
Abstract: This letter presents an unsupervised feature selection method based on machine learning. Feature selection is an important component of artificial intelligence, machine learning, which can effectively solve the curse of dimensionality problem. Since most of the labeled data is expensive to obtain, this paper focuses on the unsupervised feature selection method. The distance metric of traditional unsupervised feature selection algorithms is usually based on Euclidean distance, and it is maybe unreasonable to map high-dimensional data into low-dimensional space by using Euclidean distance. Inspired by this, this paper combines manifold learning to improve the multi-cluster unsupervised feature selection algorithm. By using geodesic distance, we propose a multi-cluster feature selection based on isometric mapping (MCFS-I) algorithm to perform unsupervised feature selection adaptively for multiple clusters. Experimental results show that the proposed method consistently improves the clustering performance compared to the existing competing methods.


Journal ArticleDOI
TL;DR: Zhang et al. as mentioned in this paper proposed a multi-label learning guided self-paced clustering (MLC) to learn discriminative features without annotations across disjoint camera views.