scispace - formally typeset
Search or ask a question

Showing papers on "Markov random field published in 2019"


Posted Content
TL;DR: It is shown that BERT (Devlin et al., 2018) is a Markov random field language model, and this formulation gives way to a natural procedure to sample sentences from BERT, which can produce high quality, fluent generations.
Abstract: We show that BERT (Devlin et al., 2018) is a Markov random field language model. This formulation gives way to a natural procedure to sample sentences from BERT. We generate from BERT and find that it can produce high-quality, fluent generations. Compared to the generations of a traditional left-to-right language model, BERT generates sentences that are more diverse but of slightly worse quality.

201 citations


Proceedings ArticleDOI
11 Feb 2019
TL;DR: This paper showed that BERT is a Markov random field language model and showed that it can produce high quality, fluent generations, compared to the generations of a traditional left-to-right language model.
Abstract: We show that BERT (Devlin et al., 2018) is a Markov random field language model. This formulation gives way to a natural procedure to sample sentences from BERT. We generate from BERT and find that it can produce high quality, fluent generations. Compared to the generations of a traditional left-to-right language model, BERT generates sentences that are more diverse but of slightly worse quality.

136 citations


Journal ArticleDOI
TL;DR: Improved change detection-based Markov random field approach is improved by integrating normalized difference vegetation index, principal component analysis, and independent component analysis generated CDIs with MRF for landslide inventory mapping from multi-sensor data.

101 citations


Posted Content
TL;DR: In this article, the reprojection error of the 3D model with respect to the image estimates is directly optimized over rays, where the cost function depends on the semantic class and depth of the first occupied voxel along the ray.
Abstract: Dense semantic 3D reconstruction is typically formulated as a discrete or continuous problem over label assignments in a voxel grid, combining semantic and depth likelihoods in a Markov Random Field framework. The depth and semantic information is incorporated as a unary potential, smoothed by a pairwise regularizer. However, modelling likelihoods as a unary potential does not model the problem correctly leading to various undesirable visibility artifacts. We propose to formulate an optimization problem that directly optimizes the reprojection error of the 3D model with respect to the image estimates, which corresponds to the optimization over rays, where the cost function depends on the semantic class and depth of the first occupied voxel along the ray. The 2-label formulation is made feasible by transforming it into a graph-representable form under QPBO relaxation, solvable using graph cut. The multi-label problem is solved by applying alpha-expansion using the same relaxation in each expansion move. Our method was indeed shown to be feasible in practice, running comparably fast to the competing methods, while not suffering from ray potential approximation artifacts.

52 citations


Journal ArticleDOI
TL;DR: The proposed approach can greatly improve the potential for semisupervised learning in hyperspectral image classification and proposes a density-peak-based augmentation strategy for pseudolabels, due to the fact that the samples without manual labels in their super pixels are out of reach for the automatic sampling.
Abstract: In this work, we propose a new semisupervised active learning approach for hyperspectral image classification. The proposed method aims at improving machine generalization by using pseudolabeled samples, both confident and informative, which are automatically and actively selected, via semisupervised learning. The learning is performed under two assumptions: a local one for the labeling via a superpixel-based constraint dedicated to the spatial homogeneity and adaptivity into the pseudolabels, and a global one modeling the data density by a multinomial logistic regressor with a Markov random field regularizer. Furthermore, we propose a density-peak-based augmentation strategy for pseudolabels, due to the fact that the samples without manual labels in their superpixel neighborhoods are out of reach for the automatic sampling. Three real hyperspectral datasets were used in our experiments to evaluate the effectiveness of the proposed superpixel-based semisupervised learning approach. The obtained results indicate that the proposed approach can greatly improve the potential for semisupervised learning in hyperspectral image classification.

46 citations


Journal ArticleDOI
TL;DR: An active deep learning approach for minimally supervised PolSAR image classification, which integrates active learning and fine-tuned convolutional neural network (CNN) into a principled framework and achieves state-of-the-art classification results with significantly reduced annotation cost.
Abstract: Recently, deep neural networks have received intense interests in polarimetric synthetic aperture radar (PolSAR) image classification. However, its success is subject to the availability of large amounts of annotated data which require great efforts of experienced human annotators. Aiming at improving the classification performance with greatly reduced annotation cost, this paper presents an active deep learning approach for minimally supervised PolSAR image classification, which integrates active learning and fine-tuned convolutional neural network (CNN) into a principled framework. Starting from a CNN trained using a very limited number of labeled pixels, we iteratively and actively select the most informative candidates for annotation, and incrementally fine-tune the CNN by incorporating the newly annotated pixels. Moreover, to boost the performance and robustness of the proposed method, we employ Markov random field (MRF) to enforce class label smoothness, and data augmentation technique to enlarge the training set. We conducted extensive experiments on four real benchmark PolSAR images, and experiments demonstrated that our approach achieved state-of-the-art classification results with significantly reduced annotation cost.

46 citations


Journal ArticleDOI
TL;DR: An improved method that combines cooperative multitemporal segmentation and hierarchical compound classification (CMS-HCC) based on previous work can effectively detect the changes in heterogeneous images, with low false positive and high accuracy.
Abstract: Change detection in heterogeneous remote sensing images is an important but challenging task because of the incommensurable appearances of the heterogeneous images. In order to solve the change detection problem in optical and synthetic aperture radar (SAR) images, this paper proposes an improved method that combines cooperative multitemporal segmentation and hierarchical compound classification (CMS-HCC) based on our previous work. Considering the large radiometric and geometric differences between heterogeneous images, first, a cooperative multitemporal segmentation method is introduced to generate multi-scale segmentation results. This method segments two images together by associating the information from the two images and thus reduces the noises and errors caused by area transition and object misalignment, as well as makes the boundaries of detected objects described more accurately. Then, a region-based multitemporal hierarchical Markov random field (RMH-MRF) model is defined to combine spatial, temporal, and multi-level information. With the RMH-MRF model, a hierarchical compound classification method is performed by identifying the optimal configuration of labels with a region-based marginal posterior mode estimation, further improving the change detection accuracy. The changes can be determined if the labels assigned to each pair of parcels are different, obtaining multi-scale change maps. Experimental validation is conducted on several pairs of optical and SAR images. It consists of two parts: comparison on different multitemporal segmentation methods and comparison on different change detection methods. The results show that the proposed method can effectively detect the changes in heterogeneous images, with low false positive and high accuracy.

45 citations


Journal ArticleDOI
TL;DR: This study provides a comprehensive study of state-of-the-art image denoising methods using CNN and shows PDNN shows the best result in terms of PSNR for both BSD-68 and Set-12 datasets.
Abstract: Convolutional neural networks (CNNs) are deep neural networks that can be trained on large databases and show outstanding performance on object classification, segmentation, image denoising etc. In the past few years, several image denoising techniques have been developed to improve the quality of an image. The CNN based image denoising models have shown improvement in denoising performance as compared to non-CNN methods like block-matching and three-dimensional (3D) filtering, contemporary wavelet and Markov random field approaches etc. which had remained state-of-the-art for years. This study provides a comprehensive study of state-of-the-art image denoising methods using CNN. The literature associated with different CNNs used for image restoration like residual learning based models (DnCNN-S, DnCNN-B, IDCNN), non-locality reinforced (NN3D), fast and flexible network (FFDNet), deep shrinkage CNN (SCNN), a model for mixed noise reduction, denoising prior driven network (PDNN) are reviewed. DnCNN-S and PDNN remove Gaussian noise of fixed level, whereas DnCNN-B, IDCNN, NN3D and SCNN are used for blind Gaussian denoising. FFDNet is used for spatially variant Gaussian noise. The performance of these CNN models is analysed on BSD-68 and Set-12 datasets. PDNN shows the best result in terms of PSNR for both BSD-68 and Set-12 datasets.

42 citations


Journal ArticleDOI
TL;DR: A disparity refinement method that directly refines the winner-take-all (WTA) disparity map by exploring its statistical significance by design a two-layer optimization to refine the disparity plane.
Abstract: In this paper, we propose a disparity refinement method that directly refines the winner-take-all (WTA) disparity map by exploring its statistical significance. According to the primary steps of the segment-based stereo matching, the reference image is over-segmented into superpixels and a disparity plane is fitted for each superpixel by an improved random sample consensus (RANSAC). We design a two-layer optimization to refine the disparity plane. In the global optimization, mean disparities of superpixels are estimated by Markov random field (MRF) inference, and then, a 3D neighborhood system is derived from the mean disparities for occlusion handling. In the local optimization, a probability model exploiting Bayesian inference and Bayesian prediction is adopted and achieves second-order smoothness implicitly among 3D neighbors. The two-layer optimization is a pure disparity refinement method because no correlation information between stereo image pairs is demanded during the refinement. Experimental results on the Middlebury and KITTI datasets demonstrate that the proposed method can perform accurate stereo matching with a faster speed and handle the occlusion effectively. It can be indicated that the “matching cost computation + disparity refinement” framework is a possible solution to produce accurate disparity map at low computational cost.

37 citations


Proceedings ArticleDOI
11 Apr 2019
TL;DR: In this article, a new geometric and probabilistic approach to synchronization of correspondences across multiple sets of objects or images is presented, based on the first order retraction operators.
Abstract: We present an entirely new geometric and probabilistic approach to synchronization of correspondences across multiple sets of objects or images. In particular, we present two algorithms: (1) Birkhoff-Riemannian L-BFGS for optimizing the relaxed version of the combinatorially intractable cycle consistency loss in a principled manner, (2) Birkhoff-Riemannian Langevin Monte Carlo for generating samples on the Birkhoff Polytope and estimating the confidence of the found solutions. To this end, we first introduce the very recently developed Riemannian geometry of the Birkhoff Polytope. Next, we introduce a new probabilistic synchronization model in the form of a Markov Random Field (MRF). Finally, based on the first order retraction operators, we formulate our problem as simulating a stochastic differential equation and devise new integrators. We show on both synthetic and real datasets that we achieve high quality multi-graph matching results with faster convergence and reliable confidence/uncertainty estimates.

36 citations


Proceedings ArticleDOI
01 Jul 2019
TL;DR: This paper presents an active deep learning approach for minimally-supervised PolSAR image classification, which integrates active learning and fine-tuning convolutional neural network (CNN) into a principled framework.
Abstract: Aiming at improving the classification performance with greatly reduced annotation cost, this paper presents an active deep learning approach for minimally-supervised PolSAR image classification, which integrates active learning and fine-tuning convolutional neural network (CNN) into a principled framework. Starting from a CNN trained using a very limited number of labeled pixels, we iteratively and actively select the most informative candidates for annotation, and incrementally fine-tune the CNN by incorporating the newly annotated pixels. Moreover, to boost the performance and robustness of the proposed method, we employ Markov random field to enforce label smoothness, and data augmentation technique to enlarge the training set. Extensive experiments demonstrated that our approach achieved state-of-the-art classification results with significantly reduced annotation cost.

Journal ArticleDOI
TL;DR: Both qualitative and quantitative experiments implemented on visible (Vis) and infrared (IR) UAV videos prove that the presented tracker can achieve better performances in terms of precision rate and success rate when compared with other state-of-the-art trackers.
Abstract: Target tracking based on unmanned aerial vehicle (UAV) video is a significant technique in intelligent urban surveillance systems for smart city applications, such as smart transportation, road traffic monitoring, inspection of stolen vehicle, etc. In this paper, a computer vision-based target tracking algorithm aiming at locating UAV-captured targets, like pedestrian and vehicle, is proposed using sparse representation theory. First of all, each target candidate is sparsely represented in the subspace spanned by a joint dictionary. Then, the sparse representation coefficient is further constrained by an $L_{2}$ regularization based on temporal consistency. To cope with the partial occlusion appearing in UAV videos, a Markov random field (MRF)-based binary support vector with contiguous occlusion constraint is introduced to our sparse representation model. For long-term tracking, the particle filter framework along with a dynamic template update scheme is designed. Both qualitative and quantitative experiments implemented on visible (Vis) and infrared (IR) UAV videos prove that the presented tracker can achieve better performances in terms of precision rate and success rate when compared with other state-of-the-art trackers.

Journal ArticleDOI
TL;DR: The proposed method to segment sub-acute ischemic stroke lesions from fluid-attenuated inversion recovery (FLAIR) magnetic resonance imaging (MRI) datasets based on 151 multi-center datasets from three different databases is developed and evaluated.
Abstract: Robust and reliable stroke lesion segmentation is a crucial step towards employing lesion volume as an independent endpoint for randomized trials. The aim of this work was to develop and evaluate a novel method to segment sub-acute ischemic stroke lesions from fluid-attenuated inversion recovery (FLAIR) magnetic resonance imaging (MRI) datasets. After preprocessing of the datasets, a Bayesian technique based on Gabor textures extracted from the FLAIR signal intensities is utilized to generate a first estimate of the lesion segmentation. Using this initial segmentation, a customized voxel-level Markov random field model based on intensity as well as Gabor texture features is employed to refine the stroke lesion segmentation. The proposed method was developed and evaluated based on 151 multi-center datasets from three different databases using a leave-one-patient-out validation approach. The comparison of the automatically segmented stroke lesions with manual ground truth segmentation revealed an average Dice coefficient of 0.582, which is in the upper range of previously presented lesion segmentation methods using multi-modal MRI datasets. Furthermore, the results obtained by the proposed technique are superior compared to the results obtained by two methods based on convolutional neural networks and three phase level-sets, respectively, which performed best in the ISLES 2015 challenge using multi-modal imaging datasets. The results of the quantitative evaluation suggest that the proposed method leads to robust lesion segmentation results using FLAIR MRI datasets only as a follow-up sequence.

Journal ArticleDOI
17 Jul 2019
TL;DR: This work proposes a general Markov Random Field (MRF) framework to incorporate coupling in network embedding which allows better detecting network communities, and improves the accuracy of existing embedding methods, and corrects most wrongly-divided statistically-significant nodes.
Abstract: Recent research on community detection focuses on learning representations of nodes using different network embedding methods, and then feeding them as normal features to clustering algorithms. However, we find that though one may have good results by direct clustering based on such network embedding features, there is ample room for improvement. More seriously, in many real networks, some statisticallysignificant nodes which play pivotal roles are often divided into incorrect communities using network embedding methods. This is because while some distance measures are used to capture the spatial relationship between nodes by embedding, the nodes after mapping to feature vectors are essentially not coupled any more, losing important structural information. To address this problem, we propose a general Markov Random Field (MRF) framework to incorporate coupling in network embedding which allows better detecting network communities. By smartly utilizing properties of MRF, the new framework not only preserves the advantages of network embedding (e.g. low complexity, high parallelizability and applicability for traditional machine learning), but also alleviates its core drawback of inadequate representations of dependencies via making up the missing coupling relationships. Experiments on real networks show that our new approach improves the accuracy of existing embedding methods (e.g. Node2Vec, DeepWalk and MNMF), and corrects most wrongly-divided statistically-significant nodes, which makes network embedding essentially suitable for real community detection applications. The new approach also outperforms other state-of-the-art conventional community detection methods.

Journal ArticleDOI
TL;DR: A flexible Markov random field model for microbial network structure is proposed and a hypothesis testing framework for detecting differences between networks, also known as differential network analysis is introduced, which is particularly powerful against sparse alternatives.
Abstract: Micro-organisms such as bacteria form complex ecological community networks that can be greatly influenced by diet and other environmental factors. Differential analysis of microbial community structures aims to elucidate systematic changes during an adaptive response to changes in environment. In this paper, we propose a flexible Markov random field model for microbial network structure and introduce a hypothesis testing framework for detecting differences between networks, also known as differential network analysis. Our global test for differential networks is particularly powerful against sparse alternatives. In addition, we develop a multiple testing procedure with false discovery rate control to identify the structure of the differential network. The proposed method is applied to data from a gut microbiome study on U.K. twins to evaluate how age affects the microbial community network.

Journal ArticleDOI
TL;DR: A Bayesian probabilistic modeling framework for the GLCM as a multivariate object as well as its application within a cancer detection context based on computed tomography is presented.
Abstract: The emerging field of cancer radiomics endeavors to characterize intrinsic patterns of tumor phenotypes and surrogate markers of response by transforming medical images into objects that yield quantifiable summary statistics to which regression and machine learning algorithms may be applied for statistical interrogation. Recent literature has identified clinicopathological association based on textural features deriving from gray-level co-occurrence matrices (GLCM) which facilitate evaluations of gray-level spatial dependence within a delineated region of interest. GLCM-derived features, however, tend to contribute highly redundant information. Moreover, when reporting selected feature sets, investigators often fail to adjust for multiplicities and commonly fail to convey the predictive power of their findings. This article presents a Bayesian probabilistic modeling framework for the GLCM as a multivariate object as well as describes its application within a cancer detection context based on computed tomography. The methodology, which circumvents processing steps and avoids evaluations of reductive and highly correlated feature sets, uses latent Gaussian Markov random field structure to characterize spatial dependencies among GLCM cells and facilitates classification via predictive probability. Correctly predicting the underlying pathology of 81% of the adrenal lesions in our case study, the proposed method outperformed current practices which achieved a maximum accuracy of only 59%. Simulations and theory are presented to further elucidate this comparison as well as ascertain the utility of applying multivariate Gaussian spatial processes to GLCM objects.

Journal ArticleDOI
TL;DR: In this paper, a hierarchical extension of the background and foreground models is proposed to better incorporate a priori knowledge about the disparity between foreground and background and refine their joint estimates under the alternating direction multipliers method.
Abstract: In conventional wisdom of video modeling, the background is often treated as the primary target and foreground is derived using the technique of background subtraction. Based on the observation that foreground and background are two sides of the same coin, we propose to treat them as peer unknown variables and formulate a joint estimation problem, called Hierarchical modeling and Alternating Optimization (HMAO). The motivation behind our hierarchical extensions of background and foreground models is to better incorporate a priori knowledge about the disparity between background and foreground. For background, we decompose it into temporally low-frequency and high-frequency components for the purpose of better characterizing the class of video with dynamic background; for foreground, we construct a Markov random field prior at a spatially low resolution as the pivot to facilitate the noise-resilient refinement at higher resolutions. Built on hierarchical extensions of both models, we show how to successively refine their joint estimates under a unified framework known as alternating direction multipliers method. Experimental results have shown that our approach produces more discriminative background and demonstrates better robustness to noise than other competing methods. When compared against current state-of-the-art techniques, HMAO achieves at least comparable and often superior performance in terms of F-measure scores, especially for video containing dynamic and complex background.

Proceedings ArticleDOI
23 Jun 2019
TL;DR: In this article, a greedy algorithm based on influence maximization was proposed to learn restricted Boltzmann machines with bounded degree, which can simulate any bounded order MRF, and is shown to be as hard as sparse parity with noise.
Abstract: Graphical models are a rich language for describing high-dimensional distributions in terms of their dependence structure. While there are algorithms with provable guarantees for learning undirected graphical models in a variety of settings, there has been much less progress in the important scenario when there are latent variables. Here we study Restricted Boltzmann Machines (or RBMs), which are a popular model with wide-ranging applications in dimensionality reduction, collaborative filtering, topic modeling, feature extraction and deep learning. The main message of our paper is a strong dichotomy in the feasibility of learning RBMs, depending on the nature of the interactions between variables: ferromagnetic models can be learned efficiently, while general models cannot. In particular, we give a simple greedy algorithm based on influence maximization to learn ferromagnetic RBMs with bounded degree. In fact, we learn a description of the distribution on the observed variables as a Markov Random Field. Our analysis is based on tools from mathematical physics that were developed to show the concavity of magnetization. Our algorithm extends straighforwardly to general ferromagnetic Ising models with latent variables. Conversely, we show that even for a contant number of latent variables with constant degree, without ferromagneticity the problem is as hard as sparse parity with noise. This hardness result is based on a sharp and surprising characterization of the representational power of bounded degree RBMs: the distribution on their observed variables can simulate any bounded order MRF. This result is of independent interest since RBMs are the building blocks of deep belief networks.

Journal ArticleDOI
TL;DR: A new supervised classification algorithm which simultaneously considers spectral and spatial information of a hyperspectral image (HSI) is proposed, verified to obtain better performance compared with other state-of-the-art methods.
Abstract: In this paper, a new supervised classification algorithm which simultaneously considers spectral and spatial information of a hyperspectral image (HSI) is proposed. Since HSI always contains complex noise (such as mixture of Gaussian and sparse noise), the quality of the extracted feature inclines to be decreased. To tackle this issue, we utilize the low-rank property of local three-dimensional, patch and adopt complex noise strategy to model the noise embedded in each local patch. Specifically, we firstly use the mixture of Gaussian (MoG) based low-rank matrix factorization (LRMF) method to simultaneously extract the feature and remove noise from each local matrix unfolded from the local patch. Then, a classification map is obtained by applying some classifier to the extracted low-rank feature. Finally, the classification map is processed by Markov random field (MRF) in order to further utilize the smoothness property of the labels. To ease experimental comparison for different HSI classification methods, we built an open package to make the comparison fairly and efficiently. By using this package, the proposed classification method is verified to obtain better performance compared with other state-of-the-art methods.

Journal ArticleDOI
TL;DR: This paper proposes a vision-based fire and smoke segmentation system which uses spatial, temporal and motion information to extract the desired regions from the video frames and achieves a frame-wise fire detection rate of 95.39%.
Abstract: This paper proposes a vision-based fire and smoke segmentation system which uses spatial, temporal and motion information to extract the desired regions from the video frames. The fusion of information is done using multiple features such as optical flow, divergence and intensity values. These features extracted from the images are used to segment the pixels into different classes in an unsupervised way. A comparative analysis is done by using multiple clustering algorithms for segmentation. Here the Markov Random Field performs more accurately than other segmentation algorithms since it characterizes the spatial interactions of pixels using a finite number of parameters. It builds a probabilistic image model that selects the most likely labeling using the maximum a posteriori (MAP) estimation. This unsupervised approach is tested on various images and achieves a frame-wise fire detection rate of 95.39%. Hence this method can be used for early detection of fire in real-time and it can be incorporated into an indoor or outdoor surveillance system.

Proceedings ArticleDOI
01 Jan 2019
TL;DR: Zhang et al. as discussed by the authors formulated fake news detection as an inference problem in a Markov random field (MRF) which can be solved by the iterative mean-field algorithm.
Abstract: Deep-learning-based models have been successfully applied to the problem of detecting fake news on social media. While the correlations among news articles have been shown to be effective cues for online news analysis, existing deep-learning-based methods often ignore this information and only consider each news article individually. To overcome this limitation, we develop a graph-theoretic method that inherits the power of deep learning while at the same time utilizing the correlations among the articles. We formulate fake news detection as an inference problem in a Markov random field (MRF) which can be solved by the iterative mean-field algorithm. We then unfold the mean-field algorithm into hidden layers that are composed of common neural network operations. By integrating these hidden layers on top of a deep network, which produces the MRF potentials, we obtain our deep MRF model for fake news detection. Experimental results on well-known datasets show that the proposed model improves upon various state-of-the-art models.

Journal ArticleDOI
Le Zhao1, Xianpei Wang1, Hongtai Yao1, Meng Tian1, Zini Jian1 
TL;DR: A novel object-based Markov random field with anisotropic weighted penalty (OMRF-AWP) method is proposed, which defines a new neighborhood system based on the irregular graph model and builds a new potential function by considering the region angle information.
Abstract: The extraction of power line plays a key role in power line inspection by Unmanned Aerial Vehicles (UAVs). While it is challenging to extract power lines in aerial images because of the weak targets and the complex background. In this paper, a novel power line extraction method is proposed. First of all, we create a line segment candidate pool which contains power line segments and large amount of other line segments. Secondly, we construct the irregular graph model with these line segments as nodes. Then a novel object-based Markov random field with anisotropic weighted penalty (OMRF-AWP) method is proposed. It defines a new neighborhood system based on the irregular graph model and builds a new potential function by considering the region angle information. With the OMRF-AWP method, we can distinguish between the power line segments and other line segments. Finally, an envelope-based piecewise fitting (EPF) method is proposed to fit the power lines. Experimental results show that the proposed method has good performance in multiple scenes with complex background.

Journal ArticleDOI
TL;DR: A novel Bayesian hierarchical model is proposed, which incorporates a hidden Potts model to project the irregularly distributed cells to a square lattice and a Markov random field prior model to identify regions in a heterogeneous pathology image, and shows that the interaction strength between tumor and stromal cells predicts patient prognosis.
Abstract: Digital pathology imaging of tumor tissues, which captures histological details in high resolution, is fast becoming a routine clinical procedure. Recent developments in deep-learning methods have enabled the identification, characterization, and classification of individual cells from pathology images analysis at a large scale. This creates new opportunities to study the spatial patterns of and interactions among different types of cells. Reliable statistical approaches to modeling such spatial patterns and interactions can provide insight into tumor progression and shed light on the biological mechanisms of cancer. In this article, we consider the problem of modeling a pathology image with irregular locations of three different types of cells: lymphocyte, stromal, and tumor cells. We propose a novel Bayesian hierarchical model, which incorporates a hidden Potts model to project the irregularly distributed cells to a square lattice and a Markov random field prior model to identify regions in a heterogeneous pathology image. The model allows us to quantify the interactions between different types of cells, some of which are clinically meaningful. We use Markov chain Monte Carlo sampling techniques, combined with a double Metropolis-Hastings algorithm, in order to simulate samples approximately from a distribution with an intractable normalizing constant. The proposed model was applied to the pathology images of $205$ lung cancer patients from the National Lung Screening trial, and the results show that the interaction strength between tumor and stromal cells predicts patient prognosis (P = $0.005$). This statistical methodology provides a new perspective for understanding the role of cell-cell interactions in cancer progression.

Journal ArticleDOI
TL;DR: A novel method for sidescan sonar image segmentation based on MRF and ELM is proposed, which performs better than other machine learning methods such as ELM, kernel-based extreme learning machine, SVM, and convolutional neural networks.
Abstract: As a widely used segmentation scheme, Markov random field (MRF) utilizes $k$ -means clustering to calculate the initial model for sidescan sonar image segmentation. However, for the noise and intensity inhomogeneity nature of the sidescan sonar images, the segmentation results of $k$ -means clustering have low accuracy, motivating us to use machine learning methods to initialize MRF. Meanwhile, an extreme learning machine (ELM), a supervised learning algorithm derived from the single-hidden-layer feedforward neural networks, learns faster than randomly generated hidden-layer parameters and is superior to a support vector machine (SVM). Therefore, in this paper, we proposed a novel method for sidescan sonar image segmentation based on MRF and ELM. The proposed method segments sidescan sonar images in object-highlight, object-shadow, and sea-bottom reverberation areas. Specifically, we intend to use an ELM to get an initial model for MRF. Moreover, to improve the stability of an ELM, a simple ensemble ELM (SE-ELM) based on an ensemble algorithm is utilized to obtain the prediction model. In an SE-ELM, we use an ensemble of ELMs and majority votes to determine the prediction of testing data sets. Then, the classification results of the SE-ELM are utilized to initialize MRF, termed as SE-ELM-MRF. With features consisting of pixels of small image patches, our experiments on real sonar data indicate that the SE-ELM performs better than other machine learning methods such as ELM, kernel-based extreme learning machine, SVM, and convolutional neural networks. Moreover, using SE-ELM as the initial method in the proposed SE-ELM-MRF, the segmentation results are smoother and the segmentation process converges faster than the traditional MRF.

Journal ArticleDOI
TL;DR: In this paper, the authors describe how scene depth can be extracted using a hyperspectral light field capture (H-LF) system consisting of a $5 \times 6$ 5 × 6 array of cameras, with each camera sampling a different narrow band in the visible spectrum.
Abstract: In this paper, we describe how scene depth can be extracted using a hyperspectral light field capture (H-LF) system. Our H-LF system consists of a $5 \times 6$ 5 × 6 array of cameras, with each camera sampling a different narrow band in the visible spectrum. There are two parts to extracting scene depth. The first part is our novel cross-spectral pairwise matching technique, which involves a new spectral-invariant feature descriptor and its companion matching metric we call bidirectional weighted normalized cross correlation (BWNCC). The second part, namely, H-LF stereo matching, uses a combination of spectral-dependent correspondence and defocus cues. These two new cost terms are integrated into a Markov Random Field (MRF) for disparity estimation. Experiments on synthetic and real H-LF data show that our approach can produce high-quality disparity maps. We also show that these results can be used to produce the complete plenoptic cube in addition to synthesizing all-focus and defocused color images under different sensor spectral responses.

Journal ArticleDOI
TL;DR: An automatic method to localize and identify vertebrae by combining deep stacked sparse autoencoder (SSAE) contextual features and structured regression forest (SRF) to achieve higher localization accuracy, low model complexity, and no need for any assumptions about visual field in CT scans is proposed.
Abstract: Automatic vertebrae localization and identification in medical computed tomography (CT) scans is of great value for computer-aided spine diseases diagnosis. In order to overcome the disadvantages of the approaches employing hand-crafted, low-level features and based on field-of-view priori assumption of spine structure, an automatic method is proposed to localize and identify vertebrae by combining deep stacked sparse autoencoder (SSAE) contextual features and structured regression forest (SRF). The method employs SSAE to learn image deep contextual features instead of hand-crafted ones by building larger-range input samples to improve their contextual discrimination ability. In the localization and identification stage, it incorporates the SRF model to achieve whole spine localization, then screens those vertebrae within the image, thus relieves the assumption that the part of spine in the field of image is visible. In the end, the output distribution of SRF and spine CT scans properties are assembled to develop a two-stage progressive refining strategy, where the mean-shift kernel density estimation and Otsu method instead of Markov random field (MRF) are adopted to reduce model complexity and refine vertebrae localization results. Extensive evaluation was performed on a challenging data set of 98 spine CT scans. Compared with the hidden Markov model and the method based on convolutional neural network (CNN), the proposed approach could effectively and automatically locate and identify spinal targets in CT scans, and achieve higher localization accuracy, low model complexity, and no need for any assumptions about visual field in CT scans.

Journal ArticleDOI
TL;DR: A turbo message passing framework is developed by incorporating the MRF model into the bilinear generalized approximate message passing (BiGAMP) algorithm and it is shown that the proposed scheme considerably outperforms the schemes without exploiting the clustered sparsity of the channel.
Abstract: In this letter, we propose a blind multiuser detection algorithm for massive MIMO systems equipped with a uniform planar antenna array. We model the channel in the angular domain as a sparse matrix with the non-zero coefficients forming clusters. The channel sparsity is described by a binary Markov random field (MRF). Then, we develop a turbo message passing framework by incorporating the MRF model into the bilinear generalized approximate message passing (BiGAMP) algorithm. We show that the proposed scheme considerably outperforms the schemes without exploiting the clustered sparsity of the channel.

Journal ArticleDOI
TL;DR: The results showed that the PolSAR features extracted from Compact Polarimetric Synthetic Aperture Radar data can provide an acceptable overall accuracy in segmentation when compared to the full polarimetry and Dual Polarimetry data.
Abstract: As the first major step in each object-oriented feature extraction approach, segmentation plays an essential role as a preliminary step towards further and higher levels of image processing. The pr...

Journal ArticleDOI
TL;DR: The presented results demonstrated the feasibility of automated segmentation of the respiratory sliding motion interface in liver MR images and the effectiveness of using the derived motion masks to preserve motion discontinuity.
Abstract: Objective: The sliding motion of the liver during respiration violates the homogeneous motion smoothness assumption in conventional nonrigid image registration and commonly results in compromised registration accuracy. This paper presents a novel approach, registration with three-dimensional (3D) active contour motion segmentation (RAMS), to improve registration accuracy with discontinuity-aware motion regularization. Methods: A Markov random field-based discrete optimization with dense displacement sampling and self-similarity context metric is used for registration, while a graph cuts-based 3D active contour approach is applied to segment the sliding interface. In the first registration pass, a mask-free L1 regularization on an image-derived minimum spanning tree is performed to allow motion discontinuity. Based on the motion field estimates, a coarse segmentation finds the motion boundaries. Next, based on magnetic resonance (MR) signal intensity, a fine segmentation aligns the motion boundaries with anatomical boundaries. In the second registration pass, smoothness constraints across the segmented sliding interface are removed by masked regularization on a minimum spanning forest and masked interpolation of the motion field. Results: For in vivo breath-hold abdominal MRI data, the motion masks calculated by RAMS are highly consistent with manual segmentations in terms of Dice similarity and bidirectional local distance measure. These automatically obtained masks are shown to substantially improve registration accuracy for both the proposed discrete registration as well as conventional continuous nonrigid algorithms. Conclusion/Significance: The presented results demonstrated the feasibility of automated segmentation of the respiratory sliding motion interface in liver MR images and the effectiveness of using the derived motion masks to preserve motion discontinuity.

Journal ArticleDOI
TL;DR: The experimental results on real HSI data indicate that the proposed method outperforms compared MRF-based and other spectral–spatial approaches in terms of classification accuracies and region uniformity.
Abstract: This paper presents a novel Markov random field (MRF) method integrating adaptive interclass-pair penalty (aICP2) and spectral similarity information (SSI) for hyper-spectral image (HSI) classification. aICP2 structurally combines $K(K - 1)/2$ (K is the number of classes) classical “Potts model” with $K(K - 1)/2$ interaction coefficients. aICP2 tries a new way to solve the key problems, insufficient correction within homogeneous regions, and over-smoothness at class boundaries, in MRF-based HSI classification. It is assumed that different class pairs should be assigned with various degrees of penalties in MRF smoothness process, according to pairwise class separability and spatial class confusion in raw classification map. The Fisher ratio is modified to measure pairwise class separability with a training set. And, gray level co-occurrence matrix is used to measure spatial class confusion degree. Then, aICP2 is constructed by combining Fisher ratio and GCLM. aICP2 applies larger penalty on class pairs that confuse with each other seriously to provide sufficient smoothness, and vice versa. In addition, to protect class edges and details, SSI is introduced to make the penalty of related neighboring pixels small. aICP2ssi denotes the integration of aICP2 and SSI. The further improved method is both interclass-pair and interpixel adaptive in spatial term. A graph-cut-based $\alpha - \beta $ -swap method is introduced to optimize the proposed energy function. The experimental results on real HSI data indicate that the proposed method outperforms compared MRF-based and other spectral–spatial approaches in terms of classification accuracies and region uniformity.