scispace - formally typeset
Search or ask a question

Showing papers on "Pipeline (computing) published in 2019"


Proceedings ArticleDOI
15 Jun 2019
TL;DR: benchmarks suggest that PointPillars is an appropriate encoding for object detection in point clouds, and proposes a lean downstream network.
Abstract: Object detection in point clouds is an important aspect of many robotics applications such as autonomous driving. In this paper, we consider the problem of encoding a point cloud into a format appropriate for a downstream detection pipeline. Recent literature suggests two types of encoders; fixed encoders tend to be fast but sacrifice accuracy, while encoders that are learned from data are more accurate, but slower. In this work, we propose PointPillars, a novel encoder which utilizes PointNets to learn a representation of point clouds organized in vertical columns (pillars). While the encoded features can be used with any standard 2D convolutional detection architecture, we further propose a lean downstream network. Extensive experimentation shows that PointPillars outperforms previous encoders with respect to both speed and accuracy by a large margin. Despite only using lidar, our full detection pipeline significantly outperforms the state of the art, even among fusion methods, with respect to both the 3D and bird’s eye view KITTI benchmarks. This detection performance is achieved while running at 62 Hz: a 2 - 4 fold runtime improvement. A faster version of our method matches the state of the art at 105 Hz. These benchmarks suggest that PointPillars is an appropriate encoding for object detection in point clouds.

1,298 citations


Journal ArticleDOI
TL;DR: A novel deep learning framework to achieve highly accurate machine fault diagnosis using transfer learning to enable and accelerate the training of deep neural network is developed.
Abstract: We develop a novel deep learning framework to achieve highly accurate machine fault diagnosis using transfer learning to enable and accelerate the training of deep neural network. Compared with existing methods, the proposed method is faster to train and more accurate. First, original sensor data are converted to images by conducting a Wavelet transformation to obtain time-frequency distributions. Next, a pretrained network is used to extract lower level features. The labeled time-frequency images are then used to fine-tune the higher levels of the neural network architecture. This paper creates a machine fault diagnosis pipeline and experiments are carried out to verify the effectiveness and generalization of the pipeline on three main mechanical datasets including induction motors, gearboxes, and bearings with sizes of 6000, 9000, and 5000 time series samples, respectively. We achieve state-of-the-art results on each dataset, with most datasets showing test accuracy near 100%, and in the gearbox dataset, we achieve significant improvement from 94.8% to 99.64%. We created a repository including these datasets located at mlmechanics.ics.uci.edu.

721 citations


Proceedings ArticleDOI
27 Oct 2019
TL;DR: PipeDream is presented, a system that adds inter-batch pipelining to intra-batch parallelism to further improve parallel training throughput, helping to better overlap computation with communication and reduce the amount of communication when possible.
Abstract: DNN training is extremely time-consuming, necessitating efficient multi-accelerator parallelization. Current approaches to parallelizing training primarily use intra-batch parallelization, where a single iteration of training is split over the available workers, but suffer from diminishing returns at higher worker counts. We present PipeDream, a system that adds inter-batch pipelining to intra-batch parallelism to further improve parallel training throughput, helping to better overlap computation with communication and reduce the amount of communication when possible. Unlike traditional pipelining, DNN training is bi-directional, where a forward pass through the computation graph is followed by a backward pass that uses state and intermediate data computed during the forward pass. Naive pipelining can thus result in mismatches in state versions used in the forward and backward passes, or excessive pipeline flushes and lower hardware efficiency. To address these challenges, PipeDream versions model parameters for numerically correct gradient computations, and schedules forward and backward passes of different minibatches concurrently on different workers with minimal pipeline stalls. PipeDream also automatically partitions DNN layers among workers to balance work and minimize communication. Extensive experimentation with a range of DNN tasks, models, and hardware configurations shows that PipeDream trains models to high accuracy up to 5.3X faster than commonly used intra-batch parallelism techniques.

493 citations


Proceedings Article
01 Jan 2019
TL;DR: TensorPipe as mentioned in this paper is a pipeline parallelism library that allows scaling any network that can be expressed as a sequence of layers by pipelining different sub-sequences of layers on separate accelerators.
Abstract: Scaling up deep neural network capacity has been known as an effective approach to improving model quality for several different machine learning tasks. In many cases, increasing model capacity beyond the memory limit of a single accelerator has required developing special algorithms or infrastructure. These solutions are often architecture-specific and do not transfer to other machine learning tasks. To address the need for efficient and task-independent model parallelism, we introduce TensorPipe, a pipeline parallelism library that allows scaling any network that can be expressed as a sequence of layers. By pipelining different sub-sequences of layers on separate accelerators, TensorPipe provides the flexibility of scaling a variety of different networks to gigantic sizes efficiently. Moreover, TensorPipe utilizes a novel batch-splitting pipelining algorithm, resulting in almost linear speedup when a model is partitioned across multiple accelerators. We demonstrate the advantages of TensorPipe by training large-scale neural networks on two different tasks with distinct network architectures: (i)Image Classification: We train a 557-million-parameter AmoebaNet model and attain a top-1 accuracy of 84.4% on ImageNet-2012, (ii)Multilingual Neural Machine Translation: We train a single 6-billion-parameter, 128-layer Transformer model on a corpus spanning over 100 languages and achieve better quality than all bilingual models.

486 citations


Journal ArticleDOI
TL;DR: The estimation error shows that the presented algorithm is comparable to the minimum mean square error (MMSE) with full knowledge of the channel statistics, and it is better than an approximation to linear MMSE.
Abstract: In this letter, we present a deep learning algorithm for channel estimation in communication systems. We consider the time–frequency response of a fast fading communication channel as a 2D image. The aim is to find the unknown values of the channel response using some known values at the pilot locations. To this end, a general pipeline using deep image processing techniques, image super-resolution (SR), and image restoration (IR) is proposed. This scheme considers the pilot values, altogether, as a low-resolution image and uses an SR network cascaded with a denoising IR network to estimate the channel. Moreover, the implementation of the proposed pipeline is presented. The estimation error shows that the presented algorithm is comparable to the minimum mean square error (MMSE) with full knowledge of the channel statistics, and it is better than an approximation to linear MMSE. The results confirm that this pipeline can be used efficiently in channel estimation.

373 citations


Proceedings ArticleDOI
05 Aug 2019
TL;DR: This work designs a system that enables high accuracy object detection for commodity AR/MR system running at 60fps, employs low latency offloading techniques, decouples the rendering pipeline from the offloading pipeline, and uses a fast object tracking method to maintain detection accuracy.
Abstract: Most existing Augmented Reality (AR) and Mixed Reality (MR) systems are able to understand the 3D geometry of the surroundings but lack the ability to detect and classify complex objects in the real world. Such capabilities can be enabled with deep Convolutional Neural Networks (CNN), but it remains difficult to execute large networks on mobile devices. Offloading object detection to the edge or cloud is also very challenging due to the stringent requirements on high detection accuracy and low end-to-end latency. The long latency of existing offloading techniques can significantly reduce the detection accuracy due to changes in the user's view. To address the problem, we design a system that enables high accuracy object detection for commodity AR/MR system running at 60fps. The system employs low latency offloading techniques, decouples the rendering pipeline from the offloading pipeline, and uses a fast object tracking method to maintain detection accuracy. The result shows that the system can improve the detection accuracy by 20.2%-34.8% for the object detection and human keypoint detection tasks, and only requires 2.24ms latency for object tracking on the AR device. Thus, the system leaves more time and computational resources to render virtual elements for the next frame and enables higher quality AR/MR experiences.

371 citations


Proceedings ArticleDOI
15 Jun 2019
TL;DR: In this paper, the authors propose a technique to "unprocess" images by inverting each step of an image processing pipeline, thereby allowing them to synthesize realistic raw sensor measurements from commonly available Internet photos.
Abstract: Machine learning techniques work best when the data used for training resembles the data used for evaluation. This holds true for learned single-image denoising algorithms, which are applied to real raw camera sensor readings but, due to practical constraints, are often trained on synthetic image data. Though it is understood that generalizing from synthetic to real images requires careful consideration of the noise properties of camera sensors, the other aspects of an image processing pipeline (such as gain, color correction, and tone mapping) are often overlooked, despite their significant effect on how raw measurements are transformed into finished images. To address this, we present a technique to “unprocess” images by inverting each step of an image processing pipeline, thereby allowing us to synthesize realistic raw sensor measurements from commonly available Internet photos. We additionally model the relevant components of an image processing pipeline when evaluating our loss function, which allows training to be aware of all relevant photometric processing that will occur after denoising. By unprocessing and processing training data and model outputs in this way, we are able to train a simple convolutional neural network that has 14%-38% lower error rates and is 9×-18× faster than the previous state of the art on the Darmstadt Noise Dataset, and generalizes to sensors outside of that dataset as well.

369 citations


Journal ArticleDOI
TL;DR: DeepISP as mentioned in this paper is a full end-to-end deep neural model of the camera image signal processing pipeline that learns a mapping from the raw low-light mosaiced image to the final visually compelling image.
Abstract: We present DeepISP, a full end-to-end deep neural model of the camera image signal processing pipeline. Our model learns a mapping from the raw low-light mosaiced image to the final visually compelling image and encompasses low-level tasks, such as demosaicing and denoising, as well as higher-level tasks, such as color correction and image adjustment. The training and evaluation of the pipeline were performed on a dedicated data set containing pairs of low-light and well-lit images captured by a Samsung S7 smartphone camera in both raw and processed JPEG formats. The proposed solution achieves the state-of-the-art performance in objective evaluation of peak signal-to-noise ratio on the subtask of joint denoising and demosaicing. For the full end-to-end pipeline, it achieves better visual quality compared to the manufacturer ISP, in both a subjective human assessment and when rated by a deep model trained for assessing image quality.

204 citations


Journal ArticleDOI
TL;DR: A multi-task deep learning approach with atlas propagation is combined to develop a shape-refined bi-ventricular segmentation pipeline for short-axis CMR volumetric images that is robust and capable of producing accurate, high-resolution, and anatomically smooth bi- ventricular 3D models, despite the presence of artifacts in input CMR volumes.
Abstract: Deep learning approaches have achieved state-of-the-art performance in cardiac magnetic resonance (CMR) image segmentation. However, most approaches have focused on learning image intensity features for segmentation, whereas the incorporation of anatomical shape priors has received less attention. In this paper, we combine a multi-task deep learning approach with atlas propagation to develop a shape-refined bi-ventricular segmentation pipeline for short-axis CMR volumetric images. The pipeline first employs a fully convolutional network (FCN) that learns segmentation and landmark localization tasks simultaneously. The architecture of the proposed FCN uses a 2.5D representation, thus combining the computational advantage of 2D FCNs networks and the capability of addressing 3D spatial consistency without compromising segmentation accuracy. Moreover, a refinement step is designed to explicitly impose shape prior knowledge and improve segmentation quality. This step is effective for overcoming image artifacts (e.g., due to different breath-hold positions and large slice thickness), which preclude the creation of anatomically meaningful 3D cardiac shapes. The pipeline is fully automated, due to network’s ability to infer landmarks, which are then used downstream in the pipeline to initialize atlas propagation. We validate the pipeline on 1831 healthy subjects and 649 subjects with pulmonary hypertension. Extensive numerical experiments on the two datasets demonstrate that our proposed method is robust and capable of producing accurate, high-resolution, and anatomically smooth bi-ventricular 3D models, despite the presence of artifacts in input CMR volumes.

165 citations


Journal ArticleDOI
TL;DR: An integrated pipeline for index sorting and massively parallel single-cell RNA sequencing (MARS-seq2.0) is developed that builds on the previously published MARS-seq approach and allows identification of unique cell types across different tissues and diseases, as well as unique model systems and organisms.
Abstract: Human tissues comprise trillions of cells that populate a complex space of molecular phenotypes and functions and that vary in abundance by 4-9 orders of magnitude. Relying solely on unbiased sampling to characterize cellular niches becomes infeasible, as the marginal utility of collecting more cells diminishes quickly. Furthermore, in many clinical samples, the relevant cell types are scarce and efficient processing is critical. We developed an integrated pipeline for index sorting and massively parallel single-cell RNA sequencing (MARS-seq2.0) that builds on our previously published MARS-seq approach. MARS-seq2.0 is based on >1 million cells sequenced with this pipeline and allows identification of unique cell types across different tissues and diseases, as well as unique model systems and organisms. Here, we present a detailed step-by-step procedure for applying the method. In the improved procedure, we combine sub-microliter reaction volumes, optimization of enzymatic mixtures and an enhanced analytical pipeline to substantially lower the cost, improve reproducibility and reduce well-to-well contamination. Data analysis combines multiple layers of quality assessment and error detection and correction, graphically presenting key statistics for library complexity, noise distribution and sequencing saturation. Importantly, our combined FACS and single-cell RNA sequencing (scRNA-seq) workflow enables intuitive approaches for depletion or enrichment of cell populations in a data-driven manner that is essential to efficient sampling of complex tissues. The experimental protocol, from cell sorting to a ready-to-sequence library, takes 2-3 d. Sequencing and processing the data through the analytical pipeline take another 1-2 d.

156 citations


Journal ArticleDOI
TL;DR: The pipeline incorporates different techniques and makes independent implementation choices in all its stages including the search design, the method to construct template banks, the automatic routines to detect bad data segments and to insulate good data from them, and the procedure to account for the non-stationary nature of the detector noise.
Abstract: In this paper, we report on the construction of a new and independent pipeline for analyzing the public data from the first observing run of Advanced LIGO for mergers of compact binary systems. The pipeline incorporates different techniques and makes independent implementation choices in all its stages including the search design, the method to construct template banks, the automatic routines to detect bad data segments (``glitches'') and to insulate good data from them, the procedure to account for the nonstationary nature of the detector noise, the signal-quality vetoes at the single-detector level and the methods to combine results from multiple detectors. Our pipeline enabled us to identify a new binary black hole merger GW151216 in the public LIGO data. This paper serves as a bird's eye view of the pipeline's important stages. Full details and derivations underlying the various stages will appear in accompanying papers.

Posted Content
TL;DR: A Meta Neural Trojan Detection pipeline that does not make assumptions on the attack strategies and only needs black-box access to models is introduced and achieves 97% detection AUC score and significantly outperforms existing detection approaches.
Abstract: In machine learning Trojan attacks, an adversary trains a corrupted model that obtains good performance on normal data but behaves maliciously on data samples with certain trigger patterns. Several approaches have been proposed to detect such attacks, but they make undesirable assumptions about the attack strategies or require direct access to the trained models, which restricts their utility in practice. This paper addresses these challenges by introducing a Meta Neural Trojan Detection (MNTD) pipeline that does not make assumptions on the attack strategies and only needs black-box access to models. The strategy is to train a meta-classifier that predicts whether a given target model is Trojaned. To train the meta-model without knowledge of the attack strategy, we introduce a technique called jumbo learning that samples a set of Trojaned models following a general distribution. We then dynamically optimize a query set together with the meta-classifier to distinguish between Trojaned and benign models. We evaluate MNTD with experiments on vision, speech, tabular data and natural language text datasets, and against different Trojan attacks such as data poisoning attack, model manipulation attack, and latent attack. We show that MNTD achieves 97% detection AUC score and significantly outperforms existing detection approaches. In addition, MNTD generalizes well and achieves high detection performance against unforeseen attacks. We also propose a robust MNTD pipeline which achieves 90% detection AUC even when the attacker aims to evade the detection with full knowledge of the system.

Proceedings ArticleDOI
Mingyu Gao1, Xuan Yang1, Jing Pu1, Mark Horowitz1, Christos Kozyrakis1 
04 Apr 2019
TL;DR: This work proposes dataflow optimizations to address the shortcomings of existing parallel dataflow techniques for tiled NN accelerators, and develops buffer sharing dataflow that turns the distributed buffers into an idealized shared buffer, eliminating excessive data duplication and the memory access overheads.
Abstract: The use of increasingly larger and more complex neural networks (NNs) makes it critical to scale the capabilities and efficiency of NN accelerators. Tiled architectures provide an intuitive scaling solution that supports both coarse-grained parallelism in NNs: intra-layer parallelism, where all tiles process a single layer, and inter-layer pipelining, where multiple layers execute across tiles in a pipelined manner. This work proposes dataflow optimizations to address the shortcomings of existing parallel dataflow techniques for tiled NN accelerators. For intra-layer parallelism, we develop buffer sharing dataflow that turns the distributed buffers into an idealized shared buffer, eliminating excessive data duplication and the memory access overheads. For inter-layer pipelining, we develop alternate layer loop ordering that forwards the intermediate data in a more fine-grained and timely manner, reducing the buffer requirements and pipeline delays. We also make inter-layer pipelining applicable to NNs with complex DAG structures. These optimizations improve the performance of tiled NN accelerators by 2x and reduce their energy consumption by 45% across a wide range of NNs. The effectiveness of our optimizations also increases with the NN size and complexity.

Journal ArticleDOI
22 Aug 2019-Proteins
TL;DR: Detailed data analyses showed that the strength of the pipeline is due to the sensitive MSA construction and the advanced strategies for coevolutionary feature ensembling, and domain splitting was also found to help enhance the contact prediction performance.
Abstract: We report the results of residue-residue contact prediction of a new pipeline built purely on the learning of coevolutionary features in the CASP13 experiment. For a query sequence, the pipeline starts with the collection of multiple sequence alignments (MSAs) from multiple genome and metagenome sequence databases using two complementary Hidden Markov Model (HMM)-based searching tools. Three profile matrices, built on covariance, precision, and pseudolikelihood maximization respectively, are then created from the MSAs, which are used as the input features of a deep residual convolutional neural network architecture for contact-map training and prediction. Two ensembling strategies have been proposed to integrate the matrix features through end-to-end training and stacking, resulting in two complementary programs called TripletRes and ResTriplet, respectively. For the 31 free-modeling domains that do not have homologous templates in the PDB, TripletRes and ResTriplet generated comparable results with an average accuracy of 0.640 and 0.646, respectively, for the top L/5 long-range predictions, where 71% and 74% of the cases have an accuracy above 0.5. Detailed data analyses showed that the strength of the pipeline is due to the sensitive MSA construction and the advanced strategies for coevolutionary feature ensembling. Domain splitting was also found to help enhance the contact prediction performance. Nevertheless, contact models for tail regions, which often involve a high number of alignment gaps, and for targets with few homologous sequences are still suboptimal. Development of new approaches where the model is specifically trained on these regions and targets might help address these problems.

Journal ArticleDOI
TL;DR: A combined operation optimization model of the air cooler and compressor through the optimization of the switching scheme of compressors and air coolers is established, which can greatly reduce the production energy consumption of the pipeline system.
Abstract: Based on the mutual coupling effect among the compressor, the air cooler and pipes in the system of natural gas pipeline, innovatively with the goal of minimum energy consumption, this paper established a combined operation optimization model of the air cooler and compressor through the optimization of the switching scheme of compressors and air coolers, which can greatly reduce the production energy consumption of the pipeline system. Moreover, when the air temperature is taken as an optimization variable, the most proper temperature to start the air cooler of each compressor station can be worked out to guide the optimized operation of the pipeline, which is of high value for promotion and application. The case analysis of west-east natural gas pipeline II showed that among genetic algorithm (GA), particle swarm optimization (PSO), and simulated annealing (SA) algorithm that are used to solve the optimization model, the genetic algorithm is the fastest, and the simulated annealing algorithm the slowest, but the optimization results of the simulated annealing algorithm is the best, in which the reduced production energy consumption accounted for 33.77%, testifying the practicability and creativity of the optimization model.

Proceedings ArticleDOI
01 Oct 2019
TL;DR: This work presents a deep-learning architecture for multi-view stereo with conditional random fields (MVSCRF), and achieves comparable results with state-of-the-art learning based methods on outdoor Tanks and Temples dataset without fine-tuning, which demonstrates the method’s generalization ability.
Abstract: We present a deep-learning architecture for multi-view stereo with conditional random fields (MVSCRF). Given an arbitrary number of input images, we first use a U-shape neural network to extract deep features incorporating both global and local information, and then build a 3D cost volume for the reference camera. Unlike previous learning based methods, we explicitly constraint the smoothness of depth maps by using conditional random fields (CRFs) after the stage of cost volume regularization. The CRFs module is implemented as recurrent neural networks so that the whole pipeline can be trained end-to-end. Our results show that the proposed pipeline outperforms previous state-of-the-arts on large-scale DTU dataset. We also achieve comparable results with state-of-the-art learning based methods on outdoor Tanks and Temples dataset without fine-tuning, which demonstrates our method’s generalization ability.

Journal ArticleDOI
TL;DR: The results found that the total oil transportation cost and the total exergy loss of the pipeline transportation system after optimization have seen an obvious improvement in the effective utilization of energy when compared to pre-optimized levels.

Journal ArticleDOI
Suleyman Mete1
TL;DR: FMEA-based AHP-MOORA integrated approach under Pythagorean fuzzy sets is proposed for assessing occupational risks in a natural gas pipeline construction project and gives more reasonable results.
Abstract: The aim of natural gas pipeline projects is generally to make it possible to transfer the natural gas produced from one place to another. There are various safety risks for these broad construction...

Journal ArticleDOI
TL;DR: This approach not only demonstrates the record 1 million mass resolution for lipid imaging from brain tissue, but explicitly show such mass resolution is required to resolve the complexity of the lipidome.
Abstract: Desorption electrospray ionisation-mass spectrometry imaging (DESI-MSI) is a powerful imaging technique for the analysis of complex surfaces. However, the often highly complex nature of biological samples is particularly challenging for MSI approaches, as options to appropriately address molecular complexity are limited. Fourier transform ion cyclotron resonance mass spectrometry (FT-ICR MS) offers superior mass accuracy and mass resolving power, but its moderate throughput inhibits broader application. Here we demonstrate the dramatic gains in mass resolution and/or throughput of DESI-MSI on an FT-ICR MS by developing and implementing a sophisticated data acquisition and data processing pipeline. The presented pipeline integrates, for the first time, parallel ion accumulation and detection, post-processing absorption mode Fourier transform and pixel-by-pixel internal re-calibration. To achieve that, first, we developed and coupled an external high-performance data acquisition system to an FT-ICR MS instrument to record the time-domain signals (transients) in parallel with the instrument’s built-in electronics. The recorded transients were then processed by the in-house developed computationally-efficient data processing and data analysis software. Importantly, the described pipeline is shown to be applicable even to extremely large, up to 1 TB, imaging datasets. Overall, this approach provides improved analytical figures of merits such as: (i) enhanced mass resolution at no cost in experimental time; and (ii) up to 4-fold higher throughput while maintaining a constant mass resolution. Using this approach, we not only demonstrate the record 1 million mass resolution for lipid imaging from brain tissue, but explicitly show such mass resolution is required to resolve the complexity of the lipidome.

Journal ArticleDOI
D. J. Brout1, Masao Sako1, Daniel Scolnic, Richard Kessler2, C. B. D'Andrea1, Tamara M. Davis3, Samuel Hinton3, Alex G. Kim4, J. Lasker2, Edward Macaulay5, Anais Möller6, Robert C. Nichol5, Mathew Smith7, Mark Sullivan7, R. Wolf8, S. Allam9, Bruce A. Bassett10, Peter de Nully Brown11, Francisco J. Castander12, M. Childress7, Ryan J. Foley13, Lluís Galbany14, Ken Herner9, E. Kasai15, M. March1, Eric Morganson16, Peter Nugent4, Yen-Chen Pan17, Yen-Chen Pan18, R. C. Thomas4, Brad E. Tucker6, William Wester9, T. M. C. Abbott, James Annis9, Santiago Avila5, Emmanuel Bertin19, David Brooks20, D. L. Burke21, D. L. Burke8, A. Carnero Rosell, M. Carrasco Kind16, J. Carretero22, Martin Crocce12, Carlos Cunha8, L. N. da Costa, C. Davis8, J. De Vicente, Shantanu Desai23, H. T. Diehl9, P. Doel20, Tim Eifler24, Tim Eifler25, B. Flaugher9, Pablo Fosalba12, J. Frieman9, Juan Garcia-Bellido26, Enrique Gaztanaga12, D. W. Gerdes27, Daniel A. Goldstein24, Daniel Gruen8, Daniel Gruen21, Robert A. Gruendl16, J. Gschwend, G. Gutierrez9, W. G. Hartley28, W. G. Hartley20, Devon L. Hollowood13, K. Honscheid29, David J. James30, Kyler Kuehn31, N. P. Kuropatkin9, Ofer Lahav20, Tenglin Li9, Marco A. P. Lima32, Jennifer L. Marshall11, Paul Martini29, Ramon Miquel22, Brian Nord9, A. A. Plazas24, A. Roodman8, A. Roodman21, Eli S. Rykoff8, Eli S. Rykoff21, E. J. Sanchez, V. Scarpine9, Rafe Schindler21, M. S. Schubnell27, S. Serrano12, I. Sevilla-Noarbe, Marcelle Soares-Santos33, Flavia Sobreira34, E. Suchyta35, M. E. C. Swanson16, Gregory Tarle27, Daniel Thomas5, Douglas L. Tucker9, Alistair R. Walker, Brian Yanny9, Yanxi Zhang9 
TL;DR: In this article, the authors present griz light curves of 251 SNe Ia from the first 3 years of the DES-SN spectroscopically classified sample, which are used in the cosmological parameter analysis by employing a scene modeling approach that simultaneously models a variable transient flux and temporally constant host galaxy.
Abstract: We present griz light curves of 251 SNe Ia from the first 3 years of the Dark Energy Survey Supernova Program's (DES-SN) spectroscopically classified sample. The photometric pipeline described in this paper produces the calibrated fluxes and associated uncertainties used in the cosmological parameter analysis by employing a scene modeling approach that simultaneously models a variable transient flux and temporally constant host galaxy. We inject artificial point sources onto DECam images to test the accuracy of our photometric method. Upon comparison of input and measured artificial supernova fluxes, we find that flux biases peak at 3 mmag. We require corrections to our photometric uncertainties as a function of host galaxy surface brightness at the transient location, similar to that seen by the DES Difference Imaging Pipeline used to discover transients. The public release of the light curves can be found at https://des.ncsa.illinois.edu/releases/sn.

Journal ArticleDOI
TL;DR: PynPoint as mentioned in this paper is a data-reduction pipeline for processing and analysis of high-contrast imaging data obtained with pupil-stabilized observations, which is particularly suitable for the 3-5 μm wavelength range where typically thousands of frames have to be processed and an accurate subtraction of the thermal background emission is critical.
Abstract: Context. The direct detection and characterization of planetary and substellar companions at small angular separations is a rapidly advancing field. Dedicated high-contrast imaging instruments deliver unprecedented sensitivity, enabling detailed insights into the atmospheres of young low-mass companions. In addition, improvements in data reduction and point spread function (PSF)-subtraction algorithms are equally relevant for maximizing the scientific yield, both from new and archival data sets.Aims. We aim at developing a generic and modular data-reduction pipeline for processing and analysis of high-contrast imaging data obtained with pupil-stabilized observations. The package should be scalable and robust for future implementations and particularly suitable for the 3–5 μ m wavelength range where typically thousands of frames have to be processed and an accurate subtraction of the thermal background emission is critical.Methods. PynPoint is written in Python 2.7 and applies various image-processing techniques, as well as statistical tools for analyzing the data, building on open-source Python packages. The current version of PynPoint has evolved from an earlier version that was developed as a PSF-subtraction tool based on principal component analysis (PCA).Results. The architecture of PynPoint has been redesigned with the core functionalities decoupled from the pipeline modules. Modules have been implemented for dedicated processing and analysis steps, including background subtraction, frame registration, PSF subtraction, photometric and astrometric measurements, and estimation of detection limits. The pipeline package enables end-to-end data reduction of pupil-stabilized data and supports classical dithering and coronagraphic data sets. As an example, we processed archival VLT/NACO L ′ and M ′ data of β Pic b and reassessed the brightness and position of the planet with a Markov chain Monte Carlo analysis; we also provide a derivation of the photometric error budget.

Posted Content
TL;DR: A pipeline is proposed that takes advantage of a pre-trained general animal detector and a smaller set of labeled images to train a classification model that can efficiently achieve accurate results in a new region.
Abstract: Biologists all over the world use camera traps to monitor biodiversity and wildlife population density. The computer vision community has been making strides towards automating the species classification challenge in camera traps, but it has proven difficult to to apply models trained in one region to images collected in different geographic areas. In some cases, accuracy falls off catastrophically in new region, due to both changes in background and the presence of previously-unseen species. We propose a pipeline that takes advantage of a pre-trained general animal detector and a smaller set of labeled images to train a classification model that can efficiently achieve accurate results in a new region.

Journal ArticleDOI
TL;DR: In this paper, a combination of distributed fiber optical sensor and conjugate beam method was used to solve the hot issue on pipeline deformation monitoring, and the results of the two tests indicate that the deformation of the pipeline can be well monitored and the method can be applied to the field applications.

Journal ArticleDOI
02 Aug 2019
TL;DR: This work develops a pipeline that can automatically generate segmentation labels for drivable areas and road anomalies using RGB-D data-based semantic segmentation neural networks and gets predicted labels and shows that the proposed automatic labeling pipeline achieves an impressive speed-up compared to manual labeling.
Abstract: The segmentation of drivable areas and road anomalies are critical capabilities to achieve autonomous navigation for robotic wheelchairs. The recent progress of semantic segmentation using deep learning techniques has presented effective results. However, the acquisition of large-scale datasets with hand-labeled ground truth is time-consuming and labor-intensive, making the deep learning-based methods often hard to implement in practice. We contribute to the solution of this problem for the task of drivable area and road anomaly segmentation by proposing a self-supervised learning approach. We develop a pipeline that can automatically generate segmentation labels for drivable areas and road anomalies. Then, we train RGB-D data-based semantic segmentation neural networks and get predicted labels. Experimental results show that our proposed automatic labeling pipeline achieves an impressive speed-up compared to manual labeling. In addition, our proposed self-supervised approach exhibits more robust and accurate results than the state-of-the-art traditional algorithms as well as the state-of-the-art self-supervised algorithms.

Journal ArticleDOI
21 Jun 2019
TL;DR: This letter presents an automated robotic system for fast, precise, and noninvasive measurements using a new deep-learning-based next-best view planning pipeline that significantly outperforms the single robot in flexibility and planning time.
Abstract: Manual plant phenotyping is slow, error prone, and labor intensive. In this letter, we present an automated robotic system for fast, precise, and noninvasive measurements using a new deep-learning-based next-best view planning pipeline. Specifically, we first use a deep neural network to estimate a set of candidate voxels for the next scanning. Next, we cast rays from these voxels to determine the optimal viewpoints. We empirically evaluate our method in simulations and real-world robotic experiments with up to three robotic arms to demonstrate its efficiency and effectiveness. One advantage of our new pipeline is that it can be easily extended to a multi-robot system where multiple robots move simultaneously according to the planned motions. Our system significantly outperforms the single robot in flexibility and planning time. High-throughput phenotyping can be made practically.


Journal ArticleDOI
TL;DR: Two estimation problems for pipeline systems in which measurements of the compressible gas flowing through a network of pipes are affected by time-varying injections, withdrawals, and compression are formulated and a rapid, scalable computational method for performing a nonlinear least squares estimation is developed.
Abstract: We formulate two estimation problems for pipeline systems in which measurements of the compressible gas flowing through a network of pipes are affected by time-varying injections, withdrawals, and compression. We consider a state estimation problem that is then extended to a joint state and parameter estimation problem that can be used for data assimilation. In both formulations, the flow dynamics are described on each pipe by space- and time-dependent densities and mass flux which evolve according to a system of coupled partial differential equations, in which momentum dissipation is modeled using the Darcy–Wiesbach friction approximation. These dynamics are first spatially discretized to obtain a system of nonlinear ordinary differential equations on which state and parameter estimation formulations are given as nonlinear least squares problems. A rapid, scalable computational method for performing a nonlinear least squares estimation is developed. Extensive simulations and computational experiments on multiple pipeline test networks demonstrate the effectiveness of the formulations in obtaining state and parameter estimates in the presence of measurement and process noise.

Journal ArticleDOI
TL;DR: A nonlinear image-based visual servo control approach for pipeline following of fully-actuated Autonomous Underwater Vehicles that makes use of the binormalized Plucker coordinates of the pipeline borders detected in the image plane while the system dynamics are exploited in a cascade manner in the control design.

Journal ArticleDOI
01 Jan 2019
TL;DR: The time is ripe to consider molecular storage seriously and explore system designs and architectural implications, presenting a computer systems perspective on molecular processing and storage, positing a hybrid molecular–electronic architecture that plays to the strengths of both domains.
Abstract: Moore’s law may be slowing, but our ability to manipulate molecules is improving faster than ever. DNA could provide alternative substrates for computing and storage as existing ones approach physical limits. In this paper, we explore the implications of this trend in computer architecture. We present a computer systems perspective on molecular processing and storage, positing a hybrid molecular–electronic architecture that plays to the strengths of both domains. We cover the design and implementation of all stages of the pipeline: encoding, DNA synthesis, system integration with digital microfluidics, DNA sequencing (including emerging technologies such as nanopores), and decoding. We first draw on our experience designing a DNA-based archival storage system, which includes the largest demonstration to date of DNA digital data storage of over three billion nucleotides encoding over 400 MB of data. We then propose a more ambitious hybrid–electronic design that uses a molecular form of near-data processing for massive parallelism. We present a model that demonstrates the feasibility of these systems in the near future. We think the time is ripe to consider molecular storage seriously and explore system designs and architectural implications.

Journal ArticleDOI
TL;DR: Shouji as mentioned in this paper is a parallel and accurate pre-alignment filter that remarkably reduces the need for computationally-costly dynamic programming algorithms and can be adapted for any bioinformatics pipeline that performs sequence alignment for verification.
Abstract: Motivation The ability to generate massive amounts of sequencing data continues to overwhelm the processing capability of existing algorithms and compute infrastructures. In this work, we explore the use of hardware/software co-design and hardware acceleration to significantly reduce the execution time of short sequence alignment, a crucial step in analyzing sequenced genomes. We introduce Shouji, a highly parallel and accurate pre-alignment filter that remarkably reduces the need for computationally-costly dynamic programming algorithms. The first key idea of our proposed pre-alignment filter is to provide high filtering accuracy by correctly detecting all common subsequences shared between two given sequences. The second key idea is to design a hardware accelerator that adopts modern field-programmable gate array (FPGA) architectures to further boost the performance of our algorithm. Results Shouji significantly improves the accuracy of pre-alignment filtering by up to two orders of magnitude compared to the state-of-the-art pre-alignment filters, GateKeeper and SHD. Our FPGA-based accelerator is up to three orders of magnitude faster than the equivalent CPU implementation of Shouji. Using a single FPGA chip, we benchmark the benefits of integrating Shouji with five state-of-the-art sequence aligners, designed for different computing platforms. The addition of Shouji as a pre-alignment step reduces the execution time of the five state-of-the-art sequence aligners by up to 18.8×. Shouji can be adapted for any bioinformatics pipeline that performs sequence alignment for verification. Unlike most existing methods that aim to accelerate sequence alignment, Shouji does not sacrifice any of the aligner capabilities, as it does not modify or replace the alignment step. Availability and implementation https://github.com/CMU-SAFARI/Shouji. Supplementary information Supplementary data are available at Bioinformatics online.