scispace - formally typeset
Search or ask a question

Showing papers by "Kevin Smith published in 2018"


Journal ArticleDOI
TL;DR: It is found that engaging players of commercial computer games provided data that augmented deep learning and enabled scalable and readily improved image classification, resulting in a boosted learner that can characterize subcellular protein distribution.
Abstract: Pattern recognition and classification of images are key challenges throughout the life sciences. We combined two approaches for large-scale classification of fluorescence microscopy images. First, using the publicly available data set from the Cell Atlas of the Human Protein Atlas (HPA), we integrated an image-classification task into a mainstream video game (EVE Online) as a mini-game, named Project Discovery. Participation by 322,006 gamers over 1 year provided nearly 33 million classifications of subcellular localization patterns, including patterns that were not previously annotated by the HPA. Second, we used deep learning to build an automated Localization Cellular Annotation Tool (Loc-CAT). This tool classifies proteins into 29 subcellular localization patterns and can deal efficiently with multi-localization proteins, performing robustly across different cell types. Combining the annotations of gamers and deep learning, we applied transfer learning to create a boosted learner that can characterize subcellular protein distribution with F1 score of 0.72. We found that engaging players of commercial computer games provided data that augmented deep learning and enabled scalable and readily improved image classification.

149 citations


Proceedings Article
15 Feb 2018
TL;DR: This article showed that training a deep network using batch normalization is equivalent to approximate inference in Bayesian models, and further demonstrated that this finding allows us to make meaningful estimi cation.
Abstract: We show that training a deep network using batch normalization is equivalent to approximate inference in Bayesian models. We further demonstrate that this finding allows us to make meaningful estim ...

128 citations


Journal ArticleDOI
TL;DR: The strengths and weaknesses of non-commercial phenotypic image analysis software are examined, recent developments in the field are covered, challenges are identified, and a perspective on future possibilities are given.
Abstract: Phenotypic image analysis is the task of recognizing variations in cell properties using microscopic image data. These variations, produced through a complex web of interactions between genes and the environment, may hold the key to uncover important biological phenomena or to understand the response to a drug candidate. Today, phenotypic analysis is rarely performed completely by hand. The abundance of high-dimensional image data produced by modern high-throughput microscopes necessitates computational solutions. Over the past decade, a number of software tools have been developed to address this need. They use statistical learning methods to infer relationships between a cell's phenotype and data from the image. In this review, we examine the strengths and weaknesses of non-commercial phenotypic image analysis software, cover recent developments in the field, identify challenges, and give a perspective on future possibilities.

72 citations


Journal ArticleDOI
TL;DR: A high-throughput, scalable workflow for microscopy-based single cell isolation using machine-learning, high- throughput microscopy and laser capture microdissection is developed.
Abstract: Quantifying heterogeneities within cell populations is important for many fields including cancer research and neurobiology; however, techniques to isolate individual cells are limited. Here, we describe a high-throughput, non-disruptive, and cost-effective isolation method that is capable of capturing individually targeted cells using widely available techniques. Using high-resolution microscopy, laser microcapture microscopy, image analysis, and machine learning, our technology enables scalable molecular genetic analysis of single cells, targetable by morphology or location within the sample.

65 citations


Posted Content
TL;DR: It is shown that training a deep network using batch normalization is equivalent to approximate inference in Bayesian models, and it is demonstrated how this finding allows us to make useful estimates of the model uncertainty.
Abstract: Deep neural networks have led to a series of breakthroughs, dramatically improving the state-of-the-art in many domains. The techniques driving these advances, however, lack a formal method to account for model uncertainty. While the Bayesian approach to learning provides a solid theoretical framework to handle uncertainty, inference in Bayesian-inspired deep neural networks is difficult. In this paper, we provide a practical approach to Bayesian learning that relies on a regularization technique found in nearly every modern network, \textit{batch normalization}. We show that training a deep network using batch normalization is equivalent to approximate inference in Bayesian models, and we demonstrate how this finding allows us to make useful estimates of the model uncertainty. With our approach, it is possible to make meaningful uncertainty estimates using conventional architectures without modifying the network or the training procedure. Our approach is thoroughly validated in a series of empirical experiments on different tasks and using various measures, outperforming baselines with strong statistical significance and displaying competitive performance with other recent Bayesian approaches.

60 citations


Posted ContentDOI
05 Oct 2018-bioRxiv
TL;DR: In this paper, the authors used a siRNA knockdown screen in HeLa cells infected with Brucella abortus and identified 426 components of the human infectome for infection, including multiple components of pathways involved in central processes such as cell cycle, actin cytoskeleton dynamics or vesicular trafficking.
Abstract: Brucella, the causing agent of brucellosis, is a major zoonotic pathogen with worldwide distribution. Brucella resides and replicates inside infected host cells in membrane-bound compartments called BCVs (Brucella-containing vacuoles). Following uptake, Brucella resides in eBCVs (endosomal BCVs) that gradually mature from early to late endosomal features. Through a poorly understood process that is key to the intracellular lifestyle of Brucella, the eBCV escapes fusion with lysosomes by transitioning to the rBCV (replicative BCV), a replicative niche directly connected to the endoplasmic reticulum (ER). Despite the notion that this complex intracellular lifestyle must depend on a multitude of host factors, a holistic view on which of these components control Brucella cell entry, trafficking and replication is still missing. Here we used a systematic cell-based siRNA knockdown screen in HeLa cells infected with Brucella abortus and identified 426 components of the human infectome for Brucella infection. These include multiple components of pathways involved in central processes such as cell cycle, actin cytoskeleton dynamics or vesicular trafficking. Using assays for pathogen entry, knockdown complementation and co-localization at single-cell resolution, we identified the requirement of VPS35 for Brucella to escape the lysosomal degradative pathway and to establish its intracellular replicative niche. We thus validated a component of the retromer as novel host factor critical for Brucella intracellular trafficking. Further, our genome-wide data shed light on the interplay between central host processes and the biogenesis of the Brucella replicative niche.