scispace - formally typeset
Search or ask a question
Browse all papers

Journal ArticleDOI
TL;DR: The early outbreak data largely follows the exponential growth and indicates the potential of 2019-nCoV to cause outbreaks, as well as the impact of the variations in disease reporting rate, modelled through theonential growth.

1,561 citations


Journal ArticleDOI
TL;DR: An efficient evaluation tool for 3D medical image segmentation is proposed using 20 evaluation metrics based on a comprehensive literature review and guidelines for selecting a subset of these metrics that is suitable for the data and the segmentation task are provided.
Abstract: Medical Image segmentation is an important image processing step. Comparing images to evaluate the quality of segmentation is an essential part of measuring progress in this research area. Some of the challenges in evaluating medical segmentation are: metric selection, the use in the literature of multiple definitions for certain metrics, inefficiency of the metric calculation implementations leading to difficulties with large volumes, and lack of support for fuzzy segmentation by existing metrics. First we present an overview of 20 evaluation metrics selected based on a comprehensive literature review. For fuzzy segmentation, which shows the level of membership of each voxel to multiple classes, fuzzy definitions of all metrics are provided. We present a discussion about metric properties to provide a guide for selecting evaluation metrics. Finally, we propose an efficient evaluation tool implementing the 20 selected metrics. The tool is optimized to perform efficiently in terms of speed and required memory, also if the image size is extremely large as in the case of whole body MRI or CT volume segmentation. An implementation of this tool is available as an open source project. We propose an efficient evaluation tool for 3D medical image segmentation using 20 evaluation metrics and provide guidelines for selecting a subset of these metrics that is suitable for the data and the segmentation task.

1,561 citations


Journal ArticleDOI
TL;DR: This work proposes a novel hybrid densely connected UNet (H-DenseUNet), which consists of a 2-D Dense UNet for efficiently extracting intra-slice features and a 3-D counterpart for hierarchically aggregating volumetric contexts under the spirit of the auto-context algorithm for liver and tumor segmentation.
Abstract: Liver cancer is one of the leading causes of cancer death. To assist doctors in hepatocellular carcinoma diagnosis and treatment planning, an accurate and automatic liver and tumor segmentation method is highly demanded in the clinical practice. Recently, fully convolutional neural networks (FCNs), including 2-D and 3-D FCNs, serve as the backbone in many volumetric image segmentation. However, 2-D convolutions cannot fully leverage the spatial information along the third dimension while 3-D convolutions suffer from high computational cost and GPU memory consumption. To address these issues, we propose a novel hybrid densely connected UNet (H-DenseUNet), which consists of a 2-D DenseUNet for efficiently extracting intra-slice features and a 3-D counterpart for hierarchically aggregating volumetric contexts under the spirit of the auto-context algorithm for liver and tumor segmentation. We formulate the learning process of the H-DenseUNet in an end-to-end manner, where the intra-slice representations and inter-slice features can be jointly optimized through a hybrid feature fusion layer. We extensively evaluated our method on the data set of the MICCAI 2017 Liver Tumor Segmentation Challenge and 3DIRCADb data set. Our method outperformed other state-of-the-arts on the segmentation results of tumors and achieved very competitive performance for liver segmentation even with a single model.

1,561 citations


Journal ArticleDOI
25 Feb 2020-JAMA
TL;DR: Yet another pathogenic HCoV, 2019 novel coronavirus (2019-nCoV), was recognized in Wuhan, China, and has caused serious illness and death, and the ultimate scope and effect of this outbreak is unclear at present.
Abstract: Human coronaviruses (HCoVs) have long been considered inconsequential pathogens, causing the “common cold” in otherwise healthy people. However, in the 21st century, 2 highly pathogenic HCoVs—severe acute respiratory syndrome coronavirus (SARS-CoV) and Middle East respiratory syndrome coronavirus (MERS-CoV)—emerged from animal reservoirs to cause global epidemics with alarming morbidity and mortality. In December 2019, yet another pathogenic HCoV, 2019 novel coronavirus (2019-nCoV), was recognized in Wuhan, China, and has caused serious illness and death. The ultimate scope and effect of this outbreak is unclear at present as the situation is rapidly evolving. Coronaviruses are large, enveloped, positivestrand RNA viruses that can be divided into 4 genera: alpha, beta, delta, and gamma, of which alpha and beta CoVs are known to infect humans.1 Four HCoVs (HCoV 229E, NL63, OC43, and HKU1) are endemic globally and account for 10% to 30% of upper respiratory tract infections in adults. Coronaviruses are ecologically diverse with the greatest variety seen in bats, suggesting that they are the reservoirs for many of these viruses.2 Peridomestic mammals may serve as intermediate hosts, facilitating recombination and mutation events with expansion of genetic diversity.

1,561 citations


Proceedings ArticleDOI
01 Jul 2017
TL;DR: This work proposes a multi-scale convolutional neural network that restores sharp images in an end-to-end manner where blur is caused by various sources and presents a new large-scale dataset that provides pairs of realistic blurry image and the corresponding ground truth sharp image that are obtained by a high-speed camera.
Abstract: Non-uniform blind deblurring for general dynamic scenes is a challenging computer vision problem as blurs arise not only from multiple object motions but also from camera shake, scene depth variation. To remove these complicated motion blurs, conventional energy optimization based methods rely on simple assumptions such that blur kernel is partially uniform or locally linear. Moreover, recent machine learning based methods also depend on synthetic blur datasets generated under these assumptions. This makes conventional deblurring methods fail to remove blurs where blur kernel is difficult to approximate or parameterize (e.g. object motion boundaries). In this work, we propose a multi-scale convolutional neural network that restores sharp images in an end-to-end manner where blur is caused by various sources. Together, we present multi-scale loss function that mimics conventional coarse-to-fine approaches. Furthermore, we propose a new large-scale dataset that provides pairs of realistic blurry image and the corresponding ground truth sharp image that are obtained by a high-speed camera. With the proposed model trained on this dataset, we demonstrate empirically that our method achieves the state-of-the-art performance in dynamic scene deblurring not only qualitatively, but also quantitatively.

1,560 citations


Journal ArticleDOI
TL;DR: Findings indicate that the opioid overdose epidemic is worsening and there is a need for continued action to prevent opioid abuse, dependence, and death, improve treatment capacity for opioid use disorders, and reduce the supply of illicit opioids, particularly heroin and illicit fentanyl.
Abstract: The United States is experiencing an epidemic of drug overdose (poisoning) deaths. Since 2000, the rate of deaths from drug overdoses has increased 137%, including a 200% increase in the rate of overdose deaths involving opioids (opioid pain relievers and heroin). CDC analyzed recent multiple cause-of-death mortality data to examine current trends and characteristics of drug overdose deaths, including the types of opioids associated with drug overdose deaths. During 2014, a total of 47,055 drug overdose deaths occurred in the United States, representing a 1-year increase of 6.5%, from 13.8 per 100,000 persons in 2013 to 14.7 per 100,000 persons in 2014. The rate of drug overdose deaths increased significantly for both sexes, persons aged 25-44 years and ≥55 years, non-Hispanic whites and non-Hispanic blacks, and in the Northeastern, Midwestern, and Southern regions of the United States. Rates of opioid overdose deaths also increased significantly, from 7.9 per 100,000 in 2013 to 9.0 per 100,000 in 2014, a 14% increase. Historically, CDC has programmatically characterized all opioid pain reliever deaths (natural and semisynthetic opioids, methadone, and other synthetic opioids) as "prescription" opioid overdoses (1). Between 2013 and 2014, the age-adjusted rate of death involving methadone remained unchanged; however, the age-adjusted rate of death involving natural and semisynthetic opioid pain relievers, heroin, and synthetic opioids, other than methadone (e.g., fentanyl) increased 9%, 26%, and 80%, respectively. The sharp increase in deaths involving synthetic opioids, other than methadone, in 2014 coincided with law enforcement reports of increased availability of illicitly manufactured fentanyl, a synthetic opioid; however, illicitly manufactured fentanyl cannot be distinguished from prescription fentanyl in death certificate data. These findings indicate that the opioid overdose epidemic is worsening. There is a need for continued action to prevent opioid abuse, dependence, and death, improve treatment capacity for opioid use disorders, and reduce the supply of illicit opioids, particularly heroin and illicit fentanyl.

1,559 citations


Posted Content
TL;DR: This work introduces a new, efficient, principled and backpropagation-compatible algorithm for learning a probability distribution on the weights of a neural network, called Bayes by Backprop, and shows how the learnt uncertainty in the weights can be used to improve generalisation in non-linear regression problems.
Abstract: We introduce a new, efficient, principled and backpropagation-compatible algorithm for learning a probability distribution on the weights of a neural network, called Bayes by Backprop. It regularises the weights by minimising a compression cost, known as the variational free energy or the expected lower bound on the marginal likelihood. We show that this principled kind of regularisation yields comparable performance to dropout on MNIST classification. We then demonstrate how the learnt uncertainty in the weights can be used to improve generalisation in non-linear regression problems, and how this weight uncertainty can be used to drive the exploration-exploitation trade-off in reinforcement learning.

1,558 citations


Proceedings Article
05 Dec 2016
TL;DR: The authors apply this variational inference based dropout technique in LSTM and GRU models, assessing it on language modelling and sentiment analysis tasks, and to the best of their knowledge improve on the single model state-of-the-art in language modelling with the Penn Treebank (73.4 test perplexity).
Abstract: Recurrent neural networks (RNNs) stand at the forefront of many recent developments in deep learning. Yet a major difficulty with these models is their tendency to overfit, with dropout shown to fail when applied to recurrent layers. Recent results at the intersection of Bayesian modelling and deep learning offer a Bayesian interpretation of common deep learning techniques such as dropout. This grounding of dropout in approximate Bayesian inference suggests an extension of the theoretical results, offering insights into the use of dropout with RNN models. We apply this new variational inference based dropout technique in LSTM and GRU models, assessing it on language modelling and sentiment analysis tasks. The new approach outperforms existing techniques, and to the best of our knowledge improves on the single model state-of-the-art in language modelling with the Penn Treebank (73.4 test perplexity). This extends our arsenal of variational tools in deep learning.

1,557 citations


Journal ArticleDOI
TL;DR: This work organizes and analyze what has been learned from the past 35 years of work on emotion and decision making and proposes the emotion-imbued choice model, which accounts for inputs from traditional rational choice theory and from newer emotion research, synthesizing scientific models.
Abstract: A revolution in the science of emotion has emerged in recent decades, with the potential to create a paradigm shift in decision theories. The research reveals that emotions constitute potent, pervasive, predictable, sometimes harmful and sometimes beneficial drivers of decision making. Across different domains, important regularities appear in the mechanisms through which emotions influence judgments and choices. We organize and analyze what has been learned from the past 35 years of work on emotion and decision making. In so doing, we propose the emotion-imbued choice model, which accounts for inputs from traditional rational choice theory and from newer emotion research, synthesizing scientific models.

1,556 citations


Journal ArticleDOI
TL;DR: Several selected pharmaceutical and biomedical applications are presented, in which chitin and chitosan are recognized as new biomaterials taking advantage of their biocompatibility and biodegradability.
Abstract: This review describes the most common methods for recovery of chitin from marine organisms. In depth, both enzymatic and chemical treatments for the step of deproteinization are compared, as well as different conditions for demineralization. The conditions of chitosan preparation are also discussed, since they significantly impact the synthesis of chitosan with varying degree of acetylation (DA) and molecular weight (MW). In addition, the main characterization techniques applied for chitin and chitosan are recalled, pointing out the role of their solubility in relation with the chemical structure (mainly the acetyl group distribution along the backbone). Biological activities are also presented, such as: antibacterial, antifungal, antitumor and antioxidant. Interestingly, the relationship between chemical structure and biological activity is demonstrated for chitosan molecules with different DA and MW and homogeneous distribution of acetyl groups for the first time. In the end, several selected pharmaceutical and biomedical applications are presented, in which chitin and chitosan are recognized as new biomaterials taking advantage of their biocompatibility and biodegradability.

1,554 citations


Journal ArticleDOI
TL;DR: ColabFold as discussed by the authors combines the fast homology search of MMseqs2 with AlphaFold2 or RoseTTAFold for protein folding and achieves 40-60fold faster search and optimized model utilization.
Abstract: ColabFold offers accelerated prediction of protein structures and complexes by combining the fast homology search of MMseqs2 with AlphaFold2 or RoseTTAFold. ColabFold's 40-60-fold faster search and optimized model utilization enables prediction of close to 1,000 structures per day on a server with one graphics processing unit. Coupled with Google Colaboratory, ColabFold becomes a free and accessible platform for protein folding. ColabFold is open-source software available at https://github.com/sokrypton/ColabFold and its novel environmental databases are available at https://colabfold.mmseqs.com .

Journal ArticleDOI
01 Apr 2016
TL;DR: ASD prevalence estimates for children aged 8 years living in catchment areas of the ADDM Network sites in 2012 are provided, overall and stratified by sex, race/ethnicity, and the type of source records (education and health records versus health records only).
Abstract: PROBLEM/CONDITION Autism spectrum disorder (ASD). PERIOD COVERED 2012. DESCRIPTION OF SYSTEM The Autism and Developmental Disabilities Monitoring (ADDM) Network is an active surveillance system that provides estimates of the prevalence and characteristics of ASD among children aged 8 years whose parents or guardians reside in 11 ADDM Network sites in the United States (Arkansas, Arizona, Colorado, Georgia, Maryland, Missouri, New Jersey, North Carolina, South Carolina, Utah, and Wisconsin). Surveillance to determine ASD case status is conducted in two phases. The first phase consists of screening and abstracting comprehensive evaluations performed by professional service providers in the community. Data sources identified for record review are categorized as either 1) education source type, including developmental evaluations to determine eligibility for special education services or 2) health care source type, including diagnostic and developmental evaluations. The second phase involves the review of all abstracted evaluations by trained clinicians to determine ASD surveillance case status. A child meets the surveillance case definition for ASD if one or more comprehensive evaluations of that child completed by a qualified professional describes behaviors that are consistent with the Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition, Text Revision diagnostic criteria for any of the following conditions: autistic disorder, pervasive developmental disorder-not otherwise specified (including atypical autism), or Asperger disorder. This report provides ASD prevalence estimates for children aged 8 years living in catchment areas of the ADDM Network sites in 2012, overall and stratified by sex, race/ethnicity, and the type of source records (education and health records versus health records only). In addition, this report describes the proportion of children with ASD with a score consistent with intellectual disability on a standardized intellectual ability test, the age at which the earliest known comprehensive evaluation was performed, the proportion of children with a previous ASD diagnosis, the specific type of ASD diagnosis, and any special education eligibility classification. RESULTS For 2012, the combined estimated prevalence of ASD among the 11 ADDM Network sites was 14.5 per 1,000 (one in 69) children aged 8 years. Estimated prevalence was significantly higher among boys aged 8 years (23.4 per 1,000) than among girls aged 8 years (5.2 per 1,000). Estimated ASD prevalence was significantly higher among non-Hispanic white children aged 8 years (15.3 per 1,000) compared with non-Hispanic black children (13.1 per 1,000), and Hispanic (10.2 per 1,000) children aged 8 years. Estimated prevalence varied widely among the 11 ADDM Network sites, ranging from 8.2 per 1,000 children aged 8 years (in the area of the Maryland site where only health care records were reviewed) to 24.6 per 1,000 children aged 8 years (in New Jersey, where both education and health care records were reviewed). Estimated prevalence was higher in surveillance sites where education records and health records were reviewed compared with sites where health records only were reviewed (17.1 per 1,000 and 10.4 per 1,000 children aged 8 years, respectively; p<0.05). Among children identified with ASD by the ADDM Network, 82% had a previous ASD diagnosis or educational classification; this did not vary by sex or between non-Hispanic white and non-Hispanic black children. A lower percentage of Hispanic children (78%) had a previous ASD diagnosis or classification compared with non-Hispanic white children (82%) and with non-Hispanic black children (84%). The median age at earliest known comprehensive evaluation was 40 months, and 43% of children had received an earliest known comprehensive evaluation by age 36 months. The percentage of children with an earliest known comprehensive evaluation by age 36 months was similar for boys and girls, but was higher for non-Hispanic white children (45%) compared with non-Hispanic black children (40%) and Hispanic children (39%). INTERPRETATION Overall estimated ASD prevalence was 14.5 per 1,000 children aged 8 years in the ADDM Network sites in 2012. The higher estimated prevalence among sites that reviewed both education and health records suggests the role of special education systems in providing comprehensive evaluations and services to children with developmental disabilities. Disparities by race/ethnicity in estimated ASD prevalence, particularly for Hispanic children, as well as disparities in the age of earliest comprehensive evaluation and presence of a previous ASD diagnosis or classification, suggest that access to treatment and services might be lacking or delayed for some children. PUBLIC HEALTH ACTION The ADDM Network will continue to monitor the prevalence and characteristics of ASD among children aged 8 years living in selected sites across the United States. Recommendations from the ADDM Network include enhancing strategies to 1) lower the age of first evaluation of ASD by community providers in accordance with the Healthy People 2020 goal that children with ASD are evaluated by age 36 months and begin receiving community-based support and services by age 48 months; 2) reduce disparities by race/ethnicity in identified ASD prevalence, the age of first comprehensive evaluation, and presence of a previous ASD diagnosis or classification; and 3) assess the effect on ASD prevalence of the revised ASD diagnostic criteria published in the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition.

Journal ArticleDOI
03 Apr 2020
TL;DR: A Distance-IoU (DIoU) loss is proposed by incorporating the normalized distance between the predicted box and the target box, which converges much faster in training than IoU and GIoU losses, thereby leading to faster convergence and better performance.
Abstract: Bounding box regression is the crucial step in object detection. In existing methods, while ln-norm loss is widely adopted for bounding box regression, it is not tailored to the evaluation metric, i.e., Intersection over Union (IoU). Recently, IoU loss and generalized IoU (GIoU) loss have been proposed to benefit the IoU metric, but still suffer from the problems of slow convergence and inaccurate regression. In this paper, we propose a Distance-IoU (DIoU) loss by incorporating the normalized distance between the predicted box and the target box, which converges much faster in training than IoU and GIoU losses. Furthermore, this paper summarizes three geometric factors in bounding box regression, i.e., overlap area, central point distance and aspect ratio, based on which a Complete IoU (CIoU) loss is proposed, thereby leading to faster convergence and better performance. By incorporating DIoU and CIoU losses into state-of-the-art object detection algorithms, e.g., YOLO v3, SSD and Faster R-CNN, we achieve notable performance gains in terms of not only IoU metric but also GIoU metric. Moreover, DIoU can be easily adopted into non-maximum suppression (NMS) to act as the criterion, further boosting performance improvement. The source code and trained models are available at https://github.com/Zzh-tju/DIoU.

Journal ArticleDOI
TL;DR: Res2Net as mentioned in this paper constructs hierarchical residual-like connections within one single residual block to represent multi-scale features at a granular level and increases the range of receptive fields for each network layer.
Abstract: Representing features at multiple scales is of great importance for numerous vision tasks. Recent advances in backbone convolutional neural networks (CNNs) continually demonstrate stronger multi-scale representation ability, leading to consistent performance gains on a wide range of applications. However, most existing methods represent the multi-scale features in a layer-wise manner. In this paper, we propose a novel building block for CNNs, namely Res2Net, by constructing hierarchical residual-like connections within one single residual block. The Res2Net represents multi-scale features at a granular level and increases the range of receptive fields for each network layer. The proposed Res2Net block can be plugged into the state-of-the-art backbone CNN models, e.g., ResNet, ResNeXt, and DLA. We evaluate the Res2Net block on all these models and demonstrate consistent performance gains over baseline models on widely-used datasets, e.g., CIFAR-100 and ImageNet. Further ablation studies and experimental results on representative computer vision tasks, i.e., object detection, class activation mapping, and salient object detection, further verify the superiority of the Res2Net over the state-of-the-art baseline methods. The source code and trained models are available on https://mmcheng.net/res2net/ .

Journal ArticleDOI
TL;DR: A comprehensive and structured analysis of various graph embedding techniques proposed in the literature, and the open-source Python library, named GEM (Graph Embedding Methods, available at https://github.com/palash1992/GEM ), which provides all presented algorithms within a unified interface to foster and facilitate research on the topic.
Abstract: Graphs, such as social networks, word co-occurrence networks, and communication networks, occur naturally in various real-world applications. Analyzing them yields insight into the structure of society, language, and different patterns of communication. Many approaches have been proposed to perform the analysis. Recently, methods which use the representation of graph nodes in vector space have gained traction from the research community. In this survey, we provide a comprehensive and structured analysis of various graph embedding techniques proposed in the literature. We first introduce the embedding task and its challenges such as scalability, choice of dimensionality, and features to be preserved, and their possible solutions. We then present three categories of approaches based on factorization methods, random walks, and deep learning, with examples of representative algorithms in each category and analysis of their performance on various tasks. We evaluate these state-of-the-art methods on a few common datasets and compare their performance against one another. Our analysis concludes by suggesting some potential applications and future directions. We finally present the open-source Python library we developed, named GEM (Graph Embedding Methods, available at https://github.com/palash1992/GEM ), which provides all presented algorithms within a unified interface to foster and facilitate research on the topic.

Posted ContentDOI
30 Oct 2015-bioRxiv
TL;DR: The aggregation and analysis of high-quality exome (protein-coding region) sequence data for 60,706 individuals of diverse ethnicities generated as part of the Exome Aggregation Consortium (ExAC) provides direct evidence for the presence of widespread mutational recurrence.
Abstract: Large-scale reference data sets of human genetic variation are critical for the medical and functional interpretation of DNA sequence changes. Here we describe the aggregation and analysis of high-quality exome (protein-coding region) sequence data for 60,706 individuals of diverse ethnicities. The resulting catalogue of human genetic diversity has unprecedented resolution, with an average of one variant every eight bases of coding sequence and the presence of widespread mutational recurrence. The deep catalogue of variation provided by the Exome Aggregation Consortium (ExAC) can be used to calculate objective metrics of pathogenicity for sequence variants, and to identify genes subject to strong selection against various classes of mutation; we identify 3,230 genes with near-complete depletion of truncating variants, 79% of which have no currently established human disease phenotype. Finally, we show that these data can be used for the efficient filtering of candidate disease-causing variants, and for the discovery of human knockout variants in protein-coding genes.

Journal ArticleDOI
TL;DR: The use of maintenance therapy with olaparib provided a substantial benefit with regard to progression‐free survival among women with newly diagnosed advanced ovarian cancer and a BRCA1/2 mutation, with a 70% lower risk of disease progression or death with olAParib than with placebo.
Abstract: Background Most women with newly diagnosed advanced ovarian cancer have a relapse within 3 years after standard treatment with surgery and platinum-based chemotherapy. The benefit of the o...

Journal ArticleDOI
TL;DR: In this paper, the authors provide an overview of the latest NOMA research and innovations as well as their applications in 5G wireless networks and discuss future challenges and future research challenges.
Abstract: Non-orthogonal multiple access (NOMA) is an essential enabling technology for the fifth-generation (5G) wireless networks to meet the heterogeneous demands on low latency, high reliability, massive connectivity, improved fairness, and high throughput. The key idea behind NOMA is to serve multiple users in the same resource block, such as a time slot, subcarrier, or spreading code. The NOMA principle is a general framework, and several recently proposed 5G multiple access schemes can be viewed as special cases. This survey provides an overview of the latest NOMA research and innovations as well as their applications. Thereby, the papers published in this special issue are put into the context of the existing literature. Future research challenges regarding NOMA in 5G and beyond are also discussed.

Journal ArticleDOI
TL;DR: Considering that the treatment goal for burnout is usually to enable people to return to their job, and to be successful in their work, psychiatry could make an important contribution by identifying the treatment strategies that would be most effective in achieving that goal.

Journal Article
TL;DR: MLlib as mentioned in this paper is an open-source distributed machine learning library for Apache Spark that provides efficient functionality for a wide range of learning settings and includes several underlying statistical, optimization, and linear algebra primitives.
Abstract: Apache Spark is a popular open-source platform for large-scale data processing that is well-suited for iterative machine learning tasks. In this paper we present MLlib, Spark's open-source distributed machine learning library. MLLIB provides efficient functionality for a wide range of learning settings and includes several underlying statistical, optimization, and linear algebra primitives. Shipped with Spark, MLLIB supports several languages and provides a high-level API that leverages Spark's rich ecosystem to simplify the development of end-to-end machine learning pipelines. MLLIB has experienced a rapid growth due to its vibrant open-source community of over 140 contributors, and includes extensive documentation to support further growth and to let users quickly get up to speed.

Proceedings Article
06 Jul 2015
TL;DR: It is found that adding a bias of 1 to the LSTM's forget gate closes the gap between the L STM and the recently-introduced Gated Recurrent Unit (GRU) on some but not all tasks.
Abstract: The Recurrent Neural Network (RNN) is an extremely powerful sequence model that is often difficult to train. The Long Short-Term Memory (LSTM) is a specific RNN architecture whose design makes it much easier to train. While wildly successful in practice, the LSTM's architecture appears to be ad-hoc so it is not clear if it is optimal, and the significance of its individual components is unclear. In this work, we aim to determine whether the LSTM architecture is optimal or whether much better architectures exist. We conducted a thorough architecture search where we evaluated over ten thousand different RNN architectures, and identified an architecture that outperforms both the LSTM and the recently-introduced Gated Recurrent Unit (GRU) on some but not all tasks. We found that adding a bias of 1 to the LSTM's forget gate closes the gap between the LSTM and the GRU.

Journal ArticleDOI
TL;DR: Among patients undergoing resection of stage IIIB, IIIC, or IV melanoma, adjuvant therapy with nivolumab resulted in significantly longer recurrence‐free survival and a lower rate of grade 3 or 4 adverse events than adjuant therapy with ipilimumab.
Abstract: BackgroundNivolumab and ipilimumab are immune checkpoint inhibitors that have been approved for the treatment of advanced melanoma. In the United States, ipilimumab has also been approved as adjuvant therapy for melanoma on the basis of recurrence-free and overall survival rates that were higher than those with placebo in a phase 3 trial. We wanted to determine the efficacy of nivolumab versus ipilimumab for adjuvant therapy in patients with resected advanced melanoma. MethodsIn this randomized, double-blind, phase 3 trial, we randomly assigned 906 patients (≥15 years of age) who were undergoing complete resection of stage IIIB, IIIC, or IV melanoma to receive an intravenous infusion of either nivolumab at a dose of 3 mg per kilogram of body weight every 2 weeks (453 patients) or ipilimumab at a dose of 10 mg per kilogram every 3 weeks for four doses and then every 12 weeks (453 patients). The patients were treated for a period of up to 1 year or until disease recurrence, a report of unacceptable toxic ef...

Journal ArticleDOI
TL;DR: A reversible photo-induced instability has been found in mixed-halide photovoltaic perovskites that limits the open circuit voltage in solar cells.
Abstract: We report on reversible, light-induced transformations in (CH3NH3)Pb(BrxI1−x)3. Photoluminescence (PL) spectra of these perovskites develop a new, red-shifted peak at 1.68 eV that grows in intensity under constant, 1-sun illumination in less than a minute. This is accompanied by an increase in sub-bandgap absorption at ∼1.7 eV, indicating the formation of luminescent trap states. Light soaking causes a splitting of X-ray diffraction (XRD) peaks, suggesting segregation into two crystalline phases. Surprisingly, these photo-induced changes are fully reversible; the XRD patterns and the PL and absorption spectra revert to their initial states after the materials are left for a few minutes in the dark. We speculate that photoexcitation may cause halide segregation into iodide-rich minority and bromide-enriched majority domains, the former acting as a recombination center trap. This instability may limit achievable voltages from some mixed-halide perovskite solar cells and could have implications for the photostability of halide perovskites used in optoelectronics.

Journal ArticleDOI
TL;DR: Among patients with persistent atrial fibrillation, there was no reduction in the rate of recurrent atrialfibrillation when either linear ablation or ablation of complex fractionated electrograms was performed in addition to pulmonary-vein isolation.
Abstract: BackgroundCatheter ablation is less successful for persistent atrial fibrillation than for paroxysmal atrial fibrillation. Guidelines suggest that adjuvant substrate modification in addition to pulmonary-vein isolation is required in persistent atrial fibrillation. MethodsWe randomly assigned 589 patients with persistent atrial fibrillation in a 1:4:4 ratio to ablation with pulmonary-vein isolation alone (67 patients), pulmonary-vein isolation plus ablation of electrograms showing complex fractionated activity (263 patients), or pulmonary-vein isolation plus additional linear ablation across the left atrial roof and mitral valve isthmus (259 patients). The duration of follow-up was 18 months. The primary end point was freedom from any documented recurrence of atrial fibrillation lasting longer than 30 seconds after a single ablation procedure. ResultsProcedure time was significantly shorter for pulmonary-vein isolation alone than for the other two procedures (P<0.001). After 18 months, 59% of patients ass...

Proceedings ArticleDOI
Konstantinos Bousmalis1, Nathan Silberman1, David Dohan1, Dumitru Erhan1, Dilip Krishnan1 
01 Jul 2017
TL;DR: In this paper, a generative adversarial network (GAN)-based method adapts source-domain images to appear as if drawn from the target domain by learning in an unsupervised manner a transformation in the pixel space from one domain to another.
Abstract: Collecting well-annotated image datasets to train modern machine learning algorithms is prohibitively expensive for many tasks. One appealing alternative is rendering synthetic data where ground-truth annotations are generated automatically. Unfortunately, models trained purely on rendered images fail to generalize to real images. To address this shortcoming, prior work introduced unsupervised domain adaptation algorithms that have tried to either map representations between the two domains, or learn to extract features that are domain-invariant. In this work, we approach the problem in a new light by learning in an unsupervised manner a transformation in the pixel space from one domain to the other. Our generative adversarial network (GAN)-based method adapts source-domain images to appear as if drawn from the target domain. Our approach not only produces plausible samples, but also outperforms the state-of-the-art on a number of unsupervised domain adaptation scenarios by large margins. Finally, we demonstrate that the adaptation process generalizes to object classes unseen during training.

Proceedings Article
Jonas Gehring1, Michael Auli1, David Grangier1, Denis Yarats1, Yann N. Dauphin1 
08 May 2017
TL;DR: The authors introduced an architecture based entirely on convolutional neural networks, where computations over all elements can be fully parallelized during training and optimization is easier since the number of nonlinearities is fixed and independent of the input length.
Abstract: The prevalent approach to sequence to sequence learning maps an input sequence to a variable length output sequence via recurrent neural networks. We introduce an architecture based entirely on convolutional neural networks. Compared to recurrent models, computations over all elements can be fully parallelized during training and optimization is easier since the number of non-linearities is fixed and independent of the input length. Our use of gated linear units eases gradient propagation and we equip each decoder layer with a separate attention module. We outperform the accuracy of the deep LSTM setup of Wu et al. (2016) on both WMT'14 English-German and WMT'14 English-French translation at an order of magnitude faster speed, both on GPU and CPU.

Proceedings Article
05 Dec 2016
TL;DR: This work proposes coupled generative adversarial network (CoGAN), which can learn a joint distribution without any tuple of corresponding images, and applies it to several joint distribution learning tasks, and demonstrates its applications to domain adaptation and image transformation.
Abstract: We propose the coupled generative adversarial nets (CoGAN) framework for generating pairs of corresponding images in two different domains. The framework consists of a pair of generative adversarial nets, each responsible for generating images in one domain. We show that by enforcing a simple weight-sharing constraint, the CoGAN learns to generate pairs of corresponding images without existence of any pairs of corresponding images in the two domains in the training set. In other words, the CoGAN learns a joint distribution of images in the two domains from images drawn separately from the marginal distributions of the individual domains. This is in contrast to the existing multi-modal generative models, which require corresponding images for training. We apply the CoGAN to several pair image generation tasks. For each task, the CoGAN learns to generate convincing pairs of corresponding images. We further demonstrate the applications of the CoGAN framework for the domain adaptation and cross-domain image generation tasks.

Journal ArticleDOI
TL;DR: Venous thromboembolism is a complex disease, involving interactions between acquired or inherited predispositions to thrombosis and VTE risk factors, including increasing patient age and obesity, hospitalization for surgery or acute illness, nursing-home confinement, active cancer, trauma or fracture, immobility or leg paresis, superficial vein thromBosis, and, in women, pregnancy and puerperium.
Abstract: Venous thromboembolism (VTE) is categorized by the U.S. Surgeon General as a major public health problem. VTE is relatively common and associated with reduced survival and substantial health-care costs, and recurs frequently. VTE is a complex (multifactorial) disease, involving interactions between acquired or inherited predispositions to thrombosis and VTE risk factors, including increasing patient age and obesity, hospitalization for surgery or acute illness, nursing-home confinement, active cancer, trauma or fracture, immobility or leg paresis, superficial vein thrombosis, and, in women, pregnancy and puerperium, oral contraception, and hormone therapy. Although independent VTE risk factors and predictors of VTE recurrence have been identified, and effective primary and secondary prophylaxis is available, the occurrence of VTE seems to be relatively constant, or even increasing.

Book ChapterDOI
08 Sep 2018
TL;DR: BiSeNet as discussed by the authors designs a spatial path with a small stride to preserve the spatial information and generate high-resolution features, while a context path with fast downsampling strategy is employed to obtain sufficient receptive field.
Abstract: Semantic segmentation requires both rich spatial information and sizeable receptive field. However, modern approaches usually compromise spatial resolution to achieve real-time inference speed, which leads to poor performance. In this paper, we address this dilemma with a novel Bilateral Segmentation Network (BiSeNet). We first design a Spatial Path with a small stride to preserve the spatial information and generate high-resolution features. Meanwhile, a Context Path with a fast downsampling strategy is employed to obtain sufficient receptive field. On top of the two paths, we introduce a new Feature Fusion Module to combine features efficiently. The proposed architecture makes a right balance between the speed and segmentation performance on Cityscapes, CamVid, and COCO-Stuff datasets. Specifically, for a 2048 \(\times \) 1024 input, we achieve 68.4% Mean IOU on the Cityscapes test dataset with speed of 105 FPS on one NVIDIA Titan XP card, which is significantly faster than the existing methods with comparable performance.

Proceedings ArticleDOI
10 Aug 2015
TL;DR: Wang et al. as discussed by the authors proposed a hierarchical Bayesian model called collaborative deep learning (CDL), which jointly performs deep representation learning for the content information and collaborative filtering for the ratings (feedback) matrix.
Abstract: Collaborative filtering (CF) is a successful approach commonly used by many recommender systems. Conventional CF-based methods use the ratings given to items by users as the sole source of information for learning to make recommendation. However, the ratings are often very sparse in many applications, causing CF-based methods to degrade significantly in their recommendation performance. To address this sparsity problem, auxiliary information such as item content information may be utilized. Collaborative topic regression (CTR) is an appealing recent method taking this approach which tightly couples the two components that learn from two different sources of information. Nevertheless, the latent representation learned by CTR may not be very effective when the auxiliary information is very sparse. To address this problem, we generalize recently advances in deep learning from i.i.d. input to non-i.i.d. (CF-based) input and propose in this paper a hierarchical Bayesian model called collaborative deep learning (CDL), which jointly performs deep representation learning for the content information and collaborative filtering for the ratings (feedback) matrix. Extensive experiments on three real-world datasets from different domains show that CDL can significantly advance the state of the art.