scispace - formally typeset
Search or ask a question
Browse all papers

Dissertation
01 Jan 2015
TL;DR: New methods and properties of the current SVM system have been found that might increase the accuracies of the between subject analysis and therefore might enable SVM to become applicable for the extra feedback options.
Abstract: The society has become more sedentary and has developed a lack of physical activity, therefore increasing health risks. Feedback is needed to change these behaviours. For this feedback, first accurate monitoring is needed: sedentary behaviour must be classified as well as the intensity of physical activity. In this report a State of the Art analysis is performed to compare different classification techniques and finally two methods, both using an accelerometer on the waist, are worked out. These methods are Integral of the Modulus of the Accelerometer (IMA) classification and a machine learning technique (MLT): support vector machine (SVM). These methods are then applied in a laboratory experiment to study their quality. A measurement setup is made to create a dataset of the following activities: standing, sitting, lying, walking (2.4 - 7.5 km/h) and cycling (10.1-19 km/h). This dataset (n=15) is analysed and classified using Matlab for both methods. The IMA method was unable to monitor sedentary behaviour, but could classify the physical activity (PA) intensity with an accuracy of 66%. The SVM method within subjects was able to monitor sedentary behaviour with an accuracy of 91±20% and the classification of the PA intensity has an accuracy of 94±5%. For between subjects the accuracies decrease to 71±13% for PA intensity accuracy and 45±35% for the sedentary behaviour classification. IMA was implemented in the old feedback system, monitoring the overall daily amount of physical activity, but can significantly be outperformed by replacing it with the current SVM implementation. At this moment however, SVM can only be used to improve the old system, it cannot yet be used to create new additions to the feedback system, such as the implementation of the feedback of the sedentary behaviour and specific physical activity intensities. New methods and properties of the current SVM system have been found that might increase the accuracies of the between subject analysis and therefore might enable SVM to become applicable for the extra feedback options.

2,148 citations


Journal ArticleDOI
TL;DR: The results show that MAGMA has significantly more power than other tools for both the gene and the gene-set analysis, identifying more genes and gene sets associated with Crohn’s Disease while maintaining a correct type 1 error rate.
Abstract: By aggregating data for complex traits in a biologically meaningful way, gene and gene-set analysis constitute a valuable addition to single-marker analysis. However, although various methods for gene and gene-set analysis currently exist, they generally suffer from a number of issues. Statistical power for most methods is strongly affected by linkage disequilibrium between markers, multi-marker associations are often hard to detect, and the reliance on permutation to compute p-values tends to make the analysis computationally very expensive. To address these issues we have developed MAGMA, a novel tool for gene and gene-set analysis. The gene analysis is based on a multiple regression model, to provide better statistical performance. The gene-set analysis is built as a separate layer around the gene analysis for additional flexibility. This gene-set analysis also uses a regression structure to allow generalization to analysis of continuous properties of genes and simultaneous analysis of multiple gene sets and other gene properties. Simulations and an analysis of Crohn’s Disease data are used to evaluate the performance of MAGMA and to compare it to a number of other gene and gene-set analysis tools. The results show that MAGMA has significantly more power than other tools for both the gene and the gene-set analysis, identifying more genes and gene sets associated with Crohn’s Disease while maintaining a correct type 1 error rate. Moreover, the MAGMA analysis of the Crohn’s Disease data was found to be considerably faster as well.

2,147 citations


Journal ArticleDOI
TL;DR: Despite anticoagulation, a high number of patients with ARDS secondary to COVID-19 developed life-threatening thrombotic complications, and higher antICOagulation targets than in usual critically ill patients should therefore probably be suggested.
Abstract: Little evidence of increased thrombotic risk is available in COVID-19 patients. Our purpose was to assess thrombotic risk in severe forms of SARS-CoV-2 infection. All patients referred to 4 intensive care units (ICUs) from two centers of a French tertiary hospital for acute respiratory distress syndrome (ARDS) due to COVID-19 between March 3rd and 31st 2020 were included. Medical history, symptoms, biological data and imaging were prospectively collected. Propensity score matching was performed to analyze the occurrence of thromboembolic events between non-COVID-19 ARDS and COVID-19 ARDS patients. 150 COVID-19 patients were included (122 men, median age 63 [53; 71] years, SAPSII 49 [37; 64] points). Sixty-four clinically relevant thrombotic complications were diagnosed in 150 patients, mainly pulmonary embolisms (16.7%). 28/29 patients (96.6%) receiving continuous renal replacement therapy experienced circuit clotting. Three thrombotic occlusions (in 2 patients) of centrifugal pump occurred in 12 patients (8%) supported by ECMO. Most patients (> 95%) had elevated D-dimer and fibrinogen. No patient developed disseminated intravascular coagulation. Von Willebrand (vWF) activity, vWF antigen and FVIII were considerably increased, and 50/57 tested patients (87.7%) had positive lupus anticoagulant. Comparison with non-COVID-19 ARDS patients (n = 145) confirmed that COVID-19 ARDS patients (n = 77) developed significantly more thrombotic complications, mainly pulmonary embolisms (11.7 vs. 2.1%, p < 0.008). Coagulation parameters significantly differed between the two groups. Despite anticoagulation, a high number of patients with ARDS secondary to COVID-19 developed life-threatening thrombotic complications. Higher anticoagulation targets than in usual critically ill patients should therefore probably be suggested.

2,147 citations


Journal ArticleDOI
TL;DR: Dabrafenib plus trametinib, as compared with vemurafenib monotherapy, significantly improved overall survival in previously untreated patients with metastatic melanoma with BRAF V600E or V600K mutations, without increased overall toxicity.
Abstract: Background The BRAF inhibitors vemurafenib and dabrafenib have shown efficacy as monotherapies in patients with previously untreated metastatic melanoma with BRAF V600E or V600K mutations. Combining dabrafenib and the MEK inhibitor trametinib, as compared with dabrafenib alone, enhanced antitumor activity in this population of patients. Methods In this open-label, phase 3 trial, we randomly assigned 704 patients with metastatic melanoma with a BRAF V600 mutation to receive either a combination of dabrafenib (150 mg twice daily) and trametinib (2 mg once daily) or vemurafenib (960 mg twice daily) orally as first-line therapy. The primary end point was overall survival. Results At the preplanned interim overall survival analysis, which was performed after 77% of the total number of expected events occurred, the overall survival rate at 12 months was 72% (95% confidence interval [CI], 67 to 77) in the combination-therapy group and 65% (95% CI, 59 to 70) in the vemurafenib group (hazard ratio for death in the combination-therapy group, 0.69; 95% CI, 0.53 to 0.89; P = 0.005). The prespecified interim stopping boundary was crossed, and the study was stopped for efficacy in July 2014. Median progression-free survival was 11.4 months in the combinationtherapy group and 7.3 months in the vemurafenib group (hazard ratio, 0.56; 95% CI, 0.46 to 0.69; P<0.001). The objective response rate was 64% in the combinationtherapy group and 51% in the vemurafenib group (P<0.001). Rates of severe adverse events and study-drug discontinuations were similar in the two groups. Cutaneous squamous-cell carcinoma and keratoacanthoma occurred in 1% of patients in the combination-therapy group and 18% of those in the vemurafenib group. Conclusions Dabrafenib plus trametinib, as compared with vemurafenib monotherapy, significantly improved overall survival in previously untreated patients with metastatic melanoma with BRAF V600E or V600K mutations, without increased overall toxicity. (Funded by GlaxoSmithKline; ClinicalTrials.gov number, NCT01597908.)

2,144 citations


Journal ArticleDOI
23 Jun 2021
TL;DR: In this article, the authors describe the state-of-the-art in the field of federated learning from the perspective of distributed optimization, cryptography, security, differential privacy, fairness, compressed sensing, systems, information theory, and statistics.
Abstract: The term Federated Learning was coined as recently as 2016 to describe a machine learning setting where multiple entities collaborate in solving a machine learning problem, under the coordination of a central server or service provider. Each client’s raw data is stored locally and not exchanged or transferred; instead, focused updates intended for immediate aggregation are used to achieve the learning objective. Since then, the topic has gathered much interest across many different disciplines and the realization that solving many of these interdisciplinary problems likely requires not just machine learning but techniques from distributed optimization, cryptography, security, differential privacy, fairness, compressed sensing, systems, information theory, statistics, and more. This monograph has contributions from leading experts across the disciplines, who describe the latest state-of-the art from their perspective. These contributions have been carefully curated into a comprehensive treatment that enables the reader to understand the work that has been done and get pointers to where effort is required to solve many of the problems before Federated Learning can become a reality in practical systems. Researchers working in the area of distributed systems will find this monograph an enlightening read that may inspire them to work on the many challenging issues that are outlined. This monograph will get the reader up to speed quickly and easily on what is likely to become an increasingly important topic: Federated Learning.

2,144 citations


Journal ArticleDOI
12 May 2020-JAMA
TL;DR: The COVID-19 pandemic represents the greatest global public health crisis of this generation and, potentially, since the pandemic influenza outbreak of 1918 and both the need and capability to produce high-quality evidence even in the middle of a pandemic.
Abstract: Importance The pandemic of coronavirus disease 2019 (COVID-19) caused by the novel severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) presents an unprecedented challenge to identify effective drugs for prevention and treatment. Given the rapid pace of scientific discovery and clinical data generated by the large number of people rapidly infected by SARS-CoV-2, clinicians need accurate evidence regarding effective medical treatments for this infection. Observations No proven effective therapies for this virus currently exist. The rapidly expanding knowledge regarding SARS-CoV-2 virology provides a significant number of potential drug targets. The most promising therapy is remdesivir. Remdesivir has potent in vitro activity against SARS-CoV-2, but it is not US Food and Drug Administration approved and currently is being tested in ongoing randomized trials. Oseltamivir has not been shown to have efficacy, and corticosteroids are currently not recommended. Current clinical evidence does not support stopping angiotensin-converting enzyme inhibitors or angiotensin receptor blockers in patients with COVID-19. Conclusions and Relevance The COVID-19 pandemic represents the greatest global public health crisis of this generation and, potentially, since the pandemic influenza outbreak of 1918. The speed and volume of clinical trials launched to investigate potential therapies for COVID-19 highlight both the need and capability to produce high-quality evidence even in the middle of a pandemic. No therapies have been shown effective to date.

2,143 citations


Journal ArticleDOI
TL;DR: The T cell–inflamed GEP contained IFN-&ggr;–responsive genes related to antigen presentation, chemokine expression, cytotoxic activity, and adaptive immune resistance, and these features were necessary, but not always sufficient, for clinical benefit.
Abstract: Programmed death-1-directed (PD-1-directed) immune checkpoint blockade results in durable antitumor activity in many advanced malignancies. Recent studies suggest that IFN-γ is a critical driver of programmed death ligand-1 (PD-L1) expression in cancer and host cells, and baseline intratumoral T cell infiltration may improve response likelihood to anti-PD-1 therapies, including pembrolizumab. However, whether quantifying T cell-inflamed microenvironment is a useful pan-tumor determinant of PD-1-directed therapy response has not been rigorously evaluated. Here, we analyzed gene expression profiles (GEPs) using RNA from baseline tumor samples of pembrolizumab-treated patients. We identified immune-related signatures correlating with clinical benefit using a learn-and-confirm paradigm based on data from different clinical studies of pembrolizumab, starting with a small pilot of 19 melanoma patients and eventually defining a pan-tumor T cell-inflamed GEP in 220 patients with 9 cancers. Predictive value was independently confirmed and compared with that of PD-L1 immunohistochemistry in 96 patients with head and neck squamous cell carcinoma. The T cell-inflamed GEP contained IFN-γ-responsive genes related to antigen presentation, chemokine expression, cytotoxic activity, and adaptive immune resistance, and these features were necessary, but not always sufficient, for clinical benefit. The T cell-inflamed GEP has been developed into a clinical-grade assay that is currently being evaluated in ongoing pembrolizumab trials.

2,142 citations


Journal ArticleDOI
TL;DR: The definition, characteristics, and classification of big data along with some discussions on cloud computing are introduced, and research challenges are investigated, with focus on scalability, availability, data integrity, data transformation, data quality, data heterogeneity, privacy, legal and regulatory issues, and governance.

2,141 citations


Journal ArticleDOI
TL;DR: In this retrospective case series, chest CT scans of 21 symptomatic patients from China infected with the 2019 novel coronavirus were reviewed, with emphasis on identifying and characterizing the most common findings.
Abstract: In this retrospective case series, chest CT scans of 21 symptomatic patients from China infected with the 2019 novel coronavirus (2019-nCoV) were reviewed, with emphasis on identifying and characterizing the most common findings. Typical CT findings included bilateral pulmonary parenchymal ground-glass and consolidative pulmonary opacities, sometimes with a rounded morphology and a peripheral lung distribution. Notably, lung cavitation, discrete pulmonary nodules, pleural effusions, and lymphadenopathy were absent. Follow-up imaging in a subset of patients during the study time window often demonstrated mild or moderate progression of disease, as manifested by increasing extent and density of lung opacities.

2,141 citations


Journal ArticleDOI
TL;DR: The current status, strengths, and weaknesses of peptides as medicines and the emerging new opportunities in peptide drug design and development are discussed.

2,136 citations


Journal ArticleDOI
TL;DR: This systematic review and meta-analysis of existing research works and findings in relation to the prevalence of stress, anxiety and depression in the general population during the COVID-19 pandemic found that it is essential to preserve the mental health of individuals and to develop psychological interventions that can improve themental health of vulnerable groups during the pandemic.
Abstract: The COVID-19 pandemic has had a significant impact on public mental health Therefore, monitoring and oversight of the population mental health during crises such as a panedmic is an immediate priority The aim of this study is to analyze the existing research works and findings in relation to the prevalence of stress, anxiety and depression in the general population during the COVID-19 pandemic In this systematic review and meta-analysis, articles that have focused on stress and anxiety prevalence among the general population during the COVID-19 pandemic were searched in the Science Direct, Embase, Scopus, PubMed, Web of Science (ISI) and Google Scholar databases, without a lower time limit and until May 2020 In order to perform a meta-analysis of the collected studies, the random effects model was used, and the heterogeneity of studies was investigated using the I2 index Moreover data analysis was conducted using the Comprehensive Meta-Analysis (CMA) software The prevalence of stress in 5 studies with a total sample size of 9074 is obtained as 296% (95% confidence limit: 243–354), the prevalence of anxiety in 17 studies with a sample size of 63,439 as 319% (95% confidence interval: 275–367), and the prevalence of depression in 14 studies with a sample size of 44,531 people as 337% (95% confidence interval: 275–406) COVID-19 not only causes physical health concerns but also results in a number of psychological disorders The spread of the new coronavirus can impact the mental health of people in different communities Thus, it is essential to preserve the mental health of individuals and to develop psychological interventions that can improve the mental health of vulnerable groups during the COVID-19 pandemic

Proceedings Article
Bishan Yang1, Wen-tau Yih2, Xiaodong He2, Jianfeng Gao2, Li Deng2 
01 May 2015
TL;DR: It is found that embeddings learned from the bilinear objective are particularly good at capturing relational semantics and that the composition of relations is characterized by matrix multiplication.
Abstract: We consider learning representations of entities and relations in KBs using the neural-embedding approach. We show that most existing models, including NTN (Socher et al., 2013) and TransE (Bordes et al., 2013b), can be generalized under a unified learning framework, where entities are low-dimensional vectors learned from a neural network and relations are bilinear and/or linear mapping functions. Under this framework, we compare a variety of embedding models on the link prediction task. We show that a simple bilinear formulation achieves new state-of-the-art results for the task (achieving a top-10 accuracy of 73.2% vs. 54.7% by TransE on Freebase). Furthermore, we introduce a novel approach that utilizes the learned relation embeddings to mine logical rules such as "BornInCity(a,b) and CityInCountry(b,c) => Nationality(a,c)". We find that embeddings learned from the bilinear objective are particularly good at capturing relational semantics and that the composition of relations is characterized by matrix multiplication. More interestingly, we demonstrate that our embedding-based rule extraction approach successfully outperforms a state-of-the-art confidence-based rule mining approach in mining Horn rules that involve compositional reasoning.

Journal ArticleDOI
TL;DR: Telemedicine for Covid-19’s payment and regulatory structures, licensing, credentialing, and implementation take time to work through, but health systems that have a...
Abstract: Virtually Perfect? Telemedicine for Covid-19 Telemedicine’s payment and regulatory structures, licensing, credentialing, and implementation take time to work through, but health systems that have a...

Proceedings ArticleDOI
22 May 2016
TL;DR: In this article, the authors introduce a defensive mechanism called defensive distillation to reduce the effectiveness of adversarial samples on DNNs, which increases the average minimum number of features that need to be modified to create adversarial examples by about 800%.
Abstract: Deep learning algorithms have been shown to perform extremely well on manyclassical machine learning problems. However, recent studies have shown thatdeep learning, like other machine learning techniques, is vulnerable to adversarial samples: inputs crafted to force adeep neural network (DNN) to provide adversary-selected outputs. Such attackscan seriously undermine the security of the system supported by the DNN, sometimes with devastating consequences. For example, autonomous vehicles canbe crashed, illicit or illegal content can bypass content filters, or biometricauthentication systems can be manipulated to allow improper access. In thiswork, we introduce a defensive mechanism called defensive distillationto reduce the effectiveness of adversarial samples on DNNs. We analyticallyinvestigate the generalizability and robustness properties granted by the useof defensive distillation when training DNNs. We also empirically study theeffectiveness of our defense mechanisms on two DNNs placed in adversarialsettings. The study shows that defensive distillation can reduce effectivenessof sample creation from 95% to less than 0.5% on a studied DNN. Such dramaticgains can be explained by the fact that distillation leads gradients used inadversarial sample creation to be reduced by a factor of 1030. We alsofind that distillation increases the average minimum number of features thatneed to be modified to create adversarial samples by about 800% on one of theDNNs we tested.

Proceedings ArticleDOI
27 Jun 2016
TL;DR: This paper generates a synthetic collection of diverse urban images, named SYNTHIA, with automatically generated class annotations, and conducts experiments with DCNNs that show how the inclusion of SYnTHIA in the training stage significantly improves performance on the semantic segmentation task.
Abstract: Vision-based semantic segmentation in urban scenarios is a key functionality for autonomous driving. Recent revolutionary results of deep convolutional neural networks (DCNNs) foreshadow the advent of reliable classifiers to perform such visual tasks. However, DCNNs require learning of many parameters from raw images, thus, having a sufficient amount of diverse images with class annotations is needed. These annotations are obtained via cumbersome, human labour which is particularly challenging for semantic segmentation since pixel-level annotations are required. In this paper, we propose to use a virtual world to automatically generate realistic synthetic images with pixel-level annotations. Then, we address the question of how useful such data can be for semantic segmentation – in particular, when using a DCNN paradigm. In order to answer this question we have generated a synthetic collection of diverse urban images, named SYNTHIA, with automatically generated class annotations. We use SYNTHIA in combination with publicly available real-world urban images with manually provided annotations. Then, we conduct experiments with DCNNs that show how the inclusion of SYNTHIA in the training stage significantly improves performance on the semantic segmentation task.

Journal ArticleDOI
TL;DR: A metasurface platform based on high-contrast dielectric elliptical nanoposts that provides complete control of polarization and phase with subwavelength spatial resolution and an experimentally measured efficiency ranging from 72% to 97%, depending on the exact design.
Abstract: Metasurfaces are planar structures that locally modify the polarization, phase and amplitude of light in reflection or transmission, thus enabling lithographically patterned flat optical components with functionalities controlled by design. Transmissive metasurfaces are especially important, as most optical systems used in practice operate in transmission. Several types of transmissive metasurface have been realized, but with either low transmission efficiencies or limited control over polarization and phase. Here, we show a metasurface platform based on high-contrast dielectric elliptical nanoposts that provides complete control of polarization and phase with subwavelength spatial resolution and an experimentally measured efficiency ranging from 72% to 97%, depending on the exact design. Such complete control enables the realization of most free-space transmissive optical elements such as lenses, phase plates, wave plates, polarizers, beamsplitters, as well as polarization-switchable phase holograms and arbitrary vector beam generators using the same metamaterial platform.

Journal ArticleDOI
TL;DR: In this article, a review of the classification schemes of both fully gapped and gapless topological materials is presented, and a pedagogical introduction to the field of topological band theory is given.
Abstract: In recent years an increasing amount of attention has been devoted to quantum materials with topological characteristics that are robust against disorder and other perturbations. In this context it was discovered that topological materials can be classified with respect to their dimension and symmetry properties. This review provides an overview of the classification schemes of both fully gapped and gapless topological materials and gives a pedagogical introduction into the field of topological band theory.

Journal ArticleDOI
TL;DR: In this article, a plant-inspired shape morphing system is presented, where a composite hydrogel architecture is encoded with localized, anisotropic swelling behavior controlled by the alignment of cellulose fibrils along prescribed four-dimensional printing pathways.
Abstract: Shape-morphing systems can be found in many areas, including smart textiles, autonomous robotics, biomedical devices, drug delivery and tissue engineering. The natural analogues of such systems are exemplified by nastic plant motions, where a variety of organs such as tendrils, bracts, leaves and flowers respond to environmental stimuli (such as humidity, light or touch) by varying internal turgor, which leads to dynamic conformations governed by the tissue composition and microstructural anisotropy of cell walls. Inspired by these botanical systems, we printed composite hydrogel architectures that are encoded with localized, anisotropic swelling behaviour controlled by the alignment of cellulose fibrils along prescribed four-dimensional printing pathways. When combined with a minimal theoretical framework that allows us to solve the inverse problem of designing the alignment patterns for prescribed target shapes, we can programmably fabricate plant-inspired architectures that change shape on immersion in water, yielding complex three-dimensional morphologies.

Journal ArticleDOI
TL;DR: During the first 3 weeks of the Covid-19 outbreak in the Seattle area, the most common reasons for admission to the ICU were hypoxemic respiratory failure leading to mechanical ventilation, hypotension requiring vasopressor treatment, or both.
Abstract: Background Community transmission of coronavirus 2019 (Covid-19) was detected in the state of Washington in February 2020. Methods We identified patients from nine Seattle-area hospitals w...

Journal ArticleDOI
TL;DR: Non-fullerene OSCs show great tunability in absorption spectra and electron energy levels, providing a wide range of new opportunities, and this Review highlights these opportunities made possible by NF acceptors.
Abstract: Organic solar cells (OSCs) have been dominated by donor:acceptor blends based on fullerene acceptors for over two decades. This situation has changed recently, with non-fullerene (NF) OSCs developing very quickly. The power conversion efficiencies of NF OSCs have now reached a value of over 13%, which is higher than the best fullerene-based OSCs. NF acceptors show great tunability in absorption spectra and electron energy levels, providing a wide range of new opportunities. The coexistence of low voltage losses and high current generation indicates that new regimes of device physics and photophysics are reached in these systems. This Review highlights these opportunities made possible by NF acceptors, and also discuss the challenges facing the development of NF OSCs for practical applications.

Journal ArticleDOI
12 Dec 2017-JAMA
TL;DR: In the setting of a challenge competition, some deep learning algorithms achieved better diagnostic performance than a panel of 11 pathologists participating in a simulation exercise designed to mimic routine pathology workflow; algorithm performance was comparable with an expert pathologist interpreting whole-slide images without time constraints.
Abstract: Importance Application of deep learning algorithms to whole-slide pathology images can potentially improve diagnostic accuracy and efficiency. Objective Assess the performance of automated deep learning algorithms at detecting metastases in hematoxylin and eosin–stained tissue sections of lymph nodes of women with breast cancer and compare it with pathologists’ diagnoses in a diagnostic setting. Design, Setting, and Participants Researcher challenge competition (CAMELYON16) to develop automated solutions for detecting lymph node metastases (November 2015-November 2016). A training data set of whole-slide images from 2 centers in the Netherlands with (n = 110) and without (n = 160) nodal metastases verified by immunohistochemical staining were provided to challenge participants to build algorithms. Algorithm performance was evaluated in an independent test set of 129 whole-slide images (49 with and 80 without metastases). The same test set of corresponding glass slides was also evaluated by a panel of 11 pathologists with time constraint (WTC) from the Netherlands to ascertain likelihood of nodal metastases for each slide in a flexible 2-hour session, simulating routine pathology workflow, and by 1 pathologist without time constraint (WOTC). Exposures Deep learning algorithms submitted as part of a challenge competition or pathologist interpretation. Main Outcomes and Measures The presence of specific metastatic foci and the absence vs presence of lymph node metastasis in a slide or image using receiver operating characteristic curve analysis. The 11 pathologists participating in the simulation exercise rated their diagnostic confidence as definitely normal, probably normal, equivocal, probably tumor, or definitely tumor. Results The area under the receiver operating characteristic curve (AUC) for the algorithms ranged from 0.556 to 0.994. The top-performing algorithm achieved a lesion-level, true-positive fraction comparable with that of the pathologist WOTC (72.4% [95% CI, 64.3%-80.4%]) at a mean of 0.0125 false-positives per normal whole-slide image. For the whole-slide image classification task, the best algorithm (AUC, 0.994 [95% CI, 0.983-0.999]) performed significantly better than the pathologists WTC in a diagnostic simulation (mean AUC, 0.810 [range, 0.738-0.884];P Conclusions and Relevance In the setting of a challenge competition, some deep learning algorithms achieved better diagnostic performance than a panel of 11 pathologists participating in a simulation exercise designed to mimic routine pathology workflow; algorithm performance was comparable with an expert pathologist interpreting whole-slide images without time constraints. Whether this approach has clinical utility will require evaluation in a clinical setting.

Journal ArticleDOI
28 Oct 2015-BMJ
TL;DR: STARD 2015 is presented, an updated list of 30 essential items that should be included in every report of a diagnostic accuracy study, which incorporates recent evidence about sources of bias and variability in diagnostic accuracy.
Abstract: Incomplete reporting has been identified as a major source of avoidable waste in biomedical research. Essential information is often not provided in study reports, impeding the identification, critical appraisal, and replication of studies. To improve the quality of reporting of diagnostic accuracy studies, the Standards for Reporting Diagnostic Accuracy (STARD) statement was developed. Here we present STARD 2015, an updated list of 30 essential items that should be included in every report of a diagnostic accuracy study. This update incorporates recent evidence about sources of bias and variability in diagnostic accuracy and is intended to facilitate the use of STARD. As such, STARD 2015 may help to improve completeness and transparency in reporting of diagnostic accuracy studies.

Journal ArticleDOI
Emily Oster1
TL;DR: This article developed an extension of the theory that connects bias explicitly to coefficient stability and showed that it is necessary to take into account coefficient and R-squared movements, and showed two validation exercises and discuss application to the economics literature.
Abstract: A common approach to evaluating robustness to omitted variable bias is to observe coefficient movements after inclusion of controls. This is informative only if selection on observables is informative about selection on unobservables. Although this link is known in theory in existing literature, very few empirical articles approach this formally. I develop an extension of the theory that connects bias explicitly to coefficient stability. I show that it is necessary to take into account coefficient and R-squared movements. I develop a formal bounding argument. I show two validation exercises and discuss application to the economics literature. Supplementary materials for this article are available online.

Journal ArticleDOI
TL;DR: A suite of ligand annotation, purchasability, target, and biology association tools, incorporated into ZINC and meant for investigators who are not computer specialists, offer new analysis tools that are easy for nonspecialists yet with few limitations for experts.
Abstract: Many questions about the biological activity and availability of small molecules remain inaccessible to investigators who could most benefit from their answers. To narrow the gap between chemoinformatics and biology, we have developed a suite of ligand annotation, purchasability, target, and biology association tools, incorporated into ZINC and meant for investigators who are not computer specialists. The new version contains over 120 million purchasable “drug-like” compounds – effectively all organic molecules that are for sale – a quarter of which are available for immediate delivery. ZINC connects purchasable compounds to high-value ones such as metabolites, drugs, natural products, and annotated compounds from the literature. Compounds may be accessed by the genes for which they are annotated as well as the major and minor target classes to which those genes belong. It offers new analysis tools that are easy for nonspecialists yet with few limitations for experts. ZINC retains its original 3D roots – ...


Journal ArticleDOI
TL;DR: This poster presents a probabilistic procedure to determine the best method for selecting a single drug to treat atrial fibrillation-like symptoms in patients with a history of atrialfibrillation.
Abstract: 2015 ESC Guidelines for the Management of Patients With Ventricular Arrhythmias and the Prevention of Sudden Cardiac Death

Journal ArticleDOI
Adam Abeshouse1, Jaeil Ahn1, Rehan Akbani1, Adrian Ally1  +308 moreInstitutions (1)
05 Nov 2015-Cell
TL;DR: The Cancer Genome Atlas (TCGA) has been used for a comprehensive molecular analysis of primary prostate carcinomas as discussed by the authors, revealing substantial heterogeneity among primary prostate cancers, evident in the spectrum of molecular abnormalities and its variable clinical course.

Journal ArticleDOI
09 Mar 2018-Science
TL;DR: The rise of fake news highlights the erosion of long-standing institutional bulwarks against misinformation in the internet age as discussed by the authors. But much remains unknown regarding the vulnerabilities of individuals, institutions, and society to manipulations by malicious actors.
Abstract: The rise of fake news highlights the erosion of long-standing institutional bulwarks against misinformation in the internet age. Concern over the problem is global. However, much remains unknown regarding the vulnerabilities of individuals, institutions, and society to manipulations by malicious actors. A new system of safeguards is needed. Below, we discuss extant social and computer science research regarding belief in fake news and the mechanisms by which it spreads. Fake news has a long history, but we focus on unanswered scientific questions raised by the proliferation of its most recent, politically oriented incarnation. Beyond selected references in the text, suggested further reading can be found in the supplementary materials.

Posted Content
TL;DR: Self-Attention Generative Adversarial Network (SAGAN) as mentioned in this paper uses attention-driven, long-range dependency modeling for image generation tasks and achieves state-of-the-art results.
Abstract: In this paper, we propose the Self-Attention Generative Adversarial Network (SAGAN) which allows attention-driven, long-range dependency modeling for image generation tasks. Traditional convolutional GANs generate high-resolution details as a function of only spatially local points in lower-resolution feature maps. In SAGAN, details can be generated using cues from all feature locations. Moreover, the discriminator can check that highly detailed features in distant portions of the image are consistent with each other. Furthermore, recent work has shown that generator conditioning affects GAN performance. Leveraging this insight, we apply spectral normalization to the GAN generator and find that this improves training dynamics. The proposed SAGAN achieves the state-of-the-art results, boosting the best published Inception score from 36.8 to 52.52 and reducing Frechet Inception distance from 27.62 to 18.65 on the challenging ImageNet dataset. Visualization of the attention layers shows that the generator leverages neighborhoods that correspond to object shapes rather than local regions of fixed shape.

Proceedings ArticleDOI
07 Dec 2015
TL;DR: The authors align books to their movie releases to provide rich descriptive explanations for visual content that go semantically far beyond the captions available in the current datasets, and propose a context-aware CNN to combine information from multiple sources.
Abstract: Books are a rich source of both fine-grained information, how a character, an object or a scene looks like, as well as high-level semantics, what someone is thinking, feeling and how these states evolve through a story. This paper aims to align books to their movie releases in order to provide rich descriptive explanations for visual content that go semantically far beyond the captions available in the current datasets. To align movies and books we propose a neural sentence embedding that is trained in an unsupervised way from a large corpus of books, as well as a video-text neural embedding for computing similarities between movie clips and sentences in the book. We propose a context-aware CNN to combine information from multiple sources. We demonstrate good quantitative performance for movie/book alignment and show several qualitative examples that showcase the diversity of tasks our model can be used for.