scispace - formally typeset
Search or ask a question
Browse all papers

Journal ArticleDOI
TL;DR: Among patients with a germline BRCA mutation and metastatic pancreatic cancer, progression-free survival was longer with maintenance olaparib than with placebo, and there was no significant between-group difference in health-related quality of life.
Abstract: Background Patients with a germline BRCA1 or BRCA2 mutation make up a small subgroup of those with metastatic pancreatic cancer. The poly(adenosine diphosphate–ribose) polymerase (PARP) in...

1,321 citations


Journal ArticleDOI
TL;DR: In this article, a hybrid bidirectional LSTM and CNN architecture was proposed to automatically detect word and character-level features, eliminating the need for feature engineering and lexicons to achieve high performance.
Abstract: Named entity recognition is a challenging task that has traditionally required large amounts of knowledge in the form of feature engineering and lexicons to achieve high performance. In this paper, we present a novel neural network architecture that automatically detects word- and character-level features using a hybrid bidirectional LSTM and CNN architecture, eliminating the need for most feature engineering. We also propose a novel method of encoding partial lexicon matches in neural networks and compare it to existing approaches. Extensive evaluation shows that, given only tokenized text and publicly available word embeddings, our system is competitive on the CoNLL-2003 dataset and surpasses the previously reported state of the art performance on the OntoNotes 5.0 dataset by 2.13 F1 points. By using two lexicons constructed from publicly-available sources, we establish new state of the art performance with an F1 score of 91.62 on CoNLL-2003 and 86.28 on OntoNotes, surpassing systems that employ heavy feature engineering, proprietary lexicons, and rich entity linking information.

1,321 citations


Proceedings ArticleDOI
27 Jun 2016
TL;DR: This paper argues that identification of structural elements in indoor spaces is essentially a detection problem, rather than segmentation which is commonly used, and proposes a method for semantic parsing the 3D point cloud of an entire building using a hierarchical approach.
Abstract: In this paper, we propose a method for semantic parsing the 3D point cloud of an entire building using a hierarchical approach: first, the raw data is parsed into semantically meaningful spaces (e.g. rooms, etc) that are aligned into a canonical reference coordinate system. Second, the spaces are parsed into their structural and building elements (e.g. walls, columns, etc). Performing these with a strong notation of global 3D space is the backbone of our method. The alignment in the first step injects strong 3D priors from the canonical coordinate system into the second step for discovering elements. This allows diverse challenging scenarios as man-made indoor spaces often show recurrent geometric patterns while the appearance features can change drastically. We also argue that identification of structural elements in indoor spaces is essentially a detection problem, rather than segmentation which is commonly used. We evaluated our method on a new dataset of several buildings with a covered area of over 6, 000m2 and over 215 million points, demonstrating robust results readily useful for practical applications.

1,320 citations


Book ChapterDOI
08 Oct 2016
TL;DR: In this paper, the authors present an approach to rapidly create pixel-accurate semantic label maps for images extracted from modern computer games, which enables rapid propagation of semantic labels within and across images synthesized by the game, without access to the source code or the content.
Abstract: Recent progress in computer vision has been driven by high-capacity models trained on large datasets. Unfortunately, creating large datasets with pixel-level labels has been extremely costly due to the amount of human effort required. In this paper, we present an approach to rapidly creating pixel-accurate semantic label maps for images extracted from modern computer games. Although the source code and the internal operation of commercial games are inaccessible, we show that associations between image patches can be reconstructed from the communication between the game and the graphics hardware. This enables rapid propagation of semantic labels within and across images synthesized by the game, with no access to the source code or the content. We validate the presented approach by producing dense pixel-level semantic annotations for 25 thousand images synthesized by a photorealistic open-world computer game. Experiments on semantic segmentation datasets show that using the acquired data to supplement real-world images significantly increases accuracy and that the acquired data enables reducing the amount of hand-labeled real-world data: models trained with game data and just \(\tfrac{1}{3}\) of the CamVid training set outperform models trained on the complete CamVid training set.

1,319 citations


Journal ArticleDOI
TL;DR: Diagonally weighted least squares was less biased and more accurate than MLR in estimating the factor loadings across nearly every condition and the proposed model tended to be over-rejected by chi-square test statistics under both MLR and WLSMV in the condition of small sample size N = 200.
Abstract: In confirmatory factor analysis (CFA), the use of maximum likelihood (ML) assumes that the observed indicators follow a continuous and multivariate normal distribution, which is not appropriate for ordinal observed variables. Robust ML (MLR) has been introduced into CFA models when this normality assumption is slightly or moderately violated. Diagonally weighted least squares (WLSMV), on the other hand, is specifically designed for ordinal data. Although WLSMV makes no distributional assumptions about the observed variables, a normal latent distribution underlying each observed categorical variable is instead assumed. A Monte Carlo simulation was carried out to compare the effects of different configurations of latent response distributions, numbers of categories, and sample sizes on model parameter estimates, standard errors, and chi-square test statistics in a correlated two-factor model. The results showed that WLSMV was less biased and more accurate than MLR in estimating the factor loadings across nearly every condition. However, WLSMV yielded moderate overestimation of the interfactor correlations when the sample size was small or/and when the latent distributions were moderately nonnormal. With respect to standard error estimates of the factor loadings and the interfactor correlations, MLR outperformed WLSMV when the latent distributions were nonnormal with a small sample size of N = 200. Finally, the proposed model tended to be over-rejected by chi-square test statistics under both MLR and WLSMV in the condition of small sample size N = 200.

1,319 citations



Journal ArticleDOI
TL;DR: In this article, a gate-tunable superconducting and correlated insulating phase diagram for bilayer graphene was presented. But the authors only considered the twisted bilayer, and the interlayer coupling can also be modified to precisely tune these phases.
Abstract: Materials with flat electronic bands often exhibit exotic quantum phenomena owing to strong correlations. Remarkably, an isolated low-energy flat band can be induced in bilayer graphene by simply rotating the layers to 1.1$^{\circ}$, resulting in the appearance of gate-tunable superconducting and correlated insulating phases. Here, we demonstrate that in addition to the twist angle, the interlayer coupling can also be modified to precisely tune these phases. We establish the capability to induce superconductivity at a twist angle larger than 1.1$^{\circ}$ $-$ in which correlated phases are otherwise absent $-$ by varying the interlayer spacing with hydrostatic pressure. Realizing devices with low disorder additionally reveals new details about the superconducting phase diagram and its relationship to the nearby insulator. Our results demonstrate twisted bilayer graphene to be a uniquely tunable platform for exploring novel correlated states.

1,318 citations


Journal ArticleDOI
TL;DR: Proposed mechanisms of antibacterial action of different metal NPs include the production of reactive oxygen species, cation release, biomolecule damages, ATP depletion, and membrane interaction.
Abstract: As the field of nanomedicine emerges, there is a lag in research surrounding the topic of nanoparticle (NP) toxicity, particularly concerned with mechanisms of action. The continuous emergence of bacterial resistance has challenged the research community to develop novel antibiotic agents. Metal NPs are among the most promising of these because show strong antibacterial activity. This review summarizes and discusses proposed mechanisms of antibacterial action of different metal NPs. These mechanisms of bacterial killing include the production of reactive oxygen species, cation release, biomolecule damages, ATP depletion, and membrane interaction. Finally, a comprehensive analysis of the effects of NPs on the regulation of genes and proteins (transcriptomic and proteomic) profiles is discussed.

1,318 citations


Proceedings Article
07 Dec 2015
TL;DR: This work introduces a robust new AutoML system based on scikit-learn, which improves on existing AutoML methods by automatically taking into account past performance on similar datasets, and by constructing ensembles from the models evaluated during the optimization.
Abstract: The success of machine learning in a broad range of applications has led to an ever-growing demand for machine learning systems that can be used off the shelf by non-experts. To be effective in practice, such systems need to automatically choose a good algorithm and feature preprocessing steps for a new dataset at hand, and also set their respective hyperparameters. Recent work has started to tackle this automated machine learning (AutoML) problem with the help of efficient Bayesian optimization methods. Building on this, we introduce a robust new AutoML system based on scikit-learn (using 15 classifiers, 14 feature preprocessing methods, and 4 data preprocessing methods, giving rise to a structured hypothesis space with 110 hyperparameters). This system, which we dub AUTO-SKLEARN, improves on existing AutoML methods by automatically taking into account past performance on similar datasets, and by constructing ensembles from the models evaluated during the optimization. Our system won the first phase of the ongoing ChaLearn AutoML challenge, and our comprehensive analysis on over 100 diverse datasets shows that it substantially outperforms the previous state of the art in AutoML. We also demonstrate the performance gains due to each of our contributions and derive insights into the effectiveness of the individual components of AUTO-SKLEARN.

1,318 citations


Posted Content
TL;DR: This paper quantifies the approximate financial and environmental costs of training a variety of recently successful neural network models for NLP and proposes actionable recommendations to reduce costs and improve equity in NLP research and practice.
Abstract: Recent progress in hardware and methodology for training neural networks has ushered in a new generation of large networks trained on abundant data. These models have obtained notable gains in accuracy across many NLP tasks. However, these accuracy improvements depend on the availability of exceptionally large computational resources that necessitate similarly substantial energy consumption. As a result these models are costly to train and develop, both financially, due to the cost of hardware and electricity or cloud compute time, and environmentally, due to the carbon footprint required to fuel modern tensor processing hardware. In this paper we bring this issue to the attention of NLP researchers by quantifying the approximate financial and environmental costs of training a variety of recently successful neural network models for NLP. Based on these findings, we propose actionable recommendations to reduce costs and improve equity in NLP research and practice.

1,318 citations


Journal ArticleDOI
TL;DR: Dupilumab improved the signs and symptoms of atopic dermatitis, including pruritus, symptoms of anxiety and depression, and quality of life, as compared with placebo in two phase 3 trials of identical design.
Abstract: BackgroundDupilumab, a human monoclonal antibody against interleukin-4 receptor alpha, inhibits signaling of interleukin-4 and interleukin-13, type 2 cytokines that may be important drivers of atopic or allergic diseases such as atopic dermatitis. MethodsIn two randomized, placebo-controlled, phase 3 trials of identical design (SOLO 1 and SOLO 2), we enrolled adults with moderate-to-severe atopic dermatitis whose disease was inadequately controlled by topical treatment. Patients were randomly assigned in a 1:1:1 ratio to receive, for 16 weeks, subcutaneous dupilumab (300 mg) or placebo weekly or the same dose of dupilumab every other week alternating with placebo. The primary outcome was the proportion of patients who had both a score of 0 or 1 (clear or almost clear) on the Investigator’s Global Assessment and a reduction of 2 points or more in that score from baseline at week 16. ResultsWe enrolled 671 patients in SOLO 1 and 708 in SOLO 2. In SOLO 1, the primary outcome occurred in 85 patients (38%) who...

Proceedings ArticleDOI
19 May 2019
TL;DR: Spectre as mentioned in this paper is a side channel attack that can leak the victim's confidential information via side channel to the adversary. And it can read arbitrary memory from a victim's process.
Abstract: Modern processors use branch prediction and speculative execution to maximize performance. For example, if the destination of a branch depends on a memory value that is in the process of being read, CPUs will try to guess the destination and attempt to execute ahead. When the memory value finally arrives, the CPU either discards or commits the speculative computation. Speculative logic is unfaithful in how it executes, can access the victim's memory and registers, and can perform operations with measurable side effects. Spectre attacks involve inducing a victim to speculatively perform operations that would not occur during correct program execution and which leak the victim's confidential information via a side channel to the adversary. This paper describes practical attacks that combine methodology from side channel attacks, fault attacks, and return-oriented programming that can read arbitrary memory from the victim's process. More broadly, the paper shows that speculative execution implementations violate the security assumptions underpinning numerous software security mechanisms, including operating system process separation, containerization, just-in-time (JIT) compilation, and countermeasures to cache timing and side-channel attacks. These attacks represent a serious threat to actual systems since vulnerable speculative execution capabilities are found in microprocessors from Intel, AMD, and ARM that are used in billions of devices. While makeshift processor-specific countermeasures are possible in some cases, sound solutions will require fixes to processor designs as well as updates to instruction set architectures (ISAs) to give hardware architects and software developers a common understanding as to what computation state CPU implementations are (and are not) permitted to leak.

Posted Content
TL;DR: This work proposes building data networks among organizations based on federated mechanisms as an effective solution to allow knowledge to be shared without compromising user privacy.
Abstract: Today's AI still faces two major challenges. One is that in most industries, data exists in the form of isolated islands. The other is the strengthening of data privacy and security. We propose a possible solution to these challenges: secure federated learning. Beyond the federated learning framework first proposed by Google in 2016, we introduce a comprehensive secure federated learning framework, which includes horizontal federated learning, vertical federated learning and federated transfer learning. We provide definitions, architectures and applications for the federated learning framework, and provide a comprehensive survey of existing works on this subject. In addition, we propose building data networks among organizations based on federated mechanisms as an effective solution to allow knowledge to be shared without compromising user privacy.

Proceedings ArticleDOI
Guorui Zhou1, Xiaoqiang Zhu1, Chenru Song1, Ying Fan1, Han Zhu1, Xiao Ma1, Yan Yanghui1, Junqi Jin1, Han Li1, Kun Gai1 
19 Jul 2018
TL;DR: A novel model: Deep Interest Network (DIN) is proposed which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad.
Abstract: Click-through rate prediction is an essential task in industrial applications, such as online advertising. Recently deep learning based models have been proposed, which follow a similar Embedding&MLP paradigm. In these methods large scale sparse input features are first mapped into low dimensional embedding vectors, and then transformed into fixed-length vectors in a group-wise manner, finally concatenated together to fed into a multilayer perceptron (MLP) to learn the nonlinear relations among features. In this way, user features are compressed into a fixed-length representation vector, in regardless of what candidate ads are. The use of fixed-length vector will be a bottleneck, which brings difficulty for Embedding&MLP methods to capture user's diverse interests effectively from rich historical behaviors. In this paper, we propose a novel model: Deep Interest Network (DIN) which tackles this challenge by designing a local activation unit to adaptively learn the representation of user interests from historical behaviors with respect to a certain ad. This representation vector varies over different ads, improving the expressive ability of model greatly. Besides, we develop two techniques: mini-batch aware regularization and data adaptive activation function which can help training industrial deep networks with hundreds of millions of parameters. Experiments on two public datasets as well as an Alibaba real production dataset with over 2 billion samples demonstrate the effectiveness of proposed approaches, which achieve superior performance compared with state-of-the-art methods. DIN now has been successfully deployed in the online display advertising system in Alibaba, serving the main traffic.

Proceedings Article
01 Jan 2018
TL;DR: Glow as mentioned in this paper is a simple type of generative flow using invertible 1x1 convolutional neural network (1x1-CNN) that is optimized towards the plain log-likelihood objective.
Abstract: Flow-based generative models are conceptually attractive due to tractability of the exact log-likelihood, tractability of exact latent-variable inference, and parallelizability of both training and synthesis. In this paper we propose Glow, a simple type of generative flow using invertible 1x1 convolution. Using our method we demonstrate a significant improvement in log-likelihood and qualitative sample quality. Perhaps most strikingly, we demonstrate that a generative model optimized towards the plain log-likelihood objective is capable of efficient synthesis of large and subjectively realistic-looking images.

Journal ArticleDOI
TL;DR: An overview of machine learning applications for the analysis of genome sequencing data sets, including the annotation of sequence elements and epigenetic, proteomic or metabolomic data is provided.
Abstract: The field of machine learning, which aims to develop computer algorithms that improve with experience, holds promise to enable computers to assist humans in the analysis of large, complex data sets. Here, we provide an overview of machine learning applications for the analysis of genome sequencing data sets, including the annotation of sequence elements and epigenetic, proteomic or metabolomic data. We present considerations and recurrent challenges in the application of supervised, semi-supervised and unsupervised machine learning methods, as well as of generative and discriminative modelling approaches. We provide general guidelines to assist in the selection of these machine learning methods and their practical application for the analysis of genetic and genomic data sets.

Journal ArticleDOI
TL;DR: The aim of this review is to emphasize with current information the importance of antioxidants which play the role in cellular responce against oxidative/nitrosative stress, which would be helpful in enhancing the knowledge of any biochemist, pathophysiologist, or medical personnel regarding this important issue.
Abstract: Remarkable interest has risen in the idea that oxidative/nitrosative stress is mediated in the etiology of numerous human diseases. Oxidative/Nitrosative stress is the result of an disequilibrium in oxidant/antioxidant which reveals from continuous increase of Reactive Oxygen and Reactive Nitrogen Species production. The aim of this review is to emphasize with current information the importance of antioxidants which play the role in cellular responce against oxidative/nitrosative stress, which would be helpful in enhancing the knowledge of any biochemist, pathophysiologist, or medical personnel regarding this important issue. Products of lipid peroxidation have commonly been used as biomarkers of oxidative/nitrosative stress damage. Lipid peroxidation generates a variety of relatively stable decomposition end products, mainly α, β-unsaturated reactive aldehydes, such as malondialdehyde, 4-hydroxy-2-nonenal, 2-propenal (acrolein) and isoprostanes, which can be measured in plasma and urine as an indirect index of oxidative/nitrosative stress. Antioxidants are exogenous or endogenous molecules that mitigate any form of oxidative/nitrosative stress or its consequences. They may act from directly scavenging free radicals to increasing antioxidative defences. Antioxidant deficiencies can develop as a result of decreased antioxidant intake, synthesis of endogenous enzymes or increased antioxidant utilization. Antioxidant supplementation has become an increasingly popular practice to maintain optimal body function. However, antoxidants exhibit pro-oxidant activity depending on the specific set of conditions. Of particular importance are their dosage and redox conditions in the cell.

Posted Content
TL;DR: This document provides a brief introduction to CNNs, discussing recently published papers and newly formed techniques in developing these brilliantly fantastic image recognition models.
Abstract: The field of machine learning has taken a dramatic twist in recent times, with the rise of the Artificial Neural Network (ANN). These biologically inspired computational models are able to far exceed the performance of previous forms of artificial intelligence in common machine learning tasks. One of the most impressive forms of ANN architecture is that of the Convolutional Neural Network (CNN). CNNs are primarily used to solve difficult image-driven pattern recognition tasks and with their precise yet simple architecture, offers a simplified method of getting started with ANNs. This document provides a brief introduction to CNNs, discussing recently published papers and newly formed techniques in developing these brilliantly fantastic image recognition models. This introduction assumes you are familiar with the fundamentals of ANNs and machine learning.

Journal ArticleDOI
TL;DR: A comprehensive review of TADF materials is presented, with a focus on linking their optoelectronic behavior with the performance of the organic light-emitting diode (OLED) and related EL devices.
Abstract: We thank the University of St Andrews for support. EZ-C thanks the Leverhulme Trust for financial support (RPG-2016-047). and the EPSRC (EP/P010482/1) for financial support.

Journal ArticleDOI
TL;DR: Experimental results based on several hyperspectral image data sets demonstrate that the proposed method can achieve better classification performance than some traditional methods, such as support vector machines and the conventional deep learning-based methods.
Abstract: Recently, convolutional neural networks have demonstrated excellent performance on various visual tasks, including the classification of common two-dimensional images. In this paper, deep convolutional neural networks are employed to classify hyperspectral images directly in spectral domain. More specifically, the architecture of the proposed classifier contains five layers with weights which are the input layer, the convolutional layer, the max pooling layer, the full connection layer, and the output layer. These five layers are implemented on each spectral signature to discriminate against others. Experimental results based on several hyperspectral image data sets demonstrate that the proposed method can achieve better classification performance than some traditional methods, such as support vector machines and the conventional deep learning-based methods.

Posted Content
TL;DR: This article addresses the key challenges in designing and implementing the new IRS-aided hybrid (with both active and passive components) wireless network, as compared to the traditional network comprising active components only.
Abstract: Although the fifth-generation (5G) technologies will significantly improve the spectrum and energy efficiency of today's wireless communication networks, their high complexity and hardware cost as well as increasingly more energy consumption are still crucial issues to be solved. Furthermore, despite that such technologies are generally capable of adapting to the space and time varying wireless environment, the signal propagation over it is essentially random and largely uncontrollable. Recently, intelligent reflecting surface (IRS) has been proposed as a revolutionizing solution to address this open issue, by smartly reconfiguring the wireless propagation environment with the use of massive low-cost, passive, reflective elements integrated on a planar surface. Specifically, different elements of an IRS can independently reflect the incident signal by controlling its amplitude and/or phase and thereby collaboratively achieve fine-grained three-dimensional (3D) passive beamforming for signal enhancement or cancellation. In this article, we provide an overview of the IRS technology, including its main applications in wireless communication, competitive advantages over existing technologies, hardware architecture as well as the corresponding new signal model. We focus on the key challenges in designing and implementing the new IRS-aided hybrid (with both active and passive components) wireless network, as compared to the traditional network comprising active components only. Furthermore, numerical results are provided to show the potential for significant performance enhancement with the use of IRS in typical wireless network scenarios.

Journal ArticleDOI
TL;DR: Among patients with type 2 diabetes with or without previous cardiovascular disease, the incidence of major adverse cardiovascular events did not differ significantly between patients who received exenatide and those who received placebo.
Abstract: BackgroundThe cardiovascular effects of adding once-weekly treatment with exenatide to usual care in patients with type 2 diabetes are unknown. MethodsWe randomly assigned patients with type 2 diabetes, with or without previous cardiovascular disease, to receive subcutaneous injections of extended-release exenatide at a dose of 2 mg or matching placebo once weekly. The primary composite outcome was the first occurrence of death from cardiovascular causes, nonfatal myocardial infarction, or nonfatal stroke. The coprimary hypotheses were that exenatide, administered once weekly, would be noninferior to placebo with respect to safety and superior to placebo with respect to efficacy. ResultsIn all, 14,752 patients (of whom 10,782 [73.1%] had previous cardiovascular disease) were followed for a median of 3.2 years (interquartile range, 2.2 to 4.4). A primary composite outcome event occurred in 839 of 7356 patients (11.4%; 3.7 events per 100 person-years) in the exenatide group and in 905 of 7396 patients (12.2...

Journal ArticleDOI
21 Oct 2019-Nature
TL;DR: AFerroptosis suppressor protein 1 (FSP1) is identified as a key component of a non-mitochondrial CoQ antioxidant system that acts in parallel to the canonical glutathione-based GPX4 pathway, and pharmacological inhibition of FSP1 may provide an effective strategy to sensitize cancer cells to ferroPTosis-inducing chemotherapeutic agents.
Abstract: Ferroptosis is a form of regulated cell death that is caused by the iron-dependent peroxidation of lipids1,2. The glutathione-dependent lipid hydroperoxidase glutathione peroxidase 4 (GPX4) prevents ferroptosis by converting lipid hydroperoxides into non-toxic lipid alcohols3,4. Ferroptosis has previously been implicated in the cell death that underlies several degenerative conditions2, and induction of ferroptosis by the inhibition of GPX4 has emerged as a therapeutic strategy to trigger cancer cell death5. However, sensitivity to GPX4 inhibitors varies greatly across cancer cell lines6, which suggests that additional factors govern resistance to ferroptosis. Here, using a synthetic lethal CRISPR–Cas9 screen, we identify ferroptosis suppressor protein 1 (FSP1) (previously known as apoptosis-inducing factor mitochondrial 2 (AIFM2)) as a potent ferroptosis-resistance factor. Our data indicate that myristoylation recruits FSP1 to the plasma membrane where it functions as an oxidoreductase that reduces coenzyme Q10 (CoQ) (also known as ubiquinone-10), which acts as a lipophilic radical-trapping antioxidant that halts the propagation of lipid peroxides. We further find that FSP1 expression positively correlates with ferroptosis resistance across hundreds of cancer cell lines, and that FSP1 mediates resistance to ferroptosis in lung cancer cells in culture and in mouse tumour xenografts. Thus, our data identify FSP1 as a key component of a non-mitochondrial CoQ antioxidant system that acts in parallel to the canonical glutathione-based GPX4 pathway. These findings define a ferroptosis suppression pathway and indicate that pharmacological inhibition of FSP1 may provide an effective strategy to sensitize cancer cells to ferroptosis-inducing chemotherapeutic agents. A synthetic lethal CRISPR–Cas9 screen identifies ferroptosis suppressor protein 1 as a key ferroptosis-resistance factor, the expression of which correlates with ferroptosis resistance in hundreds of cancer cell lines.

Proceedings ArticleDOI
01 Oct 2019
TL;DR: CCNet as mentioned in this paper proposes a recurrent criss-cross attention module to harvest the contextual information of all the pixels on its crisscross path, and then takes a further recurrent operation to finally capture the full-image dependencies from all pixels.
Abstract: Full-image dependencies provide useful contextual information to benefit visual understanding problems. In this work, we propose a Criss-Cross Network (CCNet) for obtaining such contextual information in a more effective and efficient way. Concretely, for each pixel, a novel criss-cross attention module in CCNet harvests the contextual information of all the pixels on its criss-cross path. By taking a further recurrent operation, each pixel can finally capture the full-image dependencies from all pixels. Overall, CCNet is with the following merits: 1) GPU memory friendly. Compared with the non-local block, the proposed recurrent criss-cross attention module requires 11x less GPU memory usage. 2) High computational efficiency. The recurrent criss-cross attention significantly reduces FLOPs by about 85% of the non-local block in computing full-image dependencies. 3) The state-of-the-art performance. We conduct extensive experiments on popular semantic segmentation benchmarks including Cityscapes, ADE20K, and instance segmentation benchmark COCO. In particular, our CCNet achieves the mIoU score of 81.4 and 45.22 on Cityscapes test set and ADE20K validation set, respectively, which are the new state-of-the-art results. The source code is available at https://github.com/speedinghzl/CCNet.

Book ChapterDOI
20 Aug 2017
TL;DR: “Ouroboros” is presented, the first blockchain protocol based on proof of stake with rigorous security guarantees and it is proved that, given this mechanism, honest behavior is an approximate Nash equilibrium, thus neutralizing attacks such as selfish mining.
Abstract: We present “Ouroboros”, the first blockchain protocol based on proof of stake with rigorous security guarantees. We establish security properties for the protocol comparable to those achieved by the bitcoin blockchain protocol. As the protocol provides a “proof of stake” blockchain discipline, it offers qualitative efficiency advantages over blockchains based on proof of physical resources (e.g., proof of work). We also present a novel reward mechanism for incentivizing Proof of Stake protocols and we prove that, given this mechanism, honest behavior is an approximate Nash equilibrium, thus neutralizing attacks such as selfish mining.

Journal ArticleDOI
22 Oct 2017-Foods
TL;DR: The purpose of this review is to provide a brief overview of the plethora of research regarding the health benefits ofCurcumin combined with enhancing agents provides multiple health benefits.
Abstract: Turmeric, a spice that has long been recognized for its medicinal properties, has received interest from both the medical/scientific world and from culinary enthusiasts, as it is the major source of the polyphenol curcumin. It aids in the management of oxidative and inflammatory conditions, metabolic syndrome, arthritis, anxiety, and hyperlipidemia. It may also help in the management of exercise-induced inflammation and muscle soreness, thus enhancing recovery and performance in active people. In addition, a relatively low dose of the complex can provide health benefits for people that do not have diagnosed health conditions. Most of these benefits can be attributed to its antioxidant and anti-inflammatory effects. Ingesting curcumin by itself does not lead to the associated health benefits due to its poor bioavailability, which appears to be primarily due to poor absorption, rapid metabolism, and rapid elimination. There are several components that can increase bioavailability. For example, piperine is the major active component of black pepper and, when combined in a complex with curcumin, has been shown to increase bioavailability by 2000%. Curcumin combined with enhancing agents provides multiple health benefits. The purpose of this review is to provide a brief overview of the plethora of research regarding the health benefits of curcumin.

Journal ArticleDOI
TL;DR: The importance of potentially modifiable risk factors for stroke in different regions of the world, and in key populations and primary pathological subtypes of stroke, was quantified.

Journal ArticleDOI
TL;DR: A rapid increase in coronavirus disease 2019 (Covid-19) cases due to the omicron (B.1.529) variant in highly vaccinated populations has aroused concerns about the effectiveness of current vaccines as mentioned in this paper .
Abstract: A rapid increase in coronavirus disease 2019 (Covid-19) cases due to the omicron (B.1.1.529) variant of severe acute respiratory syndrome coronavirus 2 in highly vaccinated populations has aroused concerns about the effectiveness of current vaccines.

Proceedings ArticleDOI
Steven J. Rennie1, Etienne Marcheret1, Youssef Mroueh1, Jarret Ross1, Vaibhava Goel1 
21 Jul 2017
TL;DR: This paper considers the problem of optimizing image captioning systems using reinforcement learning, and shows that by carefully optimizing systems using the test metrics of the MSCOCO task, significant gains in performance can be realized.
Abstract: Recently it has been shown that policy-gradient methods for reinforcement learning can be utilized to train deep end-to-end systems directly on non-differentiable metrics for the task at hand. In this paper we consider the problem of optimizing image captioning systems using reinforcement learning, and show that by carefully optimizing our systems using the test metrics of the MSCOCO task, significant gains in performance can be realized. Our systems are built using a new optimization approach that we call self-critical sequence training (SCST). SCST is a form of the popular REINFORCE algorithm that, rather than estimating a baseline to normalize the rewards and reduce variance, utilizes the output of its own test-time inference algorithm to normalize the rewards it experiences. Using this approach, estimating the reward signal (as actor-critic methods must do) and estimating normalization (as REINFORCE algorithms typically do) is avoided, while at the same time harmonizing the model with respect to its test-time inference procedure. Empirically we find that directly optimizing the CIDEr metric with SCST and greedy decoding at test-time is highly effective. Our results on the MSCOCO evaluation sever establish a new state-of-the-art on the task, improving the best result in terms of CIDEr from 104.9 to 114.7.

Proceedings Article
21 Jan 2020
TL;DR: This paper demonstrates the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling, and shows that FixMatch achieves state-of-the-art performance across a variety of standard semi-supervised learning benchmarks.
Abstract: Semi-supervised learning (SSL) provides an effective means of leveraging unlabeled data to improve a model's performance. In this paper, we demonstrate the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling. Our algorithm, FixMatch, first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images. For a given image, the pseudo-label is only retained if the model produces a high-confidence prediction. The model is then trained to predict the pseudo-label when fed a strongly-augmented version of the same image. Despite its simplicity, we show that FixMatch achieves state-of-the-art performance across a variety of standard semi-supervised learning benchmarks, including 94.93% accuracy on CIFAR-10 with 250 labels and 88.61% accuracy with 40 -- just 4 labels per class. Since FixMatch bears many similarities to existing SSL methods that achieve worse performance, we carry out an extensive ablation study to tease apart the experimental factors that are most important to FixMatch's success. We make our code available at this https URL.