scispace - formally typeset
Open AccessPosted Content

Diagnose like a Radiologist: Attention Guided Convolutional Neural Network for Thorax Disease Classification.

TLDR
A three-branch attention guided convolution neural network (AG-CNN) that learns from disease-specific regions to avoid noise and improve alignment, and also integrates a global branch to compensate the lost discriminative cues by local branch.
Abstract
This paper considers the task of thorax disease classification on chest X-ray images Existing methods generally use the global image as input for network learning Such a strategy is limited in two aspects 1) A thorax disease usually happens in (small) localized areas which are disease specific Training CNNs using global image may be affected by the (excessive) irrelevant noisy areas 2) Due to the poor alignment of some CXR images, the existence of irregular borders hinders the network performance In this paper, we address the above problems by proposing a three-branch attention guided convolution neural network (AG-CNN) AG-CNN 1) learns from disease-specific regions to avoid noise and improve alignment, 2) also integrates a global branch to compensate the lost discriminative cues by local branch Specifically, we first learn a global CNN branch using global images Then, guided by the attention heat map generated from the global branch, we inference a mask to crop a discriminative region from the global image The local region is used for training a local CNN branch Lastly, we concatenate the last pooling layers of both the global and local branches for fine-tuning the fusion branch The Comprehensive experiment is conducted on the ChestX-ray14 dataset We first report a strong global baseline producing an average AUC of 0841 with ResNet-50 as backbone After combining the local cues with the global information, AG-CNN improves the average AUC to 0868 While DenseNet-121 is used, the average AUC achieves 0871, which is a new state of the art in the community

read more

Citations
More filters
Journal ArticleDOI

CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison

TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation.
Journal ArticleDOI

Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists

TL;DR: CheXNeXt, a convolutional neural network to concurrently detect the presence of 14 different pathologies, including pneumonia, pleural effusion, pulmonary masses, and nodules in frontal-view chest radiographs, achieved radiologist-level performance on 11 pathologies and did not achieve radiologists' level performance on 3 pathologies.
Posted Content

CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison

TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation and different approaches to using the uncertainty labels for training convolutional neural networks that output the probability of these observations given the available frontal and lateral radiographs.
Journal ArticleDOI

PadChest: A large chest x-ray image dataset with multi-label annotated reports.

TL;DR: This dataset includes more than 160,000 images obtained from 67,000 patients that were interpreted and reported by radiologists at San Juan Hospital (Spain) from 2009 to 2017, covering six different position views and additional information on image acquisition and patient demography.
Posted Content

Attention Mechanisms in Computer Vision: A Survey.

TL;DR: A comprehensive review of attention mechanisms in computer vision can be found in this article, which categorizes them according to approach, such as channel attention, spatial attention, temporal attention and branch attention.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

ImageNet Classification with Deep Convolutional Neural Networks

TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI

Long short-term memory

TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.