Open AccessPosted Content
Diagnose like a Radiologist: Attention Guided Convolutional Neural Network for Thorax Disease Classification.
TLDR
A three-branch attention guided convolution neural network (AG-CNN) that learns from disease-specific regions to avoid noise and improve alignment, and also integrates a global branch to compensate the lost discriminative cues by local branch.Abstract:
This paper considers the task of thorax disease classification on chest X-ray images Existing methods generally use the global image as input for network learning Such a strategy is limited in two aspects 1) A thorax disease usually happens in (small) localized areas which are disease specific Training CNNs using global image may be affected by the (excessive) irrelevant noisy areas 2) Due to the poor alignment of some CXR images, the existence of irregular borders hinders the network performance In this paper, we address the above problems by proposing a three-branch attention guided convolution neural network (AG-CNN) AG-CNN 1) learns from disease-specific regions to avoid noise and improve alignment, 2) also integrates a global branch to compensate the lost discriminative cues by local branch Specifically, we first learn a global CNN branch using global images Then, guided by the attention heat map generated from the global branch, we inference a mask to crop a discriminative region from the global image The local region is used for training a local CNN branch Lastly, we concatenate the last pooling layers of both the global and local branches for fine-tuning the fusion branch The Comprehensive experiment is conducted on the ChestX-ray14 dataset We first report a strong global baseline producing an average AUC of 0841 with ResNet-50 as backbone After combining the local cues with the global information, AG-CNN improves the average AUC to 0868 While DenseNet-121 is used, the average AUC achieves 0871, which is a new state of the art in the communityread more
Citations
More filters
Journal ArticleDOI
CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison
Jeremy Irvin,Pranav Rajpurkar,Michael Ko,Yifan Yu,Silviana Ciurea-Ilcus,Christopher G. Chute,Henrik Marklund,Behzad Haghgoo,Robyn L. Ball,Katie Shpanskaya,Jayne Seekins,David A. Mong,Safwan Halabi,Jesse K. Sandberg,Ricky Jones,David B. Larson,Curtis P. Langlotz,Bhavik N. Patel,Matthew P. Lungren,Andrew Y. Ng +19 more
TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation.
Journal ArticleDOI
Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists
Pranav Rajpurkar,Jeremy Irvin,Robyn L. Ball,Kaylie Zhu,Brandon Yang,Hershel Mehta,Tony Duan,Daisy Ding,Aarti Bagul,Curtis P. Langlotz,Bhavik N. Patel,Kristen W. Yeom,Katie Shpanskaya,Francis G. Blankenberg,Jayne Seekins,Timothy J. Amrhein,David A. Mong,Safwan Halabi,Evan J. Zucker,Andrew Y. Ng,Matthew P. Lungren +20 more
TL;DR: CheXNeXt, a convolutional neural network to concurrently detect the presence of 14 different pathologies, including pneumonia, pleural effusion, pulmonary masses, and nodules in frontal-view chest radiographs, achieved radiologist-level performance on 11 pathologies and did not achieve radiologists' level performance on 3 pathologies.
Posted Content
CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison
Jeremy Irvin,Pranav Rajpurkar,Michael Ko,Yifan Yu,Silviana Ciurea-Ilcus,Christopher G. Chute,Henrik Marklund,Behzad Haghgoo,Robyn L. Ball,Katie Shpanskaya,Jayne Seekins,David A. Mong,Safwan Halabi,Jesse K. Sandberg,Ricky Jones,David B. Larson,Curtis P. Langlotz,Bhavik N. Patel,Matthew P. Lungren,Andrew Y. Ng +19 more
TL;DR: CheXpert as discussed by the authors is a large dataset of chest radiographs of 65,240 patients annotated by 3 board-certified radiologists with 14 observations in radiology reports, capturing uncertainties inherent in radiograph interpretation and different approaches to using the uncertainty labels for training convolutional neural networks that output the probability of these observations given the available frontal and lateral radiographs.
Journal ArticleDOI
PadChest: A large chest x-ray image dataset with multi-label annotated reports.
TL;DR: This dataset includes more than 160,000 images obtained from 67,000 patients that were interpreted and reported by radiologists at San Juan Hospital (Spain) from 2009 to 2017, covering six different position views and additional information on image acquisition and patient demography.
Posted Content
Attention Mechanisms in Computer Vision: A Survey.
Meng-Hao Guo,Tian-Xing Xu,Jiangjiang Liu,Zheng-Ning Liu,Peng-Tao Jiang,Tai-Jiang Mu,Song-Hai Zhang,Ralph R. Martin,Ming-Ming Cheng,Shi-Min Hu +9 more
TL;DR: A comprehensive review of attention mechanisms in computer vision can be found in this article, which categorizes them according to approach, such as channel attention, spatial attention, temporal attention and branch attention.
References
More filters
Proceedings ArticleDOI
Deep Residual Learning for Image Recognition
TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article
ImageNet Classification with Deep Convolutional Neural Networks
TL;DR: The state-of-the-art performance of CNNs was achieved by Deep Convolutional Neural Networks (DCNNs) as discussed by the authors, which consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax.
Journal ArticleDOI
Long short-term memory
TL;DR: A novel, efficient, gradient based method called long short-term memory (LSTM) is introduced, which can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units.
Proceedings Article
Very Deep Convolutional Networks for Large-Scale Image Recognition
Karen Simonyan,Andrew Zisserman +1 more
TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings Article
Very Deep Convolutional Networks for Large-Scale Image Recognition
Karen Simonyan,Andrew Zisserman +1 more
TL;DR: In this paper, the authors investigated the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting and showed that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 layers.