scispace - formally typeset
Search or ask a question
Posted ContentDOI

Clinically Applicable Deep Learning Strategy for Pulmonary Nodule Risk Prediction: Insights into HONORS

TL;DR: The combination of HONORS and FGP-NET provides well-organized stratification for pulmonary nodules and also offers the potential for reducing medical errors.
Abstract: Background and Purpose Limited optimization was clinically applicable for reducing missed diagnosis, misdiagnosis and inter-reader variability in pulmonary nodule diagnosis. We aimed to propose a deep learning-based algorithm and a practical strategy to better stratify the risk of pulmonary nodules, thus reducing medical errors and optimizing the clinical workflow. Materials and Methods A total of 2,348 pulmonary nodules (1,215 with lung cancer) containing screened nodules from National Lung Cancer Screening Trial (NLST) and incidentally detected nodules from Jinling Hospital (JLH) were used to train and evaluate a deep learning algorithm, Filter-guided pyramid network (FGP-NET). Internal and external test of FGP-NET were performed on two independent datasets (n=542). The performance of FGP-NET at Youden point which maximizing the Youden index was compared with 126 board-certificated radiologists. We further proposed Hierarchical Ordered Network ORiented Strategy (HONORS), which manipulates the emphasis either on sensitivity or specificity to target risk-stratified clinical scenarios, directly making decisions for some patients. Results FGP-NET achieved a high area under the curve (AUC) of 0.969 and 0.855 for internal and external testing, and was comparable or even outperformed the radiologists when considering sensitivity. HONORS-guided FGP-NET identified benign nodules with a high sensitivity (95.5%) in the screening scenario, and demonstrated satisfactory performance for the rest ambiguous nodules with 0.945 of AUC by the Youden point. FGP-NET also detected lung cancer with a high specificity of 94.5% in routine diagnostic scenario; an AUC of 0.809 was achieved for the rest nodules. Conclusion The combination of HONORS and FGP-NET provides well-organized stratification for pulmonary nodules and also offers the potential for reducing medical errors. Highlights Pulmonary nodules were managed for both screening and diagnostic scenarios Proposal of a hierarchical strategy for targeting risk-stratified clinical scenarios A large scale Human-deep learning contest for reliable performance evaluation

Summary (3 min read)

Introduction

  • Lung cancer is one of the commonest cancers and the leading cause of cancer mortality worldwide 1,2.
  • Every step from screening to final diagnosis is indispensable which makes the entire process time- and labor-consuming.
  • The desire to improve the efficacy and efficiency of clinical care continues to drive multiple innovations into practice, including deep learning (DL).
  • Most of previous works focused on the screening population, and studies based on incidentally detected nodules in routine diagnostic scenario with higher risk were limited.
  • The benign nodules or lung cancer can be accurately identified in screening and routine diagnostic scenarios in step-1 and further stratification of ambiguous nodules was performed to aid clinical decision making in step-2.

Datasets

  • Ethical approval was obtained for this retrospective study, and informed consent was waived for reviewing patients’ medical records.
  • The authors retrospectively analyzed 16,801 patients who underwent surgery or biopsy due to lung lesions from May 2009 to June 2018 in Jinling hospital.
  • The authors analyzed 1,060 patients confirmed as having lung cancer and 5,275 patients as not in the NLST.
  • Details for the data curation are available in the Supplementary Information, Inclusion and exclusion criteria.

Annotation and preprocessing

  • Automatic nodule detection was performed using the Dr. wise platform 17 (the detection network could be found in the supplementary information, Pulmonary nodule detection network) and the geometric centers of the enrolled nodules were further revised by two radiologists.
  • Due to the differences in pixel spacing and slice thickness, the CT images were subsequently linearly interpolated into 3D isotropic images with voxel spacing of 0.6 × 0.6 × 0.6 mm3.
  • Since the datasets were relatively small compared to traditional image classification datasets for deep learning such as cifar1018 and ImageNet19, the authors conducted heavy data augmentations on all initially generated image patches (a size of 128 × 128 × 128) containing the nodules, e.g., 0-360 degree of random rotation, random zooming in or zooming out, random cropping and flipping.

Development and training of FGP-NET

  • JLH and NLST dataset were randomly assigned into one of the following three sets: training set (JLH, 1086 nodules; NLST, 520 nodules) for optimizing network weights, validation set (JLH, 100 nodules; NLST, 100 nodules) for deciding the values of hyperparameters and internal test set (JLH, 100 nodules; NLST, 200 nodules) for evaluating the performance.
  • The patients in the training, validation and test sets were exclusive to each of the other data sets.
  • To this end, the authors aimed to capture both local and global features and their interactive relationships to better represent the nodule.
  • By concatenating the clear attention map distilled by local feature extractors and raw feature maps from early stage of network, FGP-Net was able to keep high resolution details to describe small-sized local features and to use the accurate localization of large ones to guide small ones.
  • DenseNet was one of the state-of-art network structures in computer vision.

Validation of FGP-NET

  • FGP-Net generated continuous numbers between 0 and 1 for nodule risk stratification, being consistent with the malignancy probability of the nodules, named ‘malignancy score’.
  • The corresponding sets were defined as S1, S2, S3 and S4, respectively.
  • To keep up with the input of FGP-NET, radiologists were only informed with the precise location of the nodules while blinded to the medical history and pathological results.
  • The performance of radiologists was evaluated at the average level and majority level (voting the scores rated by 126 radiologists) when compared with FGP-NET.
  • All statistical tests used in this study were 2-sided and a P-value less than .05 was considered significant.

Interpretation of learned features

  • Given the black box property of DL, the authors further conducted a two-way feature interpretation to explore whether FGP-NET learned solid and effective features.
  • More specifically, T-distributed Stochastic Neighborhood Embedding (t-SNE) 24 and probability heat-map were applied for visualization of global features and local features (Supplementary Information, Feature visualization of FGP-Net).

Evaluation of HONORS

  • To validate the performance of HONORS in screening and routine diagnostic scenarios, the authors simulated the its application on NLST test set and multi-center set.
  • The precision of the stratified nodules was evaluated using negative predictive value (NPV) and positive predictive value (PPV), respectively.
  • The authors further compared the performance between HONORS and 126 radiologists in the incidentally detected nodules using JLH test set (Supplementary information, Comparison of HONORS with radiologists).
  • In addition to the two-step way targeted on different scenarios, HONORS can also be realized in a three-step way regardless of scenarios .
  • In the first step, FPG-NET at HSen point stratified the benign nodules; in the second step, FPG-NET at HSpe point stratified the lung cancer; and in the third step, FPG-NET at Youden point stratified the rest ambiguous nodules.

Results

  • Overview of the study design and results A total of 2,348 pulmonary nodules (1,215 malignant nodules) containing screened and incidentally detected nodules found by chest CT were used to train and evaluate their DL algorithm, FGP-NET .
  • The authors further investigated whether the FGP-NET was comparable or even superior to radiologists.
  • It was still at a relatively low level even in consultant group (κw < 0.4).
  • A total of 18 nodules was misdiagnosed by radiologists’ majority opinion and FGP-NET inconsistently, accounting for 9 each .

Interpretation of Features Learned by FGP-NET

  • DL is frequently referred to as a black box—data goes in, decisions come out, but the processes between input and output are opaque 25.
  • It is crucial to enable the black box to be opened, and thus, the authors interpreted the global features by using t-SNE which is particularly well suited for the visualization of high-dimensional data.
  • To be specific, these attribution regions were mainly located in benign features for those benign nodules.
  • Nevertheless, to malignant nodules, regions were predominately situated in malignant features, suggesting that FGP-NET may concentrate on the irregular margin of the nodule and solid component within part-solid nodules .
  • These attribution regions were certainly consistent with the visual observation by radiologists.

Proposal of HONORS

  • This does not eliminate the role of radiologists in making the final decision.
  • Therefore, the authors further proposed HONORS, a novel two-step hierarchical strategy for clinical application of FGP-NET.
  • HONORS was applied to NLST test set and multi-center set to appraise its performance.
  • Additionally, a three-step way to realize HONORS was also evaluated in this context .

Discussion

  • The authors developed and validated a DL algorithm—FGP-NET that is capable of stratifying pulmonary nodules with great performance and is comparable with a large group of 10 / 20 radiologists.
  • The authors study was conceptually practical in clinics because FGP-NET and HONORS were designed to tailor both screening and routine diagnostic scenarios.
  • Pyramid structure that the authors harnessed to support FGP-NET is a method for extracting multi-scale features and is quite suitable for feature extraction of pulmonary nodules due to its large size variation 26.
  • Huang et al. developed a computer-aided diagnosis approach with a sensitivity of 95% and a specificity of 88% which outperformed three radiologists’ combined reading using 186 nodules from NLST dataset 29.
  • 12 / 20 Taken together, the authors proposed HONORS to lay groundwork toward application of the DL-based pulmonary nodule stratification algorithm in the screening and routine diagnostic scenarios.

Did you find this useful? Give us your feedback

Figures (6)

Content maybe subject to copyright    Report

1 / 20
Clinically Applicable Deep Learning Strategy for Pulmonary Nodule Risk
Prediction: Insights into HONORS
Wenhui Lv
1,2,13
, Yang Wang
3,13
, Changsheng Zhou
2,13
, Sheng Huang
4
, Xiangming Fang
5
,
Qiuzhen Xu
6
, Qirui Zhang
1,2
, Chuxi Huang
2
, Xinyu Li
1,2
, Zhen Zhou
7
, Yizhou Yu
8
,Yizhou
Wang
7
, Mengjie Lu
2
, Qiang Xu
2
, Xiuli Li
9
, Haoliang Lin
9
, Xiaofan Lu
10
, Qinmei Xu
2
, Jing
Sun
2
,Yuxia Tang
2
, Yong Song
11
, Fangrong Yan
10
, Bing Zhang
3
, Zhen Cheng
12
, Longjiang
Zhang
1,2
, Guangming Lu
1,2*
Author Affiliations:
1
Department of Medical Imaging, Jinling Hospital, Southern Medical University, Nanjing,
China
2
Department of Medical Imaging, Jinling Hospital, Nanjing University School of Medicine,
Nanjing, China
3
Department of Radiology, the Affiliated Nanjing Drum Tower Hospital of Nanjing
University Medical School, Nanjing, China
4
Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, China
5
Department of Medical Imaging, Wuxi People’s Hospital, Wuxi, China
6
Department of Radiology, Southeast University Zhongda Hospital, Nanjing, China
7
Computer Science Department, School of EECS, Peking University, Beijing, China
8
University of Hongkong, Hongkong, China
9
Deepwise AI Lab, Deepwise Inc., Beijing, China
10
Research Center of Biostatistics and Computational Pharmacy, China Pharmaceutical
University, Nanjing, China
11
Department of Respiratory Medicine, Jinling Hospital, Nanjing University School of
Medicine, Nanjing, Jiangsu, China
12
Department of Radiology, Stanford University, School of Medicine, Stanford, CA, United
States,
13
These authors contributed equally
*Corresponding author: Guangming Lu, MD, Department of Medical Imaging, Nanjing
Jinling Hospital, Nanjing University School of Medicine, 305#, Eastern Zhongshan Rd,
Nanjing 210002, China (Email: cjr.luguangming@vip.163.com)
All rights reserved. No reuse allowed without permission.
perpetuity.
preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in
The copyright holder for thisthis version posted February 7, 2020. ; https://doi.org/10.1101/2020.02.03.20020297doi: medRxiv preprint
NOTE: This preprint reports new research that has not been certified by peer review and should not be used to guide clinical practice.

2 / 20
Abstract
Background and Purpose: Limited optimization was clinically applicable for reducing
missed diagnosis, misdiagnosis and inter-reader variability in pulmonary nodule diagnosis.
We aimed to propose a deep learning-based algorithm and a practical strategy to better stratify
the risk of pulmonary nodules, thus reducing medical errors and optimizing the clinical
workflow.
Materials and Methods: A total of 2,348 pulmonary nodules (1,215 with lung cancer)
containing screened nodules from National Lung Cancer Screening Trial (NLST) and
incidentally detected nodules from Jinling Hospital (JLH) were used to train and evaluate a
deep learning algorithm, Filter-guided pyramid network (FGP-NET). Internal and external
test of FGP-NET were performed on two independent datasets (n=542). The performance of
FGP-NET at Youden point which maximizing the Youden index was compared with 126
board-certificated radiologists. We further proposed Hierarchical Ordered Network ORiented
Strategy (HONORS), which manipulates the emphasis either on sensitivity or specificity to
target risk-stratified clinical scenarios, directly making decisions for some patients.
Results: FGP-NET achieved a high area under the curve (AUC) of 0.969 and 0.855 for
internal and external testing, and was comparable or even outperformed the radiologists when
considering sensitivity. HONORS-guided FGP-NET identified benign nodules with a high
sensitivity (95.5%) in the screening scenario, and demonstrated satisfactory performance for
the rest ambiguous nodules with 0.945 of AUC by the Youden point. FGP-NET also detected
lung cancer with a high specificity of 94.5% in routine diagnostic scenario; an AUC of 0.809
was achieved for the rest nodules.
Conclusion: The combination of HONORS and FGP-NET provides well-organized
stratification for pulmonary nodules and also offers the potential for reducing medical errors.
Keywords: clinically applicable hierarchical strategy; Filter-guided pyramid network;
pulmonary nodule; human-AI contest; computed tomography
Highlights
Pulmonary nodules were managed for both screening and diagnostic scenarios
Proposal of a hierarchical strategy for targeting risk-stratified clinical scenarios
A large scale Human-deep learning contest for reliable performance evaluation
All rights reserved. No reuse allowed without permission.
perpetuity.
preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in
The copyright holder for thisthis version posted February 7, 2020. ; https://doi.org/10.1101/2020.02.03.20020297doi: medRxiv preprint

3 / 20
Introduction
Lung cancer is one of the commonest cancers and the leading cause of cancer mortality
worldwide
1,2
. Survival rate is hugely influenced by stage at diagnosis and early diagnosis is
full of challenges
3
. Every step from screening to final diagnosis is indispensable which
makes the entire process time- and labor-consuming. For better management and diagnosis of
pulmonary nodules, Lung-RADs guidelines
4
, recommendations from Fleschnier society
5
have been proposed. However, the lack of awareness of these guidelines and the inter-reader
variability have limited the broader use of them
6,7
. Besides, the detection of pulmonary
nodules is rising up dramatically and physicians are experiencing burnout at increasing rate.
Excessive workload is resulting in diagnostic errors including missed diagnosis, and
misdiagnosis
8
. Therefore, there is an unmet need for relieving working pressure of physicians
and reducing the incidence of such medical errors. To the best of our knowledge, limited
literature has delved into this place for a clinically applicable optimization.
The desire to improve the efficacy and efficiency of clinical care continues to drive multiple
innovations into practice, including deep learning (DL). In recent, DL has infiltrated the
optimization and streamlining of clinical workflows, quietly improving, changing and
reconstructing the way health care works
9-13
, especially for pulmonary nodules management
12,14
. However, most of previous works focused on the screening population, and studies based
on incidentally detected nodules in routine diagnostic scenario with higher risk were limited.
In this study, we reasoned that DL would play distinct roles in the risk-stratified clinical
scenarios for pulmonary nodule screening and diagnosis. Therefore, we proposed a clinically
applicable DL-based algorithmFilter-guided pyramid network (FGP-NET), and a practical
strategyHierarchical-Ordered Network-ORiented Strategy (HONORS), which involves two
steps for two different clinical scenarios (i.e., screening and routine diagnostic scenarios)
(Figure 1a). The benign nodules or lung cancer can be accurately identified in screening and
routine diagnostic scenarios in step-1 and further stratification of ambiguous nodules was
performed to aid clinical decision making in step-2. Consequently, HONORS would directly
make decision for some patients without any human intervention and assist physicians to
better manage other ambiguous nodules. It has great potential to provide well-organized
management for pulmonary nodules by optimizing the clinical workflow and reduce medical
errors through rapid and accurate image interpretation.
Methods
Datasets
Ethical approval was obtained for this retrospective study, and informed consent was waived
for reviewing patients’ medical records.
All rights reserved. No reuse allowed without permission.
perpetuity.
preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in
The copyright holder for thisthis version posted February 7, 2020. ; https://doi.org/10.1101/2020.02.03.20020297doi: medRxiv preprint

4 / 20
For training and validation of FGP-NET, we incorporated both incidentally detected and
screened nodules from Jinling Hospital (JLH) and National Lung Cancer Screening Trial
(NLST), respectively. We retrospectively analyzed 16,801 patients who underwent surgery or
biopsy due to lung lesions from May 2009 to June 2018 in Jinling hospital. Nodules were
classified as either malignant or benign according to the World Health Organization (WHO)
classification of lung tumors in 2015
15
. Finally, a total of 1,286 nodules (935 malignant
nodules and 351 benign nodules) found in 1,238 patients (591 males and 647 females; 56.48 ±
11.50 years of age [mean ± standard deviation (SD)]) were included. We analyzed 1,060
patients confirmed as having lung cancer and 5,275 patients as not in the NLST. A total of 820
nodules (93 malignant nodules and 727 benign nodules) found in 442 participants (244 males
and 198 females; 61.71 ± 5.32 years of age) were included. All participants enrolled in NLST
signed an informed consent developed and approved by the screening centers’ Institutional
Review Board (IRB), the National Cancer Institute (NCI) IRB and the Westat IRB. The data
and images were obtained from the National Cancer Institute Cancer Data Access System
16
.
We also conducted an external test of FGP-NET using a multi-center set. Multi-center set
contained consecutive 235 patients (104 males and 131 females; 58.48 ± 10.75 years of age)
with 242 incidentally detected nodules (187 malignant nodules and 55 benign nodules) from
three tertiary hospitals. These patients all underwent biopsy or surgery. Details for the data
curation are available in the Supplementary Information, Inclusion and exclusion criteria. All
patients included in this study underwent nonenhanced chest computed tomography (CT) and
the CT images were reconstructed with a thickness less than 2.5 mm (see more details of
acquisition and reconstruction parameters in Supplementary Table 1).
Annotation and preprocessing
Automatic nodule detection was performed using the Dr. wise platform
17
(the detection
network could be found in the supplementary information, Pulmonary nodule detection
network) and the geometric centers of the enrolled nodules were further revised by two
radiologists. Due to the differences in pixel spacing and slice thickness, the CT images were
subsequently linearly interpolated into 3D isotropic images with voxel spacing of 0.6 × 0.6 ×
0.6 mm3. Since the datasets were relatively small compared to traditional image classification
datasets for deep learning such as cifar10
18
and ImageNet
19
, we conducted heavy data
augmentations on all initially generated image patches (a size of 128 × 128 × 128) containing
the nodules, e.g., 0-360 degree of random rotation, random zooming in or zooming out,
random cropping and flipping. Processed patches were cropped to a size of 48 × 48 × 48 and
used for algorithm training.
All rights reserved. No reuse allowed without permission.
perpetuity.
preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in
The copyright holder for thisthis version posted February 7, 2020. ; https://doi.org/10.1101/2020.02.03.20020297doi: medRxiv preprint

5 / 20
Development and training of FGP-NET
JLH and NLST dataset were randomly assigned into one of the following three sets: training
set (JLH, 1086 nodules; NLST, 520 nodules) for optimizing network weights, validation set
(JLH, 100 nodules; NLST, 100 nodules) for deciding the values of hyperparameters and
internal test set (JLH, 100 nodules; NLST, 200 nodules) for evaluating the performance. The
patients in the training, validation and test sets were exclusive to each of the other data sets.
Pulmonary nodule evaluation on chest CT is of great challenge partly due to the significant
overlap and complicated interaction among visual features. To this end, we aimed to capture
both local and global features and their interactive relationships to better represent the nodule.
We designed a new U-shape network structure named FGP-Net which could produce a feature
pyramid for different-sized local features (Figure 1b). By concatenating the clear attention
map distilled by local feature extractors and raw feature maps from early stage of network,
FGP-Net was able to keep high resolution details to describe small-sized local features and to
use the accurate localization of large ones to guide small ones. Additionally, we added one
more backbone module to dig complex semantic information and relationships between these
various local ones. We applied Densely Connected Convolutional Network (DenseNet)
20
as
our backbone network and Discriminative Filter Learning (DFL)
21
modules as local feature
extractors for details of nodules. DenseNet was one of the state-of-art network structures in
computer vision. The core of Densenet was a so-called dense block, in which every layer was
connected directly to all the subsequent layers. DFL Module was a two-stream structure,
enabling network to detect discriminative mid-level patches under weakly supervised
condition.
FGP-Net was trained by using a batch size of 4 for 400 epochs and learning rate of 0.0001,
with adam optimizer and 0.0001 weight decay. After 400 epochs, the training process was
stopped due to the convergence of both the area under the curve (AUC) and cross-entropy loss
(Supplementary Figure 1).
Validation of FGP-NET
FGP-Net generated continuous numbers between 0 and 1 for nodule risk stratification, being
consistent with the malignancy probability of the nodules, named ‘malignancy score’.
Receiver operating characteristic (ROC) curves were generated by plotting sensitivity and 1 -
specificity, and AUCs were calculated to evaluate the discrimination
22
. Three operating
points were chosen for FGP-NET on the validation set; a high sensitivity (HSen)/specificity
(HSpe) point corresponded to a sensitivity/specificity of 99% to stratify the benign/malignant
nodules with high precision and a Youden point maximized the value of the Youden index for
generic discrimination between malignant and benign nodules.
Human-DL contest between FGP-NET and 126 radiologists
All rights reserved. No reuse allowed without permission.
perpetuity.
preprint (which was not certified by peer review) is the author/funder, who has granted medRxiv a license to display the preprint in
The copyright holder for thisthis version posted February 7, 2020. ; https://doi.org/10.1101/2020.02.03.20020297doi: medRxiv preprint

Citations
More filters
Journal ArticleDOI
TL;DR: A comprehensive and in-depth study of Deep Learning methodologies and applications in medicine and how, where and why Deep Learning models are applied in medicine is presented.
Abstract: New technologies are transforming medicine, and this revolution starts with data. Health data, clinical images, genome sequences, data on prescribed therapies and results obtained, data that each of us has helped to create. Although the first uses of artificial intelligence (AI) in medicine date back to the 1980s, it is only with the beginning of the new millennium that there has been an explosion of interest in this sector worldwide. We are therefore witnessing the exponential growth of health-related information with the result that traditional analysis techniques are not suitable for satisfactorily management of this vast amount of data. AI applications (especially Deep Learning), on the other hand, are naturally predisposed to cope with this explosion of data, as they always work better as the amount of training data increases, a phase necessary to build the optimal neural network for a given clinical problem. This paper proposes a comprehensive and in-depth study of Deep Learning methodologies and applications in medicine. An in-depth analysis of the literature is presented; how, where and why Deep Learning models are applied in medicine are discussed and reviewed. Finally, current challenges and future research directions are outlined and analysed.

162 citations

Journal ArticleDOI
TL;DR: A reliable diagnosis on nodule persistence can be achieved at baseline CT; thus, an early diagnosis as well as better patient care is available, and the transfer learning model presented good performance on the discrimination between transient and persistent SSNs.
Abstract: To develop and validate a deep learning model to discriminate transient from persistent subsolid nodules (SSNs) on baseline CT. A cohort of 1414 SSNs, consisting of 319 transient SSNs in 168 individuals and 1095 persistent SSNs in 816 individuals, were identified on chest CT. The cohort was assigned by examination date into a development set of 996 SSNs, a tuning set of 212 SSNs, and a validation set of 206 SSNs. Our model was built by transfer learning, which was transferred from a well-performed deep learning model for pulmonary nodule classification. The performance of the model was compared with that of two experienced radiologists. Each nodule was categorized by Lung CT Screening Reporting and Data System (Lung-RADS) to further evaluate the performance and the potential clinical benefit of the model. Two methods were employed to visualize the learned features. Our model achieved an AUC of 0.926 on the validation set with an accuracy of 0.859, a sensitivity of 0.863, and a specificity of 0.858, and outperformed the radiologists. The model performed the best among Lung-RADS 2 nodules and maintained well performance among Lung-RADS 4 nodules. Feature visualization demonstrated the model’s effectiveness in extracting features from images. The transfer learning model presented good performance on the discrimination between transient and persistent SSNs. A reliable diagnosis on nodule persistence can be achieved at baseline CT; thus, an early diagnosis as well as better patient care is available. • Deep learning can be used for the discrimination between transient and persistent subsolid nodules. • A transfer learning model can achieve good performance when it is transferred from a model with a similar task. • With the assistance of deep learning model, a reliable diagnosis on nodule persistence can be achieved at baseline CT, which can bring a better patient care strategy.

8 citations


Cites methods from "Clinically Applicable Deep Learning..."

  • ...It was finetuned on the basis of Filter-guided Pyramid Network (FGPNET), a novel 3D convolutional network structure designed for the classification of malignant and benign pulmonary nodules in our previous study [23]....

    [...]

References
More filters
Proceedings ArticleDOI
Jia Deng1, Wei Dong1, Richard Socher1, Li-Jia Li1, Kai Li1, Li Fei-Fei1 
20 Jun 2009
TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Abstract: The explosion of image data on the Internet has the potential to foster more sophisticated and robust models and algorithms to index, retrieve, organize and interact with images and multimedia data. But exactly how such data can be harnessed and organized remains a critical problem. We introduce here a new database called “ImageNet”, a large-scale ontology of images built upon the backbone of the WordNet structure. ImageNet aims to populate the majority of the 80,000 synsets of WordNet with an average of 500-1000 clean and full resolution images. This will result in tens of millions of annotated images organized by the semantic hierarchy of WordNet. This paper offers a detailed analysis of ImageNet in its current state: 12 subtrees with 5247 synsets and 3.2 million images in total. We show that ImageNet is much larger in scale and diversity and much more accurate than the current image datasets. Constructing such a large-scale database is a challenging task. We describe the data collection scheme with Amazon Mechanical Turk. Lastly, we illustrate the usefulness of ImageNet through three simple applications in object recognition, image classification and automatic object clustering. We hope that the scale, accuracy, diversity and hierarchical structure of ImageNet can offer unparalleled opportunities to researchers in the computer vision community and beyond.

49,639 citations

Proceedings ArticleDOI
21 Jul 2017
TL;DR: DenseNet as mentioned in this paper proposes to connect each layer to every other layer in a feed-forward fashion, which can alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.
Abstract: Recent work has shown that convolutional networks can be substantially deeper, more accurate, and efficient to train if they contain shorter connections between layers close to the input and those close to the output. In this paper, we embrace this observation and introduce the Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed-forward fashion. Whereas traditional convolutional networks with L layers have L connections—one between each layer and its subsequent layer—our network has L(L+1)/2 direct connections. For each layer, the feature-maps of all preceding layers are used as inputs, and its own feature-maps are used as inputs into all subsequent layers. DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters. We evaluate our proposed architecture on four highly competitive object recognition benchmark tasks (CIFAR-10, CIFAR-100, SVHN, and ImageNet). DenseNets obtain significant improvements over the state-of-the-art on most of them, whilst requiring less memory and computation to achieve high performance. Code and pre-trained models are available at https://github.com/liuzhuang13/DenseNet.

27,821 citations

Journal ArticleDOI
TL;DR: A representation and interpretation of the area under a receiver operating characteristic (ROC) curve obtained by the "rating" method, or by mathematical predictions based on patient characteristics, is presented and it is shown that in such a setting the area represents the probability that a randomly chosen diseased subject is (correctly) rated or ranked with greater suspicion than a random chosen non-diseased subject.
Abstract: A representation and interpretation of the area under a receiver operating characteristic (ROC) curve obtained by the "rating" method, or by mathematical predictions based on patient characteristics, is presented. It is shown that in such a setting the area represents the probability that a randomly chosen diseased subject is (correctly) rated or ranked with greater suspicion than a randomly chosen non-diseased subject. Moreover, this probability of a correct ranking is the same quantity that is estimated by the already well-studied nonparametric Wilcoxon statistic. These two relationships are exploited to (a) provide rapid closed-form expressions for the approximate magnitude of the sampling variability, i.e., standard error that one uses to accompany the area under a smoothed ROC curve, (b) guide in determining the size of the sample required to provide a sufficiently reliable estimate of this area, and (c) determine how large sample sizes should be to ensure that one can statistically detect difference...

19,398 citations

Journal ArticleDOI
TL;DR: The overall cancer death rate dropped continuously from 1991 to 2016 by a total of 27%, translating into approximately 2,629,200 fewer cancer deaths than would have been expected if death rates had remained at their peak.
Abstract: Each year, the American Cancer Society estimates the numbers of new cancer cases and deaths that will occur in the United States and compiles the most recent data on cancer incidence, mortality, and survival. Incidence data, available through 2015, were collected by the Surveillance, Epidemiology, and End Results Program; the National Program of Cancer Registries; and the North American Association of Central Cancer Registries. Mortality data, available through 2016, were collected by the National Center for Health Statistics. In 2019, 1,762,450 new cancer cases and 606,880 cancer deaths are projected to occur in the United States. Over the past decade of data, the cancer incidence rate (2006-2015) was stable in women and declined by approximately 2% per year in men, whereas the cancer death rate (2007-2016) declined annually by 1.4% and 1.8%, respectively. The overall cancer death rate dropped continuously from 1991 to 2016 by a total of 27%, translating into approximately 2,629,200 fewer cancer deaths than would have been expected if death rates had remained at their peak. Although the racial gap in cancer mortality is slowly narrowing, socioeconomic inequalities are widening, with the most notable gaps for the most preventable cancers. For example, compared with the most affluent counties, mortality rates in the poorest counties were 2-fold higher for cervical cancer and 40% higher for male lung and liver cancers during 2012-2016. Some states are home to both the wealthiest and the poorest counties, suggesting the opportunity for more equitable dissemination of effective cancer prevention, early detection, and treatment strategies. A broader application of existing cancer control knowledge with an emphasis on disadvantaged groups would undoubtedly accelerate progress against cancer.

16,028 citations

Journal ArticleDOI
TL;DR: Screening with the use of low-dose CT reduces mortality from lung cancer, as compared with the radiography group, and the rate of death from any cause was reduced.
Abstract: Background The aggressive and heterogeneous nature of lung cancer has thwarted efforts to reduce mortality from this cancer through the use of screening. The advent of low-dose helical computed tomography (CT) altered the landscape of lung-cancer screening, with studies indicating that low-dose CT detects many tumors at early stages. The National Lung Screening Trial (NLST) was conducted to determine whether screening with low-dose CT could reduce mortality from lung cancer. Methods From August 2002 through April 2004, we enrolled 53,454 persons at high risk for lung cancer at 33 U.S. medical centers. Participants were randomly assigned to undergo three annual screenings with either low-dose CT (26,722 participants) or single-view posteroanterior chest radiography (26,732). Data were collected on cases of lung cancer and deaths from lung cancer that occurred through December 31, 2009. Results The rate of adherence to screening was more than 90%. The rate of positive screening tests was 24.2% with low-dose CT and 6.9% with radiography over all three rounds. A total of 96.4% of the positive screening results in the low-dose CT group and 94.5% in the radiography group were false positive results. The incidence of lung cancer was 645 cases per 100,000 person-years (1060 cancers) in the low-dose CT group, as compared with 572 cases per 100,000 person-years (941 cancers) in the radiography group (rate ratio, 1.13; 95% confidence interval [CI], 1.03 to 1.23). There were 247 deaths from lung cancer per 100,000 person-years in the low-dose CT group and 309 deaths per 100,000 person-years in the radiography group, representing a relative reduction in mortality from lung cancer with low-dose CT screening of 20.0% (95% CI, 6.8 to 26.7; P=0.004). The rate of death from any cause was reduced in the low-dose CT group, as compared with the radiography group, by 6.7% (95% CI, 1.2 to 13.6; P=0.02). Conclusions Screening with the use of low-dose CT reduces mortality from lung cancer. (Funded by the National Cancer Institute; National Lung Screening Trial ClinicalTrials.gov number, NCT00047385.).

7,710 citations

Related Papers (5)
Frequently Asked Questions (2)
Q1. What have the authors contributed in "Clinically applicable deep learning strategy for pulmonary nodule risk prediction: insights into honors" ?

Wang et al. this paper proposed a clinically applicable DL-based algorithm, Filter-guided pyramid network ( FGP-NET ), and a practical strategy, Hierarchical-Ordered Network-ORiented Strategy ( HONORS ), which involves two steps for two different clinical scenarios ( i.e., screening and routine diagnostic scenarios ). 

To further applied FGP-NET to different clinical settings, a novel two-step strategy—Hierarchical-Ordered Network-ORiented Strategy ( HONORS ) was proposed. Based on FGP-NET, the authors further proposed a two-step strategy—HONORS, which is promising to optimize clinical workflow and realize personalized precise treatment of pulmonary 11 / 20 nodules. Future studies are warranted to prospectively assess the performance and generalizability of the algorithm at a variety of sites in real-world-use scenarios, and determine how the HONORS would impact diagnostic accuracy and clinical workflow. It may have potential to accelerate the process of pulmonary nodule diagnosis and to free doctors, nurses and other healthcare professionals to focus on providing real care for patients.