Deep Residual Learning for Image Recognition
Citations
12,770 citations
Cites background or methods from "Deep Residual Learning for Image Re..."
...of COCOs weird average mean AP metric it is on par with the SSD variants but is 3 faster. It is still quite a bit behind other backbone AP AP 50 AP 75 AP S AP M AP L Two-stage methods Faster R-CNN+++ [5] ResNet-101-C4 34.9 55.7 37.4 15.6 38.7 50.9 Faster R-CNN w FPN [8] ResNet-101-FPN 36.2 59.1 39.0 18.2 39.0 48.2 Faster R-CNN by G-RMI [6] Inception-ResNet-v2 [21] 34.7 55.5 36.7 13.5 38.1 52.0 Faster...
[...]
...owerful than Darknet19 but still more efficient than ResNet-101 or ResNet-152. Here are some ImageNet results: Backbone Top-1 Top-5 Bn Ops BFLOP/s FPS Darknet-19 [15] 74.1 91.8 7.29 1246 171 ResNet-101[5] 77.1 93.7 19.7 1039 53 ResNet-152 [5] 77.6 93.8 29.4 1090 37 Darknet-53 77.2 93.8 18.7 1457 78 Table 2. Comparison of backbones. Accuracy, billions of operations, billion floating point operations per...
[...]
[...]
11,343 citations
Cites background or methods from "Deep Residual Learning for Image Re..."
...We train on 8 GPUs (so effective minibatch size is 16) for 160k iterations, with a learning rate of 0.02 which is decreased by 10 at the 120k iteration....
[...]
...We hope such rapid training will remove a major hurdle in this area and encourage more people to perform research on this challenging topic....
[...]
...We also report results on test-dev [28]....
[...]
8,730 citations
Cites methods from "Deep Residual Learning for Image Re..."
...Nie et al. (2016c) CNN Survival prediction; features from a Multi-modal 3D CNN is fused with hand-crafted features to train an SVM...
[...]
...Xie et al. (2015a) Nucleus detection Ki-67 CNN model that learns the voting offset vectors and voting...
[...]
...Xie et al. (2016b) Perimysium segmentation H&E 2D spatial clockwork RNN...
[...]
...Xie et al. (2015b) Nucleus detection H&E, Ki-67 CNN-based structured regression model for cell detection...
[...]
...Xie et al. (2016a) Nucleus detection and cell counting FL and H&E Microscopy cell counting with fully convolutional regression...
[...]
7,715 citations
Cites methods from "Deep Residual Learning for Image Re..."
...Plugging in ResNeXt-32x8d-101-FPN [40] as the RetinaNet backbone TABLE 3 Object Detection Single-Model Results (Bounding Box AP), versus State-of-the-Art on COCO test-dev backbone AP AP50 AP75 APS APM APL Two-stage methods Faster R-CNN+++ [31] ResNet-101-C4 34.9 55.7 37.4 15.6 38.7 50.9 Faster R-CNNw FPN [4] ResNet-101-FPN 36.2 59.1 39.0 18.2 39.0 48.2 Faster R-CNN by G-RMI [33] Inception-ResNet-v2 [38] 34.7 55.5 36.7 13.5 38.1 52.0 Faster R-CNNw TDM [30] Inception-ResNet-v2-TDM 36.8 57.7 39.2 16.2 39.8 52.1 One-stage methods YOLOv2 [8] DarkNet-19 [8] 21.6 44.0 19.2 5.0 22.4 35.5 SSD513 [9], [10] ResNet-101-SSD 31.2 50.4 33.3 10.2 34.5 49.8 DSSD513 [10] ResNet-101-DSSD 33.2 53.3 35.2 13.0 35.4 51.1 RetinaNet (ours) ResNet-101-FPN 39.1 59.1 42.3 21.8 42.7 50.2 RetinaNet (ours) ResNeXt-101-FPN 40.8 61.1 44.1 24.1 44.2 51.2 We show results for our RetinaNet-101-800 model, trained with scale jitter and for 1.5 longer than the same model from Table 1e....
[...]
...Following [4], we build FPN on top of the ResNet architecture [31]....
[...]
...For all experiments we use depth 50 or 101 ResNets [31] with a Feature Pyramid Network [4] constructed on top....
[...]
...ResNet-101 models are pre-trained on ImageNet1k; we use the models released by [31]....
[...]
...The one-stage RetinaNet network architecture uses a Feature Pyramid Network (FPN) [4] backbone on top of a feedforward ResNet architecture [31] (a) to generate a rich, multi-scale convolutional feature pyramid (b)....
[...]
5,904 citations
Cites background or methods from "Deep Residual Learning for Image Re..."
...level, i.e. the contents of individual modules of the CNN. Now, we explore design decisions at the macroarchitecture level concerning the high-level connections among Fire modules. Inspired by ResNet [13], we explored three different architectures: Vanilla SqueezeNet (as per the prior sections). SqueezeNet with simple bypass connections between some Fire modules. SqueezeNet with complex bypass conn...
[...]
... layers that deliver even higher ImageNet accuracy [14]. The choice of connections across multiple layers or modules is an emerging area of CNN macroarchitectural research. Residual Networks (ResNet) [13] and Highway Networks [29] each propose the use of connections that skip over multiple layers, for example additively connecting the activations from layer 3 to the activations from layer 6. We refer ...
[...]
References
73,978 citations
72,897 citations
"Deep Residual Learning for Image Re..." refers background in this paper
...Concurrent with our work, “highway networks” [42, 43] present shortcut connections with gating functions [15]....
[...]
49,914 citations
40,257 citations
30,843 citations
"Deep Residual Learning for Image Re..." refers background or methods in this paper
...We do not use dropout [14], following the practice in [16]....
[...]
...Recent evidence [41, 44] reveals that network depth is of crucial importance, and the leading results [41, 44, 13, 16] on the challenging ImageNet dataset [36] all exploit “very deep” [41] models, with a depth of sixteen [41] to thirty [16]....
[...]
...These plain networks are trained with BN [16], which ensures forward propagated signals to have non-zero variances....
[...]
...We adopt batch normalization (BN) [16] right after each convolution and before activation, following [16]....
[...]
...9, and adopt the weight initialization in [13] and BN [16] but with no dropout....
[...]