Very Deep Convolutional Networks for Large-Scale Image Recognition
Citations
405 citations
Cites methods from "Very Deep Convolutional Networks fo..."
...A baseline Domain Adaptation [15] method is also evaluated using the features from the VGG fc1 layer....
[...]
...In addition, images are preprocessed similar to [25] before being fed into the VGG net....
[...]
...In all of our experiments, image features are extracted by running the 19 layer VGG [25] model pre-trained on ImageNet without fine-tuning....
[...]
...The image features from the fc1 layer of the VGG net are fed into the visual mapping gv(·)....
[...]
...We evaluate the performance of our convolutional classifier using features from different intermediate convolutional layers in the VGG net and report the results in Table (3)....
[...]
403 citations
402 citations
402 citations
Cites methods from "Very Deep Convolutional Networks fo..."
...We augment the layers of the ImageNet-pretrained VGG-16 network [29] or ResNet-101 [16] with our LRR architecture and fine-tune all layers via backpropagation....
[...]
402 citations
Cites background or methods from "Very Deep Convolutional Networks fo..."
...We will utilize this as a building block for handling LSTMs and CNN architectures such as VGG (Simonyan & Zisserman, 2014) widely used in practice....
[...]
...We empirically analyze the different choices of the three hyper-parameters and find the choice of γ0 = 7, σ 2 0 = 1, σ 2 = 1 for VGG-9 on CIFAR-10 dataset and γ0 = 10−3, σ20 = 1, σ 2 = 1 for LSTM on Shakespeare dataset lead to good performance in our experimental studies....
[...]
...We conduct an experimental study on the effect of E over FedAvg, FedProx, and FedMA on VGG-9 trained on CIFAR-10 under heterogeneous setup....
[...]
...In conducting the “freezing and retraining” process of FedMA, we notice when retraining the last FC layer while keeping all previous layers frozen, the initial learning rate we use for SGD doesn’t lead to a good convergence (this is only for the VGG-9 architecture)....
[...]
...We note that we ignore all batch normalization (Ioffe & Szegedy, 2015) layers in the VGG architecture and leave it for future work....
[...]
References
49,639 citations
21,729 citations
9,803 citations
9,775 citations
6,061 citations