-
-
Notifications
You must be signed in to change notification settings - Fork 66
Closed
Labels
Description
These changes will make it easier to port over pretrained weights for the models from PyTorch.
- Right now, GoogLeNet matches the implementation in the paper, which does not use batch normalisation. However, torchvision uses batch normalisation layers after the convolution layers in both the inception block and the stem. It also uses
bias = false
for the convolution layer andeps = 0.001
for the batch normalisation layer. We could decide if similar to VGG, we can have a toggle for the batch normalisation in the model. (closed by Tweak GoogLeNet to match the torchvision implementations #205) - InceptionV3, InceptionV4 and InceptionResNetv2 have the right placement of normalisation layers but similar to GoogLeNet, use
bias = false
for the convolution layer andeps = 0.001
for the batch normalisation layer. (should be closed by Implementation of EfficientNetv2 and MNASNet #198)