-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is the D right ? #14
Comments
Yes, it is right since these numbers are divided by the |
oh, really thanks for answer~ but, emmmm, i also refer to another implement: https://github.com/kuangliu/pytorch-cifar/blob/master/models/resnext.py And the numbers of conv with stride 3 is less than that stride 1, even though it is also divided into some groups. I check the network architecture by thanks again~ |
Please make sure that you are executing with the correct commandline parameters. For
Which exactly corresponds to the paper (https://arxiv.org/pdf/1611.05431.pdf): |
Yes, i think when but in your train.py,
I think it is wrong, in a bottlenect, the channel number of the middle conv should be half of that of the output conv. But under this parameter setting, it is twice. sorry, i don't know how to change lines, but you can test by yourself. |
Resnext bottlenecks are a bit different, if you ask for a base width of 64 and a cardinality of 8, this is 64*8 = 512. These 512 will be divided in 8 groups of 64 channels. Maybe I am wrong, could you execute the oiriginal torch code and compare with mine to make it sure? |
Running the torch code is a bit troublesome, but i think you are right. I test a total of three implement code: And your network architecture is same as the third one, but different from the second. After i finish my work, i will check again. Really thanks for your answer~ |
ResNeXt.pytorch/models/model.py
Line 39 in 48c19fb
Hi, This may be a stupid question. I did not read the original paper, but i think the channels of the conv layer with stride 3 should be less than that with stride 1, to reduce the computational complexity.
I print the channels after line 39:
print(widen_factor, in_channels, D, out_channels)
and the output:
4 64 512 256
4 256 512 256
4 256 512 256
4 256 1024 512
4 512 1024 512
4 512 1024 512
4 512 2048 1024
4 1024 2048 1024
4 1024 2048 1024
Is that right? thanks for answer
The text was updated successfully, but these errors were encountered: