Closed zsgj-Xxx closed 4 years ago
Hello,
I have not checked converge speed of models with and without CSP. However, all of my experiments follow the same setting as https://pjreddie.com/darknet/imagenet/. So the training epochs are totally same.
Thank you very much for your reply,
I want to do some small tests with CSP I tried to copy it on the pytorch, but the parameters were worse, I haven't found any problems yet How to modify the CSP method based on resnext?
the topology of resnet, resnext, and darknet are almost same. https://github.com/WongKinYiu/CrossStagePartialNetworks/issues/24#issuecomment-623125410 is for your reference.
Thank you for your work,
I just need to replace darknet_layer with resne(x)t_layer to get the result I need?:heart_eyes:
In addition, in this figure, after maxpooling, is ① CSP? But I think the parameter displayed is not split, but copy
yes.
i think there will be a convolutional layer behind ①. more details: https://github.com/WongKinYiu/CrossStagePartialNetworks/issues/18
I'm sorry that I've read the paper and the cfg file over and over again, but I still don't understand it
14 14 1024 - > whether two 7 7 1024 branches have also been trained
It looks like
14x14 is belong to partial transition layer in previous stage.
Due to the limitation of GPU devices, I only tested the model with epoch = 1, and found that compared with the traditional resnext model, the result of cspresnext model for an epoch is not satisfactory. Is it because of the residual link used that the model needs more time to learn