SimJeg / FC-DenseNet

Fully Convolutional DenseNets for semantic segmentation.
487 stars 143 forks source link

How was number of trainable parameters calculated? #31

Open Faur opened 7 years ago

Faur commented 7 years ago

I have tried to re-implement the architecture described in the paper exactly, just in TensorFlow. But don't get the correct number of trainable parameters. I can't find where this is calculated, so I was hoping someone could help me out.

Paper: 56 layer: 1.5 mil. 103 layer: 9.4 mil

My implementation: 56 layer: 1.4 mil 103 layer: 9.2 mil

The discrepancy is small, so normally I wouldn't care, but I can't quite get the same performance results as in the paper, so perhaps this could help reveal any bugs in my code.

dongzhuoyao commented 6 years ago

could you tell me your best result?

Faur commented 6 years ago

image

There were some issues with the 103 layer implementation (to small batch size and image resolution)

xiaomixiaomi123zm commented 4 years ago

@Faur @dongzhuoyao Hi, do you still remember the FLOPs of the model when you runed fc-densenet, the reviewer of my paper asked me to write the parameters and FLOPs values, but I failed to run because the dataset could not be loaded, I look forward to your reply, thank you very much

Faur commented 4 years ago

@xiaomixiaomi123zm I am sorry, but I can't help you. I has been a while since I worked on this, and I don't have access to the code base as is.

But if the issue is just the dataset you should be able to make some dumme data with np.zeros quite easily - then you should be able to get the FLOPs of the model

xiaomixiaomi123zm commented 4 years ago

Hi, I don’t quite understand how to do it. Is it convenient for you to add my qq( 907675183 ) and tell me about it? I have tried a lot of ways, and I have not been successful, thank you very much!! @Faur