Verg-Avesta / CounTR

CounTR: Transformer-based Generalised Visual Counting
https://verg-avesta.github.io/CounTR_Webpage/
MIT License
92 stars 9 forks source link

[Reproduce] Cannot Reproduce the results. #32

Open HeyLinyuhao opened 1 year ago

HeyLinyuhao commented 1 year ago

Hello. Thank you for the great work.

However, under the given instructions with the pretrained weight, I could not reproduce the reported result on FSC-147.

I followed your instructiond and my test results are "Current MAE: 23.07, RMSE: 104.73".

My results are so far away from your results in the paper. Is there any hyper-parameters different from yours?

Verg-Avesta commented 1 year ago

Could you provide further information about your results? For example, which checkpoint did you use and did you fine-tune the provided weights?

HeyLinyuhao commented 1 year ago

Hi, Chang

I download the ckpt from MAE, pretrain and fineturn in FSC with your code. The environment you provided cannot be installed exactly the same, I intsall the closest versions. I test with the 'checkpoint__finetuning_minMAE.pth'

Cheers, Yuhao

Chang Liu @.***> 于2023年7月12日周三 17:14写道:

Could you provide further information about your results? For example, which checkpoint did you use and did you fine-tune the provided weights?

— Reply to this email directly, view it on GitHub https://github.com/Verg-Avesta/CounTR/issues/32#issuecomment-1632012271, or unsubscribe https://github.com/notifications/unsubscribe-auth/ALAFVAYNAOAJN33VHGTQIRDXPZIVPANCNFSM6AAAAAA2HBULHU . You are receiving this because you authored the thread.Message ID: @.***>

Verg-Avesta commented 1 year ago

I am not sure what the problem could be now.

Maybe you can try whether the fine-tuned weights I provided could produce the correct results? If the results are close, the environment should be OK.

After that, you can also try to visualize the performance of the model after MAE pre-training, and check whether the model has learned to reconstruct the masked images.

HeyLinyuhao commented 1 year ago

Hi I use the wanb to record, the visulisation is okay, both reconstruction and density map. I dont think package version can lead so large gap. How many times did you repeat your experiment? According to the issues on github, the tolerance seems to be large. Also, i will try the ckpt you provided.

Cheers

Chang Liu @.***> 于2023年7月12日周三 17:31写道:

I am not sure what the problem could be now.

Maybe you can try whether the fine-tuned weights I provided could produce the correct results? If the results are close, the environment should be OK.

After that, you can also try to visualize the performance of the model after MAE pre-training, and check whether the model has learned to reconstruct the masked images.

— Reply to this email directly, view it on GitHub https://github.com/Verg-Avesta/CounTR/issues/32#issuecomment-1632034891, or unsubscribe https://github.com/notifications/unsubscribe-auth/ALAFVA4OZBHIR7EQ3BP6MCTXPZKWTANCNFSM6AAAAAA2HBULHU . You are receiving this because you authored the thread.Message ID: @.***>

Verg-Avesta commented 1 year ago

I think an MAE difference between -1~1 should be OK. But your result differs too much.....

Besides, I have another suggestion. You can remove this line to unfreeze the encoder, and this will lead to an improvement in performance. But I don't think this will cause such a large performance gap.

laleh-samadfam commented 10 months ago

Hello,

I am trying to reproduce the results too, but I can not achieve the same results with the provided setting and pretrained model. I get MAE: 14.98, RMSE: 106.51 on FSC dataset.

Verg-Avesta commented 10 months ago

Hello,

I am not quite sure what the problem could be. Maybe you can try to remove this line and unfreeze the image encoder? You can also use a smaller learning rate and check the results of some earlier checkpoints.