Closed dbugger29 closed 6 months ago
@dbugger29 you are right, this had me puzzled for a second, you need model.eval()
... open_clip models default to .train() mode like many torch model libs. Why is this needed? convnext doesn't have batchnorm like the Resnet models, but it does have stochastic depth active by default which will inject randomness in train mode.
Thank you :D
Hello,
@rwightman, I come back with a similar issue with: ViT-H-14-378-quickgelu
. Can you help me with this also?
Also, I am trying the latest models from https://github.com/facebookresearch/MetaCLIP/, I cannot load:
open_clip.create_model_and_transforms('ViT-bigG-14-quickgelu', pretrained='metaclip_fullcc')
Any kind of help would be appreciated :D
thank you
Can you help me with the following:
I receive different values for the same image and with the same class names.
These are the results after I ran the same code twice:
This does not seems to be the case with other models. Any help would be appreciated 🙇♀️