Closed teasgen closed 12 months ago
Approve, please
Sorry for the delay an thank you very much for the PR @teasgen . I will have a look right after fixing #109
Hi @teasgen ! Do you happen to know the best setting to use your PR for linear probe on ImageNet?
Hi @teasgen ! Do you happen to know the best setting to use your PR for linear probe on ImageNet?
Hi, unfortunately I haven't tested my PR on ImageNet. But you can efficiently find best hyperparameters using cli arguments. However, I used same setting for all datasets, so you can firstly try it
Hi @teasgen, working fine for me! the only thing that would be nice to keep is the default behavior, i.e. not specifying a validation dataset. Currently, it fails with an error message if validation set or validation proportion are not given. With this commit: https://github.com/LAION-AI/CLIP_benchmark/commit/396f8073f6c84ca230e4ecaa6d16db3e90a71d1c, I could make it working fine again, but I might have missed something. Could you please have a look/confirm ?
Thanks!
Hi @teasgen, working fine for me! the only thing that would be nice to keep is the default behavior, i.e. not specifying a validation dataset. Currently, it fails with an error message if validation set or validation proportion are not given. With this commit: 396f807, I could make it working fine again, but I might have missed something. Could you please have a look/confirm ?
Thanks!
Hi! Your commit looks good, I suppose now its alright. Could you please release new version to pypi as soon as pr will be merged?
Great, thanks @teasgen! yes, sure, will release a new version on pypi!
Merging, will add the other commit right after.
@teasgen @Danil328 available now in 1.6.0, pip install clip_benchmark==1.6.0
In the updated linear evaluation, the calculation process involves dividing the dataset into three parts: train, validation, and test. However, if the dataset does not already have a validation split, I will divide the train part into two sections based on the specified proportion. It means we will get more fair results. Also I've added regularization with openAI hyperparameter sweep (https://arxiv.org/pdf/2103.00020.pdf A.3). Now the results are more similar to openAI metrics for CLIP models (same paper, table 10)