Closed roger-rodriguez closed 9 months ago
Hi @roger-rodriguez!
I am not sure I know how to help (maybe @patrickjohncyh ?) but I think this is a good question for the HF transformers repo.
Our model has the same architecture as the HF CLIP (https://huggingface.co/openai/clip-vit-base-patch32) so what works for the general model should work also for FashionCLIP.
Thanks for the quick reply @vinid! We can close this one. I ended up downloading the model and deploying it with a docker lambda.
Realize this might be out of scope but hoping someone can point me in the right direction.
I have deployed the Hugging Face model to sagemaker and I'm call it via a lambda function. However what are the inputs the model expects for zero shot image classification? Assuming I need an image url or base64 encoded input somewhere?
What should
payload
look like?Lambda code: