aws / sagemaker-huggingface-inference-toolkit

Apache License 2.0
222 stars 60 forks source link

Using custom inference script and models from Hub #102

Open Tarun02 opened 9 months ago

Tarun02 commented 9 months ago

Hello,

Can we use custom inference script without downloading the model to the S3 ? From the link, it says that we need to download the model artifacts and push them to S3 before using the custom inference script.

This will add considerable overhead depending on the model size.

babeal commented 8 months ago

+1 if I have to download the weights and repackage them, it kind of defeats the purpose of this library. The embeddings models "feature-extraction" return the last hidden state which is missing a step. The only thing I can do is repackage everything. I can't even add a file "code/inference.py" to the huggingface repository as the FILE_LIST_NAMES prevents it from being downloaded.