futo-org / whisper-acft

MIT License
74 stars 3 forks source link

Checkpoints on the Hugging Face Hub? #2

Closed sanchit-gandhi closed 2 months ago

sanchit-gandhi commented 3 months ago

Hey @abb128 - congrats on this great project. It was super interesting reading how you tackled the problem of a fixed context length for Whisper!

I'm very excited to try the model in the Transformers library. However, I noticed that neither the Whisper cpp nor Transformers weights are not published on the Hugging Face Hub. Would you be open to pushing your weights to the Hugging Face Hub, in order to:

  1. Increase the visibility of your work through a Hub presence
  2. Reduce the barrier of entry to using your checkpoints, by enabling anyone in the community to load your model with two lines of code:
    
    from transformers import WhisperForConditionalGeneration

model = WhisperForConditionalGeneration.from_pretrained("your-username/your-model-name")


It's super easy to push the Transformers weights to the Hub with one line of code:
```python
from transformers import WhisperForConditionalGeneration

model = WhisperForConditionalGeneration.from_pretrained("./local/path/to/checkpoint")
model.push_to_hub("your-model-name")

For the Whisper cpp weights, you can follow this guide. You can also then draft a pretty model card that explains how to use the model, e.g. in the style of distil-large-v3.

abb128 commented 2 months ago

We've uploaded checkpoints to our HF: https://huggingface.co/futo-org