Open eostis opened 1 year ago
It's a bit technical, plus people want to quantize, not quantize, and so forth.
Closing inactive issue.
I think we want to stick with a curated set with a guaranteed lifecycle: https://cloud.vespa.ai/en/model-hub
Reopened!
Context @eostis ?
I think it would be useful to put some models on the shelves. The onnx generation is a real headache. Maybe by making Vespa Hub downloadable?
We can probably put a few onnx models on HF. Do you think it is still painful to use Optimum to export models?
It can be tricky:
Sure, I get that, but it's a general problem with the python world of dependencies that I don't think we can fully address :)
(As usual, I'll give the point of view of small projects like WooCommerce ...)
The current way of using HF models is quite technical: build scripts to download models, convert to ONNX, deploy (https://github.com/vespa-engine/sample-apps/blob/master/transformers/src/python/setup-model.py)
I suggest to ease the process, to be able to use any HF sentence transformer model, by configuration rather than scripts.
Some possibilities: