vespa-engine / vespa

AI + Data, online. https://vespa.ai
https://vespa.ai
Apache License 2.0
5.85k stars 604 forks source link

Plug & Play download & configuration of HuggingFace sentence transformer models #26696

Open eostis opened 1 year ago

eostis commented 1 year ago

(As usual, I'll give the point of view of small projects like WooCommerce ...)

The current way of using HF models is quite technical: build scripts to download models, convert to ONNX, deploy (https://github.com/vespa-engine/sample-apps/blob/master/transformers/src/python/setup-model.py)

I suggest to ease the process, to be able to use any HF sentence transformer model, by configuration rather than scripts.

Some possibilities:

jobergum commented 1 year ago

It's a bit technical, plus people want to quantize, not quantize, and so forth.

eostis commented 1 year ago

Closing inactive issue.

bratseth commented 1 year ago

I think we want to stick with a curated set with a guaranteed lifecycle: https://cloud.vespa.ai/en/model-hub

eostis commented 1 year ago

Reopened!

jobergum commented 1 year ago

Context @eostis ?

eostis commented 1 year ago

I think it would be useful to put some models on the shelves. The onnx generation is a real headache. Maybe by making Vespa Hub downloadable?

jobergum commented 1 year ago

We can probably put a few onnx models on HF. Do you think it is still painful to use Optimum to export models?

image

eostis commented 1 year ago

It can be tricky:

jobergum commented 1 year ago

Sure, I get that, but it's a general problem with the python world of dependencies that I don't think we can fully address :)