For some reason, specifying the dtype is still downloading all the other weights. This approach has some issues:
Is a waste of space and bandwidth (specially HF)
Prevents you to use other repos that has not converted the models to other dtypes like this one
file:///***/node_modules/@xenova/transformers/src/utils/hub.js:238
throw Error(`${message}: "${remoteURL}".`);
^
Error: Could not locate file: "https://huggingface.co/intfloat/multilingual-e5-large/resolve/main/onnx/model_quantized.onnx".
at handleError (file:///***/node_modules/@xenova/transformers/src/utils/hub.js:238:11)
at getModelFile (file:///***/node_modules/@xenova/transformers/src/utils/hub.js:471:24)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
at async constructSession (file:///***/node_modules/@xenova/transformers/src/models.js:123:18)
at async Promise.all (index 1)
at async XLMRobertaModel.from_pretrained (file:///***/node_modules/@xenova/transformers/src/models.js:793:20)
at async AutoModel.from_pretrained (file:///***/node_modules/@xenova/transformers/src/models.js:5519:20)
at async Promise.all (index 1)
at async loadItems (file:///***/node_modules/@xenova/transformers/src/pipelines.js:3279:5)
at async pipeline (file:///***/node_modules/@xenova/transformers/src/pipelines.js:3219:21)
System Info
tranformers v2.17.2 node v18.20.3
Environment/Platform
Description
For some reason, specifying the
dtype
is still downloading all the other weights. This approach has some issues:Reproduction