{
"error": "Request failed during generation: Server error: No local weights found in AP-2-lora with extension .safetensors",
"error_type": "generation"
}
However, if we use the absolute path, we are able to get a response
request:
Thanks for raising this issue @bjornjee! This sounds like a good improvement to me. Since you mentioned you would be open to submitting a PR, is this something you'd like to contribute?
Feature request/question
Expose ENV/flag in
lorax-server
andlorax-launcher
to set base path of adapter during inference.We currently tried to do a workaround by setting HUGGINGFACE_HUB_CACHE=/home/adapters . With reference to: https://github.com/predibase/lorax/blob/main/server/lorax_server/utils/sources/local.py#L26.
However, since we only save adapter weights to .bin file extension, we are getting the error during inference:
request:
error:
However, if we use the absolute path, we are able to get a response request:
Motivation
abstract away absolute path of adapters from users during inference with custom adapters which are downloaded locally into instance.
Your contribution
possible to prepare a PR.