Open maswadkar opened 5 hours ago
I think thats the holy grail of opensource LLMs
One should be able to download -> finetune -> serve via REST Endpoint (api)
and it should be done in single tool chain.
If you use our meta-reference
inference provider you can now do that using this: https://github.com/meta-llama/llama-stack/blob/main/llama_stack/providers/impls/meta_reference/inference/config.py#L34 although we don't quite support or advertise this super well (because you could technically put any random checkpoint in there with params.json not being correct or weights being wrong, etc.)
Hi, I have successfully done the cycle of
now my next assignment is to
my question is how can i change the checkpoint path to my custom downloaded model