I've successfully trained llama with unsloth using my custom dataset. Also tested inference in google colab and successfully saved the merged 16bit model to hugging face.
However, i'm stuck trying to deploy to inference endpoint. To check if this was an issue with my trained model, i tried deploying the Unsloth llama model (unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit ) directly to an inference endpoint and I keep getting the same failure
Is this a build configuration issue? Or I've done something wrong with the advanced configuration when creating the endpoint
Hi,
I've successfully trained llama with unsloth using my custom dataset. Also tested inference in google colab and successfully saved the merged 16bit model to hugging face.
However, i'm stuck trying to deploy to inference endpoint. To check if this was an issue with my trained model, i tried deploying the Unsloth llama model (unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit ) directly to an inference endpoint and I keep getting the same failure
Is this a build configuration issue? Or I've done something wrong with the advanced configuration when creating the endpoint
Thoughts?