Closed cdrage closed 2 months ago
You can curl... I'm assuming you have enough ram on your macbook pro....
It's late for me, but you could just play around with:
ramalama run llama3
to start
You started a server on port 8080 there, it's llama.cpp under the hood
You can curl... I'm assuming you have enough ram on your macbook pro....
It's late for me, but you could just play around with:
ramalama run llama3
to start
Ah that makes sense. I guess docs are on the way to show how to use it / simple hello world curl request
It's here:
https://github.com/containers/ramalama?tab=readme-ov-file#running-models
We should probably move that closer to the top of the README.md before Listing Models
Feel free to do that, I'll merge
We should also add this types of data to man ramalama-serve.
Eventually we want to look into adding ramalama to ai-lab-recipes, and hopefully the ramalama serve AI Models can be used for all of the different recipes.
After running serve:
I'm confused what to do next to be honest...
What are the next steps / CURL / webview to try it out?
Also, I am a bit confused too as to where or not this is a container being ran in the background or not, or whether this is native? I had thought that
serve
does it via containers only on the host system?It ran, but there's nothing on podman ps