Open alomrani opened 6 months ago
Hi, I see that you use a local API sever for llama. How do you initialize such a server? Thanks
We use VLLM: https://github.com/vllm-project/vllm. You can refer to its documentation about server startup.
Hi, I see that you use a local API sever for llama. How do you initialize such a server? Thanks