Open atisharma opened 1 year ago
did you properly set up your path to your model in your auth? how are you sending your request?
Yes, I believe so. I am testing with curl.
curl --location --request POST 'http://localhost:8080/v1/chat/completions' --header 'Authorization: Bearer /sol/ll
m/LLaMA/7B/ggml-model-q4_0.bin' --header 'Content-Type: application/json' --data-raw '{
"model": "gpt-3.5-turbo",
"messages": [
{
"role": "system",
"content": "You are ChatGPT, a helpful assistant developed by OpenAI."
},
{
"role": "user",
"content": "How are you doing today?"
}
]
}'
curl: (52) Empty reply from server
and
$ ls /sol/llm/LLaMA/7B/ggml-model-q4_0.bin
/sol/llm/LLaMA/7B/ggml-model-q4_0.bin
Regarding the port 443 being reserved for system processes, you could do the following:
sudo setcap cap_net_bind_service=+ep `readlink -f \`which node\``
and then
npm start
@atisharma can you try the new ./test-installation.sh script to help validate if you're giving it the right path?
Should fix the whole Port issue. I have similar changes in a dev branch and it works fine
Still curl: (52) Empty reply from server
.
Your test script also hardcodes port 443 by the way.
good point @atisharma , let me update the script to be more flexible. and @th-neu , i'll take a quick look at your PR.
for now, @atisharma , what do u get if you change the script to your port? does the curl command return the proper value?
No, the empty reply was with using 8080 in both index.js and the test script.
i just updated the test script and added support for port – PORT=8080 npm run start
should do the trick. are you still running into the same issue?
@atisharma are you still running into the same issue?
First problem is that port 443 is usually reserved. I edited index.js to 8080.
Next problem is that it crashes on first request: