igniterealtime / openfire-llama-plugin

LLaMA 2 Inference Engine for Openfire
Apache License 2.0
2 stars 2 forks source link

First Test and feedback, Pls add the possibility to connect to an independant running LLama server URL #1

Open ClaudeStabile opened 11 months ago

ClaudeStabile commented 11 months ago

Dele,,

This is my first test with this plugin. Please : do not use that with DOCKER !!! Console issue : Console language issue : my console is in French screenshot12

This is downloading a Hugging face model. Please do no do that on your docker container as it will be a killer.... app. We need to have a separate volume so large files are mounted outside the container. This is CPU based so will not be terrific for performances

Perfect world i dream about :

FYI : I am also capable to build Llama full GPU docker containers & servers. For now i run this at home but if we have serious sponsors we can do a Llama servers in datacenter. Cost for server, i quote that 30-70kCHF/1U Box with 48-96GB GPU nvidia. I have the supplier

Thanks for doing this great JOB :) I need to have the possiblity to enter my sever URL as https://chatgpt.free-solutions.ch or http://chatgpt.free-solutions.ch:8080

Congrats