Closed bbartling closed 2 days ago
It is using ctransformers . you can read how to use gpu there. For your particular use case, i would suggest you to see this repo of mine . Its currently based on openai, but i am working on using opensource model. Instead of flask frontend a streamlit frontend can be used.
Nice work! How do you setup the demo of your app? Is that hosted on GitHub?
Very cool.
On Mon, Aug 7, 2023, 12:38 PM Talha Anwar @.***> wrote:
It is using ctransformers https://github.com/marella/ctransformers . you can read how to use gpu there. For your particular use case, i would suggest you to see this repo of mine https://github.com/talhaanwarch/doc_chat. Its currently based on openai, but i am working on using opensource model. Instead of flask frontend a streamlit frontend can be used.
— Reply to this email directly, view it on GitHub https://github.com/talhaanwarch/streamlit-llama/issues/1#issuecomment-1668313240, or unsubscribe https://github.com/notifications/unsubscribe-auth/AHC4BHP7WUFTX5WON7F6BGLXUERZPANCNFSM6AAAAAA3HAL574 . You are receiving this because you authored the thread.Message ID: @.***>
its hosted on vps. though the model is hosted locally
Any chance you could send me a link for that hosting service?
Would you have any advice between using llama-cpp-python or ctramformers for better results?
the hosting server has no rule, the model is hosting locally. I dont think there will be much difference between both.
what is VPS where you say its hosted. I cant find that vendor
llama model is hosted locally on my pc
@talhaanwarch nice job on the app! Looks great.
Can this run on GPU as well as CPU? This code appears to run WAY faster than
llama-cpp-python
bindings.Curious to ask if you would be interested in collaborating at all? I am working on a similar streamlit app trying to figure out how to "chat" with your own data in creating Word embeddings. https://github.com/bbartling/my-own-llm
See make_embeddings.py in my repo, its a learning experiment for me...if you are interested I could merge my into your repo on a PR or something.