Closed tcsenpai closed 2 months ago
Thanks for preparing the ollama version, you can significantly reduce the amount of misses by using "format": "json"
for the chat completion API. Even 8b can handle the reasoning stream with that.
Thanks for the comment. I also added a perplexity version of the app, in separate scripts so if @bklieger-groq wants to merge the projects he can do it without replacing the whole app.py.
I just tested the perplexity fork and it works quite good (except formatting errors) if you want to give it a try.
I've integrated your implementation with a few tweaks (accepting more relaxed generation, more parameters) in Harbor v0.1.24, thanks!
Happy to accept PRs here with alternative model hosting options! To keep the repo simple, I may move them into a dedicated folder, so the root has g1 as default and if you want to explore other options you can try them. Alternatively, happy to link to these forks in the README, but merging with this project would be ideal so that any improvements automatically sync across the options.
Happy to accept PRs here with alternative model hosting options! To keep the repo simple, I may move them into a dedicated folder, so the root has g1 as default and if you want to explore other options you can try them. Alternatively, happy to link to these forks in the README, but merging with this project would be ideal so that any improvements automatically sync across the options.
For me too the best way is to merge!
I am preparing a PR, what do you think about having a launcher script that defaults to Grok but is able to launch Ollama and Perplexity too?
Ideally I would create three different modules and converge them in a main method, but I am really a newbie in starlit/gradio, so I am not able to do that.
Let me know, so I can prepare the pr accordingly : )
Update: I tried to draft a launcher.py file that should be simple enough to allow multi modes. I changed my repo to multi1 to keep it simple. I am opening a PR so we can continue there (pr here)
We may want to consider using LiteLLM so that we can support just about everything without custom code for vendors.
Update: I tried to draft a launcher.py file that should be simple enough to allow multi modes. I changed my repo to multi1 to keep it simple. I am opening a PR so we can continue there (https://github.com/bklieger-groq/g1/pull/6)
I would prefer to keep this repository named "g1" and keep the quickstart process simple with one default app. I like the launcher and alternative options, so I am happy to merge into a folder called "mult1" that has these options. If you make that change, I will review and merge the PR.
Update: I tried to draft a launcher.py file that should be simple enough to allow multi modes. I changed my repo to multi1 to keep it simple. I am opening a PR so we can continue there (#6)
I would prefer to keep this repository named "g1" and keep the quickstart process simple with one default app. I like the launcher and alternative options, so I am happy to merge into a folder called "mult1" that has these options. If you make that change, I will review and merge the PR.
Ok I am doing it. If that's ok with you, I will keep my repo as a standalone project with credits to your one, anyway
Sure! Actually, I am happy to link to your project in the README if thats preferable. I think its more intuitive to have mult1 in one place instead of two.
Yep I agree! Feel free to link it, so that when I will update it, it will be in one place. To @jjafuller I will work to implement that in multi1, then
Great! I will close out the PR then.
Added to the README here: https://github.com/bklieger-groq/g1/commit/62735af1400fa1f8d821c5093c88326b668ec36a
Looking forward to seeing how Mult1 advances!
multi1 doesnt allow for Issues though...
multi1 doesnt allow for Issues though...
Oh my bad! Opening them
Hello!
Not really an issue here, but just a note to inform you about a fork I made (https://github.com/tcsenpai/ol1) that uses local Ollama.
The fork is still in testing phase, but I thought it would just be fair to inform you as the original creator of the project.
If you think this can be merged in one project (maybe with configurable options in the frontend?) let me know, else we can continue as two separate projects : )
Compliments again for the wonderful idea!