front-seat / engage

Engage™️ with your City Council!
https://scc.frontseat.org/
Mozilla Public License 2.0
7 stars 0 forks source link

Summarize with Vicuna-7B and Vicuna-13B in the cloud #6

Closed davepeck closed 1 year ago

davepeck commented 1 year ago

There are sort of two approaches: run Vicuna locally, or host it in the cloud.

Running locally is fine (and easier) but forces me to manually run daily crawls/updates on my laptop. I'd really like updates to run in the cloud, preferably when I'm sleeping. :-)

So, at some point, I'll need to host Vicuna behind an API in the cloud.

Alas, the Internet has not been cooperative and completely solved this problem for me just yet — stuff is moving way too fast — so some exploration and head-scratching proved necessary. Self-hosted LLMs really are the bleeding edge.

After my exploration, I've concluded that building on top of @ggerganov's llama.cpp is the way to go. In particular:

Okay, now I've got Vicuna running in the cloud. Just a few more steps to put it all together:

davepeck commented 1 year ago

This was a bit more of an adventure than anticipated, but we are now summarizing with VIC13B, too.