prolego-team / neo-sophia

Applying the latest advancements in AI and machine learning to solve complex business problems.
BSD 3-Clause "New" or "Revised" License
66 stars 26 forks source link

Local LLM agent experimentation #116

Closed bdzimmer closed 9 months ago

bdzimmer commented 11 months ago

Updates for a version of bank_agent that uses a local LLM. I've gotten the best results by far with this model: phind-codellama-34b-v2. You can download the quantized version from HuggingFace Hub here: https://huggingface.co/TheBloke/Phind-CodeLlama-34B-v2-GGUF (phind-codellama-34b-v2.Q4_K_M.gguf) and put it in the MODELS_DIR_PATH defined in project

Additional requirements:

Entry points:

If you're not running on Apple Silicon, you should probably set GPU_ENABLE = False.

jmpounders commented 11 months ago

Is llama_cpp the only additional package that needs to be installed? Recommend updating the requirements.txt, esp if a certain version is needed.

jmpounders commented 11 months ago

Is llama_cpp the only additional package that needs to be installed? Recommend updating the requirements.txt, esp if a certain version is needed.

Reading the description closer now... can you update the requirements for llama_cpp and still get the right cmake flags, or should that always be a one-off installation?

bdzimmer commented 11 months ago

@jmpounders I'll look into it, might have to make an install script since it's setting environment variables but there may be a way to do that in requirements.txt. That being said, might be better to keep the metal flags separate since that's platform specific. And I'm realizing I should have made a note about that somewhere in the PR, heh.