daveshap / ACE_Framework

ACE (Autonomous Cognitive Entities) - 100% local and open source autonomous agents
MIT License
1.45k stars 206 forks source link

COST TO RUN & DEPENDENCE on OpenAI #98

Closed jayfalls closed 1 year ago

jayfalls commented 1 year ago

When watching Dave's demo of the project, a big standout were his remarks of timing out the API when just running the demo briefly, and seeing the amount of inferences that will need to be generated.

I don't think this limitation is necessary, and depending on a third party is not ideal. The limitation should rather be the amount of compute available, and getting this to run on consumer hardware would be the best.

As such, I suggest using the dolphin-2.1-mistral-7b model. Specifically a quantised version that can run with a maximum ram requirement of only 7.63 GB and a download size of only 5.13gb. Using the llama-cpp-python bindings, which meets the project requirements of only being in python.

There are benefits to doing it this way:

And benefits to this model specifically

This is just a suggestion, and this model will become outdated within the week. But I think that this is truly the right way to go.

daveshap commented 1 year ago

This does not belong here. Please move it to the Discussions tab.