Closed c0derm4n closed 3 months ago
We have a number of settings in GraphRAG to tune parallelization and token consumption, but they assume you are working with a single API endpoint. You can add llm
config blocks to any step and pass in new config which may help, but more generally I think you're looking for a load-balancing system that spreads requests across multiple endpoints within a single verb. We don't plan to support this specifically because it is more of an infrastructure concern. I would suggest the use of a proxy or gateway that can do this without GraphRAG needing to be aware.
Is there an existing issue for this?
Describe the issue
The speed of constructing graphs is too slow now, especially when using larger local models
Steps to reproduce
No response
GraphRAG Config Used
Logs and screenshots
No response
Additional Information