defenseunicorns / leapfrogai

Production-ready Generative AI for local, cloud native, airgap, and edge deployments.
https://leapfrog.ai
Apache License 2.0
253 stars 29 forks source link

chore(backend): unify `GPU_` deployment variables #611

Open justinthelaw opened 3 months ago

justinthelaw commented 3 months ago

Describe what should be investigated or refactored

There are mentions of GPU_LIMITS, GPU_REQUEST, and GPU_ENABLED across the repository. These should be cleaned up and unified under a single variable across the backends. My recommendation is to follow the whisper pattern for all backends.

Ensure all backends have a modifiable GPU_REQUEST Zarf variable for the delivery engineer.

Links to any relevant code

https://github.com/search?q=repo%3Adefenseunicorns%2Fleapfrogai+GPU_&type=code