dusty-nv / NanoLLM

Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector DB, and RAG.
https://dusty-nv.github.io/NanoLLM/
MIT License
196 stars 31 forks source link

Ratelimit Node #51

Open JIA-HONG-CHU opened 3 weeks ago

JIA-HONG-CHU commented 3 weeks ago

I found that switching the on_demand parameter of the ratelimit node to true or false doesn't make any difference. Has this feature been implemented? I also didn't see any implemented function in the parent node plugin. Could you help me check? Tks.