Closed dapurv5 closed 2 months ago
Any plans to make it compatible with vLLM for faster generations?
Thank you for your suggestions! Our current project does not include a module for LLM inference acceleration, but we can consider adding it in the future.
Any plans to make it compatible with vLLM for faster generations?