laelhalawani / gguf_llama

Wrapper for simplified use of Llama2 GGUF quantized models.
https://pypi.org/project/gguf_llama
Other
5 stars 1 forks source link

various refactoring changes to enable easier and more intuitive inference #9

Closed laelhalawani closed 9 months ago

laelhalawani commented 9 months ago

made the limit on input tokens optional guardrail barrier ( can be set up outside initialization with a dedicated method .set_max_tokens()