Closed virentakia closed 2 weeks ago
There's no magic solution to increasing context length beyond what the model was designed for. You can load the model with double its base context length and get it to remain coherent by setting rope alpha = ~2.63. However the further you try to stretch a model beyond the context length it was designed for, the further its quality will degrade.
The inference speed is amazing - excellent work.
Is it possible to increase the context length of models?
Using the "Solar" model -
https://huggingface.co/bartowski/Nous-Hermes-2-SOLAR-10.7B-exl2/tree/8_0
- with the following config:Wondering what possible options are available? Are there examples of these options available?