Closed tmsingson closed 9 months ago
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment.
Is there any way to get this for Exl2?
This was closed as stale. Did it ever get implemented @oobabooga? Literally this is driving me to use KoboldCpp. As soon as you hit context limits in Oobabooga it becomes obnoxious in comparison. :/
Is there any way to get this for Exl2?
Wish too know too
Thanks @oobabooga you rock
Description
About 10 days ago, KoboldCpp added a feature called Context Shifting which is supposed to greatly reduce reprocessing. Here is their official description of the feature:
Any chance this gets added to Ooba as well?
Additional Context
Reddit thread: https://www.reddit.com/r/LocalLLaMA/comments/17ni4hm/koboldcpp_v148_context_shifting_massively_reduced/ llama.cpp pull: https://github.com/ggerganov/llama.cpp/pull/3228 kobold.cpp 1.48.1 release: https://github.com/LostRuins/koboldcpp/releases/tag/v1.48.1