Closed grencez closed 1 year ago
I tried the 13B SuperCOT LoRA and it seems to improve chat quality even without the instruction/response format. Though if you let it go on its own, it'll often try to add end-of-message strings like ### End of Transcript Continuation
, presumably because its preferred format has lots of ###
sections.
That said, the LoRA takes a while for me to load because of the higher-precision base model. My test machine's disk is kinda slow, but I guess that isn't the case for everyone.
Not sure how useful this is without instruction/response format (#5), but it seems worth doing. It's also a good excuse to bump the llama.cpp version.