Closed 0wwafa closed 1 month ago
I also tried -r "<|im_end|>" same thing
Codegeex uses this prompt format:
[gMASK]
try using --chat-template zephyr
Codegeex uses this prompt format: [gMASK] <|system|> {system_prompt} <|user|> {prompt} <|assistant|>
try using --chat-template zephyr
Do you know why the [gMASK] token is required at the start?
The CodeGeex README doesn't even mention it, but without it, the model really doesn't work well.
@TheLapinMalin AFAIK it is a BOS token (begining of string token) which masks a set of instructions used for text generation.
Ok, thanks for the info, I couldn't find anything describing it in their docs. :)
This issue was closed because it has been inactive for 14 days since being marked as stale.
For anyone else struggling to get codegeex4 models working, don't bother going through the wiki instructions to add a new template because the list of supported templates is outdated. Codegeex4 IS supported. That's just not the name of the template. Codegeex4 is a fine-tuned ChatGLM4 model and happens to use the same template, so "--chat-template chatglm4" actually works. Has to be that exact string too, because that's the string llama.cpp checks for.
llama-cli -c 1024 -t 6 -m codegeex4-all-9b.q4_k.gguf -p "You are my assistant." -e -cnv --chat-template chatml
and it continues by itself. what am I missing?