nlpxucan / WizardLM

LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
9.11k stars 711 forks source link

WizardCoder hallucinations or bug in inference params? #169

Closed Extremys closed 11 months ago

Extremys commented 11 months ago

WizardCoder is really subject to expose roles from the prompt template and continue the discussion like this:

image

Do you have an idea what could cause this undesirable behavior? How to avoid that? I am using the fastchat repo for inference reusing alpaca prompt template, do you see any bug? https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py#L378C26-L378C26

Thanks.

ChiYeungLaw commented 11 months ago

Your template has a problem. The stop token of WizardCoder <|endoftext|> is different from Alpaca </s>.

Extremys commented 11 months ago

thanks for your feedback Chi, I think there is an other because switching from to <|endoftext|> do not resolve the tricky behavior :) any idea?

ChiYeungLaw commented 11 months ago

I try to say Hi!. It seems to work well. image image

ChiYeungLaw commented 11 months ago

My suggestion is to use the raw template rather than the template from FastChat. Raw template:

Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
{instruction}

### Response:
Extremys commented 11 months ago

Hello Chi, I found the root cause thanks for your feedback it was coming from the worker implementation :)