Closed seolhokim closed 3 months ago
Oh you're calling the base model, which is to be expected (it will output gibberish). The instruct model is finetuned on <|begin_of_text|><|begin_of_text|><|start_header_id|>system<|end_header_id|>...
whilst the base model is not, so the base model will output gibberish - so hence why we have to finetune it!
Also the base model does not handle the llama-3 chat template well, since the tokens are untrained - see https://unsloth.ai/blog/phi3 for more details (the blog was for Phi-3, but has Llama-3 bug fixes)
Thank you for your kind answer. I should have checked that differences. I thought those are trained on same dataset. thank you again.
Nw!
Thank you for your work. However, I've noticed some performance issues that differ significantly when compared to the Llama 3.1 model. Specifically, I've observed the following problems:
Issue Description
Special Token Handling: The Unsloth model seems to have difficulty managing special tokens, leading to incoherent and less relevant outputs.
Incomplete and Nonsensical Responses: The model frequently produces incomplete responses or outputs that are nonsensical and meaningless. This issue affects the clarity and flow of the generated content.
Issue example
1.It is saying same thing again and again.
Reproduce
Result comparison