Closed ee2110 closed 1 year ago
It is a generative model, not an instruction-following model.
Try open-llama-7b-open-instruct instead.
It is a generative model, not an instruction-following model.
Try open-llama-7b-open-instruct instead.
Hi @gjmulder , thank you for your comment and the suggestion! I see, I wonder if the Open-LLaMA can handle QA task as mentioned in their paper, I have tested Vicuna-7B with the same questions and it performs better more stable. I certainly will also take a look at the open-llama-7b-open-instruct, thank you!
Vicuna-7B is an instruction fine-tuned model.
Hi, here is some examples of the problems:
I give the prompt:
'Q: What is the largest animal?\nA:'
The output from open_llama_7b as following:
or sometimes with prompt
'What is the name of the foods with strawberry topping?'
, the generated output is weird and keep repeating:However, with
max_new_tokens=100
I expect the output should be something like this:The largest animal is the blue whale.
The person is making French Strawberry Cake.
A snippet of my code:
I would like to know how to control the model to generate desired output as expecting? Thank you!