Open empty2enrich opened 1 year ago
Hi @empty2enrich, which model are you using? Do you face the same issue when loading with transformers AutoModelForCausalLM
?
Something you could try is to add for example max_new_tokens=3
to the generate
call to check whether it is not simply the model generating a very long sequence. From your log (Using
max_length's default (4096)
) it could very well be the issue.
System Info
log: