Open oobabooga opened 1 month ago
For now, setting topP to 0.99 serves as a casual workaround.
The 5 outputs will be identical.
In your example you use seed=-1
. Could you confirm, that when not using top_p=1.0
the 5 outputs will be different, but every time it will be the same 5 outputs? I opened an issue here, which was not present in llama-cpp-python==0.2.9
Setting
top_p = 1
causes outputs to be identical even with a random seed. This was discovered by https://github.com/oobabooga/text-generation-webui/issues/6431#issuecomment-2409089861. See the full issue at https://github.com/oobabooga/text-generation-webui/issues/6431.Reproduction
The 5 outputs will be identical.
Verified with
llama-cpp-python==0.3.1
.