Open AmitRozner opened 6 months ago
It seems like all weights has not been loaded? Do you have the terminal output for the demo.
34B is relatively slow, it takes around 3 minute for a response on two A100. Mainly because 34b model provides long response.
It seems like all weights has not been loaded? Do you have the terminal output for the demo.
34B is relatively slow, it takes around 3 minute for a response on two A100. Mainly because 34b model provides long response.
A100 40G? or 80G?
Thanks for the repo and models! When trying to run demo.sh with the 34b model (commented and uncommented the relevant lines), I am getting nonsense output (with the example video and prompt):
It also takes like 10 minutes on 4 RTX 3090 GPUs. Any thoughts?