Closed PawelSzpyt closed 4 months ago
I think if you add the --use-temp-file
argument, it should work:
True, I just successfully converted 8x22b using convert-hf-to-gguf.py, thanks for help. I close the issue as I assume that is the correct way to convert mixtral and llama3. I guess I'll close this ticket.
If I got it right, we should convert Llama3 with "convert-hf-to-gguf.py". This uses a ton of memory and my Mac Studio M1 Ultra with 128GB VRAM is unable to convet Llama3-70b to f32. Luckily it worked for f16 (although it hit swap very hard even with f16). I am unable to convert Mixtral 8x22b with this script at all (process gets killed at 38 part out of 59). So I wanted to ask a few questions:
Cheers, keep up the good work :)