Open brand17 opened 2 weeks ago
I am getting this error on inference stage wiht llama-3-8b if set load_in_4bit = False: https://colab.research.google.com/drive/1RUzN1ZNjuDi4y-HyT2W9v7ePM3tLI-Lo#scrollTo=QmUBVEnvCDJv
load_in_4bit = False
Hi for some reason it says I need access? When you load in 4 bit = False, you will consume a lot more VRAM so the notebook may have gone out of memory causing this error.
I am getting this error on inference stage wiht llama-3-8b if set
load_in_4bit = False
: https://colab.research.google.com/drive/1RUzN1ZNjuDi4y-HyT2W9v7ePM3tLI-Lo#scrollTo=QmUBVEnvCDJv