When I run the GQTP W4A16 LLAMA 2 7B on A100, I have this Issue. I didn't get any bug report. Is this memory issue?
quantizing weights: 78%|█████████████████████████████████████████████████████████████████████████████████▎ | 25/32 [18:16<12:14, 104.91s/it]
collecting calibration activations in model.layers.25: 77%|█████████████████████████████████Killed
When I run the GQTP W4A16 LLAMA 2 7B on A100, I have this Issue. I didn't get any bug report. Is this memory issue? quantizing weights: 78%|█████████████████████████████████████████████████████████████████████████████████▎ | 25/32 [18:16<12:14, 104.91s/it] collecting calibration activations in model.layers.25: 77%|█████████████████████████████████Killed