Closed sophosympatheia closed 5 months ago
It doesn't make a lot of sense that this would happen sometimes. I've pushed a commit that might fix it, though.
Thanks, Turbo! I looked into it on my end too, and like you said, it seems like it's a situation that shouldn't come up when things are working properly.
When I look at some of my past measurements taken from 70b models that didn't produce the error, the accuracy for the last layer MLP is > 0.9 for even the very low bpw measurements, whereas for this model the accuracy was < 1e-6 for all bpw measurements taken on that model.layers.79.mlp. That's what triggered the error: max(1e-6, 1 - (rfn_sum / rfn_count)) returned 1e-6, which is a float and doesn't have the item() method, and I can see where that result is not expected to occur under normal circumstances.
I proceeded with quantizing the model anyway and the final result appears to be okay. ¯_(ツ)_/¯
Thank you for committing a fix to this so quickly! I greatly appreciate all your work on the exllama project.
Sometimes when using the conversion script to produce a measurement file of a model that I merged together from other models, I'll encounter the following error at the very end of the measurement pass after it measures the last model layer but before it saves the measurement json file.
Any idea what I should be checking? I can provide more context if someone can steer me towards what you need.