facebookresearch / SpinQuant

Code repo for the paper "SpinQuant LLM quantization with learned rotations"
Other
164 stars 14 forks source link

Questions about replication experiments. #5

Open BenchuYee opened 4 months ago

BenchuYee commented 4 months ago

Are the code and parameters in this repository consistent with the parameters used in the experiments described in the paper? I conducted an experiment on an A100 using the provided command "bash 10_optimize_rotation.sh meta-llama/Llama-2-7b 4 4 4" keeping the parameters unchanged in the script. However, the results I obtained differ somewhat from the results reported in the paper. This is my command "bash 10_optimize_rotation.sh meta-llama/Llama-2-7b 4 4 4", and I only change the nproc_per_node to 1. When I use the eval command "2_eval_ptq.sh $model_name 4 4 4", the ppl is about 8.49. The model I use is NousResearch/Llama-2-7b-chat-hf.

cokeshao commented 2 months ago

+1

for W4A4KV4, here is my result: I change nproc_per_node to 4 and run the code on 4*RTX 4090.

meta-llama/Llama-2-7b-hf meta-llama/Meta-Llama-3-8B
in paper 5.9 7.3
mine 6.81 8.06
WeiMa01 commented 3 weeks ago

+1 for W4A4KV4, here is my result: with the default config and run the code on 8*A100, got the following result:

meta-llama/Llama-2-7b-hf
in paper 5.9
mine 6.85

would you please share the config about the PPL 5.9 in the paper?

testworldagain commented 1 week ago

+1