Open WentaoTan opened 3 months ago
+1 I see the same issue too. tried with multiple settings but no luck. Any help would be appreciated.
+1 got same error
+1 got same error
Has anyone solved this problem?
The line attn_output = attn_output.reshape(bsz, q_len, self.hidden_size) can be modified to attn_output = attn_output.reshape(bsz, q_len, -1) to resolve the issue.
got same error
The line attn_output = attn_output.reshape(bsz, q_len, self.hidden_size) can be modified to attn_output = attn_output.reshape(bsz, q_len, -1) to resolve the issue.
I tried this method, but it doesn't work.
In the modeling_llama.py file I downloaded, the source code is:
attn_output = attn_output.reshape(bsz, q_len, -1)
The command I run: ''' python llama3.py --pruning_ratio 0.25 \ --device cuda --eval_device cuda \ --base_model home/Meta-Llama-3-8B \ --block_wise --block_mlp_layer_start 4 --block_mlp_layer_end 30 \ --block_attention_layer_start 4 --block_attention_layer_end 30 \ --save_ckpt_log_name llama3_prune \ --pruner_type taylor --taylor param_first \ --max_seq_len 2048 \ --test_after_train --test_before_train --save_model ''' When running to line 259 of the code, an error occurs: How to solve this problem? Looking forward to your reply!