harrisonvanderbyl / rwkv-cpp-accelerated

A torchless, c++ rwkv implementation using 8bit quantization, written in cuda/hip/vulkan for maximum compatibility and minimum dependencies
MIT License
303 stars 19 forks source link

converter failure #38

Open malv-c opened 11 months ago

malv-c commented 11 months ago

% python converter/convert_model.py models/RWKV-5-World-0.1B-v1-20230803-ctx4096.pth Quantizing att.key.weight: 100%|█████████████████████████████████████████████████████████| 12/12 [00:00<00:00, 58.41it/s] stacking weightsey.weight: 42%|████████████████████████▏ | 5/12 [00:00<00:00, 45.79it/s] Cleaning att.key.weight: 100%|████████████████████████████████████████████████████████| 12/12 [00:00<00:00, 60640.54it/s] Quantizing att.value.weight: 100%|███████████████████████████████████████████████████████| 12/12 [00:00<00:00, 66.89it/s] stacking weightsalue.weight: 58%|████████████████████████████████▋ | 7/12 [00:00<00:00, 65.76it/s] Cleaning att.value.weight: 100%|██████████████████████████████████████████████████████| 12/12 [00:00<00:00, 75459.74it/s] Quantizing att.receptance.weight: 100%|██████████████████████████████████████████████████| 12/12 [00:00<00:00, 55.76it/s] stacking weightseceptance.weight: 50%|█████████████████████████▌ | 6/12 [00:00<00:00, 50.13it/s] Cleaning att.receptance.weight: 100%|█████████████████████████████████████████████████| 12/12 [00:00<00:00, 83055.52it/s] Quantizing ffn.key.weight: 100%|█████████████████████████████████████████████████████████| 12/12 [00:00<00:00, 14.30it/s] stacking weightsey.weight: 100%|█████████████████████████████████████████████████████████| 12/12 [00:00<00:00, 14.85it/s] Cleaning ffn.key.weight: 100%|████████████████████████████████████████████████████████| 12/12 [00:00<00:00, 85019.68it/s] Quantizing ffn.value.weight: 100%|███████████████████████████████████████████████████████| 12/12 [00:00<00:00, 14.70it/s] stacking weightsalue.weight: 100%|███████████████████████████████████████████████████████| 12/12 [00:00<00:00, 14.81it/s] Cleaning ffn.value.weight: 100%|██████████████████████████████████████████████████████| 12/12 [00:00<00:00, 92182.51it/s] Quantizing ffn.receptance.weight: 100%|██████████████████████████████████████████████████| 12/12 [00:00<00:00, 77.98it/s] stacking weightseceptance.weight: 75%|██████████████████████████████████████▎ | 9/12 [00:00<00:00, 83.23it/s] Cleaning ffn.receptance.weight: 100%|█████████████████████████████████████████████████| 12/12 [00:00<00:00, 81180.08it/s] Quantizing att.output.weight: 100%|██████████████████████████████████████████████████████| 12/12 [00:00<00:00, 77.60it/s] stacking weightsutput.weight: 75%|█████████████████████████████████████████▎ | 9/12 [00:00<00:00, 82.97it/s] Cleaning att.output.weight: 100%|████████████████████████████████████████████████████| 12/12 [00:00<00:00, 107316.95it/s] Quantizing: 100%|██████████████████████████████████████████████████████████████████████████| 7/7 [00:03<00:00, 2.00it/s] saving: xbuf saving: embed python: /M/rwkv-cpp-accelerated/converter/cpp_save_tensor.cpp:81: void save(std::string, int64_t, int64_t, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&): Assertion `getSize(i,n_layers,n_emb) == tensors[i]->numel()' failed. Abandon (lcpu) /M/rwkv-cpp-accelerated

harrisonvanderbyl commented 11 months ago

Sorry, rwkv-cpp-accelerated does not support V5 yet :/

malv-c commented 11 months ago

yet thankd

Le ven. 18 août 2023 à 00:08, Harrison Vanderbyl @.***> a écrit :

Sorry, rwkv-cpp-accelerated does not support V5 yet :/

— Reply to this email directly, view it on GitHub https://github.com/harrisonvanderbyl/rwkv-cpp-accelerated/issues/38#issuecomment-1683041235, or unsubscribe https://github.com/notifications/unsubscribe-auth/AESIHJLBGXFRW3JGC72DPTLXV2I6JANCNFSM6AAAAAA3UFQA7U . You are receiving this because you authored the thread.Message ID: @.***>