Open PicoCreator opened 1 year ago
<= 67.2 ms on 7B with partial GPU offload (3060 Ti 8G) is a huge win win
Hi, I converted a pth file to bin format, unfortunately it crashed with following error message:
% rwkv-cpp-node --modelPath ./rwkv-7b-369-Q5_1.bin
--------------------------------------
Starting RWKV chat mode
--------------------------------------
Loading model from ./rwkv-7b-369-Q5_1.bin ...
Unsupported file version 101
/Users/eugene/Desktop/RWKV/rwkv.cpp/rwkv.cpp:211: version == RWKV_FILE_VERSION
zsh: segmentation fault rwkv-cpp-node --modelPath ./rwkv-7b-369-Q5_1.bin
Is it the compatibility issue mentioned here? Thanks (Looking forward to the world version :-) )
Yup, the new rwkv.cpp - is now merged into v2 (publishing now)
Merged in : 74655dee0d3824162d7926882ae95d1371a11fa2
This resolves all issues, EXCEPT support for world model tokenizer (needs a new JS tokenizer)
Yup, the new rwkv.cpp - is now merged into v2 (publishing now)
update the docs please
Merged in : 74655de
This resolves all issues, EXCEPT support for world model tokenizer (needs a new JS tokenizer)
waiting for it ~~
Hi, I reinstalled the package, but when I run it, I get only following result:
% rwkv-cpp-node --modelPath ./rwkv-7b-369-Q5_1.bin
--------------------------------------
Starting RWKV chat mode
--------------------------------------
Loading model with {"path":"./rwkv-7b-369-Q5_1.bin","threads":4,"gpuOffload":0} ...
The following is a conversation between the User and the Bot ...
--------------------------------------
? User: Hi how are you
Bot: <|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|>
Did I convert the model wrong?
Hi, I reinstalled the package, but when I run it, I get only following result:
% rwkv-cpp-node --modelPath ./rwkv-7b-369-Q5_1.bin -------------------------------------- Starting RWKV chat mode -------------------------------------- Loading model with {"path":"./rwkv-7b-369-Q5_1.bin","threads":4,"gpuOffload":0} ... The following is a conversation between the User and the Bot ... -------------------------------------- ? User: Hi how are you Bot: <|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|>
Did I convert the model wrong?
world models not support yet
Hi, I reinstalled the package, but when I run it, I get only following result:
% rwkv-cpp-node --modelPath ./rwkv-7b-369-Q5_1.bin -------------------------------------- Starting RWKV chat mode -------------------------------------- Loading model with {"path":"./rwkv-7b-369-Q5_1.bin","threads":4,"gpuOffload":0} ... The following is a conversation between the User and the Bot ... -------------------------------------- ? User: Hi how are you Bot: <|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|>
Did I convert the model wrong?
world models not support yet
It is just a normal finetuned raven model
Latest version of https://github.com/saharNooby/rwkv.cpp has new quantization format (breaking change?) and GPU offload (!!!) Since this might be potentially breaking changes, its gonna be a v2 update.