MegEngine / InferLLM

a lightweight LLM model inference framework
Apache License 2.0
680 stars 84 forks source link

I got the error on centos 7: failed to tokenize string! #73

Open Mignet opened 9 months ago

Mignet commented 9 months ago

I got the error like this on centos 7: failed to tokenize string!

[root@VM-0-15-centos build]# ./llama -m chinese-alpaca-7b-q4.bin -t 2 main: seed = 1700961097 model is new , version = 1 load: n_vocab = 49954 load: n_ctx = 2048 load: n_embd = 4096 load: n_mult = 256 load: n_head = 32 load: n_layer = 32 load: n_rot = 128 load: model ftype = 2 total weight length = 4304332800 main: interactive mode on. sampling parameters: temp = 0.100000, top_k = 40, top_p = 0.950000, repeat_last_n = 64, repeat_penalty = 1.300000

== Running in chat mode. ==

Originally posted by @Mignet in https://github.com/MegEngine/InferLLM/issues/28#issuecomment-1826459644

Mignet commented 9 months ago

== Running in chat mode. ==

Breakpoint 1, inferllm::ModelImp::tokenize (this=0x51e1b0, text=" Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n", bos=true) at /root/InferLLM/src/core/model_imp.cpp:83 83 std::vector res; (gdb) c Continuing.

Breakpoint 2, inferllm::ModelImp::tokenize (this=0x51e1b0, text=" Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n", bos=true) at /root/InferLLM/src/core/model_imp.cpp:92 92 for (int i = 0; i < len; i++) { (gdb) i b Num Type Disp Enb Address What 1 breakpoint keep y 0x000000000040e993 in inferllm::ModelImp::tokenize(std::string const&, bool) at /root/InferLLM/src/core/model_imp.cpp:83 breakpoint already hit 1 time 2 breakpoint keep y 0x000000000040e9fc in inferllm::ModelImp::tokenize(std::string const&, bool) at /root/InferLLM/src/core/model_imp.cpp:92 breakpoint already hit 1 time (gdb) c Continuing. failed to tokenize string!

Mignet commented 9 months ago

cpu 2 core, mem 4 G