ztxz16 / fastllm

纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行
Apache License 2.0
3.31k stars 339 forks source link

Is there any accuracy loss when converting to flm model? #305

Open empty2enrich opened 1 year ago

empty2enrich commented 1 year ago

I convert llama2-7b using fastllm_pytools.torch2flm, The inference result looks wrong, Also inconsistent with inference results using llama2-7b directly:

prompt: The president of the United States is

generate result:

### Instruction:
The president of the United States is

### Response:
The president of the United States

The president of the United States is

### Instruction:
The president of the United States is

### Response:
The president of the United States is

### Instruction:
The president of the United States is

### Response:
The president of the United States is

### Instruction:
The president of the United States is

### Response:
The president of the United States is

The president of the United States is

### Instruction:
The president of the United States is

### Response:
The president of the United States is
TylunasLi commented 1 year ago

you should pay atention to difference of the prompt building process between original python code and fastlllm code. for llama2-7b:

for llama2-7b-chat-hf,