main: build = 775 (d7d2e6a)
main: quantizing './zh-models/7B/ggml-model-f16.bin' to './zh-models/7B/ggml-model-q4_K.bin' as Q4_K
llama.cpp: loading model from ./zh-models/7B/ggml-model-f16.bin
llama.cpp: saving model to ./zh-models/7B/ggml-model-q4_K.bin
========================= Tensor sizes 4096 x 49954 are not divisible by 256
This is required to be able to use k-quants for now!
========================================================================================
llama_model_quantize: failed to quantize: Unsupported tensor size encountered
main: failed to quantize model from './zh-models/7B/ggml-model-f16.bin'
[ 1/ 291] tok_embeddings.weight - 4096 x 49954, type = f16,
提交前必须检查以下项目
问题类型
模型量化和部署
基础模型
Alpaca-Plus-7B
操作系统
None
详细描述问题
在运行这行代码时报错:
关于Colab的代码,我只修改了这里
依赖情况(代码类问题务必提供)
运行日志或截图