Closed Ther-nullptr closed 2 months ago
This version seems not support quantized backbones:
main: work_size = 768376 bytes (0.7 MB)
train_opt_callback: iter= 0 sample=1/27583 sched=0.000000 loss=0.000000>
ggml_cuda_op_bin_bcast: unsupported types: dst: f32, src0: q3_K, src1: f32
GGML_ASSERT: /home/ther-nullptr/Desktop/llama.cpp/ggml-cuda.cu:8011: false
Could not attach to process. If your uid matches the uid of the target
process, check the setting of /proc/sys/kernel/yama/ptrace_scope, or try
again as the root user. For more details, see /etc/sysctl.d/10-ptrace.conf
ptrace: Operation not permitted.
No stack.
The program is not being run.
Aborted (core dumped)
Yeah, you would have to use f32 or f16 until the Segmentation fault
is fixed. Also, GPU offloading isn't that much faster since the computation is still done on the CPU.
So compute acceleration of GPU is not implemented on finetuning algorithm yet?
This issue was closed because it has been inactive for 14 days since being marked as stale.
What happened?
I try to finetune a llama-like model use
./llama-finetune
.Name and Version
./llama-cli --version version: 3196 (7d5e8777) built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for aarch64-linux-gnu
What operating system are you seeing the problem on?
Linux
Relevant log output
gdb debug trace: