Closed huangyunxin closed 6 months ago
Based on the log analysis, it appears that the loss was not decreasing, indicating that the fine-tuning process did not yield the desired improvements. In standard machine learning practices, when faced with such an issue, it's essential to adjust the hyperparameters. This could involve extending the training duration or escalating the learning rate to achieve better model optimization.
In the specific context of your situation, the Qwen-Chat models have already undergone rigorous fine-tuning specifically to robustly reject requests involving personal information as a security measure. However, to change this behavior and adapt the model to handle such inquiries differently, you would require a more "intensive" or "precise" fine-tuning strategy, potentially exploring different hyperparameters and training techniques tailored to your new objective.
使用docker镜像
qwenllm/qwen:cu121
对Qwen-7B-Chat-Int4
进行微调,docker命令如下:微调example.json内容如下:
显卡是
RTX 4090 24G
,运行日志如下:我使用docker镜像qwenllm/qwen:cu117启动模型,我更改了openai_api.py文件,改为使用
AutoPeftModelForCausalLM
加载模型,如下:测试模型对话和微调前一样,如下:
请教是哪个环节有问题