Closed Abolfazl-kr closed 4 months ago
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your consideration.
Both data and hyperparameter will affect the final effect of the model. There may be cases of overfitting, try increasing total_batch_size
or observe the effect of the intermediate model to locate the problem, good luck.
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your consideration.
Check before submitting issues
Type of Issue
Model training and fine-tuning
Base Model
Chinese-LLaMA-2 (7B/13B)
Operating System
Linux
Describe your issue in detail
when i start pre training, the model seems to be broken. I feed the model the minimum data (less than 1 MB) after that the model cannot generate English sentence. I use your Chinese-LLaMA-Alpaca repo to create my tokenizer.
I would be so appreciated if you can help me.
Dependencies (must be provided for code-related issues)
Name: peft Version: 0.3.0 Name: transformers Version: 4.35.0
Execution logs or screenshots