-
- Feature Name: Central Management of SVD to CSP process
- Start Date: 2018-05-xx
- RFC PR:
- Rust Issue:
# Summary
[summary]: #summary
I need to make this a paragraph, but my main points a…
-
1. transformers 版本 4.31.0 最新版本;
2. 已经修改的地方有
```
--- a/reward_modeling.py
+++ b/reward_modeling.py
@@ -34,6 +34,7 @@ from transformers import (
Trainer,
TrainingArguments,
set_s…
-
Could not estimate the number of tokens of the input, floating-point operations will not be computed
Traceback (most recent call last):
File "/root/nas-share/chat/MedicalGPT-main/reward_modeling.p…
-
### 🐛 Describe the bug
Hi,
I'm very new to TRLX, PEFT, and Huggingface, so I'm not sure if I just have some simple configuration wrong, but I am trying to recreate the notebook [here](https://cola…
-
### Describe the Question
按照run_training_pipeline.ipynb的步骤执行,
Stage1,Stage2都执行OK,执行到第三阶段:RM(Reward Model)奖励模型建模时,报错,请帮忙解决。
错误:**ValueError: weight is on the meta device, we need a `value` to put …
-
Direction changed, txt will be updated soon.
Old stuff:
- 1997: [The Internet: A Future Tragedy of the Commons?](https://link.springer.com/chapter/10.1007/978-1-4757-2644-2_22)
- [Internet Securi…
-
![image](https://github.com/hiyouga/LLaMA-Efficient-Tuning/assets/26586964/b36befe3-0e2b-4954-b14d-8af7dc221b16)
![image](https://github.com/hiyouga/LLaMA-Efficient-Tuning/assets/26586964/9bddca35-ae…
-
Hello!
I am trying to get the ``reward_modeling.py`` file to work on a smaller scale by using gpt2 as a reward model.
The only changes I made to the file from its current version in the repo w…
-
### Describe the bug
Please provide a clear and concise description of what the bug is. If applicable, add screenshots to help explain your problem, especially for visualization related problems.
-
This might be a dumb question, but I am having trouble following how the readme example matches how reward modeling is described in the latest papers I've read on it.
From the readme, the example …