-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
hello, i want to profile the llama index system . my code snippet is below. My gpu is on…
-
I try to fine-tune Llama 2 and when I launch the training with :
```
trainer = SFTTrainer(
model=model,
train_dataset=dataset,
peft_config=peft_config,
dataset_text_field="text…
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/PKU-Alignment/omnisafe/issues) and [Discussions](https://github.com/PKU-A…
-
https://huggingface.co/blog/rlhf
### Background
In the section on the third step of the process, it is written:
- What multiple organizations seem to have gotten to work is **fine-tuning some…
-
Direction changed, txt will be updated soon.
Old stuff:
- 1997: [The Internet: A Future Tragedy of the Commons?](https://link.springer.com/chapter/10.1007/978-1-4757-2644-2_22)
- [Internet Securi…
-
Hello
Does llama2 provide a list of sources used for training the model.if so, where is that made available..
Is the complete code and training sources available in this github repo?
Thanks
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/PKU-Alignment/safe-rlhf/issues) and [Discussions](https://github.com/PKU-…
-
PPOTrainer throws the following error when passed argument --gradient_accumulation_steps >=2.
```
$ python trl/examples/scripts/sentiment_tuning.py --gradient_accumulation_steps 2
[2023-08-15 20:…
-
I'm curious to read an explanation of fundamental technical differences or similarities between Perl 6 macros (and, separately if appropriate, 007 macros) with Lisp FEXPRs. I thought ven especially mi…
raiph updated
3 months ago
-
## 💥 Proposal
The goal of this project is to develop an autonomous robot navigation system using reinforcement learning. The robot will learn to navigate and explore its environment efficiently wit…