-
我的微调命令就是基于本仓库提供的示例
https://github.com/FlagOpen/FlagEmbedding/tree/master/examples/unified_finetune
微调命令:
`export CUDA_VISIBLE_DEVICES=0,1
torchrun --nproc_per_node 2 \
-m FlagEmbedding.BGE_M3…
-
### Describe the issue
> [!TIP]
> ## Want to get involved?
> We'd love it if you did! Please get in contact with the people assigned to this issue, or leave a comment. See general contributing ad…
-
A lot of research in the field of RL is being done now days.
I thought it can be both interesting and productive to have a post that would bring new research from time to time that might be relevant …
-
Hola, estaba tratando de probar el código que hice para los agentes en la implementación de aprendizaje reforzado. Para ello ejecuté, primero el archivo test.py seleccionando a uno de los agentes con …
-
seems the code in chapter 4 just show some function of search
so i dont have to update my code written at chapter 1 2 3 ?
thanks for somebody's answer in the future, love you!
-
Hi,
I stumbled upon the following potential improvement, I am hacking it right now, but it would be great to have a proper solution. MCTS and other forward simulation techniques must have access to…
-
Thank you for your great work.
1. I want to use the data set to train an agent, but the data set you provided are too large. so what's the minimum data set needed to get a satisfying result?
2. th…
-
Hi, I’ve been using PPO to train an agent, and I noticed that the agent’s performance fluctuates even when it seems to have found an optimal policy. Specifically, after 200 episodes, the success rate …
-
### 🚀 The feature, motivation and pitch
Would be cool if Pytorch had smth like an agent, that we can spin up in the cloud, or even have a multi-user service, so instead of specifying cpu or gpu, we c…
-
We've received feedback from an [advanced user](https://github.com/erik-megarad) (working on [AutoPack](https://github.com/AutoPackAI/autopack) and [Beebot](https://github.com/AutoPackAI/beebot)) that…