-
@OH-ThatGuy @gylertaydos Part of your milestone for Monday:
Let's look up some resources:
* passages of dragon (with or without English translation)
* resources on language models that train for la…
-
I try to start a large version of the model using docker:
`docker run -p 10249:80 -e RUST_BACKTRACE=full -e FLASH_ATTENTION=1 -e CUDA_VISIBLE_DEVICES=4,7 --privileged --security-opt="seccomp=unconf…
-
Hello! I'm currently running bloomz-petals on a Google Colab notebook in order to make use of the free GPU. However, I've recently started receiving the following error:
MissingBlocksError: No serv…
-
Hi, when I ran ppo with bloomz-7b1-mt and bloom-560m (prompt_len = answer_len = 256) with zero stage 3 (8*A100-40G), it seems the generation time is too slow (average about 72s). When I setting zero s…
-
https://arxiv.org/pdf/2212.09535.pdf
I was reading this paper, and really interested into trying this myself. But I can't find the model weights (bloom-3b) anywhere. Can you link that? would be great…
-
I have completely fine-tuning my model by LOMO, more details i'm using bloomz-7b1-mt as the backbone and finetuning it on Alpaca Instruction Dataset. I'm using my own data processing pipeline and just…
-
I collect some chinese data about "中国云南" like this:
![0417-2](https://user-images.githubusercontent.com/52442277/232364095-2bf77e7b-f850-46ba-ae5f-5d9777404b1c.png)
And train follow the readme base …
-
-
Loading these models and being able to use them for inference or in a repl is very cool.
To make this broadly useful, we need a way to verify that models loaded by llm are acting the same way as they…
-
is it possible to use petals for inferring/prompt tuning without sharing my gpu?