bigscience-workshop / petals

🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
https://petals.dev
MIT License
8.89k stars 490 forks source link

Error trying to raise Mixtral private swarm server #569

Closed Qessia closed 2 months ago

Qessia commented 2 months ago

Reproduce:

python3 -m petals.cli.run_server mistralai/Mixtral-8x7B-v0.1 --new_swarm or python3 -m petals.cli.run_server SanjiWatsuki/TinyMixtral-32x248M --new_swarm

Got:

TypeError: WrappedMixtralBlock.__init__() missing 1 required positional argument: layer_idx

System:

artek0chumak commented 2 months ago

Hello! Thank you for reporting! We will quickly resolve this issue.

mprishchepo commented 2 months ago

Hello!

I observe the same problem. I have tried to diagnose the issue a bit by myselve.

As I understood (if you haven't found it already) the problem is in calculating block size (its parameters). The layer_idx mentioned above is used in load_pretrained_block, but it is not used when calculating block_size and when calculating rps in throughput.

Very much waiting for a solution.

artek0chumak commented 2 months ago

We resolved this issue in recent master update. Just pull new updates. Thank tou for noticing the issue and waiting fixes.

Qessia commented 2 months ago

Thank you for your quick response!

mprishchepo commented 2 months ago

Hi! Original error of this issue doesn't appear anymore, but I've got another error when I try launching private swarm with Mixtral (with GPU, CPU is ok). Also it doesn't appear when I do the same with StableBeluga2

System:

artek0chumak commented 2 months ago

Hello! This is a strange error. Can you also provide a transformers' version?

mprishchepo commented 2 months ago

Can you also provide a transformers' version?

4.38.2

artek0chumak commented 2 months ago

Thank you for the information. It seems the only change required is this: https://github.com/bigscience-workshop/petals/pull/574. We will soon merge it with the main.

mprishchepo commented 2 months ago

Hi! How is the work on the fixes going, is everything good? We are really looking for the merge

jmikedupont2 commented 2 months ago

I had that same error on master as well and had a ticket open for it, https://github.com/bigscience-workshop/petals/issues/575

artek0chumak commented 2 months ago

Sorry for taking so long; the fix is merged into the master.

jmikedupont2 commented 2 months ago

I was able to get the branch mentioned running and my docker work rebased. Screenshot_20240416_140359_Termux.jpg

Have now tinymixtral running locally in gpu. https://github.com/meta-introspector/petals

Qessia commented 2 months ago

Thank you for fixes!! It works