InternLM / InternLM-XComposer

InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output
Apache License 2.0
2.48k stars 154 forks source link

Can't run with multiple gpu #364

Open volcverse opened 3 months ago

volcverse commented 3 months ago

Hello, thanks for the great work!

I refer to the example_code/example_chat.py to run the newest InternLM-XComposer-2.5 model using 4 NVIDIA 4090 GPUs. But still meet the OOM problem. It seems that although the weights are divided successfully, the first gpu always runs into OOM when model.chat is called.

Any response will be greatly appreciated!

Uoops commented 3 months ago

same

hyyuan123 commented 3 months ago

I refer to the example_code/example_chat.py to run the newest InternLM-XComposer-2.5 model using 4 A800 GPUs. But still meet the OOM problem.

waltonfuture commented 3 months ago

same question

yhcao6 commented 3 months ago

Please try to install transformers==4.33.1 with the following command and try again:

pip install transformers==4.33.1
YerongLi commented 3 months ago

Hello, thanks for the great work!

I refer to the example_code/example_chat.py to run the newest InternLM-XComposer-2.5 model using 4 NVIDIA 4090 GPUs. But still meet the OOM problem. It seems that although the weights are divided successfully, the first gpu always runs into OOM when model.chat is called.

Any response will be greatly appreciated!

I found the model cannot take multiple images as inputs, neither can it take a list of images thus the fix is

resi1ience commented 3 months ago

Please try to install transformers==4.33.1 with the following command and try again:

pip install transformers==4.33.1

Still meet the same problem with transformers 4.33.1. I'm running the video understanding example on the huggingface. Response will be greatly appreciated