ymcui / Chinese-LLaMA-Alpaca

中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
https://github.com/ymcui/Chinese-LLaMA-Alpaca/wiki
Apache License 2.0
18.23k stars 1.86k forks source link

33b模型的pipline parallelism #832

Closed raindrop313 closed 1 year ago

raindrop313 commented 1 year ago

提交前必须检查以下项目

问题类型

模型推理

基础模型

Alpaca-33B

操作系统

Linux

详细描述问题

作者您好,目前我有八张V100,单卡的话无法跑33b的模型,直接用LlamaForCausalLM.from_pretrained()的方法会爆显存,有什么办法在导入的时候分布式的放在不同的gpu上呢?

github-actions[bot] commented 1 year ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your consideration.

github-actions[bot] commented 1 year ago

Closing the issue, since no updates observed. Feel free to re-open if you need any further assistance.