-
Suppose I want to employ a larger model for calculating embeddings such as the SFR-2 by SalesForce.
Is there a way to load the model into multiple GPUs?
Currently, it seems like only training suppor…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
- train.sh如下:
deepspeed --num_gpus=2 --master_port $MASTER_PORT test02.py \
--deepspe…
-
### Hello, does this analysis tool support Baichuan1 and Baichuan2?
Hello, I want to use this tool to analyze the Baichuan1 and Baichuan2 models. I don’t know if it is supported.
-
When will support huggingface GPT BigCode model
-
![1666232927523](https://user-images.githubusercontent.com/71402902/196841984-ac8ee470-6a82-4003-9828-6a077ad4735e.png)
Dear Yuan,
I met this issue when running the demo.py, it occurred in line…
-
Get this error
`RuntimeError: Expected tensor for argument #1 'indices' to have one of the following scalar types: Long, Int; but got torch.cuda.FloatTensor instead (while checking arguments for embe…
-
## Context
Hi everyone,
What I'm trying to achieve is to run privateGPT with some production-grade environment. To do so, I've tried to run something like :
- Create a Qdrant database in Qd…
LvffY updated
6 months ago
-
Is there a way to support pipelines with CPU offloading enabled?
It seems currently unable to handle this condition
```python
import gc
import torch
from diffusers import StableDiffusion3Pipe…
-
### **I think the front is running smoothly without bug.**
But until it runs to the fifteenth step. I have encountered the following problems.
**The order given by the author:**
`CUDA_VISIBLE_DEV…
-
Hi, I'm trying use lightseq to accelerate [MarianMTModel](https://huggingface.co/docs/transformers/model_doc/marian) seq2seq inference. This model is basically same with BartForConditionalGeneration.
…