-
To reproduce:
```
th> a = torch.Tensor({1.0})
[0.0001s]
th> b = torch.Tensor({2.0, 2.0, 2.0})
…
-
File "train_stage_1.py", line 730, in
main(config)
File "train_stage_1.py", line 601, in main
Traceback (most recent call last):
File "train_stage_1.py", line 730, in
accelerator.ba…
-
Click to expand!
### Issue Type
Bug
### Have you reproduced the bug with TF nightly?
Yes
### Source
binary
### Tensorflow Version
2.14.0-dev20230512
### Custom Code
No
### OS Platform an…
-
Were you able to find out the reason for the small numerical errors in backward pass with ring flash attention?
I found the errors increase as you increase the world size, so it does seem to be rel…
-
when i use one GPU,there is a error:
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!
-
### What is the issue?
I am running aya model locally. When i just start the model with `ollama run aya` and interact in the terminal, it works fine. But when I try using it via POSTMAN on Windows 10…
-
When loading a model across 2 GPUs, the layers are split evenly, but the GPU memory usage is quite a bit higher on the first GPU:
```
|=========================================+===================…
-
Something wrong about the dimension in deepspeed's auto-tp.py:
![image](https://github.com/dvlab-research/LISA/assets/51390653/8f8adf4a-fc87-4c54-bd00-1c9cf10199a2)
-
### Your current environment
We are working on accelerating RLHF algorithms and need to broadcast the weights of the DeepSpeed engine to the vLLM Ray worker. In v0.4.2, we were able to create an ad…
-
hi 先進,
您好.
當有2個GPU在run時, 會出現底下的error
/usr/lib/python3.8/site-packages/torch/nn/parallel/_functions.py:61: UserWarning: Was asked to gather along dimension 0, but all input tensors were scalars; wil…