-
running dpo with Qwen meet flatten problem. FSDP config as follow
```yaml
compute_environment: LOCAL_MACHINE
debug: false
distributed_type: FSDP
downcast_bf16: 'no'
fsdp_config:
fsdp_auto_w…
-
Would love to see Ollama run on a TPU not just GPU. Has this been done by anyone already?
-
Tile based operations have been quite a success for creating optimal GPU kernels. The programming model, in my understanding, offers flexibility while taking advantage of cache hierarchies.
http:/…
-
## Detailed Description
Currently, the graph neural network library dependencies don't support TPUs with pytorch geometric, or don't seem to at least because of custom kernels. We could add a Jax…
-
### Steps to reproduce
```sh
dstack run . -f train.dstack.yml -b gcp --gpu tpu-v2-8
```
### Actual behaviour
The process ran normally, including output info from installing packages, downlo…
-
## 🐛 Bug
Whenever I use `generate` function on a TPU (I use v5e litepod8), I have a crash with a C++ stack trace but no info on the python side and no way to catch/recover.
## To Reproduce
I…
-
## 🐛 Bug
I am fine tuning a transformers model - openai whisper - and using pytorch to do the training. I am training using a google cloud TPU v4, and it freezes with zero errors at iteration 34. T…
-
Because of the capacity limit, we cannot do TPU testing on every PR push. We have two options here:
1) type gcbrun/ manually to trigger TPU testing, which should happen after normal working hours.
2…
-
Any help would be appreciated.
I am running on two PC with Windows 11 and Windows 10, each with Nvidia GPU, and facing same issue on both.
All installation was done correctly but keep getting "No…
-
I'm trying to deploy Llama3 8b on GKE using optimum but running into some troubles.
Following instructions here: https://github.com/huggingface/optimum-tpu/tree/main/text-generation-inference. I bu…