-
Hello,
One of my systems, running Ubuntu 22.04, updated docker to version 27.3.1 (build ce12230)... after which running AlphaFold2 produces a neverending series of messages like the following once …
-
### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.17
### Custom code
Yes
### OS platform and distribution
Linux U…
x0w3n updated
22 hours ago
-
Recently, we have started the process to reduce the torch_xla API footprint in favor of torch API to improve the usability. This RFC focuses on the process to deprecate any functions.
## Backward c…
-
Problem: I have some Jax code that does sequence parallel, so somewhat similar to this
```
activation = jax.lax.with_sharding_constraint(activation, NamedSharding(mesh, PartitionSpec('data', 'tenso…
-
I am running SpeedPPI on GPU nodes. But some of the jobs would run out of memory even with 250GB memory. The error says `RESOURCE_EXHAUSTED: Out of memory while trying to allocate 16508718128 bytes.` …
-
Following the previous issue (https://github.com/chunmeifeng/T2Net/issues/1) updates. I am faced with the above mentioned runtime error. This happens when I run the code in the GPU runtime. The stack-…
-
### 🚀 The feature, motivation and pitch
For users that wish to run Torch on XLA devices in Python 3.11, having pre-built Torch XLA wheels will provide a smooth, reliable installation experience.
###…
-
## ❓ Questions and Help
I want to run pytorch xla on kaggle tpu v3-8 and use all core in tpu. But I always get A process in the process pool was terminated abruptly while the future was running or pe…
-
### What happened?
[StableHLO spec](https://openxla.org/stablehlo/spec#add) says that `stablehlo.add` op on booleans should behave as "logical OR".
but that's not the logical behavior nor the …
-
Thank you for the repo.
I am wondering if a recipe for TPU pods can be added. I have access to v4-32 and want to train a LLaMA model from scratch. Wondering if the repo can be extended for this us…