-
### Your current environment
```text
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Amazon Linux 2 (x86_64)
GCC vers…
-
With test `SMS_Ld1.T62_oEC60to30v3.CMPASO-NYF.pm-gpu_nvidiagpu` it has been failing for a while now. I think I mentioned this to @jonbob who said the fail dates matched a PR that recently went in. I t…
-
Install `diffusers` first.
And then do:
```python
from diffusers import DiffusionPipeline
from optimum.quanto import quantize, freeze, qint4
import torch
ckpt_id = "ptx0/pixart-900m-1024…
-
### 请提出你的问题
报错如下
Error Message Summary:
----------------------
ResourceExhaustedError:
Out of memory error on GPU 0. Cannot allocate 428.000000MB memory on GPU 0, 79.153320GB memory has been a…
-
Here is my description
```
[GIN] 2024/05/11 - 15:36:03 | 200 | 34.248274ms | 38.104.153.43 | GET "/api/v1/computing/cp"
time="2024-05-11 15:36:03.554" level=info msg="receive ubi task rece…
-
https://discord.com/channels/1107178041848909847/1252601327255683145
**Describe the bug**
The inference engine process associated with Jan does not terminate even after Jan is closed, resulting in…
-
It would be useful if it were possible to customize `nvidia-container-runtime.toml` without having to build new build assets.
We are using nvidia GPUs in our cluster, and we want to prevent users f…
-
Hi,
I'm trying to train APT using the dlc network, but the training window gets stuck on "Preprocessing" without throwing any errors.
![Screenshot](https://github.com/kristinbranson/APT/assets/…
-
Hello,
I am looking for a way to run some Machine Learning Inference within a Kubernetes cluster on Windows. Microk8s seemed to be a good solution as I saw that there was a gpu add on. I did some …
-
### NVIDIA Open GPU Kernel Modules Version
550.90.07 (latest)
### Please confirm this issue does not happen with the proprietary driver (of the same version). This issue tracker is only for bugs spe…