-
The issue started with https://github.com/openxla/xla/pull/18052
Error log:
```
Per train step:
Total TFLOPs: 377.53
split as 86.02% learnable weight flops and 13.98% attention flops
jax.er…
-
We are currently using xgboost 1.6.2 and are trying to upgrade to 2.1.1. On the way through the versions, we observed the following prediction time averages:
1.6.2: 15ms
1.7.6: 17ms
2.0.3: 43ms
…
Raemi updated
8 hours ago
-
i don't know how to run
-
Hi, great work!
I have two questions about visualization of data distributions on different sources in fig.1.
Q1: Is the generated data visualized here learned from the corresponding data source? Fo…
-
### Describe the desired feature
Currently there are hard coded indices used to pick trainers so it only really works with the original 6.
Indices hard coded to 1 to 6 and special handling of inde…
-
When I run the train.py file to train the Davis dataset, I set input_dim_drug in the config file to 212 as prompted by the author. But then a runtime error occurs:
RuntimeError: CUDA error: device-si…
-
### Problem Description
Based on the decision tree used in the Test to select a kernel, I want to see how the features in the YAML file are trained.
### Operating System
Centos
### CPU
AMD
### G…
-
I am still very new to LLMs. I have access to a large amount of GPUs, and I would like to train this model across multiple GPUs ( though, I am not sure whether this is necessary/overkill ). Previously…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
- `Accelerate` version: 0.34.2
- Platform: Linux-6.5.0-35-generic-x86_64-with-glibc2.35
- `accele…
-
Hey @Littleor I've been trying your training script here but am not getting great results so far..
Are you able to successfully train things into a token?
I've been running with the following comm…