-
Hi, there
When I try to run your codes on the Semantic-KITTI dataset, I met an error as follows
```
Reusing positional embeddings.
Traceback (most recent call last):
File "main.py", line 334,…
-
### System Info
- `transformers` version: 4.42.0.dev0
- Platform: Linux-5.4.0-172-generic-x86_64-with-glibc2.17
- Python version: 3.8.19
- Huggingface_hub version: 0.23.1
- Safetensors version:…
-
Hi,
I tried to integrate the TTLayer into transformerXL,
however I found that it consumes much more memory than usual.
Did you experience such problems? do you know anyway around this?
(BTW I a…
-
I am surveying different packages for hyperparameter optimization, and HpBandSter seems promising, especially becaues of its support for distributed training. But one thing I haven't had a clue is how…
-
I'm working on the C version of the code in preparation for (#40)
So llm.c with **no** code modifications I observe the following:
- `test_gpt2` works successfully and the loss matches
- `train_g…
-
Hi Devs,
I really liked the clean API of sparkflow for distributed training. Is it possible to run keras code using sparkflow?
-
I try to pretrain blip2 on a slurm cluster, but it seems that the current programme does not support distributed training on slurm by default. Any advice on it?
| distributed init (rank 0, world 1)…
-
LLaMa 2 is highly requested by customers. Can we ensure we have LLaMa 2 fine-tuning working with neuronx-distributed including sample code and tutorials for the 7B, 13B and 70B models ?
-
```bash
python3 -m torch.distributed.launch \
--nproc_per_node 8 \
--master_port 9527 \
train.py \
--workers 8 \
--device 0,1,2,3,4,5,6,7 \
--syn…
-
Hello, as you might know, I'm admiring your works (all of you guys, all the contributors) and love our community.
Apart from this start, here is my simple question:
Is there any plan to make it …