-
For how many epochs & steps per epoch/ how many images was the model trained on COCO dataset in order to obtain the mask_rcnn_coco.h5 weights?
-
I am trying to use the framework to continue pretraining llama3-8B. I have converted the HF checkpoint into nanotron format and the generated tokens seem reasonable.
I use the following setting to…
-
I am trying to run the pretraining scripts and encountering the following error while loading the datasets from disk.
GPU available: True (cuda), used: True
TPU available: False, using: 0 TPU core…
-
I use one node with four GPU(V100, 32G) for pretrain, but parallel training is a little weird. All **four** processes run on **one(device:0)** GPU.
Why it happened? Thanks for everyone's help!
I u…
-
Hi,
I am implementing fine-tuning exBERT for sequence classification. I already have done the pretraining for my data. However, since the pre-training python script that you have provided is only f…
-
It is mentioned in the repo that the pretraining step should run for some time, please mention after how much time i should interrupt it.
Also i can't use the pretrained npz file as i'm planning to…
-
Hello,
Thank you for your very interesting model.
I intend to use SparseBEV on a new dataset with only 1 frame , as a baseline (no previous temporal frame).
In the paper, on ablation study, y…
-
### System Info
```shell
vault.habana.ai/gaudi-docker/1.17.0/ubuntu22.04/habanalabs/pytorch-installer-2.3.1:latest
```
### Information
- [X] The official example scripts
- [ ] My own modified scri…
-
-
Are there any appropriate setups or losses in sentence-transformers for pretraining sentence embeddings in cases where I have labels as targets?
(I want to finetune the actual embeddings, not just a…