-
~/MusePose# accelerate launch train_stage_2.py --config configs/train/stage2.yaml
The following values were not passed to `accelerate launch` and had defaults used instead:
`--num_processes`…
-
Hi,
Thanks for sharing the work. Would you mind also releasing the `EgoNaoDataset` class? The` data_preproceessing/datasets/` folder is missing. And it would be helpful if you can also release the …
-
I tried this model on my own video, hoping it can somehow improve the detection performance on corner cases like Y branching, but it fail to work as what I expected when using pretrained curvelanes m…
-
**Describe the bug**
Query_input's shape is [batch, pos, n_heads, d_model], and the purpose of the code where the error occurred is to reshape query_input to [batch, pos, n_heads, d_head].
I found t…
-
Dear @sanchit-gandhi,
I was following your tutorial, [Fine-Tune Whisper For Multilingual ASR with 🤗 Transformers](https://huggingface.co/blog/fine-tune-whisper), to fine-tune Whisper with a dataset i…
-
what's matter please
Traceback (most recent call last):
File "train.py", line 90, in
student=config.create_student(nocrf=args.nocrf)#调用 config 的 create_student 方法来创建一个学生模型,指定是否启用 CRF
F…
-
pretrained word embedding
-
Hi,
Thanks for publishing and sharing the TSDAE approach.
I am reading through the paper. I have one question.
In Section 7.4 of the [paper](https://arxiv.org/pdf/2104.06979.pdf), it recommend…
-
### System Info
```
transformers==4.45.1
peft==0.13.1
accelerate==0.34.2
bitsandbytes==0.44.0
datasets==2.20.0
evaluate==0.4.1
safetensors>=0.4.3
sagemaker==2.232.2
sentencepiece==0.2.0
sci…
-
You mention that bidirectional attention is used for embedding task. But it appears that you only use the last hidden states from the pretrained LLM to generate embeddings. Is the final projection is …