-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
采用数据并行lora微调,报错如下
CUDA SETUP: CUDA runtime path found: /usr/local/cuda/lib64/libcudart.so.11…
-
/home/test/anaconda3/envs/kolors/lib/python3.8/site-packages/diffusers/models/transformers/transformer_2d.py:34: FutureWarning: `Transformer2DModelOutput` is deprecated and will be removed in version …
-
Traceback (most recent call last):
File "/content/kohya-trainer/train_network_xl_wrapper.py", line 10, in
trainer.train(args)
File "/content/kohya-trainer/train_network.py", line 251, in t…
-
I ran the demo_lazy.py as shown below (on google colab) -
```
!cd /content/APE && python3 demo/demo_lazy.py \
--config-file configs/LVISCOCOCOCOSTUFF_O365_OID_VGR_SA1B_REFCOCO_GQA_PhraseCut_Flickr…
-
**Hey, I've noticed that whenever I try to use GPTQ quantized models, it doesn't work. I've noticed it sees it as AWQ and not GPTQ, I don't know why that is because it worked fine yesterday. I tried t…
-
I'm not able to run inference for my fine tuned model - tried running the command in the documentation.
I finetuned my model using:
```
swift sft \
--model_type florence-2-large \
--s…
-
Error transcribing chunk 25 in video.mp4
The length of `decoder_input_ids`, including special start tokens, prompt tokens, and previous tokens, is 2, and `max_new_tokens` is 512. Thus, the combined …
-
With the new release of version 3.2.0, the use of ONNX has become much easier but initial local tests led to various errors, meaning that it was not possible to use ONNX Runtime via Sentence Transform…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
ChatGLMTokenizer(name_or_path='THUDM/chatglm-6b', vocab_size=64794, model_max_length=1000…
-
### System Info
- `transformers` version: 4.44.0
- Platform: Linux-5.4.0-162-generic-x86_64-with-glibc2.31
- Python version: 3.11.9
- Huggingface_hub version: 0.23.4
- Safetensors version: 0.4.…