-
I am using this command to try inference:
```
python generate.py --load pytorch_model.bin --tokenizer-type HFTokenizer --vocab-file tokenizer.json --text_gen_type interactive --temperature 0.0 --max…
-
I wanted to check out how does DirectML work with transformers but when i tried following steps in attention_is_all_you_need the error occured.
> Traceback (most recent call last):
> File "E:\P…
-
Hello !
for the gradio demo: 384p works fine but when i launch for 768p: it's fail:
```
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.4.0+…
-
I have installed trl
-
按照 readme的方式准备的环境,进行inference 报错:
Traceback (most recent call last):
File "/mnt/localdisk/tanm/miniconda/envs/table_llava/lib/python3.10/site-packages/transformers/feature_extraction_utils.py", li…
-
### Area of Improvement
Why is the following section in the docs correct:
> Here [superjson](https://github.com/blitz-js/superjson) is used for uploading and [devalue](https://github.com/Rich-Harr…
-
### Issue
tl;dr - I think there is some possible doubling in the reporting of sent tokens on the summary/cost estimate line and I think aider isn't accounting for the cached input rate when calculati…
-
I'm trying to finetune **Mistral-Nemo-Base-2407** with a `text` dataset of long inputs. Usually, the SFTrainer will truncate it to fit the specified context size.
However, I get an error when using…
-
完全按照readme文档进行的安装, 运行出现报错
具体报错信息
```
# ComfyUI Error Report
## Error Details
- **Node Type:** Joy_caption_two
- **Exception Type:** TypeError
- **Exception Message:** LoraConfig.__init__() go…
-
The current `RetrieverCache` implementation calls the `transform()` function of the wrapped `Transformer` once for each query in the input data frame, which can be costly for some retrieval models due…