-
I encountered a runtime error while using the transformers-interpret library with a fine-tuned LLama-2 model that includes LoRA adapters for sequence classification. The error occurs when invoking the…
-
**Describe the bug**
When using a SmoothQuantModifier and cpu offloading there is a conflict of tensors not being on the right device.
**Expected behavior**
cpu offloading should work w/ SmoothQu…
-
You can modify the location where the Hugging Face model should be stored (when using the `transformers` library) by setting the environment variable `TRANSFORMERS_CACHE`. The default location is the …
-
### System Info
- 2x H100 80GB on docker container (nvidia/cuda:12.4.1-devel-ubuntu22.04)
- last version of the library
### Who can help?
_No response_
### Information
- [X] The official example…
-
### Summary
There are various LLM inference libraries. WasmEdge already integrated llama.cpp, but we want to bring more to the community.
### Details
Already supported:
1. PyTorch
2. TFLi…
-
I tried to run Lab 6 on my laptop following the instructions but I received this message:
_WARNING: You are currently loading Falcon using legacy code contained in the model repository. Falcon has …
-
CommonJS compatibility
*In general, the feature you want added should be supported by HuggingFace's [transformers](https://github.com/huggingface/transformers) library:*
- *If requesting a **model*…
Emsu updated
2 months ago
-
I get this error while running the inference.ipynb code in the docs folder. I did not change anything in the file. Does anyone know how to solve it ?
Thank you for your time.
RuntimeError …
-
Salesforce is releasing there time series library for transformers training and forecasting. Thus, it might be worth to interface it.
This is also supporting a download of their time series foundat…
-
Now that we have a working PoC (#9165) of NF4 quantization through `bitsandbytes` and also [this](https://huggingface.co/blog/quanto-diffusers) through `optimum.quanto`, it's time to bring in quantiza…