-
Hello,
ViP sounds very interesting—congratulations on the paper! I wonder if you would consider adding it to the Hugging Face Hub. Doing so would increase the model's visibility and make it easier …
-
I attempted to run benchmarks for the llama-3-8b-instruct and llama-3.1-8b-instruct models using both CPU and GPU, but the process failed. (I successfully tested the llama2-7b-chatbot model)
I f…
-
### System Info
- `transformers` version: 4.44.2
- Platform: Windows-10-10.0.19041-SP0
- Python version: 3.8.0
- Huggingface_hub version: 0.24.6
- Safetensors version: 0.4.4
- Accelerate version…
-
**Description**
HuggingFace's Quanto has implemented 4 bit & 2 bit KV cache quantization compatible with Transformers. See: https://huggingface.co/blog/kv-cache-quantization
I may PR when I've t…
-
### Anything you want to discuss about vllm.
Got error
---
ValueError: The checkpoint you are trying to load has model type `cohere` but Transformers does not recognize this architecture. This coul…
-
Hello!
First of all, thanks for working on this library, super handy. I was wondering if you were thinking about implementing the repetition bracket symbols (e.g _[a-z]{5,10}_)?
It is supported …
-
this is my test code:
version is 0.6.0
```
fn sam() {
let result: Result = (|| {
let directory = "/home/foliage/model/candle-sam".to_string();
let device …
-
Hi,
Congrats on this work! I discovered it from the paper page: https://huggingface.co/papers/2408.15881 (feel free to claim the paper in case you're one of the authors, so that it appears at your …
-
Cannot import this from the library,
optimum-1.21.4
transformers-4.43.4
how to fix this issue the code i am running is from the quick start documentation of `Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4…
-
We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like microsoft/Florence-2-large is not the path to a directory containing a file na…