-
E:\ComfyUI\models\LLM\Meta-Llama-3.1-8B-bnb-4bit
Unused kwargs: ['_load_in_4bit', '_load_in_8bit', 'quant_method']. These kwargs are not used in .
transformers version 4.44 or 4.43, this warning m…
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
Hello,
How can I do multi-doc RAG using Weaviate as vector store?
-
When I run
`bash scripts/video/demo/video_demo.sh ${the path of LLaVA-NeXT-Video-7B-DPO} vicuna_v1 32 2 True ${the path of video}`
I get the error
```
Can't set vocab_size with value 32000 for …
-
The `llama_kv_cache_seq_shift` or `llama_kv_cache_seq_rm` (or all two of them) is broken with cache type q4_0 for K.
In the `main.cpp`, these functions are used for "context swapping", meaning we c…
-
-
**描述一下Bug**
使用kaggle平台部屬方式,也就是本wiki中,第三種部屬方式,部屬完之後可以進入llama的畫面
使用RPGMaker_LLM_Translator進行翻譯,大約進行250~350次左右的翻譯
整個kaggle虛擬機就會無法再進行翻譯,不接受任何API呼叫
**复现步骤**
1.先啟動kaggle平台的部屬方式
llama.cpp版本: b2859
更…
-
I've noticed most of the PRs recently are generated by LLM, and this project has had some issues.
1. No type check, no information. It takes work to maintain and understand what's going on.
For …
-
I tried two gguf conversion on M2 ultra (metal) but no luck. I converted them myself and still the same error.
Here is the first model I tried:
https://huggingface.co/guinmoon/MobileVLM-1.7B-GGUF…
-
Hello!
Commit `2badd76` appears to break `examples.models.llama2.export_llama`, specifically with Llama 3.
### Expected Behavior
```
[INFO 2024-06-14 16:04:23,366 export_llama_lib.py:390] Ap…
amqdn updated
2 months ago
-
## Goal
Run a llama model from https://github.com/nod-ai/sharktank/blob/main/sharktank/sharktank/models/llama/llama.py through IREE
Starting with `open_llama_3b_v2_f16_gguf` since we have that i…