-
Hello everyone, thank you for the great job!
I am trying to further fine-tune the LLaVA architecture using your implementation with LLaMA 3 Instruct 8B. I can already fine-tune the Vicuna model usi…
-
**Describe the bug**
The new Llama 2 70B features GQA. This causes an issue with inject_fused_attention.
When a user attempts to do inference on a Llama 2 70B model with inject_fused_attention=Tr…
-
### What happened?
When I use android NDK to compile the bin file on Linux, and then transfer the bin file to Android termux and execute it, an error will be reported:CANNOT LINK EXECUTABLE "./llama-…
-
I followed the readme, but i can't get llama-cpp to run on my 4090.
```
set CMAKE_ARGS=-DLLAMA_CUBLAS=on
set FORCE_CMAKE=1
pip install llama-cpp-python --upgrade --force-reinstall --no-cache-di…
-
### Describe the bug
When I toggle the option for multimodal the software crashes.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
Turning on …
-
Getting this message:
File "/anaconda3/lib/python3.11/site-packages/transformers/processing_utils.py", line 926, in apply_chat_template
raise ValueError(
ValueError: No chat template is set f…
-
Hello, and thanks for such a great contribution to the field of interleaved LMMs! This is really great work. I was wondering if there was an example of the format for multiple image or multiple video …
-
- cmd:
`xtuner chat LLM-Research/Meta-Llama-3-8B-Instruct \
--visual-encoder ./clip-vit-large-patch14-336 \
--llava ./LLM-Research/llava-llama-3-8b \
--prompt-template llama3_chat \
--ima…
-
![微信截图_20240713161048](https://github.com/user-attachments/assets/179a13fc-1dce-45d5-b803-69151cab8e56)
转换后的文件和说明中的不同
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [X] I am running the latest code. Development is very rapid so there are no tagged versions as of…