-
https://huggingface.co/microsoft/Phi-3-medium-128k-instruct
https://huggingface.co/microsoft/Phi-3-medium-4k-instruct
https://huggingface.co/microsoft/Phi-3-small-8k-instruct
https://huggingf…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video
The claim is it performs very well for an 8 billion size model
I am interested in learning what it takes to add suppor…
-
**The bug**
The 'quick spot check to verify we can rebuild complex multi-token unicode symbols' check in the `TransformersTokenizer` constructor is failing for Phi-3-vision-128k-instruct.
This i…
-
The Phi3 vision model is excellent and does a great job in extracting text. I am using the CPU version via C# DirectML package.
1. What is the max image filesize in kb that can be sent to the mode…
-
Looks like a good project and I would like to try this, would it be possible to add HF / Gemini options? HF has more diversity and Gemini is smarter, neither has "customer noncompete clause" like this…
-
We're seeking talented individual to develop a comprehensive tutorial for a new Llama-3.
The Plan and Structure of the Tutorial (recommended)
Here's an outline of what we're looking for in the tut…
-
### Question
I want to try changing `liuhaotian/llava-v1.5-13b` to use a different image tower instead of `clip-vit-large-patch14`.
1. After changing the vision tower, is it necessary to pretrain …
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue]…
-
Hello, I load pre-trained llava-llama3 SFT weights and fine-tune using LoRA, but get an error when merging weights:
**scripts:**
Training:
```
deepspeed --master_port=$((RANDOM + 10000)) --inclu…