-
Hi Team,
I have attempted Knowledge Distillation using Torchtune for the 8B and 1B Instruct models. However, I still need to apply KD to the Vision Instruct model. I followed the same steps and cre…
-
### What feature would you like to be added?
How Magentic-One be used with local LLMs or Ollama?
### Why is this needed?
This will enable users to use Magentic-One with open-source LLMs other than …
-
### 🚀 The feature, motivation and pitch
Llama3.2 vision (Mllama) models requires model runner as "Enocoder_Decoder_Model_Runner"
which includes:
1. prepare "encoder_seq_lens" and "encoder_seq_len…
-
## Describe the bug
Hey everyone,
I’m trying to run vision models in Rust on my M4 Pro (48GB RAM). After some research, I found [Mistral.rs](https://github.com/EricLBuehler/mistral.rs/tree/master)…
-
### What is the issue?
My initial goal is to check if specific model is available using Ollama API.
I use OpenAI library `github.com/sashabaranov/go-openai` to do that.
The problem is when I …
-
### What is the issue?
Running llama3.2-vision 11b model currently seem to throw request timed out errors.
BoltAI 1.26.1
ollama 0.4.1
In ollama console output I can see:
```
[GIN] 2024/11/…
-
Hi,
First of all, thanks for your work.
I tried out the llama3.2-vision:90b model on ollama and it seems to underperform the version available on the build.nvidia.com API, with the same prompt …
-
![image](https://github.com/user-attachments/assets/c7f60383-beab-46fe-91d7-e35a6ca77bf4)
`public static void main(String[] args) {
init();
String question = "最近3天北京的天气怎么样?火车是否允许发车?…
-
請問在訓練MiniCPM-Llama3-V 2.5時,透過以下設定只微調視覺模型,而不全量或LoRA微調LLM所訓練後的權重要如何使用呢?
--tune_vision true
--tune_llm false
--use_lora false
嘗試直接載入模型會出現下方錯誤
`AttributeError: 'MiniCPMVTokenizerFast' object has …
-
the architecture error when creating models or converting to gguf with llama.cpp
it look like the model has the same architecture as the llama3.2V, could you help me with this?