-
Please let us know what model architectures you would like to be added!
**Up to date todo list below. Please feel free to contribute any model, a PR without device mapping, ISQ, etc. will still be …
-
您好,对比了多家的vlm, minicpm-v做的还不错。
但是
1高精度图效果不行,你们为什么没采用5B的图像编码器呢?
2为什么没支持多图呢?
能从技术、效果解答我的疑问吗
-
### Checklist
- [x] 1. I have searched related issues but cannot get the expected help.
- [x] 2. The bug has not been fixed in the latest version.
### Describe the bug
minicpm-v采用W4A16量化后,显存占用确实小了,…
-
### Your current environment
```text
安装完之后,直接运行python examples/minicpmv_example.py出现的问题
INFO 06-27 10:16:32 utils.py:598] Found nccl from environment variable VLLM_NCCL_SO_PATH=/usr/local/lib/pytho…
-
请问跑minicpm-llama3-v-2_5(int4)支持并发调用接口么?2个及以上并发调用就报错了,单个没有问题。。。。
![微信截图_20240619180246](https://github.com/xorbitsai/inference/assets/167763677/caba7bf3-199d-4a24-88a3-b0e9833b50b2)
![微信截图_2024061918…
-
Hello,
Thanks for the project it look really nice ! i'm new in this world and i'm struggling to do what i want.
I have a macbook m1 pro 16gb.
I managed to install and make it run but some thin…
-
### Describe the bug
使用最新版本 xinference 部署 bge-reranker-v2-minicpm-layerwise,modescope 无法下载,更换 huggingface 后部署成功,但在使用的时候耗时特别严重,基本无法应用。
```
You're using a LlamaTokenizerFast tokenizer. Please note …
-
**Describe the feature**
Please describe the feature requested here(请在这里描述需求)
目前infer的方式和deploy部署,都是串行的。是否有某种方式可以实现并行对多条数据进行同时处理(为了加快速度)。
例如,我需要使用一个自己训练的模型(minicpm-v2.5-chat),对大约一万条文字提问+每个提问对应的一张…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
- [x] MiniCPM-Llama3-V-2_5
- [x] Florence 2
- [x] Phi-3-vision
- [x] Bunny
- [x] Dolphi-vision-72b
- [x] Llava Next
- [ ] Llava Interleave
- [ ] internlm-xcomposer2d5-7b
- [ ] InternVL
- [ ] Co…