-
hi, great work! how to implement the batch inference in offline environment.
-
Hi guys,
Amazing work you guys have done here!
However, when I'm trying to reproduce MMMU benchmarking using your model, I got low overall ACC (0.35) and my evaluation calls the .chat() function…
-
### Description / 描述
```
import torch
from PIL import Image
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained('MiniCPM-V-2.0', trust_remote_code=True)
tok…
-
## 🚀 Feature
Currently, mlc-llm lacks support for VL models. Consider supporting the following VL models:
https://huggingface.co/Qwen/Qwen-VL-Chat
https://huggingface.co/openbmb/MiniCPM-V
https://…
-
Hi 2 quick questions,
1. From the paper algorithm1, I get a sense that the algorithm can work in an online divide-n-conquer manner with updated model and I am just curious when the self-feedback co…
-
finetune/readme.md下面写的加载lora方法的方法只有一部分,请问具体要怎么加载测试呢?
-
### Description / 描述
當於LM Studio載入"minicpm-2b-dpo-fp32.Q6_K.gguf" 時,報錯:"create_tensor: tensor 'output.weight' not found",不知道應該設定甚麼Preset
### Case Explaination / 案例解释
_No response_
-
-
(infer) tskj@tskj:~/project$ python -m vllm.entrypoints.openai.api_server --model /home/tskj/MOD/MiniCPM-V-2 --trust-remote-code --host 0.0.0.0 --port 9998 --gpu-memory-utilization 0.45
INFO 05-09 07…
-