Closed LDLINGLINGLING closed 1 month ago
怎么安装的 lmdeploy ?pip install . 是不行的。
正常不会用pytorch backend来加载模型。
那应该如何安装
https://github.com/InternLM/lmdeploy/blob/main/docs/zh_cn/multi_modal/minicpmv.md
新建一个环境,然后 pip install lmdeploy
你可以按照这个说明来
(minicpmv) (base) root@gpu-62:~/ld# /root/ocnda/envs/minicpmv/bin/python /root/ld/ld_project/MiniCPM-V/lmdeploy_cli.py
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
[WARNING] gemm_config.in is not found; using default GEMM algo
卡在这里不动了,是不能使用本地pytorch路径进行加载?
创建的这一行改成这个把日志打开。 pipe = pipeline('/root/ld/ld_model_pretrain/MiniCPM-Llama3-V-2_5', log_level='INFO')
然后把完整的日志发一下。
https://github.com/InternLM/lmdeploy/issues/2004#issuecomment-2224402209
可以按照这个说明改一下这一行,run_in_executor
的时候,没有加异常处理,可能会隐藏真实的问题。
同遇到这个问题,日志如下
@Wimen
minicpmv 在 turbomind 引擎支持的。
turbomind 引擎需要编译,只靠 pip install . 是无法安装的。
可以在这里安装最新的版本 https://github.com/zhyncs/lmdeploy-build/releases
嗯,之前看到了前面的评论,所以我是新起的环境,直接pip install lmdeploy装了0.5.0,另外里面有报transformers的版本warnings,但还是会出现这个问题
请问有什么办法可以预先确认tubormind是可以正常工作的
请问有什么办法可以预先确认tubormind是可以正常工作的
from lmdeploy.turbomind import *
不报错的话,是成功的,报错的话,可以检查下原因:
python -c "import lmdeploy; print(lmdeploy)"
<module 'lmdeploy' from '/home/chenxin/ws3/vl/lmdeploy/__init__.py'>
ldd /home/chenxin/ws3/vl/lmdeploy/lib/*
linux-vdso.so.1 (0x00007ffe6c7ad000)
libnccl.so.2 => /usr/lib/x86_64-linux-gnu/libnccl.so.2 (0x00007ff8c557f000)
libcudart.so.11.0 => /usr/local/cuda/lib64/libcudart.so.11.0 (0x00007ff8c52d8000)
libcublas.so.11 => /usr/local/cuda/lib64/libcublas.so.11 (0x00007ff8bf67a000)
libcublasLt.so.11 => /usr/local/cuda/lib64/libcublasLt.so.11 (0x00007ff89b0f4000)
libdl.so.2 => /usr/lib/x86_64-linux-gnu/libdl.so.2 (0x00007ff89b0ec000)
libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007ff89af0a000)
libm.so.6 => /usr/lib/x86_64-linux-gnu/libm.so.6 (0x00007ff89adbb000)
libgcc_s.so.1 => /usr/lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007ff89ada0000)
libc.so.6 => /usr/lib/x86_64-linux-gnu/libc.so.6 (0x00007ff89abae000)
/lib64/ld-linux-x86-64.so.2 (0x00007ff8dee41000)
libpthread.so.0 => /usr/lib/x86_64-linux-gnu/libpthread.so.0 (0x00007ff89ab8b000)
librt.so.1 => /usr/lib/x86_64-linux-gnu/librt.so.1 (0x00007ff89ab7f000)
pip instal lmdeploy
默认会装cuda 12的 runtime,可以用 nvidia-smi
检查下驱动支持的cuda版本
日志如下: 2024-07-12 17:21:55,573 - lmdeploy - INFO - start ImageEncoder._forward_loop 2024-07-12 17:21:55,573 - lmdeploy - INFO - ImageEncoder received 1 images, left 1 images. 2024-07-12 17:21:55,573 - lmdeploy - INFO - ImageEncoder process 1 images, left 0 images. 在这里卡住了,而且问题狠奇怪,tp>1能跑,tp=1卡在以下bug [WARNING] gemm_config.in is not found; using default GEMM algo
@LDLINGLINGLING
可以把不能跑的图传上来看看
另外,可以按照这个 comment改一下代码看看 https://github.com/InternLM/lmdeploy/issues/2004#issuecomment-2224402209
This issue is marked as stale because it has been marked as invalid or awaiting response for 7 days without any further response. It will be closed in 5 days if the stale label is not removed or if there is no further response.
This issue is closed because it has been stale for 5 days. Please open a new issue if you have similar issues or you have any new updates now.
https://github.com/InternLM/lmdeploy/blob/main/docs/zh_cn/multi_modal/minicpmv.md
新建一个环境,然后
pip install lmdeploy
你可以按照这个说明来
支持openbmb/MiniCPM-V-2这个模型吗?我直接推理会报错:
AttributeError: 'MiniCPMVConfig' object has no attribute 'slice_mode'
支持的模型都在 https://lmdeploy.readthedocs.io/en/latest/supported_models/supported_models.html 目前只支持了 mincpm-llama3-v-2_5
Checklist
Describe the bug
官方的给的python示例代码推理Minicpmv-2.5报错
Reproduction
from lmdeploy import pipeline from lmdeploy.vl import load_image
pipe = pipeline('/root/ld/ld_model_pretrain/MiniCPM-Llama3-V-2_5')
image = load_image('https://wxls-cms.oss-cn-hangzhou.aliyuncs.com/online/2024-04-18/218da022-f4bf-456a-99af-5cb8e157f7b8.jpg') response = pipe(('describe this image', image)) print(response)
Environment
Error traceback