Open ZzzybEric opened 4 months ago
i use vllm to inference deepspeed, use flask to deploy model. When the problem enters the model, it always gets stuck for a long time in the processd prompt step,the code i use is your example code
https://huggingface.co/deepseek-ai/DeepSeek-V2/discussions/1 @ZzzybEric
whats your gpu type?
i use vllm to inference deepspeed, use flask to deploy model. When the problem enters the model, it always gets stuck for a long time in the processd prompt step,the code i use is your example code