Open goodskillprogramer opened 4 months ago
if I only input text ,the model works ok. But when I added a image ,it will hang and there will be no output. below is the consol;
llama_new_context_with_model: graph nodes = 1030 llama_new_context_with_model: graph splits = 2 INFO [main] model loaded | tid="140026374909952" timestamp=1718185245 time=2024-06-12T09:40:45.713Z level=INFO source=server.go:567 msg="llama runner started in 2.26 seconds" [GIN] 2024/06/12 - 09:40:46 | 200 | 3.363909682s | 127.0.0.1 | POST "/api/chat" [GIN] 2024/06/12 - 09:40:46 | 200 | 204.205184ms | 127.0.0.1 | POST "/v1/chat/completions"
Linux
Nvidia
No response
the same
Same here, both for CPU and GPU.
OS Linux
GPU Nvidia
Same here, using on Docker. The endpoint generate hangs forever.
same here, no help
What is the issue?
if I only input text ,the model works ok. But when I added a image ,it will hang and there will be no output. below is the consol;
llama_new_context_with_model: graph nodes = 1030 llama_new_context_with_model: graph splits = 2 INFO [main] model loaded | tid="140026374909952" timestamp=1718185245 time=2024-06-12T09:40:45.713Z level=INFO source=server.go:567 msg="llama runner started in 2.26 seconds" [GIN] 2024/06/12 - 09:40:46 | 200 | 3.363909682s | 127.0.0.1 | POST "/api/chat" [GIN] 2024/06/12 - 09:40:46 | 200 | 204.205184ms | 127.0.0.1 | POST "/v1/chat/completions"
OS
Linux
GPU
Nvidia
CPU
No response
Ollama version
No response