Open jedld opened 9 months ago
Getting the same problem using the latest jetpack and the recommended actions. Hope this comes back soon to the orin nano
I got it somewhat working by using the lowest quantization possible and using CPU only (removing the CUDA options in the minigpt4 docker build). But it would be nice if it worked out of the box.
I'm using agx xavier (which have 32G ), and can run successfully. but the response frequency is too slow. I also tried on agx orin, it's the same. is that normal?
@UserName-wang Mini-GPT4 is not the most optimized, Llava we have more optimized. You can see the benchmarks at https://www.jetson-ai-lab.com/benchmarks.html
For the most optimized VLM pipeline with Llava-1.5, see https://www.jetson-ai-lab.com/tutorial_llava.html#4-optimized-multimodal-pipeline-with-local_llm
Tried running the minigpt4 webui demo (https://www.jetson-ai-lab.com/tutorial_minigpt4.html) and my device keeps locking up after running the run script. Figured I may lack the memory resources to do it, hence I opted for a smaller model and followed the instructions in tuning the memory here:
https://www.jetson-ai-lab.com/tips_ram-optimization.html
So I setup swap, ran headless and disable some services.
Ran the benchmark again headless using the commands below:
Still got killed.
The webui shows similar behavior, after adding swap I got to the point where it can show the gradio UI, however attempting to upload an image causes the same OOM issue.
Would appreciate any tips or help.