dusty-nv / jetson-containers

Machine Learning Containers for NVIDIA Jetson and JetPack-L4T
MIT License
2.18k stars 448 forks source link

./autotag minigpt4 can't start the docker #402

Open ztbupt opened 7 months ago

ztbupt commented 7 months ago

I pulled the docker images and download the models and places them to folders. Because of the network issues, I can not download them by command.

But failed to start the docker. Below is the logs. The hardware is Orin NX 8G. Please help to fix. Thanks!

em3@em3-desktop:~/Downloads/jetson-containers$ ./run.sh $(./autotag minigpt4) /bin/bash -c 'cd /opt/minigpt4.cpp/minigpt4 && python3 webui.py /data/models/huggingface/datasets--maknee--minigpt4-7b-ggml/minigpt4-7B-f16.bin /data/models/huggingface/datasets--maknee--ggml-vicuna-v0-quantized/ggml-vicuna-7B-v0-q5_k.bin' Namespace(disable=[''], output='/tmp/autotag', packages=['minigpt4'], prefer=['local', 'registry', 'build'], quiet=False, user='dustynv', verbose=False) -- L4T_VERSION=35.4.1 JETPACK_VERSION=5.1.2 CUDA_VERSION=11.4.315 -- Finding compatible container image for ['minigpt4'] dustynv/minigpt4:r35.4.1 [sudo] password for em3: MoTTY X11 proxy: No authorisation provided xhost: unable to open display "localhost:10.0"

dusty-nv commented 7 months ago

@ztbupt I think this stopped working in 8GB memory at some point, also that maknee/minigpt4-7b-ggml model became unmaintained (and GGML format was replaced in llama.cpp by GGUF). Also Mini-GPT4 has been surpassed by other VLMs like Llava, ect.

However, I did just recently get another mini-VLM (NousResearch/Obsidian-3B-V0.5) based on Llava working on Orin Nano 8GB, and am getting it ready to post on Jetson AI Lab!

ztbupt commented 7 months ago

@ztbupt I think this stopped working in 8GB memory at some point, also that maknee/minigpt4-7b-ggml model became unmaintained (and GGML format was replaced in llama.cpp by GGUF). Also Mini-GPT4 has been surpassed by other VLMs like Llava, ect.

However, I did just recently get another mini-VLM (NousResearch/Obsidian-3B-V0.5) based on Llava working on Orin Nano 8GB, and am getting it ready to post on Jetson AI Lab!

Thanks dusty-nv! Which llm contain do you suggest to use on Orin 8G modules? I run text-generation-webui, but looks like work slowly. The minigpt4 is failed due to memory block.

dusty-nv commented 7 months ago

@ztbupt I take it that you have tried Llama-2-7B GGUF on Nano. Using MLC on Orin Nano, for Llama-2-7B I get 16.4 tokens/sec. I am preparing a tutorial page for Jetson AI Lab about Small Language Models (SLMs) that go higher.

ztbupt commented 7 months ago

@ztbupt I take it that you have tried Llama-2-7B GGUF on Nano. Using MLC on Orin Nano, for Llama-2-7B I get 16.4 tokens/sec. I am preparing a tutorial page for Jetson AI Lab about Small Language Models (SLMs) that go higher.

Got it. Looking forward your tutorial page.