Open vladmandic opened 4 months ago
Great suggestion! Making MuLan available for every one is our ultimate goal and we will make some attempts on the smaller one.
~~I used Mini-InternVL-Chat-2B-V1-5 as text_encoder, but it loaded so slow and want to enter yes to trust_remote_code. what is the problem?~~ now, it worked.
Currently MuLan internally uses
OpenGVLab/InternVL-14B-224px
as default text encoder While its possible to pass path to any downloadable encoder, which ones did you test?Note that InternVL-14B-224px is a massive model at 27GB in size and requires ~17GB of VRAM to execute in FP16 context which prohibits usage of this library on any normal consumer GPU