OSError: lmms-lab/llava-onevision-qwen2-72b-ov does not appear to have a file named preprocessor_config.json.
Checkout 'https://huggingface.co/lmms-lab/llava-onevision-qwen2-72b-ov/tree/main' for available files.
so maybe you dont offer this option at this moment;
which results in
ValueError: Model architectures ['LlavaVidForCausalLM'] are not supported for now.
Can I even use the generic launch scripts from SGLang or do I have to run controller.py and sglang_worker.py from the llava_next repo? If so, what would be the 'sgl-endpoint' argument there? I am a little lost for now :D
Hi there,
first a quick disclaimer: I did never use SGLang before, but want to serve the model on our cluster for research purposes now;
Executing
results in:
so maybe you dont offer this option at this moment;
Also tried
which results in
ValueError: Model architectures ['LlavaVidForCausalLM'] are not supported for now.
Can I even use the generic launch scripts from SGLang or do I have to run controller.py and sglang_worker.py from the llava_next repo? If so, what would be the 'sgl-endpoint' argument there? I am a little lost for now :D