Closed hyang1974 closed 3 months ago
Hi, we provide the LoRA usage and quant usage in our another similar project, you can refer to A3VLM, we also provide the 7B model in that project, check it if you need!
@hyang1974 We have an 7B version just released, which only use 1 GPU to infer
Very attractive work! We are planning to investigate visual large models for industrial robotics use cases and I think your work is a very good reference. But I found your fine-tuning scripts are using slurm and probably running on clusters with 8 A100? I am wondering such fine-tuning/inference can be implemented with less GPU resource, eg, a desktop machine with single GPU? especially for inference?
By the way, I saw you are Ph.D. candidate at SJTU, we are close. We may discuss on this offline if you have time. :)