This repository contains a custom implementation of the BERT model, fine-tuned for specific tasks, along with an implementation of Low Rank Approximation (LoRA). The models are optimized for high performance using NVIDIA's TensorRT.
Thank you for documenting everything you learned. It is very helpful. I have been trying to find a pre-coded Q-LORA for BiomedCLIP but I couldn't so I have to do it on my own. BioMedCLIP uses a BERT model as text encoder and ViT as vision encode with contrastive learning. Do you think we can apply Q-Lora to BERT and ViT separately and then join them?
Thank you for documenting everything you learned. It is very helpful. I have been trying to find a pre-coded Q-LORA for BiomedCLIP but I couldn't so I have to do it on my own. BioMedCLIP uses a BERT model as text encoder and ViT as vision encode with contrastive learning. Do you think we can apply Q-Lora to BERT and ViT separately and then join them?