Open vincent-revotech opened 5 months ago
Hi!
Exactly the same way, you would use the model locally. Choose the appropriate framework, such as transformers adjust the score.py
init
to load the model (e.g.
from transformers import AutoModelForCausalLM
model = AutoModelForCausalLM.from_pretrained(
"your/llm", device_map="auto", load_in_4bit=True
)
) and the code in score
to make the predictions (e.g. model.generate()
.
Hi,
This notebook is awesome. I have a question, what if the model I want to deploy is not the sentence-transformers list, then how can I do that? For example, with this model https://huggingface.co/liuhaotian/llava-v1.5-7b Thanks,