Closed sauravgrd closed 5 days ago
LLM Foundry supports either MPT models, or models that are on huggingface. if you come with an unknown architecture, LLM Foundry will not be able to train it. Please provide more information if I misunderstand your question.
❓ Question
I have developed a pretrained model on my own dataset using custom Attention(without using MosaicML llmfoundary). But when i try to finetune the pre-trained model via mosaicML by giving my model's local or google cloud storage(in the finetuning config yaml)), MosaicML downloads the model artifact(which it considers appropriate) from HuggingFace rather than using my model location Wanted to check how can i use my model's artifact while finetuning?
Additional context
Note: saved my pretrained model using torch.save()