-
```py
from unsloth import FastLanguageModel
from unsloth import is_bfloat16_supported
import torch
from unsloth.chat_templates import get_chat_template
from trl import SFTTrainer
from transform…
-
I'm trying to fine-tune dbrx on a single machine with 8 H100 gpus. I keep getting OOM error with different configurations, I wonder if this is even doable.
I see a note that suggests 64x80 GPUs, bu…
-
I generated my own lora adapters using the finetune executable from the llama.cpp repository, when I tried to use them in llama.cpp using .bin it works, but the .gguf returns "bad file magic". The thi…
-
Hi and thanks for the great resources.
I used "train-deploy-llama3.ipynb" and trained a similar Llama3 model as shown in the notebook.
I pushed my model on hugging face and now I want to use that …
-
```
│ 154 │ │ lens = len(data_point['input']) │
│ 155 │ │ for i in range(lens-1): …
-
thanks for providing this toolset. i am unsure how utils.extend_model_vocab is intended to be used. in current form, it takes checkpoints only. when i try to adapt it to extend a model in safetensor f…
-
It would be great to see these models work!
> NotImplementedError: Unsloth: /srv/models/Phi-3-medium-4k-instruct not supported yet!
> Make an issue to https://github.com/unslothai/unsloth!
Done…
-
### Summary
# Motivation
WasmEdge is a lightweight inference runtime for AI and LLM applications. Build specialized and finetuned models for WasmEdge community. The model should be supported by Wa…
-
The license of the finetuned checkpoints currently makes no sense.
The base model was almost certainly trained on a ton of unlicensed all-rights-reserved data. In particular, the README says that i…
-
The latest version of `open_instruct` does not contain function `encode_with_prompt_completion_format`, which is used in `utils.py`.
I found the code in:
https://github.com/allenai/open-instruct/b…