-
Hi there, I wrote two methods that allow unsloth models to be loaded into memory and unloaded into memory. To my knowledge, I believe this is the only way to do change unsloth models
```
llm_mode…
-
LLM Summarizer takes standard format data and summarizes it in English language.
-
### Description
When using memory=True for a crew that uses Azure Open AI, there is an error creating long term memory.
### Steps to Reproduce
```
import os
from chromadb.utils.embedding_…
-
A continuation from task #15. Should include an in-depth description of the technology behind the LLMs and of the training and inference. Finish the section
This issue should neatly be tied together …
-
We should update [the Training Operator ROADMAP](https://github.com/kubeflow/training-operator/blob/master/ROADMAP.md) with 2024 work items.
Let's discuss it during [the upcoming Training WG calls]…
-
Dump raw training data for the LLM-jp-3 series. For each training instance, the following fields should be included at least:
- `token_ids`: A list of token IDs for the training instance
- `training…
-
As a researcher I want to come to Hub and find tools, data and compute relevant to developing novel LLM, test it on benchmarks and public results based on comparison with other LLMs.
## User journe…
-
## 🚀 Feature Request
Supporting TP and SP seems quite easy to do with the `replication parameter:
```
replication = tp * sp
```
I have tried various ways to enable PP without success (unexp…
-
Hello. I’m very excited about the big V2 update and have managed to fine-tune your model on 5 hours of my high-quality data. However, I have a few questions:
1. I was using the Colab version here: …
-
**Command: tune run lora_finetune_single_device --config llama3_1/8B_lora_single_device**
**Output**:
```
INFO:torchtune.utils._logging:Running LoRAFinetuneRecipeSingleDevice with resolved config:…