-
Whitepaper: https://arxiv.org/pdf/2306.02707.pdf
Will be released here: https://aka.ms/orca-lm
Summary: https://www.youtube.com/watch?v=Dt_UNg7Mchg
-
```
Traceback (most recent call last):
File "", line 1, in
File "/home/mmior/.local/share/virtualenvs/json-descriptions-cLP9Kwl8/lib/python3.10/site-packages/litgpt/tokenizer.py", line 39, in …
-
Hi, I want to upload finetune Llama 3-instruct to ollama, and following the [https://docs.unsloth.ai/tutorials/how-to-finetune-llama-3-and-export-to-ollama](url) to do it, but it didn't generate the M…
-
Thank you for your valuable contributions to the community; your work looks great! I came across this 70B model: Llama-3.1-Nemotron-70B-Instruct, and its benchmark results are impressive. Could you pl…
-
If we following the script setting of long-llm, the parameter num_train_epoch is set to 1, it will give out really significant improvment over the original model. However, if we change the paramter to…
-
Training LLaMA-13B-4bit on a single RTX 4090 with `finetune.py` (using PyTorch 2 beta, to support the requisite CUDA 11.8 for compute rev 8.9) finishes 3 epochs in only a minute:
```
=============…
-
how to finetune Videollama2 chat models using QLoRA and LoRA.
...
--data_path datasets/custom_sft/custom.json
--data_folder datasets/custom_sft/
--pretrain_mm_mlp_adapter CONNECTOR_DOWNLOAD_PAT…
-
### Request for Release of Pretrained NLLB-LLM2Vec Model
Hello Team,
Could you please release the pretrained NLLB-LLM2Vec models mentioned in your paper on "Self-Distillation for Model Stacking…
-
I installed everything from `requirements.txt`, but still get this error when I run `finetune.py`:
```
Traceback (most recent call last):
File "/work/pi_hzamani_umass_edu/zarifalam_umass_edu/fi…
-
### What is the issue?
No issues with any model that fits into a single 3090 but seems to run out of memory when trying to distribute to the second 3090.
```
INFO [wmain] starting c++ runner | ti…