-
Hello,
I want to express my gratitude for your outstanding work. The powerful lm-evaluation-harness and your continuous maintenance have made LLM-evaluation much more convenient.
However, I hav…
-
**Describe the bug**
I’m experiencing an issue when fine-tuning the Llama-2-7b model from Hugging Face with Zero optimization enabled. I am running on 8 Intel Max 1550 GPUs using the code from the exa…
-
----
## 🚀 Feature
I would like to propose adding a feature to LitServe that enables models to be deployed with a keep-alive functionality, similar to what Ollama provides. This feature would…
-
code
```python
'''
conda activate beyond_scale_2_unsloth
'''
import torch
from datasets import load_dataset
from trl import SFTConfig, SFTTrainer
from unsloth import FastLanguageModel
from tr…
-
I am not very experienced, but I loveee this package. However, my GPU acceleration seems to only utilize about 1% of my GPU. Increasing the batch size made my predictions far less accurate. And I read…
-
Hi,
I really love this project and am willing to participate. This project has shown how powerful new optimization techniques are.
However, H100 have been banned in China from the day it's out on t…
-
**Describe the bug**
When running auto3d to generate 3D structures with multiple GPUs (e.g., using --gpu_idx 0,1), I encounter an OSError related to an invalid file, arated.sdf. This error does not o…
-
### Problem Description
This [page](https://rocm.blogs.amd.com/software-tools-optimization/hipify/README.html#a-unified-wrapper-header) links to https://github.com/rocm/rocm-blogs/tree/release/blogs/…
-
## Motivation
GPU resources used for storage optimization experiments aren't free and need someone to foot the bill.
## Completion Criteria
Have clear agreement on who is paying for the GPU reso…
-
From @maxdeest
runners * chunks_per_runner * chunk_size ~= 100000 * [max GPU memory in GB] * 2
For Ampere A100 GPU, max out the 40 GB GPU mem
`--chunk_size 3000 --gpu_runners_per_device 8 --ch…