-
Hi,
I met a bug on save_file when I used DeepSpeed.
How should I fix it?
```
override steps. steps for 1 epochs is / 指定エポックまでのステップ数: 18165
[2023-07-09 14:35:23,072] [INFO] [logging.py:96:log_…
-
### Describe the bug
when I was training dreambooth lora sdxl script on dag dataset, it output the errors as following:
ValueError: Attempting to unscale FP16 gradients.
### Reproduction
export MO…
-
### Checklist
- [ ] The issue exists after disabling all extensions
- [X] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a …
-
@BarakKatzir has been developing the [`types-scipy-sparse`](https://github.com/BarakKatzir/types-scipy-sparse) stub package for `scipy.sparse`. A large portion appears to more complete that the `scipy…
-
I've been using DeepSpeed successfully with my large model train jobs. But [this](https://www.microsoft.com/en-us/research/blog/zero-2-deepspeed-shattering-barriers-of-deep-learning-speed-scale/) blog…
-
Often it is needed to move model results to cpu (or inputs to gpu). Once the data structures get a bit complicated, dicts and lists appear often in model results. Often we have to roll a little utilit…
-
Starting a new topic since there is a concrete example:
You'll need to install coffea from this branch: `https://github.com/CoffeaTeam/coffea/tree/awkward2_dev` (pip install -e '.[dev]')
You'll ne…
-
## ❓ Questions and Help
### Before asking:
1. search the issues.
2. search the docs.
#### What is your question?
I want to do fine tuning of the BART summary model.
The machine spec I'…
-
### Bug description
I was able to fine-tune a 8B LLM using Huggingface training framework with PEFT+DeepSpeed stage 2 under fp16 precision(mixed precision training). Recently I would like to change…
-
The attachments contain details of the warning that I encountered while working on some dataset. Kindly review it. In case of bug, fix it.
WARNING:tensorflow:Entity could not be transformed and wi…