-
Can the problem be that I have GTX 1050 ti 4 GB? (playing with options to lower VRAM usage does not help), When I play with settings I get the same thing but the last thing changes to returned non-zer…
-
The problem: some systems have async work to do, which may yield. They don't want or simply can't do the work right away. For example, can be called via FFI, or want to collect a batch of such request…
-
**Is your feature request related to a problem? Please describe.**
I'm serving a model that supports batching (`max_batch_size` > 0) and I would like to use config autocomplete, but I don't want to u…
-
I get it at different batches in the first epoch, not always the same. But around 70-80% progress iof the first batch it seems.
```
----------------------------------------------------------------…
-
### Checklist
- [ ] The issue exists after disabling all extensions
- [ ] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a …
-
Sometimes, when training using the SimCLR method I get some divergent loss function (see attached screenshot). I wonder if anyone has ever experienced this kind of issue when training with SimCLR. Thi…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
训练命令:
llamafactory-cli train \
--stage dpo \
--do_train \
--finetuning_type full \
…
-
### Motivation.
Speculative Decoding is a crucial feature for reducing latency, currently supported by vLLM (credit to @cadedaniel !). However, when deploying Speculative Decoding in real online LL…
-
**Description**
The archiver should be integrated into the new architecture. To achieve this we will move the existing archivers from Operate and Tasklist to camunda-exporter, while merging them a…
-
Why is `batch_size*2` passed as a parameter in the following lines instead of `batch_size`?
In `train_LEP.py`:
`noisy_image, noise_level, timesteps = noisy_latent(latent_image, pipe.scheduler, b…