-
### Feature request
I want to add L1/L2 regularization to the transformer training.
### Motivation
Adding L1/L2 reg can promote sparser models that can accelerate inference and reduce storage.
###…
-
hi!when I try to running your demo in PiA part, I get an error in 'instruction tuning' step:
```
root@0de6f5c3da0f:/workspace/zt/code/Sequence-Scheduling# bash train.sh
[2024-10-02 22:24:40,711] …
-
I have a large datasets that was encoded with Clip 32 but I cannot deploy that in ES because it uses the ClipTokenizer.
Is there a way to add this?
-
### Describe the bug
When using 8-bit quantization with the LLM pipeline and a multiple GPU setup, it mostly runs fine.
After some random amount of requests however the pipeline starts failing an…
-
### System Info
When using DeepSpeed, the RLOOTrainer reports an error: "ValueError: Please make sure to properly initialize your accelerator via accelerator = Accelerator() before using any function…
-
Here's the Error:
```
python functioncall.py --query "I need the current stock price of Tesla (TSLA)"
…
-
I started a `inf2.48xlarge` ec2, pull and get into [TGI-Neuron DLC with optimum-neuron 0.0.17 installed](https://github.com/aws/deep-learning-containers/releases/tag/v1.0-hf-tgi-0.0.17-pt-1.13.1-inf-n…
-
Hello developers,
I'm trying to use xDiT (version 3.3) comfyui-xdit on 2 servers with 4 NVIDIA 3090 GPUs. I use the command below to start the service:
```
torchrun --nproc_per_node=2 --nnodes=…
-
## **First good issue**
There have been quite some issues/questions with how to use the Encoder-Decoder model, e.g.: https://github.com/huggingface/transformers/issues/4483 and https://github.com/h…
-
I'm running on Windows 10, my FLUX and many other AI repos work flawlessly even the most error prone ones like Tortoise TTS however I can't fix an error while running FLUXGYM. The AI captions generate…