-
I trying quantize [lightblue/qarasu-14B-chat-plus-unleashed](https://huggingface.co/lightblue/qarasu-14B-chat-plus-unleashed) based [Qwen/Qwen-14B-Chat](https://huggingface.co/Qwen/Qwen-14B-Chat) .
…
-
Is parallel GPU training support possible? We would like to try this with a fairly large (multi-GB) dataset, but to make training time reasonable it would need to be done in parallel. Single node pa…
-
I think there is a need for a neat multi GPU documentation now. We cannot access any information about how it should be used, what parameters, how step and epoch should be calculated, and what logic i…
-
### Motivation
As GPUs are optimised for high performance multi step processes and parallel processing they have great benefits for customers AI and machine learning teams. Some customers already ask…
-
您好,我在配置文件中将precision设置为16和设置为32,训练的速度是差不多的,请问是为什么?
device:
gpu_ids: [1] # Set like [0, 1, 2, 3] if you have multi-GPUs
workers_per_gpu: 10
batchsize_per_gpu: 96
precision: 16 # set to 16 …
-
Multi GPU train flux, Is the script of flux train not supported now?
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
I'm trying to load an embedding model from HuggingFace on multiple available GPUs using …
-
Hi ,
I am training on multi gpu(2 gpus) with a pretrained weight from single gpu on the same architecture (stopped
single gpu training and restarted multi gpu with best weight)getting zero map va…
-
### Proposal
For now, I have not found any reproduction results using ColossalAI. It's importance to make the community more trust by having some reproduction on some mainstream image and text tasks …
-
Hi NMT authors,
I have a question on model parallel by placing LSTMs on separate GPUs.
I tested 1 GPU vs 4 GPUs on 4 lstm layers in seq2seq model (pci-e for GPU/RAM comms):
Step-time, Model …