fe1ixxu / ALMA

State-of-the-art LLM-based translation models.
MIT License
440 stars 35 forks source link

GPUs used during parallel data fine-tuning #55

Closed liangyingshao closed 4 months ago

liangyingshao commented 4 months ago

Thank you for your outstanding work. Could you please provide details on the following?

  1. The number of GPUs used during the parallel data fine-tuning.
  2. The type of the GPUs used during the parallel data fine-tuning.
fe1ixxu commented 4 months ago

Hi, thanks for your interest!

  1. 16 GPUs were used during the parallel data fine-tuning.
  2. The models were trained on MI200 GPUs.