0nutation / SpeechGPT

SpeechGPT Series: Speech Large Language Models
https://0nutation.github.io/SpeechGPT.github.io/
Apache License 2.0
1.04k stars 64 forks source link

Performance on smaller models? #18

Open aquorio15 opened 8 months ago

aquorio15 commented 8 months ago

Hi,

I am trying to build a multilingual version of the model that you have used, instead of using Llama-7B, I tried training with a much smaller version of Llama which has 1.1B parameters with over 100k hours of audio data in German and English, but it does not seem to be working well in a multilingual setting, especially in German.

My question is it because the LLM has much fewer parameters as compared to Llama-7b?

Thank You