Closed pikalover6 closed 1 year ago
Do you mean the llama 33 billion weight model? If so, given it takes three times as long (twice the size and 1.4B tokens) to train as the 13b I suspect not.
Do you mean the llama 33 billion weight model? If so, given it takes three times as long (twice the size and 1.4B tokens) to train as the 13b I suspect not.
No, the creators have said that they trained a 20b model on discord/Twitter.
I think the 20B model is trained by Stability AI, and we are not involved in that, so I wouldn't consider that part of this project.
Do you plan on releasing the v1 20b model?