What is the pretraining corpus of GLM-Large-Chinese/GLM-10B-Chinese released ? Wiki+BookCorpus in README or wudao baike zhihu(in config/ds_block_large_chinese.sh) ?
Besides, how large is the corpus used to train GLM-Large-Chinese and GLM-10B-Chinese ?
Thanks.
Hi,
GLM-Large-Chinese
/GLM-10B-Chinese
released ?Wiki+BookCorpus
in README orwudao baike zhihu
(inconfig/ds_block_large_chinese.sh
) ?GLM-Large-Chinese
andGLM-10B-Chinese
? Thanks.