-
Hi, thanks for your wonderful work and congratulations on your paper being accepted by ICLR 2023.
But I have some questions about the expression (i.e., "A sentence with a large norm is usually not vi…
-
I have setup the enviroment and downloaded the dataset using the dockerfile offered in the repo, and I have already modified the data locations in config files. When I execute `python driver.py --conf…
-
I'm getting an IsADirectory error when I run the CLI demo.
```bash
python cli_demo.py
____ _ ____ __ __ …
-
Hi,
1. What is the pretraining corpus of `GLM-Large-Chinese`/`GLM-10B-Chinese` released ? `Wiki+BookCorpus` in README or `wudao baike zhihu`(in `config/ds_block_large_chinese.sh`) ?
2. Besides, how …
-
Model I am using (VLMO), I found that the text-onlt data is loaded from "wikibk.{index}.txt" where index=0,1,...,49,I want to ask I can I get the .txt files?
-
Hello!
I am wondering what the correct data preprocessing command is for the final recipe. Could you add this information to the README?
Also, is there a straight forward way to restrict memory …
-
It seem that the bookcoprus data downloaded through the library was pretokenized with NLTK's Treebank tokenizer, which changes the text in incompatible ways to how, for instance, BERT's wordpiece toke…
-
1. download the vicuna model from this: [vicuna model](https://huggingface.co/lmsys/vicuna-7b-v1.3)
2. because of network problem, i download the book corpus.tar.bz2 and uncompress it:
and …
-
Hi, thanks for your excellent work.
As described in the Experiment Settings Section, the pretraining of retromae with enhanced decoding was finished on 8*A100 GPUs. Could you please tell us how man…
-
Hi! Thank you for sharing the code for LogiGAN paper.
I'm having troubles creating training set. In particular:
1. [Here](https://github.com/microsoft/ContextualSP/blob/master/logigan/corpus_constru…