-
I just run `run_jailbreak.py` by using `pyhton run_jailbreak.py`.
but it turns out that:
``` APL
IndexError: index out of range in self
```
I find that, the vocab_size is 32000, but there are 3…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Is your proposal related to a problem or functionality gap?
Currently we are required to upload the data…
-
I am trying to use Owen-2.0 in mllm. I converted the model, vocab using the given tools. However, the outputs of Owen-2.0 was garbled. Do I need to do any further modification on the code of mllm?
-
I need some clarity regarding using a language model's word vocabulary from its training data. Is it essential to stick to the exact vocabulary during usage? Your insights would be much appreciated.
-
These are our v0.1 terms for the general vocabulary:
- funder
- institution
- model (alias: algorithm)
- licence_category
- instrument_type
- instrument
- variable
- platform_type (e.g.: "sate…
-
Hi, wei,
Thanks for your public. However, I can not found the code for write vocab.txt or the data of vocab.txt. Can you provide that for me?
Thanks!
-
Hi, I noticed that the `data.vocab` stored in the baseline model has a different vocabulary length compared to the language embedding stored in pretrained model.
For the baseline model "et_plus_h",…
-
hello,
i have some questions about the bert_qg vocab. you just given the preprocessed vocab file but the codes to get vocab and related embed are not in the preprocess.py. would you like to release …
-
what is the codebook size / vocab size for encoded snac data for the various models?
-