Closed y12uc231 closed 3 months ago
Hi, I believe that Olmo's vocab size != it's embedding size (see https://huggingface.co/allenai/OLMo-7B/blob/main/config.json). In general, this can sometimes happen when model makers want to leave extra space for new tokens, or to pad out to an even size that slightly improves training.
As for why the scores are mostly -infs, I'm not sure. It'd probably be a good idea to open an issue in the transformers repository or the core olmo repository - this repo is about instruction-tuning models, rather than the specifics of generating from olmo.
🐛 Describe the bug
Here is the code I am running. The goal is to get logprob for each token generated by the chat model.
Here is the error when I run the code above.
Here is where the weird part : the size of the output.scores[0] should be [1, vocab_size] where for olmo vocab_size = 50280 but the size of output.scores[0] = [1, 50304] . How come the outcome is not aligned with the vocab_size. Also the value of outcome.scores is mostly -infs.
Versions
Python 3.10.13