Open SefaZeng opened 6 months ago
cc @lintangsutawika
@SefaZeng this is intentional. It's inspired by how XNLI was evaluated in the XGLM paper.
The doc_to_choice
has 3 options which a decoder model is simply required to choose which is the most likely.
I try to test the XNLI results using the latest commit. And I find the inputs have a prefix_token which looks like:
This is the content of
requests
for_loglikelihood_tokens
. Thecontext
is '' andcontext_enc
is [1]. It is appended in this function fromlm_eval/api/model.py
:As XNLI's config (xnli_zh.yaml) is this:
It merges both premise and hypothesis and has no
context
. So, the code will add aprefix_token_id
to the input for any model. This input format seems wired for base model as most LLM do not usebos
oreos
in the start of inputs (maybe except Gemma).