Open sbmaruf opened 1 year ago
From the given line, it appears that the data has been transformed using a gpt2 tokenizer.
https://github.com/yizhongw/Tk-Instruct/blob/7c0e4000901e7c148fbb8964042ebfc433f8d98e/src/convert_data_to_s2s.py#L35
However, the model trained is T5.v.1.-lm-adapted checkpoint. Can you confirm that the model was trained with a T5 tokenizer?
T5.v.1.-lm-adapted
Also all the model cards are same in huggingface,
allenai/tk-instruct-11b-def-pos-neg-expl, allenai/tk-instruct-11b-def-pos, allenai/tk-instruct-11b-def.
allenai/tk-instruct-11b-def-pos-neg-expl
allenai/tk-instruct-11b-def-pos
allenai/tk-instruct-11b-def
Can you confirm that the task encoding for these models are following,
From the given line, it appears that the data has been transformed using a gpt2 tokenizer.
https://github.com/yizhongw/Tk-Instruct/blob/7c0e4000901e7c148fbb8964042ebfc433f8d98e/src/convert_data_to_s2s.py#L35
However, the model trained is
T5.v.1.-lm-adapted
checkpoint. Can you confirm that the model was trained with a T5 tokenizer?Also all the model cards are same in huggingface,
allenai/tk-instruct-11b-def-pos-neg-expl
,allenai/tk-instruct-11b-def-pos
,allenai/tk-instruct-11b-def
.Can you confirm that the task encoding for these models are following,
allenai/tk-instruct-11b-def-pos-neg-expl
: {"add_task_name": False, "add_task_definition": True, "num_pos_examples": 1, "num_neg_examples": 1, "add_explanation": True},allenai/tk-instruct-11b-def-pos
: {"add_task_name": False, "add_task_definition": True, "num_pos_examples": 1, "num_neg_examples": 0, "add_explanation": False},allenai/tk-instruct-11b-def
: {"add_task_name": False, "add_task_definition": True, "num_pos_examples": False, "num_neg_examples": 0, "add_explanation": False},