Closed Muennighoff closed 2 years ago
cc @haileyschoelkopf @lintangsutawika would be great if you could take a look at the prefixlm config. It's using @haileyschoelkopf transformers branch by directly feeding in the causal mask.
@Muennighoff What checkpoint are you using? A quick look at this PRs, my understanding is this is a different model to Bloom?
This is using the checkpoint of the bloom model after conversion to transformers. The slurm script is here: https://github.com/bigscience-workshop/bigscience/blob/d98e577e5740304e200aedb74939aff900684d83/evaluation/results/tr13/tzeroeval/evaluate_t0.slurm
This repo is for reproducing the T0 paper. If changes are needed for BLOOM evaluation, perhaps they would be better put somewhere else?
Edit: For example, we already decided against unpinning the promptsource version: https://github.com/bigscience-workshop/t-zero/commit/fd057a2fc3f3161491437c31aae91a0dfc93ebf0
Agreed good point. We could make it a fork of t-zero on the bigscience org cc @thomasw21 ?
I don't think we need to fork, especially since it'll never get merged. Essentially we can write our own evaluation script in a new repo.
gonna close that one pr since i understand you are working on another repo
Running
/gpfsscratch/rech/six/commun/experiments/muennighoff/bloomckpt/6b3t0/tr13f-6b3-ml-t0-lmtoks168b-t0toks13b-prefix
onsuper_glue,copa,None,"best_option"
resulted in:With prefix: Result: {'accuracy': 0.54} Without prefix: Result: {'accuracy': 0.53}