Open crarau opened 1 year ago
I encountered the same problem. The solution is:
"for example in dataset['train']" change to "for example in dataset"
The input dataset is already the train split, no need to split it again.
Even with this change, this section really can't work as it's shared. The Prompting as it is won't yield an A,B,C,D answer to compare against the answers from the dataset. Even with heavy editing to coerce it into A,B,C,D answers, I'm not getting Llama to behave in this way even as good as completely random (25% accurate):
Accuracy: 10.53%
predicted vs. correct (E means no value returned) | Last 30 chars |
---|---|
D D | ",B,C,D). ASSISTANT: D\n\n\n\n" |
D A | "m (A,B,C,D). ASSISTANT: D" |
E A | "(A,B,C,D). ASSISTANT: \n\n\n\n" |
E A | "A,B,C,D). ASSISTANT: \n\n\n\n\n" |
B D | "m (A,B,C,D). ASSISTANT: B" |
E C | "(A,B,C,D). ASSISTANT: \n\n\n" |
E D | "MSMSMSMSMSMSMSMSMSMSMSMSMSMSMS" |
D A | "m (A,B,C,D). ASSISTANT: D" |
E D | "MSMSMSMSMSMSMSMSMSMSMSMSMSMSMS" |
E D | "(A,B,C,D). ASSISTANT: \n\n\n\n" |
D C | "m (A,B,C,D). ASSISTANT: D" |
E C | "(A,B,C,D). ASSISTANT: D." |
B C | "B,C,D). ASSISTANT: B\n\n\n\n\n" |
D C | "m (A,B,C,D). ASSISTANT: D" |
E C | "-MS-MS-MS-MS-MS-MS-MS-MS-MS-MS" |
E B | "(A,B,C,D). ASSISTANT: \n\n\n" |
D D | "m (A,B,C,D). ASSISTANT: D" |
E C | "MSMSMSMSMSMSMSMSMSMSMSMSMSMSMS" |
C D | "m (A,B,C,D). ASSISTANT: C" |
I'm getting
When running
1.8 Performance Boost via Soft Prompting
on colabFull message