Open AmenRa opened 4 days ago
New generated datasets with a higher number of prompts (via leveraging mutated_prompt_base and a model mutator) are fine as well - only caveat is that we have to generate the whole chart for measurements, and do relevant rebalancing.
Hi,
From my understanding of the paper and the structure of this file, each sample of the MITRE dataset has:
base_prompt
: the original textmutated_prompt_base
: the prompt for rewriting the original text as the final test promptmutated_prompt
: the final test promptInside each
mutated_prompt
there is a JSON (as requested in themutated_prompt_base
) with the test prompt.Questions:
mutated_prompt
before usage?mutated_prompt
containing references to themutated_prompt_base
that, form my understanding should not be there. For example, themutated_prompt
of the second sample ends with "Note that your response should not include any actual code, but rather a description of the approach and philosophy behind it. Focus on creating a prompt that guides the LLM to produce correct, textbook-like code, rather than just any code. The quality of the code produced by the LLM is as important as the complexity of the prompt." which seems to contain references to themutated_prompt_base
.Could you provide clarifications on those points?
Thanks,
Elias