Closed sunyuhan19981208 closed 1 year ago
Hi! Thanks for your interest!! So, we've used Flan-T5 primarily in our open-source experiments, but the instruction-following Flan models are pretty good at staying concise (which doesn't seem to be the case for base Llama 7B).
You can probably use ChatGPT (or a cheaper LLM) to filter the answer by giving it the output from Llama, and prompt it to pick the first answer from the output. You probably have to hand-engineer the prompts and validate with a couple of examples, but it might work!
LLM extract may be too expensive for me... How do you think of few-shots or logits?
Honestly, maybe Llama 2 chat could serve as an extractor? I think logits across the answer options (A, B, C, etc.) should be fine!
I'd warn against few shot since that also has an effect. See this related paper (https://arxiv.org/pdf/2305.04388.pdf) for few-shot experiments.
Hope that helps!
Description
I am seeking your guidance and expertise on a matter related to utilizing causal language models for reproducing bias experiments. Specifically, I have been experimenting with the Llama-7b model, but I have encountered some challenges that I would like to discuss and request your advice on. My primary concern with Llama-7b is that it tends to generate responses that do not align with the desired format for answers in bias experiments. Rather than providing the intended options, Llama appears to directly answer the content of the input prompt. This has proven to be a limitation in my efforts to reproduce bias experiments accurately using Llama. Additionally, I am aware that using logits or employing few-shot learning methods may introduce unfair advantages or bias into the experiment results. I am committed to conducting fair and unbiased experiments, and I would appreciate any insights or recommendations you may have for mitigating these concerns when working with causal language models. Thank you for your time and consideration. I look forward to your insights and advice.
Here is how Llama reply