Closed matankley closed 8 months ago
It's hard to say what's happening. When I run the example as is with and without an Alpaca LoRA over the original Llama-7B, here is the output:
--------------------------
No LoRA:
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Write three tweets explaining that the Earth is not flat, using spaghetti and meatballs as an analogy.
### Response:
1. "Earth is round like a spaghetti ball." (Tweet)
2. "No matter how you look at it, the spaghetti ball is round." (Tweet)
3. "You can put this on your plate or eat it straight from the jar." (Tweet)
## External links
* Twitter
* SAT
* SAT 2016: A Revised Scoring Plan
--------------------------
Yes LoRA:
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Write three tweets explaining that the Earth is not flat, using spaghetti and meatballs as an analogy.
### Response:
1. Spaghetti and Meatballs are not flat. Just like the Earth, they are round.
2. The Earth is round too, just like the spaghetti and meatballs.
3. If you want to learn more about why the Earth is round, try looking up some science articles on Google Scholar.
It's very consistently sticking to the format imposed by the adapter. Without the LoRA it's trying its best to interpret the ###
tags and improvising as you'd expect a base model to do.
So it's hard to say what's happening in your case. Mistral-instruct is already very heavily tuned and may have strong preferences that whatever LoRA you're using can't overcome. Do you have a more complete example including the calls to generate_with_lora
?
@turboderp
I'm running the lora example, but im getting the same results for inference with lora adapters and for inference without adapters.
Here is the code im running
Can you please assist with that?