Closed NumberChiffre closed 2 months ago
Yeah, I agree with you on the specific prompts for smaller models.
Many developers have said that smaller models like qwen2-7B
have troubles on extracting entities and relations, I added a FAQ.md to claim this problem.
Description
Hey guys,
I'm trying out the repo with the same structure as the DeepSeek example with local llama 3.1 8B as both the best and cheap model. The problem is, I get empty dictionaries from the entity extraction results. I thought it was a token context problem, so I even reduced the max token size down to 4096 and even 1024 with the same outcome, so they were probably not the cause. I need your help to figure out why this is not working with llama 3.1 8B, so far I'm thinking the cause could be a combination of these:
I can confirm that this works with either GPT-4 or DeepSeek-v2 chat. Do we need some kind of prompt format specifically for smaller models?
Updates:
Simplifying the entity extraction prompt to something super simple for llama 3.1 8B did not work:
Error output:
Below is the error that I got with llama 3.1 8B:
Code to reproduce the error: