Closed zkx06111 closed 2 months ago
I used voyage-code-2 to embed the repos before running the evaluations. I encountered this issue and added truncation=True
to the voyage config. But it's probably needed for text-embedding-3-small
as well then.
If you want to reproduce the exact flow I've used you can use voyage-code-2
instead as it gives somewhat better results when I evaluated only the vector retrieval solution. I'm not sure how big the difference is for the whole flow though as I never run it with text-embedding-3-small
.
As it's kind of time consuming to created these indexes I plan to upload them somewhere also so it's easier to reproduce the results.
Thanks for the quick response! I'll use voyage then. Also, I think the OpenAIEmbedding in llama index does not have truncation as a parameter.
I pushed the notebook I used for ingestion https://github.com/aorwall/moatless-tools/blob/main/notebooks/ingest.ipynb Not sure it works properly though as I've done some refactorings lately.
One thing that will decrease indexing time is to sort the instances by date to get as few changes between each commit ( instances = sorted(instances, key=lambda x: x["created_at"])
by the way do you have any idea why the epicsplitter is not preventing this from happening?
I was trying to reproduce the results on SWE-Bench and got the invalid request error for exceeding maximum context after ~30 instances.
Have you run into this issue while evaluating on SWE-Bench? Do you think we can just truncate the code to go around it?