Closed paulpalmieri closed 4 months ago
In order to do more detailed benchmarks and comparison, having more details than just the total number of consumed tokens would be helpful.
Ideally a breakdown by input/output tokens with the type of llms, show the name of the models used as well.
success rate (%) input | output llm mm_llm embedding
In order to do more detailed benchmarks and comparison, having more details than just the total number of consumed tokens would be helpful.
Ideally a breakdown by input/output tokens with the type of llms, show the name of the models used as well.