kasnerz / factgenie

A Toolkit for Annotating and Visualizing LLM Hallucinations
MIT License
1 stars 0 forks source link

Simplify defining span annotations categories #26

Open oplatek opened 2 weeks ago

oplatek commented 2 weeks ago

Current state

For LLM evaluation user specifies LLMMetric config in the yaml config and possibly implements/extends existing LLM metric and changes the LLMMetric factory in factgenie/evaluate.py. Note all the metrics according to llm-eval/*yaml configs are loaded and offered. So most of the LLM annotation campaign is defined in code. With the single exception of the error_categories which one must insert via web browser dialog and they must match the categories specified in the yaml config for the LLM prompt.

Proposal