-
Beyond LLM supports, 4 evaluation metrics: Context relevancy, Answer relevancy, Groundedness, and Ground truth.
We would be looking forward to add new evaluation metric support to evaluate LLM/RAG…
-
Faithfulness is not callable
-
environment:
python 3.9.20
datasets 3.0.1
langchain 0.3.3
langchain-community 0.3.2
langchain-core 0.3.10
langchain-openai 0.2.2
la…
-
[X] I have checked the [documentation](https://docs.ragas.io/) and related resources and couldn't resolve my bug.
**Describe the bug**
I'm trying to adapt the prompts to 'italian' using "llama3.1"…
-
I am getting error "Failed to parse output. Returning None" on faithfulness metric for some inputs. This is inconsistent behavior as it is haphazard and sometimes works, sometimes doesn't for the same…
-
[ ] I have checked the [documentation](https://docs.ragas.io/) and related resources and couldn't resolve my bug.
**Describe the bug**
Message 'o statements were generated from the answer' was sen…
-
[ ] I have checked the [documentation](https://docs.ragas.io/) and related resources and couldn't resolve my bug.
**Describe the bug**
The answer generated by my model is similar to the ground_tru…
-
Hello @annahedstroem,
It is possible that I am confusing something or calling the function incorrectly but if not, then I think there might be a mistake in how Sensitivity-n is calculated.
### Descr…
-
Can the faithfulness metric can be used as a metric for explanations obtained using LIME or SHAP.
-
## Issue encountered
It would be good to have a system for evaluating both the relevance of the RAG and its use by the LLM in producing the response. My first intuition would be a multi-stage system …