irthomasthomas / undecidability

2 stars 1 forks source link

WisdomShell/kieval: A Knowledge-grounded Interactive Evaluation Framework for Large Language Models #829

Closed ShellLM closed 3 weeks ago

ShellLM commented 3 weeks ago

WisdomShell/kieval: A Knowledge-grounded Interactive Evaluation Framework for Large Language Models

Snippet

A knowledge-grounded interactive evaluation framework for large language models.

README

Description

kieval is a framework for interactively evaluating the capabilities of large language models (LLMs) in a knowledge-grounded setting. It allows users to engage in natural conversations with an LLM, while providing the model with relevant background knowledge to aid in its responses. kieval supports a variety of knowledge sources, including Wikipedia, academic papers (via arXiv), and custom knowledge bases.

Key Features

Usage

To use kieval, simply install the package and run the provided script:

pip install kieval
kieval

This will launch the interactive evaluation interface, where you can begin conversing with the LLM and exploring its capabilities.

Contributing

We welcome contributions to the kieval project. If you're interested in helping to develop new features, improve the existing codebase, or address any issues, please check out the contribution guidelines for more information.

Suggested labels

{'label-description': 'Evaluation Framework', 'label-name': 'evaluation-framework', 'gh-repo': 'WisdomShell/kieval', 'confidence': 67.2}

ShellLM commented 3 weeks ago

Related content

750 similarity score: 0.86

743 similarity score: 0.86

494 similarity score: 0.85

333 similarity score: 0.85

813 similarity score: 0.85

309 similarity score: 0.85