issues
search
onejune2018
/
Awesome-LLM-Eval
Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs. 一个由工具、基准/数据、演示、排行榜和大模型等组成的精选列表,主要面向基础大模型评测,旨在探求生成式AI的技术边界.
MIT License
396
stars
39
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Add MMLU-Pro
#8
hobbytp
opened
3 months ago
1
add trustllm benchmark
#7
HowieHwong
closed
4 months ago
0
remove wrong and lost
#6
zhimin-z
closed
9 months ago
0
make the title shine out
#5
zhimin-z
closed
10 months ago
0
Refactor leaderboard
#4
zhimin-z
closed
9 months ago
0
Add related work
#3
ChanLiang
closed
10 months ago
0
Add references
#2
ChanLiang
closed
11 months ago
0
add CMMLU
#1
yzbx
closed
12 months ago
0