Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
onejune2018/Awesome-LLM-Eval

Awesome-LLM-Eval

Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs. 一个由工具、基准/数据、演示、排行榜和大模型等组成的精选列表,主要面向基础大模型评测,旨在探求生成式AI的技术边界.

56.5/100
632Forks: 60
View on GitHubHomepage →
Loading report...

Similar Projects

Static-to-Dynamic-LLMEval

54

The official GitHub repository of the paper "Recent advances in large language model benchmarks against data contamination: From static to dynamic evaluation"

505

SuperCLUE

51

SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese

3.3K

Awesome-LLM-Long-Context-Modeling

74

📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥

2.0K

LLM-eval-survey

65

The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".

1.6K
Back to List