STACKQUADRANT

onejune2018/Awesome-LLM-Eval

Evaluation & Testing

Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs. 一个由工具、基准/数据、演示、排行榜和大模型等组成的精选列表,主要面向基础大模型评测,旨在探求生成式AI的技术边界.

GitHub Metrics
Stars
614
Forks
51
Open Issues
9
Watchers
9
Contributors
5
Weekly Commits
0
Language
License
MIT
Last Commit
Nov 24, 2025
Created
Apr 26, 2023
Latest Release
Release Date
Synced: Mar 3, 2026
Quality Scores
Documentation Qualityw: 20%
0.0
Community Healthw: 20%
0.0
Maintenance Velocityw: 15%
0.0
API Design & DXw: 20%
0.0
Production Readinessw: 15%
0.0
Ecosystem Integrationw: 10%
0.0
Tags
awsome-listawsome-listsbenchmarkbertchatglmchatgptdatasetevaluationgpt3large-language-model
Radar
No scores yet