evaluating-llms-harness

62 stars 2 forks
26

"Evaluates LLMs across 60+ academic benchmarks (MMLU, HumanEval, GSM8K, TruthfulQA, HellaSwag). Use when benchmarking model quality, comparing models, reporting academic results, or tracking training progress. Industry standard used by EleutherAI, HuggingFace, and major labs. Supports HuggingFace, vLLM, APIs."

Third-Party Agent Skill: Review the code before installing. Agent skills execute in your AI assistant's environment and can access your files. Learn more about security

Installation for Agentic Skill

View all platforms →
skilz install zechenzhangAGI/AI-research-SKILLs/evaluating-llms-harness
skilz install zechenzhangAGI/AI-research-SKILLs/evaluating-llms-harness --agent opencode
skilz install zechenzhangAGI/AI-research-SKILLs/evaluating-llms-harness --agent codex
skilz install zechenzhangAGI/AI-research-SKILLs/evaluating-llms-harness --agent gemini

First time? Install Skilz: pip install skilz

Works with 22+ AI coding assistants

Cursor, Aider, Copilot, Windsurf, Qwen, Kimi, and more...

View All Agents
Download Agent Skill ZIP

Extract and copy to ~/.claude/skills/ then restart Claude Desktop

1. Clone the repository:
git clone https://github.com/zechenzhangAGI/AI-research-SKILLs
2. Copy the agent skill directory:
cp -r AI-research-SKILLs/11-evaluation/lm-evaluation-harness ~/.claude/skills/

Need detailed installation help? Check our platform-specific guides:

Related Agentic Skills

Agentic Skill Details

Stars
62
Forks
2
Type
Technical
Meta-Domain
data ai
Primary Domain
machine learning
Market Score
26

Report Security Issue

Found a security vulnerability in this agent skill?