Back to Agents
OpenAI Evals
AgentLLM Evaluationllmai-appawesome-listawesome-gen-ai-tools
A
An open-source library for evaluating task performance of language models and prompts.
Directory Presence
Cross-referenced across 55 tracked directories
Adoption Metrics
#43
Popularity Rank
2 / 55
Listed In
Emerging
Adoption Stage
2d
Listed For
Recently added to the ecosystem
Cross-Posting Opportunities
Could also be listed in these directories:
Official MCP RegistrySmitheryPulseMCPnpm RegistryPyPIGlamaHugging Face HubAwesome MCP ServersAwesome Claude Skillsmcp.so+43 more
Security Analysis
Score: 100/100
0 dependency vulnerabilities found
Related Agents
a lightweight LLM evaluation suite that Hugging Face has been using internally.
AgentLLM Evaluation
1 dir