>_Skillful
Need help with advanced AI agent engineering?Contact FirmAdapt
Back to Agents

OpenAI Evals

AgentLLM Evaluationllmai-appawesome-listawesome-gen-ai-tools

An open-source library for evaluating task performance of language models and prompts.

Directory Presence

Cross-referenced across 55 tracked directories

DirectoryStatusLink
A
Awesome LLM Apps
A
AI Collections

Adoption Metrics

#43

Popularity Rank

2 / 55

Listed In

Emerging

Adoption Stage

2d

Listed For

Recently added to the ecosystem

Cross-Posting Opportunities

Could also be listed in these directories:

Official MCP RegistrySmitheryPulseMCPnpm RegistryPyPIGlamaHugging Face HubAwesome MCP ServersAwesome Claude Skillsmcp.so+43 more

Security Analysis

Score: 100/100

0 dependency vulnerabilities found

Related Agents

LiveBench

A Challenging, Contamination-Free LLM Benchmark.

AgentLLM Evaluation
2 dirs

lighteval

a lightweight LLM evaluation suite that Hugging Face has been using internally.

AgentLLM Evaluation
1 dir

OLMO-eval

a repository for evaluating open language models.

AgentLLM Evaluation
1 dir

simple-evals

Eval tools by OpenAI.

AgentLLM Evaluation
1 dir