>_Skillful
Need help with advanced AI agent engineering?Contact FirmAdapt
Back to Agents

instruct-eval

AgentLLM Evaluationllmai-appawesome-list

This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.

Directory Presence

Cross-referenced across 55 tracked directories

DirectoryStatusLink
A
Awesome LLM Apps

Adoption Metrics

#308

Popularity Rank

2%

Adoption Rate

Emerging

Adoption Stage

54

Unlisted Directories

Recently added to directories

Cross-Posting Gaps

Not yet listed in these active directories:

Official MCP RegistrySmitheryPulseMCPnpm RegistryPyPIGlamaHugging Face HubAwesome MCP ServersAwesome Claude Skillsmcp.soOpenToolsmcp-getCursor DirectoryGitHub SearchLobeHubBest of MCP ServersMCPMarketTensorBlock Awesome MCPCline MCP MarketplaceAnthropic Official Skillsnpm Skills RegistryPyPI Skills RegistryAwesome AI AgentsAwesome MCP ToolsComposioLangChain HubCrewAI ToolsZapier MCPModelScopeAutoGen (Microsoft)LlamaIndex ToolsOpenAI EcosystemMCP HubChatGPT PluginsSuperAgent & FrameworksDify MarketplaceRaycast AI ExtensionsCloudflare Workers AIVercel AI SDKGitHub Copilot ExtensionsWindsurf RulesOllama LibraryReplicateToolhousePhidata/AgnoAgenticAI CollectionsVS Code MarketplaceJetBrains Marketplacen8n Community NodesHugging Face SpacesOpenRouterPyPI (Real-time)npm Registry (Real-time)

Security Analysis

Score: 100/100

0 dependency vulnerabilities found

Related Agents

LiveBench

A Challenging, Contamination-Free LLM Benchmark.

AgentLLM Evaluation
2 dirs

OpenAI Evals

An open-source library for evaluating task performance of language models and prompts.

AgentLLM Evaluation
2 dirs

lighteval

a lightweight LLM evaluation suite that Hugging Face has been using internally.

AgentLLM Evaluation
1 dir

OLMO-eval

a repository for evaluating open language models.

AgentLLM Evaluation
1 dir