Back to Agents
instruct-eval
AgentLLM Evaluationllmai-appawesome-list
A
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
Directory Presence
Cross-referenced across 55 tracked directories
Adoption Metrics
#308
Popularity Rank
2%
Adoption Rate
Emerging
Adoption Stage
54
Unlisted Directories
Recently added to directories
Cross-Posting Gaps
Not yet listed in these active directories:
Official MCP RegistrySmitheryPulseMCPnpm RegistryPyPIGlamaHugging Face HubAwesome MCP ServersAwesome Claude Skillsmcp.soOpenToolsmcp-getCursor DirectoryGitHub SearchLobeHubBest of MCP ServersMCPMarketTensorBlock Awesome MCPCline MCP MarketplaceAnthropic Official Skillsnpm Skills RegistryPyPI Skills RegistryAwesome AI AgentsAwesome MCP ToolsComposioLangChain HubCrewAI ToolsZapier MCPModelScopeAutoGen (Microsoft)LlamaIndex ToolsOpenAI EcosystemMCP HubChatGPT PluginsSuperAgent & FrameworksDify MarketplaceRaycast AI ExtensionsCloudflare Workers AIVercel AI SDKGitHub Copilot ExtensionsWindsurf RulesOllama LibraryReplicateToolhousePhidata/AgnoAgenticAI CollectionsVS Code MarketplaceJetBrains Marketplacen8n Community NodesHugging Face SpacesOpenRouterPyPI (Real-time)npm Registry (Real-time)
Security Analysis
Score: 100/100
0 dependency vulnerabilities found
Related Agents
An open-source library for evaluating task performance of language models and prompts.
AgentLLM Evaluation
2 dirs
a lightweight LLM evaluation suite that Hugging Face has been using internally.
AgentLLM Evaluation
1 dir