An open-source visual programming environment for battle-testing prompts to LLMs.
Cross-referenced across 55 tracked directories
#3826
Popularity Rank
1 / 55
Listed In
Emerging
Adoption Stage
3/26/2023
Created
2,959
GitHub Stars
Score: 100/100
0 dependency vulnerabilities found
Run an AI-powered security scan to analyze this package's source code for vulnerabilities, prompt injection vectors, data exfiltration risks, and behavior mismatches.
Scans fetch actual source code from the GitHub repository, not just the README.
Jeffrey Ip
The LLM Evaluation Framework
Google, LLC
LLM Comparator: An interactive visualization tool for side-by-side LLM evaluation
Awesome Gen AI Tools: LLM Benchmarks: MMLU, HellaSwag, BBH, and Beyond - Confident AI
Awesome Gen AI Tools: Cleanlab Trustworthy Language Model: Score the trustworthiness of any LLM response
252
Forks
71
Open Issues
1/2/2026
Last Commit
Recently added to the ecosystem