by Re-Align · Agent Tool · ★ 90
Last updated: · Indexed by AgentSkillsHub · Auto-synced every 8h
A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.
| Stars | 90 |
| Forks | 7 |
| Language | Python |
| Category | Agent Tool |
| License | MIT |
| Quality Score | 42.75/100 |
| Open Issues | 2 |
| Last Updated | 2024-01-29 |
| Created | 2023-11-19 |
| Platforms | python |
| Est. Tokens | ~1201k |
Looking for a just-eval alternative? If you're comparing just-eval with other agent tool tools, these 6 projects are the closest alternatives on Agent Skills Hub — ranked by topic overlap, star count, and community traction.
AI agent simulation framework
Build, Improve Performance, and Productionize your LLM Application with an Integrated Framework
OpenClawProBench is a live-first benchmark harness for evaluating LLM agents in the OpenClaw runtime with de
A list of LLMs Tools & Projects
Bayesian Optimization as a Coverage Tool for Evaluating LLMs. Accurate evaluation (benchmarking) that's 10 tim
Official Implementation of Dynamic LLM-Agent Network: An LLM-agent Collaboration Framework with Agent Team Opt
Explore other popular agent tool tools:
just-eval is A simple GPT-based evaluation tool for multi-aspect, interpretable assessment of LLMs.. It is categorized as a Agent Tool with 90 GitHub stars.
just-eval is primarily written in Python. It covers topics such as evaluation, gpt4, llm.
You can find installation instructions and usage details in the just-eval GitHub repository at github.com/Re-Align/just-eval. The project has 90 stars and 7 forks, indicating an active community.
just-eval is released under the MIT license, making it free to use and modify according to the license terms.
The top alternatives to just-eval on Agent Skills Hub include synkro, palico-ai, OpenClawProBench. Each offers a different approach to the same problem space — compare them side-by-side by stars, quality score, and community activity.