From 4d8fd07dfd4cb12138a9f83c0e23537267aa1405 Mon Sep 17 00:00:00 2001 From: jharris1679 Date: Sun, 21 Dec 2025 16:42:32 -0500 Subject: [PATCH] Move sniffbench to new Evaluation section As requested by maintainer, created a new Evaluation section for tools that benchmark/evaluate AI coding agents, distinct from Testing tools that help generate tests. --- README.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/README.md b/README.md index 0ae7c38..b8df920 100644 --- a/README.md +++ b/README.md @@ -19,6 +19,7 @@ This is a curated list of AI-powered developer tools. These tools leverage AI to - [OpenAI plugins](#openai-plugins) - [Search](#search) - [Testing](#testing) +- [Evaluation](#evaluation) ## IDEs - [Google Antigravity](https://antigravity.google/) — An agent-first IDE that orchestrates autonomous AI agents to plan, execute, and verify complex coding tasks with deep browser integration. @@ -253,4 +254,7 @@ This is a curated list of AI-powered developer tools. These tools leverage AI to - [Test Gru](https://gru.ai/home#test-gru) — Provides enterprise-level unit test automation services. - [AgentsKB](https://agentskb.com) - Expert knowledge layer for AI assistants. Your AI searches, we research. That's the difference. + +## Evaluation + - [sniffbench](https://github.com/AnswerLayer/sniffbench) — Benchmark suite for evaluating coding agents. Compare configurations, track metrics, and A/B test with real issues from your repos.