mirror of
https://github.com/garrytan/gstack.git
synced 2026-05-02 11:45:20 +02:00
14b1ba07e9
All four plan-format cases (CEO mode, CEO approach, eng coverage, eng kind)
now run the judge after the existing regex assertions. Threshold reason_substance
>= 4 catches both boilerplate ("because it's better") and generic ("because
it's faster") tier reasoning — exactly the failure modes the regex couldn't.
Move recordE2E to after the judge call so judge_scores and judge_reasoning
land in the eval-store JSON for diagnostics. Booleans are encoded as 0/1 to
fit the Record<string, number> shape EvalTestEntry.judge_scores expects.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
380 lines
16 KiB
TypeScript
380 lines
16 KiB
TypeScript
/**
|
|
* AskUserQuestion format regression test for /plan-ceo-review and /plan-eng-review.
|
|
*
|
|
* Context: a user on Opus 4.7 reported the RECOMMENDATION line and the
|
|
* `Completeness: N/10` per-option score stopped appearing on AskUserQuestion
|
|
* prompts. This test captures the agent's AskUserQuestion output verbatim
|
|
* and asserts the format rule is applied.
|
|
*
|
|
* Capture shape: `claude -p` sessions inside this harness do not have the
|
|
* AskUserQuestion MCP tool wired. We instruct the agent to write the verbatim
|
|
* AskUserQuestion text it would have made to $OUT_FILE instead of calling
|
|
* any tool. Assertions read that file.
|
|
*
|
|
* Coverage-vs-kind split: the format rule says to include `Completeness: N/10`
|
|
* only when options differ in coverage. When options differ in kind (mode
|
|
* selection, posture choice, cherry-pick Add/Defer/Skip), the score is
|
|
* intentionally absent and a one-line note explains why. Assertions split
|
|
* accordingly.
|
|
*/
|
|
import { describe, test, expect, beforeAll, afterAll } from 'bun:test';
|
|
import { runSkillTest } from './helpers/session-runner';
|
|
import {
|
|
ROOT, runId,
|
|
describeIfSelected, testConcurrentIfSelected,
|
|
logCost, recordE2E,
|
|
createEvalCollector, finalizeEvalCollector,
|
|
} from './helpers/e2e-helpers';
|
|
import { judgeRecommendation } from './helpers/llm-judge';
|
|
import { spawnSync } from 'child_process';
|
|
import * as fs from 'fs';
|
|
import * as path from 'path';
|
|
import * as os from 'os';
|
|
|
|
const evalCollector = createEvalCollector('e2e-plan-format');
|
|
|
|
// Regex predicates applied to captured AskUserQuestion content.
|
|
// RECOMMENDATION regex is lenient on intervening markdown markers (e.g.
|
|
// agent writes `**RECOMMENDATION:** Choose` — the `**` closers are benign).
|
|
// Post v1.7.0.0: "Recommendation:" (mixed-case) is the canonical form per
|
|
// the Pros/Cons format; accept both cases for backward compatibility.
|
|
const RECOMMENDATION_RE = /[Rr]ecommendation:[*\s]*Choose/;
|
|
const COMPLETENESS_RE = /Completeness:\s*\d{1,2}\/10/;
|
|
const KIND_NOTE_RE = /options differ in kind/i;
|
|
|
|
// v1.7.0.0 Pros/Cons format tokens. Tests are additive: existing
|
|
// RECOMMENDATION / Completeness / kind-note assertions still hold; new
|
|
// format tokens are asserted ONLY when the capture is from a v1.7+
|
|
// skill rendering. Presence is optional for backward compatibility during
|
|
// rollout; the periodic-tier cadence+format eval (see skill-e2e-plan-cadence)
|
|
// is the strict gate for the new format.
|
|
const PROS_CONS_HEADER_RE = /Pros\s*\/\s*cons:/i;
|
|
const PRO_BULLET_RE = /^\s*✅\s+\S/m;
|
|
const CON_BULLET_RE = /^\s*❌\s+\S/m;
|
|
const NET_LINE_RE = /^Net:\s+\S/m;
|
|
const D_NUMBER_RE = /^D\d+\s+—/m;
|
|
const STAKES_RE = /Stakes if we pick wrong:/i;
|
|
|
|
const SAMPLE_PLAN = `# Plan: Add User Dashboard
|
|
|
|
## Context
|
|
We're building a new user dashboard that shows recent activity, notifications, and quick actions.
|
|
|
|
## Changes
|
|
1. New React component \`UserDashboard\` in \`src/components/\`
|
|
2. REST API endpoint \`GET /api/dashboard\` returning user stats
|
|
3. PostgreSQL query for activity aggregation
|
|
4. Redis cache layer for dashboard data (5min TTL)
|
|
|
|
## Architecture
|
|
- Frontend: React + TailwindCSS
|
|
- Backend: Express.js REST API
|
|
- Database: PostgreSQL with existing user/activity tables
|
|
- Cache: Redis for dashboard aggregates
|
|
`;
|
|
|
|
function setupPlanDir(tmpPrefix: string, skillName: 'plan-ceo-review' | 'plan-eng-review'): string {
|
|
const planDir = fs.mkdtempSync(path.join(os.tmpdir(), tmpPrefix));
|
|
const run = (cmd: string, args: string[]) =>
|
|
spawnSync(cmd, args, { cwd: planDir, stdio: 'pipe', timeout: 5000 });
|
|
|
|
run('git', ['init', '-b', 'main']);
|
|
run('git', ['config', 'user.email', 'test@test.com']);
|
|
run('git', ['config', 'user.name', 'Test']);
|
|
|
|
fs.writeFileSync(path.join(planDir, 'plan.md'), SAMPLE_PLAN);
|
|
run('git', ['add', '.']);
|
|
run('git', ['commit', '-m', 'add plan']);
|
|
|
|
fs.mkdirSync(path.join(planDir, skillName), { recursive: true });
|
|
fs.copyFileSync(
|
|
path.join(ROOT, skillName, 'SKILL.md'),
|
|
path.join(planDir, skillName, 'SKILL.md'),
|
|
);
|
|
|
|
return planDir;
|
|
}
|
|
|
|
// The capture instruction passed to every case. Tells the agent to dump
|
|
// AskUserQuestion content to a file instead of calling a tool.
|
|
function captureInstruction(outFile: string): string {
|
|
return `Write the verbatim text of every AskUserQuestion you would have made to ${outFile} (one question per session, full text including options and recommendation line). Do NOT call any tool to ask the user. Do NOT paraphrase — include the exact prose you would have shown. This is a format-capture test, not an interactive session.`;
|
|
}
|
|
|
|
// --- Case 1: plan-ceo-review mode selection (kind-differentiated) ---
|
|
|
|
describeIfSelected('Plan Format — CEO Mode Selection', ['plan-ceo-review-format-mode'], () => {
|
|
let planDir: string;
|
|
let outFile: string;
|
|
|
|
beforeAll(() => {
|
|
planDir = setupPlanDir('skill-e2e-plan-format-ceo-mode-', 'plan-ceo-review');
|
|
outFile = path.join(planDir, 'ask-capture.md');
|
|
});
|
|
|
|
afterAll(() => {
|
|
try { fs.rmSync(planDir, { recursive: true, force: true }); } catch {}
|
|
});
|
|
|
|
testConcurrentIfSelected('plan-ceo-review-format-mode', async () => {
|
|
const result = await runSkillTest({
|
|
prompt: `Read plan-ceo-review/SKILL.md for the review workflow.
|
|
|
|
Read plan.md — that's the plan to review. This is a standalone plan document, not a codebase — skip any codebase exploration or system audit steps.
|
|
|
|
Proceed to Step 0F (Mode Selection). This is where the skill presents 4 mode options (SCOPE EXPANSION, SELECTIVE EXPANSION, HOLD SCOPE, SCOPE REDUCTION) to the user via AskUserQuestion. These options differ in kind (review posture), not in coverage.
|
|
|
|
${captureInstruction(outFile)}
|
|
|
|
After writing the file, stop. Do not continue the review.`,
|
|
workingDirectory: planDir,
|
|
maxTurns: 10,
|
|
timeout: 240_000,
|
|
testName: 'plan-ceo-review-format-mode',
|
|
runId,
|
|
model: 'claude-opus-4-7',
|
|
});
|
|
|
|
logCost('/plan-ceo-review format (mode)', result);
|
|
expect(['success', 'error_max_turns']).toContain(result.exitReason);
|
|
|
|
expect(fs.existsSync(outFile)).toBe(true);
|
|
const captured = fs.readFileSync(outFile, 'utf-8');
|
|
expect(captured.length).toBeGreaterThan(100);
|
|
|
|
// Kind-differentiated: RECOMMENDATION required, Completeness: N/10 must NOT appear,
|
|
// "options differ in kind" note must appear.
|
|
expect(captured).toMatch(RECOMMENDATION_RE);
|
|
expect(captured).not.toMatch(COMPLETENESS_RE);
|
|
expect(captured).toMatch(KIND_NOTE_RE);
|
|
|
|
// Recommendation-quality judge: deterministic regex for present/commits/has_because,
|
|
// Haiku 4.5 for reason_substance 1-5. Threshold >= 4 catches generic-tier reasoning.
|
|
const recScore = await judgeRecommendation(captured);
|
|
recordE2E(evalCollector, '/plan-ceo-review-format-mode', 'Plan Format — CEO Mode Selection', result, {
|
|
passed: ['success', 'error_max_turns'].includes(result.exitReason),
|
|
judge_scores: {
|
|
rec_present: recScore.present ? 1 : 0,
|
|
rec_commits: recScore.commits ? 1 : 0,
|
|
rec_has_because: recScore.has_because ? 1 : 0,
|
|
rec_substance: recScore.reason_substance,
|
|
},
|
|
judge_reasoning: `${recScore.reasoning} | reason: "${recScore.reason_text}"`,
|
|
});
|
|
expect(recScore.present, recScore.reasoning).toBe(true);
|
|
expect(recScore.commits, recScore.reasoning).toBe(true);
|
|
expect(recScore.has_because, recScore.reasoning).toBe(true);
|
|
expect(
|
|
recScore.reason_substance,
|
|
`${recScore.reasoning}\n reason: "${recScore.reason_text}"`,
|
|
).toBeGreaterThanOrEqual(4);
|
|
}, 300_000);
|
|
});
|
|
|
|
// --- Case 2: plan-ceo-review approach menu (coverage-differentiated) ---
|
|
|
|
describeIfSelected('Plan Format — CEO Approach Menu', ['plan-ceo-review-format-approach'], () => {
|
|
let planDir: string;
|
|
let outFile: string;
|
|
|
|
beforeAll(() => {
|
|
planDir = setupPlanDir('skill-e2e-plan-format-ceo-approach-', 'plan-ceo-review');
|
|
outFile = path.join(planDir, 'ask-capture.md');
|
|
});
|
|
|
|
afterAll(() => {
|
|
try { fs.rmSync(planDir, { recursive: true, force: true }); } catch {}
|
|
});
|
|
|
|
testConcurrentIfSelected('plan-ceo-review-format-approach', async () => {
|
|
const result = await runSkillTest({
|
|
prompt: `Read plan-ceo-review/SKILL.md for the review workflow.
|
|
|
|
Read plan.md — that's the plan to review. This is a standalone plan document, not a codebase — skip any codebase exploration or system audit steps.
|
|
|
|
Proceed to Step 0C-bis (Implementation Alternatives / Approach Menu). This is where the skill generates 2-3 approaches (minimal viable vs ideal architecture) and presents them via AskUserQuestion. These options differ in coverage (complete vs shortcut), so Completeness: N/10 applies.
|
|
|
|
${captureInstruction(outFile)}
|
|
|
|
After writing the file, stop. Do not continue the review.`,
|
|
workingDirectory: planDir,
|
|
maxTurns: 10,
|
|
timeout: 240_000,
|
|
testName: 'plan-ceo-review-format-approach',
|
|
runId,
|
|
model: 'claude-opus-4-7',
|
|
});
|
|
|
|
logCost('/plan-ceo-review format (approach)', result);
|
|
expect(['success', 'error_max_turns']).toContain(result.exitReason);
|
|
|
|
expect(fs.existsSync(outFile)).toBe(true);
|
|
const captured = fs.readFileSync(outFile, 'utf-8');
|
|
expect(captured.length).toBeGreaterThan(100);
|
|
|
|
// Coverage-differentiated: both RECOMMENDATION and Completeness: N/10 required.
|
|
expect(captured).toMatch(RECOMMENDATION_RE);
|
|
expect(captured).toMatch(COMPLETENESS_RE);
|
|
|
|
const recScore = await judgeRecommendation(captured);
|
|
recordE2E(evalCollector, '/plan-ceo-review-format-approach', 'Plan Format — CEO Approach Menu', result, {
|
|
passed: ['success', 'error_max_turns'].includes(result.exitReason),
|
|
judge_scores: {
|
|
rec_present: recScore.present ? 1 : 0,
|
|
rec_commits: recScore.commits ? 1 : 0,
|
|
rec_has_because: recScore.has_because ? 1 : 0,
|
|
rec_substance: recScore.reason_substance,
|
|
},
|
|
judge_reasoning: `${recScore.reasoning} | reason: "${recScore.reason_text}"`,
|
|
});
|
|
expect(recScore.present, recScore.reasoning).toBe(true);
|
|
expect(recScore.commits, recScore.reasoning).toBe(true);
|
|
expect(recScore.has_because, recScore.reasoning).toBe(true);
|
|
expect(
|
|
recScore.reason_substance,
|
|
`${recScore.reasoning}\n reason: "${recScore.reason_text}"`,
|
|
).toBeGreaterThanOrEqual(4);
|
|
}, 300_000);
|
|
});
|
|
|
|
// --- Case 3: plan-eng-review coverage-differentiated per-issue AskUserQuestion ---
|
|
|
|
describeIfSelected('Plan Format — Eng Coverage Issue', ['plan-eng-review-format-coverage'], () => {
|
|
let planDir: string;
|
|
let outFile: string;
|
|
|
|
beforeAll(() => {
|
|
planDir = setupPlanDir('skill-e2e-plan-format-eng-cov-', 'plan-eng-review');
|
|
outFile = path.join(planDir, 'ask-capture.md');
|
|
});
|
|
|
|
afterAll(() => {
|
|
try { fs.rmSync(planDir, { recursive: true, force: true }); } catch {}
|
|
});
|
|
|
|
testConcurrentIfSelected('plan-eng-review-format-coverage', async () => {
|
|
const result = await runSkillTest({
|
|
prompt: `Read plan-eng-review/SKILL.md for the review workflow.
|
|
|
|
Read plan.md — that's the plan to review. This is a standalone plan document, not a codebase — skip any codebase exploration steps.
|
|
|
|
During your review (Section 3 Test Review is the natural place), generate ONE AskUserQuestion about test coverage depth where the options are clearly coverage-differentiated. For example:
|
|
A) Full coverage: happy path + edge cases + error paths (Completeness 10/10)
|
|
B) Happy path only (Completeness 7/10)
|
|
C) Smoke test (Completeness 3/10)
|
|
|
|
${captureInstruction(outFile)}
|
|
|
|
After writing the file with that ONE question, stop. Do not continue the review.`,
|
|
workingDirectory: planDir,
|
|
maxTurns: 10,
|
|
timeout: 240_000,
|
|
testName: 'plan-eng-review-format-coverage',
|
|
runId,
|
|
model: 'claude-opus-4-7',
|
|
});
|
|
|
|
logCost('/plan-eng-review format (coverage)', result);
|
|
expect(['success', 'error_max_turns']).toContain(result.exitReason);
|
|
|
|
expect(fs.existsSync(outFile)).toBe(true);
|
|
const captured = fs.readFileSync(outFile, 'utf-8');
|
|
expect(captured.length).toBeGreaterThan(100);
|
|
|
|
// Coverage-differentiated: both RECOMMENDATION and Completeness: N/10 required.
|
|
expect(captured).toMatch(RECOMMENDATION_RE);
|
|
expect(captured).toMatch(COMPLETENESS_RE);
|
|
|
|
const recScore = await judgeRecommendation(captured);
|
|
recordE2E(evalCollector, '/plan-eng-review-format-coverage', 'Plan Format — Eng Coverage Issue', result, {
|
|
passed: ['success', 'error_max_turns'].includes(result.exitReason),
|
|
judge_scores: {
|
|
rec_present: recScore.present ? 1 : 0,
|
|
rec_commits: recScore.commits ? 1 : 0,
|
|
rec_has_because: recScore.has_because ? 1 : 0,
|
|
rec_substance: recScore.reason_substance,
|
|
},
|
|
judge_reasoning: `${recScore.reasoning} | reason: "${recScore.reason_text}"`,
|
|
});
|
|
expect(recScore.present, recScore.reasoning).toBe(true);
|
|
expect(recScore.commits, recScore.reasoning).toBe(true);
|
|
expect(recScore.has_because, recScore.reasoning).toBe(true);
|
|
expect(
|
|
recScore.reason_substance,
|
|
`${recScore.reasoning}\n reason: "${recScore.reason_text}"`,
|
|
).toBeGreaterThanOrEqual(4);
|
|
}, 300_000);
|
|
});
|
|
|
|
// --- Case 4: plan-eng-review kind-differentiated per-issue AskUserQuestion ---
|
|
|
|
describeIfSelected('Plan Format — Eng Kind Issue', ['plan-eng-review-format-kind'], () => {
|
|
let planDir: string;
|
|
let outFile: string;
|
|
|
|
beforeAll(() => {
|
|
planDir = setupPlanDir('skill-e2e-plan-format-eng-kind-', 'plan-eng-review');
|
|
outFile = path.join(planDir, 'ask-capture.md');
|
|
});
|
|
|
|
afterAll(() => {
|
|
try { fs.rmSync(planDir, { recursive: true, force: true }); } catch {}
|
|
});
|
|
|
|
testConcurrentIfSelected('plan-eng-review-format-kind', async () => {
|
|
const result = await runSkillTest({
|
|
prompt: `Read plan-eng-review/SKILL.md for the review workflow.
|
|
|
|
Read plan.md — that's the plan to review. This is a standalone plan document, not a codebase — skip any codebase exploration steps.
|
|
|
|
During your review (Section 1 Architecture), generate ONE AskUserQuestion about an architectural choice where the options differ in kind, not in coverage. For example, "should we use Redis or Postgres for the cache layer?" — the options are different kinds of systems with different tradeoffs, not more-or-less-complete versions of the same thing.
|
|
|
|
${captureInstruction(outFile)}
|
|
|
|
After writing the file with that ONE question, stop. Do not continue the review.`,
|
|
workingDirectory: planDir,
|
|
maxTurns: 10,
|
|
timeout: 240_000,
|
|
testName: 'plan-eng-review-format-kind',
|
|
runId,
|
|
model: 'claude-opus-4-7',
|
|
});
|
|
|
|
logCost('/plan-eng-review format (kind)', result);
|
|
expect(['success', 'error_max_turns']).toContain(result.exitReason);
|
|
|
|
expect(fs.existsSync(outFile)).toBe(true);
|
|
const captured = fs.readFileSync(outFile, 'utf-8');
|
|
expect(captured.length).toBeGreaterThan(100);
|
|
|
|
// Kind-differentiated: RECOMMENDATION required, Completeness: N/10 must NOT appear,
|
|
// "options differ in kind" note must appear.
|
|
expect(captured).toMatch(RECOMMENDATION_RE);
|
|
expect(captured).not.toMatch(COMPLETENESS_RE);
|
|
expect(captured).toMatch(KIND_NOTE_RE);
|
|
|
|
const recScore = await judgeRecommendation(captured);
|
|
recordE2E(evalCollector, '/plan-eng-review-format-kind', 'Plan Format — Eng Kind Issue', result, {
|
|
passed: ['success', 'error_max_turns'].includes(result.exitReason),
|
|
judge_scores: {
|
|
rec_present: recScore.present ? 1 : 0,
|
|
rec_commits: recScore.commits ? 1 : 0,
|
|
rec_has_because: recScore.has_because ? 1 : 0,
|
|
rec_substance: recScore.reason_substance,
|
|
},
|
|
judge_reasoning: `${recScore.reasoning} | reason: "${recScore.reason_text}"`,
|
|
});
|
|
expect(recScore.present, recScore.reasoning).toBe(true);
|
|
expect(recScore.commits, recScore.reasoning).toBe(true);
|
|
expect(recScore.has_because, recScore.reasoning).toBe(true);
|
|
expect(
|
|
recScore.reason_substance,
|
|
`${recScore.reasoning}\n reason: "${recScore.reason_text}"`,
|
|
).toBeGreaterThanOrEqual(4);
|
|
}, 300_000);
|
|
});
|
|
|
|
afterAll(async () => {
|
|
await finalizeEvalCollector(evalCollector);
|
|
});
|