# Genesis AI Code Bench **Developed by: Within Us AI** Generated: 2026-01-01 A lightweight evaluation harness for Genesis-style datasets that focuses on the signals developers care about in practice: - **Structure validity** (JSON parsing, required fields, schema consistency) - **Tool-trace validity** (JSON array of tool calls with `tool` + `args`) - **Diff validity** (`patch_diff` blocks contain recognizable unified-diff markers) - **Self-grade validity** (score bounds, confidence bounds, presence of notes) - **Governance presence** (audit/tests flags when expected) - **Economics presence** (cost budgets + latency targets) This bench is intentionally fast and offline-friendly. It does not execute repo tests; it scores dataset quality and readiness for downstream training workflows. ## Quick start ```bash python bench.py --jsonl path/to/train.jsonl --max_rows 5000 ``` ## Metrics produced - `format_valid_rate` - `required_fields_rate` - `tool_trace_valid_rate` - `patch_diff_valid_rate` - `self_grade_valid_rate` - `governance_present_rate` - `economics_present_rate` - `uniqueness_rate` (hash-based) ## Recommended use - Run before upload to ensure Viewer-ready consistency - Run after merges to confirm schema stability - Compare v1.0 vs v1.1 addon impact