File size: 1,279 Bytes
297b174 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
# Genesis AI Code Bench
**Developed by: Within Us AI**
Generated: 2026-01-01
A lightweight evaluation harness for Genesis-style datasets that focuses on the signals
developers care about in practice:
- **Structure validity** (JSON parsing, required fields, schema consistency)
- **Tool-trace validity** (JSON array of tool calls with `tool` + `args`)
- **Diff validity** (`patch_diff` blocks contain recognizable unified-diff markers)
- **Self-grade validity** (score bounds, confidence bounds, presence of notes)
- **Governance presence** (audit/tests flags when expected)
- **Economics presence** (cost budgets + latency targets)
This bench is intentionally fast and offline-friendly. It does not execute repo tests; it
scores dataset quality and readiness for downstream training workflows.
## Quick start
```bash
python bench.py --jsonl path/to/train.jsonl --max_rows 5000
```
## Metrics produced
- `format_valid_rate`
- `required_fields_rate`
- `tool_trace_valid_rate`
- `patch_diff_valid_rate`
- `self_grade_valid_rate`
- `governance_present_rate`
- `economics_present_rate`
- `uniqueness_rate` (hash-based)
## Recommended use
- Run before upload to ensure Viewer-ready consistency
- Run after merges to confirm schema stability
- Compare v1.0 vs v1.1 addon impact
|