File size: 1,279 Bytes
297b174
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
# Genesis AI Code Bench
**Developed by: Within Us AI**  
Generated: 2026-01-01

A lightweight evaluation harness for Genesis-style datasets that focuses on the signals
developers care about in practice:

- **Structure validity** (JSON parsing, required fields, schema consistency)
- **Tool-trace validity** (JSON array of tool calls with `tool` + `args`)
- **Diff validity** (`patch_diff` blocks contain recognizable unified-diff markers)
- **Self-grade validity** (score bounds, confidence bounds, presence of notes)
- **Governance presence** (audit/tests flags when expected)
- **Economics presence** (cost budgets + latency targets)

This bench is intentionally fast and offline-friendly. It does not execute repo tests; it
scores dataset quality and readiness for downstream training workflows.

## Quick start
```bash
python bench.py --jsonl path/to/train.jsonl --max_rows 5000
```

## Metrics produced
- `format_valid_rate`
- `required_fields_rate`
- `tool_trace_valid_rate`
- `patch_diff_valid_rate`
- `self_grade_valid_rate`
- `governance_present_rate`
- `economics_present_rate`
- `uniqueness_rate` (hash-based)

## Recommended use
- Run before upload to ensure Viewer-ready consistency
- Run after merges to confirm schema stability
- Compare v1.0 vs v1.1 addon impact