| # Genesis AI Code Bench |
| **Developed by: Within Us AI** |
| Generated: 2026-01-01 |
|
|
| A lightweight evaluation harness for Genesis-style datasets that focuses on the signals |
| developers care about in practice: |
|
|
| - **Structure validity** (JSON parsing, required fields, schema consistency) |
| - **Tool-trace validity** (JSON array of tool calls with `tool` + `args`) |
| - **Diff validity** (`patch_diff` blocks contain recognizable unified-diff markers) |
| - **Self-grade validity** (score bounds, confidence bounds, presence of notes) |
| - **Governance presence** (audit/tests flags when expected) |
| - **Economics presence** (cost budgets + latency targets) |
|
|
| This bench is intentionally fast and offline-friendly. It does not execute repo tests; it |
| scores dataset quality and readiness for downstream training workflows. |
|
|
| ## Quick start |
| ```bash |
| python bench.py --jsonl path/to/train.jsonl --max_rows 5000 |
| ``` |
|
|
| ## Metrics produced |
| - `format_valid_rate` |
| - `required_fields_rate` |
| - `tool_trace_valid_rate` |
| - `patch_diff_valid_rate` |
| - `self_grade_valid_rate` |
| - `governance_present_rate` |
| - `economics_present_rate` |
| - `uniqueness_rate` (hash-based) |
|
|
| ## Recommended use |
| - Run before upload to ensure Viewer-ready consistency |
| - Run after merges to confirm schema stability |
| - Compare v1.0 vs v1.1 addon impact |
|
|