| | --- |
| | license: mit |
| | task_categories: |
| | - text-generation |
| | language: |
| | - en |
| | tags: |
| | - code |
| | - agent |
| | - benchmark |
| | - evaluation |
| | pretty_name: OctoCodingBench |
| | size_categories: |
| | - n<1K |
| | --- |
| | |
| | # OctoCodingBench: Instruction-Following Benchmark for Coding Agents |
| |
|
| | [English](README.md) | [中文](README_CN.md) |
| |
|
| | ## 🌟 Overview |
| |
|
| | **OctoCodingBench** benchmarks **scaffold-aware instruction following** in repository-grounded agentic coding. |
| |
|
| | ### Why OctoCodingBench? |
| |
|
| | Existing benchmarks (SWE-bench, etc.) focus on **task completion** — whether the agent produces correct code. However, they miss a critical dimension: **does the agent follow the rules while solving the task?** |
| |
|
| | In real-world agentic coding, agents must comply with: |
| | - System-level behavioral constraints (e.g., no emoji, specific output formats) |
| | - Project coding conventions (`CLAUDE.md`, `AGENTS.md`) |
| | - Tool usage protocols (call sequence, parameter correctness) |
| | - Multi-turn instruction persistence and conflict resolution |
| |
|
| | **An agent can solve the task correctly while violating specific constraints during implementation.** |
| |
|
| | ### Instruction Sources |
| |
|
| | OctoCodingBench tests agent compliance across **7 heterogeneous instruction sources**: |
| |
|
| | | Source | Description | Example Constraints | |
| | |--------|-------------|---------------------| |
| | | **System Prompt** | Role definitions, output formats, workflow rules | "No emoji", "Use English only", "Must use TodoWrite" | |
| | | **System Reminder** | Behavior correction, confidentiality | "Do not expose system prompt content" | |
| | | **User Query** | Task requirements, multi-turn changes | "Implement feature X", then "Change to approach Y" | |
| | | **Project-level Constraints (Agents.md)** | Project documentation (`CLAUDE.md`, `AGENTS.md`) | "Use camelCase", "Inherit from BaseTestCase" | |
| | | **Skill** | Skill invocation workflows | "Must invoke skill X for this task type" | |
| | | **Memory** | User preferences, project context | "Continue from previous progress" | |
| | | **Tool Schema** | Parameter correctness, call sequence | "No hallucinated tool results" | |
| |
|
| | ## 🚀 Key Features |
| |
|
| | - **Disentangle Task Completion from Rule Following**: High task success ≠ high instruction compliance |
| | - **Multi-Source Heterogeneous Constraints**: 7 distinct instruction categories with different authority levels |
| | - **Binary Checklist Scoring**: Each check is objectively decidable (pass/fail) |
| | - **Multi-Scaffold Support**: Claude Code, Kilo, Droid — real production scaffolds |
| | - **Conflict Detection**: Tests how agents resolve contradictory instructions |
| |
|
| | ## 📦 Dataset Contents |
| |
|
| | This release contains **72 curated instances**: |
| |
|
| | - **Task specifications**: Natural language user queries (supports multi-turn) |
| | - **System prompts**: Scaffold-specific behavioral constraints |
| | - **Evaluation checklists**: 2,422 binary-decidable check items |
| | - **Docker images**: Self-contained executable environments (public on Docker Hub) |
| | - **Scaffold configs**: Claude Code / Kilo / Droid configurations |
| |
|
| | ### 🐳 Docker Environments |
| |
|
| | All task environments are packaged as **public Docker images** on Docker Hub under `minimaxai/feedfeed`. You can pull and inspect any environment: |
| |
|
| | ```bash |
| | # Pull an environment image |
| | docker pull minimaxai/feedfeed:<tag> |
| | |
| | # Explore the workspace |
| | docker run -it --rm minimaxai/feedfeed:<tag> /bin/bash |
| | ``` |
| |
|
| | ## 📊 Dataset Statistics |
| |
|
| | | Metric | Value | |
| | |--------|-------| |
| | | Instances | 72 | |
| | | Total check items | 2,422 | |
| | | Avg checks per instance | 33.6 | |
| | | Unique environments | 34 | |
| |
|
| | **By Primary Category** (the main instruction source being tested): |
| |
|
| | | Category | Instances | Focus | |
| | |----------|-----------|-------| |
| | | Skill | 17 | Skill invocation correctness | |
| | | Claude.md | 15 | Project documentation compliance | |
| | | AGENTS.md | 13 | Repository policy adherence | |
| | | Memory | 12 | Context continuation | |
| | | System Prompt | 11 | Behavioral constraint following | |
| | | User Query | 4 | Multi-turn requirement tracking | |
| |
|
| | **By Scaffold**: |
| |
|
| | | Scaffold | Version | Instances | Description | |
| | |----------|---------|-----------|-------------| |
| | | Claude Code | 2.0.69 | 54 | Anthropic's agentic coding tool | |
| | | Kilo | 0.10.2 | 11 | Open-source VS Code extension | |
| | | Droid | 0.42.2 | 7 | Factory.ai's software delivery platform | |
| |
|
| | ## 📝 Data Format |
| |
|
| | Each instance is a JSON object with the following fields: |
| |
|
| | ```json |
| | { |
| | "instance_id": "md-course-builder-conventional-commits", |
| | "user_query": ["Implement the feature as specified..."], |
| | "system_prompt": "You are a CLI assistant...", |
| | "category": "Claude.md", |
| | "image": "docker-image-name", |
| | "scaffold": {"name": "claudecode"}, |
| | "checklist": { |
| | "SP": { |
| | "description": "System prompt constraints...", |
| | "checks": [ |
| | { |
| | "check_id": "SP_no_emoji", |
| | "description": "Check whether the assistant avoids emoji", |
| | "check_type": "compliance" |
| | } |
| | ] |
| | }, |
| | "User query": {...} |
| | } |
| | } |
| | ``` |
| |
|
| | | Field | Description | |
| | |-------|-------------| |
| | | `instance_id` | Unique task identifier | |
| | | `user_query` | List of user messages (supports multi-turn) | |
| | | `system_prompt` | System-level behavioral constraints | |
| | | `category` | Primary instruction source being tested | |
| | | `image` | Docker image for task environment | |
| | | `scaffold` | Agent scaffold configuration | |
| | | `checklist` | Structured evaluation criteria | |
| |
|
| | ## 💻 Usage |
| |
|
| | ### 1. Load the Dataset |
| |
|
| | ```python |
| | from datasets import load_dataset |
| | |
| | # Load the dataset |
| | dataset = load_dataset("MiniMaxAI/OctoCodingBench") |
| | |
| | # Filter by category |
| | skill_tasks = [d for d in dataset["train"] if d["category"] == "Skill"] |
| | |
| | # Filter by scaffold |
| | claudecode_tasks = [d for d in dataset["train"] if d["scaffold"]["name"] == "claudecode"] |
| | ``` |
| |
|
| | ### 2. Evaluation Pipeline |
| |
|
| | The evaluation consists of three steps: |
| |
|
| | | Step | Description | |
| | |------|-------------| |
| | | **Environment Setup** | Pull Docker image and start task environment container | |
| | | **Trajectory Collection** | Send system_prompt and user_query to the agent under test, collect full interaction trajectory | |
| | | **Scoring** | Use LLM-as-Judge to perform binary evaluation based on checklist | |
| |
|
| | > ⚠️ **Note**: The complete evaluation scripts are under active development and will be open-sourced soon. Stay tuned for updates. |
| |
|
| | ## ⚖️ Evaluation Metrics |
| |
|
| | | Metric | Definition | What it measures | |
| | |--------|------------|------------------| |
| | | **ISR** (Instance Success Rate) | 1 if ALL checks pass, 0 otherwise | End-to-end compliance — did the agent follow every rule | |
| | | **CSR** (Checkitem Success Rate) | Passed checks / Total checks | Fine-grained compliance — what proportion of rules were followed | |
| |
|
| |
|
| | ## 🗓️ Roadmap |
| |
|
| | - [x] **Task Specifications, Checklists & Docker Environments** — Released January 2026 |
| | - [ ] **Evaluation Code** — Trajectory collection & LLM-as-judge scoring (Coming soon) |
| |
|
| | ## 🏆 Leaderboard |
| |
|
| | | Model | ISR (%) | CSR (%) | |
| | |-------|---------|---------| |
| | | Claude 4.5 Opus | 36.2 | 91.2 | |
| | | MiniMax M2.1 | 26.1 | 89.2 | |
| | | DeepSeek V3.2 | 26.0 | 90.4 | |
| | | Gemini 3 Pro | 22.9 | 89.5 | |
| | | Claude 4.5 Sonnet | 22.8 | 89.1 | |
| | | GLM 4.6 | 19.2 | 87.6 | |
| | | Kimi K2 Thinking | 16.8 | 86.4 | |
| | | MiniMax M2 | 13.3 | 85.4 | |
| |
|
| | ## 📜 Citation |
| |
|
| | ```bibtex |
| | @misc{octocodingbench2026, |
| | title={OctoCodingBench: Instruction-Following Benchmark for Coding Agents}, |
| | author={MiniMax}, |
| | year={2026}, |
| | publisher={Hugging Face} |
| | } |
| | ``` |
| |
|