--- {} --- # **Model Summary: Mify-Coder-2.5B** ## **Overview** Mify-Coder-2.5B-v1 is a breakthrough 2.5B-parameter code model fully designed, engineered, and trained at Infosys on 4.2T tokens on Mify-2.5B base model. Despite its compact size, Mify-Coder-2.5B-v1 sets a new benchmark for small language models, achieving performance parity with frontier open-source models in code generation and tool calling, along with exemplary performance on safety metrics in helpfulness and harmlessness, and superior throughput that surpasses larger frontier models. **Developed by**: Infosys Ltd. --- ## **Architecture & Training** - **Base Model:** Mify-2.5B - **Training Phases:** - **Continual Pretraining (CPT):** Next-token prediction with Fill-in-the-Middle (FIM) for structural infilling. - **Supervised Fine-Tuning (SFT):** Instruction alignment for coding tasks, function calling, and safety. - **Optimization:** - **BF16 mixed precision**, **Grouped Query Attention (GQA)**, and **Distributed Fused Adam** optimizer. - Specialized tokenization with syntax markers and reasoning tokens for advanced behaviors. --- ## **Performance Highlights** | **Category** | **Benchmark** | **# Shots** | **Metric** | **Scores** | |----------------|--------------------------------------|-------------|--------------|--------------| | Code Gen | MBPP | 0 | pass@1 | 91.21% | | Code Gen | MBPP+ | 0 | pass@1 | 89.15% | | Code Gen | HumanEval | 0 | pass@1 | 53.66% | | Code Gen | HumanEval+ | 0 | pass@1 | 48.78% | | Code Gen | NumpyEval | 0 | pass@1 | 56.44% | | Code Gen | PandasEval | 0 | pass@1 | 53.47% | | Tool Use | BFCL v2 | 0 | overall acc | 55.26% | | Safety | AIR-Bench | 0 | pass@1 | 67.32% | | SecCode Gen | CybersecEval4-Autocomplete | 0 | pass@1 | 78.91% | --- ## **Responsible AI & Safety** - Integrated safety objectives during SFT. - Balanced harmful/general sample ratio (1:4) for secure code generation and ethical language use. - Validated against **Stanford AIR-Bench** and **CybersecEval4-Autocomplete** benchmarks. --- ## **Deployment & Future Work** - **Quantization:** The model was optimized for low latency outperforming most sub-8B SLM models. Furthermore, the quantized variants of Mify-Coder can be seamlessly deployed and inferenced on standard desktop environments, eliminating the need for specialized hardware such as GPUs. - Future work includes enhancing Mify-Coder with agentic coding competencies and scaling its context length. The model weights will be open-sourced early next year to accelerate research and real-world deployment.