Granite-4.0-Micro
Run Granite-4.0-Micro optimized for Apple NPU
Quickstart
Install NexaSDK
After you install the SDK, then run the model with one line of code:
nexa infer NexaAI/Granite-4-Micro-ANE
Model Description
Granite-4.0-Micro is a 3B parameter instruction-tuned model in the Granite 4.0 family, developed by IBM.
It’s optimized for long-context reasoning (128K tokens), efficient inference, and enterprise-ready capabilities such as tool calling and retrieval-augmented generation. The model balances compact size with strong performance across general NLP tasks, making it suitable for both experimentation and production workloads.
Features
- Compact transformer architecture: 3B parameters with GQA, RoPE, SwiGLU, and RMSNorm layers.
- Instruction-following & tool calling: Tuned with supervised finetuning, alignment (RLHF), and model merging for robust enterprise tasks.
- Multilingual support: Covers 12+ languages including English, German, Spanish, French, Japanese, Korean, Arabic, and Chinese.
- Extended context window: Supports sequences up to 128K tokens for long-form reasoning.
Use Cases
- Conversational AI and virtual assistants.
- Enterprise applications needing tool/API calling and structured outputs.
- Long-document summarization, classification, and extraction.
- Retrieval-augmented generation (RAG) for knowledge-intensive workflows.
- Lightweight coding assistants and multilingual dialog systems.
Inputs and Outputs
Input: Natural language text prompts, chat conversations, or tool-augmented requests.
Output: Natural language responses—answers, explanations, summaries, structured JSON for function calls, or code snippets.
License
This model is released under the Creative Commons Attribution–NonCommercial 4.0 (CC BY-NC 4.0) license.
Non-commercial use, modification, and redistribution are permitted with attribution.
For commercial licensing, please contact dev@nexa.ai.
- Downloads last month
- 31