--- license: mit tags: - bitnet - ternary - trillim - cpu-inference base_model: 1bitLLM/bitnet_b1_58-large --- # BitNet-Large-TRNQ Ternary-quantized version of [1bitLLM/bitnet_b1_58-large](https://huggingface.co/1bitLLM/bitnet_b1_58-large), packaged for the [Trillim DarkNet](https://huggingface.co/Trillim) inference engine. This model runs entirely on CPU — no GPU required. ## Model Details | | | |---|---| | **Architecture** | BitNet (BitnetForCausalLM) | | **Parameters** | ~700M | | **Hidden size** | 1536 | | **Layers** | 24 | | **Attention heads** | 16 | | **Context length** | 2048 | | **Quantization** | Ternary ({-1, 0, 1}) | | **Source model** | [1bitLLM/bitnet_b1_58-large](https://huggingface.co/1bitLLM/bitnet_b1_58-large) | | **License** | MIT | ## Usage ```bash pip install trillim trillim pull Trillim/BitNet-Large-TRNQ trillim serve Trillim/BitNet-Large-TRNQ ``` This starts an OpenAI-compatible API server at `http://127.0.0.1:8000`. For interactive CLI chat: ```bash trillim chat Trillim/BitNet-Large-TRNQ ``` ## What's in this repo | File | Description | |---|---| | `qmodel.tensors` | Ternary-quantized weights in Trillim format | | `rope.cache` | Precomputed RoPE embeddings | | `config.json` | Model configuration | | `tokenizer.json` | Tokenizer | | `tokenizer_config.json` | Tokenizer configuration | | `tokenizer.model` | SentencePiece model | | `tokenization_bitnet.py` | Custom tokenizer class | | `trillim_config.json` | Trillim metadata | ## License This model is released under the [MIT License](https://opensource.org/licenses/MIT), following the license of the source model.