Open4bits / Granite 4.0 H Tiny FP16
This repository provides the Granite 4.0 H Tiny model converted to FP16 (float16) precision, published by Open4bits for more efficient inference while preserving the original model behavior.
The underlying Granite model architecture and weights are owned by IBM. This repository contains only a precision-converted version of the original model.
The model is suitable for general-purpose text generation tasks and lightweight language model experimentation.
Model Overview
Granite is a family of transformer-based language models developed by IBM, designed for efficient and scalable natural language processing tasks.
This release uses the Granite 4.0 H Tiny variant, focusing on lower memory usage and faster inference while maintaining architectural fidelity to the original model.
Model Details
- Architecture: Granite 4.0 H Tiny
- Precision: float16 (FP16)
- Task: Text Generation
- Weight tying: Preserved
- Compatibility: Hugging Face Transformers, PyTorch
The FP16 conversion reduces VRAM requirements and improves inference speed compared to FP32 versions, making the model suitable for deployment on resource-constrained hardware.
Intended Use
This model is intended for:
- Text generation and completion
- Research and experimentation
- Lightweight inference workloads
- Prototyping and benchmarking language models
Limitations
- Inherits limitations of the Granite Tiny architecture
- Output quality may be lower than larger Granite variants
- Not fine-tuned for domain-specific or instruction-following tasks
License
This model is released under the Apache License 2.0. The original Granite model and associated intellectual property are owned by IBM.
Support
If you find this model useful, please consider supporting the project. Your support helps us continue releasing and maintaining high-quality open models. Support us with a heart.
- Downloads last month
- 14
Model tree for Open4bits/granite-4.0-h-tiny-f16
Base model
ibm-granite/granite-4.0-h-tiny