| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | library_name: transformers |
| | tags: |
| | - code |
| | - python |
| | - maincoder |
| | - code-generation |
| | - gguf |
| | - quantized |
| | pipeline_tag: text-generation |
| | base_model: Maincode/Maincoder-1B |
| | --- |
| | <img src="https://huggingface.co/datasets/Maincode/assets/resolve/e51154e034201be1a5dad0e9c8de31d8b9f17643/maincoder_logo.png" alt="" width="1250"> |
| |
|
| | # Maincoder-1B-GGUF |
| |
|
| | GGUF quantizations of [**Maincoder-1B**](https://huggingface.co/Maincode/Maincoder-1B), a code-focused language model optimized for code generation and completion tasks. These quantized versions are designed for efficient local deployment with [llama.cpp](https://github.com/ggerganov/llama.cpp). |
| |
|
| | Find more details in the original model card: https://huggingface.co/Maincode/Maincoder-1B |
| |
|
| | ## How to run Maincoder |
| |
|
| | Example usage with llama.cpp: |
| |
|
| | ```bash |
| | llama-cli -hf Maincode/Maincoder-1B-GGUF |
| | ``` |
| |
|
| | Or with a specific quantization: |
| |
|
| | ```bash |
| | llama-cli -hf Maincode/Maincoder-1B-GGUF -m Maincoder-1B-Q4_K_M.gguf |
| | ``` |
| |
|
| | Code completion example: |
| |
|
| | ```bash |
| | llama-cli -hf Maincode/Maincoder-1B-GGUF -p 'def fibonacci(n: int) -> int: |
| | """Return the n-th Fibonacci number.""" |
| | ' -n 256 |
| | ``` |
| |
|
| | ## Available Quantizations |
| |
|
| | | Filename | Size | Description | |
| | |----------|------|-------------| |
| | | Maincoder-1B-BF16.gguf | 1.9 GB | BFloat16 - Full precision, best quality | |
| | | Maincoder-1B-F16.gguf | 1.9 GB | Float16 - Full precision | |
| | | Maincoder-1B-Q8_0.gguf | 1.0 GB | 8-bit quantization - Highest quality quantized | |
| | | Maincoder-1B-Q6_K.gguf | 809 MB | 6-bit quantization - High quality | |
| | | Maincoder-1B-Q5_K_M.gguf | 722 MB | 5-bit quantization - Great balance | |
| | | Maincoder-1B-Q4_K_M.gguf | 641 MB | 4-bit quantization - Recommended | |
| | | Maincoder-1B-Q4_0.gguf | 614 MB | 4-bit quantization - Smallest, fastest | |
| | |
| | ## ๐ License |
| | |
| | This model is released under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0). |
| | |
| | ## ๐ Links |
| | |
| | - [Original Model](https://huggingface.co/Maincode/Maincoder-1B) |
| | - [Maincode](https://maincode.com) |
| | - [llama.cpp](https://github.com/ggerganov/llama.cpp) |
| | |
| | |