Add files using upload-large-folder tool
Browse files- README.md +1 -1
- chat_template.jinja +1 -26
- config.json +4 -2
- tokenizer_config.json +1 -1
README.md
CHANGED
|
@@ -12,7 +12,7 @@ base_model: Qwen/Qwen3-4B-Instruct-2507
|
|
| 12 |
|
| 13 |
This model [mlx-community/Qwen3-4B-Instruct-2507-5bit](https://huggingface.co/mlx-community/Qwen3-4B-Instruct-2507-5bit) was
|
| 14 |
converted to MLX format from [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507)
|
| 15 |
-
using mlx-lm version **0.
|
| 16 |
|
| 17 |
## Use with mlx
|
| 18 |
|
|
|
|
| 12 |
|
| 13 |
This model [mlx-community/Qwen3-4B-Instruct-2507-5bit](https://huggingface.co/mlx-community/Qwen3-4B-Instruct-2507-5bit) was
|
| 14 |
converted to MLX format from [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507)
|
| 15 |
+
using mlx-lm version **0.28.2**.
|
| 16 |
|
| 17 |
## Use with mlx
|
| 18 |
|
chat_template.jinja
CHANGED
|
@@ -14,14 +14,6 @@
|
|
| 14 |
{{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
|
| 15 |
{%- endif %}
|
| 16 |
{%- endif %}
|
| 17 |
-
{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
|
| 18 |
-
{%- for message in messages[::-1] %}
|
| 19 |
-
{%- set index = (messages|length - 1) - loop.index0 %}
|
| 20 |
-
{%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
|
| 21 |
-
{%- set ns.multi_step_tool = false %}
|
| 22 |
-
{%- set ns.last_query_index = index %}
|
| 23 |
-
{%- endif %}
|
| 24 |
-
{%- endfor %}
|
| 25 |
{%- for message in messages %}
|
| 26 |
{%- if message.content is string %}
|
| 27 |
{%- set content = message.content %}
|
|
@@ -31,24 +23,7 @@
|
|
| 31 |
{%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
|
| 32 |
{{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
|
| 33 |
{%- elif message.role == "assistant" %}
|
| 34 |
-
{
|
| 35 |
-
{%- if message.reasoning_content is string %}
|
| 36 |
-
{%- set reasoning_content = message.reasoning_content %}
|
| 37 |
-
{%- else %}
|
| 38 |
-
{%- if '</think>' in content %}
|
| 39 |
-
{%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
|
| 40 |
-
{%- set content = content.split('</think>')[-1].lstrip('\n') %}
|
| 41 |
-
{%- endif %}
|
| 42 |
-
{%- endif %}
|
| 43 |
-
{%- if loop.index0 > ns.last_query_index %}
|
| 44 |
-
{%- if loop.last or (not loop.last and reasoning_content) %}
|
| 45 |
-
{{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
|
| 46 |
-
{%- else %}
|
| 47 |
-
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 48 |
-
{%- endif %}
|
| 49 |
-
{%- else %}
|
| 50 |
-
{{- '<|im_start|>' + message.role + '\n' + content }}
|
| 51 |
-
{%- endif %}
|
| 52 |
{%- if message.tool_calls %}
|
| 53 |
{%- for tool_call in message.tool_calls %}
|
| 54 |
{%- if (loop.first and content) or (not loop.first) %}
|
|
|
|
| 14 |
{{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
|
| 15 |
{%- endif %}
|
| 16 |
{%- endif %}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
{%- for message in messages %}
|
| 18 |
{%- if message.content is string %}
|
| 19 |
{%- set content = message.content %}
|
|
|
|
| 23 |
{%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
|
| 24 |
{{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
|
| 25 |
{%- elif message.role == "assistant" %}
|
| 26 |
+
{{- '<|im_start|>' + message.role + '\n' + content }}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
{%- if message.tool_calls %}
|
| 28 |
{%- for tool_call in message.tool_calls %}
|
| 29 |
{%- if (loop.first and content) or (not loop.first) %}
|
config.json
CHANGED
|
@@ -19,11 +19,13 @@
|
|
| 19 |
"num_key_value_heads": 8,
|
| 20 |
"quantization": {
|
| 21 |
"group_size": 64,
|
| 22 |
-
"bits": 5
|
|
|
|
| 23 |
},
|
| 24 |
"quantization_config": {
|
| 25 |
"group_size": 64,
|
| 26 |
-
"bits": 5
|
|
|
|
| 27 |
},
|
| 28 |
"rms_norm_eps": 1e-06,
|
| 29 |
"rope_scaling": null,
|
|
|
|
| 19 |
"num_key_value_heads": 8,
|
| 20 |
"quantization": {
|
| 21 |
"group_size": 64,
|
| 22 |
+
"bits": 5,
|
| 23 |
+
"mode": "affine"
|
| 24 |
},
|
| 25 |
"quantization_config": {
|
| 26 |
"group_size": 64,
|
| 27 |
+
"bits": 5,
|
| 28 |
+
"mode": "affine"
|
| 29 |
},
|
| 30 |
"rms_norm_eps": 1e-06,
|
| 31 |
"rope_scaling": null,
|
tokenizer_config.json
CHANGED
|
@@ -231,7 +231,7 @@
|
|
| 231 |
"eos_token": "<|im_end|>",
|
| 232 |
"errors": "replace",
|
| 233 |
"extra_special_tokens": {},
|
| 234 |
-
"model_max_length":
|
| 235 |
"pad_token": "<|endoftext|>",
|
| 236 |
"split_special_tokens": false,
|
| 237 |
"tokenizer_class": "Qwen2Tokenizer",
|
|
|
|
| 231 |
"eos_token": "<|im_end|>",
|
| 232 |
"errors": "replace",
|
| 233 |
"extra_special_tokens": {},
|
| 234 |
+
"model_max_length": 1010000,
|
| 235 |
"pad_token": "<|endoftext|>",
|
| 236 |
"split_special_tokens": false,
|
| 237 |
"tokenizer_class": "Qwen2Tokenizer",
|