awni commited on
Commit
b51755b
·
verified ·
1 Parent(s): 6c59277

Add files using upload-large-folder tool

Browse files
Files changed (4) hide show
  1. README.md +1 -1
  2. chat_template.jinja +1 -26
  3. config.json +4 -2
  4. tokenizer_config.json +1 -1
README.md CHANGED
@@ -12,7 +12,7 @@ base_model: Qwen/Qwen3-4B-Instruct-2507
12
 
13
  This model [mlx-community/Qwen3-4B-Instruct-2507-5bit](https://huggingface.co/mlx-community/Qwen3-4B-Instruct-2507-5bit) was
14
  converted to MLX format from [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507)
15
- using mlx-lm version **0.26.2**.
16
 
17
  ## Use with mlx
18
 
 
12
 
13
  This model [mlx-community/Qwen3-4B-Instruct-2507-5bit](https://huggingface.co/mlx-community/Qwen3-4B-Instruct-2507-5bit) was
14
  converted to MLX format from [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507)
15
+ using mlx-lm version **0.28.2**.
16
 
17
  ## Use with mlx
18
 
chat_template.jinja CHANGED
@@ -14,14 +14,6 @@
14
  {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
  {%- endif %}
16
  {%- endif %}
17
- {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
18
- {%- for message in messages[::-1] %}
19
- {%- set index = (messages|length - 1) - loop.index0 %}
20
- {%- if ns.multi_step_tool and message.role == "user" and message.content is string and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}
21
- {%- set ns.multi_step_tool = false %}
22
- {%- set ns.last_query_index = index %}
23
- {%- endif %}
24
- {%- endfor %}
25
  {%- for message in messages %}
26
  {%- if message.content is string %}
27
  {%- set content = message.content %}
@@ -31,24 +23,7 @@
31
  {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
32
  {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
33
  {%- elif message.role == "assistant" %}
34
- {%- set reasoning_content = '' %}
35
- {%- if message.reasoning_content is string %}
36
- {%- set reasoning_content = message.reasoning_content %}
37
- {%- else %}
38
- {%- if '</think>' in content %}
39
- {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
40
- {%- set content = content.split('</think>')[-1].lstrip('\n') %}
41
- {%- endif %}
42
- {%- endif %}
43
- {%- if loop.index0 > ns.last_query_index %}
44
- {%- if loop.last or (not loop.last and reasoning_content) %}
45
- {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content.strip('\n') + '\n</think>\n\n' + content.lstrip('\n') }}
46
- {%- else %}
47
- {{- '<|im_start|>' + message.role + '\n' + content }}
48
- {%- endif %}
49
- {%- else %}
50
- {{- '<|im_start|>' + message.role + '\n' + content }}
51
- {%- endif %}
52
  {%- if message.tool_calls %}
53
  {%- for tool_call in message.tool_calls %}
54
  {%- if (loop.first and content) or (not loop.first) %}
 
14
  {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
  {%- endif %}
16
  {%- endif %}
 
 
 
 
 
 
 
 
17
  {%- for message in messages %}
18
  {%- if message.content is string %}
19
  {%- set content = message.content %}
 
23
  {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
  {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
  {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27
  {%- if message.tool_calls %}
28
  {%- for tool_call in message.tool_calls %}
29
  {%- if (loop.first and content) or (not loop.first) %}
config.json CHANGED
@@ -19,11 +19,13 @@
19
  "num_key_value_heads": 8,
20
  "quantization": {
21
  "group_size": 64,
22
- "bits": 5
 
23
  },
24
  "quantization_config": {
25
  "group_size": 64,
26
- "bits": 5
 
27
  },
28
  "rms_norm_eps": 1e-06,
29
  "rope_scaling": null,
 
19
  "num_key_value_heads": 8,
20
  "quantization": {
21
  "group_size": 64,
22
+ "bits": 5,
23
+ "mode": "affine"
24
  },
25
  "quantization_config": {
26
  "group_size": 64,
27
+ "bits": 5,
28
+ "mode": "affine"
29
  },
30
  "rms_norm_eps": 1e-06,
31
  "rope_scaling": null,
tokenizer_config.json CHANGED
@@ -231,7 +231,7 @@
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
- "model_max_length": 262144,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",
 
231
  "eos_token": "<|im_end|>",
232
  "errors": "replace",
233
  "extra_special_tokens": {},
234
+ "model_max_length": 1010000,
235
  "pad_token": "<|endoftext|>",
236
  "split_special_tokens": false,
237
  "tokenizer_class": "Qwen2Tokenizer",