nm-autobot commited on
Commit
ab0a2b7
·
verified ·
1 Parent(s): 7551e2b

Upload folder using huggingface_hub

Browse files
Files changed (5) hide show
  1. config.json +5 -3
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
  4. recipe.yaml +2 -0
  5. tokenizer.model +3 -0
config.json CHANGED
@@ -32,21 +32,23 @@
32
  "num_bits": 8,
33
  "observer": "minmax",
34
  "observer_kwargs": {},
 
35
  "strategy": "tensor",
36
  "symmetric": true,
37
- "type": "float"
 
38
  },
39
  "quant_method": "compressed-tensors",
40
  "quantization_status": "frozen",
41
  "sparsity_config": {},
42
  "transform_config": {},
43
- "version": "0.12.3.a20251030"
44
  },
45
  "rms_norm_eps": 1e-05,
46
  "rope_scaling": null,
47
  "rope_theta": 10000.0,
48
  "tie_word_embeddings": false,
49
- "transformers_version": "4.57.1",
50
  "use_cache": true,
51
  "vocab_size": 32000
52
  }
 
32
  "num_bits": 8,
33
  "observer": "minmax",
34
  "observer_kwargs": {},
35
+ "scale_dtype": null,
36
  "strategy": "tensor",
37
  "symmetric": true,
38
+ "type": "float",
39
+ "zp_dtype": null
40
  },
41
  "quant_method": "compressed-tensors",
42
  "quantization_status": "frozen",
43
  "sparsity_config": {},
44
  "transform_config": {},
45
+ "version": "0.12.3.a20251114"
46
  },
47
  "rms_norm_eps": 1e-05,
48
  "rope_scaling": null,
49
  "rope_theta": 10000.0,
50
  "tie_word_embeddings": false,
51
+ "transformers_version": "4.57.3",
52
  "use_cache": true,
53
  "vocab_size": 32000
54
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "eos_token_id": 2,
4
  "max_length": 2048,
5
  "pad_token_id": 0,
6
- "transformers_version": "4.57.1"
7
  }
 
3
  "eos_token_id": 2,
4
  "max_length": 2048,
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.57.3"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:691f114803b2ad769d4e1f32f11970fca2ecd878c5600566f1e6f991f6fac386
3
  size 2200124472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63c57c6f80f7b48b8106b8b34ca4b4c036503156def11fd5f159453ca0932f68
3
  size 2200124472
recipe.yaml CHANGED
@@ -12,5 +12,7 @@ quant_stage:
12
  block_structure: null
13
  dynamic: false
14
  actorder: null
 
 
15
  observer: minmax
16
  observer_kwargs: {}
 
12
  block_structure: null
13
  dynamic: false
14
  actorder: null
15
+ scale_dtype: null
16
+ zp_dtype: null
17
  observer: minmax
18
  observer_kwargs: {}
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723