p2o6e100 commited on
Commit
70de3ab
·
1 Parent(s): 71b008d

init commit

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 106152,
3
+ "<|im_end|>": 106154,
4
+ "<|im_start|>": 106153
5
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
mlc-chat-config.json ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "decomp_qwen2",
4
+ "quantization": "q0f16",
5
+ "model_config": {
6
+ "hidden_act": "silu",
7
+ "hidden_size": 2688,
8
+ "intermediate_size": 18944,
9
+ "num_attention_heads": 21,
10
+ "num_hidden_layers": 28,
11
+ "num_key_value_heads": 3,
12
+ "rms_norm_eps": 1e-06,
13
+ "rope_theta": 1000000.0,
14
+ "vocab_size": 106155,
15
+ "tie_word_embeddings": false,
16
+ "context_window_size": 1024,
17
+ "prefill_chunk_size": 1024,
18
+ "tensor_parallel_shards": 1,
19
+ "truncation_ranks": {
20
+ "model.layers.0.mlp.down_proj": 2109,
21
+ "model.layers.0.mlp.gate_up_proj": 819,
22
+ "model.layers.0.self_attn.o_proj": 845,
23
+ "model.layers.0.self_attn.qkv_proj": 1495,
24
+ "model.layers.1.mlp.down_proj": 1657,
25
+ "model.layers.1.mlp.gate_up_proj": 1310,
26
+ "model.layers.1.self_attn.o_proj": 154,
27
+ "model.layers.1.self_attn.qkv_proj": 176,
28
+ "model.layers.10.self_attn.o_proj": 615,
29
+ "model.layers.10.self_attn.qkv_proj": 703,
30
+ "model.layers.11.mlp.down_proj": 286,
31
+ "model.layers.11.mlp.gate_up_proj": 318,
32
+ "model.layers.11.self_attn.o_proj": 384,
33
+ "model.layers.11.self_attn.qkv_proj": 439,
34
+ "model.layers.12.mlp.down_proj": 1574,
35
+ "model.layers.12.mlp.gate_up_proj": 2705,
36
+ "model.layers.12.self_attn.o_proj": 154,
37
+ "model.layers.12.self_attn.qkv_proj": 176,
38
+ "model.layers.13.mlp.down_proj": 716,
39
+ "model.layers.13.mlp.gate_up_proj": 1273,
40
+ "model.layers.13.self_attn.o_proj": 615,
41
+ "model.layers.13.self_attn.qkv_proj": 703,
42
+ "model.layers.14.mlp.down_proj": 716,
43
+ "model.layers.14.mlp.gate_up_proj": 1751,
44
+ "model.layers.14.self_attn.o_proj": 384,
45
+ "model.layers.14.self_attn.qkv_proj": 439,
46
+ "model.layers.15.mlp.down_proj": 716,
47
+ "model.layers.15.mlp.gate_up_proj": 1273,
48
+ "model.layers.15.self_attn.o_proj": 154,
49
+ "model.layers.15.self_attn.qkv_proj": 176,
50
+ "model.layers.16.mlp.down_proj": 716,
51
+ "model.layers.16.mlp.gate_up_proj": 1273,
52
+ "model.layers.16.self_attn.o_proj": 384,
53
+ "model.layers.16.self_attn.qkv_proj": 439,
54
+ "model.layers.17.mlp.down_proj": 1145,
55
+ "model.layers.17.mlp.gate_up_proj": 1751,
56
+ "model.layers.17.self_attn.o_proj": 384,
57
+ "model.layers.17.self_attn.qkv_proj": 439,
58
+ "model.layers.18.mlp.down_proj": 1145,
59
+ "model.layers.18.mlp.gate_up_proj": 1273,
60
+ "model.layers.18.self_attn.o_proj": 384,
61
+ "model.layers.18.self_attn.qkv_proj": 439,
62
+ "model.layers.19.mlp.gate_up_proj": 1751,
63
+ "model.layers.19.self_attn.o_proj": 384,
64
+ "model.layers.19.self_attn.qkv_proj": 703,
65
+ "model.layers.2.mlp.down_proj": 1657,
66
+ "model.layers.2.mlp.gate_up_proj": 1801,
67
+ "model.layers.2.self_attn.o_proj": 154,
68
+ "model.layers.2.self_attn.qkv_proj": 176,
69
+ "model.layers.20.self_attn.o_proj": 384,
70
+ "model.layers.20.self_attn.qkv_proj": 703,
71
+ "model.layers.21.self_attn.o_proj": 615,
72
+ "model.layers.21.self_attn.qkv_proj": 703,
73
+ "model.layers.22.self_attn.o_proj": 615,
74
+ "model.layers.22.self_attn.qkv_proj": 967,
75
+ "model.layers.23.self_attn.o_proj": 615,
76
+ "model.layers.23.self_attn.qkv_proj": 703,
77
+ "model.layers.24.self_attn.o_proj": 384,
78
+ "model.layers.24.self_attn.qkv_proj": 439,
79
+ "model.layers.25.self_attn.o_proj": 154,
80
+ "model.layers.25.self_attn.qkv_proj": 439,
81
+ "model.layers.26.self_attn.o_proj": 384,
82
+ "model.layers.26.self_attn.qkv_proj": 439,
83
+ "model.layers.27.mlp.gate_up_proj": 2783,
84
+ "model.layers.27.self_attn.o_proj": 384,
85
+ "model.layers.27.self_attn.qkv_proj": 176,
86
+ "model.layers.3.mlp.down_proj": 1205,
87
+ "model.layers.3.mlp.gate_up_proj": 1310,
88
+ "model.layers.3.self_attn.o_proj": 615,
89
+ "model.layers.3.self_attn.qkv_proj": 439,
90
+ "model.layers.4.mlp.down_proj": 1145,
91
+ "model.layers.4.mlp.gate_up_proj": 1273,
92
+ "model.layers.4.self_attn.o_proj": 384,
93
+ "model.layers.4.self_attn.qkv_proj": 439,
94
+ "model.layers.5.mlp.down_proj": 286,
95
+ "model.layers.5.mlp.gate_up_proj": 318,
96
+ "model.layers.5.self_attn.o_proj": 154,
97
+ "model.layers.5.self_attn.qkv_proj": 439,
98
+ "model.layers.6.mlp.down_proj": 1574,
99
+ "model.layers.6.self_attn.o_proj": 154,
100
+ "model.layers.6.self_attn.qkv_proj": 439,
101
+ "model.layers.7.self_attn.o_proj": 615,
102
+ "model.layers.7.self_attn.qkv_proj": 703,
103
+ "model.layers.8.mlp.down_proj": 1574,
104
+ "model.layers.8.mlp.gate_up_proj": 1751,
105
+ "model.layers.8.self_attn.o_proj": 615,
106
+ "model.layers.8.self_attn.qkv_proj": 703,
107
+ "model.layers.9.mlp.gate_up_proj": 2705,
108
+ "model.layers.9.self_attn.o_proj": 384,
109
+ "model.layers.9.self_attn.qkv_proj": 703
110
+ },
111
+ "pruned_layers": [
112
+ 4,
113
+ 5,
114
+ 6,
115
+ 7,
116
+ 8,
117
+ 9,
118
+ 10,
119
+ 11,
120
+ 12,
121
+ 13,
122
+ 14,
123
+ 15,
124
+ 16,
125
+ 17,
126
+ 18,
127
+ 19,
128
+ 20,
129
+ 21,
130
+ 22,
131
+ 23,
132
+ 24
133
+ ],
134
+ "pruned_intermediate_size": 14208,
135
+ "ori_hidden_size": 3584,
136
+ "head_dim": 128,
137
+ "dtype": "float32",
138
+ "max_batch_size": 1
139
+ },
140
+ "vocab_size": 106155,
141
+ "context_window_size": 1024,
142
+ "sliding_window_size": -1,
143
+ "prefill_chunk_size": 1024,
144
+ "attention_sink_size": -1,
145
+ "tensor_parallel_shards": 1,
146
+ "pipeline_parallel_stages": 1,
147
+ "temperature": 0.7,
148
+ "presence_penalty": 0.0,
149
+ "frequency_penalty": 0.0,
150
+ "repetition_penalty": 1.05,
151
+ "top_p": 0.8,
152
+ "tokenizer_files": [
153
+ "tokenizer.json",
154
+ "vocab.json",
155
+ "merges.txt",
156
+ "added_tokens.json",
157
+ "tokenizer_config.json"
158
+ ],
159
+ "tokenizer_info": {
160
+ "token_postproc_method": "byte_level",
161
+ "prepend_space_in_encode": false,
162
+ "strip_space_in_decode": false
163
+ },
164
+ "conv_template": {
165
+ "name": "qwen2",
166
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
167
+ "system_message": "You are a helpful assistant.",
168
+ "system_prefix_token_ids": [106153],
169
+ "add_role_after_system_message": true,
170
+ "roles": {
171
+ "user": "<|im_start|>user",
172
+ "assistant": "<|im_start|>assistant"
173
+ },
174
+ "role_templates": {
175
+ "user": "{user_message}",
176
+ "assistant": "{assistant_message}",
177
+ "tool": "{tool_message}"
178
+ },
179
+ "messages": [],
180
+ "seps": [
181
+ "<|im_end|>\n"
182
+ ],
183
+ "role_content_sep": "\n",
184
+ "role_empty_sep": "\n",
185
+ "stop_str": [],
186
+ "stop_token_ids": [
187
+ 106154,
188
+ 106152
189
+ ],
190
+ "function_string": "",
191
+ "use_function_calling": false
192
+ },
193
+ "pad_token_id": 106152,
194
+ "bos_token_id": 106153,
195
+ "eos_token_id": [
196
+ 106154,
197
+ 106152
198
+ ]
199
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:974f845feeac18abbb988389f291c34cdef79bb0d3b03cc6e69a7df10cde8687
3
+ size 760919040
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11774e574324aa604e2307e4772b3dd85402fd438d6115debd33392165d00dbc
3
+ size 101842944
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b07ded152c12d2d7a9a2055cc838270f89b399c22020943bfdaa03e04091ce
3
+ size 271581184
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea914152bdba2c2cf0e0c24352cec0c1846a3a5f53618b58d6cd831fccecde3b
3
+ size 135790592
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95207e09f9d925c0ef5f2c57e3c252afa58c532de9c1ee50361bf3377adce750
3
+ size 271581184
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2081268f061413f66df08885e1d62bc20fed7a382612b54d49acc107a863ee17
3
+ size 31818496
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39a4ed2eff9f862dc367c69f69e02202bdf2fa8b752d14daa3d1a07eb9e344cb
3
+ size 135790592
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70090d8135b4c19ce1b01713c72a90aa5bcb83d7b97be9a23348352003f9ef86
3
+ size 210884608
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4527563ab9c9dee0710fc9b67ed6d6d3f2f0159ad3393ce0bdfe8bf8f48a7aad
3
+ size 33039104
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aac43a7378e5da200bdca8993a06797a06436affe3079fc9b2861aa1341cd658
3
+ size 760919040
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de837d0e39e700685dd9efa305167ea0d6ecffe745e2b3b3055618b565a7fd6a
3
+ size 79905792
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaaafc8c1a6bca3c3a02a1727ddec3128de3e42b378cd0bea99d6b29a78e088c
3
+ size 62060544
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b203233cb1d5f8efc2616d67e9cf19d1d8c549d8c575bfc54a6eb6688ddfe541
3
+ size 101842944
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1a7194e447a4fa8ed758806376ffbf0a83ac8cf00e23801ef042041b6fdbdb3
3
+ size 29544448
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56747d4dd56583939f5fcd6c763e594217e658038b5e846ddf50f0836a171199
3
+ size 25606400
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14415ebcc7211bca20985a2ed3f52e4ec7b34210ff745d09e0ff65efc6ceabc5
3
+ size 62780416
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dfb2102759fdacb95ed52ffcae23d3956abba58e6b878d7106c1d240bfe7c68
3
+ size 99266560
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa88855ddf0282562efc0f76d0b5e0e513d983d8241362a09094857244b1604a
3
+ size 101842944
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5cce82babc36bba100712fcd0225220baf864c2b9ee6a9d793b3c2f5af29b9
3
+ size 203685888
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87b1427b327c840dbc04fb13467b66d9cea4015f3d31c59729043c3c42cc1555
3
+ size 33427200
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf4e75859c9cdf8482525b907b755c1fe422505371e21793ee24e161def2b9d1
3
+ size 18072576
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a3f24f855e36ab892b1defb83b542f956819b84e3addc5481e2be0b4bdbbe1
3
+ size 33381120
params_shard_29.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:679ec3a5f36635fbc68c5c55466d2e1a085c3c1c1d16bbc1875ffffd926cd55a
3
+ size 44726784
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62d9fc63528d7b840239dbf69bf0bd6546abe011fe2259ce9b47a7e807ac1e8f
3
+ size 203685888
params_shard_30.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de965075f1fcab971a056cde6968c3601b3cbcfe5f575b394cc1f4d9cc43cf55
3
+ size 153730560
params_shard_31.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c455caa5dacc6da1b390cfa80640f26599b5cb296f2d9de076947cf11416d892
3
+ size 32617728
params_shard_32.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d51992712dc475e885fb4fbe1818012ae9266e88a037ade5a15b911814e5e7a
3
+ size 72347136
params_shard_33.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a1dfbca20a8098cf92e738e46933acf13cec771fd37f42242ce5ecc76b4db0
3
+ size 27963392
params_shard_34.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e2e45dabd14d5f56409087826a9b8536eeb0e29046e19a11a6679a69a5204df
3
+ size 20345856
params_shard_35.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9f31fe1865cce85a640e35e7af7fe347a3ab4cdc27b7316e07a897f1f6f028d
3
+ size 99512832
params_shard_36.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da80d8a21bec0b1493bc611871d08af77e41464a61023d13b1da2e599b4e1137
3
+ size 31891200
params_shard_37.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c3804820b65b834a24c3b3569e0afe3d085f41d2dceb5329296a0c89edd5dd9
3
+ size 20345856
params_shard_38.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:266d60af38df04845819ab1002cbebcfa2eb85c0aa1eed9695836ea8750585c5
3
+ size 72347136
params_shard_39.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99183bd80b2ff835dd559395aadfe7e8b77a1dd7c9c281973df6125bbf7c2879
3
+ size 28702720
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8025b6464a4a435f317c92542147fa0f00ab066155facf3d72ddad481eb396a
3
+ size 101842944
params_shard_40.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d987cd63822644eb405521cb221c0a56d511b67659854738e08fd461f9a9b8e
3
+ size 20345856
params_shard_41.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:848391b6e3ad1bede32378dfa2295f550e105b4f2846615705bb832c78599dd5
3
+ size 72347136
params_shard_42.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8025855f8d919c48e045af25a49a1306cbc645c3c0cae3b5516d0b9ca07c648e
3
+ size 32644096
params_shard_43.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c946bbb4d614983b3cbfc5aefd8588664e72a7661283fdd6375d7505bed4d2c6
3
+ size 32536320
params_shard_44.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bf96c83f17ae016da612555674a26b2719ae8383677ebd00cc036c09859f325
3
+ size 99512832
params_shard_45.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77224bef38e9e3e52a5ead2e021ee9d5c339db416622b2ddf3837e5f27670cfe
3
+ size 31777792
params_shard_46.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:172fdac0362f1f35f3f49745a5ac3c46ace5cee20f9059312d5de26fa3570d89
3
+ size 32536320
params_shard_47.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09d474703bb688b2599c16b2d2d118b88320ee1c2367f35e6372e1659981d0fc
3
+ size 72347136
params_shard_48.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cb8d673ec9003c172ed26e144976abeb0213aa316ed6b733c092dfc9e4e15aa
3
+ size 31385856
params_shard_49.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9c0a20839b9b38bbb9f0d4250cc66f8fa0f99977f47b928c11fa6006a9b6580
3
+ size 101842944
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b013a7b70d117b6b89148a368fad911ffe448928ff10e8ef72d223821375abd
3
+ size 203685888