jerryzh168 commited on
Commit
4820374
·
verified ·
1 Parent(s): 3846aef

Upload OPTForCausalLM

Browse files
Files changed (1) hide show
  1. config.json +42 -0
config.json CHANGED
@@ -208,6 +208,48 @@
208
  "_type": "Float8DynamicActivationFloat8WeightConfig",
209
  "_version": 2
210
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
211
  "re:model\\.decoder\\.layers\\..+\\.self_attn\\.q_proj": {
212
  "_data": {
213
  "activation_dtype": {
 
208
  "_type": "Float8DynamicActivationFloat8WeightConfig",
209
  "_version": 2
210
  },
211
+ "re:model\\.decoder\\.layers\\..+\\.self_attn\\.out_pro": {
212
+ "_data": {
213
+ "activation_dtype": {
214
+ "_data": "float8_e4m3fn",
215
+ "_type": "torch.dtype"
216
+ },
217
+ "activation_value_lb": null,
218
+ "activation_value_ub": null,
219
+ "granularity": [
220
+ {
221
+ "_data": {},
222
+ "_type": "PerRow",
223
+ "_version": 1
224
+ },
225
+ {
226
+ "_data": {},
227
+ "_type": "PerRow",
228
+ "_version": 1
229
+ }
230
+ ],
231
+ "kernel_preference": {
232
+ "_data": "AUTO",
233
+ "_type": "KernelPreference"
234
+ },
235
+ "mm_config": {
236
+ "_data": {
237
+ "emulate": false,
238
+ "pad_inner_dim": false,
239
+ "use_fast_accum": true
240
+ },
241
+ "_type": "Float8MMConfig",
242
+ "_version": 1
243
+ },
244
+ "set_inductor_config": true,
245
+ "weight_dtype": {
246
+ "_data": "float8_e4m3fn",
247
+ "_type": "torch.dtype"
248
+ }
249
+ },
250
+ "_type": "Float8DynamicActivationFloat8WeightConfig",
251
+ "_version": 2
252
+ },
253
  "re:model\\.decoder\\.layers\\..+\\.self_attn\\.q_proj": {
254
  "_data": {
255
  "activation_dtype": {