mingyi456 commited on
Commit
7354fa4
·
verified ·
1 Parent(s): aada5eb

Upload 2 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ lumina_2_model_bf16-DF11-workflow.png filter=lfs diff=lfs merge=lfs -text
lumina_2_model_bf16-DF11-workflow.json ADDED
@@ -0,0 +1,548 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": "14cfe0d4-af51-4a99-b160-53cee3b55d61",
3
+ "revision": 0,
4
+ "last_node_id": 19,
5
+ "last_link_id": 22,
6
+ "nodes": [
7
+ {
8
+ "id": 8,
9
+ "type": "VAEDecode",
10
+ "pos": [
11
+ 1209,
12
+ 188
13
+ ],
14
+ "size": [
15
+ 210,
16
+ 46
17
+ ],
18
+ "flags": {},
19
+ "order": 10,
20
+ "mode": 0,
21
+ "inputs": [
22
+ {
23
+ "name": "samples",
24
+ "type": "LATENT",
25
+ "link": 14
26
+ },
27
+ {
28
+ "name": "vae",
29
+ "type": "VAE",
30
+ "link": 21
31
+ }
32
+ ],
33
+ "outputs": [
34
+ {
35
+ "name": "IMAGE",
36
+ "type": "IMAGE",
37
+ "slot_index": 0,
38
+ "links": [
39
+ 16
40
+ ]
41
+ }
42
+ ],
43
+ "properties": {
44
+ "cnr_id": "comfy-core",
45
+ "ver": "0.3.75",
46
+ "Node name for S&R": "VAEDecode"
47
+ },
48
+ "widgets_values": []
49
+ },
50
+ {
51
+ "id": 13,
52
+ "type": "EmptySD3LatentImage",
53
+ "pos": [
54
+ 530,
55
+ 620
56
+ ],
57
+ "size": [
58
+ 315,
59
+ 106
60
+ ],
61
+ "flags": {},
62
+ "order": 0,
63
+ "mode": 0,
64
+ "inputs": [],
65
+ "outputs": [
66
+ {
67
+ "name": "LATENT",
68
+ "type": "LATENT",
69
+ "slot_index": 0,
70
+ "links": [
71
+ 17
72
+ ]
73
+ }
74
+ ],
75
+ "properties": {
76
+ "cnr_id": "comfy-core",
77
+ "ver": "0.3.75",
78
+ "Node name for S&R": "EmptySD3LatentImage"
79
+ },
80
+ "widgets_values": [
81
+ 1024,
82
+ 1024,
83
+ 1
84
+ ]
85
+ },
86
+ {
87
+ "id": 9,
88
+ "type": "SaveImage",
89
+ "pos": [
90
+ 1454.93896484375,
91
+ 190.9700164794922
92
+ ],
93
+ "size": [
94
+ 976.0567626953125,
95
+ 1060.9766845703125
96
+ ],
97
+ "flags": {},
98
+ "order": 11,
99
+ "mode": 0,
100
+ "inputs": [
101
+ {
102
+ "name": "images",
103
+ "type": "IMAGE",
104
+ "link": 16
105
+ }
106
+ ],
107
+ "outputs": [],
108
+ "properties": {
109
+ "cnr_id": "comfy-core",
110
+ "ver": "0.3.75"
111
+ },
112
+ "widgets_values": [
113
+ "ComfyUI"
114
+ ]
115
+ },
116
+ {
117
+ "id": 11,
118
+ "type": "ModelSamplingAuraFlow",
119
+ "pos": [
120
+ 524.8006591796875,
121
+ 81.51603698730469
122
+ ],
123
+ "size": [
124
+ 315,
125
+ 58
126
+ ],
127
+ "flags": {},
128
+ "order": 8,
129
+ "mode": 0,
130
+ "inputs": [
131
+ {
132
+ "name": "model",
133
+ "type": "MODEL",
134
+ "link": 22
135
+ }
136
+ ],
137
+ "outputs": [
138
+ {
139
+ "name": "MODEL",
140
+ "type": "MODEL",
141
+ "slot_index": 0,
142
+ "links": [
143
+ 13
144
+ ]
145
+ }
146
+ ],
147
+ "properties": {
148
+ "cnr_id": "comfy-core",
149
+ "ver": "0.3.75",
150
+ "Node name for S&R": "ModelSamplingAuraFlow"
151
+ },
152
+ "widgets_values": [
153
+ 6
154
+ ]
155
+ },
156
+ {
157
+ "id": 7,
158
+ "type": "CLIPTextEncode",
159
+ "pos": [
160
+ 420,
161
+ 400
162
+ ],
163
+ "size": [
164
+ 425.27801513671875,
165
+ 180.6060791015625
166
+ ],
167
+ "flags": {},
168
+ "order": 7,
169
+ "mode": 0,
170
+ "inputs": [
171
+ {
172
+ "name": "clip",
173
+ "type": "CLIP",
174
+ "link": 20
175
+ }
176
+ ],
177
+ "outputs": [
178
+ {
179
+ "name": "CONDITIONING",
180
+ "type": "CONDITIONING",
181
+ "slot_index": 0,
182
+ "links": [
183
+ 6
184
+ ]
185
+ }
186
+ ],
187
+ "title": "CLIP Text Encode (Negative Prompt)",
188
+ "properties": {
189
+ "cnr_id": "comfy-core",
190
+ "ver": "0.3.75",
191
+ "Node name for S&R": "CLIPTextEncode"
192
+ },
193
+ "widgets_values": [
194
+ "blurry ugly bad"
195
+ ],
196
+ "color": "#322",
197
+ "bgcolor": "#533"
198
+ },
199
+ {
200
+ "id": 15,
201
+ "type": "Note",
202
+ "pos": [
203
+ 71.95149993896484,
204
+ 192.96051025390625
205
+ ],
206
+ "size": [
207
+ 319.26513671875,
208
+ 197.89625549316406
209
+ ],
210
+ "flags": {},
211
+ "order": 1,
212
+ "mode": 0,
213
+ "inputs": [],
214
+ "outputs": [],
215
+ "properties": {},
216
+ "widgets_values": [
217
+ "The \"You are an assistant... <Prompt Start> \" text before the actual prompt is the one used in the official example.\n\nThe reason it is exposed to the user like this is because the model still works if you modify or remove it."
218
+ ],
219
+ "color": "#432",
220
+ "bgcolor": "#653"
221
+ },
222
+ {
223
+ "id": 14,
224
+ "type": "Note",
225
+ "pos": [
226
+ 860,
227
+ -50
228
+ ],
229
+ "size": [
230
+ 310,
231
+ 180
232
+ ],
233
+ "flags": {},
234
+ "order": 2,
235
+ "mode": 0,
236
+ "inputs": [],
237
+ "outputs": [],
238
+ "properties": {},
239
+ "widgets_values": [
240
+ "The official way to sample this model is: shift 6 with 36 steps\n\nSampling it with lower steps works but you might have to lower the shift value to reduce the amount of artifacts.\n\nEx: 20 steps with shift 3 seems to not produce artifacts"
241
+ ],
242
+ "color": "#432",
243
+ "bgcolor": "#653"
244
+ },
245
+ {
246
+ "id": 6,
247
+ "type": "CLIPTextEncode",
248
+ "pos": [
249
+ 420,
250
+ 190
251
+ ],
252
+ "size": [
253
+ 423.83001708984375,
254
+ 177.11770629882812
255
+ ],
256
+ "flags": {},
257
+ "order": 6,
258
+ "mode": 0,
259
+ "inputs": [
260
+ {
261
+ "name": "clip",
262
+ "type": "CLIP",
263
+ "link": 19
264
+ }
265
+ ],
266
+ "outputs": [
267
+ {
268
+ "name": "CONDITIONING",
269
+ "type": "CONDITIONING",
270
+ "slot_index": 0,
271
+ "links": [
272
+ 4
273
+ ]
274
+ }
275
+ ],
276
+ "title": "CLIP Text Encode (Positive Prompt)",
277
+ "properties": {
278
+ "cnr_id": "comfy-core",
279
+ "ver": "0.3.75",
280
+ "Node name for S&R": "CLIPTextEncode"
281
+ },
282
+ "widgets_values": [
283
+ "You are an assistant designed to generate superior images with the superior degree of image-text alignment based on textual prompts or user prompts. <Prompt Start> a cute anime girl with massive fennec ears mouth open and a big fluffy tail long blonde hair and blue eyes wearing a maid outfit with a long black dress and a large purple liquid stained white apron and white gloves and black leggings sitting on a large cushion in the middle of a kitchen in a dark victorian mansion with a stained glass window drinking a glass with a galaxy inside"
284
+ ],
285
+ "color": "#232",
286
+ "bgcolor": "#353"
287
+ },
288
+ {
289
+ "id": 17,
290
+ "type": "CLIPLoader",
291
+ "pos": [
292
+ 84.4089852709569,
293
+ 579.9749737092388
294
+ ],
295
+ "size": [
296
+ 270,
297
+ 106
298
+ ],
299
+ "flags": {},
300
+ "order": 3,
301
+ "mode": 0,
302
+ "inputs": [],
303
+ "outputs": [
304
+ {
305
+ "name": "CLIP",
306
+ "type": "CLIP",
307
+ "links": [
308
+ 19,
309
+ 20
310
+ ]
311
+ }
312
+ ],
313
+ "properties": {
314
+ "cnr_id": "comfy-core",
315
+ "ver": "0.3.75",
316
+ "Node name for S&R": "CLIPLoader"
317
+ },
318
+ "widgets_values": [
319
+ "gemma_2_2b_fp16.safetensors",
320
+ "lumina2",
321
+ "default"
322
+ ]
323
+ },
324
+ {
325
+ "id": 18,
326
+ "type": "VAELoader",
327
+ "pos": [
328
+ 82.24204408227945,
329
+ 732.7443275110128
330
+ ],
331
+ "size": [
332
+ 270,
333
+ 58
334
+ ],
335
+ "flags": {},
336
+ "order": 4,
337
+ "mode": 0,
338
+ "inputs": [],
339
+ "outputs": [
340
+ {
341
+ "name": "VAE",
342
+ "type": "VAE",
343
+ "links": [
344
+ 21
345
+ ]
346
+ }
347
+ ],
348
+ "properties": {
349
+ "cnr_id": "comfy-core",
350
+ "ver": "0.3.75",
351
+ "Node name for S&R": "VAELoader"
352
+ },
353
+ "widgets_values": [
354
+ "ae.safetensors"
355
+ ]
356
+ },
357
+ {
358
+ "id": 19,
359
+ "type": "DFloat11ModelLoader",
360
+ "pos": [
361
+ 68.859432083083,
362
+ 460.61347106132257
363
+ ],
364
+ "size": [
365
+ 281.861328125,
366
+ 58
367
+ ],
368
+ "flags": {},
369
+ "order": 5,
370
+ "mode": 0,
371
+ "inputs": [],
372
+ "outputs": [
373
+ {
374
+ "name": "MODEL",
375
+ "type": "MODEL",
376
+ "links": [
377
+ 22
378
+ ]
379
+ }
380
+ ],
381
+ "properties": {
382
+ "aux_id": "mingyi456/ComfyUI-DFloat11-Extended",
383
+ "ver": "a4538723928a03ace4c18047668c020dd32feb66",
384
+ "Node name for S&R": "DFloat11ModelLoader"
385
+ },
386
+ "widgets_values": [
387
+ "lumina_2_model_bf16-DF11.safetensors"
388
+ ]
389
+ },
390
+ {
391
+ "id": 3,
392
+ "type": "KSampler",
393
+ "pos": [
394
+ 863,
395
+ 186
396
+ ],
397
+ "size": [
398
+ 315,
399
+ 262
400
+ ],
401
+ "flags": {},
402
+ "order": 9,
403
+ "mode": 0,
404
+ "inputs": [
405
+ {
406
+ "name": "model",
407
+ "type": "MODEL",
408
+ "link": 13
409
+ },
410
+ {
411
+ "name": "positive",
412
+ "type": "CONDITIONING",
413
+ "link": 4
414
+ },
415
+ {
416
+ "name": "negative",
417
+ "type": "CONDITIONING",
418
+ "link": 6
419
+ },
420
+ {
421
+ "name": "latent_image",
422
+ "type": "LATENT",
423
+ "link": 17
424
+ }
425
+ ],
426
+ "outputs": [
427
+ {
428
+ "name": "LATENT",
429
+ "type": "LATENT",
430
+ "slot_index": 0,
431
+ "links": [
432
+ 14
433
+ ]
434
+ }
435
+ ],
436
+ "properties": {
437
+ "cnr_id": "comfy-core",
438
+ "ver": "0.3.75",
439
+ "Node name for S&R": "KSampler"
440
+ },
441
+ "widgets_values": [
442
+ 854547257397629,
443
+ "randomize",
444
+ 25,
445
+ 4,
446
+ "res_multistep",
447
+ "simple",
448
+ 1
449
+ ]
450
+ }
451
+ ],
452
+ "links": [
453
+ [
454
+ 4,
455
+ 6,
456
+ 0,
457
+ 3,
458
+ 1,
459
+ "CONDITIONING"
460
+ ],
461
+ [
462
+ 6,
463
+ 7,
464
+ 0,
465
+ 3,
466
+ 2,
467
+ "CONDITIONING"
468
+ ],
469
+ [
470
+ 13,
471
+ 11,
472
+ 0,
473
+ 3,
474
+ 0,
475
+ "MODEL"
476
+ ],
477
+ [
478
+ 14,
479
+ 3,
480
+ 0,
481
+ 8,
482
+ 0,
483
+ "LATENT"
484
+ ],
485
+ [
486
+ 16,
487
+ 8,
488
+ 0,
489
+ 9,
490
+ 0,
491
+ "IMAGE"
492
+ ],
493
+ [
494
+ 17,
495
+ 13,
496
+ 0,
497
+ 3,
498
+ 3,
499
+ "LATENT"
500
+ ],
501
+ [
502
+ 19,
503
+ 17,
504
+ 0,
505
+ 6,
506
+ 0,
507
+ "CLIP"
508
+ ],
509
+ [
510
+ 20,
511
+ 17,
512
+ 0,
513
+ 7,
514
+ 0,
515
+ "CLIP"
516
+ ],
517
+ [
518
+ 21,
519
+ 18,
520
+ 0,
521
+ 8,
522
+ 1,
523
+ "VAE"
524
+ ],
525
+ [
526
+ 22,
527
+ 19,
528
+ 0,
529
+ 11,
530
+ 0,
531
+ "MODEL"
532
+ ]
533
+ ],
534
+ "groups": [],
535
+ "config": {},
536
+ "extra": {
537
+ "ds": {
538
+ "scale": 1.0152559799477197,
539
+ "offset": [
540
+ 148.54386753733925,
541
+ 82.03785124693952
542
+ ]
543
+ },
544
+ "frontendVersion": "1.32.9",
545
+ "workflowRendererVersion": "LG"
546
+ },
547
+ "version": 0.4
548
+ }
lumina_2_model_bf16-DF11-workflow.png ADDED

Git LFS Details

  • SHA256: 82693e2e693670c4ee058548cf18999e93bf186db98120c21a3906b883bd91f9
  • Pointer size: 132 Bytes
  • Size of remote file: 1.49 MB