Spaces:
Runtime error
Runtime error
| #### Model Setting | |
| pretrained_model_name_or_path = 'stabilityai/stable-diffusion-xl-base-1.0' | |
| pretrained_vae_model_name_or_path = 'madebyollin/sdxl-vae-fp16-fix' | |
| revision = None | |
| byt5_max_length = 512 | |
| byt5_mapper_type = 'T5EncoderBlockByT5Mapper' | |
| byt5_mapper_config = dict( | |
| num_layers=4, | |
| sdxl_channels=2048, | |
| ) | |
| byt5_config = dict( | |
| byt5_name='google/byt5-small', | |
| special_token=True, | |
| color_special_token=True, | |
| font_special_token=True, | |
| color_ann_path='assets/color_idx.json', | |
| font_ann_path='assets/font_idx_512.json', | |
| multilingual=False, | |
| ) | |
| attn_block_to_modify = [ | |
| "down_blocks.1.attentions.0.transformer_blocks.0", | |
| "down_blocks.1.attentions.0.transformer_blocks.1", | |
| "down_blocks.1.attentions.1.transformer_blocks.0", | |
| "down_blocks.1.attentions.1.transformer_blocks.1", | |
| "down_blocks.2.attentions.0.transformer_blocks.0", | |
| "down_blocks.2.attentions.0.transformer_blocks.1", | |
| "down_blocks.2.attentions.0.transformer_blocks.2", | |
| "down_blocks.2.attentions.0.transformer_blocks.3", | |
| "down_blocks.2.attentions.0.transformer_blocks.4", | |
| "down_blocks.2.attentions.0.transformer_blocks.5", | |
| "down_blocks.2.attentions.0.transformer_blocks.6", | |
| "down_blocks.2.attentions.0.transformer_blocks.7", | |
| "down_blocks.2.attentions.0.transformer_blocks.8", | |
| "down_blocks.2.attentions.0.transformer_blocks.9", | |
| "down_blocks.2.attentions.1.transformer_blocks.0", | |
| "down_blocks.2.attentions.1.transformer_blocks.1", | |
| "down_blocks.2.attentions.1.transformer_blocks.2", | |
| "down_blocks.2.attentions.1.transformer_blocks.3", | |
| "down_blocks.2.attentions.1.transformer_blocks.4", | |
| "down_blocks.2.attentions.1.transformer_blocks.5", | |
| "down_blocks.2.attentions.1.transformer_blocks.6", | |
| "down_blocks.2.attentions.1.transformer_blocks.7", | |
| "down_blocks.2.attentions.1.transformer_blocks.8", | |
| "down_blocks.2.attentions.1.transformer_blocks.9", | |
| "up_blocks.0.attentions.0.transformer_blocks.0", | |
| "up_blocks.0.attentions.0.transformer_blocks.1", | |
| "up_blocks.0.attentions.0.transformer_blocks.2", | |
| "up_blocks.0.attentions.0.transformer_blocks.3", | |
| "up_blocks.0.attentions.0.transformer_blocks.4", | |
| "up_blocks.0.attentions.0.transformer_blocks.5", | |
| "up_blocks.0.attentions.0.transformer_blocks.6", | |
| "up_blocks.0.attentions.0.transformer_blocks.7", | |
| "up_blocks.0.attentions.0.transformer_blocks.8", | |
| "up_blocks.0.attentions.0.transformer_blocks.9", | |
| "up_blocks.0.attentions.1.transformer_blocks.0", | |
| "up_blocks.0.attentions.1.transformer_blocks.1", | |
| "up_blocks.0.attentions.1.transformer_blocks.2", | |
| "up_blocks.0.attentions.1.transformer_blocks.3", | |
| "up_blocks.0.attentions.1.transformer_blocks.4", | |
| "up_blocks.0.attentions.1.transformer_blocks.5", | |
| "up_blocks.0.attentions.1.transformer_blocks.6", | |
| "up_blocks.0.attentions.1.transformer_blocks.7", | |
| "up_blocks.0.attentions.1.transformer_blocks.8", | |
| "up_blocks.0.attentions.1.transformer_blocks.9", | |
| "up_blocks.0.attentions.2.transformer_blocks.0", | |
| "up_blocks.0.attentions.2.transformer_blocks.1", | |
| "up_blocks.0.attentions.2.transformer_blocks.2", | |
| "up_blocks.0.attentions.2.transformer_blocks.3", | |
| "up_blocks.0.attentions.2.transformer_blocks.4", | |
| "up_blocks.0.attentions.2.transformer_blocks.5", | |
| "up_blocks.0.attentions.2.transformer_blocks.6", | |
| "up_blocks.0.attentions.2.transformer_blocks.7", | |
| "up_blocks.0.attentions.2.transformer_blocks.8", | |
| "up_blocks.0.attentions.2.transformer_blocks.9", | |
| "up_blocks.1.attentions.0.transformer_blocks.0", | |
| "up_blocks.1.attentions.0.transformer_blocks.1", | |
| "up_blocks.1.attentions.1.transformer_blocks.0", | |
| "up_blocks.1.attentions.1.transformer_blocks.1", | |
| "up_blocks.1.attentions.2.transformer_blocks.0", | |
| "up_blocks.1.attentions.2.transformer_blocks.1", | |
| "mid_block.attentions.0.transformer_blocks.0", | |
| "mid_block.attentions.0.transformer_blocks.1", | |
| "mid_block.attentions.0.transformer_blocks.2", | |
| "mid_block.attentions.0.transformer_blocks.3", | |
| "mid_block.attentions.0.transformer_blocks.4", | |
| "mid_block.attentions.0.transformer_blocks.5", | |
| "mid_block.attentions.0.transformer_blocks.6", | |
| "mid_block.attentions.0.transformer_blocks.7", | |
| "mid_block.attentions.0.transformer_blocks.8", | |
| "mid_block.attentions.0.transformer_blocks.9", | |
| ] | |
| unet_lora_rank = 128 | |
| inference_dtype = 'fp16' | |