| dtype: bfloat16 | |
| flux_path: black-forest-labs/FLUX.1-Fill-dev | |
| model: | |
| add_cond_attn: false | |
| latent_lora: false | |
| union_cond_attn: true | |
| model_type: flux_fill | |
| train: | |
| accumulate_grad_batches: 2 | |
| batch_size: 8 | |
| condition_type: word_fill | |
| dataloader_workers: 4 | |
| dataset: | |
| condition_size: 512 | |
| drop_image_prob: 0.1 | |
| drop_text_prob: 0.1 | |
| image_size: 512 | |
| padding: 8 | |
| target_size: 512 | |
| type: word | |
| gradient_checkpointing: true | |
| lora_config: | |
| init_lora_weights: gaussian | |
| lora_alpha: 256 | |
| r: 256 | |
| target_modules: (.*x_embedder|.*(?<!single_)transformer_blocks\.[0-9]+\.norm1\.linear|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_k|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_q|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_v|.*(?<!single_)transformer_blocks\.[0-9]+\.attn\.to_out\.0|.*(?<!single_)transformer_blocks\.[0-9]+\.ff\.net\.2|.*single_transformer_blocks\.[0-9]+\.norm\.linear|.*single_transformer_blocks\.[0-9]+\.proj_mlp|.*single_transformer_blocks\.[0-9]+\.proj_out|.*single_transformer_blocks\.[0-9]+\.attn.to_k|.*single_transformer_blocks\.[0-9]+\.attn.to_q|.*single_transformer_blocks\.[0-9]+\.attn.to_v|.*single_transformer_blocks\.[0-9]+\.attn.to_out) | |
| max_steps: -1 | |
| odm_loss: | |
| input_resolution: 512 | |
| modelpath: epoch_100.pt | |
| w_loss_1: 30 | |
| w_loss_2: 30 | |
| w_loss_3: 30 | |
| w_loss_4: 30 | |
| w_loss_f: 1 | |
| optimizer: | |
| params: | |
| lr: 1 | |
| safeguard_warmup: true | |
| use_bias_correction: true | |
| weight_decay: 0.01 | |
| type: Prodigy | |
| reuse_lora_path: reuse.safetensors | |
| sample_interval: 2000 | |
| save_interval: 1000 | |
| save_path: output_dir | |