Commit
·
1ee0abf
1
Parent(s):
79a0e90
update: upload joint-train intermediate ckpt
Browse files- sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/args.json +61 -0
- sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/checkpoints/0400000.pt +3 -0
- sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k.csv +3 -0
sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/args.json
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"output_dir": "exps",
|
| 3 |
+
"exp_name": "sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k",
|
| 4 |
+
"logging_dir": "logs",
|
| 5 |
+
"report_to": "wandb",
|
| 6 |
+
"sampling_steps": 10000,
|
| 7 |
+
"resume_step": 0,
|
| 8 |
+
"continue_train_exp_dir": null,
|
| 9 |
+
"wandb_history_path": null,
|
| 10 |
+
"requeue": false,
|
| 11 |
+
"model": "SiT-XL/1",
|
| 12 |
+
"num_classes": 1000,
|
| 13 |
+
"encoder_depth": 8,
|
| 14 |
+
"fused_attn": true,
|
| 15 |
+
"qk_norm": false,
|
| 16 |
+
"bn_momentum": 0.1,
|
| 17 |
+
"bn_affine": false,
|
| 18 |
+
"compile": true,
|
| 19 |
+
"data_dir": "data/imagenet-latents",
|
| 20 |
+
"resolution": 256,
|
| 21 |
+
"batch_size": 256,
|
| 22 |
+
"allow_tf32": true,
|
| 23 |
+
"mixed_precision": "fp16",
|
| 24 |
+
"epochs": 1400,
|
| 25 |
+
"max_train_steps": 400000,
|
| 26 |
+
"checkpointing_steps": 50000,
|
| 27 |
+
"gradient_accumulation_steps": 1,
|
| 28 |
+
"learning_rate": 0.0001,
|
| 29 |
+
"adam_beta1": 0.9,
|
| 30 |
+
"adam_beta2": 0.999,
|
| 31 |
+
"adam_weight_decay": 0.0,
|
| 32 |
+
"adam_epsilon": 1e-08,
|
| 33 |
+
"max_grad_norm": 1.0,
|
| 34 |
+
"seed": 0,
|
| 35 |
+
"num_workers": 4,
|
| 36 |
+
"path_type": "linear",
|
| 37 |
+
"prediction": "v",
|
| 38 |
+
"cfg_prob": 0.1,
|
| 39 |
+
"enc_type": "dinov2-vit-b",
|
| 40 |
+
"proj_coeff": 0.5,
|
| 41 |
+
"weighting": "uniform",
|
| 42 |
+
"legacy": false,
|
| 43 |
+
"vae": "f16d32",
|
| 44 |
+
"vae_ckpt": "pretrained_models/invae-f16d32.pt",
|
| 45 |
+
"use_sd_vae_stats": false,
|
| 46 |
+
"vae_repa_align": false,
|
| 47 |
+
"freeze_bn_stats": false,
|
| 48 |
+
"vae_lora": false,
|
| 49 |
+
"vae_lora_rank": 64,
|
| 50 |
+
"vae_lora_alpha": 16,
|
| 51 |
+
"vae_lora_dropout": 0.0,
|
| 52 |
+
"vae_lora_bias": "none",
|
| 53 |
+
"disc_pretrained_ckpt": "pretrained_models/invae-discriminator-ckpt.pt",
|
| 54 |
+
"loss_cfg_path": "configs/loss_cfg/l1_lpips_kl_gan.yaml",
|
| 55 |
+
"vae_learning_rate": 0.0001,
|
| 56 |
+
"disc_learning_rate": 0.0001,
|
| 57 |
+
"vae_train_freq": 1,
|
| 58 |
+
"unfreeze_decoder": true,
|
| 59 |
+
"vae_align": true,
|
| 60 |
+
"vae_align_proj_coeff": 1.5
|
| 61 |
+
}
|
sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/checkpoints/0400000.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53b4cabdee2a3293650df5cf610d97d25d17ae01115905e9741f9c57da84ccf1
|
| 3 |
+
size 11804655505
|
sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k/sit-xl-1-dinov2-b-enc8-invae-freq1-lr1e-4-momentum0.1-l1-lpips-kl-gan-vaealign1.5-unfreezedecoder-compile-400k.csv
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a0678d8f164cfb788e776e158c0d77df73a97c364da2d050106053618e1e6bda
|
| 3 |
+
size 155328420
|