kmpartner commited on
Commit
e1e5ac0
1 Parent(s): 7298e92

Upload folder using huggingface_hub

Browse files
controlnet/config.json CHANGED
@@ -20,7 +20,7 @@
20
  256
21
  ],
22
  "controlnet_conditioning_channel_order": "rgb",
23
- "cross_attention_dim": 768,
24
  "down_block_types": [
25
  "DownBlock2D",
26
  "CrossAttnDownBlock2D",
 
20
  256
21
  ],
22
  "controlnet_conditioning_channel_order": "rgb",
23
+ "cross_attention_dim": 1024,
24
  "down_block_types": [
25
  "DownBlock2D",
26
  "CrossAttnDownBlock2D",
controlnet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d005da74cf9e639b3b9d6a49ff845690fd2081fead531c714c98d2694e490c96
3
- size 728496840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0686b783664e862fb77836c7a6cf5d116dca2aa9d40341e68a90128177245430
3
+ size 1456953560
logs/text2image-fine-tune/1717903784.86201/events.out.tfevents.1717903784.3a3e3b616396.2113.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16f8cfa3986afcf097e6b2644a7db6d8f9957da41cb1774b7a9331730d73d19e
3
+ size 2457
logs/text2image-fine-tune/1717903784.8642344/hparams.yml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ cache_dir: null
7
+ center_crop: true
8
+ checkpointing_steps: 5000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_config_name: null
12
+ enable_xformers_memory_efficient_attention: false
13
+ gradient_accumulation_steps: 4
14
+ gradient_checkpointing: true
15
+ lambda_kd_feat: 1.0
16
+ lambda_kd_output: 1.0
17
+ lambda_sd: 1.0
18
+ learning_rate: 5.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ lr_scheduler: constant
22
+ lr_warmup_steps: 0
23
+ max_grad_norm: 1.0
24
+ max_train_samples: null
25
+ max_train_steps: 400000
26
+ mixed_precision: fp16
27
+ non_ema_revision: null
28
+ num_train_epochs: 4939
29
+ num_valid_images: 2
30
+ output_dir: ./results/v2-base_kd_bk_tiny
31
+ pretrained_model_name_or_path: stabilityai/stable-diffusion-2-1-base
32
+ random_flip: true
33
+ report_to: all
34
+ resolution: 512
35
+ resume_from_checkpoint: null
36
+ revision: null
37
+ scale_lr: false
38
+ seed: 1234
39
+ train_batch_size: 32
40
+ train_data_dir: ./data/laion_aes/preprocessed_11k
41
+ unet_config_name: bk_tiny
42
+ unet_config_path: ./src/unet_config_v2-base
43
+ use_8bit_adam: false
44
+ use_copy_weight_from_teacher: true
45
+ use_ema: true
46
+ valid_prompt: a golden vase with different flowers
47
+ valid_steps: 500
logs/text2image-fine-tune/1717903925.3434956/events.out.tfevents.1717903925.3a3e3b616396.3501.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02e23a7c5660311efb1d4a72ab4129c6e946d478660020b301651b29015f8e04
3
+ size 2409
logs/text2image-fine-tune/1717903925.3458326/hparams.yml ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ cache_dir: null
7
+ center_crop: true
8
+ checkpointing_steps: 5000
9
+ checkpoints_total_limit: null
10
+ dataloader_num_workers: 0
11
+ dataset_config_name: null
12
+ enable_xformers_memory_efficient_attention: false
13
+ gradient_accumulation_steps: 4
14
+ gradient_checkpointing: true
15
+ lambda_kd_feat: 1.0
16
+ lambda_kd_output: 1.0
17
+ lambda_sd: 1.0
18
+ learning_rate: 5.0e-05
19
+ local_rank: -1
20
+ logging_dir: logs
21
+ lr_scheduler: constant
22
+ lr_warmup_steps: 0
23
+ max_grad_norm: 1.0
24
+ max_train_samples: null
25
+ max_train_steps: 0
26
+ mixed_precision: null
27
+ non_ema_revision: null
28
+ num_train_epochs: 0
29
+ num_valid_images: 2
30
+ output_dir: ./results/v2-base_kd_bk_tiny
31
+ pretrained_model_name_or_path: stabilityai/stable-diffusion-2-1-base
32
+ random_flip: true
33
+ report_to: all
34
+ resolution: 512
35
+ resume_from_checkpoint: null
36
+ revision: null
37
+ scale_lr: false
38
+ seed: 1234
39
+ train_batch_size: 32
40
+ train_data_dir: ./data/laion_aes/preprocessed_11k
41
+ unet_config_name: bk_tiny
42
+ unet_config_path: ./src/unet_config_v2-base
43
+ use_8bit_adam: false
44
+ use_copy_weight_from_teacher: true
45
+ use_ema: true
46
+ valid_prompt: a golden vase with different flowers
47
+ valid_steps: 500
logs/text2image-fine-tune/events.out.tfevents.1717903778.3a3e3b616396.2113.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fe47a550748b5bca29fae86582b59a28b7224ad77874115c8c680f265a2b8a6
3
+ size 88
logs/text2image-fine-tune/events.out.tfevents.1717903908.3a3e3b616396.3501.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73a6f78f4d4a32a59c49b1e64948d1d5e6e4225c22cdd82eda7313ca7ffa5f9a
3
+ size 88
text_encoder/config.json CHANGED
@@ -19,7 +19,7 @@
19
  "num_hidden_layers": 23,
20
  "pad_token_id": 1,
21
  "projection_dim": 512,
22
- "torch_dtype": "float16",
23
  "transformers_version": "4.27.4",
24
  "vocab_size": 49408
25
  }
 
19
  "num_hidden_layers": 23,
20
  "pad_token_id": 1,
21
  "projection_dim": 512,
22
+ "torch_dtype": "float32",
23
  "transformers_version": "4.27.4",
24
  "vocab_size": 49408
25
  }
text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:681c555376658c81dc273f2d737a2aeb23ddb6d1d8e5b3a7064636d359a22668
3
- size 680821096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e4aa519f64dc6386f88221a66c106a09fa027b47a20cc0e126687695f2a6669
3
+ size 1361597016
vae/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e4c08995484ee61270175e9e7a072b66a6e4eeb5f0c266667fe1f45b90daf9a
3
- size 167335342
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa1f43011b553a4cba7f37456465cdbd48aab7b54b9348b890e8058ea7683ec
3
+ size 334643268