kirisame commited on
Commit
ed31d35
1 Parent(s): 4836839

wd 1.4 anime model release

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. wd-1-4-anime_e1.ckpt +3 -0
  3. wd-1-4-anime_e1.yaml +67 -0
.gitattributes CHANGED
@@ -34,3 +34,4 @@ vae/kl-f8-anime.ckpt filter=lfs diff=lfs merge=lfs -text
34
  vae/kl-f8-anime2.ckpt filter=lfs diff=lfs merge=lfs -text
35
  models/wd-1-3-penultimate-ucg-cont.ckpt filter=lfs diff=lfs merge=lfs -text
36
  models/wd-1-4-float32-booru-110k.ckpt filter=lfs diff=lfs merge=lfs -text
 
 
34
  vae/kl-f8-anime2.ckpt filter=lfs diff=lfs merge=lfs -text
35
  models/wd-1-3-penultimate-ucg-cont.ckpt filter=lfs diff=lfs merge=lfs -text
36
  models/wd-1-4-float32-booru-110k.ckpt filter=lfs diff=lfs merge=lfs -text
37
+ wd-1-4-anime_e1.ckpt filter=lfs diff=lfs merge=lfs -text
wd-1-4-anime_e1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8f4b2225b8ac6464e9d13d25b708349b1ca340c92dfc3575e8d5ef18f689457
3
+ size 5160236290
wd-1-4-anime_e1.yaml ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 1.0e-4
3
+ target: ldm.models.diffusion.ddpm.LatentDiffusion
4
+ params:
5
+ linear_start: 0.00085
6
+ linear_end: 0.0120
7
+ num_timesteps_cond: 1
8
+ log_every_t: 200
9
+ timesteps: 1000
10
+ first_stage_key: "jpg"
11
+ cond_stage_key: "txt"
12
+ image_size: 64
13
+ channels: 4
14
+ cond_stage_trainable: false
15
+ conditioning_key: crossattn
16
+ monitor: val/loss_simple_ema
17
+ scale_factor: 0.18215
18
+ use_ema: False # we set this to false because this is an inference only config
19
+
20
+ unet_config:
21
+ target: ldm.modules.diffusionmodules.openaimodel.UNetModel
22
+ params:
23
+ use_checkpoint: True
24
+ use_fp16: True
25
+ image_size: 32 # unused
26
+ in_channels: 4
27
+ out_channels: 4
28
+ model_channels: 320
29
+ attention_resolutions: [ 4, 2, 1 ]
30
+ num_res_blocks: 2
31
+ channel_mult: [ 1, 2, 4, 4 ]
32
+ num_head_channels: 64 # need to fix for flash-attn
33
+ use_spatial_transformer: True
34
+ use_linear_in_transformer: True
35
+ transformer_depth: 1
36
+ context_dim: 1024
37
+ legacy: False
38
+
39
+ first_stage_config:
40
+ target: ldm.models.autoencoder.AutoencoderKL
41
+ params:
42
+ embed_dim: 4
43
+ monitor: val/rec_loss
44
+ ddconfig:
45
+ #attn_type: "vanilla-xformers"
46
+ double_z: true
47
+ z_channels: 4
48
+ resolution: 256
49
+ in_channels: 3
50
+ out_ch: 3
51
+ ch: 128
52
+ ch_mult:
53
+ - 1
54
+ - 2
55
+ - 4
56
+ - 4
57
+ num_res_blocks: 2
58
+ attn_resolutions: []
59
+ dropout: 0.0
60
+ lossconfig:
61
+ target: torch.nn.Identity
62
+
63
+ cond_stage_config:
64
+ target: ldm.modules.encoders.modules.FrozenOpenCLIPEmbedder
65
+ params:
66
+ freeze: True
67
+ layer: "penultimate"