adipanda commited on
Commit
8b51ff2
1 Parent(s): 5e067ca

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ images_0.png filter=lfs diff=lfs merge=lfs -text
37
+ images_1.png filter=lfs diff=lfs merge=lfs -text
38
+ images_2.png filter=lfs diff=lfs merge=lfs -text
39
+ images_3.png filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: John6666/anima-pencil-sdxl
3
+ library_name: diffusers
4
+ license: openrail++
5
+ tags:
6
+ - stable-diffusion-xl
7
+ - stable-diffusion-xl-diffusers
8
+ - text-to-image
9
+ - diffusers
10
+ - controlnet
11
+ - diffusers-training
12
+ inference: true
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # controlnet-adipanda/colorizer_300k_3-6
20
+
21
+ These are controlnet weights trained on John6666/anima-pencil-sdxl with new type of conditioning.
22
+ You can find some example images below.
23
+
24
+ prompt: comic, english text, blue hair, multiple girls, !, multiple boys, bandages, brown hair
25
+ ![images_0)](./images_0.png)
26
+ prompt: 1boy, male focus, orange hair, open mouth, english text, solo, speech bubble, shirt, short hair, blush, comic, teeth, white shirt, brown eyes
27
+ ![images_1)](./images_1.png)
28
+ prompt: english text, comic, black hair, red hair, weapon, sword, speech bubble, glasses, japanese clothes, multiple girls, 1boy, !, 2girls, holding weapon, ponytail, holding, bag, katana, teeth, holding sword, backpack, multiple boys
29
+ ![images_2)](./images_2.png)
30
+ prompt: comic, multiple girls, 2girls, blonde hair, breasts, cleavage, mask, large breasts, multiple boys, black hair, long hair, brown eyes, 2boys
31
+ ![images_3)](./images_3.png)
32
+
33
+
34
+
35
+ ## Intended uses & limitations
36
+
37
+ #### How to use
38
+
39
+ ```python
40
+ # TODO: add an example code snippet for running this diffusion pipeline
41
+ ```
42
+
43
+ #### Limitations and bias
44
+
45
+ [TODO: provide examples of latent issues and potential remediations]
46
+
47
+ ## Training details
48
+
49
+ [TODO: describe the data used to train the model]
checkpoint-500/controlnet/config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ControlNetModel",
3
+ "_diffusers_version": "0.30.0.dev0",
4
+ "_name_or_path": "./prev_controlnet",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": "text_time",
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": 256,
9
+ "attention_head_dim": [
10
+ 5,
11
+ 10,
12
+ 20
13
+ ],
14
+ "block_out_channels": [
15
+ 320,
16
+ 640,
17
+ 1280
18
+ ],
19
+ "class_embed_type": null,
20
+ "conditioning_channels": 3,
21
+ "conditioning_embedding_out_channels": [
22
+ 16,
23
+ 32,
24
+ 96,
25
+ 256
26
+ ],
27
+ "controlnet_conditioning_channel_order": "rgb",
28
+ "cross_attention_dim": 2048,
29
+ "down_block_types": [
30
+ "DownBlock2D",
31
+ "CrossAttnDownBlock2D",
32
+ "CrossAttnDownBlock2D"
33
+ ],
34
+ "downsample_padding": 1,
35
+ "encoder_hid_dim": null,
36
+ "encoder_hid_dim_type": null,
37
+ "flip_sin_to_cos": true,
38
+ "freq_shift": 0,
39
+ "global_pool_conditions": false,
40
+ "in_channels": 4,
41
+ "layers_per_block": 2,
42
+ "mid_block_scale_factor": 1,
43
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
44
+ "norm_eps": 1e-05,
45
+ "norm_num_groups": 32,
46
+ "num_attention_heads": null,
47
+ "num_class_embeds": null,
48
+ "only_cross_attention": false,
49
+ "projection_class_embeddings_input_dim": 2816,
50
+ "resnet_time_scale_shift": "default",
51
+ "transformer_layers_per_block": [
52
+ 1,
53
+ 2,
54
+ 10
55
+ ],
56
+ "upcast_attention": null,
57
+ "use_linear_projection": true
58
+ }
checkpoint-500/controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:feea2ade069acedc3560cc1fad727fbfc8bb5ebcee7e3a1a9db74103eb861621
3
+ size 5004167864
checkpoint-500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13b6d2e7ace6352e98896a8d09f963af20afddfa90fc43f13630e85b66d839bf
3
+ size 10008841510
checkpoint-500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67704f2a8a1d531e46cb7abb9464bcb99af3f54622fcfe3b062100ba0002e7aa
3
+ size 15060
checkpoint-500/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18b984273ea2d45b7ffb1d047bb359d93111e41fcad70d16a1b453fd38f72636
3
+ size 988
checkpoint-500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88022db35e802613d8fdda53ce01c5c68144b6adbcc6bc79ddd6cb06134eca02
3
+ size 1000
config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ControlNetModel",
3
+ "_diffusers_version": "0.30.0.dev0",
4
+ "_name_or_path": "./prev_controlnet",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": "text_time",
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": 256,
9
+ "attention_head_dim": [
10
+ 5,
11
+ 10,
12
+ 20
13
+ ],
14
+ "block_out_channels": [
15
+ 320,
16
+ 640,
17
+ 1280
18
+ ],
19
+ "class_embed_type": null,
20
+ "conditioning_channels": 3,
21
+ "conditioning_embedding_out_channels": [
22
+ 16,
23
+ 32,
24
+ 96,
25
+ 256
26
+ ],
27
+ "controlnet_conditioning_channel_order": "rgb",
28
+ "cross_attention_dim": 2048,
29
+ "down_block_types": [
30
+ "DownBlock2D",
31
+ "CrossAttnDownBlock2D",
32
+ "CrossAttnDownBlock2D"
33
+ ],
34
+ "downsample_padding": 1,
35
+ "encoder_hid_dim": null,
36
+ "encoder_hid_dim_type": null,
37
+ "flip_sin_to_cos": true,
38
+ "freq_shift": 0,
39
+ "global_pool_conditions": false,
40
+ "in_channels": 4,
41
+ "layers_per_block": 2,
42
+ "mid_block_scale_factor": 1,
43
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
44
+ "norm_eps": 1e-05,
45
+ "norm_num_groups": 32,
46
+ "num_attention_heads": null,
47
+ "num_class_embeds": null,
48
+ "only_cross_attention": false,
49
+ "projection_class_embeddings_input_dim": 2816,
50
+ "resnet_time_scale_shift": "default",
51
+ "transformer_layers_per_block": [
52
+ 1,
53
+ 2,
54
+ 10
55
+ ],
56
+ "upcast_attention": null,
57
+ "use_linear_projection": true
58
+ }
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf8884130193bfbb7285c148abe1416302422bddc2673fc68cd04585d2e08216
3
+ size 5004167864
image_control.png ADDED
images_0.png ADDED

Git LFS Details

  • SHA256: 122fc302200927e54314305eb88f9b76a2e0f3ccd9904b08c4b1f99a1f747965
  • Pointer size: 132 Bytes
  • Size of remote file: 4.98 MB
images_1.png ADDED

Git LFS Details

  • SHA256: b06089139b8d5fdf820b00069a7a59111021ebbe7b7be4150492fd93cb74cf0e
  • Pointer size: 132 Bytes
  • Size of remote file: 3.14 MB
images_2.png ADDED

Git LFS Details

  • SHA256: 1a6092853a00e349bdd67ef9aa001b39df96b7ebdc5cd1a2408d381418041c94
  • Pointer size: 132 Bytes
  • Size of remote file: 3.78 MB
images_3.png ADDED

Git LFS Details

  • SHA256: 9a7657b8b512e8dee436720ac6de46fc60c11e20dab7d56e2bdff3ece8122702
  • Pointer size: 132 Bytes
  • Size of remote file: 4.42 MB