Guizmus commited on
Commit
085c1b2
1 Parent(s): 26e03f8

version 1.0 (#1)

Browse files

- All the files for 1st version, Adding diffusers and ckpt, updating readme, adding showcase (780b6431fff5942c0319badec9246ac440935a1d)

BloodborneStyle-v1.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ce1bb306053671e3c35f30219bf347ee05b87d201b41db7fdbf2cff1c97e311
3
+ size 2132871991
README.md CHANGED
@@ -1,3 +1,55 @@
1
  ---
 
 
2
  license: creativeml-openrail-m
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: creativeml-openrail-m
5
+ thumbnail: "https://huggingface.co/Guizmus/BloodborneDiffusion/resolve/main/bloodbornestyle_showcase.jpg"
6
+ tags:
7
+ - stable-diffusion
8
+ - text-to-image
9
+ - image-to-image
10
+ library_name: "https://github.com/ShivamShrirao/diffusers"
11
+
12
  ---
13
+
14
+ # DarkSouls Diffusion
15
+ <p>
16
+ <img src="https://huggingface.co/Guizmus/BloodborneDiffusion/resolve/main/bloodbornestyle_showcase.jpg"/><br/>
17
+ This is a Dreamboothed Stable Diffusion model trained on the Bloodborne series Style.<br/>
18
+ The total dataset is made of 100 pictures, and the training has been done on runawayml 1.5 and the new VAE, with 12k steps (poly LR1e-6).<br/>
19
+ The token "Bloodborne Style" will bring in the new concept.<br/>
20
+ The recommended sampling is k_Euler_a or DPM++ 2M Karras on 20 steps, CFGS 7 .
21
+
22
+ </p>
23
+
24
+ [CKPT download link](https://huggingface.co/Guizmus/Bloodborne/resolve/main/BloodborneStyle-v1.ckpt)
25
+
26
+ ## 🧨 Diffusers
27
+
28
+ This model can be used just like any other Stable Diffusion model. For more information,
29
+ please have a look at the [Stable Diffusion](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
30
+
31
+ You can also export the model to [ONNX](https://huggingface.co/docs/diffusers/optimization/onnx), [MPS](https://huggingface.co/docs/diffusers/optimization/mps) and/or [FLAX/JAX]().
32
+
33
+ ```python
34
+ from diffusers import StableDiffusionPipeline
35
+ import torch
36
+
37
+ model_id = "Guizmus/BloodborneDiffusion"
38
+ pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
39
+ pipe = pipe.to("cuda")
40
+
41
+ prompt = "a red moon, Bloodborne Style"
42
+ image = pipe(prompt).images[0]
43
+
44
+ image.save("./BloodborneStyle.png")
45
+ ```
46
+
47
+ ## License
48
+
49
+ This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
50
+ The CreativeML OpenRAIL License specifies:
51
+
52
+ 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
53
+ 2. The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
54
+ 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
55
+ [Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
args.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "pretrained_model_name_or_path": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle/v1-10k",
3
+ "pretrained_vae_name_or_path": null,
4
+ "revision": null,
5
+ "tokenizer_name": null,
6
+ "instance_data_dir": null,
7
+ "class_data_dir": null,
8
+ "instance_prompt": null,
9
+ "class_prompt": null,
10
+ "save_sample_prompt": null,
11
+ "save_sample_negative_prompt": null,
12
+ "n_save_sample": 4,
13
+ "save_guidance_scale": 7.5,
14
+ "save_infer_steps": 50,
15
+ "pad_tokens": false,
16
+ "with_prior_preservation": true,
17
+ "prior_loss_weight": 1.0,
18
+ "num_class_images": 100,
19
+ "output_dir": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle",
20
+ "seed": null,
21
+ "resolution": 512,
22
+ "center_crop": false,
23
+ "train_text_encoder": true,
24
+ "train_batch_size": 1,
25
+ "sample_batch_size": 4,
26
+ "num_train_epochs": 50,
27
+ "max_train_steps": 5000,
28
+ "gradient_accumulation_steps": 1,
29
+ "gradient_checkpointing": true,
30
+ "learning_rate": 1e-06,
31
+ "scale_lr": false,
32
+ "lr_scheduler": "polynomial",
33
+ "lr_warmup_steps": 0,
34
+ "use_8bit_adam": false,
35
+ "adam_beta1": 0.9,
36
+ "adam_beta2": 0.999,
37
+ "adam_weight_decay": 0.01,
38
+ "adam_epsilon": 1e-08,
39
+ "max_grad_norm": 1.0,
40
+ "push_to_hub": false,
41
+ "hub_token": null,
42
+ "hub_model_id": null,
43
+ "logging_dir": "logs",
44
+ "log_interval": 10,
45
+ "save_interval": 1000,
46
+ "save_min_steps": 1000,
47
+ "mixed_precision": "no",
48
+ "not_cache_latents": false,
49
+ "hflip": false,
50
+ "local_rank": -1,
51
+ "concepts_list": [
52
+ {
53
+ "instance_prompt": "Bloodborne Style",
54
+ "instance_data_dir": "F:/AI/Data/Datasets/Soulsborne/Bloodborne/v1.1",
55
+ "class_prompt": "Horror Game Style",
56
+ "class_data_dir": "F:/AI/Data/Datasets/Regularisation/HorrorGameStyle"
57
+ }
58
+ ]
59
+ }
bloodbornestyle_showcase.jpg ADDED
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": 224,
3
+ "do_center_crop": true,
4
+ "do_convert_rgb": true,
5
+ "do_normalize": true,
6
+ "do_resize": true,
7
+ "feature_extractor_type": "CLIPFeatureExtractor",
8
+ "image_mean": [
9
+ 0.48145466,
10
+ 0.4578275,
11
+ 0.40821073
12
+ ],
13
+ "image_std": [
14
+ 0.26862954,
15
+ 0.26130258,
16
+ 0.27577711
17
+ ],
18
+ "resample": 3,
19
+ "size": 224
20
+ }
model_index.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.7.0.dev0",
4
+ "feature_extractor": [
5
+ "transformers",
6
+ "CLIPFeatureExtractor"
7
+ ],
8
+ "scheduler": [
9
+ "diffusers",
10
+ "DDIMScheduler"
11
+ ],
12
+ "text_encoder": [
13
+ "transformers",
14
+ "CLIPTextModel"
15
+ ],
16
+ "tokenizer": [
17
+ "transformers",
18
+ "CLIPTokenizer"
19
+ ],
20
+ "unet": [
21
+ "diffusers",
22
+ "UNet2DConditionModel"
23
+ ],
24
+ "vae": [
25
+ "diffusers",
26
+ "AutoencoderKL"
27
+ ]
28
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DDIMScheduler",
3
+ "_diffusers_version": "0.7.0.dev0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "set_alpha_to_one": false,
10
+ "steps_offset": 1,
11
+ "trained_betas": null
12
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle/v1-10k",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.24.0",
24
+ "vocab_size": 49408
25
+ }
text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c5b53caf2b8f820f23f1626fa35393c3e3f4035754f56de9a0921de823449e
3
+ size 246188833
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle/v1-10k\\tokenizer",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.7.0.dev0",
4
+ "_name_or_path": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle/v1-10k",
5
+ "act_fn": "silu",
6
+ "attention_head_dim": 8,
7
+ "block_out_channels": [
8
+ 320,
9
+ 640,
10
+ 1280,
11
+ 1280
12
+ ],
13
+ "center_input_sample": false,
14
+ "cross_attention_dim": 768,
15
+ "down_block_types": [
16
+ "CrossAttnDownBlock2D",
17
+ "CrossAttnDownBlock2D",
18
+ "CrossAttnDownBlock2D",
19
+ "DownBlock2D"
20
+ ],
21
+ "downsample_padding": 1,
22
+ "flip_sin_to_cos": true,
23
+ "freq_shift": 0,
24
+ "in_channels": 4,
25
+ "layers_per_block": 2,
26
+ "mid_block_scale_factor": 1,
27
+ "norm_eps": 1e-05,
28
+ "norm_num_groups": 32,
29
+ "out_channels": 4,
30
+ "sample_size": 64,
31
+ "up_block_types": [
32
+ "UpBlock2D",
33
+ "CrossAttnUpBlock2D",
34
+ "CrossAttnUpBlock2D",
35
+ "CrossAttnUpBlock2D"
36
+ ]
37
+ }
unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:201fd06a467146bd349af8b2f3431da9e36caa94bdbf024162f3efec25069115
3
+ size 1719334053
vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.7.0.dev0",
4
+ "_name_or_path": "F:/AI/Data/Outputs/Dreambooth/BloodborneStyle/v1-10k",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 512,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
vae/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36bb8e1b54aba3a0914eb35fba13dcb107e9f18d379d1df2158732cd4bf56a94
3
+ size 334711857