Kaludi commited on
Commit
d5fa843
1 Parent(s): 1d22b0e

ARTificialJourneyV1-768px

Browse files
.gitattributes CHANGED
@@ -32,3 +32,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ Images1.jpg filter=lfs diff=lfs merge=lfs -text
36
+ Images2.jpg filter=lfs diff=lfs merge=lfs -text
37
+ samples/001516.53be1b9b.2447414055.png filter=lfs diff=lfs merge=lfs -text
38
+ samples/001531.df3197d2.1766343097.png filter=lfs diff=lfs merge=lfs -text
39
+ samples/001555.64a554e5.2103898698.png filter=lfs diff=lfs merge=lfs -text
ARTificialJourneyV1-768px.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ffd08c5854ee55a2de2bc7896f9499cbded08984069b8f4b24eb10f2a523030
3
+ size 2132791380
ARTificialJourneyV1-768px.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:311df55bd0abb236faa13cd0046863ac3d6ce318e60b8d8907bea70df6732f14
3
+ size 2132625431
Images1.jpg ADDED

Git LFS Details

  • SHA256: 9639d954d949e2770eeb749a9d6e775a8b84c8c03e9f99872d8ed1f98ca9bbaf
  • Pointer size: 132 Bytes
  • Size of remote file: 1.53 MB
Images2.jpg ADDED

Git LFS Details

  • SHA256: d0617cfd095bbe467015685f9117894049752abbf3c064975e78b7aa088ad8d7
  • Pointer size: 132 Bytes
  • Size of remote file: 1.43 MB
README.md CHANGED
@@ -1,3 +1,73 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: creativeml-openrail-m
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ tags:
5
+ - stable-diffusion
6
+ - stable-diffusion-diffusers
7
+ - text-to-image
8
+ - art
9
+ - artistic
10
+ - diffusers
11
+ - midjourney
12
+ - artificial
13
+ - artificial-journey
14
+ - journey
15
+ - portrait
16
+ - art
17
+ - diffusion
18
+ - photorealistic
19
+ inference: true
20
  license: creativeml-openrail-m
21
  ---
22
+
23
+ # ARTificialJourney-1.0 768X768
24
+
25
+ ![img](https://huggingface.co/)
26
+
27
+ This is an AI model trained on ~100 hand picked 768X768 images targeted to get the best close up portrait pictures, as well as amazing looking landscapes. Full body portrait will be improved in the next update of this model. Please use the keywords **"artificial-journey style"** before or after the prompt. Also be sure to add more detailed text to the prompt to get the best results. Resolutions 1024x768px, 768x1024px, and 768x768 work best for this model.
28
+
29
+ *.safetensors file was generated using the new sd-webui-model-converter.*
30
+
31
+ # Examples (As is images, no upsclaing or face improvements were used)
32
+ ![img](https://huggingface.co/)
33
+
34
+ **Have fun generating!**
35
+
36
+ ### CKPT & Safetensors Download
37
+
38
+ [Download ARTificialJourneyV1-768px.ckpt) (2.9GB)](https://huggingface.co/)
39
+ [Download ARTificialJourneyV1-768px.safetensors) (2.9GB)](https://huggingface.co/)
40
+
41
+ ### 🧨 Diffusers
42
+
43
+ This model can be used just like any other Stable Diffusion model. For more information,
44
+ please have a look at the [Stable Diffusion Pipeline](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
45
+
46
+ ```python
47
+ from diffusers import StableDiffusionPipeline, DPMSolverMultistepScheduler
48
+ import torch
49
+
50
+ prompt = (
51
+ "artificial-journey style portrait of male dark magician, d & d, dark eyeliner, intricate, elegant, highly detailed, digital painting, artstation, concept art, matte, sharp focus, illustration")
52
+
53
+ model_id = "Kaludi/ARTificialJourney-1.0"
54
+ pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
55
+ pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
56
+ pipe = pipe.to("cuda")
57
+
58
+ image = pipe(prompt, num_inference_steps=30).images[0]
59
+
60
+ image.save("./result.jpg")
61
+ ```
62
+
63
+ ![img](https://huggingface.co/)
64
+
65
+ ## License
66
+
67
+ This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
68
+ The CreativeML OpenRAIL License specifies:
69
+
70
+ 1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
71
+ 2. The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
72
+ 3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
73
+ [Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
args.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "pretrained_model_name_or_path": "runwayml/stable-diffusion-v1-5",
3
+ "pretrained_vae_name_or_path": "stabilityai/sd-vae-ft-mse",
4
+ "revision": "fp16",
5
+ "tokenizer_name": null,
6
+ "instance_data_dir": null,
7
+ "class_data_dir": null,
8
+ "instance_prompt": null,
9
+ "class_prompt": null,
10
+ "save_sample_prompt": "artificial-journey style",
11
+ "save_sample_negative_prompt": null,
12
+ "n_save_sample": 4,
13
+ "save_guidance_scale": 7.5,
14
+ "save_infer_steps": 50,
15
+ "pad_tokens": false,
16
+ "with_prior_preservation": true,
17
+ "prior_loss_weight": 1.0,
18
+ "num_class_images": 1000,
19
+ "output_dir": "/content/drive/MyDrive/stable_diffusion_weights/AJSv2-768",
20
+ "seed": 1337,
21
+ "resolution": 768,
22
+ "center_crop": false,
23
+ "train_text_encoder": true,
24
+ "train_batch_size": 1,
25
+ "sample_batch_size": 4,
26
+ "num_train_epochs": 9,
27
+ "max_train_steps": 9000,
28
+ "gradient_accumulation_steps": 1,
29
+ "gradient_checkpointing": false,
30
+ "learning_rate": 1e-06,
31
+ "scale_lr": false,
32
+ "lr_scheduler": "polynomial",
33
+ "lr_warmup_steps": 0,
34
+ "use_8bit_adam": false,
35
+ "adam_beta1": 0.9,
36
+ "adam_beta2": 0.999,
37
+ "adam_weight_decay": 0.01,
38
+ "adam_epsilon": 1e-08,
39
+ "max_grad_norm": 1.0,
40
+ "push_to_hub": false,
41
+ "hub_token": null,
42
+ "hub_model_id": null,
43
+ "logging_dir": "logs",
44
+ "log_interval": 10,
45
+ "save_interval": 10000,
46
+ "save_min_steps": 0,
47
+ "mixed_precision": "fp16",
48
+ "not_cache_latents": false,
49
+ "hflip": false,
50
+ "local_rank": -1,
51
+ "concepts_list": [
52
+ {
53
+ "instance_prompt": "artificial-journey style",
54
+ "class_prompt": "artwork style",
55
+ "instance_data_dir": "/content/drive/MyDrive/custom/AJSv2-768-instancee",
56
+ "class_data_dir": "/content/drive/MyDrive/custom/artwork-style"
57
+ }
58
+ ]
59
+ }
model_index.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.9.0",
4
+ "requires_safety_checker": false,
5
+ "scheduler": [
6
+ "diffusers",
7
+ "DDIMScheduler"
8
+ ],
9
+ "text_encoder": [
10
+ "transformers",
11
+ "CLIPTextModel"
12
+ ],
13
+ "tokenizer": [
14
+ "transformers",
15
+ "CLIPTokenizer"
16
+ ],
17
+ "unet": [
18
+ "diffusers",
19
+ "UNet2DConditionModel"
20
+ ],
21
+ "vae": [
22
+ "diffusers",
23
+ "AutoencoderKL"
24
+ ]
25
+ }
samples/001509.80088957.696590176.png ADDED
samples/001510.328c40e1.152343323.png ADDED
samples/001516.53be1b9b.2447414055.png ADDED

Git LFS Details

  • SHA256: bcb692f426c47f2913705cc93f1ead63c97c94a3a211ff99d7ad63f474b792b8
  • Pointer size: 132 Bytes
  • Size of remote file: 1.03 MB
samples/001520.283fae42.696613010.png ADDED
samples/001524.6a09234e.3630974078.png ADDED
samples/001526.2abba9a2.4228509653.png ADDED
samples/001527.9445437b.3509991735.png ADDED
samples/001531.df3197d2.1766343097.png ADDED

Git LFS Details

  • SHA256: 1b676f9b35123fe61501f9fc4fa1fee2e38f5ba96be385b18eb4a4eff356cc5b
  • Pointer size: 132 Bytes
  • Size of remote file: 1.14 MB
samples/001532.2788076b.2558804922.png ADDED
samples/001537.0e844b7f.428005861.png ADDED
samples/001538.dc86706b.2040109847.png ADDED
samples/001539.8cfd9302.1705631662.png ADDED
samples/001540.51633820.2752599392.png ADDED
samples/001555.64a554e5.2103898698.png ADDED

Git LFS Details

  • SHA256: 6cec83ffc36a7b0984229d010f578d5b8f68c050a29d29ab0a8f657105c5bf2f
  • Pointer size: 132 Bytes
  • Size of remote file: 1 MB
samples/001565.987286f6.2786882365.png ADDED
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DDIMScheduler",
3
+ "_diffusers_version": "0.9.0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "steps_offset": 1,
12
+ "trained_betas": null
13
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "runwayml/stable-diffusion-v1-5",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 12,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 768,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.25.1",
24
+ "vocab_size": 49408
25
+ }
text_encoder/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3946f8a5a2d77dae3cbd817dad7c6747b3f42353d8545438f4c644f7367f4acc
3
+ size 492309793
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|endoftext|>",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "/root/.cache/huggingface/diffusers/models--runwayml--stable-diffusion-v1-5/snapshots/ded79e214aa69e42c24d3f5ac14b76d568679cc2/tokenizer",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.9.0",
4
+ "_name_or_path": "runwayml/stable-diffusion-v1-5",
5
+ "act_fn": "silu",
6
+ "attention_head_dim": 8,
7
+ "block_out_channels": [
8
+ 320,
9
+ 640,
10
+ 1280,
11
+ 1280
12
+ ],
13
+ "center_input_sample": false,
14
+ "cross_attention_dim": 768,
15
+ "down_block_types": [
16
+ "CrossAttnDownBlock2D",
17
+ "CrossAttnDownBlock2D",
18
+ "CrossAttnDownBlock2D",
19
+ "DownBlock2D"
20
+ ],
21
+ "downsample_padding": 1,
22
+ "dual_cross_attention": false,
23
+ "flip_sin_to_cos": true,
24
+ "freq_shift": 0,
25
+ "in_channels": 4,
26
+ "layers_per_block": 2,
27
+ "mid_block_scale_factor": 1,
28
+ "norm_eps": 1e-05,
29
+ "norm_num_groups": 32,
30
+ "num_class_embeds": null,
31
+ "only_cross_attention": false,
32
+ "out_channels": 4,
33
+ "sample_size": 64,
34
+ "up_block_types": [
35
+ "UpBlock2D",
36
+ "CrossAttnUpBlock2D",
37
+ "CrossAttnUpBlock2D",
38
+ "CrossAttnUpBlock2D"
39
+ ],
40
+ "use_linear_projection": false
41
+ }
vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.9.0",
4
+ "_name_or_path": "stabilityai/sd-vae-ft-mse",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 256,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
vae/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36bb8e1b54aba3a0914eb35fba13dcb107e9f18d379d1df2158732cd4bf56a94
3
+ size 334711857