ARTificialJourneyV1-768px
Browse files- .gitattributes +5 -0
- ARTificialJourneyV1-768px.ckpt +3 -0
- ARTificialJourneyV1-768px.safetensors +3 -0
- Images1.jpg +3 -0
- Images2.jpg +3 -0
- README.md +70 -0
- args.json +59 -0
- model_index.json +25 -0
- samples/001509.80088957.696590176.png +0 -0
- samples/001510.328c40e1.152343323.png +0 -0
- samples/001516.53be1b9b.2447414055.png +3 -0
- samples/001520.283fae42.696613010.png +0 -0
- samples/001524.6a09234e.3630974078.png +0 -0
- samples/001526.2abba9a2.4228509653.png +0 -0
- samples/001527.9445437b.3509991735.png +0 -0
- samples/001531.df3197d2.1766343097.png +3 -0
- samples/001532.2788076b.2558804922.png +0 -0
- samples/001537.0e844b7f.428005861.png +0 -0
- samples/001538.dc86706b.2040109847.png +0 -0
- samples/001539.8cfd9302.1705631662.png +0 -0
- samples/001540.51633820.2752599392.png +0 -0
- samples/001555.64a554e5.2103898698.png +3 -0
- samples/001565.987286f6.2786882365.png +0 -0
- scheduler/scheduler_config.json +13 -0
- text_encoder/config.json +25 -0
- text_encoder/pytorch_model.bin +3 -0
- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +24 -0
- tokenizer/tokenizer_config.json +34 -0
- tokenizer/vocab.json +0 -0
- unet/config.json +41 -0
- vae/config.json +30 -0
- vae/diffusion_pytorch_model.bin +3 -0
.gitattributes
CHANGED
@@ -32,3 +32,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
Images1.jpg filter=lfs diff=lfs merge=lfs -text
|
36 |
+
Images2.jpg filter=lfs diff=lfs merge=lfs -text
|
37 |
+
samples/001516.53be1b9b.2447414055.png filter=lfs diff=lfs merge=lfs -text
|
38 |
+
samples/001531.df3197d2.1766343097.png filter=lfs diff=lfs merge=lfs -text
|
39 |
+
samples/001555.64a554e5.2103898698.png filter=lfs diff=lfs merge=lfs -text
|
ARTificialJourneyV1-768px.ckpt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3ffd08c5854ee55a2de2bc7896f9499cbded08984069b8f4b24eb10f2a523030
|
3 |
+
size 2132791380
|
ARTificialJourneyV1-768px.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:311df55bd0abb236faa13cd0046863ac3d6ce318e60b8d8907bea70df6732f14
|
3 |
+
size 2132625431
|
Images1.jpg
ADDED
Git LFS Details
|
Images2.jpg
ADDED
Git LFS Details
|
README.md
CHANGED
@@ -1,3 +1,73 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: creativeml-openrail-m
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
tags:
|
5 |
+
- stable-diffusion
|
6 |
+
- stable-diffusion-diffusers
|
7 |
+
- text-to-image
|
8 |
+
- art
|
9 |
+
- artistic
|
10 |
+
- diffusers
|
11 |
+
- midjourney
|
12 |
+
- artificial
|
13 |
+
- artificial-journey
|
14 |
+
- journey
|
15 |
+
- portrait
|
16 |
+
- art
|
17 |
+
- diffusion
|
18 |
+
- photorealistic
|
19 |
+
inference: true
|
20 |
license: creativeml-openrail-m
|
21 |
---
|
22 |
+
|
23 |
+
# ARTificialJourney-1.0 768X768
|
24 |
+
|
25 |
+
![img](https://huggingface.co/)
|
26 |
+
|
27 |
+
This is an AI model trained on ~100 hand picked 768X768 images targeted to get the best close up portrait pictures, as well as amazing looking landscapes. Full body portrait will be improved in the next update of this model. Please use the keywords **"artificial-journey style"** before or after the prompt. Also be sure to add more detailed text to the prompt to get the best results. Resolutions 1024x768px, 768x1024px, and 768x768 work best for this model.
|
28 |
+
|
29 |
+
*.safetensors file was generated using the new sd-webui-model-converter.*
|
30 |
+
|
31 |
+
# Examples (As is images, no upsclaing or face improvements were used)
|
32 |
+
![img](https://huggingface.co/)
|
33 |
+
|
34 |
+
**Have fun generating!**
|
35 |
+
|
36 |
+
### CKPT & Safetensors Download
|
37 |
+
|
38 |
+
[Download ARTificialJourneyV1-768px.ckpt) (2.9GB)](https://huggingface.co/)
|
39 |
+
[Download ARTificialJourneyV1-768px.safetensors) (2.9GB)](https://huggingface.co/)
|
40 |
+
|
41 |
+
### 🧨 Diffusers
|
42 |
+
|
43 |
+
This model can be used just like any other Stable Diffusion model. For more information,
|
44 |
+
please have a look at the [Stable Diffusion Pipeline](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion).
|
45 |
+
|
46 |
+
```python
|
47 |
+
from diffusers import StableDiffusionPipeline, DPMSolverMultistepScheduler
|
48 |
+
import torch
|
49 |
+
|
50 |
+
prompt = (
|
51 |
+
"artificial-journey style portrait of male dark magician, d & d, dark eyeliner, intricate, elegant, highly detailed, digital painting, artstation, concept art, matte, sharp focus, illustration")
|
52 |
+
|
53 |
+
model_id = "Kaludi/ARTificialJourney-1.0"
|
54 |
+
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
|
55 |
+
pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
|
56 |
+
pipe = pipe.to("cuda")
|
57 |
+
|
58 |
+
image = pipe(prompt, num_inference_steps=30).images[0]
|
59 |
+
|
60 |
+
image.save("./result.jpg")
|
61 |
+
```
|
62 |
+
|
63 |
+
![img](https://huggingface.co/)
|
64 |
+
|
65 |
+
## License
|
66 |
+
|
67 |
+
This model is open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
|
68 |
+
The CreativeML OpenRAIL License specifies:
|
69 |
+
|
70 |
+
1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
|
71 |
+
2. The authors claims no rights on the outputs you generate, you are free to use them and are accountable for their use which must not go against the provisions set in the license
|
72 |
+
3. You may re-distribute the weights and use the model commercially and/or as a service. If you do, please be aware you have to include the same use restrictions as the ones in the license and share a copy of the CreativeML OpenRAIL-M to all your users (please read the license entirely and carefully)
|
73 |
+
[Please read the full license here](https://huggingface.co/spaces/CompVis/stable-diffusion-license)
|
args.json
ADDED
@@ -0,0 +1,59 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"pretrained_model_name_or_path": "runwayml/stable-diffusion-v1-5",
|
3 |
+
"pretrained_vae_name_or_path": "stabilityai/sd-vae-ft-mse",
|
4 |
+
"revision": "fp16",
|
5 |
+
"tokenizer_name": null,
|
6 |
+
"instance_data_dir": null,
|
7 |
+
"class_data_dir": null,
|
8 |
+
"instance_prompt": null,
|
9 |
+
"class_prompt": null,
|
10 |
+
"save_sample_prompt": "artificial-journey style",
|
11 |
+
"save_sample_negative_prompt": null,
|
12 |
+
"n_save_sample": 4,
|
13 |
+
"save_guidance_scale": 7.5,
|
14 |
+
"save_infer_steps": 50,
|
15 |
+
"pad_tokens": false,
|
16 |
+
"with_prior_preservation": true,
|
17 |
+
"prior_loss_weight": 1.0,
|
18 |
+
"num_class_images": 1000,
|
19 |
+
"output_dir": "/content/drive/MyDrive/stable_diffusion_weights/AJSv2-768",
|
20 |
+
"seed": 1337,
|
21 |
+
"resolution": 768,
|
22 |
+
"center_crop": false,
|
23 |
+
"train_text_encoder": true,
|
24 |
+
"train_batch_size": 1,
|
25 |
+
"sample_batch_size": 4,
|
26 |
+
"num_train_epochs": 9,
|
27 |
+
"max_train_steps": 9000,
|
28 |
+
"gradient_accumulation_steps": 1,
|
29 |
+
"gradient_checkpointing": false,
|
30 |
+
"learning_rate": 1e-06,
|
31 |
+
"scale_lr": false,
|
32 |
+
"lr_scheduler": "polynomial",
|
33 |
+
"lr_warmup_steps": 0,
|
34 |
+
"use_8bit_adam": false,
|
35 |
+
"adam_beta1": 0.9,
|
36 |
+
"adam_beta2": 0.999,
|
37 |
+
"adam_weight_decay": 0.01,
|
38 |
+
"adam_epsilon": 1e-08,
|
39 |
+
"max_grad_norm": 1.0,
|
40 |
+
"push_to_hub": false,
|
41 |
+
"hub_token": null,
|
42 |
+
"hub_model_id": null,
|
43 |
+
"logging_dir": "logs",
|
44 |
+
"log_interval": 10,
|
45 |
+
"save_interval": 10000,
|
46 |
+
"save_min_steps": 0,
|
47 |
+
"mixed_precision": "fp16",
|
48 |
+
"not_cache_latents": false,
|
49 |
+
"hflip": false,
|
50 |
+
"local_rank": -1,
|
51 |
+
"concepts_list": [
|
52 |
+
{
|
53 |
+
"instance_prompt": "artificial-journey style",
|
54 |
+
"class_prompt": "artwork style",
|
55 |
+
"instance_data_dir": "/content/drive/MyDrive/custom/AJSv2-768-instancee",
|
56 |
+
"class_data_dir": "/content/drive/MyDrive/custom/artwork-style"
|
57 |
+
}
|
58 |
+
]
|
59 |
+
}
|
model_index.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.9.0",
|
4 |
+
"requires_safety_checker": false,
|
5 |
+
"scheduler": [
|
6 |
+
"diffusers",
|
7 |
+
"DDIMScheduler"
|
8 |
+
],
|
9 |
+
"text_encoder": [
|
10 |
+
"transformers",
|
11 |
+
"CLIPTextModel"
|
12 |
+
],
|
13 |
+
"tokenizer": [
|
14 |
+
"transformers",
|
15 |
+
"CLIPTokenizer"
|
16 |
+
],
|
17 |
+
"unet": [
|
18 |
+
"diffusers",
|
19 |
+
"UNet2DConditionModel"
|
20 |
+
],
|
21 |
+
"vae": [
|
22 |
+
"diffusers",
|
23 |
+
"AutoencoderKL"
|
24 |
+
]
|
25 |
+
}
|
samples/001509.80088957.696590176.png
ADDED
samples/001510.328c40e1.152343323.png
ADDED
samples/001516.53be1b9b.2447414055.png
ADDED
Git LFS Details
|
samples/001520.283fae42.696613010.png
ADDED
samples/001524.6a09234e.3630974078.png
ADDED
samples/001526.2abba9a2.4228509653.png
ADDED
samples/001527.9445437b.3509991735.png
ADDED
samples/001531.df3197d2.1766343097.png
ADDED
Git LFS Details
|
samples/001532.2788076b.2558804922.png
ADDED
samples/001537.0e844b7f.428005861.png
ADDED
samples/001538.dc86706b.2040109847.png
ADDED
samples/001539.8cfd9302.1705631662.png
ADDED
samples/001540.51633820.2752599392.png
ADDED
samples/001555.64a554e5.2103898698.png
ADDED
Git LFS Details
|
samples/001565.987286f6.2786882365.png
ADDED
scheduler/scheduler_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "DDIMScheduler",
|
3 |
+
"_diffusers_version": "0.9.0",
|
4 |
+
"beta_end": 0.012,
|
5 |
+
"beta_schedule": "scaled_linear",
|
6 |
+
"beta_start": 0.00085,
|
7 |
+
"clip_sample": false,
|
8 |
+
"num_train_timesteps": 1000,
|
9 |
+
"prediction_type": "epsilon",
|
10 |
+
"set_alpha_to_one": false,
|
11 |
+
"steps_offset": 1,
|
12 |
+
"trained_betas": null
|
13 |
+
}
|
text_encoder/config.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPTextModel"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 0,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_factor": 1.0,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
+
"layer_norm_eps": 1e-05,
|
16 |
+
"max_position_embeddings": 77,
|
17 |
+
"model_type": "clip_text_model",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
+
"pad_token_id": 1,
|
21 |
+
"projection_dim": 768,
|
22 |
+
"torch_dtype": "float32",
|
23 |
+
"transformers_version": "4.25.1",
|
24 |
+
"vocab_size": 49408
|
25 |
+
}
|
text_encoder/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3946f8a5a2d77dae3cbd817dad7c6747b3f42353d8545438f4c644f7367f4acc
|
3 |
+
size 492309793
|
tokenizer/merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer/special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|startoftext|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|endoftext|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "<|endoftext|>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<|endoftext|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": true,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
tokenizer/tokenizer_config.json
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"bos_token": {
|
4 |
+
"__type": "AddedToken",
|
5 |
+
"content": "<|startoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": true,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false
|
10 |
+
},
|
11 |
+
"do_lower_case": true,
|
12 |
+
"eos_token": {
|
13 |
+
"__type": "AddedToken",
|
14 |
+
"content": "<|endoftext|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": true,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false
|
19 |
+
},
|
20 |
+
"errors": "replace",
|
21 |
+
"model_max_length": 77,
|
22 |
+
"name_or_path": "/root/.cache/huggingface/diffusers/models--runwayml--stable-diffusion-v1-5/snapshots/ded79e214aa69e42c24d3f5ac14b76d568679cc2/tokenizer",
|
23 |
+
"pad_token": "<|endoftext|>",
|
24 |
+
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
+
"tokenizer_class": "CLIPTokenizer",
|
26 |
+
"unk_token": {
|
27 |
+
"__type": "AddedToken",
|
28 |
+
"content": "<|endoftext|>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": true,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false
|
33 |
+
}
|
34 |
+
}
|
tokenizer/vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
unet/config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.9.0",
|
4 |
+
"_name_or_path": "runwayml/stable-diffusion-v1-5",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"attention_head_dim": 8,
|
7 |
+
"block_out_channels": [
|
8 |
+
320,
|
9 |
+
640,
|
10 |
+
1280,
|
11 |
+
1280
|
12 |
+
],
|
13 |
+
"center_input_sample": false,
|
14 |
+
"cross_attention_dim": 768,
|
15 |
+
"down_block_types": [
|
16 |
+
"CrossAttnDownBlock2D",
|
17 |
+
"CrossAttnDownBlock2D",
|
18 |
+
"CrossAttnDownBlock2D",
|
19 |
+
"DownBlock2D"
|
20 |
+
],
|
21 |
+
"downsample_padding": 1,
|
22 |
+
"dual_cross_attention": false,
|
23 |
+
"flip_sin_to_cos": true,
|
24 |
+
"freq_shift": 0,
|
25 |
+
"in_channels": 4,
|
26 |
+
"layers_per_block": 2,
|
27 |
+
"mid_block_scale_factor": 1,
|
28 |
+
"norm_eps": 1e-05,
|
29 |
+
"norm_num_groups": 32,
|
30 |
+
"num_class_embeds": null,
|
31 |
+
"only_cross_attention": false,
|
32 |
+
"out_channels": 4,
|
33 |
+
"sample_size": 64,
|
34 |
+
"up_block_types": [
|
35 |
+
"UpBlock2D",
|
36 |
+
"CrossAttnUpBlock2D",
|
37 |
+
"CrossAttnUpBlock2D",
|
38 |
+
"CrossAttnUpBlock2D"
|
39 |
+
],
|
40 |
+
"use_linear_projection": false
|
41 |
+
}
|
vae/config.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.9.0",
|
4 |
+
"_name_or_path": "stabilityai/sd-vae-ft-mse",
|
5 |
+
"act_fn": "silu",
|
6 |
+
"block_out_channels": [
|
7 |
+
128,
|
8 |
+
256,
|
9 |
+
512,
|
10 |
+
512
|
11 |
+
],
|
12 |
+
"down_block_types": [
|
13 |
+
"DownEncoderBlock2D",
|
14 |
+
"DownEncoderBlock2D",
|
15 |
+
"DownEncoderBlock2D",
|
16 |
+
"DownEncoderBlock2D"
|
17 |
+
],
|
18 |
+
"in_channels": 3,
|
19 |
+
"latent_channels": 4,
|
20 |
+
"layers_per_block": 2,
|
21 |
+
"norm_num_groups": 32,
|
22 |
+
"out_channels": 3,
|
23 |
+
"sample_size": 256,
|
24 |
+
"up_block_types": [
|
25 |
+
"UpDecoderBlock2D",
|
26 |
+
"UpDecoderBlock2D",
|
27 |
+
"UpDecoderBlock2D",
|
28 |
+
"UpDecoderBlock2D"
|
29 |
+
]
|
30 |
+
}
|
vae/diffusion_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36bb8e1b54aba3a0914eb35fba13dcb107e9f18d379d1df2158732cd4bf56a94
|
3 |
+
size 334711857
|