lodestones commited on
Commit
e251a3d
1 Parent(s): 20669c2

deleted 9 out of 10 checkpoints

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e101-terminal-snr-vpred-e74.safetensors +0 -3
  2. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e101-terminal-snr-vpred-e74.yaml +0 -71
  3. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e102-terminal-snr-vpred-e75.safetensors +0 -3
  4. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e103-terminal-snr-vpred-e76.safetensors +0 -3
  5. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e104-terminal-snr-vpred-e77.safetensors +0 -3
  6. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e104-terminal-snr-vpred-e77.yaml +0 -71
  7. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e105-terminal-snr-vpred-e78.safetensors +0 -3
  8. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e105-terminal-snr-vpred-e78.yaml +0 -71
  9. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e106-terminal-snr-vpred-e79.safetensors +0 -3
  10. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e106-terminal-snr-vpred-e79.yaml +0 -71
  11. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e107-terminal-snr-vpred-e80.safetensors +0 -3
  12. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e107-terminal-snr-vpred-e80.yaml +0 -71
  13. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e108-terminal-snr-vpred-e81.safetensors +0 -3
  14. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e108-terminal-snr-vpred-e81.yaml +0 -71
  15. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e109-terminal-snr-vpred-e82.safetensors +0 -3
  16. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e109-terminal-snr-vpred-e82.yaml +0 -71
  17. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e111-terminal-snr-vpred-e84.safetensors +0 -3
  18. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e111-terminal-snr-vpred-e84.yaml +0 -71
  19. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e112-terminal-snr-vpred-e85.safetensors +0 -3
  20. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e112-terminal-snr-vpred-e85.yaml +0 -71
  21. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e113-terminal-snr-vpred-e86.safetensors +0 -3
  22. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e113-terminal-snr-vpred-e86.yaml +0 -71
  23. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e114-terminal-snr-vpred-e87.safetensors +0 -3
  24. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e114-terminal-snr-vpred-e87.yaml +0 -71
  25. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e115-terminal-snr-vpred-e88.safetensors +0 -3
  26. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e115-terminal-snr-vpred-e88.yaml +0 -71
  27. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e116-terminal-snr-vpred-e89.safetensors +0 -3
  28. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e116-terminal-snr-vpred-e89.yaml +0 -71
  29. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e117-terminal-snr-vpred-e90.safetensors +0 -3
  30. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e117-terminal-snr-vpred-e90.yaml +0 -71
  31. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e118-terminal-snr-vpred-e91.safetensors +0 -3
  32. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e118-terminal-snr-vpred-e91.yaml +0 -71
  33. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e119-terminal-snr-vpred-e92.safetensors +0 -3
  34. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e119-terminal-snr-vpred-e92.yaml +0 -71
  35. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e121-terminal-snr-vpred-e94.safetensors +0 -3
  36. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e121-terminal-snr-vpred-e94.yaml +0 -71
  37. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e122-terminal-snr-vpred-e95.safetensors +0 -3
  38. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e122-terminal-snr-vpred-e95.yaml +0 -71
  39. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e123-terminal-snr-vpred-e96.safetensors +0 -3
  40. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e123-terminal-snr-vpred-e96.yaml +0 -71
  41. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e124-terminal-snr-vpred-e97.safetensors +0 -3
  42. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e124-terminal-snr-vpred-e97.yaml +0 -71
  43. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e125-terminal-snr-vpred-e98.safetensors +0 -3
  44. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e125-terminal-snr-vpred-e98.yaml +0 -71
  45. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e126-terminal-snr-vpred-e99.safetensors +0 -3
  46. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e126-terminal-snr-vpred-e99.yaml +0 -71
  47. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e127-terminal-snr-vpred-e100.safetensors +0 -3
  48. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e127-terminal-snr-vpred-e100.yaml +0 -71
  49. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e128-terminal-snr-vpred-e101.safetensors +0 -3
  50. fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e128-terminal-snr-vpred-e101.yaml +0 -71
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e101-terminal-snr-vpred-e74.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d723c4588f301a773f91b3310d4223f06960d983207727df09fbae2c219050e4
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e101-terminal-snr-vpred-e74.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e102-terminal-snr-vpred-e75.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:25ac25082f91abdfe979afb24e075750267c9cbc128c8e863e15d0c76834a5c2
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e103-terminal-snr-vpred-e76.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b354efeadf79327593d4404a5d647cf3aca8810dba99c2ca6d4ca622d0287bdc
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e104-terminal-snr-vpred-e77.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7787b7f8b8af195838dc89c71c199aa370803230114b54cd9e9437a5a7d2a589
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e104-terminal-snr-vpred-e77.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e105-terminal-snr-vpred-e78.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b32991906fe993b455270a92c292a37a3d3c9b976b86d7c24274b55576bb129
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e105-terminal-snr-vpred-e78.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e106-terminal-snr-vpred-e79.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:88776f7afefe1b7b73a9f96ace8ee2922ddbabe138f52cec9bf484ed3f6cbd77
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e106-terminal-snr-vpred-e79.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e107-terminal-snr-vpred-e80.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:955bf2ecab16ad43640fd8401128820a0e68de42b15e08c1b0642885fa998c24
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e107-terminal-snr-vpred-e80.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e108-terminal-snr-vpred-e81.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b231d122299ad15fecc5880c3d3de9e9fd173a96b51c058e9908953cc067c84
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e108-terminal-snr-vpred-e81.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e109-terminal-snr-vpred-e82.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8e9bbcb0410a106457f3213275544fc8e5bee63bf77937897db2e37028f2bd8
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e109-terminal-snr-vpred-e82.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e111-terminal-snr-vpred-e84.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b353fcaeaa9f59af85d967084947fca099dfc213513ff1c6a39a2923fc232e58
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e111-terminal-snr-vpred-e84.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e112-terminal-snr-vpred-e85.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:33495a702f03c170a03e7261ce7d7672f4e56b6026af7be3a12d3959631b3dc1
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e112-terminal-snr-vpred-e85.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e113-terminal-snr-vpred-e86.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:83590899bf85df34ab7d12eb549ed06390cfa8f25a9a5a06376841d684838df8
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e113-terminal-snr-vpred-e86.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e114-terminal-snr-vpred-e87.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1c2e4b8f862960a29317ea0eb5ef59c5a2da261f43e582af8bff11ba2300097
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e114-terminal-snr-vpred-e87.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e115-terminal-snr-vpred-e88.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:faf60d8865c71b50d1ee567e32f35a6e9e50d2ea96ab1a8a127a2aa080503ed1
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e115-terminal-snr-vpred-e88.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e116-terminal-snr-vpred-e89.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:598a16c31a6e1097ead3c723ba4a31772d83ad2fd8f059b02297f75cdd2f0809
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e116-terminal-snr-vpred-e89.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e117-terminal-snr-vpred-e90.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:54bd7346d97301b4099fd93f47528e0fc8cea21dbffcc01d4594b06570c2416b
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e117-terminal-snr-vpred-e90.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e118-terminal-snr-vpred-e91.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d44a686137b706e4ca68f1d14ea5803c53af24a31ee7f6cd30fbc7cf68527d06
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e118-terminal-snr-vpred-e91.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e119-terminal-snr-vpred-e92.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:33740e1e6d4910009609973a42e8e61d25f24a889ed4ef1ac231f12a1d46b572
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e119-terminal-snr-vpred-e92.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e121-terminal-snr-vpred-e94.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcecb38ca755543c55d6fed3315578c46bf8d692d435d85a754308e7339001b5
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e121-terminal-snr-vpred-e94.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e122-terminal-snr-vpred-e95.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e219e5faf4605f722853e35e27e9e20a8512b4a8976e63b1d742db3ec446516
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e122-terminal-snr-vpred-e95.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e123-terminal-snr-vpred-e96.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b75c1fc2081090c8f38710d7d7d9eb2365f5111e9ebf22aeb915adce37f3ef37
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e123-terminal-snr-vpred-e96.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e124-terminal-snr-vpred-e97.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e4dc29b0dfdd67241f544e42f90d87656a74f0c24dafe0086ad31b92b910605
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e124-terminal-snr-vpred-e97.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e125-terminal-snr-vpred-e98.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:073209b80b2bf6a0507468dfafc18e2aa02010d14d74f2f792a12914622f2ffc
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e125-terminal-snr-vpred-e98.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e126-terminal-snr-vpred-e99.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe6ea6dfc4a9a1421e8b45b9abe7e66d61255bcac9a56846f67c2243d73f1899
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e126-terminal-snr-vpred-e99.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e127-terminal-snr-vpred-e100.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8233509f21b5185438831281b04717bf90a0a1d78e373eb4a3c3239947702857
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e127-terminal-snr-vpred-e100.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e128-terminal-snr-vpred-e101.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4f3162e08cb9fe0be9568cb6b9316ee988dfaff3ce57e760171c71cf803803c
3
- size 4265096996
 
 
 
 
fluffyrock-1088-megares-terminal-snr-vpred/fluffyrock-576-704-832-960-1088-lion-low-lr-e128-terminal-snr-vpred-e101.yaml DELETED
@@ -1,71 +0,0 @@
1
- model:
2
- base_learning_rate: 1.0e-04
3
- target: ldm.models.diffusion.ddpm.LatentDiffusion
4
- params:
5
- parameterization: "v"
6
- linear_start: 0.00085
7
- linear_end: 0.0120
8
- num_timesteps_cond: 1
9
- log_every_t: 200
10
- timesteps: 1000
11
- first_stage_key: "jpg"
12
- cond_stage_key: "txt"
13
- image_size: 64
14
- channels: 4
15
- cond_stage_trainable: false # Note: different from the one we trained before
16
- conditioning_key: crossattn
17
- monitor: val/loss_simple_ema
18
- scale_factor: 0.18215
19
- use_ema: False
20
-
21
- scheduler_config: # 10000 warmup steps
22
- target: ldm.lr_scheduler.LambdaLinearScheduler
23
- params:
24
- warm_up_steps: [ 10000 ]
25
- cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases
26
- f_start: [ 1.e-6 ]
27
- f_max: [ 1. ]
28
- f_min: [ 1. ]
29
-
30
- unet_config:
31
- target: ldm.modules.diffusionmodules.openaimodel.UNetModel
32
- params:
33
- image_size: 32 # unused
34
- in_channels: 4
35
- out_channels: 4
36
- model_channels: 320
37
- attention_resolutions: [ 4, 2, 1 ]
38
- num_res_blocks: 2
39
- channel_mult: [ 1, 2, 4, 4 ]
40
- num_heads: 8
41
- use_spatial_transformer: True
42
- transformer_depth: 1
43
- context_dim: 768
44
- use_checkpoint: True
45
- legacy: False
46
-
47
- first_stage_config:
48
- target: ldm.models.autoencoder.AutoencoderKL
49
- params:
50
- embed_dim: 4
51
- monitor: val/rec_loss
52
- ddconfig:
53
- double_z: true
54
- z_channels: 4
55
- resolution: 256
56
- in_channels: 3
57
- out_ch: 3
58
- ch: 128
59
- ch_mult:
60
- - 1
61
- - 2
62
- - 4
63
- - 4
64
- num_res_blocks: 2
65
- attn_resolutions: []
66
- dropout: 0.0
67
- lossconfig:
68
- target: torch.nn.Identity
69
-
70
- cond_stage_config:
71
- target: ldm.modules.encoders.modules.FrozenCLIPEmbedder