louaaron commited on
Commit
a18d6d3
1 Parent(s): 42ef2d3

Create config.json

Browse files
Files changed (1) hide show
  1. config.json +64 -0
config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ngpus": 8,
3
+ "tokens": 50257,
4
+ "training": {
5
+ "batch_size": 512,
6
+ "accum": 1,
7
+ "n_iters": 1300001,
8
+ "snapshot_freq": 50000,
9
+ "log_freq": 50,
10
+ "eval_freq": 100,
11
+ "snapshot_freq_for_preemption": 10000,
12
+ "weight": "standard",
13
+ "snapshot_sampling": true,
14
+ "ema": 0.9999
15
+ },
16
+ "data": {
17
+ "train": "openwebtext",
18
+ "valid": "wikitext103",
19
+ "cache_dir": "data"
20
+ },
21
+ "graph": {
22
+ "type": "absorb",
23
+ "file": "data",
24
+ "report_all": false
25
+ },
26
+ "noise": {
27
+ "type": "loglinear",
28
+ "sigma_min": 0.0001,
29
+ "sigma_max": 20
30
+ },
31
+ "sampling": {
32
+ "predictor": "euler",
33
+ "steps": 128,
34
+ "noise_removal": true
35
+ },
36
+ "eval": {
37
+ "batch_size": 512,
38
+ "perplexity": true,
39
+ "perplexity_batch_size": 32
40
+ },
41
+ "optim": {
42
+ "weight_decay": 0,
43
+ "optimizer": "AdamW",
44
+ "lr": 0.0003,
45
+ "beta1": 0.9,
46
+ "beta2": 0.999,
47
+ "eps": 1e-08,
48
+ "warmup": 2500,
49
+ "grad_clip": 1.0
50
+ },
51
+ "model": {
52
+ "name": "small",
53
+ "type": "ddit",
54
+ "hidden_size": 768,
55
+ "cond_dim": 128,
56
+ "length": 1024,
57
+ "n_blocks": 12,
58
+ "n_heads": 12,
59
+ "scale_by_sigma": true,
60
+ "dropout": 0.1,
61
+ "score_strategy": 0
62
+ },
63
+ "work_dir": "absorb_small"
64
+ }