norabelrose commited on
Commit
b07ca63
1 Parent(s): 55ceecf

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sae": {"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": true}, "batch_size": 4, "grad_acc_steps": 2, "micro_acc_steps": 1, "lr": null, "lr_warmup_steps": 1000, "auxk_alpha": 0.0, "dead_feature_threshold": 10000000, "hookpoints": ["layers.29.mlp"], "layers": [], "layer_stride": 1, "distribute_modules": false, "save_every": 1000, "log_to_wandb": false, "run_name": "llama-64x", "wandb_log_frequency": 1, "model": "meta-llama/Meta-Llama-3.1-8B", "dataset": "/home/fslcollab366/sae/rpj-pretokenized.hf", "split": "train", "ctx_len": 2048, "hf_token": null, "load_in_8bit": false, "max_examples": null, "resume": false, "seed": 42, "data_preprocessing_num_proc": 64}
layers.23.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": true, "d_in": 4096}
layers.23.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3fd2305f163cf63c80b41280fd9f832e58068c60cc882c3bc903a0f8b3b3839
3
+ size 8590999896
layers.29.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"expansion_factor": 64, "normalize_decoder": true, "num_latents": 0, "k": 32, "multi_topk": true, "d_in": 4096}
layers.29.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a70ad3da4509295b2d8dba0f41cddd8a49d114d940f1933e5dcadfc9759b1707
3
+ size 8590999896