winglian commited on
Commit
5c0f5ec
1 Parent(s): df435ab

include axolotl config

Browse files
Files changed (1) hide show
  1. configs/axolotl.yml +65 -0
configs/axolotl.yml ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: teknium/OpenHermes-2.5-Mistral-7B
2
+ model_type: MistralForCausalLM
3
+ tokenizer_type: LlamaTokenizer
4
+ is_mistral_derived_model: true
5
+
6
+ load_in_8bit: false
7
+ load_in_4bit: true
8
+ strict: false
9
+
10
+ rl: true
11
+ datasets:
12
+ - path: winglian/no_robots_rlhf
13
+ split: train
14
+ test_datasets:
15
+ - path: winglian/no_robots_rlhf
16
+ split: test
17
+ dataset_prepared_path:
18
+ val_set_size: 0.0
19
+ output_dir: ./qlora-out
20
+
21
+ adapter: qlora
22
+ lora_model_dir:
23
+
24
+ sequence_len: 2048
25
+ sample_packing: false
26
+ pad_to_sequence_len: true
27
+
28
+ lora_r: 64
29
+ lora_alpha: 16
30
+ lora_dropout: 0.05
31
+ lora_target_modules:
32
+ lora_target_linear: true
33
+ lora_fan_in_fan_out:
34
+
35
+ wandb_project: oh-ipo
36
+ wandb_entity: oaaic
37
+ wandb_watch:
38
+ wandb_run_id:
39
+ wandb_log_model:
40
+
41
+ gradient_accumulation_steps: 4
42
+ micro_batch_size: 2
43
+ num_epochs: 3
44
+ optimizer: rmsprop
45
+ lr_scheduler: cosine
46
+ learning_rate: 5e-7
47
+
48
+ train_on_inputs: false
49
+ group_by_length: false
50
+ bf16: true
51
+ fp16: false
52
+ tf32: false
53
+
54
+ gradient_checkpointing: true
55
+ resume_from_checkpoint:
56
+ logging_steps: 1
57
+ flash_attention: true
58
+
59
+ warmup_steps: 10
60
+ eval_steps: 0.0
61
+ eval_table_size:
62
+ save_steps:
63
+ deepspeed: deepspeed/zero3.json
64
+ weight_decay: 0.0
65
+