chlee10 commited on
Commit
7f72575
1 Parent(s): fa54530

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -5
README.md CHANGED
@@ -25,11 +25,28 @@ This model is a fine-tuned version of upstage/SOLAR-10.7B-v1.0
25
 
26
  The following hyperparameters were used during training:
27
 
28
- - batch_size = 16
29
- - num_epochs = 1
30
- - micro_batch = 1
31
- - cutoff_len = 4096
32
- - learning_rate = 4e-4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
  ## Framework versions
35
 
 
25
 
26
  The following hyperparameters were used during training:
27
 
28
+ ```python
29
+ python finetune.py \
30
+ --base_model PracticeLLM/Twice-KoSOLAR-16.1B-test \
31
+ --data-path kyujinpy/KOR-OpenOrca-Platypus-v3 \
32
+ --output_dir ./Twice-KoSOLAR-16.1B-instruct-test \
33
+ --batch_size 64 \
34
+ --micro_batch_size 1 \
35
+ --num_epochs 1 \
36
+ --learning_rate 3e-5 \
37
+ --cutoff_len 4096 \
38
+ --val_set_size 0 \
39
+ --lora_r 16 \
40
+ --lora_alpha 16 \
41
+ --lora_dropout 0.05 \
42
+ --lora_target_modules '[q_proj, k_proj, v_proj, o_proj, gate_proj, down_proj, up_proj, lm_head]' \
43
+ --train_on_inputs False \
44
+ --add_eos_token False \
45
+ --group_by_length False \
46
+ --prompt_template_name user_prompt \
47
+ --lr_scheduler 'cosine' \
48
+ #--warmup_steps 100 \
49
+ ```
50
 
51
  ## Framework versions
52