xu3kev commited on
Commit
16b46a9
1 Parent(s): f20a385

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/README.md +151 -0
  2. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/adapter_config.json +34 -0
  3. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/adapter_model.bin +3 -0
  4. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/README.md +202 -0
  5. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/adapter_config.json +34 -0
  6. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/adapter_model.safetensors +3 -0
  7. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  8. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  9. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  10. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
  11. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt +3 -0
  12. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt +3 -0
  13. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt +3 -0
  14. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt +3 -0
  15. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/mp_rank_00_model_states.pt +3 -0
  16. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/latest +1 -0
  17. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_0.pth +3 -0
  18. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_1.pth +3 -0
  19. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_2.pth +3 -0
  20. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_3.pth +3 -0
  21. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_4.pth +3 -0
  22. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_5.pth +3 -0
  23. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_6.pth +3 -0
  24. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_7.pth +3 -0
  25. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/scheduler.pt +3 -0
  26. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/special_tokens_map.json +23 -0
  27. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/tokenizer.json +0 -0
  28. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/tokenizer_config.json +193 -0
  29. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/trainer_state.json +2055 -0
  30. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/training_args.bin +3 -0
  31. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/zero_to_fp32.py +592 -0
  32. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/README.md +202 -0
  33. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/adapter_config.json +34 -0
  34. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/adapter_model.safetensors +3 -0
  35. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  36. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  37. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
  38. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
  39. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt +3 -0
  40. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt +3 -0
  41. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt +3 -0
  42. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt +3 -0
  43. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/mp_rank_00_model_states.pt +3 -0
  44. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/latest +1 -0
  45. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_0.pth +3 -0
  46. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_1.pth +3 -0
  47. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_2.pth +3 -0
  48. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_3.pth +3 -0
  49. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_4.pth +3 -0
  50. gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_5.pth +3 -0
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/README.md ADDED
@@ -0,0 +1,151 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ base_model: deepseek-ai/deepseek-coder-33b-instruct
7
+ model-index:
8
+ - name: lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
16
+ <details><summary>See axolotl config</summary>
17
+
18
+ axolotl version: `0.4.0`
19
+ ```yaml
20
+ adapter: lora
21
+ base_model: deepseek-ai/deepseek-coder-33b-instruct
22
+ bf16: auto
23
+ dataset_prepared_path: ./logo_ds_preprocess_list_gpt35
24
+ datasets:
25
+ - path: ../logo/fix_synthetic_int_images_data.jsonl
26
+ type:
27
+ field_instruction: input
28
+ field_output: output
29
+ format: '### Instruction:
30
+
31
+ {input}
32
+
33
+ ### Response:
34
+
35
+ '
36
+ no_input_format: '{instruction}'
37
+ debug: null
38
+ deepspeed: ./deepspeed_configs/zero2.json
39
+ early_stopping_patience: null
40
+ eval_sample_packing: true
41
+ evals_per_epoch: 4
42
+ flash_attention: true
43
+ fp16: null
44
+ fsdp: null
45
+ fsdp_config: null
46
+ gradient_accumulation_steps: 4
47
+ gradient_checkpointing: true
48
+ group_by_length: false
49
+ is_llama_derived_model: true
50
+ learning_rate: 0.0002
51
+ load_in_4bit: false
52
+ load_in_8bit: true
53
+ local_rank: null
54
+ logging_steps: 1
55
+ lora_alpha: 2048
56
+ lora_dropout: 0.05
57
+ lora_fan_in_fan_out: null
58
+ lora_model_dir: null
59
+ lora_r: 1024
60
+ lora_target_linear: true
61
+ lr_scheduler: cosine
62
+ micro_batch_size: 2
63
+ model_type: AutoModelForCausalLM
64
+ num_epochs: 2
65
+ optimizer: adamw_bnb_8bit
66
+ output_dir: ./lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024
67
+ pad_to_sequence_len: true
68
+ resume_from_checkpoint: null
69
+ s2_attention: null
70
+ sample_packing: true
71
+ saves_per_epoch: 1
72
+ sequence_len: 1800
73
+ special_tokens:
74
+ bos_token: "<\uFF5Cbegin\u2581of\u2581sentence\uFF5C>"
75
+ eos_token: <|EOT|>
76
+ strict: true
77
+ tf32: false
78
+ tokenizer_type: AutoTokenizer
79
+ train_on_inputs: false
80
+ val_set_size: 0.05
81
+ wandb_entity: null
82
+ wandb_log_model: null
83
+ wandb_name: logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024
84
+ wandb_project: pbe-axo
85
+ wandb_watch: null
86
+ warmup_steps: 20
87
+ weight_decay: 0.0
88
+ xformers_attention: null
89
+
90
+ ```
91
+
92
+ </details><br>
93
+
94
+ # lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024
95
+
96
+ This model is a fine-tuned version of [deepseek-ai/deepseek-coder-33b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-33b-instruct) on the None dataset.
97
+ It achieves the following results on the evaluation set:
98
+ - Loss: 0.2402
99
+
100
+ ## Model description
101
+
102
+ More information needed
103
+
104
+ ## Intended uses & limitations
105
+
106
+ More information needed
107
+
108
+ ## Training and evaluation data
109
+
110
+ More information needed
111
+
112
+ ## Training procedure
113
+
114
+ ### Training hyperparameters
115
+
116
+ The following hyperparameters were used during training:
117
+ - learning_rate: 0.0002
118
+ - train_batch_size: 2
119
+ - eval_batch_size: 2
120
+ - seed: 42
121
+ - distributed_type: multi-GPU
122
+ - num_devices: 8
123
+ - gradient_accumulation_steps: 4
124
+ - total_train_batch_size: 64
125
+ - total_eval_batch_size: 16
126
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
127
+ - lr_scheduler_type: cosine
128
+ - lr_scheduler_warmup_steps: 20
129
+ - num_epochs: 2
130
+
131
+ ### Training results
132
+
133
+ | Training Loss | Epoch | Step | Validation Loss |
134
+ |:-------------:|:-----:|:----:|:---------------:|
135
+ | 2.206 | 0.0 | 1 | 2.1888 |
136
+ | 0.3412 | 0.25 | 72 | 0.3284 |
137
+ | 0.3118 | 0.5 | 144 | 0.3046 |
138
+ | 0.3142 | 0.75 | 216 | 0.2882 |
139
+ | 0.2749 | 1.01 | 288 | 0.2703 |
140
+ | 0.2169 | 1.24 | 360 | 0.2578 |
141
+ | 0.2324 | 1.49 | 432 | 0.2465 |
142
+ | 0.2489 | 1.74 | 504 | 0.2402 |
143
+
144
+
145
+ ### Framework versions
146
+
147
+ - PEFT 0.10.0
148
+ - Transformers 4.40.0.dev0
149
+ - Pytorch 2.1.2+cu121
150
+ - Datasets 2.15.0
151
+ - Tokenizers 0.15.0
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "deepseek-ai/deepseek-coder-33b-instruct",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 2048,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 1024,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "up_proj",
24
+ "o_proj",
25
+ "v_proj",
26
+ "down_proj",
27
+ "gate_proj",
28
+ "q_proj",
29
+ "k_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:764413cd9a473461607a6c24c13da7641136e46d972808225a2137a1e4895e49
3
+ size 15765481838
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ base_model: deepseek-ai/deepseek-coder-33b-instruct
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.10.0
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "deepseek-ai/deepseek-coder-33b-instruct",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 2048,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 1024,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "up_proj",
24
+ "o_proj",
25
+ "v_proj",
26
+ "down_proj",
27
+ "gate_proj",
28
+ "q_proj",
29
+ "k_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a03b36d3d4afde18bf44e734a5d6337ce7dbc7b1fa5ec731f4bc0450a58991b
3
+ size 15765462656
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9feefedad199a1633ff9608ed199e6513e5aaa301308029c40c3318f844d2fe6
3
+ size 11824020348
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d08f2225aaf457ac72ee3cdee372eae09cbdc335278c7c0d2bd6ba7002039a5
3
+ size 11824020540
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdea969b8a65230105943f78737055889b6d81a876efec04da2faf039d74b2e5
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e15dd21bd46e4b67ad2a95cb5e4c9c81a8ce4c01f04ce69ab6201635f01c0a38
3
+ size 11824020028
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d33be829b249550afca75a04b25ae5b36added085d99536917f9082fbe1ffa06
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61cf0293ba683518c4f458a3d68707fe15e16cb6b1cadabf964f91c2e877c6f2
3
+ size 11824020604
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46ceb78a3e0e991b5eee1b63bba8e0696539f81e02b788f377e9d8ee4db22112
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b69aa13b0e43ac97f4143dd83bb4da77b27865fc51e12f6b3a1a5512a645df8
3
+ size 11824020028
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/global_step286/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47742a0ac78410ff9819898f9e5ba7b8f21cb3b3efbc68a5a49dff1e0d4f97c3
3
+ size 15781541500
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step286
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1dd04671815f401b65c2256c3a66a1577f5a4c769d3544600a3f0fa4107f9b29
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f7f4a7bef94094980cd20dca58daf484299b6612523ebfd0e8b22e270819c80
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ac480c64a722006274d4f75c3e3a20169c4274a81671344ddb4f17399478ecd
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a8917b8f3a77a3f15cd87c2a6c125163b0d8d991900d7bf47b96c85ecc71ee6
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b112f636062fcc0e3140c7cd7f3b6fa90936bc05e1616df8fe59cd56805ce8
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8f237aff8959ddf960f7a06056851a8f6ec5100ded74f0c031d241ec6d82586
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30618de4c07c655e4df3b33c8eb718d5cff20de4e2d3209d9356e17be4699c9b
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7edf280ffabc9ca167b2c7dda571d4ff3059de3356c5e840691db168a405590
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9956ec6e7856961b226e71519552e2f3230b72ee11ef507e852a485da664ed63
3
+ size 1064
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin▁of▁sentence|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|EOT|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|end▁of▁sentence|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/tokenizer_config.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "32000": {
6
+ "content": "õ",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": false
12
+ },
13
+ "32001": {
14
+ "content": "÷",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": false
20
+ },
21
+ "32002": {
22
+ "content": "Á",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": false
28
+ },
29
+ "32003": {
30
+ "content": "ý",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": false
36
+ },
37
+ "32004": {
38
+ "content": "À",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": false
44
+ },
45
+ "32005": {
46
+ "content": "ÿ",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": false
52
+ },
53
+ "32006": {
54
+ "content": "ø",
55
+ "lstrip": false,
56
+ "normalized": true,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": false
60
+ },
61
+ "32007": {
62
+ "content": "ú",
63
+ "lstrip": false,
64
+ "normalized": true,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": false
68
+ },
69
+ "32008": {
70
+ "content": "þ",
71
+ "lstrip": false,
72
+ "normalized": true,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": false
76
+ },
77
+ "32009": {
78
+ "content": "ü",
79
+ "lstrip": false,
80
+ "normalized": true,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": false
84
+ },
85
+ "32010": {
86
+ "content": "ù",
87
+ "lstrip": false,
88
+ "normalized": true,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": false
92
+ },
93
+ "32011": {
94
+ "content": "ö",
95
+ "lstrip": false,
96
+ "normalized": true,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": false
100
+ },
101
+ "32012": {
102
+ "content": "û",
103
+ "lstrip": false,
104
+ "normalized": true,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": false
108
+ },
109
+ "32013": {
110
+ "content": "<|begin▁of▁sentence|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "32014": {
118
+ "content": "<|end▁of▁sentence|>",
119
+ "lstrip": false,
120
+ "normalized": true,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": true
124
+ },
125
+ "32015": {
126
+ "content": "<|fim▁hole|>",
127
+ "lstrip": false,
128
+ "normalized": true,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "32016": {
134
+ "content": "<|fim▁begin|>",
135
+ "lstrip": false,
136
+ "normalized": true,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "32017": {
142
+ "content": "<|fim▁end|>",
143
+ "lstrip": false,
144
+ "normalized": true,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "32018": {
150
+ "content": "<pad>",
151
+ "lstrip": false,
152
+ "normalized": true,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "32019": {
158
+ "content": "<|User|>",
159
+ "lstrip": false,
160
+ "normalized": true,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "32020": {
166
+ "content": "<|Assistant|>",
167
+ "lstrip": false,
168
+ "normalized": true,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "32021": {
174
+ "content": "<|EOT|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": true
180
+ }
181
+ },
182
+ "bos_token": "<|begin▁of▁sentence|>",
183
+ "chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{bos_token}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
184
+ "clean_up_tokenization_spaces": false,
185
+ "eos_token": "<|EOT|>",
186
+ "legacy": true,
187
+ "model_max_length": 16384,
188
+ "pad_token": "<|end▁of▁sentence|>",
189
+ "sp_model_kwargs": {},
190
+ "tokenizer_class": "LlamaTokenizer",
191
+ "unk_token": null,
192
+ "use_default_system_prompt": false
193
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/trainer_state.json ADDED
@@ -0,0 +1,2055 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.9991266375545852,
5
+ "eval_steps": 72,
6
+ "global_step": 286,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "grad_norm": 6.735667801609769,
14
+ "learning_rate": 1e-05,
15
+ "loss": 2.206,
16
+ "step": 1
17
+ },
18
+ {
19
+ "epoch": 0.0,
20
+ "eval_loss": 2.1888480186462402,
21
+ "eval_runtime": 121.3992,
22
+ "eval_samples_per_second": 12.01,
23
+ "eval_steps_per_second": 0.758,
24
+ "step": 1
25
+ },
26
+ {
27
+ "epoch": 0.01,
28
+ "grad_norm": 7.273415210005453,
29
+ "learning_rate": 2e-05,
30
+ "loss": 2.3157,
31
+ "step": 2
32
+ },
33
+ {
34
+ "epoch": 0.01,
35
+ "grad_norm": 6.065138409852167,
36
+ "learning_rate": 3e-05,
37
+ "loss": 1.9397,
38
+ "step": 3
39
+ },
40
+ {
41
+ "epoch": 0.01,
42
+ "grad_norm": 7.4973462814136935,
43
+ "learning_rate": 4e-05,
44
+ "loss": 1.5277,
45
+ "step": 4
46
+ },
47
+ {
48
+ "epoch": 0.02,
49
+ "grad_norm": 4.005938889538725,
50
+ "learning_rate": 5e-05,
51
+ "loss": 1.1011,
52
+ "step": 5
53
+ },
54
+ {
55
+ "epoch": 0.02,
56
+ "grad_norm": 3.175050395655731,
57
+ "learning_rate": 6e-05,
58
+ "loss": 0.7586,
59
+ "step": 6
60
+ },
61
+ {
62
+ "epoch": 0.02,
63
+ "grad_norm": 4.4243840624051884,
64
+ "learning_rate": 7e-05,
65
+ "loss": 0.6707,
66
+ "step": 7
67
+ },
68
+ {
69
+ "epoch": 0.03,
70
+ "grad_norm": 1.8307542722309151,
71
+ "learning_rate": 8e-05,
72
+ "loss": 0.5738,
73
+ "step": 8
74
+ },
75
+ {
76
+ "epoch": 0.03,
77
+ "grad_norm": 1.67357492034038,
78
+ "learning_rate": 9e-05,
79
+ "loss": 0.518,
80
+ "step": 9
81
+ },
82
+ {
83
+ "epoch": 0.03,
84
+ "grad_norm": 1.2285284335828708,
85
+ "learning_rate": 0.0001,
86
+ "loss": 0.4759,
87
+ "step": 10
88
+ },
89
+ {
90
+ "epoch": 0.04,
91
+ "grad_norm": 1.4025765653440176,
92
+ "learning_rate": 0.00011000000000000002,
93
+ "loss": 0.482,
94
+ "step": 11
95
+ },
96
+ {
97
+ "epoch": 0.04,
98
+ "grad_norm": 1.764363673866666,
99
+ "learning_rate": 0.00012,
100
+ "loss": 0.4528,
101
+ "step": 12
102
+ },
103
+ {
104
+ "epoch": 0.05,
105
+ "grad_norm": 1.1465976015812096,
106
+ "learning_rate": 0.00013000000000000002,
107
+ "loss": 0.4309,
108
+ "step": 13
109
+ },
110
+ {
111
+ "epoch": 0.05,
112
+ "grad_norm": 0.905812519633333,
113
+ "learning_rate": 0.00014,
114
+ "loss": 0.43,
115
+ "step": 14
116
+ },
117
+ {
118
+ "epoch": 0.05,
119
+ "grad_norm": 0.9012554260598487,
120
+ "learning_rate": 0.00015000000000000001,
121
+ "loss": 0.4018,
122
+ "step": 15
123
+ },
124
+ {
125
+ "epoch": 0.06,
126
+ "grad_norm": 0.7370091534412341,
127
+ "learning_rate": 0.00016,
128
+ "loss": 0.4183,
129
+ "step": 16
130
+ },
131
+ {
132
+ "epoch": 0.06,
133
+ "grad_norm": 0.6613195203569469,
134
+ "learning_rate": 0.00017,
135
+ "loss": 0.4095,
136
+ "step": 17
137
+ },
138
+ {
139
+ "epoch": 0.06,
140
+ "grad_norm": 0.6284882714369103,
141
+ "learning_rate": 0.00018,
142
+ "loss": 0.3891,
143
+ "step": 18
144
+ },
145
+ {
146
+ "epoch": 0.07,
147
+ "grad_norm": 0.5831826872671497,
148
+ "learning_rate": 0.00019,
149
+ "loss": 0.3803,
150
+ "step": 19
151
+ },
152
+ {
153
+ "epoch": 0.07,
154
+ "grad_norm": 0.5688120818577218,
155
+ "learning_rate": 0.0002,
156
+ "loss": 0.3833,
157
+ "step": 20
158
+ },
159
+ {
160
+ "epoch": 0.07,
161
+ "grad_norm": 0.688607797221884,
162
+ "learning_rate": 0.00019999838046468694,
163
+ "loss": 0.361,
164
+ "step": 21
165
+ },
166
+ {
167
+ "epoch": 0.08,
168
+ "grad_norm": 0.6179301706803293,
169
+ "learning_rate": 0.00019999352191120556,
170
+ "loss": 0.4071,
171
+ "step": 22
172
+ },
173
+ {
174
+ "epoch": 0.08,
175
+ "grad_norm": 0.6502118022336852,
176
+ "learning_rate": 0.00019998542449692791,
177
+ "loss": 0.4067,
178
+ "step": 23
179
+ },
180
+ {
181
+ "epoch": 0.08,
182
+ "grad_norm": 0.6204600910475001,
183
+ "learning_rate": 0.00019997408848413493,
184
+ "loss": 0.4129,
185
+ "step": 24
186
+ },
187
+ {
188
+ "epoch": 0.09,
189
+ "grad_norm": 0.6682295708321836,
190
+ "learning_rate": 0.00019995951424000807,
191
+ "loss": 0.3766,
192
+ "step": 25
193
+ },
194
+ {
195
+ "epoch": 0.09,
196
+ "grad_norm": 0.5531393394120152,
197
+ "learning_rate": 0.0001999417022366174,
198
+ "loss": 0.3925,
199
+ "step": 26
200
+ },
201
+ {
202
+ "epoch": 0.09,
203
+ "grad_norm": 0.5642129459679449,
204
+ "learning_rate": 0.0001999206530509063,
205
+ "loss": 0.3884,
206
+ "step": 27
207
+ },
208
+ {
209
+ "epoch": 0.1,
210
+ "grad_norm": 0.7545523765388521,
211
+ "learning_rate": 0.00019989636736467278,
212
+ "loss": 0.3711,
213
+ "step": 28
214
+ },
215
+ {
216
+ "epoch": 0.1,
217
+ "grad_norm": 0.7034250678747912,
218
+ "learning_rate": 0.0001998688459645473,
219
+ "loss": 0.3887,
220
+ "step": 29
221
+ },
222
+ {
223
+ "epoch": 0.1,
224
+ "grad_norm": 0.7272817271797393,
225
+ "learning_rate": 0.00019983808974196753,
226
+ "loss": 0.3851,
227
+ "step": 30
228
+ },
229
+ {
230
+ "epoch": 0.11,
231
+ "grad_norm": 0.5202650753093606,
232
+ "learning_rate": 0.00019980409969314915,
233
+ "loss": 0.3901,
234
+ "step": 31
235
+ },
236
+ {
237
+ "epoch": 0.11,
238
+ "grad_norm": 0.5035407640856523,
239
+ "learning_rate": 0.00019976687691905393,
240
+ "loss": 0.4019,
241
+ "step": 32
242
+ },
243
+ {
244
+ "epoch": 0.12,
245
+ "grad_norm": 0.7239776078180935,
246
+ "learning_rate": 0.00019972642262535377,
247
+ "loss": 0.3977,
248
+ "step": 33
249
+ },
250
+ {
251
+ "epoch": 0.12,
252
+ "grad_norm": 0.423267661940076,
253
+ "learning_rate": 0.00019968273812239184,
254
+ "loss": 0.3645,
255
+ "step": 34
256
+ },
257
+ {
258
+ "epoch": 0.12,
259
+ "grad_norm": 0.5184324303598159,
260
+ "learning_rate": 0.00019963582482514,
261
+ "loss": 0.3898,
262
+ "step": 35
263
+ },
264
+ {
265
+ "epoch": 0.13,
266
+ "grad_norm": 0.5703301361701498,
267
+ "learning_rate": 0.00019958568425315314,
268
+ "loss": 0.3572,
269
+ "step": 36
270
+ },
271
+ {
272
+ "epoch": 0.13,
273
+ "grad_norm": 0.7476520741504539,
274
+ "learning_rate": 0.00019953231803051974,
275
+ "loss": 0.3683,
276
+ "step": 37
277
+ },
278
+ {
279
+ "epoch": 0.13,
280
+ "grad_norm": 0.6166161415520125,
281
+ "learning_rate": 0.00019947572788580947,
282
+ "loss": 0.3654,
283
+ "step": 38
284
+ },
285
+ {
286
+ "epoch": 0.14,
287
+ "grad_norm": 0.6803246942355095,
288
+ "learning_rate": 0.0001994159156520171,
289
+ "loss": 0.3736,
290
+ "step": 39
291
+ },
292
+ {
293
+ "epoch": 0.14,
294
+ "grad_norm": 1.2865471082113682,
295
+ "learning_rate": 0.00019935288326650312,
296
+ "loss": 0.3654,
297
+ "step": 40
298
+ },
299
+ {
300
+ "epoch": 0.14,
301
+ "grad_norm": 0.49764384998341465,
302
+ "learning_rate": 0.000199286632770931,
303
+ "loss": 0.3958,
304
+ "step": 41
305
+ },
306
+ {
307
+ "epoch": 0.15,
308
+ "grad_norm": 0.48420075389585926,
309
+ "learning_rate": 0.00019921716631120107,
310
+ "loss": 0.366,
311
+ "step": 42
312
+ },
313
+ {
314
+ "epoch": 0.15,
315
+ "grad_norm": 0.3800370753861025,
316
+ "learning_rate": 0.00019914448613738106,
317
+ "loss": 0.3767,
318
+ "step": 43
319
+ },
320
+ {
321
+ "epoch": 0.15,
322
+ "grad_norm": 0.41063916070172707,
323
+ "learning_rate": 0.00019906859460363307,
324
+ "loss": 0.3625,
325
+ "step": 44
326
+ },
327
+ {
328
+ "epoch": 0.16,
329
+ "grad_norm": 0.7014230687297667,
330
+ "learning_rate": 0.00019898949416813756,
331
+ "loss": 0.3568,
332
+ "step": 45
333
+ },
334
+ {
335
+ "epoch": 0.16,
336
+ "grad_norm": 0.3857820514409519,
337
+ "learning_rate": 0.00019890718739301344,
338
+ "loss": 0.3689,
339
+ "step": 46
340
+ },
341
+ {
342
+ "epoch": 0.16,
343
+ "grad_norm": 0.3907216906088691,
344
+ "learning_rate": 0.0001988216769442353,
345
+ "loss": 0.3643,
346
+ "step": 47
347
+ },
348
+ {
349
+ "epoch": 0.17,
350
+ "grad_norm": 0.3862868314916432,
351
+ "learning_rate": 0.00019873296559154698,
352
+ "loss": 0.3289,
353
+ "step": 48
354
+ },
355
+ {
356
+ "epoch": 0.17,
357
+ "grad_norm": 0.4079296163127676,
358
+ "learning_rate": 0.0001986410562083718,
359
+ "loss": 0.3519,
360
+ "step": 49
361
+ },
362
+ {
363
+ "epoch": 0.17,
364
+ "grad_norm": 0.4565925285652506,
365
+ "learning_rate": 0.00019854595177171968,
366
+ "loss": 0.3545,
367
+ "step": 50
368
+ },
369
+ {
370
+ "epoch": 0.18,
371
+ "grad_norm": 0.4151471704043633,
372
+ "learning_rate": 0.00019844765536209045,
373
+ "loss": 0.3439,
374
+ "step": 51
375
+ },
376
+ {
377
+ "epoch": 0.18,
378
+ "grad_norm": 0.5325434590008968,
379
+ "learning_rate": 0.0001983461701633742,
380
+ "loss": 0.3465,
381
+ "step": 52
382
+ },
383
+ {
384
+ "epoch": 0.19,
385
+ "grad_norm": 0.40099633154653697,
386
+ "learning_rate": 0.00019824149946274825,
387
+ "loss": 0.3471,
388
+ "step": 53
389
+ },
390
+ {
391
+ "epoch": 0.19,
392
+ "grad_norm": 0.463564547141941,
393
+ "learning_rate": 0.00019813364665057052,
394
+ "loss": 0.3522,
395
+ "step": 54
396
+ },
397
+ {
398
+ "epoch": 0.19,
399
+ "grad_norm": 0.39803220667592676,
400
+ "learning_rate": 0.0001980226152202697,
401
+ "loss": 0.3242,
402
+ "step": 55
403
+ },
404
+ {
405
+ "epoch": 0.2,
406
+ "grad_norm": 0.43886397818532347,
407
+ "learning_rate": 0.00019790840876823232,
408
+ "loss": 0.3621,
409
+ "step": 56
410
+ },
411
+ {
412
+ "epoch": 0.2,
413
+ "grad_norm": 0.3860678927424882,
414
+ "learning_rate": 0.00019779103099368595,
415
+ "loss": 0.3521,
416
+ "step": 57
417
+ },
418
+ {
419
+ "epoch": 0.2,
420
+ "grad_norm": 0.4176549714108472,
421
+ "learning_rate": 0.00019767048569857963,
422
+ "loss": 0.359,
423
+ "step": 58
424
+ },
425
+ {
426
+ "epoch": 0.21,
427
+ "grad_norm": 0.38573299348714896,
428
+ "learning_rate": 0.00019754677678746063,
429
+ "loss": 0.3682,
430
+ "step": 59
431
+ },
432
+ {
433
+ "epoch": 0.21,
434
+ "grad_norm": 0.3715542195121731,
435
+ "learning_rate": 0.00019741990826734794,
436
+ "loss": 0.3503,
437
+ "step": 60
438
+ },
439
+ {
440
+ "epoch": 0.21,
441
+ "grad_norm": 0.5325807845038206,
442
+ "learning_rate": 0.00019728988424760248,
443
+ "loss": 0.3595,
444
+ "step": 61
445
+ },
446
+ {
447
+ "epoch": 0.22,
448
+ "grad_norm": 0.5535862965718638,
449
+ "learning_rate": 0.00019715670893979414,
450
+ "loss": 0.3488,
451
+ "step": 62
452
+ },
453
+ {
454
+ "epoch": 0.22,
455
+ "grad_norm": 0.40196725210678685,
456
+ "learning_rate": 0.0001970203866575652,
457
+ "loss": 0.343,
458
+ "step": 63
459
+ },
460
+ {
461
+ "epoch": 0.22,
462
+ "grad_norm": 0.39577392291226743,
463
+ "learning_rate": 0.00019688092181649065,
464
+ "loss": 0.3279,
465
+ "step": 64
466
+ },
467
+ {
468
+ "epoch": 0.23,
469
+ "grad_norm": 0.4317509436848305,
470
+ "learning_rate": 0.00019673831893393518,
471
+ "loss": 0.3734,
472
+ "step": 65
473
+ },
474
+ {
475
+ "epoch": 0.23,
476
+ "grad_norm": 0.4257454638359886,
477
+ "learning_rate": 0.00019659258262890683,
478
+ "loss": 0.3464,
479
+ "step": 66
480
+ },
481
+ {
482
+ "epoch": 0.23,
483
+ "grad_norm": 0.41845639027353915,
484
+ "learning_rate": 0.0001964437176219075,
485
+ "loss": 0.3538,
486
+ "step": 67
487
+ },
488
+ {
489
+ "epoch": 0.24,
490
+ "grad_norm": 0.36818626195219933,
491
+ "learning_rate": 0.00019629172873477995,
492
+ "loss": 0.3392,
493
+ "step": 68
494
+ },
495
+ {
496
+ "epoch": 0.24,
497
+ "grad_norm": 0.35715662384747937,
498
+ "learning_rate": 0.00019613662089055148,
499
+ "loss": 0.344,
500
+ "step": 69
501
+ },
502
+ {
503
+ "epoch": 0.24,
504
+ "grad_norm": 0.3928528389137624,
505
+ "learning_rate": 0.00019597839911327476,
506
+ "loss": 0.3398,
507
+ "step": 70
508
+ },
509
+ {
510
+ "epoch": 0.25,
511
+ "grad_norm": 0.35373447030288285,
512
+ "learning_rate": 0.0001958170685278649,
513
+ "loss": 0.3306,
514
+ "step": 71
515
+ },
516
+ {
517
+ "epoch": 0.25,
518
+ "grad_norm": 0.4081232823310914,
519
+ "learning_rate": 0.0001956526343599335,
520
+ "loss": 0.3412,
521
+ "step": 72
522
+ },
523
+ {
524
+ "epoch": 0.25,
525
+ "eval_loss": 0.32837027311325073,
526
+ "eval_runtime": 122.3443,
527
+ "eval_samples_per_second": 11.917,
528
+ "eval_steps_per_second": 0.752,
529
+ "step": 72
530
+ },
531
+ {
532
+ "epoch": 0.26,
533
+ "grad_norm": 0.40010484870998836,
534
+ "learning_rate": 0.00019548510193561939,
535
+ "loss": 0.3582,
536
+ "step": 73
537
+ },
538
+ {
539
+ "epoch": 0.26,
540
+ "grad_norm": 0.40224368508421876,
541
+ "learning_rate": 0.00019531447668141608,
542
+ "loss": 0.3444,
543
+ "step": 74
544
+ },
545
+ {
546
+ "epoch": 0.26,
547
+ "grad_norm": 0.37113340065591016,
548
+ "learning_rate": 0.00019514076412399612,
549
+ "loss": 0.3379,
550
+ "step": 75
551
+ },
552
+ {
553
+ "epoch": 0.27,
554
+ "grad_norm": 0.3774332536653395,
555
+ "learning_rate": 0.00019496396989003193,
556
+ "loss": 0.3308,
557
+ "step": 76
558
+ },
559
+ {
560
+ "epoch": 0.27,
561
+ "grad_norm": 0.42643318661506063,
562
+ "learning_rate": 0.00019478409970601358,
563
+ "loss": 0.383,
564
+ "step": 77
565
+ },
566
+ {
567
+ "epoch": 0.27,
568
+ "grad_norm": 0.3802734954943659,
569
+ "learning_rate": 0.00019460115939806338,
570
+ "loss": 0.3088,
571
+ "step": 78
572
+ },
573
+ {
574
+ "epoch": 0.28,
575
+ "grad_norm": 0.43650689829786143,
576
+ "learning_rate": 0.00019441515489174708,
577
+ "loss": 0.3268,
578
+ "step": 79
579
+ },
580
+ {
581
+ "epoch": 0.28,
582
+ "grad_norm": 0.37882459881558067,
583
+ "learning_rate": 0.00019422609221188207,
584
+ "loss": 0.3309,
585
+ "step": 80
586
+ },
587
+ {
588
+ "epoch": 0.28,
589
+ "grad_norm": 0.3410760526657497,
590
+ "learning_rate": 0.000194033977482342,
591
+ "loss": 0.3321,
592
+ "step": 81
593
+ },
594
+ {
595
+ "epoch": 0.29,
596
+ "grad_norm": 0.3741929987128578,
597
+ "learning_rate": 0.0001938388169258587,
598
+ "loss": 0.3408,
599
+ "step": 82
600
+ },
601
+ {
602
+ "epoch": 0.29,
603
+ "grad_norm": 0.3330884930823651,
604
+ "learning_rate": 0.00019364061686382044,
605
+ "loss": 0.314,
606
+ "step": 83
607
+ },
608
+ {
609
+ "epoch": 0.29,
610
+ "grad_norm": 0.3798944427357396,
611
+ "learning_rate": 0.00019343938371606712,
612
+ "loss": 0.3688,
613
+ "step": 84
614
+ },
615
+ {
616
+ "epoch": 0.3,
617
+ "grad_norm": 0.35343720425122044,
618
+ "learning_rate": 0.00019323512400068262,
619
+ "loss": 0.3424,
620
+ "step": 85
621
+ },
622
+ {
623
+ "epoch": 0.3,
624
+ "grad_norm": 0.30331935354303496,
625
+ "learning_rate": 0.0001930278443337833,
626
+ "loss": 0.3328,
627
+ "step": 86
628
+ },
629
+ {
630
+ "epoch": 0.3,
631
+ "grad_norm": 0.3482731799478109,
632
+ "learning_rate": 0.00019281755142930407,
633
+ "loss": 0.3388,
634
+ "step": 87
635
+ },
636
+ {
637
+ "epoch": 0.31,
638
+ "grad_norm": 0.3586206396987296,
639
+ "learning_rate": 0.00019260425209878052,
640
+ "loss": 0.3342,
641
+ "step": 88
642
+ },
643
+ {
644
+ "epoch": 0.31,
645
+ "grad_norm": 0.2909096774596904,
646
+ "learning_rate": 0.0001923879532511287,
647
+ "loss": 0.309,
648
+ "step": 89
649
+ },
650
+ {
651
+ "epoch": 0.31,
652
+ "grad_norm": 0.32004191437042734,
653
+ "learning_rate": 0.00019216866189242097,
654
+ "loss": 0.3183,
655
+ "step": 90
656
+ },
657
+ {
658
+ "epoch": 0.32,
659
+ "grad_norm": 0.33385732810933305,
660
+ "learning_rate": 0.00019194638512565936,
661
+ "loss": 0.3283,
662
+ "step": 91
663
+ },
664
+ {
665
+ "epoch": 0.32,
666
+ "grad_norm": 0.3345269597654594,
667
+ "learning_rate": 0.00019172113015054532,
668
+ "loss": 0.3214,
669
+ "step": 92
670
+ },
671
+ {
672
+ "epoch": 0.32,
673
+ "grad_norm": 0.3662026773156966,
674
+ "learning_rate": 0.0001914929042632466,
675
+ "loss": 0.3378,
676
+ "step": 93
677
+ },
678
+ {
679
+ "epoch": 0.33,
680
+ "grad_norm": 0.3334365948036441,
681
+ "learning_rate": 0.00019126171485616084,
682
+ "loss": 0.3255,
683
+ "step": 94
684
+ },
685
+ {
686
+ "epoch": 0.33,
687
+ "grad_norm": 0.3621308765797038,
688
+ "learning_rate": 0.00019102756941767623,
689
+ "loss": 0.3318,
690
+ "step": 95
691
+ },
692
+ {
693
+ "epoch": 0.34,
694
+ "grad_norm": 0.3399305671206559,
695
+ "learning_rate": 0.0001907904755319289,
696
+ "loss": 0.3396,
697
+ "step": 96
698
+ },
699
+ {
700
+ "epoch": 0.34,
701
+ "grad_norm": 0.32397556376658004,
702
+ "learning_rate": 0.00019055044087855726,
703
+ "loss": 0.3221,
704
+ "step": 97
705
+ },
706
+ {
707
+ "epoch": 0.34,
708
+ "grad_norm": 0.3545140871791968,
709
+ "learning_rate": 0.00019030747323245327,
710
+ "loss": 0.3337,
711
+ "step": 98
712
+ },
713
+ {
714
+ "epoch": 0.35,
715
+ "grad_norm": 0.3633137760162914,
716
+ "learning_rate": 0.00019006158046351059,
717
+ "loss": 0.3111,
718
+ "step": 99
719
+ },
720
+ {
721
+ "epoch": 0.35,
722
+ "grad_norm": 0.38270678325554436,
723
+ "learning_rate": 0.0001898127705363696,
724
+ "loss": 0.3366,
725
+ "step": 100
726
+ },
727
+ {
728
+ "epoch": 0.35,
729
+ "grad_norm": 0.3432620117707175,
730
+ "learning_rate": 0.00018956105151015965,
731
+ "loss": 0.3379,
732
+ "step": 101
733
+ },
734
+ {
735
+ "epoch": 0.36,
736
+ "grad_norm": 0.4622433156323975,
737
+ "learning_rate": 0.00018930643153823777,
738
+ "loss": 0.3046,
739
+ "step": 102
740
+ },
741
+ {
742
+ "epoch": 0.36,
743
+ "grad_norm": 0.3719507661138522,
744
+ "learning_rate": 0.00018904891886792463,
745
+ "loss": 0.3161,
746
+ "step": 103
747
+ },
748
+ {
749
+ "epoch": 0.36,
750
+ "grad_norm": 0.3273813699486369,
751
+ "learning_rate": 0.0001887885218402375,
752
+ "loss": 0.3245,
753
+ "step": 104
754
+ },
755
+ {
756
+ "epoch": 0.37,
757
+ "grad_norm": 0.48172713072913986,
758
+ "learning_rate": 0.00018852524888962008,
759
+ "loss": 0.3399,
760
+ "step": 105
761
+ },
762
+ {
763
+ "epoch": 0.37,
764
+ "grad_norm": 0.7705038285726575,
765
+ "learning_rate": 0.00018825910854366912,
766
+ "loss": 0.335,
767
+ "step": 106
768
+ },
769
+ {
770
+ "epoch": 0.37,
771
+ "grad_norm": 0.37045920479977706,
772
+ "learning_rate": 0.0001879901094228584,
773
+ "loss": 0.3278,
774
+ "step": 107
775
+ },
776
+ {
777
+ "epoch": 0.38,
778
+ "grad_norm": 0.6053731781378583,
779
+ "learning_rate": 0.00018771826024025946,
780
+ "loss": 0.3431,
781
+ "step": 108
782
+ },
783
+ {
784
+ "epoch": 0.38,
785
+ "grad_norm": 0.3888129931803796,
786
+ "learning_rate": 0.00018744356980125924,
787
+ "loss": 0.3475,
788
+ "step": 109
789
+ },
790
+ {
791
+ "epoch": 0.38,
792
+ "grad_norm": 0.3466135418611391,
793
+ "learning_rate": 0.00018716604700327514,
794
+ "loss": 0.3582,
795
+ "step": 110
796
+ },
797
+ {
798
+ "epoch": 0.39,
799
+ "grad_norm": 0.3563721673197914,
800
+ "learning_rate": 0.00018688570083546656,
801
+ "loss": 0.3371,
802
+ "step": 111
803
+ },
804
+ {
805
+ "epoch": 0.39,
806
+ "grad_norm": 0.3626204602924834,
807
+ "learning_rate": 0.00018660254037844388,
808
+ "loss": 0.3395,
809
+ "step": 112
810
+ },
811
+ {
812
+ "epoch": 0.39,
813
+ "grad_norm": 0.3381328706653806,
814
+ "learning_rate": 0.00018631657480397428,
815
+ "loss": 0.3415,
816
+ "step": 113
817
+ },
818
+ {
819
+ "epoch": 0.4,
820
+ "grad_norm": 0.34730897124431775,
821
+ "learning_rate": 0.0001860278133746847,
822
+ "loss": 0.3351,
823
+ "step": 114
824
+ },
825
+ {
826
+ "epoch": 0.4,
827
+ "grad_norm": 0.3536370000546,
828
+ "learning_rate": 0.00018573626544376176,
829
+ "loss": 0.3295,
830
+ "step": 115
831
+ },
832
+ {
833
+ "epoch": 0.41,
834
+ "grad_norm": 0.3778843618413747,
835
+ "learning_rate": 0.00018544194045464886,
836
+ "loss": 0.3241,
837
+ "step": 116
838
+ },
839
+ {
840
+ "epoch": 0.41,
841
+ "grad_norm": 0.33047425466150854,
842
+ "learning_rate": 0.00018514484794074026,
843
+ "loss": 0.3058,
844
+ "step": 117
845
+ },
846
+ {
847
+ "epoch": 0.41,
848
+ "grad_norm": 0.38984034886529695,
849
+ "learning_rate": 0.00018484499752507234,
850
+ "loss": 0.3349,
851
+ "step": 118
852
+ },
853
+ {
854
+ "epoch": 0.42,
855
+ "grad_norm": 0.35279205205570496,
856
+ "learning_rate": 0.0001845423989200118,
857
+ "loss": 0.3407,
858
+ "step": 119
859
+ },
860
+ {
861
+ "epoch": 0.42,
862
+ "grad_norm": 0.33614868357115635,
863
+ "learning_rate": 0.00018423706192694116,
864
+ "loss": 0.3414,
865
+ "step": 120
866
+ },
867
+ {
868
+ "epoch": 0.42,
869
+ "grad_norm": 0.3137553036382076,
870
+ "learning_rate": 0.00018392899643594133,
871
+ "loss": 0.3165,
872
+ "step": 121
873
+ },
874
+ {
875
+ "epoch": 0.43,
876
+ "grad_norm": 0.3302671128473347,
877
+ "learning_rate": 0.0001836182124254711,
878
+ "loss": 0.3257,
879
+ "step": 122
880
+ },
881
+ {
882
+ "epoch": 0.43,
883
+ "grad_norm": 0.33379087363924165,
884
+ "learning_rate": 0.00018330471996204408,
885
+ "loss": 0.3173,
886
+ "step": 123
887
+ },
888
+ {
889
+ "epoch": 0.43,
890
+ "grad_norm": 0.3323003015422687,
891
+ "learning_rate": 0.00018298852919990252,
892
+ "loss": 0.3193,
893
+ "step": 124
894
+ },
895
+ {
896
+ "epoch": 0.44,
897
+ "grad_norm": 0.3361870371178332,
898
+ "learning_rate": 0.00018266965038068856,
899
+ "loss": 0.326,
900
+ "step": 125
901
+ },
902
+ {
903
+ "epoch": 0.44,
904
+ "grad_norm": 0.3609552882711658,
905
+ "learning_rate": 0.0001823480938331124,
906
+ "loss": 0.3235,
907
+ "step": 126
908
+ },
909
+ {
910
+ "epoch": 0.44,
911
+ "grad_norm": 0.30841187054669544,
912
+ "learning_rate": 0.0001820238699726177,
913
+ "loss": 0.3453,
914
+ "step": 127
915
+ },
916
+ {
917
+ "epoch": 0.45,
918
+ "grad_norm": 0.34361923504902736,
919
+ "learning_rate": 0.0001816969893010442,
920
+ "loss": 0.3174,
921
+ "step": 128
922
+ },
923
+ {
924
+ "epoch": 0.45,
925
+ "grad_norm": 0.3794710335011336,
926
+ "learning_rate": 0.00018136746240628783,
927
+ "loss": 0.3124,
928
+ "step": 129
929
+ },
930
+ {
931
+ "epoch": 0.45,
932
+ "grad_norm": 0.4095952511654621,
933
+ "learning_rate": 0.0001810352999619574,
934
+ "loss": 0.3173,
935
+ "step": 130
936
+ },
937
+ {
938
+ "epoch": 0.46,
939
+ "grad_norm": 0.33467410070558223,
940
+ "learning_rate": 0.00018070051272702905,
941
+ "loss": 0.306,
942
+ "step": 131
943
+ },
944
+ {
945
+ "epoch": 0.46,
946
+ "grad_norm": 0.5072397964037874,
947
+ "learning_rate": 0.00018036311154549784,
948
+ "loss": 0.3168,
949
+ "step": 132
950
+ },
951
+ {
952
+ "epoch": 0.46,
953
+ "grad_norm": 0.5356013542306316,
954
+ "learning_rate": 0.00018002310734602624,
955
+ "loss": 0.3473,
956
+ "step": 133
957
+ },
958
+ {
959
+ "epoch": 0.47,
960
+ "grad_norm": 0.3244291679820382,
961
+ "learning_rate": 0.00017968051114159047,
962
+ "loss": 0.312,
963
+ "step": 134
964
+ },
965
+ {
966
+ "epoch": 0.47,
967
+ "grad_norm": 0.32439122725379815,
968
+ "learning_rate": 0.00017933533402912354,
969
+ "loss": 0.3353,
970
+ "step": 135
971
+ },
972
+ {
973
+ "epoch": 0.48,
974
+ "grad_norm": 0.3145723176231699,
975
+ "learning_rate": 0.00017898758718915586,
976
+ "loss": 0.3228,
977
+ "step": 136
978
+ },
979
+ {
980
+ "epoch": 0.48,
981
+ "grad_norm": 0.4556724694723649,
982
+ "learning_rate": 0.00017863728188545326,
983
+ "loss": 0.3236,
984
+ "step": 137
985
+ },
986
+ {
987
+ "epoch": 0.48,
988
+ "grad_norm": 0.32584082198297704,
989
+ "learning_rate": 0.00017828442946465187,
990
+ "loss": 0.3051,
991
+ "step": 138
992
+ },
993
+ {
994
+ "epoch": 0.49,
995
+ "grad_norm": 0.28601591306628654,
996
+ "learning_rate": 0.00017792904135589084,
997
+ "loss": 0.2958,
998
+ "step": 139
999
+ },
1000
+ {
1001
+ "epoch": 0.49,
1002
+ "grad_norm": 0.32983504605069464,
1003
+ "learning_rate": 0.000177571129070442,
1004
+ "loss": 0.3284,
1005
+ "step": 140
1006
+ },
1007
+ {
1008
+ "epoch": 0.49,
1009
+ "grad_norm": 0.30738845071531173,
1010
+ "learning_rate": 0.00017721070420133702,
1011
+ "loss": 0.3339,
1012
+ "step": 141
1013
+ },
1014
+ {
1015
+ "epoch": 0.5,
1016
+ "grad_norm": 0.3204453471960076,
1017
+ "learning_rate": 0.00017684777842299205,
1018
+ "loss": 0.3351,
1019
+ "step": 142
1020
+ },
1021
+ {
1022
+ "epoch": 0.5,
1023
+ "grad_norm": 0.30021314102671764,
1024
+ "learning_rate": 0.00017648236349082928,
1025
+ "loss": 0.3257,
1026
+ "step": 143
1027
+ },
1028
+ {
1029
+ "epoch": 0.5,
1030
+ "grad_norm": 0.295043162485709,
1031
+ "learning_rate": 0.00017611447124089649,
1032
+ "loss": 0.3118,
1033
+ "step": 144
1034
+ },
1035
+ {
1036
+ "epoch": 0.5,
1037
+ "eval_loss": 0.30461451411247253,
1038
+ "eval_runtime": 124.0044,
1039
+ "eval_samples_per_second": 11.758,
1040
+ "eval_steps_per_second": 0.742,
1041
+ "step": 144
1042
+ },
1043
+ {
1044
+ "epoch": 0.51,
1045
+ "grad_norm": 0.3063805929371497,
1046
+ "learning_rate": 0.00017574411358948347,
1047
+ "loss": 0.3226,
1048
+ "step": 145
1049
+ },
1050
+ {
1051
+ "epoch": 0.51,
1052
+ "grad_norm": 0.2779054729022279,
1053
+ "learning_rate": 0.00017537130253273613,
1054
+ "loss": 0.297,
1055
+ "step": 146
1056
+ },
1057
+ {
1058
+ "epoch": 0.51,
1059
+ "grad_norm": 0.3124473288970843,
1060
+ "learning_rate": 0.00017499605014626788,
1061
+ "loss": 0.3115,
1062
+ "step": 147
1063
+ },
1064
+ {
1065
+ "epoch": 0.52,
1066
+ "grad_norm": 0.30686516042888795,
1067
+ "learning_rate": 0.00017461836858476856,
1068
+ "loss": 0.3073,
1069
+ "step": 148
1070
+ },
1071
+ {
1072
+ "epoch": 0.52,
1073
+ "grad_norm": 0.368058632241991,
1074
+ "learning_rate": 0.00017423827008161068,
1075
+ "loss": 0.3276,
1076
+ "step": 149
1077
+ },
1078
+ {
1079
+ "epoch": 0.52,
1080
+ "grad_norm": 0.33165127947152323,
1081
+ "learning_rate": 0.00017385576694845325,
1082
+ "loss": 0.3121,
1083
+ "step": 150
1084
+ },
1085
+ {
1086
+ "epoch": 0.53,
1087
+ "grad_norm": 0.49145751891877903,
1088
+ "learning_rate": 0.00017347087157484282,
1089
+ "loss": 0.3161,
1090
+ "step": 151
1091
+ },
1092
+ {
1093
+ "epoch": 0.53,
1094
+ "grad_norm": 0.3250857217236141,
1095
+ "learning_rate": 0.00017308359642781242,
1096
+ "loss": 0.3239,
1097
+ "step": 152
1098
+ },
1099
+ {
1100
+ "epoch": 0.53,
1101
+ "grad_norm": 0.36289323573821,
1102
+ "learning_rate": 0.0001726939540514776,
1103
+ "loss": 0.3244,
1104
+ "step": 153
1105
+ },
1106
+ {
1107
+ "epoch": 0.54,
1108
+ "grad_norm": 0.43817187557785664,
1109
+ "learning_rate": 0.00017230195706663008,
1110
+ "loss": 0.305,
1111
+ "step": 154
1112
+ },
1113
+ {
1114
+ "epoch": 0.54,
1115
+ "grad_norm": 0.36385656762097696,
1116
+ "learning_rate": 0.00017190761817032909,
1117
+ "loss": 0.3088,
1118
+ "step": 155
1119
+ },
1120
+ {
1121
+ "epoch": 0.54,
1122
+ "grad_norm": 0.37945349982504095,
1123
+ "learning_rate": 0.00017151095013548994,
1124
+ "loss": 0.341,
1125
+ "step": 156
1126
+ },
1127
+ {
1128
+ "epoch": 0.55,
1129
+ "grad_norm": 0.32208080193699745,
1130
+ "learning_rate": 0.0001711119658104705,
1131
+ "loss": 0.321,
1132
+ "step": 157
1133
+ },
1134
+ {
1135
+ "epoch": 0.55,
1136
+ "grad_norm": 0.3414633840595063,
1137
+ "learning_rate": 0.00017071067811865476,
1138
+ "loss": 0.3198,
1139
+ "step": 158
1140
+ },
1141
+ {
1142
+ "epoch": 0.56,
1143
+ "grad_norm": 0.3161410527231836,
1144
+ "learning_rate": 0.00017030710005803455,
1145
+ "loss": 0.3265,
1146
+ "step": 159
1147
+ },
1148
+ {
1149
+ "epoch": 0.56,
1150
+ "grad_norm": 0.29063337939498246,
1151
+ "learning_rate": 0.00016990124470078822,
1152
+ "loss": 0.309,
1153
+ "step": 160
1154
+ },
1155
+ {
1156
+ "epoch": 0.56,
1157
+ "grad_norm": 0.28694210240074125,
1158
+ "learning_rate": 0.00016949312519285746,
1159
+ "loss": 0.3187,
1160
+ "step": 161
1161
+ },
1162
+ {
1163
+ "epoch": 0.57,
1164
+ "grad_norm": 0.3912913170835531,
1165
+ "learning_rate": 0.0001690827547535214,
1166
+ "loss": 0.3184,
1167
+ "step": 162
1168
+ },
1169
+ {
1170
+ "epoch": 0.57,
1171
+ "grad_norm": 0.3046534103494425,
1172
+ "learning_rate": 0.00016867014667496836,
1173
+ "loss": 0.3013,
1174
+ "step": 163
1175
+ },
1176
+ {
1177
+ "epoch": 0.57,
1178
+ "grad_norm": 0.35663179718824367,
1179
+ "learning_rate": 0.00016825531432186543,
1180
+ "loss": 0.3245,
1181
+ "step": 164
1182
+ },
1183
+ {
1184
+ "epoch": 0.58,
1185
+ "grad_norm": 0.34732909293058967,
1186
+ "learning_rate": 0.00016783827113092548,
1187
+ "loss": 0.3262,
1188
+ "step": 165
1189
+ },
1190
+ {
1191
+ "epoch": 0.58,
1192
+ "grad_norm": 0.29322806008617414,
1193
+ "learning_rate": 0.00016741903061047203,
1194
+ "loss": 0.3031,
1195
+ "step": 166
1196
+ },
1197
+ {
1198
+ "epoch": 0.58,
1199
+ "grad_norm": 0.32546021723630353,
1200
+ "learning_rate": 0.00016699760634000165,
1201
+ "loss": 0.3187,
1202
+ "step": 167
1203
+ },
1204
+ {
1205
+ "epoch": 0.59,
1206
+ "grad_norm": 0.34240344040353254,
1207
+ "learning_rate": 0.00016657401196974405,
1208
+ "loss": 0.354,
1209
+ "step": 168
1210
+ },
1211
+ {
1212
+ "epoch": 0.59,
1213
+ "grad_norm": 0.2886828042363015,
1214
+ "learning_rate": 0.00016614826122022015,
1215
+ "loss": 0.3177,
1216
+ "step": 169
1217
+ },
1218
+ {
1219
+ "epoch": 0.59,
1220
+ "grad_norm": 0.33420409398270845,
1221
+ "learning_rate": 0.00016572036788179727,
1222
+ "loss": 0.3274,
1223
+ "step": 170
1224
+ },
1225
+ {
1226
+ "epoch": 0.6,
1227
+ "grad_norm": 0.28739463492504613,
1228
+ "learning_rate": 0.00016529034581424292,
1229
+ "loss": 0.3311,
1230
+ "step": 171
1231
+ },
1232
+ {
1233
+ "epoch": 0.6,
1234
+ "grad_norm": 0.30669805875722933,
1235
+ "learning_rate": 0.0001648582089462756,
1236
+ "loss": 0.3133,
1237
+ "step": 172
1238
+ },
1239
+ {
1240
+ "epoch": 0.6,
1241
+ "grad_norm": 0.3032406910617735,
1242
+ "learning_rate": 0.00016442397127511364,
1243
+ "loss": 0.3154,
1244
+ "step": 173
1245
+ },
1246
+ {
1247
+ "epoch": 0.61,
1248
+ "grad_norm": 0.2565178152008159,
1249
+ "learning_rate": 0.0001639876468660219,
1250
+ "loss": 0.2729,
1251
+ "step": 174
1252
+ },
1253
+ {
1254
+ "epoch": 0.61,
1255
+ "grad_norm": 0.31104756424855723,
1256
+ "learning_rate": 0.00016354924985185614,
1257
+ "loss": 0.3152,
1258
+ "step": 175
1259
+ },
1260
+ {
1261
+ "epoch": 0.61,
1262
+ "grad_norm": 0.28520281947044374,
1263
+ "learning_rate": 0.00016310879443260528,
1264
+ "loss": 0.2937,
1265
+ "step": 176
1266
+ },
1267
+ {
1268
+ "epoch": 0.62,
1269
+ "grad_norm": 0.31612903318852875,
1270
+ "learning_rate": 0.00016266629487493144,
1271
+ "loss": 0.3081,
1272
+ "step": 177
1273
+ },
1274
+ {
1275
+ "epoch": 0.62,
1276
+ "grad_norm": 0.30064240260520836,
1277
+ "learning_rate": 0.00016222176551170778,
1278
+ "loss": 0.3133,
1279
+ "step": 178
1280
+ },
1281
+ {
1282
+ "epoch": 0.63,
1283
+ "grad_norm": 0.3158776023109944,
1284
+ "learning_rate": 0.00016177522074155436,
1285
+ "loss": 0.325,
1286
+ "step": 179
1287
+ },
1288
+ {
1289
+ "epoch": 0.63,
1290
+ "grad_norm": 0.2793129602882142,
1291
+ "learning_rate": 0.00016132667502837165,
1292
+ "loss": 0.268,
1293
+ "step": 180
1294
+ },
1295
+ {
1296
+ "epoch": 0.63,
1297
+ "grad_norm": 0.2940935033818059,
1298
+ "learning_rate": 0.00016087614290087208,
1299
+ "loss": 0.2859,
1300
+ "step": 181
1301
+ },
1302
+ {
1303
+ "epoch": 0.64,
1304
+ "grad_norm": 0.31475837771133863,
1305
+ "learning_rate": 0.00016042363895210946,
1306
+ "loss": 0.314,
1307
+ "step": 182
1308
+ },
1309
+ {
1310
+ "epoch": 0.64,
1311
+ "grad_norm": 0.30842305544778004,
1312
+ "learning_rate": 0.0001599691778390063,
1313
+ "loss": 0.2811,
1314
+ "step": 183
1315
+ },
1316
+ {
1317
+ "epoch": 0.64,
1318
+ "grad_norm": 0.3418375917433244,
1319
+ "learning_rate": 0.00015951277428187898,
1320
+ "loss": 0.3191,
1321
+ "step": 184
1322
+ },
1323
+ {
1324
+ "epoch": 0.65,
1325
+ "grad_norm": 0.29715524797293696,
1326
+ "learning_rate": 0.0001590544430639611,
1327
+ "loss": 0.309,
1328
+ "step": 185
1329
+ },
1330
+ {
1331
+ "epoch": 0.65,
1332
+ "grad_norm": 0.3089962157670486,
1333
+ "learning_rate": 0.00015859419903092447,
1334
+ "loss": 0.2984,
1335
+ "step": 186
1336
+ },
1337
+ {
1338
+ "epoch": 0.65,
1339
+ "grad_norm": 0.2798581119773162,
1340
+ "learning_rate": 0.0001581320570903984,
1341
+ "loss": 0.2716,
1342
+ "step": 187
1343
+ },
1344
+ {
1345
+ "epoch": 0.66,
1346
+ "grad_norm": 0.35444678654257317,
1347
+ "learning_rate": 0.00015766803221148673,
1348
+ "loss": 0.3109,
1349
+ "step": 188
1350
+ },
1351
+ {
1352
+ "epoch": 0.66,
1353
+ "grad_norm": 0.33245785178956694,
1354
+ "learning_rate": 0.000157202139424283,
1355
+ "loss": 0.2925,
1356
+ "step": 189
1357
+ },
1358
+ {
1359
+ "epoch": 0.66,
1360
+ "grad_norm": 0.35798123891117034,
1361
+ "learning_rate": 0.00015673439381938366,
1362
+ "loss": 0.3157,
1363
+ "step": 190
1364
+ },
1365
+ {
1366
+ "epoch": 0.67,
1367
+ "grad_norm": 0.32845162982835363,
1368
+ "learning_rate": 0.00015626481054739917,
1369
+ "loss": 0.3147,
1370
+ "step": 191
1371
+ },
1372
+ {
1373
+ "epoch": 0.67,
1374
+ "grad_norm": 0.3588192830988658,
1375
+ "learning_rate": 0.00015579340481846336,
1376
+ "loss": 0.3155,
1377
+ "step": 192
1378
+ },
1379
+ {
1380
+ "epoch": 0.67,
1381
+ "grad_norm": 0.3208631225582573,
1382
+ "learning_rate": 0.00015532019190174072,
1383
+ "loss": 0.3046,
1384
+ "step": 193
1385
+ },
1386
+ {
1387
+ "epoch": 0.68,
1388
+ "grad_norm": 0.3579967857115315,
1389
+ "learning_rate": 0.00015484518712493187,
1390
+ "loss": 0.303,
1391
+ "step": 194
1392
+ },
1393
+ {
1394
+ "epoch": 0.68,
1395
+ "grad_norm": 0.3202268439415007,
1396
+ "learning_rate": 0.00015436840587377702,
1397
+ "loss": 0.2895,
1398
+ "step": 195
1399
+ },
1400
+ {
1401
+ "epoch": 0.68,
1402
+ "grad_norm": 0.4216304882867414,
1403
+ "learning_rate": 0.00015388986359155758,
1404
+ "loss": 0.3131,
1405
+ "step": 196
1406
+ },
1407
+ {
1408
+ "epoch": 0.69,
1409
+ "grad_norm": 0.31007088464120297,
1410
+ "learning_rate": 0.00015340957577859605,
1411
+ "loss": 0.2972,
1412
+ "step": 197
1413
+ },
1414
+ {
1415
+ "epoch": 0.69,
1416
+ "grad_norm": 0.29845328913359065,
1417
+ "learning_rate": 0.0001529275579917539,
1418
+ "loss": 0.28,
1419
+ "step": 198
1420
+ },
1421
+ {
1422
+ "epoch": 0.7,
1423
+ "grad_norm": 0.3102256024747912,
1424
+ "learning_rate": 0.0001524438258439277,
1425
+ "loss": 0.285,
1426
+ "step": 199
1427
+ },
1428
+ {
1429
+ "epoch": 0.7,
1430
+ "grad_norm": 0.33201884078230004,
1431
+ "learning_rate": 0.00015195839500354335,
1432
+ "loss": 0.2947,
1433
+ "step": 200
1434
+ },
1435
+ {
1436
+ "epoch": 0.7,
1437
+ "grad_norm": 0.3118500029268336,
1438
+ "learning_rate": 0.00015147128119404862,
1439
+ "loss": 0.2997,
1440
+ "step": 201
1441
+ },
1442
+ {
1443
+ "epoch": 0.71,
1444
+ "grad_norm": 0.320967620710203,
1445
+ "learning_rate": 0.00015098250019340387,
1446
+ "loss": 0.3197,
1447
+ "step": 202
1448
+ },
1449
+ {
1450
+ "epoch": 0.71,
1451
+ "grad_norm": 0.47483959815019977,
1452
+ "learning_rate": 0.00015049206783357083,
1453
+ "loss": 0.3065,
1454
+ "step": 203
1455
+ },
1456
+ {
1457
+ "epoch": 0.71,
1458
+ "grad_norm": 0.7741392785789275,
1459
+ "learning_rate": 0.00015000000000000001,
1460
+ "loss": 0.3103,
1461
+ "step": 204
1462
+ },
1463
+ {
1464
+ "epoch": 0.72,
1465
+ "grad_norm": 0.3288840302235636,
1466
+ "learning_rate": 0.00014950631263111613,
1467
+ "loss": 0.3023,
1468
+ "step": 205
1469
+ },
1470
+ {
1471
+ "epoch": 0.72,
1472
+ "grad_norm": 0.373883432823129,
1473
+ "learning_rate": 0.00014901102171780174,
1474
+ "loss": 0.2824,
1475
+ "step": 206
1476
+ },
1477
+ {
1478
+ "epoch": 0.72,
1479
+ "grad_norm": 0.2752164871808897,
1480
+ "learning_rate": 0.00014851414330287928,
1481
+ "loss": 0.2964,
1482
+ "step": 207
1483
+ },
1484
+ {
1485
+ "epoch": 0.73,
1486
+ "grad_norm": 0.2879668684767028,
1487
+ "learning_rate": 0.00014801569348059157,
1488
+ "loss": 0.2869,
1489
+ "step": 208
1490
+ },
1491
+ {
1492
+ "epoch": 0.73,
1493
+ "grad_norm": 0.32080956015920353,
1494
+ "learning_rate": 0.00014751568839608037,
1495
+ "loss": 0.3157,
1496
+ "step": 209
1497
+ },
1498
+ {
1499
+ "epoch": 0.73,
1500
+ "grad_norm": 0.29357679211507176,
1501
+ "learning_rate": 0.00014701414424486353,
1502
+ "loss": 0.3197,
1503
+ "step": 210
1504
+ },
1505
+ {
1506
+ "epoch": 0.74,
1507
+ "grad_norm": 0.30290484397244655,
1508
+ "learning_rate": 0.00014651107727231032,
1509
+ "loss": 0.2939,
1510
+ "step": 211
1511
+ },
1512
+ {
1513
+ "epoch": 0.74,
1514
+ "grad_norm": 0.28746944348035675,
1515
+ "learning_rate": 0.00014600650377311522,
1516
+ "loss": 0.3151,
1517
+ "step": 212
1518
+ },
1519
+ {
1520
+ "epoch": 0.74,
1521
+ "grad_norm": 0.5666350067167297,
1522
+ "learning_rate": 0.00014550044009077032,
1523
+ "loss": 0.3153,
1524
+ "step": 213
1525
+ },
1526
+ {
1527
+ "epoch": 0.75,
1528
+ "grad_norm": 0.36451534136785196,
1529
+ "learning_rate": 0.00014499290261703565,
1530
+ "loss": 0.2868,
1531
+ "step": 214
1532
+ },
1533
+ {
1534
+ "epoch": 0.75,
1535
+ "grad_norm": 0.3067404226224104,
1536
+ "learning_rate": 0.00014448390779140844,
1537
+ "loss": 0.3139,
1538
+ "step": 215
1539
+ },
1540
+ {
1541
+ "epoch": 0.75,
1542
+ "grad_norm": 0.2950437432928818,
1543
+ "learning_rate": 0.00014397347210059057,
1544
+ "loss": 0.3142,
1545
+ "step": 216
1546
+ },
1547
+ {
1548
+ "epoch": 0.75,
1549
+ "eval_loss": 0.28815656900405884,
1550
+ "eval_runtime": 123.7277,
1551
+ "eval_samples_per_second": 11.784,
1552
+ "eval_steps_per_second": 0.744,
1553
+ "step": 216
1554
+ },
1555
+ {
1556
+ "epoch": 0.76,
1557
+ "grad_norm": 0.2643412676048768,
1558
+ "learning_rate": 0.00014346161207795462,
1559
+ "loss": 0.2775,
1560
+ "step": 217
1561
+ },
1562
+ {
1563
+ "epoch": 0.76,
1564
+ "grad_norm": 0.3015962433589898,
1565
+ "learning_rate": 0.0001429483443030082,
1566
+ "loss": 0.3079,
1567
+ "step": 218
1568
+ },
1569
+ {
1570
+ "epoch": 0.77,
1571
+ "grad_norm": 0.43877781160353735,
1572
+ "learning_rate": 0.00014243368540085702,
1573
+ "loss": 0.3169,
1574
+ "step": 219
1575
+ },
1576
+ {
1577
+ "epoch": 0.77,
1578
+ "grad_norm": 0.27563035959754884,
1579
+ "learning_rate": 0.00014191765204166643,
1580
+ "loss": 0.2735,
1581
+ "step": 220
1582
+ },
1583
+ {
1584
+ "epoch": 0.77,
1585
+ "grad_norm": 0.3262926209405409,
1586
+ "learning_rate": 0.00014140026094012137,
1587
+ "loss": 0.29,
1588
+ "step": 221
1589
+ },
1590
+ {
1591
+ "epoch": 0.78,
1592
+ "grad_norm": 0.2903306439440674,
1593
+ "learning_rate": 0.00014088152885488502,
1594
+ "loss": 0.2915,
1595
+ "step": 222
1596
+ },
1597
+ {
1598
+ "epoch": 0.78,
1599
+ "grad_norm": 0.37122491225172793,
1600
+ "learning_rate": 0.00014036147258805604,
1601
+ "loss": 0.3074,
1602
+ "step": 223
1603
+ },
1604
+ {
1605
+ "epoch": 0.78,
1606
+ "grad_norm": 0.3357499064589517,
1607
+ "learning_rate": 0.00013984010898462416,
1608
+ "loss": 0.3062,
1609
+ "step": 224
1610
+ },
1611
+ {
1612
+ "epoch": 0.79,
1613
+ "grad_norm": 0.28867360322303315,
1614
+ "learning_rate": 0.00013931745493192473,
1615
+ "loss": 0.2843,
1616
+ "step": 225
1617
+ },
1618
+ {
1619
+ "epoch": 0.79,
1620
+ "grad_norm": 0.30040496970502784,
1621
+ "learning_rate": 0.00013879352735909163,
1622
+ "loss": 0.3197,
1623
+ "step": 226
1624
+ },
1625
+ {
1626
+ "epoch": 0.79,
1627
+ "grad_norm": 0.30150702465313145,
1628
+ "learning_rate": 0.000138268343236509,
1629
+ "loss": 0.3062,
1630
+ "step": 227
1631
+ },
1632
+ {
1633
+ "epoch": 0.8,
1634
+ "grad_norm": 0.30822955744620706,
1635
+ "learning_rate": 0.00013774191957526143,
1636
+ "loss": 0.2984,
1637
+ "step": 228
1638
+ },
1639
+ {
1640
+ "epoch": 0.8,
1641
+ "grad_norm": 0.34488799554108235,
1642
+ "learning_rate": 0.00013721427342658322,
1643
+ "loss": 0.2989,
1644
+ "step": 229
1645
+ },
1646
+ {
1647
+ "epoch": 0.8,
1648
+ "grad_norm": 0.2992279270448066,
1649
+ "learning_rate": 0.00013668542188130566,
1650
+ "loss": 0.3026,
1651
+ "step": 230
1652
+ },
1653
+ {
1654
+ "epoch": 0.81,
1655
+ "grad_norm": 0.29283269902098913,
1656
+ "learning_rate": 0.00013615538206930387,
1657
+ "loss": 0.3022,
1658
+ "step": 231
1659
+ },
1660
+ {
1661
+ "epoch": 0.81,
1662
+ "grad_norm": 0.2963532832735345,
1663
+ "learning_rate": 0.00013562417115894172,
1664
+ "loss": 0.284,
1665
+ "step": 232
1666
+ },
1667
+ {
1668
+ "epoch": 0.81,
1669
+ "grad_norm": 0.3110099434926565,
1670
+ "learning_rate": 0.0001350918063565157,
1671
+ "loss": 0.2976,
1672
+ "step": 233
1673
+ },
1674
+ {
1675
+ "epoch": 0.82,
1676
+ "grad_norm": 0.30746347128990004,
1677
+ "learning_rate": 0.0001345583049056978,
1678
+ "loss": 0.3068,
1679
+ "step": 234
1680
+ },
1681
+ {
1682
+ "epoch": 0.82,
1683
+ "grad_norm": 0.4270470727863756,
1684
+ "learning_rate": 0.0001340236840869768,
1685
+ "loss": 0.3145,
1686
+ "step": 235
1687
+ },
1688
+ {
1689
+ "epoch": 0.82,
1690
+ "grad_norm": 0.3393173086542703,
1691
+ "learning_rate": 0.00013348796121709862,
1692
+ "loss": 0.283,
1693
+ "step": 236
1694
+ },
1695
+ {
1696
+ "epoch": 0.83,
1697
+ "grad_norm": 0.31037606634125064,
1698
+ "learning_rate": 0.00013295115364850534,
1699
+ "loss": 0.3002,
1700
+ "step": 237
1701
+ },
1702
+ {
1703
+ "epoch": 0.83,
1704
+ "grad_norm": 0.3325464403016721,
1705
+ "learning_rate": 0.00013241327876877326,
1706
+ "loss": 0.3232,
1707
+ "step": 238
1708
+ },
1709
+ {
1710
+ "epoch": 0.83,
1711
+ "grad_norm": 0.28871857313160415,
1712
+ "learning_rate": 0.0001318743540000496,
1713
+ "loss": 0.3046,
1714
+ "step": 239
1715
+ },
1716
+ {
1717
+ "epoch": 0.84,
1718
+ "grad_norm": 0.292955131214224,
1719
+ "learning_rate": 0.00013133439679848823,
1720
+ "loss": 0.309,
1721
+ "step": 240
1722
+ },
1723
+ {
1724
+ "epoch": 0.84,
1725
+ "grad_norm": 0.2970827028598667,
1726
+ "learning_rate": 0.0001307934246536843,
1727
+ "loss": 0.2804,
1728
+ "step": 241
1729
+ },
1730
+ {
1731
+ "epoch": 0.85,
1732
+ "grad_norm": 0.27123196621026363,
1733
+ "learning_rate": 0.0001302514550881076,
1734
+ "loss": 0.2867,
1735
+ "step": 242
1736
+ },
1737
+ {
1738
+ "epoch": 0.85,
1739
+ "grad_norm": 0.24967181002480376,
1740
+ "learning_rate": 0.00012970850565653514,
1741
+ "loss": 0.2614,
1742
+ "step": 243
1743
+ },
1744
+ {
1745
+ "epoch": 0.85,
1746
+ "grad_norm": 0.28346057896722426,
1747
+ "learning_rate": 0.0001291645939454825,
1748
+ "loss": 0.3079,
1749
+ "step": 244
1750
+ },
1751
+ {
1752
+ "epoch": 0.86,
1753
+ "grad_norm": 0.27005348199073226,
1754
+ "learning_rate": 0.00012861973757263414,
1755
+ "loss": 0.2621,
1756
+ "step": 245
1757
+ },
1758
+ {
1759
+ "epoch": 0.86,
1760
+ "grad_norm": 0.27231955192598256,
1761
+ "learning_rate": 0.00012807395418627278,
1762
+ "loss": 0.2939,
1763
+ "step": 246
1764
+ },
1765
+ {
1766
+ "epoch": 0.86,
1767
+ "grad_norm": 0.27894707762925364,
1768
+ "learning_rate": 0.00012752726146470774,
1769
+ "loss": 0.2905,
1770
+ "step": 247
1771
+ },
1772
+ {
1773
+ "epoch": 0.87,
1774
+ "grad_norm": 0.3008410159096051,
1775
+ "learning_rate": 0.00012697967711570242,
1776
+ "loss": 0.3044,
1777
+ "step": 248
1778
+ },
1779
+ {
1780
+ "epoch": 0.87,
1781
+ "grad_norm": 0.2823937392828475,
1782
+ "learning_rate": 0.00012643121887590065,
1783
+ "loss": 0.2728,
1784
+ "step": 249
1785
+ },
1786
+ {
1787
+ "epoch": 0.87,
1788
+ "grad_norm": 0.2745336224622263,
1789
+ "learning_rate": 0.00012588190451025207,
1790
+ "loss": 0.284,
1791
+ "step": 250
1792
+ },
1793
+ {
1794
+ "epoch": 0.88,
1795
+ "grad_norm": 0.27690024577274897,
1796
+ "learning_rate": 0.00012533175181143704,
1797
+ "loss": 0.2996,
1798
+ "step": 251
1799
+ },
1800
+ {
1801
+ "epoch": 0.88,
1802
+ "grad_norm": 0.2737330882627236,
1803
+ "learning_rate": 0.00012478077859929,
1804
+ "loss": 0.2767,
1805
+ "step": 252
1806
+ },
1807
+ {
1808
+ "epoch": 0.88,
1809
+ "grad_norm": 0.3101244758838313,
1810
+ "learning_rate": 0.0001242290027202224,
1811
+ "loss": 0.3276,
1812
+ "step": 253
1813
+ },
1814
+ {
1815
+ "epoch": 0.89,
1816
+ "grad_norm": 0.28572219705697965,
1817
+ "learning_rate": 0.00012367644204664468,
1818
+ "loss": 0.2833,
1819
+ "step": 254
1820
+ },
1821
+ {
1822
+ "epoch": 0.89,
1823
+ "grad_norm": 0.2853774297270767,
1824
+ "learning_rate": 0.00012312311447638732,
1825
+ "loss": 0.2697,
1826
+ "step": 255
1827
+ },
1828
+ {
1829
+ "epoch": 0.89,
1830
+ "grad_norm": 0.3037625235263882,
1831
+ "learning_rate": 0.00012256903793212107,
1832
+ "loss": 0.2985,
1833
+ "step": 256
1834
+ },
1835
+ {
1836
+ "epoch": 0.9,
1837
+ "grad_norm": 0.2938437936059258,
1838
+ "learning_rate": 0.00012201423036077657,
1839
+ "loss": 0.3045,
1840
+ "step": 257
1841
+ },
1842
+ {
1843
+ "epoch": 0.9,
1844
+ "grad_norm": 0.30030089577237346,
1845
+ "learning_rate": 0.00012145870973296287,
1846
+ "loss": 0.2688,
1847
+ "step": 258
1848
+ },
1849
+ {
1850
+ "epoch": 0.9,
1851
+ "grad_norm": 0.29990272285936814,
1852
+ "learning_rate": 0.00012090249404238547,
1853
+ "loss": 0.2962,
1854
+ "step": 259
1855
+ },
1856
+ {
1857
+ "epoch": 0.91,
1858
+ "grad_norm": 0.3184289797652873,
1859
+ "learning_rate": 0.0001203456013052634,
1860
+ "loss": 0.2967,
1861
+ "step": 260
1862
+ },
1863
+ {
1864
+ "epoch": 0.91,
1865
+ "grad_norm": 0.31427077224590927,
1866
+ "learning_rate": 0.00011978804955974572,
1867
+ "loss": 0.2849,
1868
+ "step": 261
1869
+ },
1870
+ {
1871
+ "epoch": 0.92,
1872
+ "grad_norm": 0.2821920697210796,
1873
+ "learning_rate": 0.00011922985686532726,
1874
+ "loss": 0.2824,
1875
+ "step": 262
1876
+ },
1877
+ {
1878
+ "epoch": 0.92,
1879
+ "grad_norm": 0.3170239341572326,
1880
+ "learning_rate": 0.00011867104130226363,
1881
+ "loss": 0.2744,
1882
+ "step": 263
1883
+ },
1884
+ {
1885
+ "epoch": 0.92,
1886
+ "grad_norm": 0.3008849225406949,
1887
+ "learning_rate": 0.00011811162097098558,
1888
+ "loss": 0.2756,
1889
+ "step": 264
1890
+ },
1891
+ {
1892
+ "epoch": 0.93,
1893
+ "grad_norm": 0.30357641202979463,
1894
+ "learning_rate": 0.00011755161399151276,
1895
+ "loss": 0.3138,
1896
+ "step": 265
1897
+ },
1898
+ {
1899
+ "epoch": 0.93,
1900
+ "grad_norm": 0.36443146767004536,
1901
+ "learning_rate": 0.00011699103850286669,
1902
+ "loss": 0.2843,
1903
+ "step": 266
1904
+ },
1905
+ {
1906
+ "epoch": 0.93,
1907
+ "grad_norm": 0.28266041813229104,
1908
+ "learning_rate": 0.00011642991266248338,
1909
+ "loss": 0.2992,
1910
+ "step": 267
1911
+ },
1912
+ {
1913
+ "epoch": 0.94,
1914
+ "grad_norm": 0.31289966774968875,
1915
+ "learning_rate": 0.00011586825464562514,
1916
+ "loss": 0.275,
1917
+ "step": 268
1918
+ },
1919
+ {
1920
+ "epoch": 0.94,
1921
+ "grad_norm": 0.3002566183775796,
1922
+ "learning_rate": 0.00011530608264479179,
1923
+ "loss": 0.2729,
1924
+ "step": 269
1925
+ },
1926
+ {
1927
+ "epoch": 0.94,
1928
+ "grad_norm": 0.28876872212538823,
1929
+ "learning_rate": 0.00011474341486913145,
1930
+ "loss": 0.2756,
1931
+ "step": 270
1932
+ },
1933
+ {
1934
+ "epoch": 0.95,
1935
+ "grad_norm": 0.29145467524565366,
1936
+ "learning_rate": 0.00011418026954385082,
1937
+ "loss": 0.2736,
1938
+ "step": 271
1939
+ },
1940
+ {
1941
+ "epoch": 0.95,
1942
+ "grad_norm": 0.2905116613947892,
1943
+ "learning_rate": 0.00011361666490962468,
1944
+ "loss": 0.2738,
1945
+ "step": 272
1946
+ },
1947
+ {
1948
+ "epoch": 0.95,
1949
+ "grad_norm": 0.27003932832020283,
1950
+ "learning_rate": 0.00011305261922200519,
1951
+ "loss": 0.285,
1952
+ "step": 273
1953
+ },
1954
+ {
1955
+ "epoch": 0.96,
1956
+ "grad_norm": 0.2809966721346155,
1957
+ "learning_rate": 0.0001124881507508305,
1958
+ "loss": 0.2636,
1959
+ "step": 274
1960
+ },
1961
+ {
1962
+ "epoch": 0.96,
1963
+ "grad_norm": 0.3273824964730481,
1964
+ "learning_rate": 0.00011192327777963312,
1965
+ "loss": 0.2978,
1966
+ "step": 275
1967
+ },
1968
+ {
1969
+ "epoch": 0.96,
1970
+ "grad_norm": 0.31420393363965815,
1971
+ "learning_rate": 0.00011135801860504749,
1972
+ "loss": 0.2888,
1973
+ "step": 276
1974
+ },
1975
+ {
1976
+ "epoch": 0.97,
1977
+ "grad_norm": 0.2951497088424898,
1978
+ "learning_rate": 0.00011079239153621752,
1979
+ "loss": 0.2735,
1980
+ "step": 277
1981
+ },
1982
+ {
1983
+ "epoch": 0.97,
1984
+ "grad_norm": 0.2795628834795779,
1985
+ "learning_rate": 0.00011022641489420342,
1986
+ "loss": 0.2659,
1987
+ "step": 278
1988
+ },
1989
+ {
1990
+ "epoch": 0.97,
1991
+ "grad_norm": 0.26725143481923,
1992
+ "learning_rate": 0.0001096601070113884,
1993
+ "loss": 0.2677,
1994
+ "step": 279
1995
+ },
1996
+ {
1997
+ "epoch": 0.98,
1998
+ "grad_norm": 0.2933108430584371,
1999
+ "learning_rate": 0.0001090934862308847,
2000
+ "loss": 0.2982,
2001
+ "step": 280
2002
+ },
2003
+ {
2004
+ "epoch": 0.98,
2005
+ "grad_norm": 0.3007901797269533,
2006
+ "learning_rate": 0.00010852657090593961,
2007
+ "loss": 0.3006,
2008
+ "step": 281
2009
+ },
2010
+ {
2011
+ "epoch": 0.99,
2012
+ "grad_norm": 0.2669472182828603,
2013
+ "learning_rate": 0.00010795937939934088,
2014
+ "loss": 0.2681,
2015
+ "step": 282
2016
+ },
2017
+ {
2018
+ "epoch": 0.99,
2019
+ "grad_norm": 0.28720121147539796,
2020
+ "learning_rate": 0.00010739193008282203,
2021
+ "loss": 0.296,
2022
+ "step": 283
2023
+ },
2024
+ {
2025
+ "epoch": 0.99,
2026
+ "grad_norm": 0.2895808855589094,
2027
+ "learning_rate": 0.0001068242413364671,
2028
+ "loss": 0.2857,
2029
+ "step": 284
2030
+ },
2031
+ {
2032
+ "epoch": 1.0,
2033
+ "grad_norm": 0.31151272268015245,
2034
+ "learning_rate": 0.0001062563315481156,
2035
+ "loss": 0.2923,
2036
+ "step": 285
2037
+ },
2038
+ {
2039
+ "epoch": 1.0,
2040
+ "grad_norm": 0.25642831169307667,
2041
+ "learning_rate": 0.00010568821911276669,
2042
+ "loss": 0.2677,
2043
+ "step": 286
2044
+ }
2045
+ ],
2046
+ "logging_steps": 1,
2047
+ "max_steps": 572,
2048
+ "num_input_tokens_seen": 0,
2049
+ "num_train_epochs": 2,
2050
+ "save_steps": 286,
2051
+ "total_flos": 8.356035693876609e+18,
2052
+ "train_batch_size": 2,
2053
+ "trial_name": null,
2054
+ "trial_params": null
2055
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e51dc7b769da3aeed2febee8c34db52e73499c0ec9899f446f972273af0d42f
3
+ size 7224
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-286/zero_to_fp32.py ADDED
@@ -0,0 +1,592 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python
2
+
3
+ # Copyright (c) Microsoft Corporation.
4
+ # SPDX-License-Identifier: Apache-2.0
5
+
6
+ # DeepSpeed Team
7
+
8
+ # This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets
9
+ # copied into the top level checkpoint dir, so the user can easily do the conversion at any point in
10
+ # the future. Once extracted, the weights don't require DeepSpeed and can be used in any
11
+ # application.
12
+ #
13
+ # example: python zero_to_fp32.py . pytorch_model.bin
14
+
15
+ import argparse
16
+ import torch
17
+ import glob
18
+ import math
19
+ import os
20
+ import re
21
+ from collections import OrderedDict
22
+ from dataclasses import dataclass
23
+
24
+ # while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with
25
+ # DeepSpeed data structures it has to be available in the current python environment.
26
+ from deepspeed.utils import logger
27
+ from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS,
28
+ FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES,
29
+ FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS)
30
+
31
+
32
+ @dataclass
33
+ class zero_model_state:
34
+ buffers: dict()
35
+ param_shapes: dict()
36
+ shared_params: list
37
+ ds_version: int
38
+ frozen_param_shapes: dict()
39
+ frozen_param_fragments: dict()
40
+
41
+
42
+ debug = 0
43
+
44
+ # load to cpu
45
+ device = torch.device('cpu')
46
+
47
+
48
+ def atoi(text):
49
+ return int(text) if text.isdigit() else text
50
+
51
+
52
+ def natural_keys(text):
53
+ '''
54
+ alist.sort(key=natural_keys) sorts in human order
55
+ http://nedbatchelder.com/blog/200712/human_sorting.html
56
+ (See Toothy's implementation in the comments)
57
+ '''
58
+ return [atoi(c) for c in re.split(r'(\d+)', text)]
59
+
60
+
61
+ def get_model_state_file(checkpoint_dir, zero_stage):
62
+ if not os.path.isdir(checkpoint_dir):
63
+ raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist")
64
+
65
+ # there should be only one file
66
+ if zero_stage <= 2:
67
+ file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt")
68
+ elif zero_stage == 3:
69
+ file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt")
70
+
71
+ if not os.path.exists(file):
72
+ raise FileNotFoundError(f"can't find model states file at '{file}'")
73
+
74
+ return file
75
+
76
+
77
+ def get_checkpoint_files(checkpoint_dir, glob_pattern):
78
+ # XXX: need to test that this simple glob rule works for multi-node setup too
79
+ ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys)
80
+
81
+ if len(ckpt_files) == 0:
82
+ raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'")
83
+
84
+ return ckpt_files
85
+
86
+
87
+ def get_optim_files(checkpoint_dir):
88
+ return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt")
89
+
90
+
91
+ def get_model_state_files(checkpoint_dir):
92
+ return get_checkpoint_files(checkpoint_dir, "*_model_states.pt")
93
+
94
+
95
+ def parse_model_states(files):
96
+ zero_model_states = []
97
+ for file in files:
98
+ state_dict = torch.load(file, map_location=device)
99
+
100
+ if BUFFER_NAMES not in state_dict:
101
+ raise ValueError(f"{file} is not a model state checkpoint")
102
+ buffer_names = state_dict[BUFFER_NAMES]
103
+ if debug:
104
+ print("Found buffers:", buffer_names)
105
+
106
+ # recover just the buffers while restoring them to fp32 if they were saved in fp16
107
+ buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names}
108
+ param_shapes = state_dict[PARAM_SHAPES]
109
+
110
+ # collect parameters that are included in param_shapes
111
+ param_names = []
112
+ for s in param_shapes:
113
+ for name in s.keys():
114
+ param_names.append(name)
115
+
116
+ # update with frozen parameters
117
+ frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None)
118
+ if frozen_param_shapes is not None:
119
+ if debug:
120
+ print(f"Found frozen_param_shapes: {frozen_param_shapes}")
121
+ param_names += list(frozen_param_shapes.keys())
122
+
123
+ # handle shared params
124
+ shared_params = [[k, v] for k, v in state_dict["shared_params"].items()]
125
+
126
+ ds_version = state_dict.get(DS_VERSION, None)
127
+
128
+ frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None)
129
+
130
+ z_model_state = zero_model_state(buffers=buffers,
131
+ param_shapes=param_shapes,
132
+ shared_params=shared_params,
133
+ ds_version=ds_version,
134
+ frozen_param_shapes=frozen_param_shapes,
135
+ frozen_param_fragments=frozen_param_fragments)
136
+ zero_model_states.append(z_model_state)
137
+
138
+ return zero_model_states
139
+
140
+
141
+ def parse_optim_states(files, ds_checkpoint_dir):
142
+
143
+ total_files = len(files)
144
+ state_dicts = []
145
+ for f in files:
146
+ state_dict = torch.load(f, map_location=device)
147
+ # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights
148
+ # and also handle the case where it was already removed by another helper script
149
+ state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None)
150
+ state_dicts.append(state_dict)
151
+
152
+ if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]:
153
+ raise ValueError(f"{files[0]} is not a zero checkpoint")
154
+ zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE]
155
+ world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT]
156
+
157
+ # For ZeRO-2 each param group can have different partition_count as data parallelism for expert
158
+ # parameters can be different from data parallelism for non-expert parameters. So we can just
159
+ # use the max of the partition_count to get the dp world_size.
160
+
161
+ if type(world_size) is list:
162
+ world_size = max(world_size)
163
+
164
+ if world_size != total_files:
165
+ raise ValueError(
166
+ f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. "
167
+ "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes."
168
+ )
169
+
170
+ # the groups are named differently in each stage
171
+ if zero_stage <= 2:
172
+ fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS
173
+ elif zero_stage == 3:
174
+ fp32_groups_key = FP32_FLAT_GROUPS
175
+ else:
176
+ raise ValueError(f"unknown zero stage {zero_stage}")
177
+
178
+ if zero_stage <= 2:
179
+ fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))]
180
+ elif zero_stage == 3:
181
+ # if there is more than one param group, there will be multiple flattened tensors - one
182
+ # flattened tensor per group - for simplicity merge them into a single tensor
183
+ #
184
+ # XXX: could make the script more memory efficient for when there are multiple groups - it
185
+ # will require matching the sub-lists of param_shapes for each param group flattened tensor
186
+
187
+ fp32_flat_groups = [
188
+ torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts))
189
+ ]
190
+
191
+ return zero_stage, world_size, fp32_flat_groups
192
+
193
+
194
+ def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir):
195
+ """
196
+ Returns fp32 state_dict reconstructed from ds checkpoint
197
+
198
+ Args:
199
+ - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are)
200
+
201
+ """
202
+ print(f"Processing zero checkpoint '{ds_checkpoint_dir}'")
203
+
204
+ optim_files = get_optim_files(ds_checkpoint_dir)
205
+ zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir)
206
+ print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}")
207
+
208
+ model_files = get_model_state_files(ds_checkpoint_dir)
209
+
210
+ zero_model_states = parse_model_states(model_files)
211
+ print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}')
212
+
213
+ if zero_stage <= 2:
214
+ return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states)
215
+ elif zero_stage == 3:
216
+ return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states)
217
+
218
+
219
+ def _zero2_merge_frozen_params(state_dict, zero_model_states):
220
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
221
+ return
222
+
223
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
224
+ frozen_param_fragments = zero_model_states[0].frozen_param_fragments
225
+
226
+ if debug:
227
+ num_elem = sum(s.numel() for s in frozen_param_shapes.values())
228
+ print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
229
+
230
+ wanted_params = len(frozen_param_shapes)
231
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
232
+ avail_numel = sum([p.numel() for p in frozen_param_fragments.values()])
233
+ print(f'Frozen params: Have {avail_numel} numels to process.')
234
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
235
+
236
+ total_params = 0
237
+ total_numel = 0
238
+ for name, shape in frozen_param_shapes.items():
239
+ total_params += 1
240
+ unpartitioned_numel = shape.numel()
241
+ total_numel += unpartitioned_numel
242
+
243
+ state_dict[name] = frozen_param_fragments[name]
244
+
245
+ if debug:
246
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
247
+
248
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
249
+
250
+
251
+ def _has_callable(obj, fn):
252
+ attr = getattr(obj, fn, None)
253
+ return callable(attr)
254
+
255
+
256
+ def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
257
+ param_shapes = zero_model_states[0].param_shapes
258
+
259
+ # Reconstruction protocol:
260
+ #
261
+ # XXX: document this
262
+
263
+ if debug:
264
+ for i in range(world_size):
265
+ for j in range(len(fp32_flat_groups[0])):
266
+ print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}")
267
+
268
+ # XXX: memory usage doubles here (zero2)
269
+ num_param_groups = len(fp32_flat_groups[0])
270
+ merged_single_partition_of_fp32_groups = []
271
+ for i in range(num_param_groups):
272
+ merged_partitions = [sd[i] for sd in fp32_flat_groups]
273
+ full_single_fp32_vector = torch.cat(merged_partitions, 0)
274
+ merged_single_partition_of_fp32_groups.append(full_single_fp32_vector)
275
+ avail_numel = sum(
276
+ [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups])
277
+
278
+ if debug:
279
+ wanted_params = sum([len(shapes) for shapes in param_shapes])
280
+ wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes])
281
+ # not asserting if there is a mismatch due to possible padding
282
+ print(f"Have {avail_numel} numels to process.")
283
+ print(f"Need {wanted_numel} numels in {wanted_params} params.")
284
+
285
+ # params
286
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
287
+ # out-of-core computing solution
288
+ total_numel = 0
289
+ total_params = 0
290
+ for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups):
291
+ offset = 0
292
+ avail_numel = full_single_fp32_vector.numel()
293
+ for name, shape in shapes.items():
294
+
295
+ unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape)
296
+ total_numel += unpartitioned_numel
297
+ total_params += 1
298
+
299
+ if debug:
300
+ print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ")
301
+ state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape)
302
+ offset += unpartitioned_numel
303
+
304
+ # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and
305
+ # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex
306
+ # paddings performed in the code it's almost impossible to predict the exact numbers w/o the
307
+ # live optimizer object, so we are checking that the numbers are within the right range
308
+ align_to = 2 * world_size
309
+
310
+ def zero2_align(x):
311
+ return align_to * math.ceil(x / align_to)
312
+
313
+ if debug:
314
+ print(f"original offset={offset}, avail_numel={avail_numel}")
315
+
316
+ offset = zero2_align(offset)
317
+ avail_numel = zero2_align(avail_numel)
318
+
319
+ if debug:
320
+ print(f"aligned offset={offset}, avail_numel={avail_numel}")
321
+
322
+ # Sanity check
323
+ if offset != avail_numel:
324
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
325
+
326
+ print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements")
327
+
328
+
329
+ def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states):
330
+ state_dict = OrderedDict()
331
+
332
+ # buffers
333
+ buffers = zero_model_states[0].buffers
334
+ state_dict.update(buffers)
335
+ if debug:
336
+ print(f"added {len(buffers)} buffers")
337
+
338
+ _zero2_merge_frozen_params(state_dict, zero_model_states)
339
+
340
+ _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
341
+
342
+ # recover shared parameters
343
+ for pair in zero_model_states[0].shared_params:
344
+ if pair[1] in state_dict:
345
+ state_dict[pair[0]] = state_dict[pair[1]]
346
+
347
+ return state_dict
348
+
349
+
350
+ def zero3_partitioned_param_info(unpartitioned_numel, world_size):
351
+ remainder = unpartitioned_numel % world_size
352
+ padding_numel = (world_size - remainder) if remainder else 0
353
+ partitioned_numel = math.ceil(unpartitioned_numel / world_size)
354
+ return partitioned_numel, padding_numel
355
+
356
+
357
+ def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states):
358
+ if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0:
359
+ return
360
+
361
+ if debug:
362
+ for i in range(world_size):
363
+ num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values())
364
+ print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}')
365
+
366
+ frozen_param_shapes = zero_model_states[0].frozen_param_shapes
367
+ wanted_params = len(frozen_param_shapes)
368
+ wanted_numel = sum(s.numel() for s in frozen_param_shapes.values())
369
+ avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size
370
+ print(f'Frozen params: Have {avail_numel} numels to process.')
371
+ print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params')
372
+
373
+ total_params = 0
374
+ total_numel = 0
375
+ for name, shape in zero_model_states[0].frozen_param_shapes.items():
376
+ total_params += 1
377
+ unpartitioned_numel = shape.numel()
378
+ total_numel += unpartitioned_numel
379
+
380
+ param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states)
381
+ state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape)
382
+
383
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
384
+
385
+ if debug:
386
+ print(
387
+ f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
388
+ )
389
+
390
+ print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements")
391
+
392
+
393
+ def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states):
394
+ param_shapes = zero_model_states[0].param_shapes
395
+ avail_numel = fp32_flat_groups[0].numel() * world_size
396
+ # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each
397
+ # param, re-consolidating each param, while dealing with padding if any
398
+
399
+ # merge list of dicts, preserving order
400
+ param_shapes = {k: v for d in param_shapes for k, v in d.items()}
401
+
402
+ if debug:
403
+ for i in range(world_size):
404
+ print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}")
405
+
406
+ wanted_params = len(param_shapes)
407
+ wanted_numel = sum(shape.numel() for shape in param_shapes.values())
408
+ # not asserting if there is a mismatch due to possible padding
409
+ avail_numel = fp32_flat_groups[0].numel() * world_size
410
+ print(f"Trainable params: Have {avail_numel} numels to process.")
411
+ print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.")
412
+
413
+ # params
414
+ # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support
415
+ # out-of-core computing solution
416
+ offset = 0
417
+ total_numel = 0
418
+ total_params = 0
419
+ for name, shape in param_shapes.items():
420
+
421
+ unpartitioned_numel = shape.numel()
422
+ total_numel += unpartitioned_numel
423
+ total_params += 1
424
+
425
+ partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size)
426
+
427
+ if debug:
428
+ print(
429
+ f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}"
430
+ )
431
+
432
+ # XXX: memory usage doubles here
433
+ state_dict[name] = torch.cat(
434
+ tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)),
435
+ 0).narrow(0, 0, unpartitioned_numel).view(shape)
436
+ offset += partitioned_numel
437
+
438
+ offset *= world_size
439
+
440
+ # Sanity check
441
+ if offset != avail_numel:
442
+ raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong")
443
+
444
+ print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements")
445
+
446
+
447
+ def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states):
448
+ state_dict = OrderedDict()
449
+
450
+ # buffers
451
+ buffers = zero_model_states[0].buffers
452
+ state_dict.update(buffers)
453
+ if debug:
454
+ print(f"added {len(buffers)} buffers")
455
+
456
+ _zero3_merge_frozen_params(state_dict, world_size, zero_model_states)
457
+
458
+ _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states)
459
+
460
+ # recover shared parameters
461
+ for pair in zero_model_states[0].shared_params:
462
+ if pair[1] in state_dict:
463
+ state_dict[pair[0]] = state_dict[pair[1]]
464
+
465
+ return state_dict
466
+
467
+
468
+ def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None):
469
+ """
470
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with
471
+ ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example
472
+ via a model hub.
473
+
474
+ Args:
475
+ - ``checkpoint_dir``: path to the desired checkpoint folder
476
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14``
477
+
478
+ Returns:
479
+ - pytorch ``state_dict``
480
+
481
+ Note: this approach may not work if your application doesn't have sufficient free CPU memory and
482
+ you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with
483
+ the checkpoint.
484
+
485
+ A typical usage might be ::
486
+
487
+ from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint
488
+ # do the training and checkpoint saving
489
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu
490
+ model = model.cpu() # move to cpu
491
+ model.load_state_dict(state_dict)
492
+ # submit to model hub or save the model to share with others
493
+
494
+ In this example the ``model`` will no longer be usable in the deepspeed context of the same
495
+ application. i.e. you will need to re-initialize the deepspeed engine, since
496
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
497
+
498
+ If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead.
499
+
500
+ """
501
+ if tag is None:
502
+ latest_path = os.path.join(checkpoint_dir, 'latest')
503
+ if os.path.isfile(latest_path):
504
+ with open(latest_path, 'r') as fd:
505
+ tag = fd.read().strip()
506
+ else:
507
+ raise ValueError(f"Unable to find 'latest' file at {latest_path}")
508
+
509
+ ds_checkpoint_dir = os.path.join(checkpoint_dir, tag)
510
+
511
+ if not os.path.isdir(ds_checkpoint_dir):
512
+ raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist")
513
+
514
+ return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir)
515
+
516
+
517
+ def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None):
518
+ """
519
+ Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be
520
+ loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed.
521
+
522
+ Args:
523
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
524
+ - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin)
525
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
526
+ """
527
+
528
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag)
529
+ print(f"Saving fp32 state dict to {output_file}")
530
+ torch.save(state_dict, output_file)
531
+
532
+
533
+ def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None):
534
+ """
535
+ 1. Put the provided model to cpu
536
+ 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict``
537
+ 3. Load it into the provided model
538
+
539
+ Args:
540
+ - ``model``: the model object to update
541
+ - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``)
542
+ - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14``
543
+
544
+ Returns:
545
+ - ``model`: modified model
546
+
547
+ Make sure you have plenty of CPU memory available before you call this function. If you don't
548
+ have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it
549
+ conveniently placed for you in the checkpoint folder.
550
+
551
+ A typical usage might be ::
552
+
553
+ from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint
554
+ model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir)
555
+ # submit to model hub or save the model to share with others
556
+
557
+ Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context
558
+ of the same application. i.e. you will need to re-initialize the deepspeed engine, since
559
+ ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it.
560
+
561
+ """
562
+ logger.info(f"Extracting fp32 weights")
563
+ state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag)
564
+
565
+ logger.info(f"Overwriting model with fp32 weights")
566
+ model = model.cpu()
567
+ model.load_state_dict(state_dict, strict=False)
568
+
569
+ return model
570
+
571
+
572
+ if __name__ == "__main__":
573
+
574
+ parser = argparse.ArgumentParser()
575
+ parser.add_argument("checkpoint_dir",
576
+ type=str,
577
+ help="path to the desired checkpoint folder, e.g., path/checkpoint-12")
578
+ parser.add_argument(
579
+ "output_file",
580
+ type=str,
581
+ help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)")
582
+ parser.add_argument("-t",
583
+ "--tag",
584
+ type=str,
585
+ default=None,
586
+ help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1")
587
+ parser.add_argument("-d", "--debug", action='store_true', help="enable debug")
588
+ args = parser.parse_args()
589
+
590
+ debug = args.debug
591
+
592
+ convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, args.output_file, tag=args.tag)
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ base_model: deepseek-ai/deepseek-coder-33b-instruct
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.10.0
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "deepseek-ai/deepseek-coder-33b-instruct",
5
+ "bias": "none",
6
+ "fan_in_fan_out": null,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 2048,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 1024,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "up_proj",
24
+ "o_proj",
25
+ "v_proj",
26
+ "down_proj",
27
+ "gate_proj",
28
+ "q_proj",
29
+ "k_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acdd69f418d595ede6edd6df0d2820a50546c217bf17fe739acee9ebc5b2e678
3
+ size 15765462656
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7f0b8bfb911b53f6424012c6ef8f471ecfa68e6f5fcab62b0cfb9b5ba42f76b
3
+ size 11824020348
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdea1c2c26db62cdc26a9b975c0eab4a49ffc3cf3b7ef19d5f2fc9744a992402
3
+ size 11824020540
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffdced5fc4062746f8d77ed328d0f266701510c7b64694111067c8d94219f9b4
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29306605bf266932730b5b08328db797507f7d0f0eab7fb98259b904c0dc39cb
3
+ size 11824020028
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e02fb8ca7698be2ee1f6fdf79fbc51f11c7f0a1b4b1e3be6e1ba94b624f7f87
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14527941d6a72652cc8f17e2b1e4aac4c3a7d2124fdf0d4a28da8d1095728aae
3
+ size 11824020604
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d0c8c3d38119a8823eccd8ab1e89dcb11e2524cf37b2af063755b1aa418d139
3
+ size 11824020476
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ab035e284f88cc531d8d7661193f6002becafd8d3b40f526326c77c1d6f674b
3
+ size 11824020028
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/global_step572/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1c4662eace3ef3cff4151b53694be12b95359fd577527a36d432d17d02195bc
3
+ size 15781541500
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step572
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4629ebff7d228617d587a013934402576f65da427bcb9fe06b983e6b9562468f
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b9707493cc68b1b68c9314d01e770b4ec323722d68600cbc80f3739513ac766
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f20c6b6a252e98a003dcca2c258a5b431dcc0ff26a118b939d58a0f62f87f6d
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3626b70afe54ddfc1aeaeadbfff329ff523e4372f8de80e7f8c85d025d592a07
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:127add9441b3fc080435b1af204cc5d8309388c514f9cf99aaf73d88bdbbe82b
3
+ size 15984
gcp_models/lora-logo_fix_deepseek33b_gpt35i_lr_0.0002_alpha_2048_r_1024/checkpoint-572/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abb04d3774e3c229e88e8341a8fb8b99e2243b081b0a9090320b5905eb552c36
3
+ size 15984