MohamedAhmedAE commited on
Commit
d85e2ee
1 Parent(s): 7793ca8

Training in progress, step 103200

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8d7cc20b396cca3e8d34e141799f9fb7d26e17cc8eff473c1afd251d8d9b6cc
3
  size 5544997664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b73cfbe42b394b528fb9372ce03600810b6d2284ea81e5e376bf7b26e1b27e9f
3
  size 5544997664
last-checkpoint/adapter_config.json CHANGED
@@ -20,12 +20,12 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "up_proj",
25
- "q_proj",
26
  "v_proj",
27
- "down_proj",
28
  "gate_proj",
 
 
 
29
  "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "v_proj",
24
+ "k_proj",
25
  "gate_proj",
26
+ "down_proj",
27
+ "q_proj",
28
+ "up_proj",
29
  "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8d7cc20b396cca3e8d34e141799f9fb7d26e17cc8eff473c1afd251d8d9b6cc
3
  size 5544997664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ff50e9a0eef14c00f32c5e550257295427f2d666e009aac32472aef43b0c78f
3
  size 5544997664
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0932653c53bd9f5af69d8d728513a659e32e307161fd0359d520ee70c7aa3c5
3
  size 674093138
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b1f3c8b9d1c8514057a760685418307853b2172387d395e371d81516debcc99
3
  size 674093138
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6474451a1e2f3e46bed2d21701e092cafc22a415c811eebe38a7a07c535fa7ba
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d52e9778ae961a843d4efe5adba669832146332ec663eac9df46d71427724e3
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:212c085ef0256e6cd45cc46be3fe004a003548c018452df3f4f4b8de50ed9745
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1dd725a3e5295711459643d6e1204a1d04a7f905cc6416544fa87ecdfb18228
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.07667365728139455,
5
  "eval_steps": 200,
6
- "global_step": 103100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -7070,160 +7070,6 @@
7070
  "learning_rate": 1.9988888715649357e-05,
7071
  "loss": 1.5441,
7072
  "step": 100900
7073
- },
7074
- {
7075
- "epoch": 0.07511192420388797,
7076
- "grad_norm": 1.068562388420105,
7077
- "learning_rate": 1.998886668338319e-05,
7078
- "loss": 1.4998,
7079
- "step": 101000
7080
- },
7081
- {
7082
- "epoch": 0.075186292445674,
7083
- "grad_norm": 0.6331286430358887,
7084
- "learning_rate": 1.998884462930723e-05,
7085
- "loss": 1.5633,
7086
- "step": 101100
7087
- },
7088
- {
7089
- "epoch": 0.07526066068746003,
7090
- "grad_norm": 1.3566038608551025,
7091
- "learning_rate": 1.998882255342152e-05,
7092
- "loss": 1.4621,
7093
- "step": 101200
7094
- },
7095
- {
7096
- "epoch": 0.07533502892924605,
7097
- "grad_norm": 0.9672004580497742,
7098
- "learning_rate": 1.998880045572611e-05,
7099
- "loss": 1.5249,
7100
- "step": 101300
7101
- },
7102
- {
7103
- "epoch": 0.07540939717103208,
7104
- "grad_norm": 0.36732280254364014,
7105
- "learning_rate": 1.9988778336221045e-05,
7106
- "loss": 1.574,
7107
- "step": 101400
7108
- },
7109
- {
7110
- "epoch": 0.07548376541281811,
7111
- "grad_norm": 0.4788234829902649,
7112
- "learning_rate": 1.998875619490638e-05,
7113
- "loss": 1.5418,
7114
- "step": 101500
7115
- },
7116
- {
7117
- "epoch": 0.07555813365460413,
7118
- "grad_norm": 0.8955681324005127,
7119
- "learning_rate": 1.9988734031782157e-05,
7120
- "loss": 1.5568,
7121
- "step": 101600
7122
- },
7123
- {
7124
- "epoch": 0.07563250189639016,
7125
- "grad_norm": 0.8049163222312927,
7126
- "learning_rate": 1.9988711846848427e-05,
7127
- "loss": 1.4838,
7128
- "step": 101700
7129
- },
7130
- {
7131
- "epoch": 0.07570687013817619,
7132
- "grad_norm": 0.7558008432388306,
7133
- "learning_rate": 1.9988689640105235e-05,
7134
- "loss": 1.4955,
7135
- "step": 101800
7136
- },
7137
- {
7138
- "epoch": 0.07578123837996222,
7139
- "grad_norm": 0.4749026596546173,
7140
- "learning_rate": 1.9988667411552635e-05,
7141
- "loss": 1.5929,
7142
- "step": 101900
7143
- },
7144
- {
7145
- "epoch": 0.07585560662174824,
7146
- "grad_norm": 0.6597522497177124,
7147
- "learning_rate": 1.998864516119067e-05,
7148
- "loss": 1.5584,
7149
- "step": 102000
7150
- },
7151
- {
7152
- "epoch": 0.07592997486353428,
7153
- "grad_norm": 0.7412188053131104,
7154
- "learning_rate": 1.9988622889019395e-05,
7155
- "loss": 1.5842,
7156
- "step": 102100
7157
- },
7158
- {
7159
- "epoch": 0.07600434310532031,
7160
- "grad_norm": 0.5564984679222107,
7161
- "learning_rate": 1.9988600595038853e-05,
7162
- "loss": 1.5764,
7163
- "step": 102200
7164
- },
7165
- {
7166
- "epoch": 0.07607871134710634,
7167
- "grad_norm": 1.0488529205322266,
7168
- "learning_rate": 1.9988578279249097e-05,
7169
- "loss": 1.458,
7170
- "step": 102300
7171
- },
7172
- {
7173
- "epoch": 0.07615307958889236,
7174
- "grad_norm": 1.40269136428833,
7175
- "learning_rate": 1.998855594165017e-05,
7176
- "loss": 1.4588,
7177
- "step": 102400
7178
- },
7179
- {
7180
- "epoch": 0.07622744783067839,
7181
- "grad_norm": 0.8488138318061829,
7182
- "learning_rate": 1.9988533582242127e-05,
7183
- "loss": 1.522,
7184
- "step": 102500
7185
- },
7186
- {
7187
- "epoch": 0.07630181607246442,
7188
- "grad_norm": 0.5191701054573059,
7189
- "learning_rate": 1.9988511201025015e-05,
7190
- "loss": 1.5036,
7191
- "step": 102600
7192
- },
7193
- {
7194
- "epoch": 0.07637618431425044,
7195
- "grad_norm": 0.6648279428482056,
7196
- "learning_rate": 1.9988488797998878e-05,
7197
- "loss": 1.4929,
7198
- "step": 102700
7199
- },
7200
- {
7201
- "epoch": 0.07645055255603647,
7202
- "grad_norm": 1.8600202798843384,
7203
- "learning_rate": 1.9988466373163774e-05,
7204
- "loss": 1.5692,
7205
- "step": 102800
7206
- },
7207
- {
7208
- "epoch": 0.0765249207978225,
7209
- "grad_norm": 0.7583739757537842,
7210
- "learning_rate": 1.9988443926519743e-05,
7211
- "loss": 1.5145,
7212
- "step": 102900
7213
- },
7214
- {
7215
- "epoch": 0.07659928903960853,
7216
- "grad_norm": 0.6128048300743103,
7217
- "learning_rate": 1.998842145806684e-05,
7218
- "loss": 1.5729,
7219
- "step": 103000
7220
- },
7221
- {
7222
- "epoch": 0.07667365728139455,
7223
- "grad_norm": 0.7574602365493774,
7224
- "learning_rate": 1.998839896780511e-05,
7225
- "loss": 1.4356,
7226
- "step": 103100
7227
  }
7228
  ],
7229
  "logging_steps": 100,
@@ -7243,7 +7089,7 @@
7243
  "attributes": {}
7244
  }
7245
  },
7246
- "total_flos": 1.4045543416669962e+18,
7247
  "train_batch_size": 1,
7248
  "trial_name": null,
7249
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.07503755596210195,
5
  "eval_steps": 200,
6
+ "global_step": 100900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
7070
  "learning_rate": 1.9988888715649357e-05,
7071
  "loss": 1.5441,
7072
  "step": 100900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7073
  }
7074
  ],
7075
  "logging_steps": 100,
 
7089
  "attributes": {}
7090
  }
7091
  },
7092
+ "total_flos": 1.3747108667853128e+18,
7093
  "train_batch_size": 1,
7094
  "trial_name": null,
7095
  "trial_params": null