MohamedAhmedAE
commited on
Commit
•
0ebc3e1
1
Parent(s):
f4f54ca
Training in progress, step 104300
Browse files
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5544997664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e85942234ea7e7dab0e7cc03a56a969e30bc21546e008c54113522aa21fba1e
|
3 |
size 5544997664
|
last-checkpoint/adapter_config.json
CHANGED
@@ -20,12 +20,12 @@
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
-
"k_proj",
|
24 |
-
"up_proj",
|
25 |
-
"q_proj",
|
26 |
"v_proj",
|
27 |
-
"
|
28 |
"gate_proj",
|
|
|
|
|
|
|
29 |
"o_proj"
|
30 |
],
|
31 |
"task_type": "CAUSAL_LM",
|
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
|
|
|
|
|
|
23 |
"v_proj",
|
24 |
+
"k_proj",
|
25 |
"gate_proj",
|
26 |
+
"down_proj",
|
27 |
+
"q_proj",
|
28 |
+
"up_proj",
|
29 |
"o_proj"
|
30 |
],
|
31 |
"task_type": "CAUSAL_LM",
|
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5544997664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ff50e9a0eef14c00f32c5e550257295427f2d666e009aac32472aef43b0c78f
|
3 |
size 5544997664
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 674093138
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b1f3c8b9d1c8514057a760685418307853b2172387d395e371d81516debcc99
|
3 |
size 674093138
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d52e9778ae961a843d4efe5adba669832146332ec663eac9df46d71427724e3
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1dd725a3e5295711459643d6e1204a1d04a7f905cc6416544fa87ecdfb18228
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch": 0.
|
5 |
"eval_steps": 200,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -7070,237 +7070,6 @@
|
|
7070 |
"learning_rate": 1.9988888715649357e-05,
|
7071 |
"loss": 1.5441,
|
7072 |
"step": 100900
|
7073 |
-
},
|
7074 |
-
{
|
7075 |
-
"epoch": 0.07511192420388797,
|
7076 |
-
"grad_norm": 1.068562388420105,
|
7077 |
-
"learning_rate": 1.998886668338319e-05,
|
7078 |
-
"loss": 1.4998,
|
7079 |
-
"step": 101000
|
7080 |
-
},
|
7081 |
-
{
|
7082 |
-
"epoch": 0.075186292445674,
|
7083 |
-
"grad_norm": 0.6331286430358887,
|
7084 |
-
"learning_rate": 1.998884462930723e-05,
|
7085 |
-
"loss": 1.5633,
|
7086 |
-
"step": 101100
|
7087 |
-
},
|
7088 |
-
{
|
7089 |
-
"epoch": 0.07526066068746003,
|
7090 |
-
"grad_norm": 1.3566038608551025,
|
7091 |
-
"learning_rate": 1.998882255342152e-05,
|
7092 |
-
"loss": 1.4621,
|
7093 |
-
"step": 101200
|
7094 |
-
},
|
7095 |
-
{
|
7096 |
-
"epoch": 0.07533502892924605,
|
7097 |
-
"grad_norm": 0.9672004580497742,
|
7098 |
-
"learning_rate": 1.998880045572611e-05,
|
7099 |
-
"loss": 1.5249,
|
7100 |
-
"step": 101300
|
7101 |
-
},
|
7102 |
-
{
|
7103 |
-
"epoch": 0.07540939717103208,
|
7104 |
-
"grad_norm": 0.36732280254364014,
|
7105 |
-
"learning_rate": 1.9988778336221045e-05,
|
7106 |
-
"loss": 1.574,
|
7107 |
-
"step": 101400
|
7108 |
-
},
|
7109 |
-
{
|
7110 |
-
"epoch": 0.07548376541281811,
|
7111 |
-
"grad_norm": 0.4788234829902649,
|
7112 |
-
"learning_rate": 1.998875619490638e-05,
|
7113 |
-
"loss": 1.5418,
|
7114 |
-
"step": 101500
|
7115 |
-
},
|
7116 |
-
{
|
7117 |
-
"epoch": 0.07555813365460413,
|
7118 |
-
"grad_norm": 0.8955681324005127,
|
7119 |
-
"learning_rate": 1.9988734031782157e-05,
|
7120 |
-
"loss": 1.5568,
|
7121 |
-
"step": 101600
|
7122 |
-
},
|
7123 |
-
{
|
7124 |
-
"epoch": 0.07563250189639016,
|
7125 |
-
"grad_norm": 0.8049163222312927,
|
7126 |
-
"learning_rate": 1.9988711846848427e-05,
|
7127 |
-
"loss": 1.4838,
|
7128 |
-
"step": 101700
|
7129 |
-
},
|
7130 |
-
{
|
7131 |
-
"epoch": 0.07570687013817619,
|
7132 |
-
"grad_norm": 0.7558008432388306,
|
7133 |
-
"learning_rate": 1.9988689640105235e-05,
|
7134 |
-
"loss": 1.4955,
|
7135 |
-
"step": 101800
|
7136 |
-
},
|
7137 |
-
{
|
7138 |
-
"epoch": 0.07578123837996222,
|
7139 |
-
"grad_norm": 0.4749026596546173,
|
7140 |
-
"learning_rate": 1.9988667411552635e-05,
|
7141 |
-
"loss": 1.5929,
|
7142 |
-
"step": 101900
|
7143 |
-
},
|
7144 |
-
{
|
7145 |
-
"epoch": 0.07585560662174824,
|
7146 |
-
"grad_norm": 0.6597522497177124,
|
7147 |
-
"learning_rate": 1.998864516119067e-05,
|
7148 |
-
"loss": 1.5584,
|
7149 |
-
"step": 102000
|
7150 |
-
},
|
7151 |
-
{
|
7152 |
-
"epoch": 0.07592997486353428,
|
7153 |
-
"grad_norm": 0.7412188053131104,
|
7154 |
-
"learning_rate": 1.9988622889019395e-05,
|
7155 |
-
"loss": 1.5842,
|
7156 |
-
"step": 102100
|
7157 |
-
},
|
7158 |
-
{
|
7159 |
-
"epoch": 0.07600434310532031,
|
7160 |
-
"grad_norm": 0.5564984679222107,
|
7161 |
-
"learning_rate": 1.9988600595038853e-05,
|
7162 |
-
"loss": 1.5764,
|
7163 |
-
"step": 102200
|
7164 |
-
},
|
7165 |
-
{
|
7166 |
-
"epoch": 0.07607871134710634,
|
7167 |
-
"grad_norm": 1.0488529205322266,
|
7168 |
-
"learning_rate": 1.9988578279249097e-05,
|
7169 |
-
"loss": 1.458,
|
7170 |
-
"step": 102300
|
7171 |
-
},
|
7172 |
-
{
|
7173 |
-
"epoch": 0.07615307958889236,
|
7174 |
-
"grad_norm": 1.40269136428833,
|
7175 |
-
"learning_rate": 1.998855594165017e-05,
|
7176 |
-
"loss": 1.4588,
|
7177 |
-
"step": 102400
|
7178 |
-
},
|
7179 |
-
{
|
7180 |
-
"epoch": 0.07622744783067839,
|
7181 |
-
"grad_norm": 0.8488138318061829,
|
7182 |
-
"learning_rate": 1.9988533582242127e-05,
|
7183 |
-
"loss": 1.522,
|
7184 |
-
"step": 102500
|
7185 |
-
},
|
7186 |
-
{
|
7187 |
-
"epoch": 0.07630181607246442,
|
7188 |
-
"grad_norm": 0.5191701054573059,
|
7189 |
-
"learning_rate": 1.9988511201025015e-05,
|
7190 |
-
"loss": 1.5036,
|
7191 |
-
"step": 102600
|
7192 |
-
},
|
7193 |
-
{
|
7194 |
-
"epoch": 0.07637618431425044,
|
7195 |
-
"grad_norm": 0.6648279428482056,
|
7196 |
-
"learning_rate": 1.9988488797998878e-05,
|
7197 |
-
"loss": 1.4929,
|
7198 |
-
"step": 102700
|
7199 |
-
},
|
7200 |
-
{
|
7201 |
-
"epoch": 0.07645055255603647,
|
7202 |
-
"grad_norm": 1.8600202798843384,
|
7203 |
-
"learning_rate": 1.9988466373163774e-05,
|
7204 |
-
"loss": 1.5692,
|
7205 |
-
"step": 102800
|
7206 |
-
},
|
7207 |
-
{
|
7208 |
-
"epoch": 0.0765249207978225,
|
7209 |
-
"grad_norm": 0.7583739757537842,
|
7210 |
-
"learning_rate": 1.9988443926519743e-05,
|
7211 |
-
"loss": 1.5145,
|
7212 |
-
"step": 102900
|
7213 |
-
},
|
7214 |
-
{
|
7215 |
-
"epoch": 0.07659928903960853,
|
7216 |
-
"grad_norm": 0.6128048300743103,
|
7217 |
-
"learning_rate": 1.998842145806684e-05,
|
7218 |
-
"loss": 1.5729,
|
7219 |
-
"step": 103000
|
7220 |
-
},
|
7221 |
-
{
|
7222 |
-
"epoch": 0.07667365728139455,
|
7223 |
-
"grad_norm": 0.7574602365493774,
|
7224 |
-
"learning_rate": 1.998839896780511e-05,
|
7225 |
-
"loss": 1.4356,
|
7226 |
-
"step": 103100
|
7227 |
-
},
|
7228 |
-
{
|
7229 |
-
"epoch": 0.07674802552318058,
|
7230 |
-
"grad_norm": 1.4134727716445923,
|
7231 |
-
"learning_rate": 1.9988376455734606e-05,
|
7232 |
-
"loss": 1.5048,
|
7233 |
-
"step": 103200
|
7234 |
-
},
|
7235 |
-
{
|
7236 |
-
"epoch": 0.0768223937649666,
|
7237 |
-
"grad_norm": 0.7592337727546692,
|
7238 |
-
"learning_rate": 1.9988353921855374e-05,
|
7239 |
-
"loss": 1.4988,
|
7240 |
-
"step": 103300
|
7241 |
-
},
|
7242 |
-
{
|
7243 |
-
"epoch": 0.07689676200675263,
|
7244 |
-
"grad_norm": 0.522486686706543,
|
7245 |
-
"learning_rate": 1.9988331366167465e-05,
|
7246 |
-
"loss": 1.5654,
|
7247 |
-
"step": 103400
|
7248 |
-
},
|
7249 |
-
{
|
7250 |
-
"epoch": 0.07697113024853866,
|
7251 |
-
"grad_norm": 0.6535342335700989,
|
7252 |
-
"learning_rate": 1.9988308788670925e-05,
|
7253 |
-
"loss": 1.4593,
|
7254 |
-
"step": 103500
|
7255 |
-
},
|
7256 |
-
{
|
7257 |
-
"epoch": 0.07704549849032469,
|
7258 |
-
"grad_norm": 0.6663926243782043,
|
7259 |
-
"learning_rate": 1.9988286189365808e-05,
|
7260 |
-
"loss": 1.477,
|
7261 |
-
"step": 103600
|
7262 |
-
},
|
7263 |
-
{
|
7264 |
-
"epoch": 0.07711986673211071,
|
7265 |
-
"grad_norm": 0.5006215572357178,
|
7266 |
-
"learning_rate": 1.998826356825216e-05,
|
7267 |
-
"loss": 1.5326,
|
7268 |
-
"step": 103700
|
7269 |
-
},
|
7270 |
-
{
|
7271 |
-
"epoch": 0.07719423497389674,
|
7272 |
-
"grad_norm": 0.6826842427253723,
|
7273 |
-
"learning_rate": 1.9988240925330032e-05,
|
7274 |
-
"loss": 1.5102,
|
7275 |
-
"step": 103800
|
7276 |
-
},
|
7277 |
-
{
|
7278 |
-
"epoch": 0.07726860321568277,
|
7279 |
-
"grad_norm": 0.2680438756942749,
|
7280 |
-
"learning_rate": 1.9988218260599477e-05,
|
7281 |
-
"loss": 1.4773,
|
7282 |
-
"step": 103900
|
7283 |
-
},
|
7284 |
-
{
|
7285 |
-
"epoch": 0.07734297145746881,
|
7286 |
-
"grad_norm": 0.9159733057022095,
|
7287 |
-
"learning_rate": 1.9988195574060536e-05,
|
7288 |
-
"loss": 1.4984,
|
7289 |
-
"step": 104000
|
7290 |
-
},
|
7291 |
-
{
|
7292 |
-
"epoch": 0.07741733969925484,
|
7293 |
-
"grad_norm": 1.0930269956588745,
|
7294 |
-
"learning_rate": 1.9988172865713266e-05,
|
7295 |
-
"loss": 1.4196,
|
7296 |
-
"step": 104100
|
7297 |
-
},
|
7298 |
-
{
|
7299 |
-
"epoch": 0.07749170794104086,
|
7300 |
-
"grad_norm": 0.6656064391136169,
|
7301 |
-
"learning_rate": 1.998815013555771e-05,
|
7302 |
-
"loss": 1.5282,
|
7303 |
-
"step": 104200
|
7304 |
}
|
7305 |
],
|
7306 |
"logging_steps": 100,
|
@@ -7320,7 +7089,7 @@
|
|
7320 |
"attributes": {}
|
7321 |
}
|
7322 |
},
|
7323 |
-
"total_flos": 1.
|
7324 |
"train_batch_size": 1,
|
7325 |
"trial_name": null,
|
7326 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 0.07503755596210195,
|
5 |
"eval_steps": 200,
|
6 |
+
"global_step": 100900,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
7070 |
"learning_rate": 1.9988888715649357e-05,
|
7071 |
"loss": 1.5441,
|
7072 |
"step": 100900
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7073 |
}
|
7074 |
],
|
7075 |
"logging_steps": 100,
|
|
|
7089 |
"attributes": {}
|
7090 |
}
|
7091 |
},
|
7092 |
+
"total_flos": 1.3747108667853128e+18,
|
7093 |
"train_batch_size": 1,
|
7094 |
"trial_name": null,
|
7095 |
"trial_params": null
|