Femboyuwu2000 commited on
Commit
3bf072f
1 Parent(s): 92b0d29

Training in progress, step 20, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -20,6 +20,7 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "query_key_value"
24
  ],
25
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "word_embeddings",
24
  "query_key_value"
25
  ],
26
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82a235dcf79ae46d5d74eb2c277b2a88d917b4d0fe58e9c0483a1c6ef77bd3d0
3
- size 4725640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b717fc47fc6ca0f43cd64801c3f49ad862250ba32a036c80639250fcb7e18cd
3
+ size 12803224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cafd0997844310ce6894d5bb35d3a49ea4d5dabe1e810d400c91685f3a3abb0
3
- size 2423738
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5295139bb5ea53c3cb17ab006d58608d0220293d8748ba7d455a435ffff93806
3
+ size 6472698
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac4bf1426427b0e2a018f4b1750c36bb7e3e5ebba9e8a477845308163b2a7fb5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df81fc9d078b3589f3e74ce10344ae2b9ba2f47edb0c855a81de1c3ca1991673
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba9088914de78dc99d767c21058d110a360c0bb50564c9ed4886490601d97b6d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fcf032e6256b2cd4d659b09985077ab1961c6c13c26975c094dd36ab0c1ff74
3
  size 1064
last-checkpoint/tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12cb93e09fed5cf770f42b645926821cd300de44a57db3ca5e284142571864c2
3
- size 14500638
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
3
+ size 14500471
last-checkpoint/trainer_state.json CHANGED
@@ -1,26 +1,19 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.005451076587626056,
5
  "eval_steps": 500,
6
- "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
- "grad_norm": 2.1512844562530518,
14
- "learning_rate": 1.5833333333333336e-06,
15
- "loss": 4.4875,
16
  "step": 20
17
- },
18
- {
19
- "epoch": 0.01,
20
- "grad_norm": 2.2963316440582275,
21
- "learning_rate": 3.2500000000000002e-06,
22
- "loss": 4.1113,
23
- "step": 40
24
  }
25
  ],
26
  "logging_steps": 20,
@@ -28,7 +21,7 @@
28
  "num_input_tokens_seen": 0,
29
  "num_train_epochs": 3,
30
  "save_steps": 20,
31
- "total_flos": 334860439191552.0,
32
  "train_batch_size": 1,
33
  "trial_name": null,
34
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.002725538293813028,
5
  "eval_steps": 500,
6
+ "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.0,
13
+ "grad_norm": 36.511024475097656,
14
+ "learning_rate": 1.0000000000000002e-06,
15
+ "loss": 4.6323,
16
  "step": 20
 
 
 
 
 
 
 
17
  }
18
  ],
19
  "logging_steps": 20,
 
21
  "num_input_tokens_seen": 0,
22
  "num_train_epochs": 3,
23
  "save_steps": 20,
24
+ "total_flos": 167893794816000.0,
25
  "train_batch_size": 1,
26
  "trial_name": null,
27
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2144efe638cdc54910486057ddff9e1420f8f369906d08c03aeb6ed693951c8e
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e025a4e68c1d22e7dc695a50138fa18070a7258d30cfd192237c2cea9e5c360
3
  size 4984