hammamwahab commited on
Commit
6e6a5cb
1 Parent(s): 8f9d272

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -19,12 +19,12 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "k_proj",
23
  "up_proj",
24
- "gate_proj",
25
  "q_proj",
26
  "o_proj",
27
  "v_proj",
 
28
  "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "up_proj",
23
+ "k_proj",
24
  "q_proj",
25
  "o_proj",
26
  "v_proj",
27
+ "gate_proj",
28
  "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9492fc4bc31843c29a4fe872483bf78357710419af68475b78644551e1a5cc1f
3
  size 3443619040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a3b2bdaba5379297a84ec7a0bc298c8f07c3c25d877c1b12cae48775670a5b
3
  size 3443619040
runs/Jun01_08-22-29_e8ec2b83eda2/events.out.tfevents.1717230156.e8ec2b83eda2.4223.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3ff9fbcc1260aaad8ded83a206f870ddadcd0e3d2dad0d89ffdd103a90966ca
3
+ size 6411
tokenizer.json CHANGED
@@ -2347,10 +2347,69 @@
2347
  ]
2348
  },
2349
  "post_processor": {
2350
- "type": "ByteLevel",
2351
- "add_prefix_space": true,
2352
- "trim_offsets": false,
2353
- "use_regex": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2354
  },
2355
  "decoder": {
2356
  "type": "ByteLevel",
 
2347
  ]
2348
  },
2349
  "post_processor": {
2350
+ "type": "Sequence",
2351
+ "processors": [
2352
+ {
2353
+ "type": "ByteLevel",
2354
+ "add_prefix_space": true,
2355
+ "trim_offsets": false,
2356
+ "use_regex": true
2357
+ },
2358
+ {
2359
+ "type": "TemplateProcessing",
2360
+ "single": [
2361
+ {
2362
+ "SpecialToken": {
2363
+ "id": "<|begin_of_text|>",
2364
+ "type_id": 0
2365
+ }
2366
+ },
2367
+ {
2368
+ "Sequence": {
2369
+ "id": "A",
2370
+ "type_id": 0
2371
+ }
2372
+ }
2373
+ ],
2374
+ "pair": [
2375
+ {
2376
+ "SpecialToken": {
2377
+ "id": "<|begin_of_text|>",
2378
+ "type_id": 0
2379
+ }
2380
+ },
2381
+ {
2382
+ "Sequence": {
2383
+ "id": "A",
2384
+ "type_id": 0
2385
+ }
2386
+ },
2387
+ {
2388
+ "SpecialToken": {
2389
+ "id": "<|begin_of_text|>",
2390
+ "type_id": 1
2391
+ }
2392
+ },
2393
+ {
2394
+ "Sequence": {
2395
+ "id": "B",
2396
+ "type_id": 1
2397
+ }
2398
+ }
2399
+ ],
2400
+ "special_tokens": {
2401
+ "<|begin_of_text|>": {
2402
+ "id": "<|begin_of_text|>",
2403
+ "ids": [
2404
+ 128000
2405
+ ],
2406
+ "tokens": [
2407
+ "<|begin_of_text|>"
2408
+ ]
2409
+ }
2410
+ }
2411
+ }
2412
+ ]
2413
  },
2414
  "decoder": {
2415
  "type": "ByteLevel",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b7cf36829dd26d504423ee8ea0fb007ed93636b301e49356d7f07c61ace88a0
3
- size 4347
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c940ec85e55cb34809c3b7c7232c9a89a1e3e8e4d35706929438ceaee6ae8b7
3
+ size 4792