aariz120 commited on
Commit
561dcbc
1 Parent(s): 8aae21b

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
+ "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5afd1d9afc398730f93a89bf7e1026812e40aa45693a26b9d33709ded501a76f
3
  size 4517152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43d0898db8ec827d1cc44f1a4b2a39f228d9c7298e8cc72f3f24a5c997fdd622
3
  size 4517152
runs/May28_06-41-04_11a1ae0cd486/events.out.tfevents.1716878469.11a1ae0cd486.743.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48a7dd272b1733b97c59589491b8848f37df541e5f250c18e633fbebe8e505c
3
+ size 7075
tokenizer_config.json CHANGED
@@ -31,7 +31,6 @@
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
- "legacy": false,
35
  "max_length": 512,
36
  "model_max_length": 2048,
37
  "pad_token": "</s>",
 
31
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
 
34
  "max_length": 512,
35
  "model_max_length": 2048,
36
  "pad_token": "</s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d450fdc6d9f438d013ba74651fd66cf0c1b08211d10794a26dbc1124277d2acf
3
- size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5af128cee958bda0b3b438e21783eca22b890475ae80b372f6f949e294eb0178
3
+ size 5112