Federic commited on
Commit
b0a84bf
β€’
1 Parent(s): 2745b8a

Training in progress, step 25

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "defog/sqlcoder-7b-2",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -9,7 +9,7 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 32,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
@@ -20,9 +20,9 @@
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
23
- "q_proj",
24
  "k_proj",
25
- "o_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-2-13b-hf",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 64,
13
  "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
 
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
 
23
  "k_proj",
24
+ "o_proj",
25
+ "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f567278abac0a182c45835aeeebcf74ff1ae9af34c0dbacb7ab3fe490aa5b1e
3
- size 536906096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:335ff3ab02635313a29bd60a29e36f5945243b709320d83cfa5ece896a4ce0b6
3
+ size 838904832
special_tokens_map.json CHANGED
@@ -1,14 +1,4 @@
1
  {
2
- "additional_special_tokens": [
3
- "▁<PRE>",
4
- "▁<MID>",
5
- "▁<SUF>",
6
- "▁<EOT>",
7
- "▁<PRE>",
8
- "▁<MID>",
9
- "▁<SUF>",
10
- "▁<EOT>"
11
- ],
12
  "bos_token": {
13
  "content": "<s>",
14
  "lstrip": false,
 
1
  {
 
 
 
 
 
 
 
 
 
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
tokenizer.json CHANGED
@@ -34,42 +34,6 @@
34
  "rstrip": false,
35
  "normalized": false,
36
  "special": true
37
- },
38
- {
39
- "id": 32007,
40
- "content": "▁<PRE>",
41
- "single_word": false,
42
- "lstrip": false,
43
- "rstrip": false,
44
- "normalized": false,
45
- "special": true
46
- },
47
- {
48
- "id": 32008,
49
- "content": "▁<SUF>",
50
- "single_word": false,
51
- "lstrip": false,
52
- "rstrip": false,
53
- "normalized": false,
54
- "special": true
55
- },
56
- {
57
- "id": 32009,
58
- "content": "▁<MID>",
59
- "single_word": false,
60
- "lstrip": false,
61
- "rstrip": false,
62
- "normalized": false,
63
- "special": true
64
- },
65
- {
66
- "id": 32010,
67
- "content": "▁<EOT>",
68
- "single_word": false,
69
- "lstrip": false,
70
- "rstrip": false,
71
- "normalized": false,
72
- "special": true
73
  }
74
  ],
75
  "normalizer": {
@@ -32175,23 +32139,7 @@
32175
  "μ™•": 31996,
32176
  "ζ”Ά": 31997,
32177
  "弘": 31998,
32178
- "η»™": 31999,
32179
- "▁<SU": 32000,
32180
- "▁<SUF": 32001,
32181
- "▁<PRE": 32002,
32182
- "▁<M": 32003,
32183
- "▁<MID": 32004,
32184
- "▁<E": 32005,
32185
- "▁<EOT": 32006,
32186
- "▁<PRE>": 32007,
32187
- "▁<SUF>": 32008,
32188
- "▁<MID>": 32009,
32189
- "▁<EOT>": 32010,
32190
- "▁<EOT><EOT>": 32011,
32191
- "▁<EOT><EOT><EOT>": 32012,
32192
- "▁<EOT><EOT><EOT><EOT>": 32013,
32193
- "▁<EOT><EOT><EOT><EOT><EOT>": 32014,
32194
- "▁<EOT><EOT><EOT><EOT><EOT><EOT>": 32015
32195
  },
32196
  "merges": [
32197
  "▁ t",
@@ -93442,18 +93390,7 @@
93442
  "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93443
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93444
  "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93445
- "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
93446
- "▁< SU",
93447
- "▁<SU F",
93448
- "▁< PRE",
93449
- "▁< M",
93450
- "▁<M ID",
93451
- "▁< E",
93452
- "▁<E OT",
93453
- "▁<PRE >",
93454
- "▁<SUF >",
93455
- "▁<MID >",
93456
- "▁<EOT >"
93457
  ]
93458
  }
93459
  }
 
34
  "rstrip": false,
35
  "normalized": false,
36
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  }
38
  ],
39
  "normalizer": {
 
32139
  "μ™•": 31996,
32140
  "ζ”Ά": 31997,
32141
  "弘": 31998,
32142
+ "η»™": 31999
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32143
  },
32144
  "merges": [
32145
  "▁ t",
 
93390
  "▁▁▁▁▁▁▁▁▁ ▁▁▁▁▁▁",
93391
  "▁▁▁▁▁▁▁ ▁▁▁▁▁▁▁▁",
93392
  "▁▁▁▁▁▁▁▁▁▁▁ ▁▁▁▁",
93393
+ "▁ ▁▁▁▁▁▁▁▁▁▁▁▁▁▁"
 
 
 
 
 
 
 
 
 
 
 
93394
  ]
93395
  }
93396
  }
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45ccb9c8b6b561889acea59191d66986d314e7cbd6a78abc6e49b139ca91c1e6
3
- size 500058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json CHANGED
@@ -25,63 +25,17 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
- },
29
- "32007": {
30
- "content": "▁<PRE>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "32008": {
38
- "content": "▁<SUF>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "32009": {
46
- "content": "▁<MID>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "32010": {
54
- "content": "▁<EOT>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
  }
61
  },
62
- "additional_special_tokens": [
63
- "▁<PRE>",
64
- "▁<MID>",
65
- "▁<SUF>",
66
- "▁<EOT>",
67
- "▁<PRE>",
68
- "▁<MID>",
69
- "▁<SUF>",
70
- "▁<EOT>"
71
- ],
72
  "bos_token": "<s>",
73
  "clean_up_tokenization_spaces": false,
74
  "eos_token": "</s>",
75
- "eot_token": "▁<EOT>",
76
- "fill_token": "<FILL_ME>",
77
- "legacy": null,
78
- "middle_token": "▁<MID>",
79
  "model_max_length": 1000000000000000019884624838656,
80
  "pad_token": "</s>",
81
- "prefix_token": "▁<PRE>",
82
  "sp_model_kwargs": {},
83
- "suffix_token": "▁<SUF>",
84
- "tokenizer_class": "CodeLlamaTokenizer",
85
  "unk_token": "<unk>",
86
  "use_default_system_prompt": false
87
  }
 
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  }
29
  },
 
 
 
 
 
 
 
 
 
 
30
  "bos_token": "<s>",
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
33
+ "legacy": false,
 
 
 
34
  "model_max_length": 1000000000000000019884624838656,
35
  "pad_token": "</s>",
36
+ "padding_side": "right",
37
  "sp_model_kwargs": {},
38
+ "tokenizer_class": "LlamaTokenizer",
 
39
  "unk_token": "<unk>",
40
  "use_default_system_prompt": false
41
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bd37a7bed4b29db3131947c878454b234a832069e5051a4e513a682b0b6b1f6
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:229196146870c6d4028c2c7478b67450b5d84079445d42ab8c78010679048a83
3
  size 4856