Naveen9402 commited on
Commit
4c3cde2
1 Parent(s): 761f7bf

End of training

Browse files
README.md CHANGED
@@ -1,10 +1,11 @@
1
  ---
2
  license: mit
3
- base_model: TheBloke/zephyr-7B-alpha-GPTQ
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
 
8
  model-index:
9
  - name: zephyr-support-chatbot
10
  results: []
@@ -49,7 +50,8 @@ The following hyperparameters were used during training:
49
 
50
  ### Framework versions
51
 
52
- - Transformers 4.35.2
53
- - Pytorch 2.1.0+cu121
54
- - Datasets 2.17.1
55
- - Tokenizers 0.15.2
 
 
1
  ---
2
  license: mit
3
+ library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
+ base_model: TheBloke/zephyr-7B-alpha-GPTQ
9
  model-index:
10
  - name: zephyr-support-chatbot
11
  results: []
 
50
 
51
  ### Framework versions
52
 
53
+ - PEFT 0.10.0
54
+ - Transformers 4.38.2
55
+ - Pytorch 2.2.1+cu121
56
+ - Datasets 2.18.0
57
+ - Tokenizers 0.15.2
adapter_config.json CHANGED
@@ -6,6 +6,7 @@
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
@@ -23,5 +24,6 @@
23
  "q_proj"
24
  ],
25
  "task_type": "CAUSAL_LM",
 
26
  "use_rslora": false
27
  }
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
+ "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
 
24
  "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
+ "use_dora": false,
28
  "use_rslora": false
29
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b5abe7724ec1cbf95b57ea0d87f899dee5106c677de35ee7481761f4919bc39
3
  size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc9f8405f7ea0f046e619b1df2a73d9d04144e413e52fe0b1001542938c04a74
3
  size 27280152
runs/Mar26_08-43-59_532080ae23f2/events.out.tfevents.1711442653.532080ae23f2.757.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6119073929474ee4dda62aba026503a9d178f919f2c6f921bfcd72edaa9cdf
3
+ size 5614
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<unk>",
 
1
  {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba6a9ac7bc293bda740788137ef5e495c6e106a14292821626df088cf321df2d
3
- size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f36fd0f1fc7a06f82aa73446f7761c7b70930e258c9bc92ac2878761f2e5264e
3
+ size 4920