bilkultheek commited on
Commit
6410431
1 Parent(s): be3bcd6

End of training

Browse files
README.md CHANGED
@@ -1,23 +1,23 @@
1
  ---
2
- base_model: ahxt/llama1_s_1.8B_experimental
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  model-index:
9
- - name: Cold-Rec-LLama-2-7B
10
  results: []
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- # Cold-Rec-LLama-2-7B
17
 
18
- This model is a fine-tuned version of [ahxt/llama1_s_1.8B_experimental](https://huggingface.co/ahxt/llama1_s_1.8B_experimental) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.0777
21
 
22
  ## Model description
23
 
@@ -36,7 +36,7 @@ More information needed
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - learning_rate: 0.001
40
  - train_batch_size: 16
41
  - eval_batch_size: 32
42
  - seed: 42
@@ -51,11 +51,11 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
- | 0.2217 | 1.992 | 249 | 0.2119 |
55
- | 0.1599 | 3.984 | 498 | 0.1558 |
56
- | 0.1177 | 5.976 | 747 | 0.1157 |
57
- | 0.0842 | 7.968 | 996 | 0.0875 |
58
- | 0.075 | 9.96 | 1245 | 0.0776 |
59
 
60
 
61
  ### Framework versions
 
1
  ---
2
+ base_model: NousResearch/Llama-2-7b-hf
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
  model-index:
9
+ - name: Cold-Data-LLama-2-7B
10
  results: []
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # Cold-Data-LLama-2-7B
17
 
18
+ This model is a fine-tuned version of [NousResearch/Llama-2-7b-hf](https://huggingface.co/NousResearch/Llama-2-7b-hf) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.0526
21
 
22
  ## Model description
23
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - learning_rate: 0.002
40
  - train_batch_size: 16
41
  - eval_batch_size: 32
42
  - seed: 42
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 0.1019 | 1.992 | 249 | 0.1022 |
55
+ | 0.0542 | 3.984 | 498 | 0.0540 |
56
+ | 0.0508 | 5.976 | 747 | 0.0513 |
57
+ | 0.0479 | 7.968 | 996 | 0.0515 |
58
+ | 0.0472 | 9.96 | 1245 | 0.0537 |
59
 
60
 
61
  ### Framework versions
runs/Aug12_10-44-55_fastgpuserv/events.out.tfevents.1723720304.fastgpuserv.2458899.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c8160446538513f3946805cc0ef72e0e251a33474478fadbe680f3557397b78
3
+ size 359