ShushantLLM commited on
Commit
ba224c5
1 Parent(s): 88347e0

Training complete

Browse files
Files changed (2) hide show
  1. README.md +8 -6
  2. generation_config.json +1 -1
README.md CHANGED
@@ -1,9 +1,11 @@
1
  ---
 
 
2
  base_model: meta-llama/Llama-2-7b-hf
3
  tags:
4
  - trl
5
  - sft
6
- - missing lyric Llama2 Large
7
  - generated_from_trainer
8
  datasets:
9
  - generator
@@ -45,7 +47,7 @@ The following hyperparameters were used during training:
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: constant_with_warmup
47
  - lr_scheduler_warmup_ratio: 0.04
48
- - num_epochs: 1
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
@@ -54,7 +56,7 @@ The following hyperparameters were used during training:
54
 
55
  ### Framework versions
56
 
57
- - Transformers 4.39.3
58
- - Pytorch 2.2.2+cu121
59
- - Datasets 2.18.0
60
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
+ license: llama2
4
  base_model: meta-llama/Llama-2-7b-hf
5
  tags:
6
  - trl
7
  - sft
8
+ - LyricGeneration Again
9
  - generated_from_trainer
10
  datasets:
11
  - generator
 
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: constant_with_warmup
49
  - lr_scheduler_warmup_ratio: 0.04
50
+ - num_epochs: 2
51
  - mixed_precision_training: Native AMP
52
 
53
  ### Training results
 
56
 
57
  ### Framework versions
58
 
59
+ - Transformers 4.44.2
60
+ - Pytorch 2.4.0+cu121
61
+ - Datasets 3.0.0
62
+ - Tokenizers 0.19.1
generation_config.json CHANGED
@@ -6,5 +6,5 @@
6
  "pad_token_id": 0,
7
  "temperature": 0.6,
8
  "top_p": 0.9,
9
- "transformers_version": "4.39.3"
10
  }
 
6
  "pad_token_id": 0,
7
  "temperature": 0.6,
8
  "top_p": 0.9,
9
+ "transformers_version": "4.44.2"
10
  }