lucyknada commited on
Commit
704f2cc
1 Parent(s): 81d2cdb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -59
README.md CHANGED
@@ -1,28 +1,61 @@
1
  ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: mistralai/Mistral-Nemo-Instruct-2407
 
5
  tags:
6
- - axolotl
7
- - generated_from_trainer
8
- model-index:
9
- - name: magnum-v4-12b-r2
10
- results: []
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
 
16
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  <details><summary>See axolotl config</summary>
18
 
19
- axolotl version: `0.4.1`
20
  ```yaml
21
  base_model: mistralai/Mistral-Nemo-Instruct-2407
22
  model_type: AutoModelForCausalLM
23
  tokenizer_type: AutoTokenizer
24
 
25
- hub_model_id: anthracite-core/magnum-v4-12b-r2
26
  hub_strategy: "all_checkpoints"
27
  push_dataset_to_hub:
28
  hf_use_auth_token: true
@@ -39,17 +72,17 @@ load_in_4bit: false
39
  strict: false
40
 
41
  datasets:
42
- - path: anthracite-core/c2_logs_32k_llama3_qwen2_v1.2_no_system
43
  type: custommistralv3tekken
44
- - path: anthracite-core/kalo-opus-instruct-22k-no-refusal-no-system
45
  type: custommistralv3tekken
46
- - path: anthracite-core/kalo-opus-instruct-3k-filtered-no-system
47
  type: custommistralv3tekken
48
  - path: anthracite-org/nopm_claude_writing_fixed
49
  type: custommistralv3tekken
50
- - path: anthracite-core/kalo_opus_misc_240827_no_system
51
  type: custommistralv3tekken
52
- - path: anthracite-core/kalo_misc_part2_no_system
53
  type: custommistralv3tekken
54
  #chat_template: chatml
55
  shuffle_merged_datasets: true
@@ -110,51 +143,25 @@ fsdp_config:
110
  special_tokens:
111
  pad_token: <pad>
112
  ```
113
-
114
  </details><br>
115
 
116
- # magnum-v4-12b-r2
117
-
118
- This model is a fine-tuned version of [mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) on the None dataset.
119
-
120
- ## Model description
121
-
122
- More information needed
123
-
124
- ## Intended uses & limitations
125
-
126
- More information needed
127
-
128
- ## Training and evaluation data
129
-
130
- More information needed
131
-
132
- ## Training procedure
133
-
134
- ### Training hyperparameters
135
-
136
- The following hyperparameters were used during training:
137
- - learning_rate: 1e-05
138
- - train_batch_size: 1
139
- - eval_batch_size: 1
140
- - seed: 42
141
- - distributed_type: multi-GPU
142
- - num_devices: 8
143
- - gradient_accumulation_steps: 2
144
- - total_train_batch_size: 16
145
- - total_eval_batch_size: 8
146
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
147
- - lr_scheduler_type: cosine
148
- - lr_scheduler_warmup_steps: 40
149
- - num_epochs: 2
150
 
151
- ### Training results
152
 
 
 
 
 
 
 
 
153
 
 
 
154
 
155
- ### Framework versions
156
 
157
- - Transformers 4.45.0.dev0
158
- - Pytorch 2.3.1+cu121
159
- - Datasets 2.21.0
160
- - Tokenizers 0.19.1
 
1
  ---
2
+ license: other
3
+ license_name: mrl
4
+ language:
5
+ - en
6
  tags:
7
+ - chat
8
+ pipeline_tag: text-generation
9
+ library_name: transformers
 
 
10
  ---
11
 
 
 
12
 
13
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/-UC6YN1Gt3e1FDh8EqyaB.png)
14
+
15
+
16
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
17
+
18
+
19
+
20
+ This model is fine-tuned on top of [mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407).
21
+
22
+ ## Prompting
23
+ A typical input would look like this:
24
+
25
+ ```py
26
+ <s>[INST] SYSTEM MESSAGE
27
+ USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
28
+ ```
29
+
30
+ ## SillyTavern templates
31
+
32
+ Below are Instruct and Context templates for use within SillyTavern.
33
+
34
+ <details><summary>context template</summary>
35
+
36
+ ```yaml
37
+ default SillyTavern template works fine
38
+ ```
39
+
40
+ </details><br>
41
+ <details><summary>instruct template</summary>
42
+
43
+ ```yaml
44
+ default SillyTavern template works fine
45
+ ```
46
+
47
+ </details><br>
48
+
49
+ ## Axolotl config
50
+
51
  <details><summary>See axolotl config</summary>
52
 
 
53
  ```yaml
54
  base_model: mistralai/Mistral-Nemo-Instruct-2407
55
  model_type: AutoModelForCausalLM
56
  tokenizer_type: AutoTokenizer
57
 
58
+ hub_model_id: anthracite-org/magnum-v4-12b-r2
59
  hub_strategy: "all_checkpoints"
60
  push_dataset_to_hub:
61
  hf_use_auth_token: true
 
72
  strict: false
73
 
74
  datasets:
75
+ - path: anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system
76
  type: custommistralv3tekken
77
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
78
  type: custommistralv3tekken
79
+ - path: anthracite-org/kalo-opus-instruct-3k-filtered-no-system
80
  type: custommistralv3tekken
81
  - path: anthracite-org/nopm_claude_writing_fixed
82
  type: custommistralv3tekken
83
+ - path: anthracite-org/kalo_opus_misc_240827_no_system
84
  type: custommistralv3tekken
85
+ - path: anthracite-org/kalo_misc_part2_no_system
86
  type: custommistralv3tekken
87
  #chat_template: chatml
88
  shuffle_merged_datasets: true
 
143
  special_tokens:
144
  pad_token: <pad>
145
  ```
 
146
  </details><br>
147
 
148
+ ## Credits
149
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
150
 
151
+ We would also like to thank all members of Anthracite who made this finetune possible.
152
 
153
+ ## Datasets
154
+ - [anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system](https://huggingface.co/datasets/anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system)
155
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system)
156
+ - [anthracite-org/kalo-opus-instruct-3k-filtered-no-system](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-3k-filtered-no-system)
157
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
158
+ - [anthracite-org/kalo_opus_misc_240827_no_system](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827_no_system)
159
+ - [anthracite-org/kalo_misc_part2_no_system](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2_no_system)
160
 
161
+ ## Training
162
+ The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
163
 
164
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
165
 
166
+ ## Safety
167
+ ...