Delta-Vector commited on
Commit
7367e7d
1 Parent(s): 5bc2301

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -18
README.md CHANGED
@@ -17,9 +17,13 @@ datasets:
17
  - anthracite-org/kalo_misc_part2
18
  tags:
19
  - chat
 
 
 
 
20
  ---
21
 
22
-
23
 
24
 
25
  # Quants
@@ -52,7 +56,6 @@ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay
52
  Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
53
 
54
  <Guidelines>
55
- • Write upto 200 words.
56
  • Maintain the character persona but allow it to evolve with the story.
57
  • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
58
  • All types of outputs are encouraged; respond accordingly to the narrative.
@@ -66,7 +69,6 @@ Currently, your role is {{char}}, described in detail below. As {{char}}, contin
66
  </Guidelines>
67
 
68
  <Forbidden>
69
- • Writing more then 200 words.
70
  • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
71
  • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
72
  • Repetitive and monotonous outputs.
@@ -102,7 +104,7 @@ load_in_4bit: false
102
  strict: false
103
 
104
  datasets:
105
- - path: anthracite-core/c2_logs_16k_llama_v1.1
106
  type: sharegpt
107
  conversation: chatml
108
  - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
@@ -201,21 +203,10 @@ special_tokens:
201
  - [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
202
  - [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
203
  - [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
204
- - [anthracite-core/c2_logs_16k_llama_v1.1](https://huggingface.co/datasets/anthracite-core/c2_logs_16k_llama_v1.1)
205
 
206
 
207
  ## Training
208
- The training was done for 2 epochs. We used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
209
-
210
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
211
-
212
- ## Safety
213
-
214
- Avoid misusing this model, or you’ll need a ‘clicker’ to reset reality. ;)
215
-
216
- ## Musings
217
-
218
- One of the members of Anthracite had quite an interesting idea, to finetune a smaller model for 4 epochs at a lower Learning rate as quote "Smaller models learn slower" - [Kalomaze](https://huggingface.co/kalomaze) provided access to 10 X A40s and We finetuned what now is [Tor-8B]() for 2.5 epochs (and it's 4 Epoch version released as [Darkens-8B]()) and the result was quite impressive and the same configuration being used to train [Magnum=9B-V4] & [Odin-9B]. We also finetuned the model at above the 8192 context length to see if the model could "heal" in a way to a context length of 16384 with Needle tests coming soon ;)
219
-
220
-
221
 
 
 
17
  - anthracite-org/kalo_misc_part2
18
  tags:
19
  - chat
20
+ language:
21
+ - en
22
+ base_model:
23
+ - nvidia/Mistral-NeMo-Minitron-8B-Base
24
  ---
25
 
26
+ An earlier checkpoint of [Darkens-8B] using the same configuration, Finetuned ontop of the Prune/Distill NeMo 8B done by Nvidia, This model aims to have generally good prose and writing.
27
 
28
 
29
  # Quants
 
56
  Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
57
 
58
  <Guidelines>
 
59
  • Maintain the character persona but allow it to evolve with the story.
60
  • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
61
  • All types of outputs are encouraged; respond accordingly to the narrative.
 
69
  </Guidelines>
70
 
71
  <Forbidden>
 
72
  • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
73
  • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
74
  • Repetitive and monotonous outputs.
 
104
  strict: false
105
 
106
  datasets:
107
+ - path: PRIVATE CLAUDE LOG FILTER
108
  type: sharegpt
109
  conversation: chatml
110
  - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
 
203
  - [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
204
  - [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
205
  - [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
206
+ - [Private Claude Log filter](https://google.com)
207
 
208
 
209
  ## Training
210
+ The training was done for 4 epochs. (This model is the 2 epoch checkpoint), I used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
 
 
 
 
 
 
 
 
 
 
 
 
211
 
212
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)