Text Generation
Transformers
Safetensors
English
llama
llama-factory
Not-For-All-Audiences
conversational
text-generation-inference
Inference Endpoints
aaronday3 commited on
Commit
6eface1
1 Parent(s): d37ed12

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -158,16 +158,14 @@ When convenient, say screenplay phrases like "cut to"
158
  The split was as follows:
159
 
160
  - **2K rows from r/WritingPrompts**
161
- - **2K rows from r/DirtyWritingPrompts**
162
- - **2K rows from Opus Instruct 15K (specifically the 6.5K jsonl)**
163
  - **2K rows from c2 logs cleaned**
164
 
165
  While we did train all system prompts from c2 logs we also have our own system prompts.
166
  <details>
167
  <summary>List of trained system prompts. Note: c2 logs system prompts and char cards were also included.</summary>
168
 
169
- Here's the information in a markdown table:
170
-
171
  | Dataset | System Prompt |
172
  |--------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------|
173
  | reddit_dirty_writing_prompts.jsonl | "You are a short story writer. Write a story based on prompt provided by user below. Mode: NSFW" |
@@ -191,7 +189,7 @@ We think there is too much secrecy around what data is being used, and different
191
  We found that increasing the amount of ranks from 64 to 256 has reduced repetition but also led to the language used resembling Claude more than the 64 rank version. No worries, it's still far enough from Claude.
192
  <br>It also led to increased coherency but reduced instruction following, likely because the model started diverging more away from L3 8B Instruct.
193
  <br>**The model is uncensored for RP. For Instruct it needs 2-3 words of prefill for the first message.**
194
- <br>We found that increasing the amount of data from 1K to 8K reduced repetition aswell.
195
  <br>**The prose is much better than other synthetic data generations. The model also demonstrates increased style copying abilities likely a result of the long data and varying writing styles found in WritingPrompts.**
196
  <br>**The model is exceptional at being creative in roleplaying**, knows different persona's and even a single character will change persona in different contexts, persona is tied to last few messages rather than system message or character card. **This is great as it often means the model can do impressive things without you needing to explicitly specify.**
197
 
@@ -206,8 +204,8 @@ Grad norm kept increasing throughout the run which is concerning, albeit it coul
206
 
207
  ## Graphs
208
  Colors:
209
- <p style="color: #F0B899;">256 rank on 8K rows</p>
210
- <p style="color: #5BC5DB;">64 rank on 8K rows</p>
211
  <p style="color: #5387DD;">64 rank on 1K rows</p>
212
 
213
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630cf5d14ca0a22768bbe10c/y9hC4bGq-Lt7sDQ23q5db.png)
 
158
  The split was as follows:
159
 
160
  - **2K rows from r/WritingPrompts**
161
+ - **1.1K rows from r/DirtyWritingPrompts**
162
+ - **1.4K rows from Opus Instruct 15K (specifically the 6.5K jsonl)**
163
  - **2K rows from c2 logs cleaned**
164
 
165
  While we did train all system prompts from c2 logs we also have our own system prompts.
166
  <details>
167
  <summary>List of trained system prompts. Note: c2 logs system prompts and char cards were also included.</summary>
168
 
 
 
169
  | Dataset | System Prompt |
170
  |--------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------|
171
  | reddit_dirty_writing_prompts.jsonl | "You are a short story writer. Write a story based on prompt provided by user below. Mode: NSFW" |
 
189
  We found that increasing the amount of ranks from 64 to 256 has reduced repetition but also led to the language used resembling Claude more than the 64 rank version. No worries, it's still far enough from Claude.
190
  <br>It also led to increased coherency but reduced instruction following, likely because the model started diverging more away from L3 8B Instruct.
191
  <br>**The model is uncensored for RP. For Instruct it needs 2-3 words of prefill for the first message.**
192
+ <br>We found that increasing the amount of data from 1K to 6.5K reduced repetition aswell.
193
  <br>**The prose is much better than other synthetic data generations. The model also demonstrates increased style copying abilities likely a result of the long data and varying writing styles found in WritingPrompts.**
194
  <br>**The model is exceptional at being creative in roleplaying**, knows different persona's and even a single character will change persona in different contexts, persona is tied to last few messages rather than system message or character card. **This is great as it often means the model can do impressive things without you needing to explicitly specify.**
195
 
 
204
 
205
  ## Graphs
206
  Colors:
207
+ <p style="color: #F0B899;">256 rank on 6.5K rows</p>
208
+ <p style="color: #5BC5DB;">64 rank on 6.5K rows</p>
209
  <p style="color: #5387DD;">64 rank on 1K rows</p>
210
 
211
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630cf5d14ca0a22768bbe10c/y9hC4bGq-Lt7sDQ23q5db.png)