rAIfle commited on
Commit
bc260e4
1 Parent(s): accb54a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -0
README.md ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-2.0
3
+ tags:
4
+ - not-for-all-audiences
5
+ ---
6
+
7
+ ```
8
+ e88 88e d8
9
+ d888 888b 8888 8888 ,"Y88b 888 8e d88
10
+ C8888 8888D 8888 8888 "8" 888 888 88b d88888
11
+ Y888 888P Y888 888P ,ee 888 888 888 888
12
+ "88 88" "88 88" "88 888 888 888 888
13
+ b
14
+ 8b,
15
+
16
+ e88'Y88 d8 888
17
+ d888 'Y ,"Y88b 888,8, d88 ,e e, 888
18
+ C8888 "8" 888 888 " d88888 d88 88b 888
19
+ Y888 ,d ,ee 888 888 888 888 , 888
20
+ "88,d88 "88 888 888 888 "YeeP" 888
21
+
22
+ PROUDLY PRESENTS
23
+ ```
24
+ # MN-12B-Tarsus-exl2-longcal
25
+
26
+ Quantized using 115 rows of 8192 tokens from the default ExLlamav2-calibration dataset.
27
+
28
+ Branches:
29
+ - `main` -- `measurement.json`
30
+ - `8b8h` -- 8bpw, 8bit lm_head
31
+ - `6b6h` -- 6bpw, 6bit lm_head
32
+ - `4b6h` -- 4bpw, 6bit lm_head
33
+ - `3b6h` -- 3bpw, 6bit lm_head
34
+ - `2.25b6h` -- 2.25bpw, 6bit lm_head
35
+
36
+ Original model link: [Envoid/MN-12B-Tarsus](https://huggingface.co/Envoid/MN-12B-Tarsus)
37
+
38
+ Original model README below.
39
+
40
+ -----
41
+ ## CAUTION: This model was finetuned on a corpus that includes adult content and may produce mature content without warning.
42
+ ![](https://files.catbox.moe/1k5ama.jpg)
43
+
44
+ # MN-12B-Tarsus
45
+
46
+ Is a full-weight finetune of [mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)
47
+
48
+ Which underwent several intermediate steps.
49
+
50
+ This finetune was made with chatting/roleplaying via SillyTavern in mind and thus all of the testing was done there, with the goals being to:
51
+
52
+ -Reduce shiver-slop
53
+
54
+ -Make the model more conversationally proactive
55
+
56
+ -Give it a more human-like output (i.e. less gratuitous purple prose)
57
+
58
+ -Reducing overall positivity bias
59
+
60
+ It still responds well to Mistral-Instruct formatting.
61
+
62
+ The results are imperfect and its assistant capabilities suffered somewhat as a result but in quick testing it definitely seems to have achieved all of the goals to varying degrees.
63
+
64
+ It sometimes fumbles with tokens in odd places so it's certainly not perfect. Possibly best used as merge-fodder.
65
+
66
+ Trained using [qlora-pipe](https://github.com/tdrussell/qlora-pipe)