Triangle104 commited on
Commit
fdfe893
1 Parent(s): 637185e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -0
README.md CHANGED
@@ -107,6 +107,78 @@ model-index:
107
  This model was converted to GGUF format from [`lemon07r/Gemma-2-Ataraxy-v2-9B`](https://huggingface.co/lemon07r/Gemma-2-Ataraxy-v2-9B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
108
  Refer to the [original model card](https://huggingface.co/lemon07r/Gemma-2-Ataraxy-v2-9B) for more details on the model.
109
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  ## Use with llama.cpp
111
  Install llama.cpp through brew (works on Mac and Linux)
112
 
 
107
  This model was converted to GGUF format from [`lemon07r/Gemma-2-Ataraxy-v2-9B`](https://huggingface.co/lemon07r/Gemma-2-Ataraxy-v2-9B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
108
  Refer to the [original model card](https://huggingface.co/lemon07r/Gemma-2-Ataraxy-v2-9B) for more details on the model.
109
 
110
+ ---
111
+ Model details:
112
+ -
113
+ Gemma 2 Ataraxy v2 9B
114
+
115
+ Finally, after much testing, a sucessor to the first Gemma 2 Ataraxy 9B. Same kind of recipe, using the same principles, same concept as the last Ataraxy. It's not quite a better overall model, v1 is more well rounded, v2 is a little better at writing but has a little more slop and some other issues. consider this a sidegrade.
116
+
117
+ Ataraxy
118
+ GGUF / EXL2 Quants
119
+
120
+ Bartowski quants (imatrix): https://huggingface.co/bartowski/Gemma-2-Ataraxy-v2-9B-GGUF
121
+
122
+ Mradermacher quants (static): https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v2-9B-GGUF
123
+
124
+ Mradermacher quants (imatrix): https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v2-9B-i1-GGUF
125
+
126
+ Bartowski and mradermacher use different calibration data for their imatrix quants I believe, and the static quant of course uses none. Pick your poison.
127
+
128
+ More coming soon.
129
+ Format
130
+
131
+ Use Gemma 2 format.
132
+ Merge Details
133
+ Merge Method
134
+
135
+ This model was merged using the SLERP merge method.
136
+ Models Merged
137
+
138
+ This is a merge of pre-trained language models created using mergekit.
139
+
140
+ The following models were included in the merge:
141
+
142
+ ifable/gemma-2-Ifable-9B
143
+ jsgreenawalt/gemma-2-9B-it-advanced-v2.1
144
+
145
+ Configuration
146
+
147
+ The following YAML configuration was used to produce this model:
148
+
149
+ base_model: ifable/gemma-2-Ifable-9B
150
+ dtype: bfloat16
151
+ merge_method: slerp
152
+ parameters:
153
+ t:
154
+ - filter: self_attn
155
+ value: [0.0, 0.5, 0.3, 0.7, 1.0]
156
+ - filter: mlp
157
+ value: [1.0, 0.5, 0.7, 0.3, 0.0]
158
+ - value: 0.5
159
+ slices:
160
+ - sources:
161
+ - layer_range: [0, 42]
162
+ model: jsgreenawalt/gemma-2-9B-it-advanced-v2.1
163
+ - layer_range: [0, 42]
164
+ model: ifable/gemma-2-Ifable-9B
165
+
166
+ Open LLM Leaderboard Evaluation Results
167
+
168
+ Detailed results can be found here
169
+ Metric Value
170
+ Avg. 19.16
171
+ IFEval (0-Shot) 21.36
172
+ BBH (3-Shot) 39.80
173
+ MATH Lvl 5 (4-Shot) 0.83
174
+ GPQA (0-shot) 12.30
175
+ MuSR (0-shot) 4.88
176
+ MMLU-PRO (5-shot) 35.79
177
+
178
+ Second highest ranked open weight model in EQ-Bench.
179
+
180
+ ---
181
+
182
  ## Use with llama.cpp
183
  Install llama.cpp through brew (works on Mac and Linux)
184