RichardErkhov commited on
Commit
a3a7c1e
1 Parent(s): fd68261

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +323 -0
README.md ADDED
@@ -0,0 +1,323 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ athene-noctua-13b - GGUF
11
+ - Model creator: https://huggingface.co/ibivibiv/
12
+ - Original model: https://huggingface.co/ibivibiv/athene-noctua-13b/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [athene-noctua-13b.Q2_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q2_K.gguf) | Q2_K | 4.52GB |
18
+ | [athene-noctua-13b.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.IQ3_XS.gguf) | IQ3_XS | 4.99GB |
19
+ | [athene-noctua-13b.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.IQ3_S.gguf) | IQ3_S | 5.27GB |
20
+ | [athene-noctua-13b.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q3_K_S.gguf) | Q3_K_S | 5.27GB |
21
+ | [athene-noctua-13b.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.IQ3_M.gguf) | IQ3_M | 5.57GB |
22
+ | [athene-noctua-13b.Q3_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q3_K.gguf) | Q3_K | 5.9GB |
23
+ | [athene-noctua-13b.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q3_K_M.gguf) | Q3_K_M | 5.9GB |
24
+ | [athene-noctua-13b.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q3_K_L.gguf) | Q3_K_L | 6.45GB |
25
+ | [athene-noctua-13b.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.IQ4_XS.gguf) | IQ4_XS | 6.54GB |
26
+ | [athene-noctua-13b.Q4_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q4_0.gguf) | Q4_0 | 6.86GB |
27
+ | [athene-noctua-13b.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.IQ4_NL.gguf) | IQ4_NL | 6.9GB |
28
+ | [athene-noctua-13b.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q4_K_S.gguf) | Q4_K_S | 6.91GB |
29
+ | [athene-noctua-13b.Q4_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q4_K.gguf) | Q4_K | 7.33GB |
30
+ | [athene-noctua-13b.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q4_K_M.gguf) | Q4_K_M | 7.33GB |
31
+ | [athene-noctua-13b.Q4_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q4_1.gguf) | Q4_1 | 7.61GB |
32
+ | [athene-noctua-13b.Q5_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q5_0.gguf) | Q5_0 | 8.36GB |
33
+ | [athene-noctua-13b.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q5_K_S.gguf) | Q5_K_S | 8.36GB |
34
+ | [athene-noctua-13b.Q5_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q5_K.gguf) | Q5_K | 8.6GB |
35
+ | [athene-noctua-13b.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q5_K_M.gguf) | Q5_K_M | 8.6GB |
36
+ | [athene-noctua-13b.Q5_1.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q5_1.gguf) | Q5_1 | 9.1GB |
37
+ | [athene-noctua-13b.Q6_K.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q6_K.gguf) | Q6_K | 9.95GB |
38
+ | [athene-noctua-13b.Q8_0.gguf](https://huggingface.co/RichardErkhov/ibivibiv_-_athene-noctua-13b-gguf/blob/main/athene-noctua-13b.Q8_0.gguf) | Q8_0 | 12.88GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ language:
46
+ - en
47
+ license: llama2
48
+ tags:
49
+ - logic
50
+ - reasoning
51
+ model-index:
52
+ - name: athene-noctua-13b
53
+ results:
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: AI2 Reasoning Challenge (25-Shot)
59
+ type: ai2_arc
60
+ config: ARC-Challenge
61
+ split: test
62
+ args:
63
+ num_few_shot: 25
64
+ metrics:
65
+ - type: acc_norm
66
+ value: 57.17
67
+ name: normalized accuracy
68
+ source:
69
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
70
+ name: Open LLM Leaderboard
71
+ - task:
72
+ type: text-generation
73
+ name: Text Generation
74
+ dataset:
75
+ name: HellaSwag (10-Shot)
76
+ type: hellaswag
77
+ split: validation
78
+ args:
79
+ num_few_shot: 10
80
+ metrics:
81
+ - type: acc_norm
82
+ value: 81.52
83
+ name: normalized accuracy
84
+ source:
85
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
86
+ name: Open LLM Leaderboard
87
+ - task:
88
+ type: text-generation
89
+ name: Text Generation
90
+ dataset:
91
+ name: MMLU (5-Shot)
92
+ type: cais/mmlu
93
+ config: all
94
+ split: test
95
+ args:
96
+ num_few_shot: 5
97
+ metrics:
98
+ - type: acc
99
+ value: 55.91
100
+ name: accuracy
101
+ source:
102
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
103
+ name: Open LLM Leaderboard
104
+ - task:
105
+ type: text-generation
106
+ name: Text Generation
107
+ dataset:
108
+ name: TruthfulQA (0-shot)
109
+ type: truthful_qa
110
+ config: multiple_choice
111
+ split: validation
112
+ args:
113
+ num_few_shot: 0
114
+ metrics:
115
+ - type: mc2
116
+ value: 47.49
117
+ source:
118
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
119
+ name: Open LLM Leaderboard
120
+ - task:
121
+ type: text-generation
122
+ name: Text Generation
123
+ dataset:
124
+ name: Winogrande (5-shot)
125
+ type: winogrande
126
+ config: winogrande_xl
127
+ split: validation
128
+ args:
129
+ num_few_shot: 5
130
+ metrics:
131
+ - type: acc
132
+ value: 73.4
133
+ name: accuracy
134
+ source:
135
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
136
+ name: Open LLM Leaderboard
137
+ - task:
138
+ type: text-generation
139
+ name: Text Generation
140
+ dataset:
141
+ name: GSM8k (5-shot)
142
+ type: gsm8k
143
+ config: main
144
+ split: test
145
+ args:
146
+ num_few_shot: 5
147
+ metrics:
148
+ - type: acc
149
+ value: 15.31
150
+ name: accuracy
151
+ source:
152
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=ibivibiv/athene-noctua-13b
153
+ name: Open LLM Leaderboard
154
+ ---
155
+ # Athene Noctua 13B
156
+
157
+ ![img](./athene_noctua.png)
158
+
159
+ # Model Details
160
+ * **Trained by**: [ibivibiv](https://huggingface.co/ibivibiv)
161
+ * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
162
+ * **Model type:** **athene-noctua-13b** is an auto-regressive language model fine tuned on the Llama 2 transformer architecture.
163
+ * **Language(s)**: English
164
+ * **Purpose**: Has specific training for logic enforcement, will do well in ARC or other logic testing as well as critical thinking tasks. This model is targeted towards planning exercises.
165
+ * **Comments**: This little guy does pretty well in my logic puzzle testing for a 13B model. I've been using it for test runs to prime for larger models, but it is worth uploading now as it is doing very well on the tests. Again, this a 13B model so tricky logic does still trip it up but for its size it is doing well.
166
+
167
+ # Prompting
168
+
169
+ ## Prompt Template for alpaca style
170
+
171
+ ```
172
+ ### Instruction:
173
+
174
+ <prompt> (without the <>)
175
+
176
+ ### Response:
177
+ ```
178
+
179
+ ## Sample Code
180
+
181
+ ```python
182
+ import torch
183
+ from transformers import AutoModelForCausalLM, AutoTokenizer
184
+
185
+ torch.set_default_device("cuda")
186
+
187
+ model = AutoModelForCausalLM.from_pretrained("ibivibiv/athene-noctua-13b", torch_dtype="auto", device_config='auto')
188
+ tokenizer = AutoTokenizer.from_pretrained("ibivibiv/athene-noctua-13b")
189
+
190
+ inputs = tokenizer("### Instruction: Create a plan for developing the game of snake in python using pygame.\n### Response:\n", return_tensors="pt", return_attention_mask=False)
191
+
192
+ outputs = model.generate(**inputs, max_length=200)
193
+ text = tokenizer.batch_decode(outputs)[0]
194
+ print(text)
195
+ ```
196
+
197
+ ## Citations
198
+
199
+ ```
200
+ @misc{open-llm-leaderboard,
201
+ author = {Edward Beeching and Clémentine Fourrier and Nathan Habib and Sheon Han and Nathan Lambert and Nazneen Rajani and Omar Sanseviero and Lewis Tunstall and Thomas Wolf},
202
+ title = {Open LLM Leaderboard},
203
+ year = {2023},
204
+ publisher = {Hugging Face},
205
+ howpublished = "\url{https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard}"
206
+ }
207
+ ```
208
+ ```
209
+ @software{eval-harness,
210
+ author = {Gao, Leo and
211
+ Tow, Jonathan and
212
+ Biderman, Stella and
213
+ Black, Sid and
214
+ DiPofi, Anthony and
215
+ Foster, Charles and
216
+ Golding, Laurence and
217
+ Hsu, Jeffrey and
218
+ McDonell, Kyle and
219
+ Muennighoff, Niklas and
220
+ Phang, Jason and
221
+ Reynolds, Laria and
222
+ Tang, Eric and
223
+ Thite, Anish and
224
+ Wang, Ben and
225
+ Wang, Kevin and
226
+ Zou, Andy},
227
+ title = {A framework for few-shot language model evaluation},
228
+ month = sep,
229
+ year = 2021,
230
+ publisher = {Zenodo},
231
+ version = {v0.0.1},
232
+ doi = {10.5281/zenodo.5371628},
233
+ url = {https://doi.org/10.5281/zenodo.5371628}
234
+ }
235
+ ```
236
+ ```
237
+ @misc{clark2018think,
238
+ title={Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge},
239
+ author={Peter Clark and Isaac Cowhey and Oren Etzioni and Tushar Khot and Ashish Sabharwal and Carissa Schoenick and Oyvind Tafjord},
240
+ year={2018},
241
+ eprint={1803.05457},
242
+ archivePrefix={arXiv},
243
+ primaryClass={cs.AI}
244
+ }
245
+ ```
246
+ ```
247
+ @misc{zellers2019hellaswag,
248
+ title={HellaSwag: Can a Machine Really Finish Your Sentence?},
249
+ author={Rowan Zellers and Ari Holtzman and Yonatan Bisk and Ali Farhadi and Yejin Choi},
250
+ year={2019},
251
+ eprint={1905.07830},
252
+ archivePrefix={arXiv},
253
+ primaryClass={cs.CL}
254
+ }
255
+ ```
256
+ ```
257
+ @misc{hendrycks2021measuring,
258
+ title={Measuring Massive Multitask Language Understanding},
259
+ author={Dan Hendrycks and Collin Burns and Steven Basart and Andy Zou and Mantas Mazeika and Dawn Song and Jacob Steinhardt},
260
+ year={2021},
261
+ eprint={2009.03300},
262
+ archivePrefix={arXiv},
263
+ primaryClass={cs.CY}
264
+ }
265
+ ```
266
+ ```
267
+ @misc{lin2022truthfulqa,
268
+ title={TruthfulQA: Measuring How Models Mimic Human Falsehoods},
269
+ author={Stephanie Lin and Jacob Hilton and Owain Evans},
270
+ year={2022},
271
+ eprint={2109.07958},
272
+ archivePrefix={arXiv},
273
+ primaryClass={cs.CL}
274
+ }
275
+ ```
276
+ ```
277
+ @misc{DBLP:journals/corr/abs-1907-10641,
278
+ title={{WINOGRANDE:} An Adversarial Winograd Schema Challenge at Scale},
279
+ author={Keisuke Sakaguchi and Ronan Le Bras and Chandra Bhagavatula and Yejin Choi},
280
+ year={2019},
281
+ eprint={1907.10641},
282
+ archivePrefix={arXiv},
283
+ primaryClass={cs.CL}
284
+ }
285
+ ```
286
+ ```
287
+ @misc{DBLP:journals/corr/abs-2110-14168,
288
+ title={Training Verifiers to Solve Math Word Problems},
289
+ author={Karl Cobbe and
290
+ Vineet Kosaraju and
291
+ Mohammad Bavarian and
292
+ Mark Chen and
293
+ Heewoo Jun and
294
+ Lukasz Kaiser and
295
+ Matthias Plappert and
296
+ Jerry Tworek and
297
+ Jacob Hilton and
298
+ Reiichiro Nakano and
299
+ Christopher Hesse and
300
+ John Schulman},
301
+ year={2021},
302
+ eprint={2110.14168},
303
+ archivePrefix={arXiv},
304
+ primaryClass={cs.CL}
305
+ }
306
+ ```
307
+
308
+
309
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
310
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ibivibiv__athene-noctua-13b)
311
+
312
+ | Metric |Value|
313
+ |---------------------------------|----:|
314
+ |Avg. |55.13|
315
+ |AI2 Reasoning Challenge (25-Shot)|57.17|
316
+ |HellaSwag (10-Shot) |81.52|
317
+ |MMLU (5-Shot) |55.91|
318
+ |TruthfulQA (0-shot) |47.49|
319
+ |Winogrande (5-shot) |73.40|
320
+ |GSM8k (5-shot) |15.31|
321
+
322
+
323
+