Fizzarolli
commited on
Commit
•
05e7a35
1
Parent(s):
38eeb55
Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,21 @@ datasets:
|
|
11 |
# phi 3 4x4b
|
12 |
a continually pretrained phi3-mini sparse moe upcycle
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
## support me on ko-fi!
|
15 |
[~~please i need money to stay alive and keep making models~~](https://ko-fi.com/fizzai)
|
16 |
|
|
|
11 |
# phi 3 4x4b
|
12 |
a continually pretrained phi3-mini sparse moe upcycle
|
13 |
|
14 |
+
## benchmarks
|
15 |
+
### ran locally
|
16 |
+
|
17 |
+
| | Microsoft/phi-3-4k-instruct | Fizzarolli/phi3-4x4b-v1 |
|
18 |
+
| ----------------------- | --------------------------- | ----------------------- |
|
19 |
+
| MMLU acc. (0-shot) | **0.6799** | 0.6781 |
|
20 |
+
| Hellaswag acc. (0-shot) | **0.6053** | 0.5962 |
|
21 |
+
| ARC-E acc. (0-shot) | 0.8325 | **0.8367** |
|
22 |
+
| ARC-C acc. (0-shot) | 0.5546 | **0.5606** |
|
23 |
+
|
24 |
+
honestly i was expecting it to do worse :p, but those are all within a margin of error! so it didn't *lose* any performance, at least
|
25 |
+
|
26 |
+
### open llm leaderboard
|
27 |
+
todo!
|
28 |
+
|
29 |
## support me on ko-fi!
|
30 |
[~~please i need money to stay alive and keep making models~~](https://ko-fi.com/fizzai)
|
31 |
|