Adding Evaluation Results

#2
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -29,4 +29,17 @@ The script I used to perform the merge is available [here](https://github.com/cg
29
  The command that produced this model:
30
  ```
31
  python ties_merge.py TheBloke/Llama-2-13B-fp16 ./Chronorctypus-Limarobormes-13b --merge elinas/chronos-13b-v2 --merge Open-Orca/OpenOrca-Platypus2-13B --merge Oniichat/limarp-13b-merged --merge jondurbin/airoboros-l2-13b-gpt4-1.4.1 --merge NousResearch/Nous-Hermes-Llama2-13b --cuda
32
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  The command that produced this model:
30
  ```
31
  python ties_merge.py TheBloke/Llama-2-13B-fp16 ./Chronorctypus-Limarobormes-13b --merge elinas/chronos-13b-v2 --merge Open-Orca/OpenOrca-Platypus2-13B --merge Oniichat/limarp-13b-merged --merge jondurbin/airoboros-l2-13b-gpt4-1.4.1 --merge NousResearch/Nous-Hermes-Llama2-13b --cuda
32
+ ```
33
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
34
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_chargoddard__Chronorctypus-Limarobormes-13b)
35
+
36
+ | Metric | Value |
37
+ |-----------------------|---------------------------|
38
+ | Avg. | 49.88 |
39
+ | ARC (25-shot) | 59.9 |
40
+ | HellaSwag (10-shot) | 82.75 |
41
+ | MMLU (5-shot) | 58.45 |
42
+ | TruthfulQA (0-shot) | 51.9 |
43
+ | Winogrande (5-shot) | 74.43 |
44
+ | GSM8K (5-shot) | 3.87 |
45
+ | DROP (3-shot) | 17.89 |