Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)
MODEL HellaSwag EQ_Bench % Parsed (EQ)
argilla/CapybaraHermes-2.5-Mistral-7B 83.03 66.52 100
athirdpath/NSFW_DPO_vmgb-7b 85.36 74.83 100
berkeley-nest/Starling-LM-7B-alpha 82.17 69.78 100
cgato/Thespis-Mistral-7b-v0.6 81.83 29.13 79.53
chargoddard/loyal-piano-m7 83.28 72.15 100
cognitivecomputations/dolphin-2.2.1-mistral-7b 81.47 67.33 100
cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser 83.61 71.57 99.42
Crimvael/Raphael-7B 83.47 68.73 100
Delcos/Mistral-Pygmalion-7b 76.61 17.34 100
FallenMerick/Iced-Lemon-Cookie-7B 85.54 71.54 100
FallenMerick/Smart-Lemon-Cookie-7B 85.41 68.12 100
HuggingFaceH4/zephyr-7b-beta 82.02 58.14 100
Intel/neural-chat-7b-v3-1 79.76 62.26 100
jondurbin/airoboros-m-7b-3.1.2 81.34 38.52 100
jondurbin/cinematika-7b-v0.1 80.31 44.85 100
KatyTheCutie/LemonadeRP-4.5.3 82.65 63.23 100
KoboldAI/Mistral-7B-Erebus-v3 76.65 18.19 97.66
KoboldAI/Mistral-7B-Holodeck-1 79.19 2.10 98.25
maywell/Synatra-7B-v0.3-RP 80.46 54.93 100
migtissera/Synthia-7B-v3.0 81.74 15.03 94.74
mistralai/Mistral-7B-Instruct-v0.1 74.67 46.82 100
mistralai/Mistral-7B-Instruct-v0.2 83.66 65.50 99.42
mlabonne/NeuralBeagle14-7B 86.46 74.21 99.42
Norquinal/Mistral-7B-claude-chat 83.06 16.34 99.42
NousResearch/Hermes-2-Pro-Mistral-7B 80.56 65.93 100
NousResearch/Nous-Capybara-7B-V1.9 78.71 19.25 100
Open-Orca/Mistral-7B-OpenOrca 81.67 63.98 99.42
rwitz/go-bruins 84.92 73.62 100
SanjiWatsuki/Kunoichi-7B 85.25 72.36 100
SanjiWatsuki/Loyal-Macaroni-Maid-7B 84.53 73.67 100
SanjiWatsuki/Silicon-Maid-7B 84.55 68.74 100
senseable/WestLake-7B-v2 87.42 77.87 100
TeeZee/DarkSapling-7B-v2.0 82.56 62.19 100
teknium/Hermes-Trismegistus-Mistral-7B 79.21 64.86 98.83
teknium/OpenHermes-2.5-Mistral-7B 81.68 65.75 100
Undi95/Toppy-M-7B 83.52 66.57 100

MODEL HellaSwag EQ_Bench % Parsed (EQ)
ABX-AI/Silver-Sun-v2-11B 86.40 69.92 100
backyardai/Fimbulvetr-Holodeck-Erebus-Westlake-10.7B 86.00 69.25 100
BlueNipples/SnowLotus-v2-10.7B 83.42 60.54 99.42
FallenMerick/Chewy-Lemon-Cookie-11B 84.39 76.24 100
FallenMerick/Chunky-Lemon-Cookie-11B 84.36 76.29 100
froggeric/WestLake-10.7B-v2 86.74 73.35 95.32
head-empty-ai/Mytho-Lemon-11B 79.77 60.03 100
Himitsui/KuroMitsu-11B 86.33 70.50 98.83
kyujinpy/SOLAR-Platypus-10.7B-v2 82.05 25.11 45.61
migtissera/Tess-10.7B-v1.5b 83.82 61.83 99.42
NeverSleep/Mistral-11B-SynthIAirOmniMix 81.58 55.19 100
NousResearch/Nous-Hermes-2-SOLAR-10.7B 83.24 63.52 100
saishf/Fimbulvetr-Kuro-Lotus-10.7B 86.25 65.85 100
Sao10K/Fimbulvetr-10.7B-v1 85.81 65.42 100
Sao10K/Fimbulvetr-11B-v2 86.61 70.00 99.42
Sao10K/Frostwind-10.7B-v1 84.15 55.73 99.42
Sao10K/Solstice-11B-v1 86.42 68.24 99.42
TheDrummer/Moistral-11B-v3 86.65 69.75 99.42
Undi95/Borealis-10.7B 79.58 8.27 44.44
upstage/SOLAR-10.7B-Instruct-v1.0 86.35 68.65 98.25
upstage/SOLAR-10.7B-v1.0 83.10 28.66 100
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .