Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
backyardlab
/
Benchmark-Results
like
0
Follow
Backyard AI Labs
11
Model card
Files
Files and versions
Community
Edit model card
YAML Metadata Warning:
empty or missing yaml metadata in repo card (
https://huggingface.co/docs/hub/model-cards#model-card-metadata
)
MODEL
HellaSwag
EQ_Bench
% Parsed (EQ)
argilla/CapybaraHermes-2.5-Mistral-7B
83.03
66.52
100
athirdpath/NSFW_DPO_vmgb-7b
85.36
74.83
100
berkeley-nest/Starling-LM-7B-alpha
82.17
69.78
100
cgato/Thespis-Mistral-7b-v0.6
81.83
29.13
79.53
chargoddard/loyal-piano-m7
83.28
72.15
100
cognitivecomputations/dolphin-2.2.1-mistral-7b
81.47
67.33
100
cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser
83.61
71.57
99.42
Crimvael/Raphael-7B
83.47
68.73
100
Delcos/Mistral-Pygmalion-7b
76.61
17.34
100
FallenMerick/Iced-Lemon-Cookie-7B
85.54
71.54
100
FallenMerick/Smart-Lemon-Cookie-7B
85.41
68.12
100
HuggingFaceH4/zephyr-7b-beta
82.02
58.14
100
Intel/neural-chat-7b-v3-1
79.76
62.26
100
jondurbin/airoboros-m-7b-3.1.2
81.34
38.52
100
jondurbin/cinematika-7b-v0.1
80.31
44.85
100
KatyTheCutie/LemonadeRP-4.5.3
82.65
63.23
100
KoboldAI/Mistral-7B-Erebus-v3
76.65
18.19
97.66
KoboldAI/Mistral-7B-Holodeck-1
79.19
2.10
98.25
maywell/Synatra-7B-v0.3-RP
80.46
54.93
100
migtissera/Synthia-7B-v3.0
81.74
15.03
94.74
mistralai/Mistral-7B-Instruct-v0.1
74.67
46.82
100
mistralai/Mistral-7B-Instruct-v0.2
83.66
65.50
99.42
mlabonne/NeuralBeagle14-7B
86.46
74.21
99.42
Norquinal/Mistral-7B-claude-chat
83.06
16.34
99.42
NousResearch/Hermes-2-Pro-Mistral-7B
80.56
65.93
100
NousResearch/Nous-Capybara-7B-V1.9
78.71
19.25
100
Open-Orca/Mistral-7B-OpenOrca
81.67
63.98
99.42
rwitz/go-bruins
84.92
73.62
100
SanjiWatsuki/Kunoichi-7B
85.25
72.36
100
SanjiWatsuki/Loyal-Macaroni-Maid-7B
84.53
73.67
100
SanjiWatsuki/Silicon-Maid-7B
84.55
68.74
100
senseable/WestLake-7B-v2
87.42
77.87
100
TeeZee/DarkSapling-7B-v2.0
82.56
62.19
100
teknium/Hermes-Trismegistus-Mistral-7B
79.21
64.86
98.83
teknium/OpenHermes-2.5-Mistral-7B
81.68
65.75
100
Undi95/Toppy-M-7B
83.52
66.57
100
MODEL
HellaSwag
EQ_Bench
% Parsed (EQ)
ABX-AI/Silver-Sun-v2-11B
86.40
69.92
100
backyardai/Fimbulvetr-Holodeck-Erebus-Westlake-10.7B
86.00
69.25
100
BlueNipples/SnowLotus-v2-10.7B
83.42
60.54
99.42
FallenMerick/Chewy-Lemon-Cookie-11B
84.39
76.24
100
FallenMerick/Chunky-Lemon-Cookie-11B
84.36
76.29
100
froggeric/WestLake-10.7B-v2
86.74
73.35
95.32
head-empty-ai/Mytho-Lemon-11B
79.77
60.03
100
Himitsui/KuroMitsu-11B
86.33
70.50
98.83
kyujinpy/SOLAR-Platypus-10.7B-v2
82.05
25.11
45.61
migtissera/Tess-10.7B-v1.5b
83.82
61.83
99.42
NeverSleep/Mistral-11B-SynthIAirOmniMix
81.58
55.19
100
NousResearch/Nous-Hermes-2-SOLAR-10.7B
83.24
63.52
100
saishf/Fimbulvetr-Kuro-Lotus-10.7B
86.25
65.85
100
Sao10K/Fimbulvetr-10.7B-v1
85.81
65.42
100
Sao10K/Fimbulvetr-11B-v2
86.61
70.00
99.42
Sao10K/Frostwind-10.7B-v1
84.15
55.73
99.42
Sao10K/Solstice-11B-v1
86.42
68.24
99.42
TheDrummer/Moistral-11B-v3
86.65
69.75
99.42
Undi95/Borealis-10.7B
79.58
8.27
44.44
upstage/SOLAR-10.7B-Instruct-v1.0
86.35
68.65
98.25
upstage/SOLAR-10.7B-v1.0
83.10
28.66
100
Downloads last month
-
Downloads are not tracked for this model.
How to track
Inference API
Unable to determine this model's library. Check the
docs
.