Edit model card

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 76.10
AI2 Reasoning Challenge (25-Shot) 72.27
HellaSwag (10-Shot) 89.31
MMLU (5-Shot) 66.55
TruthfulQA (0-shot) 78.55
Winogrande (5-shot) 83.82
GSM8k (5-shot) 66.11

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 21.29
IFEval (0-Shot) 52.12
BBH (3-Shot) 33.99
MATH Lvl 5 (4-Shot) 4.76
GPQA (0-shot) 6.60
MuSR (0-shot) 4.11
MMLU-PRO (5-shot) 26.19
Downloads last month
13,690
Safetensors
Model size
10.7B params
Tensor type
FP16
·
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for vicgalle/CarbonBeagle-11B-truthy

Merges
1 model
Quantizations
2 models

Dataset used to train vicgalle/CarbonBeagle-11B-truthy

Space using vicgalle/CarbonBeagle-11B-truthy 1

Evaluation results