LLM-Neo
Collection
Model hub for LLM-Neo, including Llama3.1-Neo-1B-100w and Minitron-4B-Depth-Neo-10w.
•
3 items
•
Updated
•
3
We fine-tune nvidia/Llama-3.1-Minitron-4B-Depth-Base with LLM-Neo method,which combines LoRA and KD in one. Training data is sampling from BAAI/Infinity-Instruct for 100k lines.
In this section, we report the results for Llama-3.1-Minitron-4B-Depth-Neo-10w on standard automatic benchmarks. For all the evaluations, we use lm-evaluation-harness library.
Category | Benchmark | Version | n-shot | Metric | Value | Stderr |
BBH | BBH (General) | N/A | 3 | exact_match | 0.4729 | ± 0.0055 |
BBH (Boolean Expressions) | 2 | 3 | exact_match | 0.8120 | ± 0.0248 | |
BBH (Date Understanding) | 2 | 3 | exact_match | 0.6600 | ± 0.0300 | |
CEVAL | CEVAL (General) | N/A | 0 | acc | 0.4413 | ± 0.0135 |
CEVAL (Accountant) | 1 | 0 | acc | 0.3469 | ± 0.0687 | |
CEVAL (Advanced Mathematics) | 1 | 0 | acc | 0.4737 | ± 0.1177 | |
CEVAL (Art Studies) | 1 | 0 | acc | 0.4545 | ± 0.0880 | |
MMLU | MMLU (General) | N/A | 0 | acc | 0.6048 | ± 0.0039 |
MMLU (Humanities) | N/A | 0 | acc | 0.5552 | ± 0.0067 | |
MMLU (STEM) | N/A | 0 | acc | 0.5214 | ± 0.0086 | |
CMMLU | CMMLU (General) | N/A | 0 | acc | 0.3548 | ± 0.0044 |
CMMLU (Normalized) | N/A | 0 | acc_norm | 0.3548 | ± 0.0044 |