Model Details
- Model Description: This model is test for data ordering.
- Developed by: Juhwan Lee
- Model Type: Large Language Model
Model Architecture
This model is based on Mistral-7B-v0.1. We fine-tuning this model for data ordering task.
Mistral-7B-v0.1 is a transformer model, with the following architecture choices:
- Grouped-Query Attention
- Sliding-Window Attention
- Byte-fallback BPE tokenizer
Dataset
We random sample SlimOrca dataset.
Guthub
License
Apache License 2.0
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 61.05 |
AI2 Reasoning Challenge (25-Shot) | 60.41 |
HellaSwag (10-Shot) | 82.76 |
MMLU (5-Shot) | 62.15 |
TruthfulQA (0-shot) | 47.13 |
Winogrande (5-shot) | 78.85 |
GSM8k (5-shot) | 35.03 |
Dataset used to train NLUHOPOE/experiment2-cause-qLoRa
Evaluation results
- normalized accuracy on AI2 Reasoning Challenge (25-Shot)test set Open LLM Leaderboard60.410
- normalized accuracy on HellaSwag (10-Shot)validation set Open LLM Leaderboard82.760
- accuracy on MMLU (5-Shot)test set Open LLM Leaderboard62.150
- mc2 on TruthfulQA (0-shot)validation set Open LLM Leaderboard47.130
- accuracy on Winogrande (5-shot)validation set Open LLM Leaderboard78.850
- accuracy on GSM8k (5-shot)test set Open LLM Leaderboard35.030