3rd-Degree-Burn's picture
Update README.md
fabb7b4 verified
|
raw
history blame
1.39 kB
---
license: apache-2.0
tags:
- merge
- mergekit
- lazymergekit
- NousResearch/Meta-Llama-3.1-8B-Instruct
- EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta
- nvidia/OpenMath2-Llama3.1-8B
base_model:
- NousResearch/Meta-Llama-3.1-8B-Instruct
- EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta
- nvidia/OpenMath2-Llama3.1-8B
---
# Llama-3.1-8B-Squareroot
This is a TIES merge that combines the performance of the following models:
* [NousResearch/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct)
* [EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta](https://huggingface.co/EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta)
* [nvidia/OpenMath2-Llama3.1-8B](https://huggingface.co/nvidia/OpenMath2-Llama3.1-8B)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6479f6dbed75e95d3e97bb4d/LpWI-ug9WZdpcrjBy44iw.png)
# Description
This one's a failed attempt. Working on a better version, so check back soon!
# Benchmarks
This model ranks in the top 5 for MATH benchmarks, but is severely bad on the rest (which isn't quite what I was expecting). I’m hoping to improve its general abilities without losing its math skills. Qwen still has the top spot :(
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6479f6dbed75e95d3e97bb4d/IPC7gTS4wJPOXVm1nCqLV.png)