akjindal53244
commited on
Commit
•
aff4c18
1
Parent(s):
f43c596
Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,15 @@ datasets:
|
|
8 |
- akjindal53244/Arithmo-Data
|
9 |
---
|
10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
# Model Card for Model ID
|
13 |
|
@@ -164,6 +173,9 @@ Building LLMs takes time and resources; if you find my work interesting, your su
|
|
164 |
<a href="https://www.buymeacoffee.com/a_little_learner" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
|
165 |
|
166 |
|
|
|
|
|
|
|
167 |
|
168 |
<h2 id="References">References</h2>
|
169 |
|
|
|
8 |
- akjindal53244/Arithmo-Data
|
9 |
---
|
10 |
|
11 |
+
## [January 2024] New Model Release: Arithmo2-Mistral-7B
|
12 |
+
|
13 |
+
**Arithmo2-Mistral-7B** model improves initially released Arithmo-Mistral-7B model on both GSM8K and MATH benchmarks. Specifically, there is **absolute** improvement of:
|
14 |
+
- +1.7% on GSM8K
|
15 |
+
- +3.0% on GSM8K PoT
|
16 |
+
- +1.9% on MATH
|
17 |
+
|
18 |
+
<b>Note</b>: <span style="color:red"><b>It is recommended to use Arithmo2-Mistral-7B model</b></span>. Here is the [merged model](https://huggingface.co/upaya07/Arithmo2-Mistral-7B) and corresponding [LoRA Adapter](https://huggingface.co/upaya07/Arithmo2-Mistral-7B-adapter).
|
19 |
+
|
20 |
|
21 |
# Model Card for Model ID
|
22 |
|
|
|
173 |
<a href="https://www.buymeacoffee.com/a_little_learner" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
|
174 |
|
175 |
|
176 |
+
### Citation
|
177 |
+
Ashvini Jindal, "Arithmo-Mistral-7B", Oct, 2023, https://huggingface.co/akjindal53244/Arithmo-Mistral-7B
|
178 |
+
|
179 |
|
180 |
<h2 id="References">References</h2>
|
181 |
|