mav23 commited on
Commit
c876a59
1 Parent(s): 8cfa8e5

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +52 -0
  3. mathcoder2-llama-3-8b.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ mathcoder2-llama-3-8b.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - MathGenie/MathCode-Pile
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ base_model:
10
+ - meta-llama/Meta-Llama-3-8B
11
+ pipeline_tag: text-generation
12
+ tags:
13
+ - math
14
+ ---
15
+
16
+ # MathCoder2
17
+
18
+ ### Introduction
19
+
20
+ The MathCoder2 models are created by conducting continued pretraining on [MathCode-Pile](https://huggingface.co/datasets/MathGenie/MathCode-Pile). They are introduced in the paper [MathCoder2: Better Math Reasoning from Continued Pretraining on Model-translated Mathematical Code](https://arxiv.org/abs/2410.08196).
21
+
22
+ The mathematical pretraining dataset includes mathematical code accompanied with natural language reasoning steps, making it a superior resource for models aimed at performing advanced mathematical reasoning tasks.
23
+
24
+ ### Evaluation
25
+
26
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65dd9e7b4a4fce1ec96dc6b7/BEZoDZLjp-fPFlt7oFXBa.png)
27
+
28
+ ### Citation
29
+
30
+ If you find this repository helpful, please consider citing our papers:
31
+
32
+ ```
33
+ @misc{lu2024mathcoder2bettermathreasoning,
34
+ title={MathCoder2: Better Math Reasoning from Continued Pretraining on Model-translated Mathematical Code},
35
+ author={Zimu Lu and Aojun Zhou and Ke Wang and Houxing Ren and Weikang Shi and Junting Pan and Mingjie Zhan and Hongsheng Li},
36
+ year={2024},
37
+ eprint={2410.08196},
38
+ archivePrefix={arXiv},
39
+ primaryClass={cs.CL},
40
+ url={https://arxiv.org/abs/2410.08196},
41
+ }
42
+ ```
43
+ ```
44
+ @inproceedings{
45
+ wang2024mathcoder,
46
+ title={MathCoder: Seamless Code Integration in {LLM}s for Enhanced Mathematical Reasoning},
47
+ author={Zimu Lu and Aojun Zhou and Zimu Lu and Sichun Luo and Weikang Shi and Renrui Zhang and Linqi Song and Mingjie Zhan and Hongsheng Li},
48
+ booktitle={The Twelfth International Conference on Learning Representations},
49
+ year={2024},
50
+ url={https://openreview.net/forum?id=z8TW0ttBPp}
51
+ }
52
+ ```
mathcoder2-llama-3-8b.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7924dcd954b3f484c7bc77987dda9449e49367276de5768d6db724f189b096f1
3
+ size 4661212064