Adding Mistral-Small-2409 and Qwen2.5
Browse files- static/leaderboard.csv +25 -20
- static/models_data/Mistral-Large-Instruct-2407/model_detail.html +6 -1
- static/models_data/Mistral-Small-Instruct-2409/cfa_metrics.csv +10 -0
- static/models_data/Mistral-Small-Instruct-2409/matrix.svg +1965 -0
- static/models_data/Mistral-Small-Instruct-2409/model_detail.html +5 -0
- static/models_data/Mistral-Small-Instruct-2409/ranks.svg +0 -0
- static/models_data/Mistral-Small-Instruct-2409/structure.svg +0 -0
- static/models_data/Qwen2.5-0.5B-Instruct/cfa_metrics.csv +10 -0
- static/models_data/Qwen2.5-0.5B-Instruct/matrix.svg +2004 -0
- static/models_data/Qwen2.5-0.5B-Instruct/model_detail.html +7 -0
- static/models_data/Qwen2.5-0.5B-Instruct/ranks.svg +0 -0
- static/models_data/Qwen2.5-0.5B-Instruct/structure.svg +0 -0
- static/models_data/Qwen2.5-32B-Instruct/cfa_metrics.csv +10 -0
- static/models_data/Qwen2.5-32B-Instruct/matrix.svg +1963 -0
- static/models_data/Qwen2.5-32B-Instruct/model_detail.html +7 -0
- static/models_data/Qwen2.5-32B-Instruct/ranks.svg +0 -0
- static/models_data/Qwen2.5-32B-Instruct/structure.svg +0 -0
- static/models_data/Qwen2.5-72B-Instruct/cfa_metrics.csv +10 -0
- static/models_data/Qwen2.5-72B-Instruct/matrix.svg +1967 -0
- static/models_data/Qwen2.5-72B-Instruct/model_detail.html +7 -0
- static/models_data/Qwen2.5-72B-Instruct/ranks.svg +0 -0
- static/models_data/Qwen2.5-72B-Instruct/structure.svg +0 -0
- static/models_data/Qwen2.5-7B-Instruct/cfa_metrics.csv +10 -0
- static/models_data/Qwen2.5-7B-Instruct/matrix.svg +1964 -0
- static/models_data/Qwen2.5-7B-Instruct/model_detail.html +7 -0
- static/models_data/Qwen2.5-7B-Instruct/ranks.svg +0 -0
- static/models_data/Qwen2.5-7B-Instruct/structure.svg +0 -0
- static/models_data/cardinal.svg +590 -352
- static/models_data/ordinal.svg +685 -415
static/leaderboard.csv
CHANGED
@@ -1,21 +1,26 @@
|
|
1 |
Model,Ordinal (Win rate),Cardinal (Score),RO Stability,Stress,CFI,SRMR,RMSEA
|
2 |
-
phi-3-mini-128k-instruct,0.
|
3 |
-
phi-3-medium-128k-instruct,0.
|
4 |
-
Mistral-7B-Instruct-v0.1,0.
|
5 |
-
Mistral-7B-Instruct-v0.2,0.
|
6 |
-
Mistral-7B-Instruct-v0.3,0.
|
7 |
-
Mixtral-8x7B-Instruct-v0.1,0.
|
8 |
-
Mixtral-8x22B-Instruct-v0.1,0.
|
9 |
-
command_r_plus,0.
|
10 |
-
llama_3_8b_instruct,0.
|
11 |
-
llama_3_70b_instruct,0.
|
12 |
-
llama_3.1_8b_instruct,0.
|
13 |
-
llama_3.1_70b_instruct,0.
|
14 |
-
llama_3.1_405b_instruct_4bit,0.
|
15 |
-
Qwen2-7B-Instruct,0.
|
16 |
-
Qwen2-72B-Instruct,0.
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
1 |
Model,Ordinal (Win rate),Cardinal (Score),RO Stability,Stress,CFI,SRMR,RMSEA
|
2 |
+
phi-3-mini-128k-instruct,0.34490740740740744,0.39387631706052895,0.039299993295009855,0.281800547806919,0.7509527777777777,0.25489166666666674,0.22045000000000003
|
3 |
+
phi-3-medium-128k-instruct,0.34317129629629634,0.4026069526718651,0.09692037989916814,0.2651981204439735,0.6727694444444445,0.2984500000000001,0.2759472222222221
|
4 |
+
Mistral-7B-Instruct-v0.1,0.1996527777777778,0.30674462188144647,0.027216280472015988,0.2829498135031582,0.500288888888889,0.45314444444444446,0.4191027777777777
|
5 |
+
Mistral-7B-Instruct-v0.2,0.3755787037037038,0.4028886762146369,0.14417876497818388,0.265188983528973,0.5787944444444445,0.35010277777777776,0.3171083333333333
|
6 |
+
Mistral-7B-Instruct-v0.3,0.2708333333333333,0.34429493368035685,0.07960539866974455,0.2742399030139009,0.5231444444444444,0.4214972222222223,0.3914694444444443
|
7 |
+
Mixtral-8x7B-Instruct-v0.1,0.4496527777777778,0.47204265176392696,0.21473356319081474,0.2624402608740656,0.6766166666666665,0.25611666666666666,0.24065277777777772
|
8 |
+
Mixtral-8x22B-Instruct-v0.1,0.26620370370370366,0.3496962191659786,0.1414001940345544,0.2548838005881672,0.45902777777777776,0.4849916666666666,0.4871833333333333
|
9 |
+
command_r_plus,0.5815972222222222,0.5698450422762357,0.3429686514651868,0.23811982320641845,0.7772111111111112,0.17755277777777778,0.17465277777777777
|
10 |
+
llama_3_8b_instruct,0.48900462962962954,0.5066363890459272,0.24527785038654715,0.245806400289881,0.7348277777777779,0.20952222222222228,0.20751944444444437
|
11 |
+
llama_3_70b_instruct,0.7291666666666666,0.7270613281502669,0.607020698814379,0.18525883672204868,0.8298166666666668,0.10965277777777771,0.14649722222222217
|
12 |
+
llama_3.1_8b_instruct,0.5434027777777778,0.5599895255443657,0.4295080949846363,0.22060228669473025,0.6379333333333334,0.3225500000000001,0.3328972222222223
|
13 |
+
llama_3.1_70b_instruct,0.7847222222222222,0.7630277652278956,0.691365862744007,0.1709718847084183,0.8203805555555554,0.14023055555555552,0.17041944444444446
|
14 |
+
llama_3.1_405b_instruct_4bit,0.6886574074074073,0.6993503239272297,0.7232098126552619,0.1702199925365422,0.6062611111111111,0.3538527777777777,0.38022500000000004
|
15 |
+
Qwen2-7B-Instruct,0.43287037037037035,0.46812644016430927,0.25108519506513916,0.25776537005719313,0.6248583333333334,0.32358611111111113,0.3028361111111111
|
16 |
+
Qwen2-72B-Instruct,0.5810185185185186,0.64867678910782,0.6465993243020925,0.20297742879025626,0.5559722222222221,0.3575638888888889,0.39241388888888884
|
17 |
+
Qwen2.5-0.5B-Instruct,0.28877314814814814,0.3796838812739187,0.002970456550606876,0.2928913315666324,0.7497416666666666,0.24648888888888887,0.18477222222222223
|
18 |
+
Qwen2.5-7B-Instruct,0.6186342592592592,0.5896473181421169,0.333554494486959,0.2505866550331236,0.8311222222222222,0.10302222222222213,0.09455277777777782
|
19 |
+
Qwen2.5-32B-Instruct,0.7442129629629629,0.731635015756055,0.6724190751477237,0.1806656189868978,0.7584111111111111,0.19748055555555544,0.21686111111111106
|
20 |
+
Qwen2.5-72B-Instruct,0.7991898148148148,0.754401345305127,0.6974116787371809,0.16176650806326276,0.7859583333333332,0.177875,0.2007527777777779
|
21 |
+
gpt-3.5-turbo-0125,0.21643518518518517,0.328243163867074,0.08240359836763214,0.28728574920060357,0.4998916666666666,0.47583055555555553,0.4404444444444445
|
22 |
+
gpt-4o-0513,0.7025462962962963,0.6713251724661671,0.5122163952167618,0.19201420113771173,0.7998694444444445,0.14606111111111109,0.1400583333333334
|
23 |
+
gpt-4o-mini-2024-07-18,0.3628472222222222,0.40825697940501954,0.13575309046266867,0.2707065266105181,0.6141777777777777,0.32648055555555555,0.29394722222222214
|
24 |
+
Mistral-Large-Instruct-2407,0.8217592592592592,0.7808285247091349,0.7644582301049158,0.16944638941325085,0.7604888888888888,0.18767499999999993,0.21457222222222228
|
25 |
+
Mistral-Small-Instruct-2409,0.7083333333333335,0.7319149695591499,0.6416815833333804,0.1894343546381,0.7891722222222222,0.1387222222222222,0.17242222222222225
|
26 |
+
dummy,0.14872685185185186,0.2784036220050126,-0.009004148398032956,0.2928877637010999,0.5076361111111111,0.4973388888888889,0.4541638888888889
|
static/models_data/Mistral-Large-Instruct-2407/model_detail.html
CHANGED
@@ -1 +1,6 @@
|
|
1 |
-
<p>
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
You can find the release blog post <a href="https://mistral.ai/news/mistral-large-2407/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mistral-Large-Instruct-2407">https://huggingface.co/mistralai/Mistral-Large-Instruct-2407</a>.
|
5 |
+
The 123B model supports up to 128K token context windows.
|
6 |
+
</p>
|
static/models_data/Mistral-Small-Instruct-2409/cfa_metrics.csv
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Context chunk,CFI,TLI,SRMR,RMSEA
|
2 |
+
chunk_0,0.789825,0.71305,0.12095,0.167425
|
3 |
+
chunk_1,0.862975,0.8278750000000001,0.086675,0.110275
|
4 |
+
chunk_2,0.851325,0.796775,0.108275,0.132775
|
5 |
+
chunk_3,0.6166499999999999,0.572825,0.33632500000000004,0.355075
|
6 |
+
chunk_4,0.830525,0.769525,0.11827499999999999,0.15587499999999999
|
7 |
+
chunk_chess_0,0.823325,0.765625,0.10767499999999999,0.147825
|
8 |
+
chunk_grammar_1,0.77625,0.696125,0.12664999999999998,0.1973
|
9 |
+
chunk_no_conv,0.8255,0.7659750000000001,0.12215000000000001,0.163975
|
10 |
+
chunk_svs_no_conv,0.726175,0.6696500000000001,0.121525,0.121275
|
static/models_data/Mistral-Small-Instruct-2409/matrix.svg
ADDED
static/models_data/Mistral-Small-Instruct-2409/model_detail.html
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mistral-Small-Instruct-2409">https://huggingface.co/mistralai/Mistral-Small-Instruct-2409</a>.
|
4 |
+
The 22B model supports up to 32K token sequences.
|
5 |
+
</p>
|
static/models_data/Mistral-Small-Instruct-2409/ranks.svg
ADDED
static/models_data/Mistral-Small-Instruct-2409/structure.svg
ADDED
static/models_data/Qwen2.5-0.5B-Instruct/cfa_metrics.csv
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Context chunk,CFI,TLI,SRMR,RMSEA
|
2 |
+
chunk_0,0.7500249999999999,-2.21615,0.10572500000000001,0.0545
|
3 |
+
chunk_1,0.75,1.25775,0.3161,0.25
|
4 |
+
chunk_2,0.73175,0.9472249999999999,0.31827500000000003,0.2564
|
5 |
+
chunk_3,0.75,1.0417,0.32295,0.25
|
6 |
+
chunk_4,0.8639749999999999,0.9359999999999999,0.10375,0.0408
|
7 |
+
chunk_chess_0,0.904775,0.28604999999999997,0.09230000000000001,0.020275
|
8 |
+
chunk_grammar_1,0.75,5.354475,0.30865,0.25
|
9 |
+
chunk_no_conv,0.600475,0.04259999999999997,0.322475,0.27175000000000005
|
10 |
+
chunk_svs_no_conv,0.646675,0.6581250000000001,0.328175,0.269225
|
static/models_data/Qwen2.5-0.5B-Instruct/matrix.svg
ADDED
static/models_data/Qwen2.5-0.5B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2.5/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct">https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct</a>.
|
5 |
+
The 0.5B model was pretrained on 18 trillion tokens spanning 29 languages.
|
6 |
+
It supports up to 128K tokens and can generate up to 8K tokens.
|
7 |
+
</p>
|
static/models_data/Qwen2.5-0.5B-Instruct/ranks.svg
ADDED
static/models_data/Qwen2.5-0.5B-Instruct/structure.svg
ADDED
static/models_data/Qwen2.5-32B-Instruct/cfa_metrics.csv
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Context chunk,CFI,TLI,SRMR,RMSEA
|
2 |
+
chunk_0,0.407525,0.379375,0.54825,0.5520999999999999
|
3 |
+
chunk_1,0.662925,0.6309750000000001,0.324725,0.33095
|
4 |
+
chunk_2,0.8544,0.804925,0.09639999999999999,0.1127
|
5 |
+
chunk_3,0.910725,0.8782500000000001,0.08855,0.0821
|
6 |
+
chunk_4,0.63385,0.5929,0.320975,0.34572499999999995
|
7 |
+
chunk_chess_0,0.8669749999999999,0.8261499999999999,0.09075,0.1154
|
8 |
+
chunk_grammar_1,0.8252,0.7672,0.090675,0.1505
|
9 |
+
chunk_no_conv,0.8589749999999999,0.81545,0.110475,0.13344999999999999
|
10 |
+
chunk_svs_no_conv,0.805125,0.782975,0.106525,0.128825
|
static/models_data/Qwen2.5-32B-Instruct/matrix.svg
ADDED
static/models_data/Qwen2.5-32B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2.5/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2.5-32B-Instruct">https://huggingface.co/Qwen/Qwen2.5-32B-Instruct</a>.
|
5 |
+
The 32B model was pretrained on 18 trillion tokens spanning 29 languages.
|
6 |
+
It supports up to 128K tokens and can generate up to 8K tokens.
|
7 |
+
</p>
|
static/models_data/Qwen2.5-32B-Instruct/ranks.svg
ADDED
static/models_data/Qwen2.5-32B-Instruct/structure.svg
ADDED
static/models_data/Qwen2.5-72B-Instruct/cfa_metrics.csv
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Context chunk,CFI,TLI,SRMR,RMSEA
|
2 |
+
chunk_0,0.848975,0.804025,0.102925,0.09962499999999999
|
3 |
+
chunk_1,0.87305,0.83225,0.10135,0.122
|
4 |
+
chunk_2,0.82955,0.7724,0.106,0.12817499999999998
|
5 |
+
chunk_3,0.86325,0.8176,0.11270000000000001,0.13125
|
6 |
+
chunk_4,0.615525,0.568375,0.32222500000000004,0.388625
|
7 |
+
chunk_chess_0,0.6388,0.59945,0.32495,0.3749
|
8 |
+
chunk_grammar_1,0.66355,0.6322749999999999,0.33272500000000005,0.33187500000000003
|
9 |
+
chunk_no_conv,0.889775,0.8521249999999999,0.09684999999999999,0.14735
|
10 |
+
chunk_svs_no_conv,0.85115,0.8130499999999999,0.10115,0.082975
|
static/models_data/Qwen2.5-72B-Instruct/matrix.svg
ADDED
static/models_data/Qwen2.5-72B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2.5/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2.5-72B-Instruct">https://huggingface.co/Qwen/Qwen2.5-72B-Instruct</a>.
|
5 |
+
The 72B model was pretrained on 18 trillion tokens spanning 29 languages.
|
6 |
+
It supports up to 128K tokens and can generate up to 8K tokens.
|
7 |
+
</p>
|
static/models_data/Qwen2.5-72B-Instruct/ranks.svg
ADDED
static/models_data/Qwen2.5-72B-Instruct/structure.svg
ADDED
static/models_data/Qwen2.5-7B-Instruct/cfa_metrics.csv
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Context chunk,CFI,TLI,SRMR,RMSEA
|
2 |
+
chunk_0,0.7448499999999999,0.6389,0.1069,0.11257499999999998
|
3 |
+
chunk_1,0.83535,0.8164499999999999,0.10495,0.064675
|
4 |
+
chunk_2,0.8623,-3.591025,0.09625,0.071425
|
5 |
+
chunk_3,0.88345,0.886125,0.097575,0.05395
|
6 |
+
chunk_4,0.9301250000000001,0.9451499999999999,0.089075,0.049049999999999996
|
7 |
+
chunk_chess_0,0.8341,0.764,0.112125,0.122975
|
8 |
+
chunk_grammar_1,0.865475,0.877975,0.0904,0.0895
|
9 |
+
chunk_no_conv,0.780725,0.695675,0.10905000000000001,0.15252499999999997
|
10 |
+
chunk_svs_no_conv,0.743725,0.6745749999999999,0.120875,0.1343
|
static/models_data/Qwen2.5-7B-Instruct/matrix.svg
ADDED
static/models_data/Qwen2.5-7B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2.5/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2.5-7B-Instruct">https://huggingface.co/Qwen/Qwen2.5-7B-Instruct</a>.
|
5 |
+
The 7B model was pretrained on 18 trillion tokens spanning 29 languages.
|
6 |
+
It supports up to 128K tokens and can generate up to 8K tokens.
|
7 |
+
</p>
|
static/models_data/Qwen2.5-7B-Instruct/ranks.svg
ADDED
static/models_data/Qwen2.5-7B-Instruct/structure.svg
ADDED
static/models_data/cardinal.svg
CHANGED
static/models_data/ordinal.svg
CHANGED