Adding model descriptions
Browse files- static/models_data/Mistral-7B-Instruct-v0.1/model_detail.html +6 -0
- static/models_data/Mistral-7B-Instruct-v0.2/model_detail.html +6 -0
- static/models_data/Mistral-7B-Instruct-v0.3/model_detail.html +5 -0
- static/models_data/Mixtral-8x22B-Instruct-v0.1/model_detail.html +6 -0
- static/models_data/Mixtral-8x7B-Instruct-v0.1/model_detail.html +6 -0
- static/models_data/Qwen2-72B-Instruct/model_detail.html +6 -0
- static/models_data/Qwen2-7B-Instruct/model_detail.html +6 -0
- static/models_data/command_r_plus/model_detail.html +5 -0
- static/models_data/dummy/model_detail.html +3 -0
- static/models_data/gpt-3.5-turbo-0125/model_detail.html +4 -0
- static/models_data/gpt-4o-0513/model_detail.html +4 -0
- static/models_data/gpt-4o-mini-2024-07-18/model_detail.html +4 -0
- static/models_data/llama_3.1_405b_instruct_4bit/model_detail.html +8 -0
- static/models_data/llama_3.1_70b_instruct/model_detail.html +6 -0
- static/models_data/llama_3.1_8b_instruct/model_detail.html +6 -0
- static/models_data/llama_3_70b_instruct/model_detail.html +6 -0
- static/models_data/llama_3_8b_instruct/model_detail.html +6 -0
- static/models_data/phi-3-medium-128k-instruct/model_detail.html +6 -0
- static/models_data/phi-3-mini-128k-instruct/model_detail.html +6 -0
static/models_data/Mistral-7B-Instruct-v0.1/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
You can find the release blog post <a href="https://mistral.ai/news/announcing-mistral-7b/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1">https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1</a>.
|
5 |
+
The model has 7.3B parameters, and supports up to 8K token contexts.
|
6 |
+
</p>
|
static/models_data/Mistral-7B-Instruct-v0.2/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
You can find the release blog post <a href="https://mistral.ai/news/la-plateforme/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2">https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2</a>.
|
5 |
+
The model has 7.3B parameters, and supports up to 8K token contexts.
|
6 |
+
</p>
|
static/models_data/Mistral-7B-Instruct-v0.3/model_detail.html
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3">https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3</a>.
|
4 |
+
The model has 7.3B parameters, and supports up to 8K token contexts.
|
5 |
+
</p>
|
static/models_data/Mixtral-8x22B-Instruct-v0.1/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
You can find the release blog post <a href="https://mistral.ai/news/mixtral-8x22b/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1">https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1</a>.
|
5 |
+
The model has 141B total and 39B active parameters. It supports up to 64K token contexts.
|
6 |
+
</p>
|
static/models_data/Mixtral-8x7B-Instruct-v0.1/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
|
3 |
+
You can find the release blog post <a href="https://mistral.ai/news/mixtral-of-experts/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1">https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1</a>.
|
5 |
+
The model has 46.7B total and 12.9B active parameters. It supports up to 32K token contexts.
|
6 |
+
</p>
|
static/models_data/Qwen2-72B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2-72B-Instruct">https://huggingface.co/Qwen/Qwen2-72B-Instruct</a>.
|
5 |
+
The 72B model was pretrained on 29 different languages, and supports up to 128K tokens.
|
6 |
+
</p>
|
static/models_data/Qwen2-7B-Instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://qwenlm.github.io/">The Qwen Team of Alibaba cloud <a>.
|
3 |
+
You can find the release blog post <a href="https://qwenlm.github.io/blog/qwen2/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/Qwen/Qwen2-7B-Instruct">https://huggingface.co/Qwen/Qwen2-7B-Instruct</a>.
|
5 |
+
The 7B model was pretrained on 29 different languages, and supports up to 128K tokens.
|
6 |
+
</p>
|
static/models_data/command_r_plus/model_detail.html
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://cohere.com/">Cohere<AI<a>.
|
3 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/CohereForAI/c4ai-command-r-plus">https://huggingface.co/CohereForAI/c4ai-command-r-plus</a>.
|
4 |
+
The model has 104B parameters, and supports up to 128K token contexts.
|
5 |
+
</p>
|
static/models_data/dummy/model_detail.html
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This is a dummy model. It returns a random answer (of the suggested ones) to the multiple choice question.
|
3 |
+
</p>
|
static/models_data/gpt-3.5-turbo-0125/model_detail.html
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This proprietary model was created by <a href="https://openai.com/">OpenAI<a>.
|
3 |
+
You can find the release blog post <a href="https://openai.com/index/chatgpt/">here</a>.
|
4 |
+
</p>
|
static/models_data/gpt-4o-0513/model_detail.html
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This proprietary model was created by <a href="https://openai.com/">OpenAI<a>.
|
3 |
+
You can find the release blog post <a href="https://openai.com/index/hello-gpt-4o/">here</a>.
|
4 |
+
</p>
|
static/models_data/gpt-4o-mini-2024-07-18/model_detail.html
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This proprietary model was created by <a href="https://openai.com/">OpenAI<a>.
|
3 |
+
You can find the release blog post <a href="https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/">here</a>.
|
4 |
+
</p>
|
static/models_data/llama_3.1_405b_instruct_4bit/model_detail.html
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://ai.meta.com/">Meta AI</a>.
|
3 |
+
You can find the release blog post <a href="https://ai.meta.com/blog/meta-llama-3-1/">here</a>.
|
4 |
+
The 16bit precision model is available on the huggingface hub: <a href="https://huggingface.co/meta-llama/Meta-Llama-3.1-405B-Instruct">https://huggingface.co/meta-llama/Meta-Llama-3.1-405B-Instruct</a>.
|
5 |
+
Due to computational constrains we use the 4bit quantized version, which is also available on the huggingfacehub: <a href="unsloth/Meta-Llama-3.1-405B-Instruct-bnb-4bit">unsloth/Meta-Llama-3.1-405B-Instruct-bnb-4bit</a>.
|
6 |
+
It is relevant to note that we compared with a 16bit version hosted by <a href="https://www.together.ai/">TogetherAI</a> on a subset of problems that fall in the 4k tokens limit defined by the TogetherAI API, and we did not see drastic changes in performance.
|
7 |
+
The 405B model was pretrained on 15 trillion tokens spanning 8 different languages, and supports up to 128K token contexts.
|
8 |
+
</p>
|
static/models_data/llama_3.1_70b_instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://ai.meta.com/">Meta AI</a>.
|
3 |
+
You can find the release blog post <a href="https://ai.meta.com/blog/meta-llama-3-1/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct">https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct</a>.
|
5 |
+
The 70B model was pretrained on 15 trillion tokens spanning 8 different languages, and supports up to 128K token contexts.
|
6 |
+
</p>
|
static/models_data/llama_3.1_8b_instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://ai.meta.com/">Meta AI</a>.
|
3 |
+
You can find the release blog post <a href="https://ai.meta.com/blog/meta-llama-3-1/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct">https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct</a>.
|
5 |
+
The 70B model was pretrained on 15 trillion tokens spanning 8 different languages, and supports up to 128K token contexts.
|
6 |
+
</p>
|
static/models_data/llama_3_70b_instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://ai.meta.com/">Meta AI</a>.
|
3 |
+
You can find the release blog post <a href="https://ai.meta.com/blog/meta-llama-3/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct">https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct</a>.
|
5 |
+
The 70B model was pretrained on 15 trillion tokens spanning 30 different languages in sequences of 8,192 tokens.
|
6 |
+
</p>
|
static/models_data/llama_3_8b_instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://ai.meta.com/">Meta AI</a>.
|
3 |
+
You can find the release blog post <a href="https://ai.meta.com/blog/meta-llama-3/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct">https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct</a>.
|
5 |
+
The 8B model was pretrained on 15 trillion tokens spanning 30 different languages in sequences of 8,192 tokens.
|
6 |
+
</p>
|
static/models_data/phi-3-medium-128k-instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://www.microsoft.com/">Microsoft<a>.
|
3 |
+
You can find the release blog post <a href="https://azure.microsoft.com/en-us/blog/introducing-phi-3-redefining-whats-possible-with-slms/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/microsoft/Phi-3-medium-128k-instruct">https://huggingface.co/microsoft/Phi-3-medium-128k-instruct</a>.
|
5 |
+
The model has 14B parameters, and supports up to 128K token contexts.
|
6 |
+
</p>
|
static/models_data/phi-3-mini-128k-instruct/model_detail.html
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<p>
|
2 |
+
This open-source model was created by <a href="https://www.microsoft.com/">Microsoft<a>.
|
3 |
+
You can find the release blog post <a href="https://azure.microsoft.com/en-us/blog/introducing-phi-3-redefining-whats-possible-with-slms/">here</a>.
|
4 |
+
The model is available on the huggingface hub: <a href="https://huggingface.co/microsoft/Phi-3-mini-128k-instruct">https://huggingface.co/microsoft/Phi-3-mini-128k-instruct</a>.
|
5 |
+
The model has 3.8B parameters, and supports up to 128K token contexts.
|
6 |
+
</p>
|