File size: 483 Bytes
99de6d0 |
1 2 3 4 5 6 7 |
<p>
This open-source model was created by <a href="https://mistral.ai/">Mistral AI<a>.
You can find the release blog post <a href="https://mistral.ai/news/mixtral-of-experts/">here</a>.
The model is available on the huggingface hub: <a href="https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1">https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1</a>.
The model has 46.7B total and 12.9B active parameters. It supports up to 32K token contexts.
</p>
|