Update README.md
Browse files
README.md
CHANGED
@@ -16,8 +16,8 @@ tags:
|
|
16 |
|
17 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
|
18 |
|
19 |
-
|
20 |
-
|
21 |
|
22 |
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
|
23 |
|
@@ -47,8 +47,8 @@ u/SomeOddCodeGuy wrote on r/LocalLLaMA:
|
|
47 |
|
48 |
## Model Details
|
49 |
|
50 |
-
|
51 |
-
|
52 |
|
53 |
### Prompt template: Mistral
|
54 |
|
@@ -67,7 +67,8 @@ This model was merged using the passthrough merge method.
|
|
67 |
### Models Merged
|
68 |
|
69 |
The following models were included in the merge:
|
70 |
-
|
|
|
71 |
|
72 |
### Configuration
|
73 |
|
@@ -102,15 +103,16 @@ slices:
|
|
102 |
|
103 |
## Credits & Special Thanks
|
104 |
|
105 |
-
|
106 |
-
|
107 |
-
|
108 |
-
|
109 |
-
|
|
|
110 |
|
111 |
### Support
|
112 |
|
113 |
-
|
114 |
|
115 |
## Disclaimer
|
116 |
|
|
|
16 |
|
17 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
|
18 |
|
19 |
+
- EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
|
20 |
+
- GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/) | [IQ3_XXS](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
|
21 |
|
22 |
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
|
23 |
|
|
|
47 |
|
48 |
## Model Details
|
49 |
|
50 |
+
- Max Context: 32764 tokens (kept the weird number from the original/base model)
|
51 |
+
- Layers: 140
|
52 |
|
53 |
### Prompt template: Mistral
|
54 |
|
|
|
67 |
### Models Merged
|
68 |
|
69 |
The following models were included in the merge:
|
70 |
+
|
71 |
+
- [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
|
72 |
|
73 |
### Configuration
|
74 |
|
|
|
103 |
|
104 |
## Credits & Special Thanks
|
105 |
|
106 |
+
- original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
|
107 |
+
- ⭐⭐⭐ **[Use their newer, better, official models here!](https://console.mistral.ai/)** ⭐⭐⭐
|
108 |
+
- leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
|
109 |
+
- f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
|
110 |
+
- mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
|
111 |
+
- mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
|
112 |
|
113 |
### Support
|
114 |
|
115 |
+
- [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
|
116 |
|
117 |
## Disclaimer
|
118 |
|